profile-pic
Vetted Talent

Omjee Mishra

Vetted Talent
Strong experience in Hadoop and Spark ecosystems including Spark Core, Pyspark, SparkSQL, HDFS, Map-Reduce, Hive, Hbase, Sqoop Having good knowledge on programming language as Python and Scala. Proficient in performing SQL queries. Easily adopt, explore, learn and understand newer business domains and technology based on the requirements. Develop best practices for developing and deploying Hadoop applications and assist the team to manage compliance to the standards. Experienced with different file formats like Parquet, ORC, Sequence, CSV, JSON, Text files. Experienced Big Data/Hadoop and spark Developer with strong background with file distribution system in big data arena Execute change management activities supporting production deployment to Developers. Design, plan, and develop programs to perform automated extract, transform and load data between data sources when working with large data sets. Performed Hive operations on large datasets with proficiency in writing HiveQL queries using transactional and performance efficient concepts: Partitioning, Bucketing, efficient and effective Join operations. Gained exposure on various AWS services like Glue, S3, Lambda, RDS, Athena. Having sound knowledge on Data warehousing concepts.
  • Role

    AI Engineer / Software Developer

  • Years of Experience

    8.6 years

Skillsets

  • Google Cloud AI
  • UX/UI Design
  • TypeScript
  • Supabase
  • Streamlit
  • react
  • PostgreSQL
  • Pinecone
  • MongoDB
  • MLOps
  • LangChain
  • JavaScript
  • Git
  • FastAPI
  • Event-Driven Architectures
  • Docker
  • CI/CD
  • Aws ai tools
  • API development
  • JavaScript - 8 Years
  • JSON
  • Python - 8 Years

Vetted For

10Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    Python Developer (AI/ML & Cloud Services) - RemoteAI Screening
  • 47%
    icon-arrow-down
  • Skills assessed :GCP/Azure, Micro services, Django /Flask, Neo4j, Restful APIs, AWS, Docker, Kubernetes, machine_learning, Python
  • Score: 42/90

Professional Summary

8.6Years
  • Jul, 2025 - Present 2 months

    AI Engineer / Software Developer

    MTI
  • Jul, 2025 - Jul, 2025

    AI Solutions Developer

    Tech Mahindra

Applications & Tools Known

  • icon-tool

    CSV

  • icon-tool

    JSON

  • icon-tool

    AWS Glue

  • icon-tool

    S3

  • icon-tool

    Lambda

  • icon-tool

    RDS

  • icon-tool

    PostgreSQL

  • icon-tool

    Python

  • icon-tool

    Databricks

  • icon-tool

    Hadoop

  • icon-tool

    HDFS

  • icon-tool

    Hive

  • icon-tool

    SQL

  • icon-tool

    C#

  • icon-tool

    Oracle 11g

Work History

8.6Years

AI Engineer / Software Developer

MTI
Jul, 2025 - Present 2 months
    Designed and developed modular conversational AI systems leveraging private knowledge bases to deliver accurate, context-aware responses, enhancing customer interactions by 30%. Created operational AI tools and automations that streamlined business processes across web, mobile, and messaging platforms, resulting in a 25% reduction in processing time. Developed and maintained API integrations using FastAPI, ensuring seamless communication between various services and enhancing overall system performance. Implemented secure and scalable solutions that protected sensitive data while maintaining high performance, achieving compliance with industry standards. Collaborated with cross-functional teams to create user-configurable tools adaptable across diverse markets, increasing customer satisfaction scores by 20%.

AI Solutions Developer

Tech Mahindra
Jul, 2025 - Jul, 2025
    Led the development of a self-service platform utilizing conversational AI and natural language processing, enabling clients to manage their own AI solutions effectively. Developed modular components for integration across various channels, ensuring extensibility and ease of use for end users. Participated in the design of multi-tenant architectures and webhook implementations, enhancing the platform's capabilities for diverse client needs. Contributed to the establishment of CI/CD pipelines, improving deployment efficiency and reducing time-to-market for new features.

Major Projects

2Projects

Conversational AI Chatbot

Jan, 1970 - Jan, 1970
    Developed a chatbot using LangChain and FastAPI that integrated with multiple communication platforms, improving customer engagement by 40%.

Operational AI Automation Tool

Jan, 1970 - Jan, 1970
    Created an automation tool that streamlined data processing workflows, resulting in a 50% increase in operational efficiency for a marketing firm.

Education

  • Bachelor of Engineering

    VTU

Certifications

  • Big Data Developer Certification from Simplilearn

  • Big data developer certification from simplilearn

  • Data science with python from simplilearn

  • Machine learning advanced certification training from simplilearn

  • Tableau desktop 10 certification from simplilearn

AI-interview Questions & Answers

Yes, so, so I'm

in data engineering, where I worked on data collection, data storage, data transformation, okay. And actually, I rely on data to make informed decision. As a data engineer, I always play a vital role in providing the necessary infrastructure and tools. And as a business grows, the data needs grow too. So as a data engineer, we build a scalable system to handle increasing data volume. And I'm good into Python SQL and Scala. And I do have knowledge of relational database and NoSQL. And I do have experience with cloud platform like AWS, Azure, and GCP.

So, uh, in terms of, uh, you know, uh, implementing CI CD pipeline, uh, for deploying, uh, uh, Python application. Okay. Uh, Python application on AWS. Oh, so the service that. So basically on, in terms of, you know, uh, we can use service AWS code, uh, build and test AWS code build, uh, AWS code deploy. Okay. Infrastructure as code that is optional. AWS cloud watch. Okay. AWS code pipeline.

So, basically, if I'll be talking about how we use Docker container to manage, you know, dependency and streamline deployment for Django based machine learning service on AWS. So, basically, the entire application involvement, including all dependency like Python library system. So, outline the various AWS services that can be used for deploying a Dockerite service like Amazon Elastic Container Services, Amazon Elastic Kubernetes Services, Amazon Elastic. So, basically, we can say like, streamline the deployment process by providing consistent. Okay. So, some of the we can consider including a simple diagram. Okay.

So basically, the benefit we can say in terms of Neo4j integration in machine learning workflow, it can be graph-based feature engineering that excel at representing complex relationship between entities. This can be leveraged to engineer powerful features for machine learning model like network embedding, path-based features, community detection. It can also improve model performance as well as scalability and performance. So the approach for integration are data loading, feature engineering, model training, integrating with model training, model evaluation, and we can use Python code with Neo4j.

so basically we can break down how we can implement serverless microservice in AWS lambda using python so basically first we can say microservice is a small independent service serverless architecture where we don't manage server directly okay so first we need to create an AWS lambda function where we can choose runtime write function code configure handler then we can create an API gateway then we can configure API gateway then we can deploy and test so these are the process

So, there can be some of the potential issue like, you know, lack of entirety context in query, hard-coded entity URL, no error handling, potential framework bottleneck, ok. So, we can recommend like refining the query, parse entity URL argument, we can add error handling, we can consider indexing and we can optimize SparkQL queries.

so first we need to do logging then matrix then tracing and in terms of reliability it can be error handling and retry, idempotency, fault tolerance, monitoring and alerting and some of the AWS specific considerations like AWS step function, lambda, batch.

so basically if I will be talking about like you know some of the service component like API gateway lambda function model serving and second one is high availability and fault tolerance in terms of high availability and fault tolerance it can be API gateway lambda function model serving okay data storage and in terms of python implementation it can be request handler inference handler monitoring and logging security testing and deployment

So, yeah, in my last project, actually, you know, I worked on the Django application that hosted a set of RESTful API for a large e-commerce platform where we encountered a performance issue as the application failed to handle increasing traffic. To address this, I implemented several optimization strategies, like database optimization, API endpoint optimization, server-side optimization, cloud-specific optimization. So by implementing these optimizations, I was able to significantly improve the performance and scalability of the Django application.

so basically if I will be talking about in my previous project as I told you I worked on larger scale e-commerce platform where we needed to process a high volume of order and user interaction to achieve this I implemented a task process using Salary a popularly attributed task queue in python so some of the key aspect of implementation are task definition, task queue integration, message broker, task scheduling, error handling and retry.