profile-pic

Akshay Kumar

Cloud Solutioning | Machine Learning | Prompt Engineering

Data Science Professional with experience in FMCG, E-commerce and Insurance Domain

Experience in creating E2E pipeline for Data Science based Analytical solutions and Insights starting from understanding BRD and client data to building data pipeline for modelling framework to building model application and publishing output

Worked on building Campaign Attribution, Forecasting and Market Mix models

ML skill ranging from Stastical models to Regression , Decision tree based algo, Clustering and Classification algorithms

Have worked on OCI resources auto-provisioning IAC scripts using Terraform

Filled a patent with Oracle on ML Models Performance Tracking

PGDP in AIML from BITS Pilani

  • Role

    Principal MLops Engineer

  • Years of Experience

    7.67 years

Skillsets

  • HTML
  • HDFS
  • k8s
  • MongoDB
  • Oracle
  • SAS
  • SQL Server
  • Agentic-AI
  • GCP
  • GitLab CI/CD
  • Git
  • Kubeflow
  • LLM
  • MCP
  • Sagemaker
  • Scikit-learn
  • Vertex-ai
  • Grafana
  • rag
  • Ray
  • Time Series
  • SQL - 9 Years
  • PySpark - 2 Years
  • Deep Learning - 1 Years
  • Forecasting - 5 Years
  • Unix Shell Scripting - 5 Years
  • Prompt Engineering - 1 Years
  • Classification - 4 Years
  • clustering - 4 Years
  • Regression - 9 Years
  • AWS - 6.0 Years
  • OCI - 4 Years
  • HiveQL
  • Python
  • PyTorch
  • Tableau
  • Airflow
  • Ansible
  • Docker
  • Elasticsearch

Professional Summary

7.67Years
  • Feb, 2025 - Present1 yr 3 months

    Principal MLops Engineer

    Factspan
  • Dec, 2020 - Jan, 20254 yr 1 month

    Principal MLops Engineer

    Oracle
  • Jun, 2018 - Nov, 20202 yr 5 months

    Principal MLops Engineer

    TEG Analytics

Applications & Tools Known

  • icon-tool

    SAS

  • icon-tool

    PyTorch

  • icon-tool

    Tableau

  • icon-tool

    Docker

  • icon-tool

    k8s

  • icon-tool

    Airflow

  • icon-tool

    Terraform

  • icon-tool

    Terragrunt

  • icon-tool

    Kafka

  • icon-tool

    Ansible

Work History

7.67Years

Principal MLops Engineer

Factspan
Feb, 2025 - Present1 yr 3 months
    Models migration to Vertex-AI from Datarobot AutoML. Migrated models from Datarobot to Vertex-AI ensuring scalability, retaining logging & monitoring via Vertex-AI (both v1 & v2 version of monitoring). Optimized Training infrastructure with KubeRay. Implemented KubeRay on GKE to parallelize 48 category-specific training loops for a product propensity model previously orchestrated in Kubeflow. Alligned training to leverage distributed Ray clusters for resource elasticity, enabling efficient hardware utilization across workloads. Ruduced training runtime and infrastructure costs by migrating from Vertex AI custom jobs to KubeRay on GKE, cutting compute spend by ~40%. Automated Patient Record Synthesis via Vertex AI Pipelines. Implemented a cost effective and scalable batch prediction system using Vertex AI Pipeline to process high volumes of clinical documentation and generate automated patient summaries for Nursing staff. Implemented a structured document-to-folder workflow that automated the ingestion of diagnostics and prescriptions, utilizing Python and GCP services to maintain organized, HIPAA-compliant data hierarchies.

Principal MLops Engineer

Oracle
Dec, 2020 - Jan, 20254 yr 1 month
    Automated retraining & Monitoring for object detection model in Autonomous Driving System. Designed retraining pipelines that bench new data against existing models, storing Model-IDs and evaluation metrics in BigQuery for tracking. Implemented automated alerts when retrained models outperform production, enabling data-scientists to review and approve deployment to production. Built continuous monitoring for data drift and skew by instrumenting inference pipelines with structured logging in cloud logging, and then visualizing derived metrics in Grafana Dashboard. ML Model Training/Scoring Data Pipeline. Designed and implemented a robust data pipeline for a machine learning project, ensuring efficient data flow from the source database to the data-warehouse. Established a Kafka-based synchronization mechanism to maintain real-time consistency between the source database and the data warehouse. Multi-Channel Pipeline for Dashboard refresh and Disaster Recovery. Created schema and data model to push data from spark to sql server using pyspark jdbc connector. Established synchronous schema on sql-server to ingest incoming streams of data from pyspark ETL scripts. Stored procedure to have backup of DB on remote server. Model High Performance Testing (HPT). Created an HPT framework to keep checking model health for Data Drift and Model Drift. Patent accepted for this novel HPT project.

Principal MLops Engineer

TEG Analytics
Jun, 2018 - Nov, 20202 yr 5 months
    ETL migration from SAS to Spark. Migrated a sas based ETL for process entire US healthcare benefits data from SAS based ETL to pyspark. Built an ROI simulator to use Market Mix Model output and elasticities to predict impact on sales by changing & promotional spends. Tableau based Healthcare dashboards for Healthcare providers across US. Automated dashboard refresh through Tableau server command-line-utility for any pre-defined cadence.

Achievements

  • Patent accepted for novel project
  • Genie of the year for customer experience
  • Patent accepted for a novel project

Major Projects

4Projects

Models migration to Vertex-AI from Datarobot AutoML

    Migrated models from Datarobot to Vertex-AI ensuring scalability, retaining, logging & monitoring via Vertex-AI.

Automated Email Response (LLM based integration)

    Integrated an in-house LLM model APIs to generate response to MOS tickets from client, passing client specific model params to suit the response body.

Enabled versioning for all existing models

    Enabled versioning for all available models to have version specific instance of model run the training and then score for the same version.

Created a custom image to create a specific instance (EC2 equivalent of AWS)

    Built a new image with oraclelinux-9 slim as base image with only the required patches and tools to get CSSAP approval (OCI approval to use custom image). Used this image as the main orchestrator for running all the API requests.

Education

  • PGDP in AIML

    Birla Institute of Technology & Science, Pilani (2021)
  • B.Tech in Materials Science

    National Institute of Technology, Jamshedpur (2015)
  • AISSCE (Senior Secondary)

    Jharkhand Academic Council (2010)
  • AISSE (High School)

    Jharkhand Academic Council (2008)

Certifications

  • Pgdp in aiml