profile-pic

Supratik Chakraborty

Enthusiastic hard working team-builder with 10 years of experience in delivering high quality projects in IT. Top-performer and motivator who has the added advantage of technology as leverage to achieve timely results along with a proven track record of sticking to objectives and making the most out of available resources.
  • Role

    Senior Data Engineer

  • Years of Experience

    11.10 years

Skillsets

  • ETL
  • AWS IAM
  • AWS Lambda
  • AWS S3
  • Azure
  • Azure CLI
  • Azure Data Factory
  • Azure DevOps
  • BigQuery
  • Control-M
  • cosmos Db
  • Databricks
  • Delta Lake
  • Delta Live Tables
  • Docker
  • Erwin data modeler
  • AWS Glue
  • GCS
  • Github
  • GitHub Actions
  • Google Cloud Storage
  • HP ALM
  • Jira
  • Logic Apps
  • Microsoft Fabric
  • MongoDB
  • Phoenix
  • RPA
  • Serverless sql
  • ServiceNow
  • SQL
  • HBase
  • Big Data
  • Business Objects
  • Data Architecture
  • Hive
  • Informatica
  • Kafka
  • Pega
  • Tableau
  • Teradata
  • Unix
  • AWS
  • Azure Synapse Analytics
  • Bitbucket
  • Deep Learning
  • Hadoop
  • Python - 6 Years
  • Power BI
  • Scala
  • SVN
  • Synapse analytics studio
  • Dmxpress-h
  • Adls2
  • Openspan
  • PySpark
  • Spark
  • Airflow
  • Amazon Athena
  • Amazon Kinesis
  • Amazon Redshift
  • AutoSys

Professional Summary

11.10Years
  • Aug, 2024 - Present1 yr 6 months

    Senior Data Engineer

    Glencore Information Services
  • Mar, 2014 - Aug, 202410 yr 5 months

    Data Engineer/Senior Data Engineer/Project Lead

    TCS

Applications & Tools Known

  • icon-tool

    Python

  • icon-tool

    AWS (Amazon Web Services)

  • icon-tool

    Azure

  • icon-tool

    Unix

  • icon-tool

    Bitbucket

  • icon-tool

    Apache HBase

  • icon-tool

    Microsoft Power BI

  • icon-tool

    Scala

  • icon-tool

    Apache Spark

  • icon-tool

    Azure DevOps Server

  • icon-tool

    Apache Airflow

  • icon-tool

    Apache Kafka

  • icon-tool

    MongoDB

  • icon-tool

    Tableau CRM

  • icon-tool

    Teradata

Work History

11.10Years

Senior Data Engineer

Glencore Information Services
Aug, 2024 - Present1 yr 6 months
    Architected multi-cloud data platforms (Azure, AWS, GCP), engineering ingestion and transformations with Synapse/ADF, Databricks, AWS Glue, and BigQuery. Established medallion architecture with Delta Lake and Unity Catalog on ADLS2, elevating governance and reusable data products. Standardized orchestration via Apache Airflow, decoupling workloads from Synapse; automated Spark job registration using Azure CLI wrappers. Built end-to-end CI/CD pipelines with GitHub Actions and Azure DevOps; containerized local development using Docker and shipped Python libraries through JFrog (Poetry/PyPI), reducing release cycles by more than 50%. Enhanced resilience with fault-tolerant designs, Python Function Apps, and Logic Apps for alerting, improving SLA adherence to more than 50%. Advanced BI by modernizing Power BI and Tableau assets; enabled self-service insights, cutting time-to-decision by more than 50%. Authored HLD/LLD using data modeling best practices; templatized Synapse artifacts with configuration-driven parameters to boost maintainability by 100%. Integrated SharePoint via SMB to ADLS2 and standardized Delta tables with Serverless SQL distribution; partnered with CloudOps to institutionalize Airflow. Delivered proofs of concept with Microsoft Fabric and integrated GCS into the ingestion framework; designed an AWS stack (S3, Glue, Lambda, Kinesis, Athena/Redshift, IAM) to scale to 2 million records/day. Planned initiatives, refined estimates, and supported BAU, reducing incident MTTR by more than 50%.

Data Engineer/Senior Data Engineer/Project Lead

TCS
Mar, 2014 - Aug, 202410 yr 5 months
    Delivered end-to-end ETL and streaming workloads using PySpark/Scala across Hadoop and Azure; integrated ADF/Synapse and Databricks with ADLS2 and Unity Catalog. Migrated legacy feeds to ADLS2 + Synapse using medallion layering for analytics and visualization, hardening data reliability. Engineered enterprise data warehousing on Teradata via Informatica and Unix schedulers (Autosys/Control-M/Unix scripts) and DMXpress, reducing batch duration by 25%. Produced conceptual/logical data models with Erwin; enforced coding standards and reusable modules, lowering defect leakage by 10%. Mentored and led agile squads, improved planning and delivery predictability, increasing team velocity by 20%. Automated operational runbooks using Python/Shell and improved monitoring via ServiceNow, enhancing SLA compliance to 60%. Built RPA solutions with OpenSpan and delivered BI using Business Objects to remove manual steps and elevate reporting adoption.

Major Projects

2Projects

Teradata EDW to Azure Data Lake Migration

    Modernized an on-prem Teradata enterprise data warehouse to an ADLS2/Synapse medallion data lake to enable scalable analytics and BI. Orchestrated ingestion and transformation with ADF and Databricks, cutting manual workflows and improving job success rate. Implemented partitioning and Delta Lake optimizations, reducing query latency.

Real-time Streaming Analytics Platform

    Built a low-latency streaming platform using Kafka, Spark Structured Streaming, and HBase/Phoenix to power near real-time dashboards. Delivered exactly-once streaming pipelines with checkpointing and watermarking. Automated data quality checks and alerting via Airflow and Control-M.

Education

  • Bachelor of Technology (Electronics & Communications Engineering)

    Siliguri Institute of Technology (2013)
  • Senior Secondary (12th)

    Shishu Bihar Higher Secondary School (2009)
  • Secondary (10th)

    Shishu Bihar Higher Secondary School (2007)