profile-pic
Vetted Talent

Nivin Srinivas S

Vetted Talent
Senior Data Engineer with over four years of experience in building scalable data platforms and pipelines and maintaining optimal data infrastructure. Interested to solve challenging problems in a data-driven company.
  • Role

    Senior Data Engineer

  • Years of Experience

    6 years

Skillsets

  • Python - 5 Years
  • SQL - 5 Years
  • Spark
  • Data Warehousing
  • Airflow
  • BigQuery - 4 Years
  • Relational DB - 5 Years
  • dbt
  • Rest APIs
  • Terraform
  • data quality
  • Software Development

Vetted For

13Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    Data Engineer || (Remote)AI Screening
  • 80%
    icon-arrow-down
  • Skills assessed :Airflow, Data Governance, machine learning and data science, BigQuery, ETL processes, Hive, Relational DB, Snowflake, Hadoop, Java, Postgre SQL, Python, SQL
  • Score: 72/90

Professional Summary

6Years
  • Feb, 2023 - Present2 yr 6 months

    Senior Data Engineer

    CommerceIQ.AI
  • Mar, 2021 - Jan, 20231 yr 10 months

    Software Engineer - Data

    Unravel Data
  • Jul, 2019 - Mar, 20211 yr 8 months

    Data Engineer

    Tabsquare.AI
  • Sep, 2018 - Nov, 2018 2 months

    Software Developer Intern

    Yellow.AI

Applications & Tools Known

  • icon-tool

    BigQuery

  • icon-tool

    Apache Airflow

  • icon-tool

    Docker

  • icon-tool

    Terraform

  • icon-tool

    Github Actions

  • icon-tool

    Python

  • icon-tool

    dbt

  • icon-tool

    MySQL

  • icon-tool

    Google Cloud Platform

  • icon-tool

    BigQuery

  • icon-tool

    Git

  • icon-tool

    Unit testing

  • icon-tool

    REST API

  • icon-tool

    Apache Spark

  • icon-tool

    Terraform

  • icon-tool

    Flask

  • icon-tool

    SQLAlchemy

  • icon-tool

    Github Actions

  • icon-tool

    Slack

  • icon-tool

    Retool

  • icon-tool

    Github Actions

  • icon-tool

    AWS

Work History

6Years

Senior Data Engineer

CommerceIQ.AI
Feb, 2023 - Present2 yr 6 months
    Constructed efficient data pipelines using PySpark, DBT, BigQuery, and Apache Airflow, resulting in improved data processing speed and reliability. Developed Enterprise-Level Data Pipeline Template Repository Architected a comprehensive template repository integrating Apache Airflow with AWS Batch and Databricks for scalable data processing, enabling standardized deployment of both batch and data-intensive workflows Implemented robust monitoring system with PagerDuty and Slack integrations for real-time alerts, reducing incident response time and improving system reliability Established development best practices by incorporating pre-commit hooks with Black and Flake8, ensuring consistent code quality and maintainability across teams Containerized the solution using Docker and automated deployment to Amazon ECR, enabling seamless CI/CD pipeline integration and standardized environment management Engineered infrastructure-as-code using Terraform to provision and manage AWS resources (ECR, Batch, IAM roles, networking), ensuring consistent and reproducible infrastructure deployment across environments Spearheaded the development and productization of a data lake solution, providing critical data access that enhanced decision-making for customers.

Software Engineer - Data

Unravel Data
Mar, 2021 - Jan, 20231 yr 10 months
    Developed ETL pipelines using PySpark, Airflow, and BigQuery for seamless data processing. Engineered the backend infrastructure to efficiently manage and forecast AWS EMR and Databricks cloud costs, utilizing REST APIs, Flask, MySQL, and SQLAlchemy. Established a robust CI/CD system from scratch for the Data Science team, automating code deployment with Github Actions. Implemented a comprehensive healthcheck and reporting mechanism within the Unravel data platform, enhancing telemetry and system monitoring. Created REST APIs to deliver real-time critical budget notifications via Slack and email.

Data Engineer

Tabsquare.AI
Jul, 2019 - Mar, 20211 yr 8 months
    Responsible for designing and implementing a Data warehouse using BigQuery. Responsible for building ETL pipelines using Apache Spark. Implemented data pipelines for transforming and ingesting user events data in JSON to BigQuery for customer journey analytics. Set up Airflow and successfully migrated data pipelines and cron jobs to Airflow for efficiently scheduling and monitoring workflows. Created Dashboards and sales reports using Google Datastudio and Retool. Automated app deployments using Bitbucket pipelines CI / CD. Automated the process of onboarding new restaurant outlets using Python, Google APIs and Slack.

Software Developer Intern

Yellow.AI
Sep, 2018 - Nov, 2018 2 months
    Built, tested and deployed chatbots for multiple clients of Yellow Messenger. Gained experience in building web applications using Python.

Achievements

  • Led the development and productization of data lake, delivering raw data to customers.
  • Built pipeline to process Google Analytics events and create dashboard to track website metrics.
  • Developed and executed data validation strategies, implementing comprehensive data quality checks.
  • Created scorecard diagnostic tool to proactively identify and address anomalies within data-driven dashboards.
  • Engineered backend infrastructure to manage and forecast AWS EMR and Databricks cloud costs.
  • Established robust CI/CD system for the Data Science team, automating code deployment.

Major Projects

1Projects

Data Pipeline Template Repository

    Architected a comprehensive template repository integrating Apache Airflow with AWS Batch and Databricks for scalable data processing, enabling standardized deployment of both batch and data-intensive workflows.

Education

  • Bachelor of Engineering in Information Science

    Visvesvarayya Technological University (VTU) (2019)