profile-pic

Kartiki Prashant Dange

I mostly worked on Migration projects in telecom and e-commerce domain.

I am working with e-commerce client Macy's from US in Datametica. Here we migrated the Teradata objects, processes, lineages from Teradata to Google Cloud Platform with the help of Cloud storage, Google Composer(Airflow), Bigquery, Stackdriver, Dataproc, Spark-SQL.

While working on Bigdata technologies in Cognizant for teelecom client KPN, we migrated many RDBMS, file sources to Hadoop with the use of Hive, Sqoop, Pig, Oozie, Hcatalog

  • Role

    Data warehouse & Google Cloud PlatformDeveloper

  • Years of Experience

    9 years

Skillsets

  • Google Cloud Platform
  • Java - 1 Years
  • Core Java - 2 Years
  • Delta Lake
  • Unix
  • SQL - 8 Years
  • Spark - 5 Years
  • Python - 6 Years
  • PySpark
  • Jira
  • Hive
  • Hadoop - 8 Years
  • Agile
  • Git
  • GCP - 6 Years
  • Databricks
  • Data warehouse
  • Data Pipeline
  • Data Migration - 5 Years
  • Data lake
  • CI/CD
  • BigQuery - 6 Years
  • AWS - 2 Years
  • Airflow

Professional Summary

9Years
  • May, 2020 - Present5 yr 6 months

    Senior Data Engineer

    Uplight (prev. UST Global)
  • Nov, 2018 - Mar, 20201 yr 4 months

    Google Cloud Platform Developer

    Datametica Solution pvt ltd
  • Apr, 2015 - Oct, 20183 yr 6 months

    Hadoop Developer with Dataproc

    Cognizant Technology Solutions; pvt ltd

Applications & Tools Known

  • icon-tool

    Apache Hadoop

  • icon-tool

    Apache Spark

  • icon-tool

    Dataproc

  • icon-tool

    Google BigQuery

  • icon-tool

    Cloud SQL

  • icon-tool

    Cloud Function

  • icon-tool

    Data Studio

  • icon-tool

    Spark SQL

  • icon-tool

    Python

  • icon-tool

    SQL

  • icon-tool

    Unix

  • icon-tool

    Jira

  • icon-tool

    Oracle

  • icon-tool

    Teradata

  • icon-tool

    Databricks

  • icon-tool

    Control-M

  • icon-tool

    Kubernetes

  • icon-tool

    Container

  • icon-tool

    Scala

  • icon-tool

    Hive

  • icon-tool

    Sqoop

  • icon-tool

    Eclipse

  • icon-tool

    Spark Streaming

Work History

9Years

Senior Data Engineer

Uplight (prev. UST Global)
May, 2020 - Present5 yr 6 months
    Datalake creation and migration from legacy system to a standard platform for catering multi-tenant architecture and serving different customers in the Energy Domain. Designed ELT solutions and modeled Data structures. Developed Data Pipeline using Google Cloud Platform utilities (CDF, BigQuery, Airflow, and CI/CD) for continuous deployment. Sent customer-specific data to end users and created shims for inter-cloud data migration from GCP to AWS, visualization, and dashboard creation in Looker for tenant analysis.

Google Cloud Platform Developer

Datametica Solution pvt ltd
Nov, 2018 - Mar, 20201 yr 4 months
    Migrated over 55,000 Teradata objects, tables, views, and processes consuming data from Teradata into GCP BigQuery. Ensured process automation using Composer (Apache Airflow) and BigQuery. Conducted requirement analysis, verified existing schedules and dependencies, converted Teradata scripts into BigQuery compatible queries, created DAGs in Google Composer, and performed unit testing and bug-fixing.

Hadoop Developer with Dataproc

Cognizant Technology Solutions; pvt ltd
Apr, 2015 - Oct, 20183 yr 6 months
    Built a reusable framework to ingest data from multiple sources into Hadoop for KPN, Netherland. Conducted data analysis using Hive, Pig, Spark Streaming, SparkSQL, and various Hadoop tools like Sqoop and Flume. Developed analytics engine for batch and real-time data analysis.

Achievements

  • Learned datalake and Deltalake concept in depth
  • Explored Databricks and end-to-end development cycle
  • Pioneer for the Composer and did multiple knowledge transfer sessions to upskill the team

Major Projects

3Projects

GCP Platform creation on datalake for renewable energy client UpLight

May, 2020 - Present5 yr 6 months
    Datalake creation and migration from legacy system to a standard platform for catering multi-tenant architecture and serving different customers in the Energy Domain. Designed ELT solutions and modeled Data structures. Developed Data Pipeline using Google Cloud Platform utilities (CDF, BigQuery, Airflow, and CI/CD) for continuous deployment. Sent customer-specific data to end users and created shims for inter-cloud data migration from GCP to AWS, visualization, and dashboard creation in Looker for tenant analysis.

Migration of over 55,000 Teradata objects into GCP BigQuery

Nov, 2018 - Mar, 20201 yr 4 months
    Migrated over 55,000 Teradata objects, tables, views, and processes consuming data from Teradata into GCP BigQuery. Ensured process automation using Composer (Apache Airflow) and BigQuery.

Creating Data Lake by migrating existing databases for KPN, Netherland

Apr, 2015 - Oct, 20183 yr 6 months
    Built a reusable framework to ingest data from multiple sources into Hadoop for KPN, Netherland. Conducted data analysis using Hive, Pig, Spark Streaming, SparkSQL, and various Hadoop tools like Sqoop and Flume.

Education

  • BE in Computer Science Engineering

    SSPACE Wardha, Wardha (2014)
  • Higher Secondary Certificate

    Maharashtra Board (2010)
  • Secondary School Certificate

    Maharashtra Board (2008)