Strong experience in Hadoop and Spark ecosystems including Spark Core, Pyspark, SparkSQL, HDFS, Map-Reduce, Hive, Hbase, Sqoop Having good knowledge on programming language as Python and Scala. Proficient in performing SQL queries. Easily adopt, explore, learn and understand newer business domains and technology based on the requirements. Develop best practices for developing and deploying Hadoop applications and assist the team to manage compliance to the standards. Experienced with different file formats like Parquet, ORC, Sequence, CSV, JSON, Text files. Experienced Big Data/Hadoop and spark Developer with strong background with file distribution system in big data arena Execute change management activities supporting production deployment to Developers. Design, plan, and develop programs to perform automated extract, transform and load data between data sources when working with large data sets. Performed Hive operations on large datasets with proficiency in writing HiveQL queries using transactional and performance efficient concepts: Partitioning, Bucketing, efficient and effective Join operations. Gained exposure on various AWS services like Glue, S3, Lambda, RDS, Athena. Having sound knowledge on Data warehousing concepts.
View more