We provide IT Staff Augmentation Services!

Project Lead, Technical Lead, Etl Consultant Resume

2.00/5 (Submit Your Rating)

SUMMARY:

  • Technology oriented professional Over 13 Years of experience in Hadoop/ Big Data/ ETL/ DWH/ BI domains across various projects. Worked in different roles like Project Lead, Technical Lead, ETL Consultant and On - site Coordinator in various business verticals.
  • Around 4 years of experience with full project life cycle using Big Data, Hadoop Ecosystem Analytics in E-commerce, Cloud Services, Data center services and Marketing analytics.
  • Hands on experience in Big data stack Hadoop, Spark, Streaming, HDFS, HIVE, HBase, Sqoop, Oozie, Kafka, MapReduce.
  • Significant business domain knowledge in Cloud Services, Data Centers, Retail, and Insurance.
  • Experienced in building data pipelines using Kafka, Spark Streaming.
  • Experienced in design and implementation of scalable data lakes using Hadoop, Hive and HBase.
  • Experienced in both batch and real-time streaming.
  • Good knowledge of NoSQL databases HBase/Cassandra and CAP theorem.
  • Experienced in ingesting structured data from RDBMS to HDFS/Hive using Sqoop.
  • Expert in Spark RDD and Data Frame API.
  • Working with Hive external, internal tables, Partitioning, bucketing, UDF’s and complex query patterns.
  • Expert in dealing with performance optimization in Hive, Spark, and MR.
  • Spark, effective and efficient joins, transformations during ingestion process.
  • Experience in using different file formats like Parquet, JSON, AVRO, ORC and CSV.
  • Python script to call Rest API’s to bring in customer feedback JSON files, created ETL Pipeline to store in Hive database.
  • REST API’s calls - Salesforce objects, Cassandra, External vendors.
  • ELT - Building the Sisense Elastic cube and dashboards, Tableau data visualizations.
  • Ad-hoc data analytics using Spark SQL via Jupiter, Hive queries via Hue.
  • UNIX shell scripting - Arrays, wrapper scripts - automate file loads processes, ETL Integrations.
  • Source file validation Framework - Validating the business agreed formats and user notifications.
  • Experience in using Git, Putty, Maven, and Jenkins.

PROFESSIONAL EXPERIENCE:

Confidential

Project Lead, Technical Lead, ETL Consultant

Responsibilities:

  • Experienced in design and implementation of scalable data lakes using Hadoop, Hive and HBase.
  • Experienced in both batch and real-time streaming.
  • Good knowledge of NoSQL databases HBase/Cassandra and CAP theorem.
  • Experienced in ingesting structured data from RDBMS to HDFS/Hive using Sqoop.
  • Expert in Spark RDD and Data Frame API.
  • Working with Hive external, internal tables, Partitioning, bucketing, UDF’s and complex query patterns.
  • Expert in dealing with performance optimization in Hive, Spark, and MR.
  • Spark, effective and efficient joins, transformations during ingestion process.
  • Experience in using different file formats like Parquet, JSON, AVRO, ORC and CSV.
  • Python script to call Rest API’s to bring in customer feedback JSON files, created ETL Pipeline to store in Hive database.
  • REST API’s calls - Salesforce objects, Cassandra, External vendors.
  • ELT - Building the Sisense Elastic cube and dashboards, Tableau data visualizations.
  • Ad-hoc data analytics using Spark SQL via Jupiter, Hive queries via Hue.
  • UNIX shell scripting - Arrays, wrapper scripts - automate file loads processes, ETL Integrations.
  • Source file validation Framework - Validating the business agreed formats and user notifications.
  • Experience in using Git, Putty, Maven, and Jenkins.

We'd love your feedback!