Project Lead, Technical Lead, Etl Consultant Resume
2.00/5 (Submit Your Rating)
SUMMARY:
- Technology oriented professional Over 13 Years of experience in Hadoop/ Big Data/ ETL/ DWH/ BI domains across various projects. Worked in different roles like Project Lead, Technical Lead, ETL Consultant and On - site Coordinator in various business verticals.
- Around 4 years of experience with full project life cycle using Big Data, Hadoop Ecosystem Analytics in E-commerce, Cloud Services, Data center services and Marketing analytics.
- Hands on experience in Big data stack Hadoop, Spark, Streaming, HDFS, HIVE, HBase, Sqoop, Oozie, Kafka, MapReduce.
- Significant business domain knowledge in Cloud Services, Data Centers, Retail, and Insurance.
- Experienced in building data pipelines using Kafka, Spark Streaming.
- Experienced in design and implementation of scalable data lakes using Hadoop, Hive and HBase.
- Experienced in both batch and real-time streaming.
- Good knowledge of NoSQL databases HBase/Cassandra and CAP theorem.
- Experienced in ingesting structured data from RDBMS to HDFS/Hive using Sqoop.
- Expert in Spark RDD and Data Frame API.
- Working with Hive external, internal tables, Partitioning, bucketing, UDF’s and complex query patterns.
- Expert in dealing with performance optimization in Hive, Spark, and MR.
- Spark, effective and efficient joins, transformations during ingestion process.
- Experience in using different file formats like Parquet, JSON, AVRO, ORC and CSV.
- Python script to call Rest API’s to bring in customer feedback JSON files, created ETL Pipeline to store in Hive database.
- REST API’s calls - Salesforce objects, Cassandra, External vendors.
- ELT - Building the Sisense Elastic cube and dashboards, Tableau data visualizations.
- Ad-hoc data analytics using Spark SQL via Jupiter, Hive queries via Hue.
- UNIX shell scripting - Arrays, wrapper scripts - automate file loads processes, ETL Integrations.
- Source file validation Framework - Validating the business agreed formats and user notifications.
- Experience in using Git, Putty, Maven, and Jenkins.
PROFESSIONAL EXPERIENCE:
Confidential
Project Lead, Technical Lead, ETL ConsultantResponsibilities:
- Experienced in design and implementation of scalable data lakes using Hadoop, Hive and HBase.
- Experienced in both batch and real-time streaming.
- Good knowledge of NoSQL databases HBase/Cassandra and CAP theorem.
- Experienced in ingesting structured data from RDBMS to HDFS/Hive using Sqoop.
- Expert in Spark RDD and Data Frame API.
- Working with Hive external, internal tables, Partitioning, bucketing, UDF’s and complex query patterns.
- Expert in dealing with performance optimization in Hive, Spark, and MR.
- Spark, effective and efficient joins, transformations during ingestion process.
- Experience in using different file formats like Parquet, JSON, AVRO, ORC and CSV.
- Python script to call Rest API’s to bring in customer feedback JSON files, created ETL Pipeline to store in Hive database.
- REST API’s calls - Salesforce objects, Cassandra, External vendors.
- ELT - Building the Sisense Elastic cube and dashboards, Tableau data visualizations.
- Ad-hoc data analytics using Spark SQL via Jupiter, Hive queries via Hue.
- UNIX shell scripting - Arrays, wrapper scripts - automate file loads processes, ETL Integrations.
- Source file validation Framework - Validating the business agreed formats and user notifications.
- Experience in using Git, Putty, Maven, and Jenkins.