We provide IT Staff Augmentation Services!

Sr. Hadoop Consultant Resume

0/5 (Submit Your Rating)

SUMMARY

  • 8 years of experience in Information Technology including 5+ years in Data Warehousing/Data Mart development using ETL/Informatica Power Center and over 3 years of experience in Hadoop working environment.
  • Having exposure in various fields such as Oil&Gas, Retail, Insurance and Finance.
  • Good Knowledge in overall SDLC comprising requirements gathering, modelling, development, integration, testing, debugging, deployment, documentation and maintenance support.
  • Worked in Software methodologies like Agile and Waterfall models.
  • Good Understanding in various Hadoop Distributions like Cloudera and MapR to implement and make use of them.
  • Experienced in developing Applications using Hadoop ecosystem like MapReduce, Hive, Pig, Sqoop, Flume and HBase.
  • Excellent understanding / knowledge ofHadooparchitecture and various components ofsuch as HDFS, JobTracker, TaskTracker, NameNode, DataNode, MapReduce & YARN.
  • Proficient in writing Ad - hoc queries for moving data from HDFS to HIVE and analyzing teh data using HIVE QL.
  • Experience in importing and exporting data using Sqoop from Relational Database Systems to HDFS and vice-versa.
  • Developed Kafka producer and consumers, HBase clients, Spark and Hadoop MapReduce jobs along wif components on HDFS, Hive.
  • Collection and Analysis on large set of log data is done using Custom built Input Adapters and Sqoop.
  • Knowledge in converting Hive or SQL queries into Spark transformations using Python and Scala.
  • Worked on teh core and Spark SQL modules of Spark extensively.
  • Extending HIVE and PIG core functionality by using custom User Defined Functions(UDF), User Defined Tables-Generating Functions(UDTF) and User Defined Aggregating Functions(UDAF).
  • Experience in creating sqoop jobs wif incremental load to populate Hive External tables.
  • Good understanding of Partitioning, Bucketing, Join optimizations and query optimizations in Hive.
  • Experience in working wif flume to load teh log data from multiple sources directly into HDFS
  • Experience in teh successful implementation of ETL solution between an OLTP and OLAP database in support of Decision Support Systems wif expertise in all phases of SDLC.
  • Strong in Data warehousing Concepts, Star Schema and Snowflakes Schema methodologies, understanding business process/requirements.
  • Expertise in doing Unit Testing, Integration Testing, System Testing and Data Validation for Developed Informatica Mappings.
  • Experience in Software Development on UNIX, Red Hat Linux, Ubuntu, Sun Solaris.
  • A team player and self-motivator possessing excellent analytical, communication, problem solving, decision-making and Organizational skills

TECHNICAL SKILLS

Programming Languages: C, C++, Java, Python,Scala

Java EE Technologies: JSP, Servlets, EJB, JDBC

Hadoop Ecosystem: HDFS, MapReduce, Flume, Sqoop, Hive, Pig, Spark, and Oozie

ETL Tools: InformaticaPower Center 9.1/9.5 (Designer, Workflow Manager, Workflow Monitor, Repository manager andInformaticaServer), SSIS and SSRS

Databases: Oracle, Mysql, DB2, SQL 2000, Sybase

IDE Tools: Eclipse, NetBeans IDE, RAD, BEA WebLogic studio

Operating Systems: UNIX, Red Hat Linux, Ubuntu, MS-DOS, Sun Solaris, Windows

Testing: Junit

Design/Build Tools: Ant and Maven

Tools: & Utilities: CVS, SVN

We'd love your feedback!