Sr. Hadoop Consultant Resume
0/5 (Submit Your Rating)
SUMMARY
- 8 years of experience in Information Technology including 5+ years in Data Warehousing/Data Mart development using ETL/Informatica Power Center and over 3 years of experience in Hadoop working environment.
- Having exposure in various fields such as Oil&Gas, Retail, Insurance and Finance.
- Good Knowledge in overall SDLC comprising requirements gathering, modelling, development, integration, testing, debugging, deployment, documentation and maintenance support.
- Worked in Software methodologies like Agile and Waterfall models.
- Good Understanding in various Hadoop Distributions like Cloudera and MapR to implement and make use of them.
- Experienced in developing Applications using Hadoop ecosystem like MapReduce, Hive, Pig, Sqoop, Flume and HBase.
- Excellent understanding / knowledge ofHadooparchitecture and various components ofsuch as HDFS, JobTracker, TaskTracker, NameNode, DataNode, MapReduce & YARN.
- Proficient in writing Ad - hoc queries for moving data from HDFS to HIVE and analyzing teh data using HIVE QL.
- Experience in importing and exporting data using Sqoop from Relational Database Systems to HDFS and vice-versa.
- Developed Kafka producer and consumers, HBase clients, Spark and Hadoop MapReduce jobs along wif components on HDFS, Hive.
- Collection and Analysis on large set of log data is done using Custom built Input Adapters and Sqoop.
- Knowledge in converting Hive or SQL queries into Spark transformations using Python and Scala.
- Worked on teh core and Spark SQL modules of Spark extensively.
- Extending HIVE and PIG core functionality by using custom User Defined Functions(UDF), User Defined Tables-Generating Functions(UDTF) and User Defined Aggregating Functions(UDAF).
- Experience in creating sqoop jobs wif incremental load to populate Hive External tables.
- Good understanding of Partitioning, Bucketing, Join optimizations and query optimizations in Hive.
- Experience in working wif flume to load teh log data from multiple sources directly into HDFS
- Experience in teh successful implementation of ETL solution between an OLTP and OLAP database in support of Decision Support Systems wif expertise in all phases of SDLC.
- Strong in Data warehousing Concepts, Star Schema and Snowflakes Schema methodologies, understanding business process/requirements.
- Expertise in doing Unit Testing, Integration Testing, System Testing and Data Validation for Developed Informatica Mappings.
- Experience in Software Development on UNIX, Red Hat Linux, Ubuntu, Sun Solaris.
- A team player and self-motivator possessing excellent analytical, communication, problem solving, decision-making and Organizational skills
TECHNICAL SKILLS
Programming Languages: C, C++, Java, Python,Scala
Java EE Technologies: JSP, Servlets, EJB, JDBC
Hadoop Ecosystem: HDFS, MapReduce, Flume, Sqoop, Hive, Pig, Spark, and Oozie
ETL Tools: InformaticaPower Center 9.1/9.5 (Designer, Workflow Manager, Workflow Monitor, Repository manager andInformaticaServer), SSIS and SSRS
Databases: Oracle, Mysql, DB2, SQL 2000, Sybase
IDE Tools: Eclipse, NetBeans IDE, RAD, BEA WebLogic studio
Operating Systems: UNIX, Red Hat Linux, Ubuntu, MS-DOS, Sun Solaris, Windows
Testing: Junit
Design/Build Tools: Ant and Maven
Tools: & Utilities: CVS, SVN