Bigdata/hadoop Developer Resume
Atlanta, GA
SUMMARY:
- Around 4 years of experience in IT industry which includes 2 years of experience in development using Big Data and Hadoop ecosystem tools in retail business.
- Expertise in designing & developing applications using Big Data core technologies Map Reduce, Hive, Spark & Spark SQL.
- Good knowledge on cloud bases AWS EC2, EMR & S3.
- Expertise in writing Map Reduce using advanced features as distributed cache, partitioner, combiner and chained jobs using sequence file format.
- Expertise in implementing Hive simple, generic custom UDF’s.
- Expertise in doing performance tuning for map reduce jobs & hive complex queries.
- Experience in importing and exporting data using SQOOP from HDFS to Relational Database Systems and vice - versa.
- Experience in installation, configuration, supporting and managing Horton works HDP 2.3 Hadoop platform.
- Experience in new Hadoop 2.0 architecture of YARN.
- Have good knowledge in Oozie, Zookeeper & PostgreSQL.
- Have knowledge on new Hadoop 2.2 analytical eco systems like Spark in memory -processing, Spark SQL and Spark streaming using Kafka distributed messaging system.
- Strong understanding in Object-Oriented Programming concepts and implementation.
- Expertise in working Agile based applications in different phases like design, construction, testing and live in retail business.
- Ability to work effectively with associates at all levels within the organization.
- Strong background in mathematics and have very good analytical and problem solving skills.
- Proficient in working Retail domain applications.
- Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems application architecture.
- Able to assess business rules, collaborate with stakeholders and perform source-to-target data mapping, design and review.
TECHNICAL SKILLS:
Big Data Ecosystem: Hadoop 2.0, HDFS, Map Reduce, Hive, Sqoop, HBase, Mongo DB, YARN, Spark, Spark SQL, Kafka, Zookeeper and Oozie
Programming Languages: C, C++, Python, SQL, PL/SQL, UNIX/Linux Shell Scripts
Cloud Technologies: AWS EC2, EMR, S3
User Interfaces: HTML5, Java script, AngularJS
Operating Systems: Linux, Mac OS, WINDOWS
Methodologies: Agile, UML, Design Patterns
PROFESSIONAL EXPERIENCE:
Confidential, Atlanta, GA
BIgData/Hadoop Developer
Responsibilities:
- Played a lead role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
- Involved in designing different components of system like big-data event processing framework Spark, distributed messaging system Kafka and SQL database PostgreSQL.
- I have actively participated in HDP 2.3 YARN cluster set up using Ambari.
- I have implemented Spark Streaming and Spark SQL using Data Frames.
- I have integrated product data feeds from Kafka to Spark processing system and store the order details in PostgreSQL data base.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Worked on scheduling all jobs using TWS & maestro scripts.
Environment: Red Hat Enterprise Linux 5, HDP 2.3, Kafka, Spark, Spark SQL PostgreSQL, Shell Script, SQOOP1.4.4, Python 2.7 Tools: Eclipse 4.3, GIT, Jenkins
Confidential, Atlanta, GA
BIgData/Hadoop Developer
Responsibilities:
- Played a lead role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
- Involved in designing different components of system like sqoop, hadoop process involves map reduce & hive, FTP integration to down systems.
- Expertise in tuning complex hive queries to reduce the total run time of hive queries while processing terabytes of data.
- Designing interface validation process to provide validate incoming data arrival in hadoop HDFS before kicking off hadoop process.
- Involved in coordinating with team to assign tasks and monitor deliverables to meet project time lines.
- Have written hive queries using optimized ways like using window functions, customizing hadoop shuffle & sort parameters, ORC file format.
- Developing map reduce programs using Combiners, Sequence Files, Compression techniques, Chained Jobs, multiple input and output API.
- Maintain System integrity of all sub-components (primarily HDFS, MR and Hive).
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Worked on scheduling all jobs using TWS & maestro scripts.
Environment: Red Hat Enterprise Linux 5, Hadoop 1.0.4, Map Reduce, HDFS, Hive 0.14, Shell Script, SQOOP1.4.4, Python 2.7, PostgreSQL Tools: Eclipse 4.3, GIT, Jenkins.
Confidential
BIgData/Hadoop Developer
Responsibilities:
- Played a key role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
- Designed and developed dynamic web pages using HTML, AngularJS, Bootstrap, materialize CSS.
- Developed restful web services using jersey implementation.
- Developed Junit test cases to test web services.
Environment: AngularJS, Restful WS, HTML, JavaScript MVC, AJAX, JSON
