We provide IT Staff Augmentation Services!

Bigdata/hadoop Developer Resume

2.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY:

  • Around 4 years of experience in IT industry which includes 2 years of experience in development using Big Data and Hadoop ecosystem tools in retail business.
  • Expertise in designing & developing applications using Big Data core technologies Map Reduce, Hive, Spark & Spark SQL.
  • Good knowledge on cloud bases AWS EC2, EMR & S3.
  • Expertise in writing Map Reduce using advanced features as distributed cache, partitioner, combiner and chained jobs using sequence file format.
  • Expertise in implementing Hive simple, generic custom UDF’s.
  • Expertise in doing performance tuning for map reduce jobs & hive complex queries.
  • Experience in importing and exporting data using SQOOP from HDFS to Relational Database Systems and vice - versa.
  • Experience in installation, configuration, supporting and managing Horton works HDP 2.3 Hadoop platform.
  • Experience in new Hadoop 2.0 architecture of YARN.
  • Have good knowledge in Oozie, Zookeeper & PostgreSQL.
  • Have knowledge on new Hadoop 2.2 analytical eco systems like Spark in memory -processing, Spark SQL and Spark streaming using Kafka distributed messaging system.
  • Strong understanding in Object-Oriented Programming concepts and implementation.
  • Expertise in working Agile based applications in different phases like design, construction, testing and live in retail business.
  • Ability to work effectively with associates at all levels within the organization.
  • Strong background in mathematics and have very good analytical and problem solving skills.
  • Proficient in working Retail domain applications.
  • Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems application architecture.
  • Able to assess business rules, collaborate with stakeholders and perform source-to-target data mapping, design and review.

TECHNICAL SKILLS:

Big Data Ecosystem: Hadoop 2.0, HDFS, Map Reduce, Hive, Sqoop, HBase, Mongo DB, YARN, Spark, Spark SQL, Kafka, Zookeeper and Oozie

Programming Languages: C, C++, Python, SQL, PL/SQL, UNIX/Linux Shell Scripts

Cloud Technologies: AWS EC2, EMR, S3

User Interfaces: HTML5, Java script, AngularJS

Operating Systems: Linux, Mac OS, WINDOWS

Methodologies: Agile, UML, Design Patterns

PROFESSIONAL EXPERIENCE:

Confidential, Atlanta, GA

BIgData/Hadoop Developer

Responsibilities:

  • Played a lead role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
  • Involved in designing different components of system like big-data event processing framework Spark, distributed messaging system Kafka and SQL database PostgreSQL.
  • I have actively participated in HDP 2.3 YARN cluster set up using Ambari.
  • I have implemented Spark Streaming and Spark SQL using Data Frames.
  • I have integrated product data feeds from Kafka to Spark processing system and store the order details in PostgreSQL data base.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Worked on scheduling all jobs using TWS & maestro scripts.

Environment: Red Hat Enterprise Linux 5, HDP 2.3, Kafka, Spark, Spark SQL PostgreSQL, Shell Script, SQOOP1.4.4, Python 2.7 Tools: Eclipse 4.3, GIT, Jenkins

Confidential, Atlanta, GA

BIgData/Hadoop Developer

Responsibilities:

  • Played a lead role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
  • Involved in designing different components of system like sqoop, hadoop process involves map reduce & hive, FTP integration to down systems.
  • Expertise in tuning complex hive queries to reduce the total run time of hive queries while processing terabytes of data.
  • Designing interface validation process to provide validate incoming data arrival in hadoop HDFS before kicking off hadoop process.
  • Involved in coordinating with team to assign tasks and monitor deliverables to meet project time lines.
  • Have written hive queries using optimized ways like using window functions, customizing hadoop shuffle & sort parameters, ORC file format.
  • Developing map reduce programs using Combiners, Sequence Files, Compression techniques, Chained Jobs, multiple input and output API.
  • Maintain System integrity of all sub-components (primarily HDFS, MR and Hive).
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Worked on scheduling all jobs using TWS & maestro scripts.

Environment: Red Hat Enterprise Linux 5, Hadoop 1.0.4, Map Reduce, HDFS, Hive 0.14, Shell Script, SQOOP1.4.4, Python 2.7, PostgreSQL Tools: Eclipse 4.3, GIT, Jenkins.

Confidential

BIgData/Hadoop Developer

Responsibilities:

  • Played a key role in gathering requirements, analysis of entire system and providing estimation on development, testing efforts.
  • Designed and developed dynamic web pages using HTML, AngularJS, Bootstrap, materialize CSS.
  • Developed restful web services using jersey implementation.
  • Developed Junit test cases to test web services.

Environment: AngularJS, Restful WS, HTML, JavaScript MVC, AJAX, JSON

We'd love your feedback!