We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Bentonville, AR

PROFESSIONAL SUMMARY:

  • Having 7+ years’ experience using object oriented design methodologies (OOD), experience in core Java, Big Data and allied technologies.
  • Experience in developing applications using Java and Big data solutions.
  • 3+ years in scalable, distributed software systems like Hadoop (MR1 & YARN), HDFS, Pig, MapReduce, Hive, HBase, Flume and Sqoop.
  • 2+ years of experience in Real time streaming frameworks like Kafka, Storm and Spark.
  • Experience with various Hadoop Distributions like CDH, HDP and Apache.
  • Led end - end design, development and production deployment for critical projects like real time and Batch Data replication.
  • 2+ years of experience in NoSQL databases like MongoDB and HBase.
  • 1+ years of experience in AWS (S3, EC2).
  • 5+ years of experience in Agile Methodology.
  • Solid understanding of relational database concepts.
  • Solid understanding of data warehouse star schema concepts.
  • Experience in creating workflows sessions in Workflow Manager and running teh workflows in Workflow Monitor and analyzing them.
  • Expertise in unit testing using JUnit. Experience in error logging and debugging using Log4J.
  • Goal oriented, organized, team player with good interpersonal skills; thrives well within group environment as well as individually.
  • Strong business and application analysis skills with excellent communication and professional abilities.

TECHNICAL SKILLS:

Big Data Stack: Hadoop, Hive, Sqoop, Pig, HBase (NoSQL), Apache Spark (Spark SQL, Spark streaming) and Oozie

Programming Languages: Java, Scala - Beginner

Databases: PostgreSQL, MS SQL server and Oracle

IDE s & Utilities: Eclipse, SQL DeveloperOperating Systems: UNIX, LINUX and Windows

PROFESSIONAL EXPERIENCE:

Confidential, Bentonville, AR

Hadoop Developer

Responsibilities:

  • Lead team during initial set up of Hadoop platform, designed teh data ingestion and validation approach with ETL flow.
  • Worked on extracting data from Oracle database and load to Hive database
  • Worked extensively on optimizing transformations for better performance
  • Involved in designing approach to handle cursors, functions, stored procedures in PL/SQL to Hive Query Language
  • Developed extraction logic using Sqoop scripts to move data from relational databases to HDFS
  • Developed complex transformations using Hive QL to build aggregate/summary tables
  • Optimized teh performance of Hive Queries by looking into teh Explain plan and also tuning various hive parameters.
  • Developed UDF’s to implement functions dat were not present in Hive
  • Handled Update operations using Hive Queries
  • Involved in using HCATALOG to access Hive table metadata from Map Reduce or Pig code.
  • Performing data validation
  • Developed deployment scripts for Development, QA and Production systems through SVN.
  • Developed workflows using Oozie
  • Expertise in working in Agile environment

Environment: Hadoop, HDFS, Map Reduce, Hive, Oozie, SQL Developer, TOAD, Oracle, PL/SQL, Tortoise, SVN, Eclipse, JAVA, Shell scripting, Spark SQL, Scala, Unix and Tableau

Confidential

HadoopDeveloper

Responsibilities:

  • Designed an approach for Data Validation and Data Ingestion Framework.
  • Configured Pentaho DI with Hadoop to run teh ETL process
  • Worked on validating and converting raw data from various streams like Email, Instant messaging, IP
  • Phone, Desktop Video and Immersive Video to csv files
  • Develop Pentaho jobs to load teh data from source systems to HDFS
  • Develop transformation logic using Hive Queries to build dimension and fact tables
  • Developed schema in HBase for faster scans
  • Used HBase for storing aggregated data used for reporting
  • Worked on Unit testing by creating test data and comparing expected results
  • Developed deployment scripts for production release.
  • Participated in Daily scrum calls and track day to day activities using Rally
  • Expertise in working in Agile environment.

Environment: Hadoop, MapReduce, Yarn, Hive, Pig, HBase, Oozie, Sqoop, Flume, Oracle 11g, Core Java Cloudera HDFS, Eclipse.

Confidential

Hadoop Developer

Responsibilities:

  • Involved in all phases of teh Big Data Implementation including requirement analysis, design, development, building, testing, and deployment of Hadoop cluster in fully distributed mode
  • Created Linux and Python Scripts to automate teh daily ingestion of raw data
  • Processed teh raw data using Hive jobs and scheduling them in Crontab
  • Developed HIVE UDFs to get teh MDK and GeoIp values
  • Moved data to appropriate partition based on record-level timestamp(as we have more than one day’s worth of data in log files)
  • Compressed transformed/enriched data files with bzip2Codec
  • Experience in Daily production support to monitor and trouble shoots Hadoop/Hive jobs
  • Developed Housekeeping process purge teh old data from Edge node and HDFS.
  • Implemented two different process for Internal and External Weblogs
  • Manage and review Hadoop log files.
  • Support/Troubleshoot hive programs running on teh cluster
  • Involved in fixing issues arising out of duration testing
  • Handling structured, semi structured and unstructured data
  • Automated teh History and Purge Process
  • Involved in teh process of migrating to teh new cluster.

Environment: Hadoop 2.x, Hive 0.13.1, Python, Unix Scripts, HDP 2.3, Redhat Linux

Confidential

Java/J2EE Developer

Responsibilities:

  • Responsible for teh development of new requirements for teh presentation layer developed according to FSA style Guide andUSDAstyle guide.
  • Preparing technical specifications, work assignments, coding and unit testing.
  • Responsible for teh development of Assessment Calculation reports module.
  • Junit used for unit testing of teh application.
  • Maven build is used to build teh application on Eclipse 3.2 IDE.
  • Responsible for implementing reports module by using Springs MVC FrameWork.
  • Responsible for reviewing and approving teh projects related documents like Design document, Database design documents and so on.
  • Responsible for designing teh process configuration and event representation.
  • Responsible for teh requirement gathering from teh customer and finalizing teh features.
  • Responsible for updating project status to teh senior project manager in a regular interval.

Environment: JDK 6.0, EJB, JSP, Net beans IDE 6.0, JBoss, Dojo, Web Services, VB Script, Apache POI, Hibernate 3.0, JQuery.

Confidential

Java/J2EE Developer

Responsibilities:

  • Involved in walking through teh functional requirements and estimating teh effort for delivering these requirements.
  • Developed teh Backend code, which has teh business logic and interacts with teh database.
  • Unit Testing and delivering a quality piece of code.
  • Perform Build and deployment activities by using Maven build.
  • Involved in Database schema design, developed stored procedures.

Environment: Java 1.6, Amdocs Smart Client Framework 8.0, Log 4j, Quality Control, MKS

We'd love your feedback!