We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

LA

SUMMARY

  • 5+ years of overall experience in IT Industry which includes experience in Java, Big data technologies (Hadoop).
  • 2 years of comprehensive experience in Big Data Analytics.
  • Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and Map Reduce concepts.
  • Experience in using Pig, Hive, Sqoop and Cloudera Manager.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Hands on experience in RDBMS, and Linux shell scripting
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Experience in analyzing data using HiveQL, Pig Latin and Map Reduce.
  • Expertise in Domain Retail and Finance.
  • Knowledge in job work-flow scheduling and monitoring tools like Oozie.
  • Knowledge of data warehousing and ETL tools like Informatica and Pentaho.
  • Techno-functional responsibilities include interfacing with users, identifying functional and technical gaps, estimates, designing custom solutions, development, leading developers, producing documentation, and production support.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.
  • Trained in Cloudera for Apache Hadoop.
  • Extensive experience spans with implementing SOA architecture using WebServices using both SOAP and Rest.
  • Project delivery of various sizes, performance tuning, production support, trouble shooting and development.
  • Extensive work experience as onsite coordinator attending the project review meetings with Clients, gathering requirements, preparing HLD, LLD, requirements transition and design to offshore for the Build activity.
  • Strong problem solving skills and a thirst to learn the latest Technologies in the industry.
  • Ability to handle multiple projects concurrently, meet strict deadlines, coordination with multiple teams, Systems and Applications.
  • Trained in multiple Soft Skills like Time Management, Cross Culture, Client Interaction and Communications.
  • Java developer with experience withHadoopand Kafka.
  • Have knowledge in projects like Data Integration & Migration with functional knowledge in Banking & Finance, Retail and Healthcare & life science projects.

TECHNICAL SKILLS

Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Sqoop, Oozie

Build Tool & IDE: Eclipse

Languages: Java, JDBC, Sevlets, JSP, Java Script, Shell scripting, COBOL, JCL, PL/SQL, REXX, Ezytrieve

Mainframe Tools: Changeman, Endevor, Fila, File Manager, ISPF, Xpeditor, IBM Utilities, CA - 7.

Testing Tools: HP Quality Center

PROFESSIONAL EXPERIENCE

Confidential, LA

Hadoop Developer

Responsibilities:

  • Work on analyzing and writing Map Reduce jobs based on the requirements.
  • Involve in creating Hive tables, loading with data and writing hive queries that will internally run as Map Reduce job.
  • Work on Flume to stream Twitter data in to HDFS to analyze Black Friday promotions.
  • Work on creating Hive scripts for data analysts based on the Ad-hoc requirements.
  • Involve in writing UDF for Hive.
  • Schedule the workflow using Oozie workflow scheduler.
  • Extract the data from external data sources into HDFS using Sqoop.
  • Develop Map Reduce programs using combiner and custom partition.
  • Create Sqoop job to populate Hive external tables.
  • Involve in configuring Flume to import Sofi and DGW logs to HDFS.
  • Using Cloudera Manager for monitoring purpose.
  • Involve in loading data from Linux file system to HDFS
  • Involve in all the phases of the SDLC using Agile Scrum methodology.
  • Working in Agile environment with 3 weeks sprint, involving in grooming meeting, sprint planning meeting, retrospective meeting and daily stand up meeting.
  • Using Version One tool for tracking all agile related activities like stories, tasks, estimated story points and actual burn hours.

Environment: Hadoop, Map Reduce, HDFS, Hive, Sqoop, Flume, Oozie, Cloudera ManagerHue UI, Java (JDK 1.6), Eclipse, Maven, JSON, MySQL, Ubuntu, Version One and Cloudera Distribution

Confidential, Washington D.C

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing
  • Involved in data extraction from distributed RDBMS like Teradata and Oracle.
  • Involved in loading data from UNIX file system to HDFS.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Used Map Reduce JUnit for unit testing.
  • Troubleshooting the cluster by reviewing Hadoop LOG files. Involved in managing and reviewing Hadoop log files.
  • Installed and configured Pig for ETL jobs.
  • Used Oozie to manage the Hadoop jobs.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Used CDH3 and CDH4 distributions for development and deployment.
  • Imported data using Sqoop from Tera data using Tera data connector.
  • Implemented Partitioning, Dynamic Partitioning, and Bucketing in HIVE.
  • Exported the result set from HIVE to MySQL using Shell scripts.
  • Used Zookeeper for various types of centralized configurations.
  • Involved in maintaining various Unix Shell scripts.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
  • Used SVN for version control.
  • Maintain System integrity of all subcomponents (primarily HDFS, MR, HBase, and Flume).
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.

Environment: Hadoop, Map Reduce, Hive, Pig, Sqoop,HBase(NoSQL database), Java 1.6,and UNIX Shell Scripting.

Confidential

System Analyst

Responsibilities:

  • Involved in software development life cycle (SDLC) of the project (analysis, design, development, testing and implementation).
  • Used MS-Visio for analysis & design flow and monitored the initial applications prototype development for the project.
  • Leading a team of 4 developers and point of contact for onshore / offshore communication.
  • Designed and developed web pages using JSP, JDBC and Servlets.
  • Developed modules using Java.
  • Used HTML, CSS, XML and JavaScript to design a page.
  • Developed Web Services to get data from the external system in terms of .txt file to load into the database.
  • Developed DTS/SSIS packages to load employee details into row Mony tables of the SQL server for further processing.
  • Wrote Stored Procedures, functions and complex SQL queries for database operations.
  • Used Java Scripts to perform client side validations.
  • Worked on performance tuning of queries.
  • Developed reports using Crystal Reports reporting tool.
  • Have used LDAP and Active Directory Search Interface (ADSI) to authenticate and authorize user.
  • Involved in unit testing and production support of the application.
  • Defects were managed through the Remedy

Environment: JEE, JSP, Servlets, SQL Server 2005, T-SQL/PL-SQL, HTML, XML, Web Services, JavaScript, Tomcat, Visual Source Safe (VSS), Remedy and Crystal Reports

Confidential

Java Developer & Support

Responsibilities:

  • Provide L3 application support as primary on call.
  • Involved in the development of Report Generation module which includes volume statistics report, Sanctions Monitoring Metrics report, and TPS report.
  • Implemented Online List Management (OLM) and FMM module using spring and Hibernate.
  • Wrote various SQL, PL/SQL queries and stored procedures for data retrieval.
  • Created Configuration files for the application using Spring framework.

Environment: JEE, JSP, Servlets, JQuery, JavaScript, CSS, HTML, SQL

We'd love your feedback!