We provide IT Staff Augmentation Services!

Hadoop/big Data Developer Resume

2.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • 7+ years of IT experience in requirement analysis, design, development, testing and deployment of software systems from development stage to production stage in Big Data and Java technologies.
  • 2+ years of experience in Big Data and Hadoop Ecosystem including MapReduce, Pig, Hive, Sqoop, Oozie, Zookeeper and Flume, good knowledge in components such as Hadoop Distributes File System (HDFS), MRv1, MRv2, Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager and Node Manager.
  • Experience in creating PIG Latin Scripts and UDFs using JAVA for analysis of data efficiently.
  • Experience in creating Hive Queries and UDFs using Java for analysis of data efficiently.
  • Knowledge of Hadoop GEN2 Federation, High Availability and YARN architecture.
  • Experience in using Sqoop for fetching data from different systems into HDFS, and again putting it back to the previous system for further processing.
  • Experience in using NoSQL Databases like Hbase.
  • Experience with Oozie Workflow Engine in running workflow jobs that run Hadoop Map/Reduce and Pig jobs.
  • Good experience in optimizing MapReduce algorithms using Mappers, Reducers, Combiners and Partitioners to deliver best results for the large dataset.
  • Strong expertise on MapReduce programming model with XML, JSON, CSV file formats.
  • Extensive experience on CVS, Clear Case and SVN for Source Controlling.
  • Used Flume to process real time processing data.
  • Worked in Windows, UNIX/Linux platform with different technologies such as SQL, XML and Core Java.
  • Knowledge in using Hadoop ecosystem on Cloudera and AWS.
  • Experience in using IDEs like Eclipse and NetBeans.
  • Extensive programming experience in developing web based applications using Core Java, JSP and JDBC.
  • Experience in deploying applications in Web/Application Servers like Tomcat and WebLogic
  • Strong knowledge of Software Development Life Cycle and expertise in detailed design documentation.
  • Extensive experience with Waterfall and Agile Scrum Methodologies.
  • Experience in development of logging standards and mechanism based on Log4J.
  • Working knowledge of database such as Oracle 8i/9i/10g, Microsoft SQL Server.
  • Experience in writing numerous test cases using JUnit framework.
  • Experience in creating Product Documentation & Presentations.
  • Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.
  • Highly proficient in Object Oriented Programming concepts.

TECHNICAL SKILLS:

Hadoop/Big Data Technologies: Hadoop 2.x, HDFS, MapReduce, Hbase 0.94.8, Pig 0.14.0, Hive 1.1.0, Sqoop 1.4.6, Flume 1.5.2, Cloudera CDH 4, Oozie, YARN

Programming Languages: Java

Scripting/Web Technologies: JavaScript, HTML, XML, Shell Scripting, JDBC, JSP, CSS, JSON

Databases: Oracle, MySQL and NoSQL

Operating Systems: Linux, UNIX and Windows

Java IDE: Eclipse and NetBeans

PROFESSIONAL EXPERIENCE:

Confidential, Chicago, IL

Hadoop/Big Data Developer

Roles and Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Involved in loading data from LINUX file system to HDFS.
  • Created Pig Latin scripts to sort, group, join and filter the data.
  • Worked on tuning the performance of Pig queries.
  • Mentored analyst and test team for writing Hive Queries.
  • Installed Oozie workflow engine to run multiple MapReduce jobs.
  • Extensive Working knowledge of partitioned table, UDFs, performance tuning, compression - related properties, thrift server in Hive.
  • Developed multiple Map Reduce jobs in Java for data cleaning and preprocessing.
  • Involved in installation and configuration of Hadoop Ecosystem components.
  • Experience in Importing and exporting data into HDFS and Hive using Sqoop.
  • Load and transformed large sets of structured, semi structured and unstructured data.
  • Worked on different file formats like Sequence files, XML files, JSON and CSV files using Map Reduce Programs.
  • Responsible for managing data coming from different sources.
  • Gained good experience with NOSQL database.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in collecting data from servers to HDFS using Apache Flume.
  • Worked with the Data Science team to gather requirements for various data mining projects.

Environment: Cloudera CDH 4, HDFS, Hadoop 2.2.0 (Yarn), Flume 1.5.2, Eclipse, Map Reduce, Hive 1.1.0, Pig Latin 0.14.0, Java, SQL, Sqoop 1.4.6, Centos and NOSQL database.

Confidential, Minneapolis, MN

Hadoop/Big Data Developer

Roles and Responsibilities:

  • As a Big Data Developer, implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies such as Hadoop, MapReduce Frameworks, HBase, Hive, Oozie, Flume, Sqoop etc.
  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Experience in optimization of Map reduce algorithm using combiners and partitions to deliver the best results and worked on Application performance optimization for a HDFS cluster.
  • Involved in Unit testing and delivered Unit test plans and results documents using Junit and MRUnit.
  • Involved in collecting, aggregating and moving data from servers to HDFS using Apache Flume.
  • Written Hive scripts to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Designed and implemented Incremental Imports into Hive tables.
  • Involved with File Processing using Pig Latin and implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Imported Bulk Data into HBase Using Map Reduce programs.
  • Developed and written Apache PIG scripts and HIVE scripts to process the HDFS data.
  • Worked on debugging, performance tuning of Hive & Pig Jobs.
  • Created and maintained Technical documentation for executing Hive queries and Pig Scripts.
  • Configured Oozie workflow to run multiple Hive and Pig jobs which run independently with time and data availability.
  • Active participation in code review sessions with the team.

Environment: Java, Hadoop 2.1.0, Map Reduce, Pig 0.12.0, Hive 0.13.0, Linux, Sqoop 1.4.2, Flume 1.3.1, Eclipse, AWS EC2, and Cloudera CDH 4.

Confidential

Java Developer

Roles and Responsibilities:

  • Implemented the presentation layer using Struts MVC framework.
  • Designed HTML Web pages utilizing JavaScript and CSS.
  • Wrote JavaScript functions for various validation purposes.
  • Involved in developing distributed, transactional, secure and portable applications based on Java using EJB.
  • Developed MDB for asynchronous messaging.
  • Deployed web applications in weblogic server by creating Data source and uploading jars.
  • Implemented Multithread concepts in Java classes to avoid deadlocking.
  • Developed unit test cases using Junit.
  • Involved in code review of the modules and improved performance.
  • Supported the deployment of the application changes into production.
  • Involved in the implementation of logical and physical database design by creating suitable tables, views and triggers.
  • Developed related procedures and functions used by JDBC calls in the above components.
  • Extensively involved in performance tuning of Oracle queries.
  • Created components to extract application messages stored in xml files.
  • Mentored junior developers in project development and assignments.

Environment: Java, HTML, JavaScript, CSS, XML, JSP, Servlets, JDBC, Oracle, Eclipse

Confidential

Software Engineer

Roles and Responsibilities:

  • Engaged in the project from the initial expansion of KAP (knowledge Acquisition Phase) session and in every phase of SDLC.
  • Used the Java Struts validation framework for client side validations.
  • Developed SQL queries to access the database using JDBC.
  • Developed Presentation Layer using HTML, JavaScript and CSS.
  • Involved in code review of the modules and improved performance.
  • Supported the deployment of the application changes into production.
  • Enhanced the functionality of the application using COBOL, NATURAL and JCL.
  • Primary On-Call person to resolve the production ABENDS.
  • Resolved high severity L1 and L2 tickets within the SLA.
  • Analysed and fixed critical Out of Balance issues effectively on time.
  • Prepared technical documentation and generated reports for the Business Manager.
  • Solved customer requests on daily basis.

Environment: Java, STRUTS, EJB, Eclipse, Apache Tomcat, HTML, JavaScript, CSS, XML, JSP, JDBC, Oracle, Windows, COBOL, Natural, JCL

We'd love your feedback!