We provide IT Staff Augmentation Services!

Java Hadoop Consultant Resume

5.00/5 (Submit Your Rating)

Franklin Lakes, NJ

SUMMARY

  • Over 5.2 years of professional IT experience with Over 5 Years of Hadoop/Spark experience in ingestion, storage, querying, processing and analysis of big data and 5 Years of Java
  • Proficient in Installation, Configuration and migrating and upgrading of data from Hadoop MapReduce, HIVE, HDFS, HBase, Sqoop, Pig, Cloudera, YARN.
  • Excellent understanding/knowledge of Hadoop architecture and various components such as Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Experience with leveraging Hadoop ecosystem components including Pig and Hive for data analysis, Sqoop for data migration, Oozie for scheduling and HBase as a NoSQL data store.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Experience in Hadoop Shell commands, writing MapReduce Programs, verifying managing and reviewing Hadoop Log files.
  • Hands on experience on Cloudera Hadoop environments.
  • Experienced in application design using Unified Modeling Language (UML), Sequence diagrams, Case diagrams, Entity Relationship Diagrams (ERD), Data Flow Diagrams (DFD).
  • Proficiency in programming with different Java IDE's like Eclipse, and NetBeans.
  • Experience in database development using SQL and PL/SQL and experience working on databases like Oracle, SQL Server and MySQL.
  • Developed Spark jobs and Hive Jobs to summarize and transform data.
  • Experience in Writing Producers/Consumers and creating messaging centric applications using Apache Kafka.
  • Hands on experience in Amazon Web Services (AWS) provisioning tools like EC2, Simple Storage Service (S3), Elastic Map Reduce.
  • Extensive Experience in Java development skills using J2SE, J2EE technologies like Servlets, Spring Hibernate, JSP, JDBC.
  • Experienced in Java components like Frame work collection, Exception handling, and Multithreading and I/O system.
  • Experience in SOA using Soap and RESTful.
  • Knowledge on writing Hadoop Jobs for analyzing data using Hive and Pig.
  • Experience in NoSQL database.
  • Strong team player, ability to work independently and in a team as well, ability to adapt to a rapidly changing environment, commitment towards learning.
  • Ability to blend technical expertise with strong Conceptual, Business and Analytical skills to provide quality solutions and result-oriented problem-solving technique and leadership skills.

TECHNICAL SKILLS

Big Data Eco Systems: Hadoop (HDFS & Map Reduce), PIG, HIVE, HBASE, Sqoop, Kafka, Apache Spark

Databases: Oracle 9i/10g/11g, SQL Server 2008, MS-SQL Server.

Hadoop Distributions: Cloudera, Horton works.

Languages: Java, SQL, JavaScript, XML

Web Technologies: JavaScript, J-Query, Boot Strap, AJAX, XML, CSS, HTML, AngularJS.

Web Services: REST, SOAP, JAX-WS, JAX-RPC, JAX-RS, WSDL, Axis2, Apache HTTP, CVS, SVN.

IDE: Eclipse, Net beans, IntelliJ.

Operating Systems: Windows Variants, Linux, UNIX.

PROFESSIONAL EXPERIENCE

Confidential, Franklin Lakes, NJ

Java Hadoop consultant

Responsibilities:

  • Utilized object-oriented programming and Java for creating business logic.
  • Developed Spark scripts by using Scala Shell commands as per the requirement.
  • Dev Testing on Different Models.
  • Used Hive to do analysis on the data and identify different correlations.
  • Wrote SQL queries to retrieve data from Database using JDBC.
  • Performed file transfers using Tectia SSH Client.
  • Implemented Hadoop framework to capture user navigation across the application to validate the user interface and provide analytic feedback/result to the UI team.
  • Developed Spark scripts by using Scala Shell commands as per the requirement.
  • Developed Map-Reduce jobs on Yarn and Hadoop clusters to produce daily and monthly reports.
  • Automated all the jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Loaded the data from Teradata to HDFS using Teradata Hadoop connectors.
  • Wrote Map Reduce jobs using Java API and Pig Latin.
  • Wrote Pig scripts to run ETL jobs on the data in HDFS.
  • Used Hive to do analysis on the data and identify different correlations.
  • Deployment and administration of Splunk and Hortonworks Distribution.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Involved in creating Hive tables and working on them using Hive QL.
  • Wrote various queries using SQL and used SQL server as the database.
  • Utilized Agile Scrum Methodology to help manage and organize a team of 4 developers with regular code review sessions.

Environment: Hadoop, Map Reduce, Spark, Scala, HDFS, Pig, Hive, HBase, Sqoop, Hortonworks, Zookeeper, Cloudera, Oracle, agile, Windows, UNIX Shell Scripting.

Confidential - Seattle Washington

Java Hadoop consultant

Responsibilities:

  • Implemented Java/J2EE Design patterns like Business Delegate and Data Transfer Object (DTO), Data Access Object.
  • Developed data pipeline using Sqoop, Pig and Java MapReduce to ingest customer behavioral data and financial histories into HDFS for analysis.
  • Worked on the Hortonworks environment.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed several new Map Reduce programs to analyze and transform the data to uncover insights into the customer usage patterns.
  • Developed Hive UDFs to validate against business rules before data move to hive table
  • Developed MapReduce jobs in both PIG and Hive for data cleaning and pre-processing.
  • Developed Sqoop scripts for loading data into HDFS from DB2 and preprocessed with PIG.
  • Created Hive External tables and loaded the data in to tables and query data using SQL.
  • Involved in writing Flume and Hive scripts to extract, transform and load the data into Database.
  • Performed data analysis in Hive by creating tables, loading it with data and writing hive queries which will run internally in a MapReduce way.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase, NoSQL database and Sqoop.
  • Developed shell script to pull the data from third party system's into Hadoop file system.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig.
  • Involved in Database design and developing SQL Queries, stored procedures on MySQL.
  • Involved in Database design with Oracle as backend.

Environment: Hadoop, MapReduce, HDFS, Sqoop, Pig, HBase, Hive, Horton Works, Cassandra, Zookeeper, Cloudera, Oozie, MongoDB, Sqoop, NoSQL, SQL, Oracle, UNIX/LINUX.

Confidential

Hadoop Administrator

Responsibilities:

  • Involved in various phases of Software Development Life Cycle (SDLC) of the application like Requirement gathering, Design, Analysis and Code development.
  • Prepared Use Cases, sequence diagrams, class diagrams and deployment diagrams based on UML to enforce Rational Unified Process using Rational Rose.
  • Developed a prototype of the application and demonstrated to business users to verify the application functionality.
  • Developed and implemented the MVC Architectural Pattern using Struts Framework including JSP, Servlets, EJB, Form Bean and Action classes.
  • Developed JSP's with Custom Tag Libraries for control of the business processes in the middle-tier and was involved in their integration.
  • Developed the User Interface using spring, html, logic, bean, JSP, Java Script, HTML and CSS.
  • Designed and developed backend java Components residing on different machines to exchange information and data using JMS.
  • Developed the war/ear file using Ant script and deployed into Web Logic Application Server.
  • Used parsers like SAX and DOM for parsing XML documents.
  • Implemented Java/J2EE Design patterns like Business Delegate and Data Transfer Object (DTO), Data Access Object.
  • Used Rational Clear Case as Version control.
  • Written stored procedures, triggers, and cursors using Oracle PL/SQL.
  • Worked with QA team for testing and resolving defects.
  • Used ANT automated build scripts to compile and package the application.
  • Used Jira for bug tracking and project management.

Environment: J2EE, JSP, JDBC, Spring Core, Struts, Hibernate, Design Patterns, XML, WebLogic, Apache Axis, ANT, Clear case, Junit, UML, Webservices, SOAP, XSLT, Jira, Oracle, PL/SQL Developer and Windows.

We'd love your feedback!