We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

New, YorK

SUMMARY

  • Having total work experience of 7+ years in Information Technology with skills in analysis, design, development, testing and deploying various software applications, which include Web related and windows applications with emphasis on Object Oriented Programming.
  • 3+ years of work experience on Big Data development with excellent experience in implementing Map Reduce Algorithms/programs in Hadoop.
  • Hands on experience in installing, configuring and using ecosystem components like Hadoop Map Reduce, HDFS, Pig, Hive, Sqoop and Oozie.
  • Excellent understanding/knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node&Map Reduce programming paradigm.
  • Handling and further processing schema oriented and non - schema oriented data using Pig.
  • Experience in analyzing data using HiveQL, Pig Latin and custom Map Reduce programs in Java.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Good Knowledge on NoSql databases like HBase, Mongo DB, and Mahout.
  • Good Knowledge on Impala, Strom and Yarn.
  • Work experience as a Java/J2EE programmer developing applications using Servlets, JSP, spring, Java Beans, JDBC and MVC architecture.
  • Good knowledge of database connectivity (JDBC) for databases like Oracle, DB2, SQL Server, MySQL, MS Access, Netezza.
  • Experience on Core java technology, which includes multithreading, JDBC, Collections, network programming.
  • Experience working with BI teams for reports generation.
  • Good knowledge on ETL processing and data science solutions like SAS, Micro Strategy.
  • Good knowledge on Python.
  • Experienced in coding PL/SQL, Procedures/Functions, Triggers, and Packages on database (RDBMS) packages like Oracle and also experience on SSIS and SSRS.
  • Worked on Agile methodology, SOA for many of the applications.
  • Excellent analytical, problem solving, communication and interpersonal skills with ability to interact with individuals at all levels and can work as a part of a team as well as independently.
  • Strong Communication skills of written, oral, interpersonal and presentation.
  • Implemented Unit Testing using JUNIT testing during the projects.
  • Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.

TECHNICAL SKILLS

Hadoop/Big Data: HDFS, Map Reduce, Hbase, Pig, Hive, Sqoop, Flume, OozieZookeeper, YARN

Technologies: Servlets, JSP, JDBC, XML, Java, SQL, PL/SQL.

Programming Languages: C, C++, Java, Visual Basic, SQL, PL/SQL, Shell Scripting

Scripting Languages: HTML, CSS, JavaScript, Python

J2EE Technologies: Servlets, JSP, Custom Tag Libraries, JDBC, JNDI, Collections, Multi-threading, JMS, Java Mail, JAXB, JAXP.EJB

Web Technologies: HTML, DHTML, JavaScript, Ajax, CSS, XML, DTD, XSD, XSLT, SAXDOM, JSON

Databases: Oracle 8.x/9.x, MySQL, SQL Server.

Application/Web Servers: Apache Tomcat, JBoss, Web Sphere, Web Logic

Web Services: Apache CXF/XFire, Apache Axis, SOAP, Jersey

IDE: Eclipse, Net Beans, JDeveloper, WSAD

Frameworks: Struts, spring, Hibernate

Testing/Logging tools: JUnit, Easy Mock, JMock, log4J

Operating Systems: Windows XP/Vista/7, 8, 10, Unix/Linux, Mac.

PROFESSIONAL EXPERIENCE

Confidential, New York

Hadoop Developer

Responsibilities:

  • Moved all crawl data flat files generated from various retailers to HDFS for further processing.
  • Developed the PIG code for loading, filtering and storing the data.
  • Written the Apache PIG scripts to process the HDFS data.
  • Created Hive tables to store the processed results in a tabular format.
  • Developed Hive scripts to denormalize and aggregate the disparate data.
  • Implemented external tables and partitions using HIVE.
  • Developed the Sqoop scripts in order to make the interaction between Pig (ETL) and MySQL Database.
  • Experience on Cloudera distributions.
  • Involved in gathering the requirements, designing, development and testing.
  • Writing the script files for processing data and loading to HDFS.
  • Setting Password less Hadoop and working on MR1 and MR2.
  • Automated workflows using shell scripts and Oozie jobs.
  • Implemented HBASE for creating tabular data.
  • Integrated and worked with the reporting teams like SAS and Micro Strategy.
  • Worked on implement Hadoop streaming through Apache Kafka and Spark.
  • Moved all log/text files generated by various products into HDFS location
  • Created External Hive Table on top of parsed data.

Environment: JDK,JRE, Red Hat Linux, HDFS, Map-Reduce, Hive, Pig, Sqoop, Oozie, Oracle, HBase.

Confidential, Livingston, NJ

Hadoop Developer

Responsibilities:

  • Developed Pig program for loading and filtering the streaming data into HDFS.
  • Imported data from Oracle database to HDFS using Sqoop.
  • Worked on Data cleansing using apache Avro schema and implementing it in Pig.
  • Developed Hive scripts to denormalize and aggregate the disparate data.
  • Automated workflows using shell scripts and Oozie jobs to pull data from various databases into Hadoop.
  • Implemented external tables and dynamic partitions using Hive.
  • Worked on custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSV, etc.
  • Loaded the created HFiles into HBase for faster access of large customer base without taking performance hit.
  • Used Oozie scheduler system to automate the pipeline workflow.
  • Actively participated in software development lifecycle (scope, design, implement, deploy, test), including design and code reviews.
  • Implemented data serialization using apache Avro.
  • Involved in story-driven agile development methodology and actively participated in daily scrum meetings.

Environment: Horton works, Hadoop, Map Reduce, HDFS, Hive, Sqoop, Avro, Oozie, Java (jdk6), Informaticaand Oracle 4.0.8.

Confidential, Schaumburg, IL

Hadoop and Java Developer

Responsibilities:

  • Installed and configured Hadoop Map Reduce, HDFS and developed multiple Map Reduce jobs in Java for data cleansing and preprocessing
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup and disaster recovery systems and procedures
  • Extracted files from Couch DB, Mongo DB through Sqoop and placed in HDFS for processed
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS
  • Developed Puppet scripts to install Hive, Sqoop, etc. on the nodes
  • Data back up and synchronization using Amazon Web Services
  • Worked on Amazon Web Services as the primary cloud platform
  • Load and transform large sets of structured, semi structured and unstructured data
  • Supported Map Reduce Programs those are running on the cluster
  • Load log data into HDFS using Flume, Kafka and performing ETL integrations
  • Designed and implemented DR and OR procedures
  • Wrote shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
  • Utilized Java and MySQL from day to day to debug and fix issues with client processes
  • Used JAVA, J2EE application development skills with Object Oriented Analysis and extensively involved throughout Software Development Life Cycle (SDLC)
  • Hands-on experience of Sun One Application Server, Web logic Application Server, Web Sphere Application Server, Web Sphere Portal Server, and J2EE application deployment technology
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process, etc.
  • Monitoring Hadoop cluster using tools like Nagios, Ganglia, Ambari and Cloudera Manager
  • Automation script to monitor HDFS and HBase through Cron jobs
  • Used MRUnit for debugging Map Reduce that uses sequence files containing key value pairs.
  • Develop high-performance cache, making the site stable and improving its performance
  • Create a complete processing engine, based on Cloudera's distribution
  • Proficient with SQL languages and good understanding of Informatica and Talend Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster

Environment: Hadoop, Map Reduce, HDFS, Hive, Apache Spark, Kafka, CouchDB, Flume, AWS, Cassandra, Java, Struts, Servlets, HTML, XML, SQL, J2EE, MRUnit, JUnit, JDBC, SQL, XML, Eclipse.

Confidential, Warren, NJ

JAVA Developer

Responsibilities:

  • Developed the application using Struts Framework that leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used
  • Worked in an Agile work environment with Content Management system for workflow management and content versioning
  • Involved in designing user screens and validations using HTML, jQuery, Ext JS and JSP as per user requirements
  • Familiar with SQL Server Integration SSIS and SSRS
  • Responsible for validation of Client interface JSP pages using Struts form validations
  • Integrating Struts with Spring IOC
  • Used Spring Dependency Injection properties to provide loose-coupling between layers
  • Implemented the Web Service client for the login authentication, credit reports and applicant information using Apache Axis 2 Web Service
  • UsedHibernateORM framework withSpringframework for data persistence and transaction management
  • Used Hibernate 3.0 object relational data mapping framework to persist and retrieve the data from database
  • Created physical and logical data model, design and implementation
  • Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
  • Developed ANT Scripts to do compilation, packaging and deployment in the Web Sphere server
  • Implemented the logging mechanism using Log4j framework
  • Designing and implementing algorithms
  • Wrote test cases in JUnit for unit testing of classes

Environment: JDK 1.6, J2EE 1.4, Agile Development Process, Struts 1.3, Spring 2.0, Web Services, Hibernate 3.0, RSA, JMS, JSP, Servlets 2.5, Web Sphere 6.1, SQL Server 2005, DB2, Windows XP, HTML, XML, IBM Rational Application Developer (RAD), ANT 1.6, Log4J, XML, XSLT, XSD, jQuery, JavaScript, Ext JS, JSON, JUnit 3.8.

Confidential

Junior JAVA Developer

Responsibilities:

  • Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
  • Reviewed the functional, design, source code and test specifications
  • Involved in developing the complete front end development using Java Script and CSS
  • Author for Functional, Design and Test Specifications
  • Implemented Backend, Configuration DAO, XML generation modules of DIS
  • Analyzed, designed and developed the component
  • Used JDBC for database access
  • Used Data Transfer Object (DTO) design patterns
  • Unit testing and rigorous integration testing of the whole application
  • Written and executed the Test Scripts using JUNIT
  • Actively involved in system testing
  • Developed XML parsing tool for regression testing
  • Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product

Environment: Java, JavaScript, HTML, CSS, JDK 1.5.1, JDBC, Oracle10g, XML, XSL, Solaris and UML

We'd love your feedback!