We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Bethlehem, PA

SUMMARY

  • Around 8 years of experience in Analysis, Architecture, Design, Development, Testing, Maintenance and User training of software application which includes over 3+ Years in Big Data, Hadoop and HDFS environment and around 4+ Years of experience in JAVA, J2EE.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per requirement.
  • Good working knowledge on Data Transformations and Loading using Export and Import.
  • Hands on experience using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Used different Hive Serde's like Regex Serde and HBase Serde.
  • Experience in analyzing data using Hive, Pig Latin, and custom MR programs in Java.
  • Hands on experience in writing Spark SQL scripting.
  • Sound knowledge in programming Spark using Scala.
  • Good understanding in processing of real-time data using Spark.
  • Hands-on experience in provisioning and managing multi-tenant Cassandra cluster on public cloud environment - Amazon Web Services(AWS) - EC2, Open Stack.
  • Hands on using job scheduling and monitoring tools like Kafka, Oozie and Zookeeper.
  • Developed small distributed applications in our projects using Zookeeper and Scheduled the work flows using Oozie.
  • Used Pig as ETL tool to do transformations, event joins, filter and some pre-aggregation.
  • Clear understanding onHadoop architecture and various components such as HDFS, Job and Task Tracker, Name and Data Node, Secondary Name Node and Map Reduce programming.
  • Expertise writing custom UDFs for extending Hive and Pig core functionality.
  • Hands on dealing with log files to extract data and to copy into HDFS using flume.
  • Wrote Hadoop Test Cases in Hadoop for checking Input and Outputs.
  • Hands on integrating Hive and HBase.
  • Experience in NOSQL databases: MongoDB, HBase, Cassandra.
  • Experience inHadoopadministration activities such as installation and configuration of clusters using Apache and Cloudera.
  • Knowledge on installing, configuring and using Hadoop components like Hadoop Map Reduce(MR1), YARN(MR2), HDFS, Hive, Pig, Flume and Sqoop.
  • Experience in analyzing, designing and developing ETL strategies and processes, writing ETL specifications, Informatica development.
  • Extensively used Informatica Power Center for Extraction, Transformation and Loading process.
  • Experience in Dimensional Data Modeling using Star and Snow Flake Schema.
  • Worked on reusable code known as Tie outs to maintain the data consistency. .
  • More than 4 years of experience in JAVA, J2EE, Web Services, SOAP, HTML and XML related technologies demonstrating strong analytical and problem solving skills, computer proficiency and ability to follow through with projects from inception to completion.
  • Extensive experience working in Oracle, DB2, SQL Server and My SQL database and Java Core concepts like OOPS, Multithreading, Collections and IO.
  • Hands on JAXWS, JSP, Servlets, Struts, Web Logic, Web Sphere, Hibernate, Spring, JBoss, JDBC, RMI, Java Script, Ajax, jQuery, Linux, Unix, WSDL, XML, HTML,AWS and Scala and Vertica.
  • Developed applications using Java, RDBMS, and Linux shell scripting.
  • Experience in complete project life cycle of Client Server and Web applications.
  • Good understanding of Data Mining and Machine Learning techniques.
  • Have good interpersonal, communicational skills, strong problem solving skills, explore/adopt to new technologies with ease and a good team member.

TECHNICAL SKILLS

Hadoop/BigData Technologies: Hadoop (Horton works, Cloudera and Pivotal) HDFS, Map Reduce, Pig, HBase, Spark,Zookeeper, Hive, Oozie, Sqoop, Flume, Impala,Kafka, Solr and Tableau.Programming Languages Java JDK1.4/1.5/1.6 (JDK 5/JDK 6), C/C++, HTML, SQL, PL/SQL, AVS & JVS

Frameworks: Hibernate 2.x/3.x, Spring 2.x/3.x,Struts 1.x/2.x

Web Services: WSDL, SOAP, Apache CXF/XFire, Apache Axis, REST, Jersey

Client Technologies: JQUERY, Java Script, AJAX, CSS, HTML 5, XHTML

Operating Systems: UNIX, Windows, LINUX

Application Servers: IBM Web sphere, Tomcat, Web Logic, Web Sphere

Web technologies: JSP, Servlets, JNDI, JDBC, Java Beans, JavaScript

Databases: Greenplum,Oracle 8i/9i/10g,Netezza & MySQL 4.x/5.x

Java IDE: Eclipse 3.x, IBM Web Sphere Application Developer, IBM RAD 7.0

Development Tools: TOAD, SQL Developer, SOAP UI, ANT, Maven,Visio, Rational Rose, Endur 8.x/10.x/11.x, Informatica 9.1.

PROFESSIONAL EXPERIENCE

Confidential - Bethlehem, PA

Hadoop Developer

Responsibilities:

  • Worked with Business Analyst and helped representing the business domain details.
  • Hands on experience in gathering information from different nodes into Greenplum database and then Sqoop incremental load into HDFS.
  • Experience in importing the real time data to hadoop using Kafka and implemented the Oozie job.
  • Involved in loading data from LINUX file system to HDFS
  • Experience in Writing Map Reduce jobs for text mining and worked with predictive analysis team to check the output and requirement.
  • Hands on experience in writing hive UDF's for the requirements and to handle different schema’s and xml data.
  • Used Pig as ETL tool to do transformations, event joins, filter both traffic and some pre-aggregations before storing the data onto HDFS.
  • Wrote Hive and Pig scripts for joining the raw data with the lookup data and for some aggregative operations as per the business requirement.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Involved in writing Flume and Hive scripts to extract, transform and load the data into Database
  • Implemented Partitioning and bucketing in Hive based on the requirement.
  • Connected Tableau from client end with AWS ip addresses and view the end results.
  • Coordinator and Oozie workflows are developed to automate Hive, Map Reduce, Pig and other jobs.
  • Creation of test cases as part of enhancement rollouts and Involved in Unit level and Integration level testing.
  • Hands on experience in working with snappy compression and also different file formats.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports For the BI team.

Environment: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, Sqoop, Spark, Oozie, Impala, Greenplum, kafka, SQL, Java (jdk 1.6), Eclipse.

Confidential - Cleveland, OH

Big Data/Hadoop Developer

Responsibilities:

  • Worked closely with the business analysts to convert the Business Requirements into Technical Requirements and prepared low and high level documentation.
  • Hands on experience on writing MR jobs for encryption and also for converting text data into Avro format.
  • Hands on experience in joining raw data with the reference data using Pig scripting.
  • Hands on experience in writing scripting for copying data between different clusters and also between different Unix file systems.
  • Hands on experience in writing MR jobs for cleansing the data and to copy it to AWS cluster form our cluster.
  • Developed Spark SQL script for handling different data sets and verified its performance over MR jobs.
  • Connected Tableau from client end with AWS ip addresses and view the end results.
  • Developed Coordinator and Oozie workflows to automate the jobs.
  • Hands on experience in writing hive UDF's to handle different Avro schema's.
  • Experience with moving large datasets hourly with AVRO file format and imposing hive and impala queries.
  • Hands on experience in working with snappy compression and also different file formats.
  • Developed shell script to backup the name node meta data.
  • Cloudera Manger was used to Monitor the health of Jobs which are running on the cluster.

Environment: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, Sqoop, Spark, Oozie, Impala, SQL, Java (jdk 1.6), Eclipse.

Confidential - Moline, IL

Big Data/HadoopDeveloper

Responsibilities:

  • Responsible for building scalable distributed data solutions usingHadoop and migrate legacy Retail applications ETL to Hadoop.
  • Accessed information through mobile networks and satellites from the equipment.
  • Hands on extracting data from different databases and to copy into HDFS file system using Sqoop.
  • Implemented ETL code to load data from multiple sources into HDFS using pig scripts.
  • Hands on creating different applications in social networking websites and obtaining access data from them.
  • Wrote Map Reduce jobs using the access tokens to get the data from the customers.
  • Developed simple to complex Map Reduce jobs using Hive and Pig for analyzing the data.
  • Used different Serde's for converting JSON data into pipe separated data.
  • Implemented some business logics by writing UDFs in Java and used various UDFs from Piggybanks and other sources to get some results from the data.
  • Worked with cloud administrations like Amazon web services (AWS).
  • Used Oozie workflow engine to run multiple Hive and Pig jobs.
  • Hands on exporting the analyzed data into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Involved in installing and configuring Hive, Pig, Sqoop, Flume and Oozie on theHadoopcluster.
  • Worked with application teams to install operating system,Hadoopupdates, patches, version upgrades as required.
  • Continuously monitored and managed the Hadoop Cluster using Cloudera Manager.

Environment: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, Sqoop, Oozie,Impala, SQL, Java (jdk 1.6), Eclipse and Informatica 9.1.

Confidential - New York City, NY

JavaHadoop Developer

Responsibilities:

  • Hands on using log files and to copy them into HDFS using flume
  • Hands on writing Map Reduce code to make unstructured data as structured data and for inserting data into HBase from HDFS.
  • Experience in creating integration between Hive and HBase.
  • Experience indefining jobflows and wrote some simple to complex Map Reduce jobs as per the requirement.
  • Involved increating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Implemented business logic by writing Pig and Hive UDFs for some aggregative operations and to get the results from them.
  • Hands on experience in exporting the results into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Work closely with the business analysts to convert the Business Requirements into Technical Requirements and to make sure that the correct source table attributes are identified as per Dimensional Data Modeling(Fact Table Attributes and Dimensional Table Attributes).
  • Experienced withNoSQLdatabase and handled using the queries.
  • Design, develop and provide unit test data loading and data transformation programs related to Netezza Data Warehouses.
  • Monitored the health of MapReduce Programs which are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configuredHadoopMap Reduce, HDFS and Hive, Pig, Sqoop, Flume and Oozie on theHadoopcluster.
  • Cloudera Manger was used to monitor and manage the Hadoop Cluster.

Environment: Hadoop, Map Reducer, Cloudera Manager, HDFS, Hive, Pig, HBase, Sqoop, Flume, Oozie, SQL, Java (jdk 1.6), Eclipse.

Confidential - San Antonio, TX

JAVA Developer

Responsibilities:

  • Responsible for understanding the businessrequirement.
  • Worked with Business Analyst and helped representing the business domain details.
  • Actively involved in setting coding standards and writing related documentation.
  • Prepared the High and Low level design document.
  • Created Preferred Vehicle Web Service using JAXWS.
  • The web service is created using top down approach and tested using SOAP UI tool
  • This service is used at enterprise level and Confidential preferred vehicle are returned in different vehicle formats supported by specific LOB.
  • Used Hibernate 3.3.1 to interact with Data base.
  • Developed JSPs & Servlets to dynamically generate HTML and display data to client side.
  • An Admin tool is created in struts MVC design pattern to add preferred vehicle to Database.
  • Basic authentication is provided for preferred web service.
  • Designed Web Applications using MVC design pattern.
  • Developed Shell script to retrieve the vendor files dynamically and used Cron tab to execute these scripts periodically.
  • Designed the Batch Process for processing vendor data files using IBM Web sphere Application Server’s Task Manager Framework.
  • Developed and scheduled jobs using Cron.
  • Performed unit testing usingJUnitTesting Framework andLog4Jto monitor the error log.

Environment: IBM RAD, IBM Web Sphere App Server 7.0,Java/J2EE, Spring 3.0, JDK 1.5, Web services, SOAP, Servlets, JSP, ANT 1.6.x, Ajax, Hibernate 3.3.1,Custom tags

Confidential

JAVA Developer

Responsibilities:

  • Involved in various stages of Enhancements in the Application by doing the required analysis, development, and testing.
  • Prepared the High and Low level design document and Generating Digital Signature
  • For analysis and design of application created Use Cases, Class and Sequence Diagrams.
  • For the registration and validation of the enrolling customer developed logic and code.
  • Developed web-based user interfaces using struts frame work.
  • Handled Client side Validations used JavaScript and
  • Involved in integration of various Struts actions in the framework.
  • Used Validation Framework for Server side Validations
  • Created test cases for the Unit and Integration testing.
  • Front-end was integrated with Oracle database usingJDBCAPI through JDBC-ODBC Bridge driver at server side.

Environment: Java Servlets, JSP, Java Script, XML, HTML, UML, Apache Tomcat, JDBC, Oracle, SQL.

Confidential

JAVA Developer

Responsibilities:

  • Involved in Analysis of the requirements.
  • Prepared the High and Low level design document.
  • Created Creating UML artifacts - Use Cases, Class and Sequence Diagrams.
  • Developed Oracle PL/SQL procedures and functions.
  • Implemented Connection pool object for database connectivity.
  • Wrote hbm files and BO classes using Hibernate 3.3.1
  • Used XML parsers to parse incoming data and populating the database with the data
  • Designed the GUI screens using Struts and Configured log4j to debug the Application.
  • Involved in the development of test cases for the testing phase..
  • Performed End to end integration testing of online scenarios and unit testing usingJUnitTesting Framework.

Environment: Java, Windows XP, RAD 7.0.1, IBM Web Sphere Application Server 6.1, JSP, Struts, SQL, JDBC, Java Script, CSS, Log4j, Hibernate,HQL.

We'd love your feedback!