We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Cleveland, OH

PROFESSIONAL SUMMARY:

  • Big Data professional with 7 years of combined experience in the fields of Java and Hadoop.
  • Ability to work in high pressure kind of environments and achieving organization expectations & Strong analytical, problem solving and leadership skills.
  • Hands on experience on major components in Hadoop Ecosystem like Hadoop Map Reduce, HDFS, YARN, Cassandra, IMPALA, Hive, Pig, HBase, Sqoop, Oozie, Flume.
  • Good Knowledge on Map Reduce design patterns& Extensively worked on Hive, Pig,HBase for performing data analysis.
  • Strong understanding in distributed systems, RDBMS, large scale & small scale non - relational Data stores, No-SQL, Map-Reduce systems, Database performance, Data modeling and Multi-terabyte data warehouses.
  • Hands on experience in Hadoop Distributed file systems, Hadoop framework and Parallel processing implementation.
  • Strong command over developing applications using Java, RDBMS, Linux/Unix shell scripting and Linux internals.
  • Good Understanding in writing UDF's and Map-Reduce in Java for pig and hive.
  • Strategic knowledge in cleansing & analyzing data using Hive QL, Pig Latin and custom Map-Reduce program in java.
  • Hands on experience on Data importing and exporting tools like Sqoop and flume.
  • Comprehensively trained on Big Data processing and Hadoop Development.
  • Good experience in developing software and designing E-commerce applications, highly distributed applications using Java, J2EE, Servlets, JSP, Java Beans, EJB, Struts, Springs, JDBC, Apache Tomcat server, JavaScript, HTML5, XML, SQL.
  • Distributed Object Oriented Component analysis and Design per industry leading J2EE platform.
  • Wide range experience working in Oracle, DB2, SQL Server and My-SQL Databases.
  • Hands on experience in Object Oriented Analysis Design (OOAD) and developing a software using UML Methodology.
  • Good skills on J2EE design patterns and core Java design patterns.
  • Experience in implementing large scale applications using leading technologies in Business Intelligence.
  • Strong command on designing software which enables a system which is secured and enforce authentication, authorization, confidentiality, Data integrity and availability.
  • Excellent working knowledge of System Development Life Cycle (SDLC) and Software Testing Life Cycle (STLC) and Defect Life Cycle.
  • Setting up automated 24x7 monitoring and escalation infrastructure for Hadoop clusters using Nagios and Ganglia.
  • Experience in benchmarking, backup and disaster recovery of Name node metadata.
  • Performed major and minor upgrades, commissioning and decommissioning of data nodes on Hadoop cluster.
  • Team player and self-starter with excellent communication skills and proven abilities to finish tasks on time. Hadoop and Big Data:
  • Good knowledge on Hadoop Architecture and HDFS.
  • Strong Command on different Data Loading Techniques and Data Analysis using Java Map-Reduce programs, Hive and Pig.
  • Firm understanding of Advance Map-Reduce and perform operations like Map-side joins, reduce side joins, MRUnit Testing Framework, Disturbed cache and Sequence Files.
  • Knowledge in writing custom UDF's and UDAF's for extending Hive and Pig core components.
  • Importing and exporting the data using Sqoop from RDBMS to HDFS and HDFS to RDBMS.
  • Knowledge in executing Flume to load the long data from multiple sources directly into HDFS.
  • Procedural knowledge in cleansing and analyzing’s data using Hive QL, Pig Latin and custom Map.

TECHNICAL SKILLS:

Technology: Hadoop Eco System, J2EE,J2SE,Oracle

Big Data Ecosystem: HDFS, Map Reduce, Oozie, Hive, Pig, Sqoop, Flume,Zookeeper and HBase

ETL Tools: Knowledge on Infomatica& IBM Data stage 8.1

Data Modeling: StarSchema and Snowflake schema

Reporting Tools: Cogon s, Report studio, Query Studio, Analysis Studio,Event Studio and Framework Manager

Servers: Application Servers (WAS, Tomcat), Web Servers (IISS6, IHS)

Operating Systems: Windows 2003 Enterprise Server, XP,2000, UNIX.

Databases: SQL Server2005, SQL 2008, Oracle9i/10g, DB2, MS Access2003, Teradata

Languages: C++,JSE/Servets,Struts,spring,HTML,Python,PHP,JavaScript,JQuery,WebServices.

Application Servers: Apache Tomcat server, Apache HTTP webserver

PROFESSIONAL EXPERIENCE:

Confidential, Cleveland, OH

Sr. Hadoop Developer

Responsibilities:

  • Developed MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Used Bash Shell Scripting, Sqoop, AVRO, Hive, HDP, Redshift, Pig, Java, Map/Reduce daily to develop ETL, batch processing, and data storage functionality.
  • Responsible for developing data pipeline using Flume, Sqoop and Pig to extract the data from weblogs and store in HDFS.
  • Worked on NoSQL databases including Hbase and Cassandra.
  • Worked on loading all tables from the reference source database schema through Sqoop.
  • Worked on designed, coded and configured server side J2EE components like JSP, AWS and JAVA.
  • Collected data from different databases ( i.e. Oracle, My Sql) to Hadoop.
  • Used Oozie and Zookeeper for workflow scheduling and monitoring.
  • Worked on Designing and Developing ETL Workflows using Java for processing data in HDFS/Hbase using Oozie.
  • Experienced in managing and reviewing Hadoop log files.
  • Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
  • Involved in loading and transforming large sets of structured, semi structured and unstructured data from relational databases into HDFS using Sqoop imports.
  • Developed Sqoop scripts to import export data from relational sources and handled incremental loading on the customer, transaction data by date.
  • Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
  • Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Worked on partitioning HIVE tables and running the scripts in parallel to reduce run-time of the scripts.
  • Worked on Data Serialization formats for converting Complex objects into sequence bits by using AVRO, PARQUET, JSON, CSV formats.
  • Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
  • Created Hive tables, loaded data and wrote Hive queries that run within the map.
  • Implemented business logic by writing Pig UDF's in Java and used various UDFs from Piggybanks and other sources.
  • Used OOZIE Operational Services for batch processing and scheduling workflows dynamically.
  • Extensively worked on creating End-End data pipeline orchestration using Oozie.
  • Populated HDFS and Cassandra with huge amounts of data using Apache Kafka.
  • Worked in a language agnostic environment with exposure to multiple web platforms such as AWS, databases like Cassandra.
  • Processed the source data to structured data and store in NoSQL database Cassandra.
  • Extracted files fromMySQLthroughSqoopand placed in HDFS and processed.
  • Successfully loaded files to Hive and HDFS fromCassandra.Processed the source data to structured data and store in NoSQL databaseCassandra.Created alter, insert and delete queries involving lists, sets and maps inDataStaxCassandra.Worked in a language agnostic environment with exposure to multiple web platforms such as AWS, databases like Cassandra.
  • Cassandra through Java services.
  • Responsible for continuous monitoring and managing Elastic MapReduce cluster through AWS console.

Environment: Map Reduce, HDFS, Hive, Pig, HBase, SQL, Sqoop, Flume, Oozie, Apache Kafka, Zookeeper, J2EE, Eclipse, Cassandra.

Confidential, Green bay, WI

Hadoop Developer

Responsibilities:

  • Responsible for loading customer's data and event logs into HBase using Java API.
  • Created HBase tables to store variable data formats of input data coming from different portfolios.
  • Involved in adding huge volumes of data in rows and columns to store data in HBase.
  • Responsible for architecting Hadoop clusters with CDH4 on CentOS, managing with Cloudera Manager.
  • Involved in initiating and successfully completing Proof of Concept on FLUME for Pre-Processing, Increased
  • Reliability and Ease of Scalability over traditional MSMQ.
  • Use Flume to collect the log data from different resources and transfer the data type to hive tables using different Serde to store in JSON, XML and Sequence file formats.
  • Used Hive to find correlations between customer's browser logs in different sites and analyzed them to build risk profile for such sites.
  • End-to-end performance tuning of Hadoop clusters and Hadoop Map/Reduce routines against very large data sets.
  • Implemented Hadoop distribution using Horton Works.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
  • Created User accounts and given the users the access to the Hadoop Cluster
  • Implemented the secure authentication for the Hadoop Cluster using Kerberos Authentication protocol.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed the Pig UDF'S to pre-process the data for analysis.
  • Familiarity with a NoSQL database such as Mongo DB Solr.
  • Experience in working with various kinds of data sources such as Mongo DB Solar and Oracle.
  • Successfully loaded files to Hive and HDFS from Mongo DB Solr.
  • Experience in managing development time, bug tracking, project releases, development speed, release forecast, scheduling and many more. Using a custom framework of Nodes and Mongo DB to take care of the back-end calls with a lightning fast speed.
  • Monitored Hadoop cluster job performance and performed capacity planning and managed nodes on
  • Hadoop cluster.
  • Responsible for using Cloudera Manager, an end to end tool to manage Hadoop operations.

Environment: Hadoop, Big Data, HDFS, Pig, Hive, MapReduce, Sqoop, Cloudera manager, LINUX, CDH4, FLUME, HBase, Pig, Hive.

Confidential

Hadoop Developer

Responsibilities:

  • Configured finely tuned the Cloudera Hadoop (CDH3 & CDH4) and other Hadoop ecosystem tools like Sqoop, Hive, Pig, Hbase, Zookeeper and Flume.
  • As a teammember was assigned to write MapReduce jobs using Java API.
  • In the process of handling large sets of data extracted, loaded and transformed different structured, unstructured and semi-structured data.
  • By using Sqoop, imported and exported data into HDFS and Hive.
  • Installed and configured Hive and also wrote Hive UDFs in Java.
  • Created several Hive tables, loaded with data and wrote Hive Queries in order to run internally in MapReduce.
  • Actively involved in loading data from UNIX file system to HDFS.
  • Involved in developing Hive queries in order to pre-process the data for analysis by imposing read only structure on the streamed data.
  • Through Oozie, created workflows for automation.
  • Also automated all jobs so that data can be pulled from FTP server and loaded into Hive tables using Oozie Workflows.
  • As an active team member supported MapReduce programs which were running on the cluster.

Environment: Hadoop Framework, MapReduce, Hive, Sqoop, Flume, Oozie, Java(JDK1.6), UNIX Shell, Oracle 10g.

Confidential

Java/J2EE Developer

Responsibilities:

  • Developed the application using Struts Framework that leverages classical Model View Layer (MVC)
  • Architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used
  • Gathered business requirements and wrote functional specifications and detailed design documents
  • Extensively used Core Java, Servlets, JSP and XML
  • Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i.
  • Implemented Enterprise Logging service using JMS and apache CXF.
  • Developed Unit Test Cases, and used JUNIT for unit testing of the application
  • Implemented Framework Component to consume ELS service.
  • Involved in designing user screens and validations using HTML, jQuery, Ext JS and JSP as per user requirements.
  • Implemented JMS producer and Consumer using Mule ESB.
  • Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
  • Designed Low Level design documents for ELS Service.
  • Closely worked with QA, Business and Architect to solve various Defects in quick and fast to meet deadlines .
  • Worked on Python scripting.

Environment: Java, Spring core, JMS Web services, JMS, JDK, SVN, Maven, Mule ESB Mule, Junit, WAS7, JQuery, Ajax, SAX.

Confidential

Java Developer

Responsibilities:

  • Designed and developed Web Services usingJava/J2EE in WebLogic environment. Developed web pages usingJavaServlet, JSP, CSS,JavaScript, DHTML, HTML5, and HTML. Added extensive Struts validation.
  • Involve in the Analysis, Design, and Development and testing of business requirements.
  • Developed business logic in JAVA/J2EE technology.
  • Implemented business logic and generated WSDL for those web services using SOAP.
  • Worked on Developing JSP pages.
  • Implemented Struts Framework.
  • Developed Business Logic using Java/J2EE
  • Modified Stored Procedures in MYSQL Database.
  • Developed the application using Spring Web MVC framework.
  • Worked with Spring Configuration files to add new content to the website.
  • Worked on the Spring DAO module and ORM using Hibernate. Used Hibernate Template and Hibernate DaoSupport for Spring-Hibernate Communication.
  • Configured Association Mappings such as one-one and one-many in Hibernate
  • Worked with JavaScript calls as the Search is triggered through JS calls when a Search key is entered in the Search window
  • Worked on analyzing other Search engines to make use of best practices.
  • Collaborated with the Business team to fix defects.
  • Worked on XML, XSL and XHTML files.
  • Interacted with project management to understand, learn and to perform analysis of the Search Techniques.
  • Used Ivy for dependency management.
  • As part of the team to develop and maintain an advanced search engine, would be able to attain expertise on a variety of new software technologies.

We'd love your feedback!