We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Dallas, TexaS

SUMMARY:

  • Comprehensive experience of 9+ years, with 4years in Hadoop and Spark Development along with 4 years of experience in Java/J2EE enterprise application design, development and maintenance.
  • Extensive experience implementing Big Data solutions using various distributions of Hadoop and its ecosystem tools.
  • Hands - on experience in installing, configuring and monitoring Hadoop clusters (on premise & cloudAWS).
  • HadoopDeveloper with 4+ years of working experience in designing and implementing complete end-to-endHadoopbased data analytical solutions using HDFS, MapReduce, Spark, Yarn, Kafka, PIG, HIVE, Sqoop, Storm, Flume, Oozie, Impala, HBase etc.
  • Good experience in creating data ingestion pipelines, data transformations, data management, data governance and real time streaming Confidential an enterprise level.
  • Profound experience in creating real time data streaming solutions using Apache Spark/Spark Streaming, Kafka.
  • Expertise developing MapReduce jobs to scrub, sort, filter, join and summarize data.
  • Experience developing PigLatin and HiveQL scripts for Data Analysis and ETL purposes and also extended the default functionality by writing User Defined Functions (UDFs), (UDAFs) for custom data specific processing.
  • Good Hands-on experience on full life cycle implementation using CDH (Cloudera) and HDP (Hartonworks Data Platform) distributions.
  • In depth understanding of Hadoop Architecture and its various components such as Resource Manager, Application Master, Name Node, Data Node, HBase design principles etc.,
  • Strong Knowledge on Architecture of Distributed systems and Parallel processing, In-depth understanding of MapReduce programing paradigm and Spark execution framework.
  • Profound understanding of Partitions and Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Experience in handling messaging services using Apache Kafka.
  • Experience with migrating data to and from RDBMS and unstructured sources into HDFS using Sqoop.
  • Experience in job workflow scheduling and monitoring tools like Oozie.
  • Worked on NoSQL databases including HBase, Cassandra and Mongo DB.
  • Strong experience in collecting and storing stream data like log data in HDFS using ApacheFlume.
  • Experience in working with Java HBaseAPI for ingestion processed data to Hbase tables.
  • Experience withOozie Workflow Engineto automate and parallelize Hadoop Map/Reduce, Hive and Pig jobs.
  • Proficient in using Cloudera Manager, an end-to-end tool to manage Hadoop operations in Cloudera Cluster.
  • Assisted in Cluster maintenance, Cluster Monitoring, Managing and Reviewing data backups and log files.
  • Experience writing Shell scripts in Linux OS and integrating them with other solutions.
  • Intensive work experience in developing enterprise solutions using Java, J2EE, Servlets, JSP, JDBC, Struts, Spring, Hibernate, JavaBeans, JSF, MVC.
  • Fluent and strong with the core Java concepts like I/O, Multi-threading, Exceptions, RegEx, Collections, Data-structures and Serialization.
  • Excellent problem-solving, analytical, communication, presentation and interpersonal skills that help me to be a core member of any team.
  • Experience mentoring and working with offshore and distributed teams.

TECHNICAL SKILLS:

Hadoop Ecosystem: HDFS, MapReduce, Sqoop, Flume, Hive, Pig, HBase, Impala, HUEZookeeper, Kafka, Cloudera Manager, Ambari, Storm, Solr, CouchbaseCassandra, Spark

Hadoop Distributions: Apache Hadoop, CDH3, CDH4, Hortonworks.

Languages/Technologies: Java, C, C++, HTML, CSS3, Python, J2EE, JSP, Servlets, HTMLXHTML, CSS, JavaScript, JQuery, AJAX, Scala

Scripting Languages: Java Script

Development Tools: Eclipse IDE, MS Visual Studio 2010, Amazon Web services, Open Stack

Version Control Tools: Git

Operating Systems: LINUX and UNIX, Windows 2008

RDBMS: Oracle 10g,11g, MySQL, PostgreSQL, Teradata, Greenplum

Application/Web Servers: Apache, Tomcat, MSIIS, Splunk

Oracle utilities: EXP, IMP, EXPDP, AWR, ADDM

WORK EXPERIENCE:

Confidential, Dallas, Texas

Sr. Hadoop Developer

Responsibilities:

  • Developed simple to complex MapReduce jobs using Java language for processing and validating the data.
  • Developed data pipeline using Sqoop, Spark, MapReduce, and Hive to ingest, transform and analyze operational data.
  • Developed Map Reduce and Spark jobs to summarize and transform+ raw data.
  • Implemented Spark using Scala and utilizing Data frames and Spark SQL API for faster processing of data.
  • Used Spark for interactive queries, processing of streaming data and integration with popular NoSQL database for huge volume of data.
  • Real time streaming the data using Spark with Kafka
  • Handled importing data from different data sources into HDFS using Sqoop and also performing transformations using Hive, Map Reduce and then loading data into HDFS.
  • Exported the analyzed data to the relational databases using Sqoop, to further visualize and generate reports for the BI team.
  • Collecting and aggregating large amounts of log data using Flume and staging data in HDFS for further analysis
  • Analyzed the data by performing Hive queries (Hive QL) and running Pig scripts (Pig Latin) to study customer behavior.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Developed Hive scripts in Hive QL to de-normalize and aggregate the data.
  • Created HBase tables and column families to store the user event data.
  • Scheduled and executed workflows in Oozie to run Hive and Pig jobs.
  • Used Impala to read, write and query the Hadoop data in Hive.

Environment: Hadoop, HDFS, HBase, Pig, Hive, MapReduce, Sqoop, Flume, ETL, REST, Java, Python, PL/SQL, Oracle 11g, Unix/Linux, CDH5.4

Confidential, Concord, California

Hadoop Developers.

Responsibilities:

  • Lead a team of 3 developers that built a scalable distributed data solution using Hadoop on a 30-node cluster using AWS cloud to run analysis on 25+ Terabytes of customer usage data.
  • Developed several new MapReduce and Spark programs to analyze and transform the data to uncover insights into the customer usage patterns.
  • Altered existing Scalaprograms to enhance performance and obtain partitioned results using Spark tool
  • Used MapReduce to Index the large amount of data to easily access specific records.
  • Performed ETL using Pig, Hive and MapReduce to transform transactional data to de-normalized form.
  • Configured periodic incremental imports of data from DB2 into HDFS using Sqoop.
  • Worked extensively with importing metadata into Hive using Scala and migrated existing tables and applications to work on Hive and AWS cloud.
  • Wrote Pig and HiveUDFs to analyze the complex data to find specific user behavior.
  • Used Kafka and Solr workflow engine to schedule multiple recurring and ad-hoc Hive and Pig jobs.
  • Created HBase tables to store various data formats coming from different portfolios.
  • TibcoJasperSoft was used for the embedding BI reports
  • Assisted the team responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, managing and reviewing data backups and Hadoop log files.
  • Conversion of Teradata, RDBMS was formulated in Hadoop backlog files.
  • Worked with teams in various locations nationwide and internationally to understand and accumulate data from different sources.
  • Worked with the testing teams to fix bugs and ensure smooth and error-free code.

Environment: Hadoop, MapReduce, HDFS, Hive, Java, SQL, Cloudera Manager, Pig, Sqoop, Oozie, HBase, ZooKeeper, PL/SQL, MySQL, DB2, Teradata.

Confidential, Denver, CO

Hadoop Developer

Responsibilities:

  • Responsible for developing efficient MapReduce on AWS cloud programs for more than 20 years’ worth of claim data to detect and separate fraudulent claims.
  • Uploaded and processed more than 30 terabytes of data from various structured and unstructured sources into HDFS (AWS cloud) using Sqoop and Flume.
  • Played a key-role is setting up a 40 node Hadoop cluster utilizing Apache MapReduce by working closely with the Hadoop Administration team.
  • Worked with the advanced analytics team to design fraud detection algorithms and then developed MapReduce programs to efficiently run the algorithm on the huge datasets.
  • Developed Java programs to perform data scrubbing for unstructured data.
  • Responsible for designing and managing the Sqoop jobs that uploaded the data from Oracle to HDFS and Hive.
  • Used Flume to collect the logs data with error messages across the cluster.
  • Designed and Maintained Oozie workflows to manage the flow of jobs in the cluster.
  • Played a key role in installation and configuration of the various Hadoop ecosystem tools such as, Hive, Pig, andHBase.
  • Actively updated the upper management with daily updates on the progress of project that include the classification levels that were achieved on the data.

Environment: Java, Hadoop, Hive, Pig, Sqoop, Flume, HBase, Oracle 10g, Teradata, Cassandra

Confidential, Atlanta, GA

Java Developer

Responsibilities:

  • Developed an end to end vertical slice for a JEE based application using popular frameworks Spring, Hibernate, JSF, Facelets, XHTML, Maven2, and AJAX by applying OO Design Concepts, JEE, and GoF Design Patterns and hosted on AWS cloud
  • Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database.
  • Tuned SQL statements, Hibernate Mapping, and Web Sphere application server to improve performance, and consequently met the SLAs.
  • Collected business requirements and wrote functional specifications and detailed design documents.
  • Detected and fixed transactional issues due to wrong exception handling and concurrency issues because of unsynchronized block of code.
  • Employed MVCStruts framework for application design.
  • Assisted in designing, building, and maintaining database to analyze life cycle of checking and debit transactions.
  • Used Web Sphere to develop JAX-RPC web services.
  • Developed Unit Test Cases, and used JUNIT for Unit Testing of the application.
  • Involved in the design team for designing the Java Process Flow architecture.
  • Used Web Sphere to develop JAX-RPC web services.
  • Worked with QA, Business and Architect to solve various Defects in to meet deadlines

Environment: Spring, Hibernate, Struts MVC, AJAX, Web Sphere, Maven2, Java, Java Script, JUnit, XHTML, HTML, DB2, SQL, UML, Oracle, Eclipse, Windows.

Confidential, Chattanooga, TN

Java/J2EE Developer

Responsibilities:

  • Effective role in the team by interacting with welfare business analyst/program specialists and transformed business requirements into System Requirements.
  • Developed analysis level documentation such as Use Case, Business Domain Model, Activity, Sequence and Class Diagrams.
  • Handling of design reviews and technical reviews with other project stake holders.
  • Implemented services using Core Java.
  • Developed and deployed UI layer logics of sites using JSP.
  • Spring MVC is used for implementation of business model logic.
  • Worked with StrutsMVC objects like action Servlet, controllers, and validators, web application context, Handler Mapping, message resource bundles, and JNDIfor look-up for J2EE components.
  • Developed dynamic JSP pages with Struts.
  • Employed built-in/custom interceptors, and validators of Struts.
  • Developed the XML data object to generate the PDF documents, and reports.
  • Employed Hibernate, DAO, and JDBC for data retrieval and medications from database.
  • Messaging and interaction of web services is done using SOAP.
  • Developed Junittest cases for Unit Test cases and as well as system, and user test scenarios

Environment: Struts, Hibernate, Spring MVC, SOAP, WSDL, Web Logic, Java, JDBC, Java Script, Servlets, JSP, JUnit, XML, UML, Eclipse, Windows.

Confidential

Jr. Java Developer

Responsibilities:

  • Involved in designing the Project Structure, System Design and every phase in the project.
  • Responsible for developing platform related logic and resource classes, controller classes to access the domain and service classes.
  • Involved in Technical Discussions, Design, and Workflow.
  • Participate in the Requirement Gathering and Analysis.
  • Employed JAXB to unmarshalXML into Java Objects.
  • Developed Unit Testing cases using JUnit Framework.
  • Implemented the data access using Hibernate and wrote the domain classes to generate the Database Tables.
  • Involved in implementation of view pages based on XML attributes using normal Java classes.
  • Involved in integration of App Builder and UI modules with the platform.

Environment: Hibernate, Java, JAXB, JUnit, XML, UML, Oracle11g, Eclipse, Windows XP.

Java/J2EE Consultant

Confidential

Responsibilities:

  • Involved in designing and developing modules Confidential both Client and Server Side.
  • Developed the UI using JSP, JavaScript and HTML.
  • Responsible for validating the data Confidential the client side using JavaScript.
  • Involved in developing Server Side Java Servlets to capture the posted form data and process it on the Server side.
  • Developed web components using JSP, Servlets and JDBC.
  • Designed the controller using Servlets.
  • Accessed backend database Oracle using JDBC.
  • Developed and wrote UNIX Shell scripts to automate various tasks.
  • Developed user and technical documentation.

Environment: Java, Servlets, JSP, JavaScript, JDBC, Unix Shell scripting, HTML, Eclipse, Oracle 9i, WebLogic, WindowsNT.

We'd love your feedback!