We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

2.00/5 (Submit Your Rating)

San Diego, CA

PROFESSIONAL SUMMARY:

  • Around 8 years of IT experience in design, development, testing, maintenance, and integration.
  • Experience in Hadoop Ecosystem (MapReduce, Impala, Zookeeper, Hive, Pig, Flume, Sqoop, Spark, HBase, Oozie, Yarn, Kafka) as Developer.
  • Hands on experience in installing, configuring, and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, PIG, HIVE, HBASE, ZOOKEEPER, Sqoop, HUE, JSON.
  • Good knowledge of Hadoop Development and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and MapReduce and YARN.
  • Experience in installation, configuration, Management, supporting and monitoring Hadoop cluster using various distributions such as Cloudera and Horton works distribution.
  • Good hands on expertise with Hadoop admin roles like cluster development and configuration of all services through AMBARI.
  • Experience in managing and reviewing Hadoop log files.
  • Proficient in writing Ad - hoc queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL.
  • Experience in creating UDF's, UDAF's for Hive and Pig.
  • Experience in partitioning the Big Data according the business requirements using Hive Indexing, partitioning and Bucketing.
  • Working with data extraction, transformation and load in Hive, Pig and HBase.
  • Designed HIVE queries & Pig scripts to perform data analysis, data transfer and table design.
  • Experience in manipulating the streaming data to clusters through Flume.
  • Knowledge in converting Hive or SQL queries into Spark transformations using Python and Scala.
  • Experienced with open source NoSQL technologies such as HBase, Cassandra, MongoDB.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Good Knowledge in Spark and Scala.
  • Knowledge in core Java concepts like OOP, JAVA SWING, JDBC, JMS, Multi-Threading, Junit, and advanced Java concepts like JSP, Servlets, Struts, HTML, XML, CSS, Hibernate, AJAX, SVN, Java Beans.
  • Involved extensively in designing/developing web based applications using HTML and MVC design patterns.
  • Solid experience in communicating with all levels of enterprise.
  • Adept in handling the team in untoward situations and capable of sailing the team to deliver the quality output.
  • Highly motivated and versatile team player with the ability to work independently & adapt quickly to new emerging technologies.

TECHNICAL SKILLS:

Hadoop Ecosystem: MapReduce, Hive, Yarn, Pig, Flume, Scoop, Oozie, HBase, Spark, Cloudera Manager, Zookeeper, Horton Works Distribution

Java Technologies: Java, J2EE, Servlets, Custom Tag Libraries, EJB, JDBC, JMS, JSF, RMI, JNDI, JSTL.

Web Technologies: HTML/DHTML, JSPs, CSS, AJAX, JavaScript

Programming Languages: Java, Scala, Python, C, C++, SQL

Tools /Software: Apache Axis, Log4j, Ant, Toad, Altova XMLSpy, Maven, UML, Rational Rose, MVC architecture, Unix Shell Script.

Databases: Oracle, MySQL, SQL Server …

NoSQL Databases: HBase, Cassandra, MongoDB

Methodologies: Agile, SCRUM

Application/Web Servers: WebLogic, IBM WebSphere, Apache Tomcat, JBoss

Development tools: RAD, Eclipse, Net Beans

Operating Systems: Windows Vista/XP, Linux (Cent OS, RHEL, Ubuntu), UNIX

PROFESSIONAL EXPERIENCE:

Sr. Hadoop Developer

Confidential - San Diego, CA

Responsibilities:

  • Involved in Requirement Analysis, Development, and Documentation.
  • Responsible for planning, organizing, and implementation of complex business solutions, producing deliverables within stipulated time.
  • Assist with configuring Big Data solutions to meet client requirements.
  • Participated in Design reviews, Test Case reviews, and Production Support readiness reviews for new releases, provide inputs for Go/No Go decision.
  • Developed complex MapReduce programs in Java for Data Analysis on different data formats.
  • Developed MapReduce programs to clean and aggregate the data.
  • Developed complex MapReduce jobs, PIG scripts and HIVE queries.
  • Analyzed data using HIVE and PIG queries.
  • Worked exclusively on HBase, Sqoop and HIVE.
  • Extracted the data from the flat files and other RDBMS databases into staging area and populated onto HDFS.
  • Used PIG predefined functions to convert the fixed width file to delimited file.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Python and Scala.
  • Developed Sqoop scripts to handle change data capture for processing incremental records between new arrived and existing data in RDBMS tables.
  • Involved in resolving performance issues in Pig and Hive with understanding of Map Reduce physical plan execution and using debugging commands to run code in optimized way.
  • Developed Kafka producer and consumers, Cassandra clients and Spark along with components on HDFS, Hive.
  • Automated different workflows, which are initiated manually with Python scripts and Unix shell scripting.
  • Ingested data from RDBMS and performed data transformations, and then export the transformed data to Cassandra as per the business requirement
  • Wrote Flume configuration files for importing streaming log data into HBase with Flume.
  • Developed a data pipeline using Kafka and Strom to store data into HDFS.
  • Implemented Spark RDD Transformations and Actions.
  • Developed spark scripts by using Scala shell commands as per the requirement.
  • Experience in performance tuning a Cassandra cluster to optimize it for writes and reads.
  • Used Spark API over Hadoop YARN to perform analytics on data in Hive.
  • Generated queries using SQL to check for consistency of the data in the tables.
  • Wrote and tested Python scripts to create new data files for Linux server configuration using a Python template tool.

Environment: Hadoop, Yarn, HDFS, CDH, Flume, Python, Sqoop, Hive, Pig, Spark, Spark-SQL, Kafka, HBase, Java, Eclipse, Maven, GIT, JSON, Cassandra, Linux, Avro.

Hadoop Developer

Confidential - Louisville, KY

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing.
  • Involved in data extraction from distributed RDBMS like Teradata and Oracle.
  • Working extensively on HIVE, SQOOP, MAPREDUCE, SHELL, PIG, and PYTHON.
  • Using SQOOP to move the structured data from MySQL to HDFS, HIVE, PIG and HBase.
  • Involved in loading data from UNIX file system to HDFS.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Used Oozie to manage the Hadoop jobs.
  • Using Oozie to schedule the workflows to perform shell action and hive actions.
  • Used CDH3 and CDH4 distributions for development and deployment.
  • Implemented Partitioning, Dynamic Partitioning, and Bucketing in HIVE.
  • Develop HiveQL scripts to perform the incremental loads.
  • Exported the result set from HIVE to MySQL using Shell scripts.
  • Using PIG predefined functions to convert the fixed width file to delimited file.
  • Installed and configured Pig for ETL jobs.
  • Used agent E2E Chain for reliability and failover in flume.
  • Experience in using Flume to stream data into HDFS - from various sources
  • Experience in writing HIVE JOIN Queries.
  • Is Responsible for the implementation and ongoing administration of Big Data platforms on Hortonworks.
  • Worked with the Hortonworks support for resolving the issues.
  • Using HIVE join queries to join multiple tables of a source system and load them into Elastic Search Tables.
  • Analyzed the SQL scripts and designed the solution to implement using Scala.
  • Involved in data migration from one cluster to another cluster.
  • Experience in writing aggression logics in different combinations to perform complex data analytics to the business needs.

Environment: Hadoop, Map Reduce, Hive, Hortonworks, Pig, Sqoop, HBase (NoSQL database), Java 1.6, CentOS, and UNIX Shell Scripting.

Hadoop Developer

Confidential - Walla Walla, WA

Responsibilities:

  • As a developer am responsible to understand, document, develop and deliver the business requirements.
  • Involved in requirement sessions along with the business stakeholders.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from LINUX file system to HDFS.
  • Created HBase tables to store variable data formats of PII data coming from different portfolios.
  • Worked on tuning the performance of Pig queries.
  • Responsible to manage data coming from different sources.
  • Involved in loading data from UNIX file system to HDFS.
  • Involved in loading and transforming large sets of structured, semi structured, and unstructured data from relational databases into HDFS using Sqoop imports.
  • Developed Sqoop scripts to import and export data from and to relational sources by handling incremental data loading on the customer transaction data by date.
  • Responsible for creating complex tables using Hive.
  • Worked on data transformation and extraction for normalization.
  • Created partitioned tables in Hive for best performance and faster querying.
  • Developed JavaAPIs for invocation in Pig Scripts to solve complex problems.
  • Worked on Data Serialization formats for converting Complex objects into sequence bits by using AVRO, JSON, CSV formats.
  • Experience in writing the business logics for defining the DAT, CSV files for MapReduce.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Working in Agile environment with 3 weeks sprint, involving in grooming meeting, sprint planning meeting, retrospective meeting and daily stand up meeting.

Environment: Horton works, Map Reduce, HBase, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Zookeeper, Oozie, Java.

Java/J2EE Developer

Confidential

Responsibilities:

  • Involved in Analyzing, preparing technical design specification documents as per the Requirements,
  • Involved in study of User Requirement Specification, Communicated with Business Analysts to resolve ambiguity in Requirements document. Handled performance issues and worked on background job, which executes huge records.
  • Wrote SQL queries, stored procedures, and triggers to perform back-end database operations.
  • Developed the application using Struts Framework that leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used.
  • Developed nightly batch jobs which involved interfacing with external third party state agencies.
  • Involved in configuration of Spring MVC and Integration with Hibernate.
  • Normalized Oracle database, conforming to design concepts and best practices.
  • Used JDBC to connect to backend databases, Oracle and SQL Server 2005.
  • Used Core java and object-oriented concepts.
  • Developed JavaScript behavior code for user interaction.
  • Created database program in SQL server to manipulate data accumulated by internet transactions.
  • Wrote Servlets class to generate dynamic HTML pages.
  • Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
  • Developed the XML Schema and Web services for the data maintenance and structures Wrote test cases in JUnit for unit testing of classes.
  • Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
  • Debugged the application using Firebug to traverse the documents.
  • Involved in developing web pages using HTML and JSP.
  • Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects.
  • Involved in writing SQL Queries, Stored Procedures and used JDBC for database connectivity with MySQL Server.
  • Developed the presentation layer using CSS and HTML taken from bootstrap to develop for browsers.

Environment: Java, XML, HTML, JavaScript, JDBC, CSS, SQL, PL/SQL, XML, Web MVC, Eclipse, Ajax, JQuery, spring with Hibernate, Active MQ, Jasper Reports, JIRA, Ant as build tool and My SQL and Apache Tomcat

JAVA DEVELOPER

Confidential

Responsibilities:

  • Involved in complete requirement analysis, design, coding, and testing phases of the project.
  • Participated in JAD meetings to gather the requirements and understand the End Users System.
  • Developed user interfaces using JSP, HTML, XML, and JavaScript.
  • Created Stored Procedures & Functions. Used JDBC to process database calls for DB2/AS400 and SQL Server databases.
  • Developed the code which will create XML files and Flat files with the data retrieved from Databases and XML files.
  • Created Data sources and Helper classes which will be utilized by all the interfaces to access the data and manipulate the data.
  • Developed the interfaces using Eclipse and JBoss Involved in integrated testing, Bug fixing and in Production Support.
  • Designed user-interface and checking validations using JavaScript.
  • Implemented JDBC for mapping an object-oriented domain model to a traditional relational database.
  • Created Stored Procedures to manipulate the database and to apply the business logic according to the user's specifications.
  • Used CVS for maintaining the Source Code.

Environment: J2EE, Java/JDK, JDBC, JSP, Servlets, JavaScript, EJB, JNDI, JavaBeans, XML, XSLT, Oracle 9i, Eclipse, SQL

JAVA DEVELOPER

Confidential

Responsibilities:

  • Involved in analysis of the specifications from the client and actively participated in SRS Documentation.
  • Developed Servlets and JDBC were used in retrieving data.
  • Designed and developed dynamic Web pages using HTML and JSP.
  • Implemented Object Relational mapping in the persistence layer using Hibernate Framework in conjunction with Spring Functionality.
  • Involved in planning process of iterations under the Agile Scrum methodology.
  • Analyzed and designed a scalable system based on Object oriented concepts and the various J2EE design patterns. Implementation of Spring MVC Architecture.
  • Involved in writing PL/SQL, SQL queries.
  • Implemented web services using REST, JSON and XML.
  • Developed entire application in Spring tool suite IDE.
  • Validated the user input using Struts Validation Framework.
  • Client side validations were implemented using JavaScript.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Implemented the mechanism of logging and debugging with Log4j
  • Involved in testing the Business Logic layer and Data Access layer using JUnit.
  • Used Oracle DB for writing SQL scripts, PL/SQL code for procedures and functions.
  • Wrote JUnit test cases to test the functionality of each method in the DAO layer. Configured and deployed the WebSphere application Server.
  • Prepared technical reports and documentation manuals for efficient program development.

Environment: Java 1.5, J2EE, WebLogic, Struts 1.2.9, Spring 2.5, PL/SQL, Hibernate 3.0, JSP 2.1, JavaScript.

We'd love your feedback!