We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

0/5 (Submit Your Rating)

Dallas, TX

SUMMARY

  • Over 6 years of programming and software development experience with skills in data analysis, design and development, testing and deployment of software systems from development stage to production stage with giving emphasis on Object oriented paradigm.
  • More than 3+ years experience with the tools in Hadoop Ecosystem including Pig, Hive, HDFS, MapReduce, Sqoop, Oozie, Zookeeper.
  • Excellent knowledge on Hadoop Architecture and ecosystems such as HDFS, Name Node, Data Node and MapReduce programming paradigm.
  • Experience in migrating the data using Sqoop from HDFS to Relational Database System and vice - versa according to client’s requirement.
  • Strong experience in collecting and storing log data in HDFS using Apache Flume.
  • Experience in analyzing data using HiveQL, Pig Latin, HBase and custom MapReduce programs in Java.
  • Experience in extending Hive and Pig core functionalities by writing custom UDFs.
  • Experienced in job workflow scheduling and monitoring tools like Oozie and Zookeeper.
  • Good understanding of NoSQL databases and hands on work experience in writing applications on NoSQL databases like HBase and Cassandra.
  • Knowledge of manipulating/analyzing large datasets and finding patterns and insights within structured and unstructured data.
  • Ability to analyze and performance tune a Hadoop cluster.
  • Expert in Database, RDBMS concepts and using MS Access, MS SQL Server and Oracle 10g.
  • Expert in managing/working in multiple projects simultaneously. Worked on medium to large-scale enterprise software projects.
  • Used RDBMS concepts for the manipulation of the data and to validate the results.
  • Strong Experience in all the phases of software development life cycle (SDLC) including requirements gathering, analysis, design implementation and support.
  • Experienced in developing GUI automation scripts for Web based applications, Windows based applications, Point of Sales (POS) applications using Selenium driver with JAVA.
  • Experienced in developing Methods, Procedures and Utilities as part of the Automation Framework.
  • Experienced in coordinating with offshore teams.
  • Very good analytical and problem solving skills, very effective as an individual and as a team player, ability to perform multiple roles.
  • Experienced in working with different methodologies like Waterfall and Agile.

TECHNICAL SKILLS

Operating Systems: Windows XP/2000/Vista/7, Linux, Unix (Sun Solaris, Red Hat)

Hadoop Ecosystem: Apache Hadoop (MapReduce and HDFS), Pig (Pig Latin), Hive, HBase, Sqoop, Spark, Flume, Oozie, Zookeeper, CDH, Horton Works

Languages: C, C++, Java, JavaScript, Python, SQL, PL/SQL

Scripting: JavaScript, UNIX Shell Scripting, AJAX, HTML, DHTML, VBScript

RDBMS: ORACLE 8i/9i/10g, Sybase, MS Access, MS SQL Server, Toad

Web/Application Servers: REST, WISDL, SOAP Web Services.

Tools: Eclipse IDE, Application Lifecycle Management (ALM), Maven, JIRA, Jenkins, Firebug and Rally

PROFESSIONAL EXPERIENCE

Confidential, Dallas, TX

Sr. Hadoop Developer

Responsibilities:

  • Involved in Requirement gathering, Business Analysis and translated business requirements into Technical design in Hadoop and Big Data.
  • Designed and developed multiple MapReduce Jobs in JAVA for data cleaning and preprocessing.
  • Importing and exporting data into HDFS from database and vice versa using Sqoop.
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Involved in creating Hive tables, loading with data and writing hive queries that will run internally in map reduce way.
  • Implemented business logic by writing UDF’s in Java and used UDF’s from Piggybanks and other sources.
  • Used Pig as an ETL tool to do Transformations, even joins and some pre-aggregations before storing data into HDFS.
  • Worked on analyzing Hadoop clusters and different Big Data analytics tools including Pig, HBase database and SQOOP.
  • Responsible for operational support of Production system.
  • Involved in creating workflow to run multiple Hive and Pig Jobs, which run independently with time and data availability.
  • Installed and Configured the Spark cluster integrating with Hadoop cluster to run the programs faster than Hadoop MapReduce in memory.
  • Involved in collection and aggregating large amounts of web log data from different sources such as web servers, mobile and network devices using Apache Flume and stored the data into HDFS for analysis.
  • Skilled in handling and reviewingHadoop log files.
  • Used Oozie and Zookeeper operational services for coordinating cluster and scheduling workflows.
  • Developed Pig scripts to generate MapReduce jobs and performed ETL procedures on HDFS data.
  • Involved in UML, Package, State Diagrams and Class.
  • Responsible in loading and transforming huge sets of structured, semi structured and unstructured data.

Environment: Agile SCRUM, Apache Hadoop, CDH, Map Reduce, HDFS, Hive, Java, SQL, PIG, Zookeeper, Oozie, Apache Spark, Java (jdk1.6), Oracle 11g/10g, MySQL, Windows 7, UNIX, Sqoop, Oozie.

Confidential, Grand Rapids, MI

Hadoop Consultant

Responsibilities:

  • Worked on analyzing Hadoop cluster and different big data analytic tools including Hive, Pig, Hbase data base and Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Configured nine nodes CDH4 Hadoop cluster on Red hat LINUX.
  • Involved in loading data from LINUX/UNIX file system to HDFS.
  • Support data analysis projects using Elastic MapReduce on the Amazon Web Services (AWS) cloud.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries.
  • Developed Simple to complex MapReduce Jobs using Hive and Pig.
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Responsible to manage data coming from different sources.
  • Load and transform large sets of structured, semi structured and unstructured data Cluster coordination services through Zookeeper.
  • Experience in managing and reviewing Hadoop log files.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and Troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Experience in Automation Testing, Software Development Life Cycle (SDLC) using the Waterfall Model and good understanding of Agile Methodology.

Environment: Java 6, Eclipse, Oracle 10g, HBase, Linux,, MapReduce, HDFS, Hive, Java (JDK 1.6), Hadoop Distribution of Hortonworks, Amazon web services, EC2, Data Stax, Oracle 11g / 10g, PL/SQL, Windows NT, UNIX Shell.

Confidential, SC

Sr. Java Developer

Responsibilities:

  • Involved in developing the front end and processing the business logic using MVC based Struts framework and Oracle Application Development Framework.
  • Involved in the design of the application.
  • Implemented the spring dependency injection of the Database helper instance to the action objects.
  • Created UML diagrams based on the business requirements.
  • Have prepared UML Sequence diagrams, Class Diagram, Activity diagrams to be incorporated in the Technical Design document using Microsoft VISIO tool.
  • Designed front-end using HTML, JavaScript, CSS, and JSP.
  • Hibernate frame work is used in persistence layer for mapping an object-oriented domain model to a relational database (oracle).
  • Designed enterprise level applications on XML schema, SOAP and WSDL technologies.
  • Used XSLT for arranging the Data (XML Data) in the order required by the Client.
  • Used AJAX in user validation and suggestion list and to display dialog boxes.
  • Developed Action Form and Action Classes in struts framework.
  • Designed Stateless Session Beans for implementing the Business logic.
  • Used JMS, MDB for server side messaging.
  • Used Web Services for getting Customer information and the plans.
  • Designed stored procedures and triggers along with Performance tuning for SQL.
  • Developed modules that integrate with web services that provide global information support such as customer and account information.
  • Wrote JUnit test cases to test the functionality of each method in the DAO classes developed.
  • Configuration and deployment of the web application using Web Logic.
  • Used VSS for version control and source code management.
  • Used UNIX Shell Scripting to update the databases and Start up the application server.
  • Prepared technical reports and documentation manuals during the program development.

Environment: JSP, JSF, Struts, SOA, HTML, XSLT, AJAX, XML, JQuery, CSS, XML, EJB, JMS, SQL, PL/SQL, Web Logic, JUnit, UNIX, Shell Script, SAX Parser, Web Services, SOAP UI.

We'd love your feedback!