We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Kansas, MO

SUMMARY:

  • Over 7 years of experience in software development with experience in phases of Hadoop and HDFS development.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Hands on experience in writing MapReduce jobs in Java, Pig and Python.
  • Experienced on major Hadoop ecosystem’s projects such as PIG, HIVE and HBASE.
  • Good working experience using Sqoop to import data into HDFS from RDBMS and vice - versa.
  • Good knowledge in using job scheduling and monitoring tools like Oozie and ZooKeeper.
  • Knowledge of NoSQL databases such as HBase, and MongoDB.
  • Good understanding/knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Have a good understanding to ETL concepts.
  • Hands on experience in installing, configuring and using ecosystem components like HadoopMapReduce, HDFS, Hbase, ZooKeeper, Oozie, Flume, Sqoop, Pig & Hive.
  • Experience in database development using SQL and PL/SQL and experience working on databases like MySQL and SQL Server.
  • Performeddata analysisusingMySQL & Oracle.
  • Experienced the integration of various data sources like Java, RDBMS, Shell Scripting, Spreadsheets, and Text files.
  • Understanding of Data Structures and Algorithms.
  • UNIX shell scripting, resource Extensive experience in Java and J2EE technologies like JSP, JDBC.
  • Extensively worked ondebuggingusing Eclipse debugger.
  • Good Knowledge in Flume and Zoo Keeper Architecture.Good understanding of Data Mining and Machine Learning techniques.
  • Extensive experience in working with different databases such as RDBMS, SQL Server, MySQL and writing Stored Procedures, Functions, Joins and Triggers for different Data Models.
  • Strong work ethic with desire to succeed and make significant contributions to the organization.
  • Strong problem solving skills, good communication, interpersonal skills and a good team player.
  • Have the motivation to take independent responsibility as well as ability to contribute and be a productive team member.
  • A pleasing personality with the ability to build great rapport with clients and customers.
  • Illustrates excellent verbal and written communication, paired with great presentation and interpersonal skills.
  • Portrays strong leadership qualities, backed with a great track record as a team player.
  • Adept with the latest business/technological trends.
  • Possesses sharp business acumen as well as great analytical skills, with a penchant for improvisation and perfection.
  • Exhibits a pleasing personality infused with bold confidence.

TECHNICAL SKILLS:

Hadoop/Big Data Technologies: HDFS, Map Reduce Hive, Pig, Sqoop, Flume, Oozie, and Zookeeper.

No SQL Databases: Hbase, MongoDB.

Languages: Java, J2EE, PL/SQL, Pig Latin, HiveQL, Unix shell scripts.

Java/J2EE Technologies: Swing, JDBC, JSON, Java Script, JSP, JQuery

Web Services: HTML,XML, AJAX, SOAP, REST.

Operating Systems: HP-UNIX, Ubuntu Linux and Windows XP/Vista/7/8/10.

Databases: DB2, SQL Server, MySQL.

Web technologies: JSP, JDBC, Java Beans, JavaScript, Web Services (JAX-WS).

Java IDE: Eclipse 3.x, NetBeans.

Tools: SQL Developer, Visio

PROFFESIONAL EXPERIENCE:

Confidential, Kansas, MO

Sr. Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and processing.
  • Importing and exporting data into HDFS, Pig, Hive and HBase using Sqoop.
  • Managing and reviewing Hadoop log files.
  • Worked on loading and transformation of large sets of structured, semi structured and unstructured data into Hadoop system.
  • Responsible to manage data coming from different data sources.
  • Developed simple and complex MapReduce programs in Java for Data Analysis.
  • Load data from various data sources into HDFS using Flume.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Developed Java MapReduce programs for the analysis of sample log file stored in cluster.
  • Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.
  • Responsible for spooling data from DB2 sources to HDFS using sqoop.
  • Created HIVE tables and provided analytical queries for business user analysis
  • Extensive knowledge on PIG scripts using bags and tuples.
  • Created tables in HIVE by partitioning and bucketing for granularity and optimization of HIVEQL.
  • Involved in identifying job dependencies to design workflow for Oozie and resource management for YARN.
  • Capturing data from existing databases that provide SQL interfaces using Sqoop.
  • Efficient in building pig, hive and map-reduce scripts.
  • Cluster coordination services through Zoo Keeper.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Pig, Hive and also written Pig and Hive UDFs.
  • Automated all the jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map way.
  • Exported analyzed data to relational databases using Sqoop for visualization to generate reports for the BI team.

Environment: Amazon EC2, Apache Hadoop 1.0.1, MapReduce, HDFS, Cassandra, Hive, Pig, Oozie, Flume, Java 1.7, JSP, Spring4.x, Spring MVC, Spring Batch, Hibernate, IBM WebSphere, WebLogic application server8.1, JMS, Oracle 9i, Eclipse, HTML, CSS, JSON, Maven, GitHub.

Confidential, Allen TX

Hadoop Developer

Responsibilities:

  • Handled importing of data from various data sources, performed data transformations using HAWQ, Map Reduce.
  • Involved in creating Hive Internal and External tables, loading data and writing hive queries which will run internally in map reduce way.
  • Implemented complex map reduce programs to perform joins on the Map side using Distributed Cache in Java.
  • Designed and implemented Customization of Keys, Values, Partitioners, Combiners, InputFormats and RecordReaders in JAVA.
  • Developing Scripts and Batch Jobs to schedule various Hadoop Programs.
  • Implemented UDFS, UDAFS, UDTFS in java for hive to process the data that can’t be performed using Hive inbuilt functions.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Worked on complex data types Array, Map and Struct in Hive.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Analyzed JSON and XML files using Hive Built in functions and SerDe’s.
  • Transformed the log files into structured data using Hive SerDe’s and Pig Loaders.
  • Parsed JSON and XML files in PIG using Pig Loader functions and extracted meaningful information from Pig Relations by providing a regex using the built-in functions in Pig.
  • Extensively used Pig for data cleansing.
  • Exported the analyzed data to the relational databases using Sqoop and generated reports for the BI team.
  • Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce Hive, Pig, and Sqoop.
  • Deployed and configured Flume agents to stream log events into HDFS for analysis.
  • Familiarity in using NoSQL database, HBase on top of HDFS.
  • Load and transform large sets of structured, semi structured using Hive and Impala.
  • Connected Hive and Impala to Tableau reporting tool and generated graphical reports.

Environment: Hortonworks Hadoop, MapReduce, HDFS, Hive, Java, Pig, Linux, HBase, Zookeeper, Sqoop, Flume, Oozie, kafka, Tableau, Spark, Scala, PL/SQL, Snap Logic.

Confidential, Patskala, OH

Hadoop Developer

Responsibilities:

  • Developed multiple MapReduce Jobs in java for data cleaning and pre-processing.
  • Developed efficient MapReduce programs for filtering out the unstructured data.
  • Experience on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Developed unit test cases for mapper, reducer and driver classes.
  • Developed Hive queries for data sampling and analysis to the analysts.
  • Involved in creating Hive tables, loading with data and writing Hive queries that will run internally in map reduce way.
  • Involved in developing Pig scripts.
  • Used Pig as ETL tool to do transformations, even joins and some pre-aggregations before storing the data onto HDFS.
  • Experience in migrating the Data warehouse from oracle to teradata.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Involved in moving all log files generated from various sources into Hadoop HDFS using Flume for further processing.
  • Good Knowledge of analyzing data in HBase using Hive and Pig. Experienced in defining job flows using Oozie.
  • Used Agile/Scrum method for requirements gathering.
  • Developed Java Map Reduce programs using Mahout to apply on different datasets.
  • Extensive usage of Maven for building jar files of Map Reduce programs and deployed to cluster.
  • Identified several PL/SQL batch applications in General Ledger processing and conducted performance comparison to demonstrate the benefits of migrating to Hadoop.
  • Experienced in managing and reviewing Hadoop log files.
  • Configured Sentry to secure access to purchase information stored in Hadoop.
  • Involved in several POCs for different LOBs to benchmark the performance of data-mining using
  • Hadoop.

Environment: CloudEra Hadoop, MS SQL Server, Oracle, Hadoop CDH 3/4/5, PIG, Hive, ZooKeeper, Mahout, HDFS, HBase, Sqoop, Java, Oozie, Hue, Tez, UNIX Shell Scripting, PL/SQL, Maven, Ant.

Confidential

Hadoop developer

Responsibilities

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Involved in loading data from UNIX file system to HDFS.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries.
  • Responsible for managing data from multiple source.
  • Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Assisted in exporting analyzed data to relational databases using Sqoop.
  • Developed Hive queries to process the data and generate the data cubes for visualizing.
  • Extracted feeds form social media sites such as Facebook, Twitter.
  • Created Pig Latin scripts to sort, group, join and filter the enterprise wise data.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Gained experience in managing and reviewing Hadoop log files.
  • Gained experience in managing and reviewing Hadoop log files.

Environment: Java, J2EE, HTML, CSS, JSP, JDBC, JavaBeans, JavaScript, XML, Ajax, Web sphere 5.1, SQL,PL/SQL, Oracle, Windows & Unix, J unit,.

Confidential

Hadoop/Java Developer

Responsibilities:

  • Analyzed Object Oriented Design and presented with UML Sequence, Class Diagrams.
  • Developed Admission & Census module, which monitors a wide range of detailed information for each resident upon pre-admission or admission to your facility.
  • Developed UI using HTML, JavaScript, JSP, and developed Business Logic and Interfacing components using Business Objects, XML, and JDBC.
  • CreatedRESTfulweb services interface to Java-based runtime engine.
  • Used Spring Transactional Management, JDBC Template.
  • Designed user-interface and checking validations using JavaScript.
  • Developed components using Java multithreading concept.
  • Managed connectivity using JDBC for querying/inserting & data management including triggers and stored procedures.
  • Involved in design of JSP's and Servlets for navigation among the modules.
  • Developed various EJBs (session and entity beans) for handling business logic and data manipulations from database.
  • Designed cascading style sheets and XSLT and XML part of Order entry Module & Product Search Module and did client side validations with java script.
  • Hosted the application on WebSphere.

Environment: J2EE, Java/JDK, JMS, PL/SQL, JDBC, JSP, Servlets, JavaScript, EJB, JNDI, JavaBeans, XML, XSLT, Oracle 9i, RAD 6.0, HTML/ DHTML, UML, JavaScript.

We'd love your feedback!