We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Birmingham, AL

SUMMARY

  • 8+ years of overall experience in IT industry, which includes hands on experience in Big data eco - system related technologies like Map Reduce, Hive, HBase, Pig, SQOOP, Oozie, flume, Zookeeper and HDFS.
  • 3 years of comprehensive experience as a Big Data & Analytics (Hadoop).
  • Very Strong Object-oriented concepts with complete software development life cycle experience - Requirements gathering, Conceptual Design, Analysis, Detail design, Development, System and User Acceptance Testing.
  • Familiarity with Datastax Java and Python Client drivers for building client side applications interacting with Cassandra database.
  • Experience in working with MapReduce programs using Apache Hadoop for working with Big Data.
  • Hands on experience in big data ingestion tools like Flume and SQOOP.
  • Integrated the HIVE and HBase to access the hive data randomly using NOSQL database from UI.
  • Strong knowledge of Pig and Hive's functions, extending Hive and Pig core functionality by writing UDFs.
  • Proficiency in different Hadoop distributions Cloudera and Hortonworks Distributions.
  • Expertise on Cassandra internals like Gossip communication, Peer-Peer architecture, Incremental Repairs, Replication & Compaction strategies.
  • Experience in analyzing data using HiveQL, Pig Latin, Impala and custom Map Reduce programs in Java.
  • Hands on NoSQL database experience with Apache HBase& Mongo DB.
  • Knowledge in job workflow scheduling and coordinating tools like Oozie and Zookeeper.
  • Experience with various scripting languages like Linux/Unix shell scripts, Python 2.7 and Scala.
  • Good Knowledge on Apache Spark & Scala.
  • Good experience on EMS, RV, JDK 5, JDBC, Servlets, JSP, Struts 2.1, Spring 2, AJAX, DOJO, JSON Hibernate 3, XML Processing (JAXB), JMS, and TIBCO
  • Strong front-end UI development skills using scripting languages like JSP, HTML, JavaScript, jQuery and CSS.
  • Familiar with writing Map Reduce jobs for processing the data over Cassandra cluster, Hbase.
  • Good experience on application server WebSphere 6.0 and JBoss 4.x.
  • Experience to build tool like Ant.
  • Good Knowledge on life cycle of ETL (Extraction, Transformation and Loading) using Informatica Power Center (Repository Manager, Mapping Designer, Workflow Manager, Workflow monitor).
  • Good database experience using SQL Server, Stored Procedures, Cursors, Constraints and Triggers.
  • Highly motivated, adaptive and quick learner.
  • Domain Knowledge: Credit Cards domain experience with expertise in Non-Monetary & Monetary process, Payments, Fees, Collections, Statements functionalities, Banking, SMS Alerts & Email gateway solutions, IVR, Cards Origination, GL Accounting and middleware integration.
  • Strong Analytical and Problem-Solving skills, Multi-Tasking abilities, with proven experience in utilizing people and process knowledge to assist enterprises in making critical decision.

TECHNICAL SKILLS

  • Hadoop/Big Data Technologies: HDFS, Map Reduce, HBase, Hive, Pig, Impala, SQOOP, Flume, OOZIE, Spark, SparkQL and Zookeeper.
  • Languages: Core Java, XML, HTML and HiveQL
  • J2EE Technologies: Servlets, JSP, JMS, JSTL, AJAX, DOJO, JSON and Blaze DS.
  • Frameworks: Spring 2, Struts 2 and Hibernate 3
  • XML Processing: JAXB
  • Reporting Tools: BIRT 2.2.
  • Application & Web Services: WebSphere 6.0, JBoss 4.X and Tomcat 5.
  • Scripting Languages: Java Script, Angular JS, Pig Latin, Python 2.7and Scala.
  • Database (SQL/No SQL): Oracle 9i, SQL Server 2005, MySQL, HBase and Mongo DB 2.2
  • IDE: Eclipse and Edit plus
  • Tools: SVN, Apache Ant, Junit and Star UML
  • EAI Tools: TIBCO 5.6
  • Bug tracking/ Ticketing: Mercury Quality Center and Service Now
  • Operating System: Windows 98/2000 and Linux /Unix

PROFESSIONAL EXPERIENCE

Confidential, Birmingham, AL

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Developing Scripts and Batch Job to schedule various Hadoop Program.
  • Written Hive queries for data analysis to meet the business requirements.
  • Creating Hive tables and working on them using Hive QL. Importing and exporting data into HDFS from Oracle Database and vice versa using Sqoop.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Configured AWS RDS Aurora database users to allow each individual user privileges to perform their related tasks.
  • Developed the Apache Storm, Kafka, and HDFS integration project to do a real-time data analyses.
  • Developed a data pipeline using Kafka and Strom to store data into HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Experience in managing and reviewing Hadoop log files.
  • Worked on Hive for exposing data for further analysis and for generating transforming files from different analytical formats to text files.
  • Created Talend ETL jobs for data transformation, data sourcing and mapping.
  • Wrote MapReduce job using Pig Latin.
  • Created tables in Teradata to export the data from HDFS using Sqoop after all the transformations and wrote Bteq scripts to handle updates and inserts of the records.
  • Have solid understanding of REST architecture style and its application to well performing web sites for global usage.
  • Involved in ETL, Data Integration and Migration. Imported data using Sqoop to load data from Oracle to HDFS on regular basis.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.
  • Worked on AWS to create EC2 instance and installed Java, Zookeeper and Kafka on those instances.
  • Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS.
  • Written Hive queries for data analysis to meet the business requirements.
  • Involved in writing Hive scripts to extract, transform and load the data into Database.
  • Used JIRA for bug tracking.
  • Used CVS for version control.

Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Hive, Pig, Sqoop, Shell Scripting, Java (JDK 1.6), Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Toad 9.6, Linux, JIRA 5.1, CVS, JIRA 5.2.

Confidential, Atlanta, GA

Hadoop Developer

Responsibilities:

  • Worked extensively with Amazon Web Services.
  • Created Amazon Elastic MapReduce cluster in 1.0.3.
  • Developed Pig program for loading and filtering the streaming data into HDFS.
  • Imported data from Oracle database to HDFS using Sqoop.
  • Worked on Data cleansing using apache Avro schema and implementing it in Pig.
  • Developed Hive scripts to de-normalize and aggregate the disparate data.
  • Automated workflows using shell scripts and Oozie jobs to pull data from various databases into Hadoop.
  • Deployed Hadoop cluster in Azure HDInsight to compare scalability and cost-effectiveness, Queried Hadoop cluster using PowerShell, Hue and as well as the remote console
  • Worked on Talend with Hadoop. Worked in migrating from Informatica Talend jobs
  • Implemented external tables and dynamic partitions using Hive.
  • Worked on custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSV, etc.

Environment: : Hortonworks, Hadoop, MapReduce, HDFS, Hive, Sqoop, Avro, Oozie, Java (jdk1.6), Informatica, Oracle 4.0.8 .

Confidential, Livingston, NJ

Hadoop Developer

Responsibilities:

  • Installed, implemented and administered Hadoop & Hive cluster
  • Setup Flume to stream application web log and twitter data to HDFS
  • Parsed JSON data using Apache Pig Latin scripts
  • Analyze escalated incidences within the Azure SQL database
  • Deployed the application on Amazon Web Services (AWS)
  • Developed Pig Latin scripts to parse JSON data and create analytics reports
  • Processed data in form of tuples, bags and char arrays
  • Used Secondary Sort feature to tune Map Reduce applications
  • Configured Flume agent with Flume receive the data from twitter
  • Auto Populate HDFS with data coming from Flume sink
  • Developed Map/Reduce Jobs of various complexity using Hive and Pig
  • Analyzed and troubleshot Hadoop logs
  • Used Secondary Sort feature to tune MapReduce applications
  • Sqoop incremental loads, import and export between Oracle and HDFS
  • Worked on Sqoop to populate data from Oracle RDBMS to HDFS
  • Transformation of data within Hadoop framework.
  • Component unit testing using Azure Emulator

Environment: Hadoop Map Reduce, Hive QL, Pig Latin, ETL, Flume, Administration, Shell Script.

Confidential, Charlotte, NC

Hadoop Developer

Responsibilities:

  • Worked on Hortonworks (HDP 2.X) platform to perform Hadoop operations.
  • Installed and configured Hive, Pig, Sqoop and Flume on the Hadoop cluster.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • The logs and semi structured content that are stored on HDFS were preprocessed using PIG and the processed data is imported into Hive warehouse which enabled business analysts to write Hive queries.
  • Imported data from AWS S3 into Spark RDD, Performed transformations and actions on RDD's.
  • Created Hive tables to store the processed results in a tabular format.
  • Developed Hive Scripts for implementing dynamic Partitions.
  • Involved in importing the real time data to Hadoop using Kafka and implemented the Oozie job for daily imports.
  • Developed Pig scripts for data analysis and extended its functionality by developing custom UDF's.
  • Extensive knowledge on PIG scripts using bags and tuples.
  • Experience in managing and reviewing Hadoop log files.
  • Hands on experience in loading data from UNIX file system and Teradata to HDFS
  • Developed workflow in middleware to automate the tasks of loading the data into HDFS and pre-processing with PigInvolved in loading data from UNIX/LINUX file system to HDFS.

Environment: Hadoop (HDP 2.X), UNIX, Eclipse, TIBCO, HDFS, Java, MapReduce, Apache Pig, Hive, Flume, Python 2.7, SQOOP and MySQL.

Confidential

Java Developer

Responsibilities:

  • Developed Servlets and back-end Java classes using Web Sphere application server.
  • Developed an API to write XML documents from a database.
  • Performed usability testing for the application using JUnit Test.
  • Maintenance of a Java GUI application using JFC/Swing.
  • Created complex SQL and used JDBC connectivity to access the database.
  • Involved in the design and coding of the data capture templates, presentation and component templates.
  • Part of the team that designed, customized and implemented metadata search and database synchronization.
  • Developed JavaScript behavior code for user interaction.
  • Created database program in SQL server to manipulate data accumulated by internet transactions.
  • Wrote Servlets class to generate dynamic HTML pages.
  • Used Oracle as Database and used Toad for queries execution and also Involved in writing SQL scripts, PL SQL code for procedures and functions.

Environment: Java, Web Sphere 3.5, EJB, Servlets, JavaScript, JDBC, SQL, JUnit, Eclipse IDE and Apache Tomcat 6 .

Confidential

SQL Server/Oracle Developer

Responsibilities:

  • Involved in various stages of Enhancements in the Application by doing the required analysis, development, and testing.
  • Prepared the High and Low-level design document and Generating Digital Signature
  • For analysis and design of application created Use Cases, Class and Sequence Diagrams.
  • For the registration and validation of the enrolling customer developed logic and code.
  • Developed web-based user interfaces using struts frame work.
  • Handled Client-side Validations used JavaScript and
  • Involved in integration of various Struts actions in the framework.
  • Used Validation Framework for Server-side Validations
  • Created test cases for the Unit and Integration testing.
  • Front-end was integrated with Oracle database using JDBC API through JDBC-ODBC Bridge driver Confidential server side.

Environment: Java Servlets, JSP, Java Script, XML, HTML, UML, Apache Tomcat, JDBC, Oracle, SQL .

We'd love your feedback!