We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

San Francisco, CA

SUMMARY

  • Over 6 years of professional experience in the field of IT that includes 3 years of experience as Java Developer.
  • Over 3 years of experience as Hadoop Developer, hands on experience in Hadoop Ecosystem technologies.
  • Good understanding of Hadoop Distributed File System and Eco System (Map Reduce, PIG, HIVE, HBase, Flume, Sqoop, Zoo Keeper and Oozie).
  • Well versed in configuring the Hadoop Cluster using major Hadoop Distributions like Cloudera.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing big data as per the requirements.
  • Performed analytics on Big Data using PIG, HIVE and Map Reduce.
  • Experience in analyzing large amounts of data writing PIG Latin Scripts using and using Hive Query Language.
  • Successfully done in importing and exporting data between RDBMS into HDFS using Sqoop.
  • Used Flume to channel data from different resources into HDFS.
  • Good in writing Map Reduce programs and user defined functions (UDF’s) for both PIG and HIVE in java.
  • Worked with Oozie workflow engine to schedule time based jobs to perform multiple actions.
  • Experience in writing Logical implementation and interaction with HBase.
  • Experience in Hive partitioning, bucketing and perform different types of joins on Hive tables and implementing Hive SerDe.
  • Worked on developing Spark jobs using Scala in test environment for faster data processing and used Spark SQL for querying.
  • Good working knowledge on NoSQL databases like Cassandra and HBase.
  • Experience in using Version Controls systems like Git, SVN.
  • Well versed in writing SQL queries, stored Procedures, Functions, Cursors, Index, Triggers and packages.
  • Involved all aspects of Software Development Life Cycle (Analysis, System Design, Development, testing and maintenance) using Waterfall and Agile methodologies.
  • Hands on experience in Built tools like ANT, MAVEN and used Jenkins for continuous Integration.
  • Highly adept at promptly and thoroughly mastering new technologies with a keen awareness of new industry developments and the evolution of next generation programming solutions.

TECHNICAL SKILLS

Hadoop/Big Data Technologies: HDFS, Map Reduce, HBase, HCatalog, Pig, Hive, Sqoop, Spark, Impala, Cassandra, Oozie, YARN, Flume, Kafka

Programming / Scripting Language: Java, Python, SQL, PL/SQL, Shell Scripting, Storm, PIG Latin.

Frameworks: MVC, Spring, Struts, Hibernate

Web Technologies: HTML, XML, Ajax, SOAP, Java Script, CSS, JSP

Databases: Cassandra, Oracle 9i/10g/11g, SQL Server, MySQL

Database Tools: MS SQL Server, Oracle, My SQL

Operating Systems: Linux, Unix, Windows, Mac, CentOS

Other Concepts: OOPS, Data Structures, Algorithms, Software Engineering.

NoSQL Databases: HBase, Cassandra and MongoDB

Application Server: Apache Tomact, JBoss, Web Logic

Methodologies: Scrum, Agile, Waterfall

PROFESSIONAL EXPERIENCE

Confidential, San Francisco, CA

Hadoop Developer

Responsibilities:

  • Developing and running Map - Reduce jobs on YARN and Hadoop clusters to produce daily and monthly reports as per user’s need.
  • Debugging/Troubleshoot issues on UDF’s in Hive.
  • Scheduling and managing jobs on a Hadoop cluster using Oozie work flow.
  • Experience in developing multiple MapReduce programs in java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other file formats.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Transforming unstructured data into structured data using PIG.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Designed and developed PIG latin Scripts to process data in a batch to perform trend analysis.
  • Good experience on Hadoop tools like MapReduce, Hive and HBase.
  • Worked on both External and Managed HIVE tables for optimized performance.
  • Developed HIVE scripts for analyst requirements for analysis.
  • Hands-on experience in using Hive partitioning, bucketing and execute different types of joins on Hive tables and implementing Hive SerDes like JSON and Avro.
  • Worked on Developing custom MapReduce programs and User Defined Functions (UDFs) in Hive to transform the large volumes of data with respect to business requirement.
  • Maintenance of data importing scripts using Hive and Mapreduce jobs.
  • Data design and analysis in order to handle huge amount of data.
  • Cross examining data loaded in Hive table with the source data in oracle.
  • Working close together with QA and Operations teams to understand, design, develop and end-to-end data flow requirements.
  • Developing structured, efficient and error free codes for Big Data requirements using my knowledge in Hadoop and its Eco-system.
  • Storing, processing and analyzing huge data-set for getting valuable insights from them.

Environment: Hadoop, HDFS, Pig, Hive, Oozie, HBase, Map Reduce, Sqoop, Storm, LINUX, Cloudera, BigData, Java, SQL.

Confidential, Dover, New Hampshire

Hadoop Developer

Responsibilities:

  • Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem
  • Datasets will be loaded from two different sources like Oracle, MySQL to HDFS and Hive respectively on daily basis.
  • Installed and configured Hive on the Hadoop cluster
  • Developed complex Mapreduce streaming jobs using Java language that are implemented Using Hive and Pig.
  • Optimized Mapreduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
  • Analyzed the data by performing Hive queries (HiveQL) and running Pig Latin scripts to study customer behavior.
  • Used Impala to query the Hadoop data stored in HDFS.
  • Working as aHadoopconsultant for converting the Oracle Stored Procedures based DataWarehouse Solution to Hadoop based Solution.
  • Filtered, transformed and combined data from multiple providers based on payer filter criteria using custom Pig UDFs.
  • Used the RegEx, JSON and Avro SerDe's for serialization and de-serialization packaged with Hive to parse the contents of streamed log data and implemented Hive custom UDF's.
  • Worked on NoSQL database including MongoDB, Cassandra and HBase.
  • Extensively used Informatica Power Center in end-to-end of Data warehousing ETL routines, which includes writing custom scripts, data mining and data quality process.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Experience in using Sqoop to migrate data to and fro from HDFS and My SQL or Oracle and deployed Hive and HBase integration to perform OLAP operations on HBase data.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Exported the analyzed data to the relational databases using HIVE for visualization and to generate report.
  • Perform data analysis on large datasets.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Experience in writing MapReduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other file formats.

Environment: Hadoop, HDFS, Pig, Eclipse, Hive, Map Reduce, JIRA, HBase, Sqoop, Storm, LINUX, Cloudera, Big Data, Java, SQL, NoSQL, MongoDB.

Confidential, Kansas, MO

Hadoop Developer

Responsibilities:

  • Experienced in migrating the huge volume of data from EDW to IDW Environment.
  • Hands on Experience in managing and reviewing Hadoop log files.
  • Extensive hands-on Hadoop development configuration, management, unit testing, debugging and efficient data processing.
  • Hands on Experience in writing Core Java level programming in order to perform data cleaning, pre-processing and data validation.
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Responsible in loading and transforming large sets of structured, semi structured and unstructured data.
  • Developing and supporting of Map Reduce Programs those are running on the cluster.
  • Involved in using SQOOP for Importing and Exporting of data from Relational Database RDBMS to HDFS.
  • Involved in developing Pig latin scripts in the areas where coding needs to be reduced to analyze large data sets.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading the data and writing hive queries which will run internally in map reduce.
  • Worked on creating Hive queries for data to meet the business requirements.
  • Analyzed the data using Pig and written Pig scripts by grouping, joining and sorting the data.
  • Managing and scheduling jobs to remove the duplicate log data files in HDFS using Oozie.
  • Hands on experience with NoSQL Database.
  • Actively participated in weekly meetings with the technical teams to review the code.
  • Involved in indentifying possible ways to improve the efficiency of the system.
  • Participating in the requirement analysis, design, development and Unit Testing.

Environment: Hadoop, Cloudera Manager, CDH3, Hive, Pig, NoSQL, MongoDB, SQOOP, SQL, Oozie, Java, MapReduce, HDFS, Shell Scripting.

Confidential

Java Developer

Responsibilities:

  • Involved in the design and implementation of the architecture for the project using OOAD, UML design patterns.
  • Developed Action class and Action Form for business logic with support of spring framework and Presentation tier.
  • Involved in design and development of server side layer using XML, JSP, JDBC, JNDI, EJB and DAO patterns using Eclipse IDE.
  • Designed and developed JSP Pages using Struts Frame work and Tag libraries.
  • Involved in implementation of Spring MVC framework and developed DAO and Service layers. Configured the controllers, and different beans such as Handler Mapping, View Resolver etc.
  • Involved in developing Action Servlets classes and Action classes.
  • Created Hibernate configuration files, Struts Application context file.
  • Designed and developed various modules of the application with frameworks like Spring MVC, Web Flow, architecture and Spring Bean Factory using IOC, AOP concepts.
  • Used Angular.JS for developing Single Paged Applications and Bootstrap for responsive web.
  • Used Log4j for logging and debugging.
  • Using Spring-AOP module implemented features like logging, user session validation.
  • Used Hibernate3 with annotation to handle all database operations.
  • Worked on generating the Web Services classes by using Service oriented architecture (SOA).
  • Used JSP and Servlets for server side transactions.
  • Worked in deadline driven environment with immediate feature release cycles.

Environment: Java, spring, Hibernate, JavaScript, Angular.js, Bootstrap, XML, HTML, UML, JBoss, Log4j, Junit, Eclipse, Subversion, Web Services, SQL, Oracle, Windows.

Confidential

Java/J2EE Developer

Responsibilities:

  • Involved in SCRUM Confidential planning and daily standup meetings throughout the process of development.
  • Used OO techniques such as UML methodology (use cases, sequence diagrams and activity diagrams) and developed class diagrams that depict the code's design and its compliance with the functional requirements
  • Developed the web tier using Spring MVC framework. Used spring for dependency injection and integrated spring with Hibernate ORM framework
  • Used HTML, jQuery, JSP, JSF and AJAX in the presentation tier. Developed business delegates and Service Locators to communicate with the EnterpriseJavaBeans (EJB).
  • Developed and consumed REST web services using Jersey framework.
  • Used J2EE design patterns like Session Facade, Business Delegate, Service Locator, Command delegate extensively.
  • Used Hibernate ORM (Object Relational Mapping) Framework to interact with the database to overcome manual result set handling. Developed hibernate configuration files, mapping files and mapping classes.
  • Performed Data Transformations using XSLT and developed SOAP web services using Apache CXF.
  • Used Maven to build and deploy the application onto WebSphere Application Server.
  • Was also involved in migration of the application from WebSphere to JBOSS application server.
  • Used Eclipse IDE for development and SVN for Version Control.

Environment: JDK, Web Sphere, JBOSS, Spring, Hibernate ORM, HTML, XML, JSF, JSP, AJAX, JDBC, XSLT JavaScript, SOAP, REST, SoapUI, JMS, SVN, JUnit, EasyMock, JQuery, Maven, Jenkins, Ajax, Eclipse, DB2, Spring MVC, EJB, Design Patterns, CXF, J2EE.

We'd love your feedback!