We provide IT Staff Augmentation Services!

Hadoopdeveloper Resume

4.00/5 (Submit Your Rating)

San Jose, CA

SUMMARY:

  • IT professional with 8+ years of experience in Analysis, Design, Development, Integration, Testing and maintenance of various applications using JAVA /J2EE technologies along with 3 + years of Big Data /Hadoopexperience.
  • Experienced in building highly scalable Big - data solutions usingHadoopand multiple distributions i.e. Cloudera, Horton works and NoSQL platforms (Hbase & Cassandra).
  • Expertise in big data architecture withHadoopFile system and its eco system tools Map Reduce, HBase, Hive, Pig, Zookeeper, Oozie, Flume, Avro, Impala and Apache spark.
  • Hands on experience on performing Data Quality checks on petabytes of data
  • Solid understanding ofHadoopMRV1 andHadoopMRV2 (or) YARN Architecture.
  • Good knowledge on Amazon AWS concepts like EMR & EC2 web services which provides fast and efficient processing of Big Data.
  • Developed, deployed and supported several Map Reduce applications in Java to handle semi and unstructured data.
  • Experience in writing Map Reduce programs and using ApacheHadoopAPI for analyzing teh data.
  • Strong experience in developing, debugging and tuning Map Reduce jobs inHadoopenvironment.
  • Expertise in developing PIG and HIVE scripts for data analysis
  • Hands on experience in data mining process, implementing complex business logic and optimizing teh query using Hive QL and controlling teh data distribution by partitioning and bucketing techniques to enhance performance
  • Experience working with Hive data, extending teh Hive library using custom UDF's to query data in non-standard formats
  • Experience in performance tuning of Map Reduce, Pig jobs and Hive queries
  • Involved in teh Ingestion of data from various Databases like TERADATA( Sales Data Warehouse), AS400, DB2, SQL-SERVER using Sqoop
  • Experience working with Flume to handle large volume of streaming data
  • Extensive experience in migrating ETL operations into HDFS systems using Pig Scripts.
  • Good knowledge in evaluating big data analytics libraries (MLlib) and use of Spark-SQL for data exploratory
  • Experience with Oozie Workflow Engine in running workflow jobs with actions dat runHadoopMap Reduce and Pig jobs.
  • Worked with different File Formats like TEXTFILE, AVROFILE, ORC for HIVE Querying and Processing
  • Used Compression Techniques (snappy ) with file formats to leverage teh storage in HDFS
  • Working knowledge inHadoopHDFS Admin Shell commands.
  • Developed core modules in large cross-platform applications using JAVA, J2EE, Hibernate, Spring, JSP, Servlets, EJB, JDBC, JavaScript, XML, and HTML.
  • Experienced with build tools Maven, ANT and continuous integrations like Jenkins.
  • Working Knowledge in configuring and monitoring tools like Ganglia and Nagios.
  • Hands-on experience in using relational databases like Oracle, MySQL, PostgreSQL and MS-SQL Server.
  • Extensive experience in developing and deploying applications using Web Logic, Apache Tomcat and JBOSS.
  • Developed Unit test cases using Junit, Easy Mock and MRUnit testing frameworks.
  • Experienced with version controller systems like SVN, Clear case.
  • Experience using IDEs tools Eclipse 3.0, My Eclipse, RAD and NetBeans
  • Hands on development experience with RDBMS, including writing SQL queries, PLSQL, views, stored procedure, triggers, etc.
  • Participated in all Business Intelligence activities related to data warehouse, ETL and report development methodology
  • Expertise in Waterfall and Agile software development model & project planning using Microsoft Project Planner and JIRA.

TECHNICAL SKILLS:

Big Data Technologies: HDFS, Map Reduce, Hive, Pig, Sqoop, Flume, Oozie, Avro,HadoopStreaming, Zookeeper, Kafka, Impala, Apache Spark

HadoopDistributions: Cloudera (CDH4/CDH5),Horton Works

Languages: Java, C, SQL, PL/SQL,PIG-Latin, HQL

IDE Tools: Eclipse, NetBeans, RAD

Framework: Hibernate, Spring, Struts, Junit

Web Technologies: HTML5, CSS3, JavaScript, JQuery, AJAX, Servlets, JSP,JSON, XML, XHTML, JSF

Web Services: SOAP,REST, WSDL, JAXB, and JAXP

Operating Systems: Windows (XP,7,8), UNIX, LINUX, Ubuntu, CentOS

Application Servers: Jboss, Tomcat, Web Logic, Web Sphere

Reporting Tools /ETL Tools: Tableau, Power view for Microsoft Excel, Informatica

Databases: Oracle, MySQL, DB2, Derby, PostgreSQL, No-SQL Database (Hbase, Cassandra)

WORK EXPERIENCE:

HadoopDeveloper

Confidential, San Jose, CA

Responsibilities:

  • Responsible for building scalable distributed data solutions usingHadoop. Written multiple Map Reduce programs in Java for Data Analysis Wrote Map Reduce job using Pig Latin and Java API Performed performance tuning and troubleshooting of Map Reduce jobs by analyzing and reviewing Hadooplog files Developed pig scripts for analyzing large data sets in teh HDFS.
  • Collected teh logs from teh physical machines and teh OpenStack controller and integrated into HDFS using Flume Designed and presented plan for POC on impala. Experienced in migrating Hive QL into Impala to minimize query response time. Knowledge on handling Hive queries using Spark SQL dat integrate with Spark environment.
  • Implemented Avro and parquet data formats for apache Hive computations to handle custom business requirements Responsible for creating Hive tables, loading teh structured data resulted from Map Reduce jobs into teh tables and writing hive queries to further analyze teh logs to identify issues and behavioral patterns. Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement. Implemented Daily Cron jobs dat automate parallel tasks of loading teh data into HDFS using autosys and Oozie coordinator jobs.
  • Responsible for performing extensive data validation using Hive Sqoop jobs, PIG and Hive scripts were created for data ingestion from relational databases to compare with historical data. Involved in loading data from Teradata database into HDFS using Sqoop queries. Involved in submitting and tracking Map Reduce jobs using Job Tracker. Involved in creating
  • Oozie workflow and Coordinator jobs to kick off teh jobs on time for data availability. Used Pig as ETL tool to do transformations, event joins, filter and some pre-aggregations Used Visualization tools such as Power view for excel, Tableau for visualizing and generating reports. Exported data to Tableau and excel with Power view for presentation and refining Implemented business logic by writing Pig UDFs in Java and used various UDFs from Piggybanks and other sources Implemented Hive Generic UDF's to implement business logic.
  • Coordinated with end users for designing and implementation of analytics solutions for User Based Recommendations using R as per project proposals. Implemented test scripts to support test driven development and continuous integration. Involved in story-driven agile development methodology and actively participated in daily scrum meetings.

Environment: Hadoop, Map Reduce, HDFS, Pig, Hive, Sqoop, Flume, Oozie, Java, Linux, Maven, Teradata, Zookeeper, SVN, autosys, Tableau

HadoopDeveloper

Confidential, Albuquerque, NM

Responsibilities:

  • Worked on writing transformer/mapping Map-Reduce pipelines using Java. Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Map Reduce jobs in teh backend
  • Designed and implemented Incremental Imports into Hive tables. Worked in Loading and transforming large sets of structured, semi structured and unstructured data Involved in collecting, aggregating and moving data from servers to HDFS using Apache Flume Written Hive jobs to parse teh logs and structure them in tabular format to facilitate TEMPeffective querying on teh log data Involved in creating Hive tables loading with data and writing hive queries dat will run internally in map reduce way. Experienced in managing and reviewing theHadooplog files. Migrated ETL jobs to Pig scripts do Transformations, even joins and some pre-aggregations before storing teh data onto HDFS. Implemented teh workflows using Apache Oozie framework to automate tasks Worked with Avro Data Serialization system to work with JSON data formats.
  • Worked on different file formats like Sequence files, XML files and Map files using Map Reduce Programs. Involved in Unit testing and delivered Unit test plans and results documents using Junit and MRUnit. Developed scripts and automated data management from end to end and sync up between all teh clusters. Involved in Setup and benchmark ofHadoop/HBase clusters for internal use.
  • SetupHadoopcluster on Amazon EC2 using whirr for POC Created and maintained Technical documentation for launchingHADOOPClusters and for executing pig Scripts.

Environment: Hadoop, Big Data, HDFS, Map Reduce, Sqoop, Oozie, Pig, Hive, Flume, LINUX, Java, Eclipse, Cassandra,HadoopDistribution of Cloudera., PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting, Putty and Eclipse

JavaDeveloper

Confidential, Houston, TX

Responsibilities:

  • Worked with business analyst in understanding business requirements, design and development of teh project. Implemented teh Struts frame work with MVC architecture.
  • Created new JSP's for teh front end using HTML, Java Script, Jquery, and Ajax. Developing JSP pages and configuring teh module in teh application. Developed teh presentation layer using JSP, HTML, CSS and client side validations using JavaScript. Involved in designing, creating, reviewing Technical Design Documents. Developed DAOs (Data Access Object) using Hibernate as ORM to interact with DBMS - Oracle. Applied J2EE design patterns like Business Delegate, DAO and Singleton. Deployed and tested teh application using
  • Tomcat web server. Using java scripts did client side validation. Involved in developing DAO's using JDBC. Involved in coding, code reviews, JUnit testing, Prepared and executed Unit Test Cases. JBOSS for application deployment and MySQL for database Worked with QA team in preparation and review of test cases. JUnit was used for unit testing for teh integration testing tool.
  • Writing SQL queries to fetch teh business data using Oracle as database. Developed UI for Customer Service Modules and Reports using JSF, JSP's and My Faces Components Log4j used for logging teh application log of teh running system to trace teh errors and certain automated routine functions. CVS was used as configuration management tool.

Environment: and Tools: Java, JSP, JavaScript, Servlets, Struts, Hibernate, EJB, JSF, JSP, Ant, Tomcat, CVS, Eclipse, SQLDeveloper, Oracle.

JavaDeveloper

Confidential, Hanover, MD

Responsibilities:

  • Developed teh application using Struts Framework dat leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used Gathered business requirements and wrote functional specifications and detailed design documents Extensively used Core Java, Servlets, JSP and XML
  • Designed teh logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database Implemented Enterprise Logging service using JMS and apache CXF. Developed Unit Test Cases, and used JUNIT for unit testing of teh application Implemented Framework Component to consume ELS service. Involved in designing user screens and validations using
  • HTML, jQuery, Ext JS and JSP as per user requirements Implemented JMS producer and Consumer using Mule ESB. Wrote SQL queries, stored procedures, and triggers to perform back-end database operations Sending Email Alerts to supporting team using BMC m send. Designed Low Level design documents for ELS Service.

Environment: Java, Spring core, JMS Web services, JMS, JDK, SVN, Maven, Mule ESB Mule, Junit, WAS7, Jquery, Ajax, SAX.

JavaDeveloper

Confidential

Responsibilities:

  • Used Hibernate ORM tool as persistence Layer - using teh database and configuration data to provide persistence services (and persistent objects) to teh application. Implemented Oracle Advanced Queuing using JMS and Message driven beans. Responsible for developing DAO layer using Spring MVC and configuration XML's for
  • Hibernate and to also manage CRUD operations (insert, update, and delete). Implemented Dependency injection of spring frame work. Developed and implemented teh DAO and service classes. Developed reusable services using BPEL to transfer data. Participated in Analysis, interface design and development of JSP. Configured log4j to enable/disable logging in application. Developed Rich user interface using HTML, JSP, AJAX, JSTL, Java Script, JQuery and CSS. Implemented PL/SQL queries, Procedures to perform data base operations. Wrote UNIX Shell scripts and used
  • UNIX environment to deploy teh EAR and read teh logs. Implemented Log4j for logging purpose in teh application. Involved in code deployment activities for different environments. Implemented agile development methodology.

Environment: Java, Spring, Hibernate, JMS, EJB, Web logic Server, JDeveloper, SQLDeveloper, Maven, XML, CSS, JavaScript, JSON.

We'd love your feedback!