Hadoop Developer Resume
Charlotte, NC
SUMMARY
- 7+ years of professional experience inHadoopand Java technologies like HDFS, MapReduce,Apache Pig, Hive, Hbase, Sqoop, Oracle, JSP, JDBC and Spring
- 3 years of experience in Client - Server and Web based application development using Java technologies Java, J2EE, JSP, Java Script, Servlets, Springs.
- 4+ years of working experience inHadoopeco-system technologies like Apache Pig, Apache Hive, Apache Sqoop, Apache Flume and Apache Hbase.
- Experience in analyzing data using HiveQL, Pig latin and custom MapReduce programs in Java.
- Hands on experience in writing Pig UDFs, Hive UDFs and UDAFs in the analysis of data.
- Worked with NoSQL database like HBase.
- Experience in importing and exporting data from relational database into HDFS using Sqoop.
- Developed MapReduce jobs, Used different optimization techniques to improve performance in Map Reduce Programs.
- Good knowledge in using job scheduling and monitoring tools like Oozie and ZooKeeper
- Extensive experience in configuring Flume to stream data into HDFS.
- Experience in implementing Spark in Scala and SparkSQL for faster analyzing and processing of data.
- Excellent understanding ofHadooparchitecture and its components like HDFS, MapReduce, Job Tracker, Task Tracker, Name Node and Data Node.
- Used Apache Impala to read, write and query theHadoopdata in HDFS, Hbase and Cassandra.
- Hands on experience in Application Development using Java,Hadoop, RDBMS and Linux shell scripting.
- Extensive experience in developing and deploying Java based applications.
- Involved in working on all phases of software development life cycle (SDLC) from requirements gathering to programming, testing and maintenance.
- Strong skills in designing, developing and testing of Client-Server products and distributed programming using Java, J2EE and Web related technologies.
- Expertise in J2EE Application development using JSP, Servlets, JDBC, XML, Spring.
- Strong experience in handling different Web Servers like Tomcat, and Application Servers like Web logic, Websphere and JBOSS.
- Experienced in GUI design. Extensively used HTML, XML, Java Script, and JSP.
- Involved in the development of Spring JDBC DAO Support for data base interactions.
- Involved in developing Static and Dynamic pages using JSP and Servlets.
- Setup Struts framework in Web logic Server, Tomcat.
- Involved in Software development Lifecycle.
- Functional skills include, project management, Leadership skills, Quality control, L&D
- Good knowledge with PL/SQL stored procedures using Oracle.
- Ability to master new Concepts.
- Excellent problem solving capabilities and communication skills.
- Have proficient time-management skills, able to deliver assignments with superior quality on schedule and within budget.
TECHNICAL SKILLS:
Big Data Technologies: MapReduce, Pig, Hive, Impala, Sqoop, Oozie, Zookeeper, Kafka, Flume, Spark, Yarn, Storm
Hadoop Distributed Systems: Cloudera, Hortonworks, MapR
Java & J2EE Technologies: Spring, Hibernate, Servlets, JDBC, JSP, Struts, JMS, EJB
IDE Tools: Eclipse, Net Beans
Web Technologies: HTML, CSS, Javascript, Jquery, Bootstrap, XML, JSON
Databases: Oracle, MySQL, DB2, PostgreSQL, SQL Server
Programming Languages: Java, C++, Python, Scala
NoSQL Databases: HBase, Cassandra, MongoDB
Operating Systems: Windows, Unix, Linux
Version Control: SVN, CVS, Tortoise, GITOther Tools: Talend, Maven, Ant, Jenkins
PROFESSIONAL EXPERIENCE
Confidential , Charlotte, NC
Hadoop Developer
Responsibilities:
- Built an ETL system in Hadoop to process raw data arriving from hospitals using custom Spark jobs and applying complex transformations to store it on HDFS.
- Built a process based on RabbitMQ, Spark and Spring-boot to send/receive data to Symedical for Reference Data Management.
- Worked on optimizing ETL jobs based on Spark in Scala to improve performance in order to meet demanding client requirements.
- Worked on Proof of concepts for emerging big data technologies
- Wrote shell scripts to automate the data pipeline
- Interacted with architects and business teams to gather requirements
- Followed agile best practices.
- Designed and Maintained Oozie workflows to manage the flow of jobs in the cluster.
- Responsible for building scalable distributed data solutions using Hadoop.
- Developed Simple to complex Map/reduce Jobs using Hive and Pig.
- Provide mentorship and guidance to other architects to help them become independent.
- Provide review and feedback for existing physical architecture, data architecture and individual code.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Involved in Hadoop cluster task like commissioning & decommissioning Nodes without any effect to running jobs and data.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Analyzed large data sets by running Hive queries and Pig scripts.
- Involved in ETL, Data Integration and Migration.
- Experience in using Pig Scripts to implement ad-hoc Mapreduce programs.
- Job management using Fair scheduler.
- Develop Core Framework based on Hadoop to Migrate Existing ETL (RDBMS) Solution.
- Wrote Pig Scripts to generate Map Reduce jobs and performed ETL procedures on the data in HDFS.
- A deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment.
- Worked extensively with Sqoop for importing metadata from Oracle.
- Involved in creating Hive tables, and loading and analyzing data using hive queries.
- Responsible for managing data from multiple sources.
Environment: Hadoop, Hive, MapReduce, Hdfs, Pig, Spark, Scala, Oozie, Impala, Distribution Cloudera
Confidential - Phoenix, AZ
Hadoop Developer
Responsibilities:
- Extensively implemented various Big Data strategies in all stages of SDLC by following Agile.
- Developed Pig Scripts for validating and cleansing the data.
- Developed MapReduce programs to pharse the raw data, and stored the refined data in Hbase.
- Created Hive queries for extracting data from Cornerstone (Data Lake) to HDFS locations.
- Managed and reviewedHadooplogs.
- Tested raw data and executed performance scripts.
- Shared responsibility for administration ofHadoop, Hive and Pig.
- Exported data from HDFS to RDBMS for visualization and user report generation using Tableau.
- Involved in the process of load, transform and analyze Transactions data from various providers intoHadoopon an on-going basis.
- Filtered, transformed and combined data which came from Cornerstone (Data Lake) based on business requirements using custom Pig Scripts and stored in Cognition (downstream DB)
- Responsible for design and creation of Test cases (in Rally) and tested the Tableau dashboards using Functional testing, system testing, Integration testing, regression testing and UAT.
- Participated and conducted Issue Log weekly status meetings, Report status meetings and Project status meetings to discuss issues and workarounds.
- Communicated withdevelopers(On-shore and Off-shore) throughout all the phases of Development to eliminate Roadblocks.
- Generated daily progress report and represented in daily Agile Scrum meetings.
- Hands on experience on Apache Pig.
- Creating Unit Test plans and Test cases.
Environment: ApacheHadoop, PIG, Hive, SQL, Map Reduce, Core Java, Rally, MAPR, UNIX/LINUX, Tableau, Windows, MS Office, Microsoft Outlook.
Confidential, Hartford, CT
Hadoop Developer
Responsibilities:
- Extracted and updated the data into HDFS using Sqoop import and export command line utility interface.
- Responsible for developing data pipeline using Flume, Sqoop, and Pig to extract the data from weblogs and store in HDFS.
- Involved in using HCATLOG to access Hive table metadata from MapReduce and Pig code.
- Involved in developing Hive UDFs for the needed functionality.
- Involved in creating Hive tables, loading with data and writing Hive queries.
- Managed works including indexing data, tuning relevance, developing custom tokenizes and filters, adding functionality includes playlist, custom sorting and regionalization with Solr search engine.
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
- Used pig to do transformations, event joins, filter boot traffic and some pre-aggregations before storing the data onto HDFS.
- Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like spark.
- Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
- Extending Hive and Pig core functionality by writing custom UDFs.
- Experience in managing and reviewing Hadoop log files
- Developed data pipeline using Flume,Sqoop, pig and java Mapreduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Involved in emitting processed data from Hadoop to relational databases and external file systems using Sqoop.
- Orchestrated hundreds of Sqoop scripts, pig scripts, Hive queries using Oozie workflows and sub-workflows.
- Loaded cache data into Hbase using Sqoop.
- Experience in custom talend jobs to ingest, entich and distribute data in MapR, Cloudera Hadoop ecosystem.
- Created lots of external tables on Hive pointed to Hbase tables.
- Analyzed Hbase data in Hive by creating external partitioned and bucketed tables.
- Worked with cache data stored in Cassandra.
- Injected the data from External and Internals Flow Organizations.
- Used the external tables in Impala for data analysis.
- Supported MapReduce Programs those are running on the cluster.
- Participated in apache Spark POCS for analyzing the sales data based on several business factors
- Participated in daily scrum meetings and iterative development.
Environment: Hadoop, MapReduce, Hdfs, Pig, Hive, Hbase, Impala, Sqoop, Flume, Oozie, Apache Spark, Java, Linux, SQL Server, Zookeeper, Autosys, Tableau, Cassandra.
Confidential
Java Developer
Responsibilities:
- Gathered requirements, designed and implemented the application that utilizes Struts, Spring, JSP and Oracle database.
- Implemented J2EE design patterns like MVC and Front Controller.
- Implemented Static and Dynamic web pages using JSP, Java Script, CSS.
- Involved in Requirement analysis, design and provide the estimation.
- Responsibilities include designing and delivering web based J2EE solutions. Used JavaScript for Client Side validations.
- Involved in writing PL/SQL queries and stored procedures.
- Responsible for setup the environment and Production Environments in Server and Database level.
- Involved in developing portlets and deploying in Weblogic Portal Server.
- Involved in writing of release notes to deploy in various environments and production.
- Monitored the Server load average and prepare status report on it.
- Point of Contact to the client for all technical aspects.
- Prepared status reports.
Environment: JSP, Struts, Spring, Tomcat, JavaScript, Oracle10G, Weblogic, XML, NeatBeans
Confidential
Responsibilities:
- Developed Code Generator GUI tool used to create Java files.
- Involved in gathering requirements, design and implemented the application using Java/J2EE technologies.
- Created Static and dynamic web pages using Struts, JSP and Servlets.
- As a part of bug fixing team played a key role fixing issues.
- Implemented presentation tier on Servlets, JSP, and HTML for custom business requirements.
- Played an active role in testing the application for different scenarios and extensively involved in bug fixing.
- Involved in deploying the application in different environments.
- Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support.
- Used J2EE for the development of business layer services.
- Gathered business requirements from the end users and wrote functional specifications and design documents.
- Involved backend-stored procedures, functions and triggers.
- Handled the issues that came during the Production time and resolved them.
- JDBC framework has been used to connect the application with the Database.
- Used Net Beans for the Development, Testing and Debugging of the application.
- SQL Developer was used as a database client.
- Involved in gathering requirements, design and implemented the application using Java/J2EE technologies.
- Created Static and dynamic web pages using JSP and Servlets.
- As a part of bug fixing team played a key role fixing issues.
- Implemented presentation tier on Servlets, JSP, and HTML for custom business requirements.
- Handled the issues that came during the Production time and resolved them.
- Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
- Extensively worked on Core java.
- Supported, Testing and coding issues in Production/QA environment.
Environment: JSP, HTML, CoreJava, Servlets, JavaScript, SQL Developer, Eclipse and Apache Tomcat.