Hadoop Developer Resume
CA
SUMMARY
- 8.6 Years of extensive IT experience in Analysis, Design, Development, Implementation and experience in Big Data using Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Cassandra, ZooKeeper,Impala, Spark, Kafka.Java/J2EE and Reporting.
- High Exposure on Big Data technologies andHadoopecosystem, In - depth understanding of Map Reduce and theHadoopInfrastructure.
- Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems application architecture.
- Importing and exporting data into HDFS. Having experience in writing PIG Scripts.
- Involved in creating HIVE tables, loading with data and writing HIVE queries.
- Excellent understanding of Hadoop architecture and its components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
- Experience in providing support to data analyst in running Pig and Hive queries.
- Experience in working with different data sources like xml files, Json files, Sql server, Oracle to load data into Hive and HBasetables.
- Worked on Performance Tuning of Hadoop jobs by applying techniques such as Map Side Joins, Partitioning and Bucketing.
- Having good knowledge in NoSQL databases like Hbase, Cassandra.
- Experience in real-time Big Data solutions using HBase handling billions of records.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Experienced in implementing unified data platforms using Kafka producers/ consumers, implement pre-processing using storm topologies.
- Experienced in migrating map reduce programs into Spark RDD transformations, actions to improve performance.
- Experience utilizing Java tools in Business, Web, and Client-Server environments including Java, J2ee, Hibernate, restful webservices, Jasper Reports, report builder,Jboss Fuse ESB and oracle.
- Expertise in developing reports by using reporting tools like Report builder 3.0 and BIRT.
- Expertise in developing application by using JBoss Fuse ESB 6.1.
- Strong experience in writing database objects like Stored Procedures, Functions, Triggers, PL/SQL packages and Cursors for Oracle, SQL Server, MySQL and TelescopeEnterprise8.4.
- Experience on Source control repositories like SVN, CVS and GITHUB.
- Experienced in build/deploy multi module applications using Maven, Jira, Docker and Jenkins.
- Worked primarily in the domains of Medical, Manufacturing and telecommunication and main area of experience has been involved in project development of java and j2ee applications.
TECHNICAL SKILLS
Platforms: Windows (2000/XP), Linux, CentOS and Mac
Big Data Ecosystems: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Flume, impala, CDH4, Spark and Kafka.
Programming Languages: Java and j2ee
Scripting Languages: Jsp, Servlets, HTML, Jquery, Angular JS.
Databases: Oracle, SQL Server, NoSQL.
Frameworks: Struts, Hibernate, Fuse ESB 6.1 and RestfulWeb Services
Tools: My Eclipse, Maven Build Tool, Report builder 3.0, Hortonworks, MapR, Jira, Docker.
Servers: JBoss, Tomcat, BEA Web Logic 8.1, WebSphere 6.
Methodologies: UML, Design Patterns.
Concepts: JMS, Birt Reports, Jasper Reports
PROFESSIONAL EXPERIENCE
Confidential - CA
Hadoop Developer
Responsibilities:
- Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments.
- Developed Java Map Reduce programs on log data and analysed to support IOT.
- Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
- Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
- Good knowledge about Cassandra architecture, read, write paths and quering using Cassandra shell.
- Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
- Configuring Spark Streaming to receive real time data from the Kafka and Store the stream data to HDFS.
- Extracted the data from multiple data sources into Hbase,Hive using Sqoop.
- Monitored workload, job performance and capacity planning using Cloudera Manager.
- Involved in Agile methodologies, daily scrum meetings, sprint planning.
- Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from Sql Server and a variety of portfolios.
- Supported code/design analysis, strategy development and project planning.
- Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Produced and REST based web services.
- Implemented command design pattern to support micro services.
- Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
Environment: Hadoop,HDFS, MapReduce, Hive, Java, J2ee, Restful web services, Report Builder, Birt, HBase, Sqoop, CentOS, Cloudera.
Confidential
Hadoop Developer
Responsibilities:
- Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments.
- Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
- Developed Java Map Reduce programs on mainframe data to transform into structured way.
- Performed data analysis in Hive by creating tables, loading it with data and writing hive queries which will run internally in a MapReduce way.
- Created Hive External tables and loaded the data in to tables and query data using HQL.
- Developed optimal strategies for distributing the mainframe data over the cluster. Importing and exporting the stored mainframe data into HDFS and Hive.
- Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
- Implemented Hbase API to store the data into Hbase table from hive tables.
- Writing Hive queries for joining multiple tables based on business requirement.
- Monitored workload, job performance and capacity planning using Cloudera Manager.
- Implemented JZOS API to convert mainframe data to text file using core java.
- Involved in build applications using Maven and integrated with CI servers like Jenkins to build jobs.
- Involved in Agile methodologies, daily scrum meetings, spring planning.
- Involved in Build, Deployment and Integration.
Environment: Hadoop,HDFS, MapReduce,Hbase, Hive, Impala, Cloudera,Maven, Java, Jzos, Mainframe.
Confidential
Hadoop Developer
Responsibilities:
- Experience in developing solutions to analyze large data sets efficiently.
- Developed Map Reduce application to find out the useful metrics from the data. Did a thorough testing in local mode and distributed mode found bugs with the code and ensured 100% issue free delivery to production.
- Developed Java Map Reduce programs on log data to transform into structured way.
- Developed optimal strategies for distributing the data over the cluster; importing and exporting the stored RDBMS data into HDFS and Hive using Impala.
- Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
- Writing Hive queries for joining multiple tables based on business requirement.
- Integrated impala with the rest of theHadoopstack supporting several types ofHadoopjobs out of the box (such as Map-Reduce, Hive, and impala) as well as system specific jobs (such as Java programs).
- Created web UI with angular JS to interact with spring application to execute hive quires.
- Monitored workload, job performance and capacity planning using Cloudera Manager.
- Involved in Agile methodologies, daily scrum meetings, spring planning.
- Prepared design documents and functional documents.
- Involved in Build, Deployment and Integration.
Environment: Hadoop,HDFS,MapReduce, Hive, Impala, Cloudera, Java, Spring, Maven, Angular JS
Confidential
Hadoop Developer
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Setup and benchmarked Hadoop clusters for internal use.
- Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior. Used UDF’s to implement business logic in Hadoop.
- Implemented business logic by writing UDFs in Java and used various UDFs from other sources.
- Experienced on loading and transforming of large sets of structured and semi structured data.
- Managing and Reviewing Hadoop Log Files, deploy and Maintaining Hadoop Cluster.
- Involved in implementation of JBoss Fuse ESB 6.1 .
- Consumed REST based web services.
Environment: Hortonworks,Hadoop-Hdfs, Hive, Impala, Java, J2ee, RestServices, MapReduse, Jboss Fuse ESB 6.1.
Confidential
Hadoop Developer/ Sr. Java Developer
Responsibilities:
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
- Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
- Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
- Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
- Involved in implementation of JBoss Fuse ESB 6.1 and Restful web servies.
- Supported code/design analysis, strategy development and project planning.
- Created reports for BI using Sqoop to export data into HDFS and Hive.
- Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Assisted with data capacity planning and node forecasting.
- Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
Environment: Hadoop, HDFS, HIVE, PIG, Map Reduce, Sqoop, Java /J2ee,Restservices, JBoss Fuse ESB 6.1,BIRT
Confidential
Sr.Java Developer
Responsibilities:
- Involved in end to end development and code review.
- Developed all reports based Report Builder.
- Developed object relational mapping model using Hibernate In persistence DAO layer.
- Developed Web application based on Struts MVC framework.
- Implemented various Design patterns like Singleton, MVC, DAO.
- Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Implemented business logic by using Struts Action classes and Struts controller components.
- Coordinating the team regularly to discuss on the better functionality.
- Followed good practices while developing code.
- Analyzing issues and performs impact analysis.
- Debugging, Defect Fixing and Maintenance.
- Identifying test cases and Unit testing.
- Developed stored procedures, functions and triggers in SQL server 2008.
Environment: Struts, Hibernate, Java, JSP, Servlets, WebSphere, XML, TelescopeEnterprise, MS Sql Report Builder 3.0, MSSql Server,Eclipse, MapReduce.