Sr. Hadoop Developer Resume
Summit, NJ
SUMMARY
- Over 8 years of progressive experience in the IT industry wif proven expertise in architecting and implementing Software Solutions using Java&Big Data technologies.
- 4 years of experience on Hadoop working environment includesMap Reduce, HDFS, HBase, Zookeeper, Oozie, Hive, Sqoop, Pig, Yarn, Cassandra, Kafka, Sparkand Flume.
- Solid understanding of Hadoop Distributed File System.
- Good experience wif MapReduce (MR), Hive, Pig, HBase, Sqoop, Oozie,Flume, Spark, Zookeeper for data extraction, processing, storage and analysis.
- In - depth understanding on how MapReduce works and Hadoop infrastructure.
- In depth understanding of Hadoop Architecture and its various components such as Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager and Map Reduce concepts.
- Experience in developing custom MapReduce Programs in Java using ApacheHadoop for analyzing BigData.
- Extensively worked on Hive for ETL Transformations and optimized Hive Queries.
- Experience in importing and exporting data using Sqoop from Relational Database Systems to HDFS and vice-versa.
- Extending HIVE and PIG core functionality by using custom User Defined Function’s (UDF), User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions (UDAF) for Hive and Pig.
- Experience in analyzing data using HiveQL, Pig Latin, and custom MapReduce programs in Java.
- Developed Pig Latin scripts for data cleansing and Transformation.
- Used Flume to channel data from different sources to HDFS.
- Job workflow scheduling and monitoring using tools like Oozie.
- Good experience in Cloudera, Hortonworks&ApacheHadoopdistributions.
- Worked wif relational database systems (RDBMS) such as MySQL, MSSQL, Oracle and NoSQL database systems like HBase and Cassandra.
- Good Knowledge on HadoopCluster architecture and monitoring the cluster.
- Used Shell Scripting to move log files into HDFS.
- Good understanding in processing of real-time data usingSpark.
- Import the data from different sources like HDFS/HBase intoSpark RDD.
- Experienced wif different file formats like CSV, Text files, Sequence files, XML, JSON and Avro files.
- Good knowledge on Data Modelling and Data Mining to model the data as per business requirements.
- Involved in unit testing of Map Reduce programs using Apache MRunit.
- Good knowledge on python scripting, bash Scripting languages.
- Generated ETL reports using Tableau and created statistics dashboards for Analytics.
- Reported the bugs by classifying them and has played major role in carrying out different types of tests viz. Smoke, Functional, Integration, System, Data Comparison and Regression testing.
- Experience in creating Master Test Plan, Test Cases, Test Result Reports, Requirements Traceability Matrix and creating Status Reports and submitting to the Project management.
- Strong hands on experience in MVC frameworks and Spring MVC.
- Good in Designing and developing the Data Access Layer modules wif the halp of Hibernate Framework for the new functionalities.
- Extensively experience in working on IDEs like Eclipse, Net Beans and Edit Plus.
- Working knowledge of Agile and waterfall development models.
- Working experience in all SDLC Phases.
- Extensively used Java and J2EE technologies like Core Java, Java Beans, Servlet, JSP, spring, Hibernate, JDBC, JSON Object, and Design Patterns.
- Experienced in Application Development using Java, J2EE, JSP, Servlets, RDBMS, Tag Libraries, JDBC, Hibernate, XML and Linux shell scripting.
- Worked wif different software version control, bug tracking and code review systems like CVS, Clear Case.
TECHNICAL SKILLS
Big data/Hadoop Ecosystem: HDFS, Map Reduce, HIVE, PIG, HBase, Sqoop, Flume, Oozie, Spark,Storm,Kafka, HCatalog, Impala, Datameer.
Java / J2EE Technologies: Core Java, Servlets, JSP, JDBC, JNI, XML, REST, SOAP, WSDL
Programming Languages: C, C++, Java, Scala, SQL, PL/SQL, Linux shell scripts,HL7.
NoSQL Databases: MongoDB, Cassandra, HBase
Database: Oracle 11g/10g, DB2, MS-SQL Server, MySQL, Teradata.
Web Technologies: HTML, XML, JDBC, JSP, CSS, JavaScript, AJAX, SOAP, Angular JS
Frameworks: MVC, Hibernate 3, Spring 3/2.5/2, DJango.
Tools: Used: Eclipse,IntelliJ,Putty,Winscp,NetBeans,QC,QlikView,IssueTrack,Selenium,SplunkTableau.
Operating System: Ubuntu (Linux), Win 95/98/2000/XP, Mac OS, RedHat
Methodologies: Agile/Scrum, Rational Unified Process and Waterfall
Distributed plat forms: Hortonworks, Cloudera.
Monitoring tools: Ganglia, Nagios.
PROFESSIONAL EXPERIENCE
Confidential, Summit, NJ
Sr. Hadoop Developer
Responsibilities:
- Imported data from different relational data sources like RDBMS, Teradata to HDFS using Sqoop.
- Imported bulk data into HBase Using Map Reduce programs.
- Perform analytics on Time Series Data exists in HBase using HBase API.
- Designed and implemented Incremental Imports into Hive tables.
- Used Rest ApI to Access HBase data to perform analytics.
- DevelopedSparkcode using Scala andSpark-SQL/Streaming for faster testing and processing of data.
- Involved in converting Map Reduce programs intoSparktransformations usingSparkRDD's on Scala.
- Experienced wif batch processing of data sources using ApacheSpark, Elastic search.
- Worked in Loading and transforming large sets of structured, semi structured and unstructured data
- Involvedin collecting, aggregating and moving data from servers to HDFS using Apache Flume
- Written Hive jobs to parse the logs and structure them in tabular format to facilitate TEMPeffective querying on the log data.
- Involved in creating Hive tables, loading wif data and writing hive queries dat will run internally in mapreduce way.
- Experienced in managing and reviewing theHadooplog files.
- Designed and developed Dashboards for Analytical purposes using Tableau.
- Migrated ETL jobs to Pig scripts do Transformations, even joins and some pre-aggregations before storing the data onto HDFS.
- Worked wif Avro Data Serialization system to work wif JSON data formats.
- Worked on different file formats like Sequence files, XML files and Map files usingMap ReducePrograms.
- Involved in Unit testing and delivered Unit test plans and results documents using Junit and MRUnit.
- Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
- Worked on Oozie workflow engine for job scheduling.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Pig Scripts.
Environment: CDH 5.3, Map Reduce, Hive0.14, Spark 1.4.1, Oozie, Sqoop, Pig0.11, Java, Rest API, Maven, MRUnit, Junit, Cloudera.
Confidential, Dallas, TX
Hadoop Developer
Responsibilities:
- Involved in Automation of clickstream data collection and store into HDFS using Flume.
- Involved in creating Data Lake by extracting customer's data from various data sources into HDFS.
- Used Sqoop to load data from Oracle Database into HDFS.
- Developed MapReduce programs to cleanse the data in HDFS obtained from multiple data sources.
- Involved in creating Hive tables as per requirement defined wif appropriate static and dynamic partitions.
- Used Hive to analyze the data in HDFS to identify issues and behavioral patterns.
- Involved in production Hadoop cluster setup, administration, maintenance, monitoring and support.
- Logical implementation and interaction wif HBase.
- Cluster coordination services through Zookeeper.
- Efficiently put and fetched data to/from HBase by writing MapReduce job.
- Developed MapReduce jobs to automate transfer of data from/to HBase.
- Created data queries and reports using Qlik view and Excel. Created Customs queries/reports designed for qualifying verification and information sharing.
- Assisted wif the addition of Hadoop processing to the IT infrastructure.
- Used flume to collect the entire web log from the online ad-servers and push into HDFS.
- Implemented MapReduce job and execute the MapReduce job to process the log data from the ad servers.
- Load and transform large sets of structured, semi structured and unstructured data.
- Back-endJava developer for Data Management Platform (DMP) and building RESTful APIs to build and let other groups build dashboards.
- Worked closely wif architect and clients to define and prioritize use cases and develop APIs.
- Involve in monitoring job performance, capacity planning and workload using Cloudera Manager.
Environment: Hadoop, Pig 0.10, Sqoop, Oozie, MapReduce, HDFS, HBase. Hive 0.10, Core Java, Eclipse,Qlik view, Flume, Cloudera, Oracle 10g, UNIX Shell Scripting, Cassandra.
Confidential -Plano, TX
Sr. Java Developer/Hadoop Developer
Responsibilities:
- Participated in requirement gathering and converting the requirements into technical specifications.
- Created UML diagrams like use cases, class diagrams, interaction diagrams, and activity diagrams.
- Developed the application using Spring Framework dat leverages classical Model View Controller (MVC) architecture.
- Extensively worked on User Interface for few modules using JSPs, JavaScript and Ajax.
- Created Business Logic using Servlets, POJO’s and deployed them on Web logic server.
- Wrote complex SQL queries and stored procedures.
- Developed the XML Schema and Web services for the data maintenance and structures.
- Implemented the Web Service client for the login authentication, credit reports and applicant information using Apache Axis 2 Web Service.
- Responsible to manage data coming from different sources.
- Developed map reduce algorithms.
- Got good experience wif NOSQL database.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Worked wif cloud services like Amazon web services (AWS)
- Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 10g database.
- UsedHibernateORM framework wifSpringframework for data persistence and transaction management.
- Wrote test cases in JUnit for unit testing of classes.
- Involved in creating templates and screens in HTML and JavaScript.
- Involved in integrating Web Services using SOAP.
Environment: Hive 0.7.1, Apache Solr - 3.x, HBase-0.90.x/0.20.x, JDK, Spring MVC, WebSphere 6.1, HTML, XML, JavaScript, JUnit 3.8, Oracle 10g, Amazon Web Services.
Confidential, Chicago, IL
Sr. Java Developer
Responsibilities:
- Developed the application using Spring Framework dat leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used
- Gatheird business requirements and wrote functional specifications and detailed design documents
- Extensively used Core Java, Servlets, JSP and XML
- Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database
- Implemented Enterprise Logging service using JMS and apache CXF.
- Developed Unit Test Cases, and used JUNIT for unit testing of the application
- Implemented Framework Component to consume ELS service.
- Involved in designing user screens and validations using HTML, jQuery, Ext JS and JSP as per user requirements
- Implemented JMS producer and Consumer using Mule ESB.
- Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
- Sending Email Alerts to supporting team using BMC msend
- Designed Low Level design documents for ELS Service.
- Closely worked wif QA, Business and Architect to solve various Defects in quick and fast to meet deadlines
Environment: Java, Spring core, JMS Web services, JMS, JDK, SVN, Maven, Mule ESB Mule, Junit, WAS7, JQuery, Ajax, SAX.
Confidential
Java Developer
Responsibilities:
- Using JAVA developed a website for e-Recruitment consists of many modules.
- Followed MVC Architecture for implementing the functionality.
- Designed and reviewed the test scenarios and scripts for given functional requirements.
- Implemented Services using Core Java.
- Involved in development of classes using java.
- Designed and built user interface using spring and JavaScript& employed collection libraries.
- Designed and involved in preparing activity diagrams, usecasediagrams, sequence diagrams as per the business requirement.
- Used JavaScript for Client validation.
- Designed a website for understanding the user requirements and validated the web page using JQuery in Conjunction wif Java Spring/ Hibernate/ PHP.
- Developed user interfaces using Servlets, CSS, XSLT, XML, HTML and JavaScript.
- Good proficiency in developing algorithms for serial interfaces.
- Involved in testing of CAN protocols.
- Developed the flow of algorithm in UML.
- Developed verification and validation scripts in java.
Environment: Java, JSP, Servlets, JDBC, JavaScript, MySQL, JUnit, Eclipse IDE, Windows 7/XP/Vista, UNIX, LINUX.