Hadoop Developer Resume
Richmond, VA
SUMMARY
- Have 6+ years of programming experience with skills in analysis, design, development, testing and deploying for large scale distributed data processing usingHadoop, Pig, Hive and Java and other various software applications with emphasis on Object oriented programming.
- About 2+ years of work experience on Big Data Analytics.
- Strong hands on experience with Big Data Technologies includingHadoop(HDFS & MapReduce), PIG, HIVE, HBASE, ZOOKEEPER, and SQOOP.
- Developing Map - Reduce programs to perform Data Transformation.
- Have hands on experience in writing Map Reduce jobs on Hadoop Ecosystem including Hive and Pig for different file formats like JSON,XML.
- Have hands on experience in developing batch file processing in hadoop.
- Hands on experience in installing, configuring and using ecosystem components likeHadoop MapReduce, HDFS, Pig, Hive, and Sqoop.
- Has experience in High performance computing.
- Experience with distributed systems, large-scale non-relational data stores, mapreduce systems, data modeling, and big data systems.
- Experience in installing, configuring and administratingHadoopcluster for majorHadoop distributions.
- Experience in working withHadoopin Standalone, pseudo and distributed modes.
- Hands-on experience in products development withHadoopapplications.
- Experience in working with NoSQL databases like HBase.
- Importing and exporting data from different databases like MySQL, Oracle into HDFS and Hive using Sqoop.
- Hands-on experience in writing Pig Latin scripts, working with grunt shells and job scheduling with Oozie.
- Has development experience in RDBMS like Oracle.
- Expertise in Java/J2EE technologies such as Core Java, Spring, Hibernate, JDBC, JSP, JSTL, HTML, JavaScript, JSON.
- Experience in deploying applications in heterogeneous application servers TOMCAT and WebLogic.
- Good understanding of service oriented architecture (SOA) and web services like XML, XSD, XSDL, SOAP.
- Experience in object oriented analysis and design (OOAD), unified modeling language (UML) and design patterns.
- In-depth understanding of Data Structure and Algorithms.
- Strong Communication skills of written, oral, interpersonal and presentation.
- Implemented Unit Testing using JUNIT and MRUNIT testing during the projects.
- Strong desire and ability to perform at a high level for a fast-paced, flexible environment.
- Excellent analytical, problem solving, communication and interpersonal skills with ability to interact with individuals at all levels and can work as a part of a team as well as independently.
- A quick learner organized and highly motivated as well as a keen interest in the emerging technologies.
TECHNICAL SKILLS
Hadoop/Big Data: HDFS, Mapreduce, HBase, Pig, Hive, Sqoop, Flume, Oozie, Zookeeper, kafka, Navigator,Cloudera Manager, Hue
Application Servers: Apache Tomcat
RDBMS: Oracle, MySQL,NoSQL
Database Languages: SQL, PL/SQL
NoSQL Databases: HBase
Programming Languages: C, C++, Java, J2EE,Python, Ant, Shell scripts.
Web Technologies: HTML, XML, JavaScript, AJAX, SOAP,REST, WSDL
Platform: UNIX, Linux, Windows vista/XP/2000/98/95/7, MS DOS
PROFESSIONAL EXPERIENCE
Confidential, Richmond,VA
Hadoop Developer
Responsibilities:
- Developed various data cleansing features like Schema validation, Row Count and data profiling using mapreduce jobs.
- Created hive tables for storing the logs, whenever a mapreduce job is executed.
- Created a hive aggregator to update the hive table after running the data profiling job.
- Extracted data from Teradata to HDFS using sqoop.
- Analyzed the data by performing Hive queries
- Implemented Partitioning, Dynamic Partitioning and Bucketing in HIVE.
- Developed Hive queries to process the data and generate the data cubes for visualizing.
Environment: Hadoop Yarn architecture, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Oozie, Java (jdk 1.7), Eclipse.
Confidential, Sunnyvale, CA
Hadoop Admin/Developer
Responsibilities:
- Involved in start to end process ofhadoopcluster installation, configuration and monitoring.
- Monitored Hadoop cluster using tools like Nagios, Ganglia and Cloudera Manager.
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on theHadoopcluster.
- Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
- Prepared multi-cluster test harness to exercise the system for performance and failover.
- Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
- Created HBase tables to store variable data formats of data coming from different applications.
- Involved in transforming data from legacy tables to HDFS and HBASE tables using Sqoop.
- Responsible for building scalable distributed data solutions usingHadoop.
- Developed Simple to complex Map/reduce Jobs using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS.
- Analyzed the data by performing Hive queries and running Pig scripts to study behavior of lab equipment.
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
- Continuous monitoring and managing theHadoopcluster using Cloudera Manager.
- Worked on Oozie workflow engine to run multiple Hive and Pig jobs.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
Environment: Hadoop Yarn architecture, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse, Linux, NoSql.
Confidential, Melville, NY
Hadoop Admin/Developer
Responsibilities:
- Work closely with various levels of individuals to coordinate and prioritize multiple projects throughout SDLC.
- Worked in the BI team in the area of Big DataHadoopcluster implementation and data integration in developing large-scale system software.
- Configured Hadoop cluster with NameNode and slaves and formatted HDFS.
- Worked extensively in creating Map Reduce jobs to power data for search and aggregation.
- Designed a data warehouse using Hive.
- Worked extensively with Sqoop for importing and exporting the data from HDFS to Relational Database systems/mainframe and vice-versa.
- Extensively used Pig for data cleansing.
- Created partitioned tables in Hive.
- Each node used Ubuntu and Hadoop 0.20.
- Implemented test scripts to support test driven development and continuous integration.
- Worked on tuning the performance Hive queries.
Environment: Hadoop, MapReduce, HDFS, Pig, Hive, Java (jdk1.7), Flat files, Oracle 11g/10g, PL/SQL, SQL*PLUS, Windows NT, Sqoop.
Confidential, Monroe, MI
Java Developer
Responsibilities:
- Involved in Analysis, Design, Coding and Development of custom Interfaces.
- Involved in the feasibility study of the project.
- Gathered requirements from the client for designing the Web Pages.
- Participated in designing the user interface for the application using HTML, DHTML, and Java Server Pages (JSP).
- Involved in writing Client side Scripts using Java Scripts and Server Side scripts using Java Beans and used Servlets for handling the business.
- Developed the Form Beans and Data Access Layer classes.
- XML was used to transfer the data between different layers.
- Involved in writing complex sub-queries and used Oracle for generating on-screen reports.
- Worked on database interaction layer for insertions, updating and retrieval operations on data.
- Deployed EJB Components on Web Logic.
- Involved in deploying the application in test environment using Tomcat.
Environment: Java, J2EE, JSP, Servlets, EJB, Java Beans, JavaScript, JDBC, Web Logic Server, Oracle, HTML, DHTML, XMLCSS, Java Script, Eclipse, Servlets, CVS, Windows 2000.
Confidential
Java Developer
Responsibilities:
- Developed the user interface screens using Swing for accepting various system inputs such as contractual terms, monthly data pertaining to production, inventory and transportation.
- Involved in designing Database Connections using JDBC.
- Involved in design and Development of UI using HTML, JavaScript and CSS.
- Involved in creating tables, stored procedures in SQL for data manipulation and retrieval using SQL SERVER 2000, Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in Oracle.
- Developed the business components (in core Java) used for the calculation module (calculating various entitlement attributes).
- Involved in the logical and physical database design and implemented it by creating suitable tables, views and triggers.
- Created the related procedures and functions used by JDBC calls in the above components.
- Involved in fixing bugs and minor enhancements for the front-end modules.
Environment: JDK 1.3, Swing, JDBC, JavaScript, HTML, Resin, SQL Server 2000, Textpad, Toad, MS Visual SourceSafeWindows 2000, HP UNIX.
Confidential
Java Developer
Responsibilities:
- Study and analyze the business process flow and existing system.
- Identified the areas of improvement and the scope of enhancements in the system.
- Upgraded this application from application wise storage to Lever-wise storage for better relation and feasibility with client.
- Developed various Servlets and Java Interfaces as part of the integration and process flow required for the system.
- Created a loosely couple component model to give the better flexibility to handle frequent business changes.
- Participated in design discussions and object model development workshop sessions for different lines of business.
- Involved in developing Use Case, Class, Sequence & Components Diagrams as part of documentation.
- Performed unit testing, and comprehensive scenario testing for the modules I worked on.
Environment: Core Java 1.5, JSP, Servlets, EJB 2.0, Websphere Application Server, Oracle 10g Database, Toad 9.7, SVN (Subversion Version Control System), MyEclipse IDE.