Sr. Hadoop Developer/administrator Resume
Rancho Cordova, CA
PROFESSIONAL SUMMARY
- Overall 8+ years of IT Experience in Analysis, Design, Development, Implementation and Testing of enterprise wide application and Web - Based Applications.
- Having 6+ years of experience in dealing with Apache Hadoop components like HDFS, MapReduce, HIVE, Hbase, PIG, SQOOP, Spark, Impala, OOZIE, Kafka and Flume Big Data and Big Data Analytics.
- Very good understanding of Hadoop architecture and the daemons of Hadoop -Name Node, Data Node, Resource Manager, Node Manager, Task Tracker, Job Tracker.
- Experience working with Hadoop ecosystems - Installation, configuration, administration and monitoring of Hadoop clusters.
- MapReduce programming model was JAVA coded with design patterns and used Hadoop APIs for data transformation.
- Experience in Hadoop ecosystem components such as YARN, Pig, Hive, Sqoop, Hbase, Flume, Oozie and Shell scripting.
- Experience with Sqoop to import/export data from a RDBMS into Hadoop and used Flume to collect data and populate to Hadoop.
- Experience with HA/DR solutions for Hadoop, HBase and the underlying components.
- Good skill set in writing Pig Latin and used Pig Interpreter to run Map Reduce jobs.
- Hands on experience on Hortonworks and Cloudera Hadoop environments.
- Indepth knowledge in task automation using Oozie, cluster coordination through Pentaho and MR job scheduling using Fair Scheduler.
- Hands on Experience in SPARK, used for data transformation for larger data sets.
- Writing custom UDF's in JAVA for extended Hive and Pig core functionality.
- Experience on Transform Components such as Aggregate, Router, Sorted, Filter by Expression, Join, Normalize and Scan Components. Created appropriate DMLs and Automation of load processes using Autosys.
- Advanced skills in J2EE technologies such as JDBC, Servlet, MVC Design Pattern, Spring Core components, Spring MVC, JSP, Struts, Hibernate for development of web applications in IDEs such as STS, Eclipse, NetBeans.
- Extensive knowledge in writing and executing CRUD operations using JDBC with MySQL.
- Proficiency with the application servers like WebSphere, WebLogic, JBOSS and Tomcat.
- Extensively worked on several ETL assignments to extract, transform and load data into tables as part of Data Warehouse development.
- Good Knowledge in Amazon AWS concepts like EMR and EC2 web services which provides fast and efficient processing of Big Data.
- Experience on Front-end designing using JavaScript, JQuery, CSS, XML and HTML5.
- Proficient in Linux (or Unix) System Operations.
- Technical document preparation such as functional and design documents. Conducting code walkthrough with the team.
- Excellent global exposure to various work cultures and client interaction with diverse teams. Proven ability to learn quickly and apply new technologies
- Good communication and written skills and can lead a team and work as a team member.
TECHNICAL SKILLS
Big Data Ecosystem: Hadoop 2.7.2, MapReduce 2.0, Hive, Pig, Sqoop, FlumeOozie, Spark, Kafka and HBase
Programming Language: C, C++, C#.NET, Java, SQL, XML, Python3.4, Pig LatinHiveQL, Unix Shell Scripts
Java Technologies: Spring MVC, JDBC, JSP, JSON, Applets, Swing, JDBC, JNDIJSTL, RMI, JMS, Servlets, EJB and JSF.
Web Technologies: HTML5, XML, CSS, JavaScript, JQuery, AJAX
Web Server: Apache Tomcat, IIS
Databases: MS SQL server, MySQL
Web Service Technologies: JSON, Restful
Build Tools: Maven 2.x, 3.x
UML Tools: MS-Visio 2010/2013/2016
Development Tools: Spring Tool Suite, Eclipse, NetBeans, Toad, ANT
Versioning Tools: SVN, GitHub, CVS
Operating Systems: Ubuntu, CHD3, CHD5, Windows XP/7/8/10
Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP
CMS Packages/Servers: Microsoft Office SharePoint Server 2013
Microsoft Office: MS Word/Excel/PowerPoint/Outlook - 2010/2013/2016
WORK EXPERIENCE
Confidential, Rancho Cordova, CA
Sr. Hadoop Developer/Administrator
Responsibilities:
- Coordinating as Hadoop Developer/Admin in Hadoop distribution for 4 clusters ranges from POC to PROD.
- Using Spark for data transformation of larger data sets.
- And Hive to analyze the data and to extract report.
- Involving in Business Requirements clarification by coordination with Business Analysts. Analyzing the requirements by conducting workshops with the Business Users.
- Working with technical teams for Installation and configuration of Hadoop Cluster. Implemented Cloudera Manager on existing cluster.
- Involved in installing cluster, commissioning & decommissioning of DataNodes, NameNode recovery, capacity planning, and slots configuration.
- Used MapReduce in Hadoop Cluster and implemented task and job managers to process the huge data sets from HDFS.
- Integrated Oozie with the rest of the Hadoop stack, supporting several types of Hadoop jobs (Map-Reduce, Pig, Hive, and Sqoop) as well as system specific jobs (Java programs and shell scripts).
- Experience in retrieving data from databases like MYSQL and Oracle into HDFS using Sqoop and ingesting them into HBase and Cassandra.
- Troubleshooting, Managing and Reviewing data backups and providing Infra support.
- Coordinated with technical teams for importing and exporting of data onto HDFS using Flume and Sqoop.
- Worked on setting up KAFKA for streaming data and monitoring for the KAFKA Cluster.
- Assisted with the transfer of data from RDBMS to HDFS.
- Worked on upgrading cluster and patches without any data loss and with proper backup plans.
- Programmed shell scripts to monitor the health check of Hadoop daemon services and responded accordingly to any warning or failure conditions
- Wrote multiple Pig Latin scripts to extract the data from the web server output files to load into HDFS.
- Designed Hive tables, wrote Hive queries and added Oozie action which will Run automatically to load data into HDFS.
- Configuration of MySQL Database to store Hive metadata.
- Used Change management and Incident management process following the company standards.
- Responsible for developing data pipeline using HDInsight, flume, Sqoop and pig to extract the data from weblogs and store in HDFS.
- Extensively worked with Oozie workflow engine to run multiple Hive and Pig Jobs.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios.
- Installation and upgrading of ClouderaHadoop(CDH) and Cloudera Manager (CM).
Environment: Hadoop, Java,Python, HDFS, MapReduce, Spark, Pig, Hive, Ruby, Hadoop (Cloudera), Sqoop, Yarn, Hbase, Maven/Jenkins, UNIX/LINUX, GIT, Shell scripting, Scala, Flume, Oozie.
Confidential, Memphis, TN
Hadoop Developer
Responsibilities:- Involved in building scalable distributed data solutions using Hadoop.
- Created HBase tables to load data of structured, semi-structured and unstructured data coming from UNIX file system.
- Implemented Big Data tools withinHadoopecosystem (HDFS, Hbase, Hive, Pig, Sqoop, Oozie, Flume).
- Wrote MapReduce/Pig/UDF (Java) jobs in performing data cleaning, transformations and joins.
- Integrated Oozie with the rest of the Hadoop stack, supporting several types of Hadoop jobs (Map-Reduce, Pig, Hive, and Sqoop) as well as system specific jobs (Java programs and shell scripts).
- Developed MR programs to analyze the data, populate staging tables and store the refined data in partitioned tables in the Enterprise Data Warehouse (EDW).
- Created Hive External tables and loaded the data in to tables and queried data using HQL that helped market analysts to look for emerging trends by comparing fresh data with EDW reference tables.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Collected the logs data from web servers and integrated in to HDFS using Flume.
- Developed PIG Latin scripts to extract the data from the web server output files to load into HDFS. Used Pig as ETL tool to do transformations, event joins and some pre-aggregations before storing the data onto HDFS.
- Developed the Pig UDF’s to preprocess the data for analysis.
- Used Sqoop to import and export data from HDFS to RDBMS for visualization and to generate reports.
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
Environment: Hadoop, MapReduce, Hive, HDFS, PIG, Sqoop, Oozie, Cloudera, Flume, HBase, ZooKeeper, Cloudera, MongoDB, Oracle and Unix/Linux.
Confidential, Los Angeles, CAHadoop Developer
Responsibilities:- Involved in the elaboration, construction and transition phases of the Rational Unified Process.
- Worked with Big Data IT pilot team responsible for buildingHadoopstack and different big data analytic tools, migration from RDBMS toHadoop using Sqoop.
- ETL: Data extraction, managing, aggregations and loading into HBase.
- Contributed in writing java code for several MapReduce programs on data transformation using Hadoop API.
- Used Bash shell scripting to perform Hadoop operations.
- Collected logs from web and application servers and integrated in to HDFS using Flume.
- Used Hive to analyze the data and to extract report.
- Designed the application using DAO Layer, Service Controller, MVC, Factory, Data Access
- Object, Service Locator design patterns.
- Designed and developed necessary UML Diagrams (Use Case, Class, Sequence, State) and Activity diagrams using IBM Rational Rose.
- Used IBM Rational Application Developer (RAD) for development.
- Developed MVC architectural pattern using spring Framework such as spring Bean Factory, IOC (Inversion of control), and AOP concepts.
- Developed RESTful web services with Jersey framework and used of various components like WSDL, UDDI, XML-RPC
- Extensively applied various design patterns such as MVC-2, Front Controller, Factory, Singleton, Business Delegate, Session Façade, Service Locator, DAO etc. throughout the application for a clear and manageable distribution of roles.
- Implemented the project as a multi-tier application using Jakarta Struts Framework along with JSP for the presentation tier.
- Used the Spring Tags, Struts Tag Libraries, Struts Validation Framework for validation and Struts Tiles Framework for reusable presentation components at the presentation tier.
- Used Spring-Core and Spring-AOP. Developed various Action Classes that route requests to appropriate handlers.
- Developed Session Beans to process user requests and Entity Beans to load and store information from database.
- Used JMS (MQSeries) for reliable and asynchronous messaging the different components.
- Worked with Web Service concepts like XML, WSDL, UDDI and SOAP with SAX parser to transfer and transform data between applications
- Developed Spring MVC classes for handling requests received from front end logic such as JSP pages.
- Wrote Stored Procedures and complicated queries for IBM DB2
- Designed and used JUnit test cases during the development phase.
- Involved in design and development of UI component, using frameworks Angular JS,
- JavaScript, HTML5, CSS and Bootstrap.
- Extensively used log4j for logging throughout the application.
- Used Tortoise SVN for version control of the source code.
- Open source cross-platform WebTesting (WSDL) using SoapUI.
Environment: Hadoop, MapReduce, Hive, HDFS, PIG, Sqoop, Flume, HBase and Unix/Linux, JDK, J2EE, Web Services (SOAP, WSDL, JAX-WS), Hibernate, Spring, Servlets, JSP, Java Beans, NetBeans, SQL server, JUnit, Clover, Tortoise SVN, Log4j, PL/SQL, Web sphere Application Server, Tomcat Web Server
Confidential, Houston, TX
Java Developer
Responsibilities:- Involved in Software development life cycle (SDLC) as requirement gathering, modeling, analysis, architecture design & development.
- Used J2EE technologies (Servlets, Struts, MVC design patterns and JDBC) in business module applications.
- Implementation of JMS between workflow module and backend Message forwarding Engine.
- Used Servlet programming and JSP scripting for the communication between web browser and server.
- MVC architecture was implemented using struts framework and developed ActionServlet, ActionForm bean, configured the struts-config descriptor, implemented validator framework.
- Developed Session Beans to process requests from the user interface and CMP entity beans to retrieve and update customer information.
- Involved in OOA and OOD Techniques using Unified Modeling Language.
- Involved in writing Hibernate queries and Hibernate specific configuration and mapping files.
- Designed and developed User interface with the use of JSF.
- Written code for JDBC logic to perform CRUD operations.
- Created queries for Stored Procedures, Triggers, Functions and actively involved in tuning SQL queries.
- Used Oracle10g as the backend database and Hibernate as ORM Tool to retrieve Data Access Objects
- Production support, Bug fixing and monitoring production API failures as the part of team responsibilities.
- Quality Control Activities which are defect tracking, bug fixing using JIRA.
- Used Ant for starting up the application server in various modes.
- Used CVS as Version Control to check-in and checkout Data.
Environment: Java, JSF, ANT, struts, JMS, Servlets, MySQL, JavaScript, HTML, CSS, JDK 1.5.1, JDBC, JUnit, Oracle10g, XML, XSL and UML
