Hadoop Developer Resume
Pittsburgh, PA
SUMMARY
- Over 6 years of professional experience in designing, developing, debugging Web - based and Enterprise applications using OOA, OOD, OOPS and JAVA/J2EE technologies.
- Over 3 years of experience on major Hadoop ecosystem’s projects such as Map-Reduce, PIG, HIVE and HBASE.
- Extensive experience in developing PIG Latin Scripts and using Hive Query Language for data analytics.
- Good working experience using Sqoop to import data into HDFS from RDBMS and vice versa.
- Good knowledge in using job scheduling and monitoring tools like Oozie and ZooKeeper.
- Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera.
- Involved in developing distributed Enterprise and Web applications using UML, Java/J2EE, Web technologies that include EJB, JSP, Servlets, Struts II, JMS, JDBC, HTML, XML, XSL, XSLT, Java Script, Spring and Hibernate.
- Experienced in using Agile software methodology (scrum).
- Designed Use Case diagrams, Class diagrams, Activity diagram, Sequence diagrams, Flow Charts, and deployment diagrams using Rational Rose Tool.
- Experience in Web application development using Java, Servlets, JSP, JSTL, Java Beans, EJB, JNDI, JDBC, Struts, HTML, DHTML, CSS, PHP, XML, XSL/XSLT and AJAX.
- Experience with IDE’s like Eclipse, NetBeans, RAD, and JBuilder for developing J2EE/JAVA applications.
- Good working knowledge with Struts II, Web Logic, Web sphere and JDK
- Experience with design Patterns like MVC, Singleton, Factory, Proxy, DAO, Abstract, Prototype and Adaptor.
- Proficient in writing and handling SQL Queries, Stored Procedures, and triggers.
- Hands on experience in knowledge of user acceptance, Black Box, White box and Unit testing.
- Knowledge of multi vendor operating systems including Linux, Windows and UNIX Shell Script.
- Willing to learn and adapt to new technologies.
TECHNICAL SKILLS
Programming Languages: Java, C, C++
Hadoop Ecosystem: Map-Reduce, Hive, Pig, HBASE, Flume, Sqoop, ZooKeeper, Oozie, Cloudera Java API/FrameworksServlets, web services, SOAP, Rest, JSP, DBC, JNDI, EJB, RMI, JPA, JMS, Struts II, Spring, Hibernate.
Web Technologies: HTML, DHTML, DOM, JavaScript, php, AJAX, CSS
XML Technologies: XML, DOM, SAX, XSL, XSLT, JNDI, WSDL
Application/Web Servers: WebLogic, GlassFish, IBM WebSphere, Apache Tomcat
Databases: SQL Server, MySQL, PL/SQL, Oracle, DB2
IDE Tools: Eclipse, JBuilder, NetBeans, Visual Studio
Operating Systems: Windows, UNIX, LINUX.
Tools: Tableau, Rational Rose, JUnit, Log4j, JConsole, JProbeDesign MethodologiesUML, Rational Rose
PROFESSIONAL EXPERIENCE
Confidential, Pittsburgh, PA
Hadoop Developer
Responsibilities:
- Installed and configured multi node Hadoop cluster.
- Involved in installing Hadoop Ecosystem components.
- Responsible for managing data ingestion from multiple source
- Involved in Hadoop cluster maintenance that includes security, cluster connectivity, adding and removing nodes to the cluster.
- Importing and exporting data into HDFS from Oracle 10.2 database and vice versa using SQOOP.
- Involved in running Hadoop streaming jobs to processes terabytes of structured and semi structured data.
- Gained experience in reviewing and managing Hadoop log files.
- Involved in writing Hive queries for data analysis with respect to business requirements.
- Wrote Map-Reduce jobs to generate reports for the number of activities created on a particular day, during a dumped from the multiple sources and the output was written back to HDFS.
- Worked with Flume to import the log data from the reaper logs and syslog’s into the Hadoop cluster.
- Created workflow and coordinator using Oozie for regular jobs.
- Performed cluster capacity planning, performance tuning and monitoring and trouble shooting.
- Wrote Hive and Pig scripts as ETL tool to do transformations, event joins, filter both traffic and some pre-aggregations before storing into the HDFS.
- Worked on a Recommendation Engine, supervised by research team, built using Apache Mahout by classifying users into groups based on their attributes.
- Involved in managing running and pending tasks Map Reduce through Cloudera manager console.
- Worked on streaming analyzed stream of data in to relational database system using Sqoop for making for making it available for visualization and report generation by the BI team.
- Hands on experience with NoSQL databases like HBase, Cassandra for POC (proof of concept) in storing URL's, images, products and supplements information at real time.
- Created adhoc reports to users in Tableau by connecting HDFS data sources using Impala.
- Involved in preparing project maintenance, Test results, Test summary and Go-Live plan documents.
Environment: Apache Hadoop, Java, Eclipse, Hive, MapReduce, Pig, Sqoop, Oozie, Cassandra, NoSQL, HBase, Map-Reduce, Cloudera, HDFS, Flume, Impala, Apache Mahout, Oracle 10.2, Tableau, Agile, SVN, Git, SourceTree, Jira, Confluence
Confidential, Houston, TX
Hadoop Developer
Responsibilities:
- Setup and monitoring Hadoop development environment.
- Involved in Importing and exporting data from MySQL database into HDFS using Sqoop.
- Designed and developed Map-Reduce programs
- Involved in running Hadoop streaming jobs to process Terabytes of data.
- Load and Transform large sets of structured data.
- Worked on analyzing Hadoop Cluster and other Big Data analytics tools including Pig, Hbase database and Sqoop.
- Coordinated clusters services using ZooKeeper.
- Involved in Writing Data Refinement Pig Scripts and Hive Queries.
- Used Flume to transport logs to HDFS.
- Assessed and optimized the legacy scripts and programs along with solving existing production issues.
- Involved in implementing web interface for UAS using Oracle ADF and JDeveloper.
- Designed the web interface for the marketing user to enter the data and submit the request.
- Configured connection between HDFS and Tableau using Impala for Tableau developer team.
- Designed the page navigation flow using the ADF task flow.
- Involved in optimizing the search query and designed the database link, materialized views and oracle scheduler.
Environment: Apache Hadoop, Java, Eclipse, Hive, Pig, Sqoop, ZooKeeper, NoSQL, HBase, Map-Reduce, Cloudera, HDFS, Flume, Impala, Tableau, MySQL, Oracle ADF, JDeveloper.
Confidential, New York, NY
Java/Hadoop Developer
Responsibilities:
- Modeling conceptual design using Use Case, UML Class and Activity diagrams using Rational Rose.
- Created user interface views using JSP, JavaScript, HTML, AJAX, CSS, and XSLT.
- Wrote requirement specific SQL and PL/SQL scripts including Stored Procedures, functions, packages and triggers.
- Implemented Database access through JDBC at Server end with Oracle.
- Used Spring Aspect Oriented Programming (AOP) for addressing cross cutting concerns.
- Developed request/response paradigm by using Spring Controllers, Inversion of Control and Dependency Injection with Spring MVC.
- Used Web Services like SOAP and WSDL to communicate over internet.
- Involved in implementation of the JMS Connection Pool, including publish and subscribe using Spring JMS.
- Used CVS for version control and Log4j for logging.
- Used JProbe and JConsole to profile application for memory leaks and resource utilization.
- Developed test classes in JUnit for implementing unit testing.
- Deployed the application using WebLogic Application Server
- Switched to Hadoop technologies by setting up Hadoop Ecosystem components on a cluster.
- Developed Java MapReduce programs for the analysis of sample log file stored in cluster.
- Developed Simple to complex MapReduce Job scripts using Hive and Pig.
- Used Sqoop to import data into HDFS and Hive from other data systems.
- Used Hive and created Hive tables and involved in data loading and writing requirement specific Hive UDFs.
- Developed Map Reduce Programs for data analysis and data cleaning.
Environment: Java, JDK1.5, J2EE, JDBC, Servlets, JSP, MapReduce, JSTL, HTML, AJAX, Spring Framework, Java Script, CSS, XSLT, XML. JUnit, WebServices, SOAP, WSDL, Weblogic, CVS, Rational Application Developer RAD, Hibernate, Rational Rose, JMS, Hadoop, HDFS, Map Reduce, Hive, Pig, UDF, Sqoop.
Confidential, Cleveland, Ohio
J2EE/JAVA Developer
Responsibilities:
- Involved in class diagrams and sequence diagrams for the design phase of the project using UML.
- Responsible for writing code for dispatch action, form bean classes using Struts Framework.
- Designed applications using Struts with Tiles and Validator, implementing MVC design pattern and writing Custom Tag Libraries, JSP, Java Beans, Struts Controller, Action and Action Form classes using Struts tag libraries.
- Used XML Technologies like DOM for transferring data.
- Object relational mapping and Persistence mechanism is executed using Hibernate ORM.
- Developed custom validator in Struts and implemented server side validations using annotations.
- Created struts-config.xml file for the Action Servlet to extract the data from specified Action form so as to send it to specified instance of action class.
- Implemented the business logic using Session Beans of the Enterprise application and for integration of EJBs into Struts Framework worked with the Service Locator and Business Delegate layers.
- Used JMS Java Messaging Service for messaging and also worked with Message Driven Beans MDBs for Synchronous and Asynchronous Messaging.
- Used Oracle for the database and WebLogic as the application server.
- Involved in coding for DAO Objects using JDBC (using DAO pattern).
- Involved in writing Oracle Stored Procedures and access them through Java Programs.
Environment: Windows, WebLogic, Eclipse, Struts, Oracle, J2EE, Java, JDBC, JSP, JavaScript, EJB, JDBC, JNDI, DOM, JMS.
Confidential, Milipitas, CA
Java/J2EE Developer
Responsibilities:
- Used MVC based Spring framework for Web tier development using Eclipse IDE.
- Participated in client's demo sessions and involved in integration and planning of project.
- Created Data Access Objects using Spring Framework.
- Used Spring Aspect Oriented Programming (AOP) for addressing cross cutting concerns.
- Consumed Web Services for transferring data between different applications.
- Developed Web based (JSP, Servlets, Java Beans, JavaScript, CSS, XHTML) console for feed monitoring, reporting and life cycle management.
- Used XSLT to transform XML documents into HTML Documents.
- JDBC connectivity was established using Oracle thin client.
- Developed Ant build scripts for compiling and building the project.
- Performed comprehensive unit, system and regression testing.
Environment: JDK 1.5, J2EE, Eclipse, JSP, JSTL, Servlets, Spring, JDBC, JavaScript, Ant, JQuery, XML, XSL, Junit, HTML, XHTML, DHTML, CSS, Oracle, Windows XP, Web 2.0, Web Logic Application Server, Web Services, CVS.