Hadoop Developer Resume
Chicago, IL
SUMMARY
- Almost 7 years of professional experience in IT industry with hands - on experience in Developing, Implementing and maintenance of various applications using Java, J2EE technologies.
- Over 3+ years of experience in Hadoop ecosystem's implementation, maintenance and Big Data analysis operations.
- Excellent understanding of HDFS, Map Reduce, YARN, and tools including Pig and Hive for data analysis, Sqoop for data migration, Flume for data ingestion, Oozie for scheduling and Zookeeper for coordinating cluster resources.
- Worked on analyzing Hadoop cluster and different big data analytic like Hbase.
- Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Involved in creating Hive Tables, loading with data and writing Hive queries, which will invoke and run MapReduce jobs in the backend.
- ETL testing experience withTalendOpen Studio
- Experience in writing HiveQL queries to store processed data into Hive tables for analysis.
- Experience in building Pig scripts to extract, transform and load data onto HDFS for processing.
- Knowledge and understanding on industry latest Hadoop ecosystems like Apache Spark integration with Hadoop.
- Loaded streaming log data from various webservers into HDFS using Flume.
- Experience in data migration from RDBMS to Cassandra.
- Have the motivation to take independent responsibility as well as ability to contribute and be a productive team member.
- Experience in monitoring and managing 100+ node Hadoop cluster.
- Created a complete processing engine based on Cloudera distribution.
- Experienced in automating job flows using Oozie.
- Supported Map Reduce programs running on the cluster.
- Worked with application team via scrum to provide operational support, install Hadoop updates, patches and version upgrades as required.
- Monitor Hadoop cluster using tools like Nagios, Ganglia Ambari and Cloudera Manager.
- Worked with system engineering team to plan and deploy Hadoop hardware and software environments.
- Worked on disaster management with Hadoop cluster.
- Built ingestion framework using flume for streaming logs and aggregating the data into HDFS.
- Built data transform framework using Map Reduce and Pig.
- Designed, delivered and helped manage a device data analytics at a very large storage vendor.
- Worked with business users to extract clear requirements to create business value.
- Worked with big data teams to move ETL tasks to Hadoop.
- Experienced in Linux Administration tasks like IP Management (IP Addressing, Sub netting, Ethernet Bonding, and Static IP)
- Good communication and interpersonal skills, a committed team player and a quick learner.
- Manage and Review Hadoop Log.
TECHNICAL SKILLS
Languages: Java, Hadoop, Hadoop Cascading, Elasticsearch, JUnit, C, C++, SQL, PL/SQL
Web Technologies: JSP, Spring, Spring REST, HTML, CSS, JavaScript, JQuery
J2EE Technologies: Servlets, Web Services, SOAP
Databases: NoSQL, Oracle, DB2, MySQL, SQLite, MS SQL Server, MS Access
Tools: Eclipse, IntelliJ, Dia, NetBeans, GitHub, Dropbox, Visual Studio 2010/2012
Platforms: Windows 2003/2008/ XP/7/8/10, UNIX
SDLC: Agile, Rapid Application Development, Waterfall Model, Iterative Model
Design Patterns: Singleton, Adapter, Builder, Iterator, Template
Web Services: WebLogic, WebSphere, Apache Tomcat, JBoss
Frameworks: Hibernate, EJB, Struts, Spring, Grails
Other: SVN, Maven, ANT
PROFESSIONAL EXPERIENCE
Confidential, Chicago, IL
Hadoop Developer
Responsibilities:
- Helped the team to increase cluster size from 35 nodes to 113 nodes. The configuration for additional data nodes was managed using Puppet.
- Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
- Importing and exporting data into RDBMS and Hive using Sqoop.
- Able to partitioning a Hive table, creating an external table and differences between the managed and external tables.
- Optimized HIVE analytics SQL queries and achieve job performance.
- Created and worked Sqoop jobs with incremental load to populate Hive External tables.
- Developed Pig scripts in the areas where extensive coding needs to be reduced.
- Developed ETL jobs on Informatica &Talendusing Oracle, Vectorwise & Greenplum database
- Extensive experience in writing Pig scripts to transform raw data from several data sources into forming baseline data.
- Created HBase tables to store variable data formats of data coming from different portfolios.
- Experienced in Map Reduce programs to load the data from system generated log file to HBase database.
- Very good understanding of Partitions, Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
- Developed backend (server side) in Scala.
- Design technical solution for real-time analytics using Kafka and HBase.
- Solved performance issues in Hive and Pig scripts with understanding of Joins, Group and aggregation and how does it translate to Map Reduce jobs.
- Designing conceptual model with Spark for performance optimization.
- Developed Oozie workflow for scheduling and orchestrating the ETL process
- Developed Map Reduce programs to parse the raw data and store the refined data in tables.
- Analyzing data with Hive, Pig and Hadoop Streaming
- Worked on creating the Data Model for Cassandra from the current Oracle Data model.
- Worked with CQL to execute queries on the data persisting in the Cassandra cluster.
- To analyze data migrated to HDFS, used Hive data warehouse tool and developed Hive queries.
- Used Tableau for visualizing and to generate reports.
- Used Flume to collect, aggregate, and store the log data from different web servers.
Environment: Hadoop, Map Reduce, Hive, Pig, Hbase, Sqoop, Flume, Cassandra, Scala, Spark, Oozie, Kafka, Linux, Java, Tableau, Eclipse, HDFS,Talend, PIG, Java (JDK), MySQL and Ubuntu.
Confidential, Englewood, CO
Hadoop Developer
Responsibilities:
- Implemented code according to the business requirements.
- Contributed to reporting and analytic solutions to enhance healthcare management. Implemented features by using Hadoop Cascading/Map Reduce
- Involved in migration of Datastage jobs intoTalendopen source
- Responsible for building scalable distributed data solutions using Hadoop
- Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring and Troubleshooting, Manage and review data backups and log files.
- Responsible for running Hadoop streaming jobs to process terabytes of csv data.
- Load and transform large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
- Involved in loading data from UNIX file system to HDFS.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
- Extracted the data from Teradata into HDFS using the Sqoop.
- Exported the patterns analyzed back to Teradata using Sqoop.
- Involved in production support.
- Used Java code conventions and language standards for maintainable and documented code.
- Resolved bugs via replicating bug, applying query in Elasticsearch
- Developed automated test tools using JUnit.
- Used Maven for automated building of the Java files
- Used GIT as version control to checkout and check-in of files.
Environment: Java, Python, Big Data, Hadoop, Bash, Linux, Protobuff, Apache API, Hadoop Cascading, Map Reduce, Groovy on Grails, NoSQL (Elasticsearch), Dia, IntelliJ, IDEA, Maven, GIT, Talend, Sense (Chrome Extension).
Confidential, Columbus, OH
Java Developer
Responsibilities:
- Designed and Developed the Solution Architecture package to identify all possible solutions for the project using UML methodologies.
- Used RUP (Rational Unified Process) methodologies in design and implementation of the application
- Used J2EE Design patterns (Singleton, Service Locator and DAO).
- Implemented MVC Architecture using JSF and Struts framework and custom tag libraries.
- Developed a multi-user web application using JSP, JSTL, Action Forms, Action Classes, JavaScript and JDBC to provide the needed functionality.
- Responsible for creating the server side POJO’s and mapped these in Hibernate configuration files.
- Used Hibernate for data manipulation and access.
- Used Spring AOP and dependency injection framework for business aspect of the application.
- Used Spring DAO classes for data access. Testing - unit testing & integration testing of the application.
- Developed Java beans and JSP's by using Spring and JSTL tag libs for supplements.
- Designed and developed middleware, using RESTful web services based on a centralized schema which is consumed by thousands of users.
- Used DOM/SAX parsers and JAXB for the conversion of XML files to java objects and vice-versa.
- Used RAD 7.0 for implementing Static and Dynamic web services for consuming and providing services related to the business.
- Created creating tables, data manipulation and retrieval using Oracle.
- Used MQ Series to send, receive and read the messages and also Created Queues.
- Involved in the deployment process of the application on WebSphere Application Server.
- SVN are used as source code controls and developed test cases using JUnit.
- Deployed the Web application using ANT build script
- Used Mercury Quality Center for requirement management, executing the test plans and defect management.
- Used SOA architecture for communication between various business entities.
- Communicated with the business people to get more clarity on the requirements/changes.
Environment: Java, Spring, JSF, JavaScript, JDBC, XML, JAXB, Web services, SOAP, WSDL, Hibernate, Spring, IBM MQ, PL/SQL, WebSphere, RAD 7.0, JUnit, Oracle 10g, Mercury Quality Center, SVN
Confidential, Dallas, TX
Java Developer
Responsibilities:
- Participate in project planning sessions with business analysts and team members to analyze business IT requirements and translate business requirements into working model.
- Involve in Initial designing and creating Use case diagrams, Sequence Diagrams and class diagrams using the MS Visio Tool.
- Used Agile software development methodology which helps to develop project plans
- Implemented MVC, DAO J2EE design patterns as a part of application development.
- Involved in developing applications using Java, J2EE and Struts.
- Developed Action Form classes, Form beans, Action Classes using Struts.
- Designed and developed user interfaces using JSP, HTML and CSS.
- Used JavaScript and AJAX validation framework for performing front end validations.
- Hibernate is used to persist the objects in the Database.
- Developed DAO design pattern for hiding the access to data source objects.
- Used Eclipse as IDE tool to develop the application and JIRA for bug and issue tracking
- Used CVS for software configuration management and version control
- Deploy the application on the WebLogic Application Server.
- Tested the applications with the JUnit, Also written some test cases.
Environment: OOAD, UML, Java, J2EE, Servlets, Struts (MVC), Spring, Hibernate 3.0, WebLogic, Eclipse IDE, JSF, JSP, EJB, HTML, JavaScript, DHTML, CSS, JQuery, SOAP, WSDL, Junit
Confidential, The Woodlands, TX
Java Developer
Responsibilities:
- Designed and developed various modules of the application with J2EE design architecture and frameworks like Spring MVC architecture and Spring Bean Factory using IOC, AOP concept.
- Followed agile software development with Scrum methodology.
- Wrote application front end with HTML, JSP, JSF, Ajax, JQuery and XHTML.
- Used J Query for UI centric Ajax behavior.
- Implemented JAVA/J2EE design patterns such as Factory, DAO, Session Façade and Singleton.
- Used Hibernate in persistence layer and developed POJO's, Data Access Object (DAO) to handle all database operations.
- Developed the Form Beans and Data Access Layer classes.
- XML was used to transfer the data between different layers.
- Involved in writing complex sub-queries and used Oracle for generating on-screen reports.
- Worked on database interaction layer for insertions, updating and retrieval operations on data. Used JMS for messaging.
- Used Log4j to assign, track, report and audit the issues in the application.
- Develop and execute Unit Test plans using JUnit, ensuring that results are documented and reviewed.
- Quality Assurance teams responsible for integrated testing.
- Worked in deadline driven environment with immediate feature release cycles.
Environment: Java, J2EE, JSP, Servlets, Hibernate, UML, HTML, XHTML, DHTML, JavaScript, JQuery, CSS, XML, JBoss, Log4j, Oracle, JUnit, Eclipse.