Hadoop Developer Resume
Trenton, NJ
SUMMARY
- Over 7 years of professional IT experience with Big Data Ecosystem experience in ingestion, storage, querying, processing and analysis of big data.
- Hands on experience in installing, configuring and using ecosystem components like Hadoop MapReduce, HDFS, HBase, Zoo Keeper, Hive, Sqoop, Pig, Flume, Cassandra, Cloudera and Horton Works.
- Experience in building, maintaining multiple Hadoop clusters of different sizes and configuration and setting up the rack topology for large clusters.
- Good Understanding of Hadoop architecture and Hands - on experience with Hadoop components such as JobTracker, TaskTracker, NameNode, DataNode and MapReduce concepts and HDFS Framework.
- Well versed with Developing and Implementing MapReduce programs using Hadoop to work with BigData.
- Experience with NoSQL databases like HBase and Cassandra.
- Wrote Java based api to access Cassandra using the DataStax API and the JDBC API to Cassandra.
- Working with relative ease with different working strategies like Agile, Waterfall and Scrum methodologies
- Detailed knowledge and experience of Design, Development and Testing Software solutions using Java and J2EE technologies.
- Expertise in design and development of various web and enterprise applications using various technologies like JSP, Servlets, Struts, Hibernate, Spring, JDBC, XML, AJAX, SOAP and Web Services.
- Experience in Database design, Entity relationships, Database analysis, Programming SQL, Stored procedure’s PL/ SQL, Packages and Triggers in Oracle and SQL Server on Windows and LINUX.
- Strong understanding of Data warehouse concepts, ETL, data modeling experience using Normalization, Business Process Analysis, Reengineering, Dimensional Data modeling, physical & logical data modeling.
- Worked in Informatica for the Extraction, Transformation and Loading from various sources to the enterprise data warehouse Developed and tested extraction, transformation and load (ETL) processes.
- Experience with front end technologies like HTML, CSS and Javascript.
- Hands on experience on tools such as Eclipse, Git, Cygwin, Toad, Xml SPY, Rational Rose, Linux vi-editor and project management tools like Clear case and SVN.
- Experience in writing Shell scripts using ksh, bash, and perl, for process automation of databases, applications, backup and scheduling.
- Strong analytical skills with ability to quickly understand clients business needs. Involved in meetings to gather information and requirements from the clients. Leading the Team and involved in Onsite, Offshore co-ordination.
- Research-oriented, motivated, proactive, self-starter with strong technical, analytical and interpersonal skills.
TECHNICAL SKILLS
Big Data Technologies: Hadoop, HDFS, Hive, Map Reduce, Pig, Sqoop, Flume, ZookeeperScripting Languages Python, Perl, Shell
Programming Languages: C, C++, Java
Web Technologies: HTML, J2EE, CSS, JavaScript, AJAX, Servlets, JSP, DOM, XML, XSLTXPATH.
Java Frameworks: Struts, Spring, Hibernate
Application Server: Websphere, Weblogic Server, Apache Tomcat.
DB Languages: SQL, PL/SQL
Databases /ETL: Oracle 9i/10g/11g, MySQL 5.2, DB2, Informatica v 8.x
NoSQL Databases: Hbase, Cassandra, Mango DB.
Operating Systems: Linux, UNIX, Windows 2003 Server
PROFESSIONAL EXPERIENCE
Confidential, Trenton, NJ
Hadoop Developer
Responsibilities:
- Responsible for loading the customer’s data and event logs from MSMQ into HBase using REST API.
- Responsible for architecting Hadoop clusters with CDH4 on CentOS, managing with Cloudera Manager.
- Involved in initiating and successfully completing Proof of Concept on FLUME for Pre-Processing, Increased Reliability and Ease of Scalability over traditional MSMQ.
- Involved in loading data from LINUX file system to HDFS.
- Importing and exporting data into HDFS and Hive using Flume.
- Used Hive to find correlations between customer’s browser logs in different sites and analyzed them to build risk profile for such sites.
- End-to-end performance tuning of Hadoop clusters and Hadoop Map/Reduce routines against very large data sets.
- Developed the Pig UDF’S to pre-process the data for analysis
- Monitored Hadoop cluster job performance and performed capacity planning and managed nodes on Hadoop cluster.
- Proficient in using Cloudera Manager, an end to end tool to manage Hadoop operations.
Environment: Cloudera Distribution, CDH4, FLUME, HBase, HDFS, Pig, MapReduce, Hive
Confidential, Alpharetta, GA
Sr. Hadoop Developer
Responsibilities:
- Involved in architecture design, development and implementation of Hadoop deployment, backup and recovery systems.
- Worked on the proof-of-concept for Apache Hadoop framework initiation on Amazon Web Sevices.
- Experience in HDFS, MapReduce and Hadoop Framework
- Trained and guided the team on Hadoop framework, HDFS, MapReduce concepts.
- Developed MapReduce jobs for Log Analysis, Recommendation and Analytics.
- Wrote MapReduce jobs to generate reports for the number of activities created on a particular day, during a dumped from the multiple sources and the output was written back to HDFS
- Reviewed the HDFS usage and system design for future scalability and fault-tolerance.
- Installed and configured Hadoop HDFS, MapReduce, Pig, Hive, Sqoop.
- Wrote Pig Scripts to generate MapReduce jobs and performed ETL procedures on the data in HDFS.
- Processed HDFS data and created external tables using Hive, in order to analyze visitors per day, page views and most purchased products.
- Exported analyzed data to HDFS using Sqoop for generating reports.
- Used MapReduce and Sqoop to load, aggregate, store and analyze web log data from different web servers.
- Developed Hive queries for the analysts.
- Cluster co-ordination services through ZooKeeper
- Experience in optimization of Map reduce algorithm using combiners and partitions to deliver the best results and worked on Application performance optimization for a HDFS/Cassandra cluster.
Environment: Mapreduce,AWS,EMR,S3,EC2,Hive,Pig,Sqoop,Oracle,MapR,Cloudera,ManagerZooKeeper.
Confidential, Chantilly, VA
Senior Java Developer
Responsibilities:
- Responsible for understanding the scope of the project and requirement gathering.
- Developed the web tier using JSP, StrutsMVC to show account details and summary.
- Created and maintained the configuration of the SpringApplicationFramework (IoC).
- Implemented various design patterns - Singleton, Business Delegate, Value Object and Spring DAO.
- Used SpringJDBC to write some DAO classes to interact with the database to access account information.
- Mapped business objects to database using Hibernate.
- Involved in writing SpringConfigurationXML files that contains declarations and other dependent objects declaration.
- Used Tomcat web server for development purpose.
- Involved in creation of Test Cases for JUnit Testing.
- Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
- Used CVS, Perforce as configuration management tool for code versioning and release.
- Developed application using Eclipse and used build and deploy tool as Maven.
- Used Log4J to print the logging, debugging, warning, info on the server console.
Environment: Java, J2EE, XML, Spring, Struts, Hibernate, Design Patterns,Maven, Eclipse, Toad, Apache Tomcat and Oracle 11g.
Confidential, Columbus, Ohio
Java Developer
Responsibilities:
- Involved in requirement gathering, functional and technical specifications.
- Monitoring and fine tuning IDM performance
- Enhancements in the self-registration process.
- Fixing the existing bugs in various releases
- Global deployment of the application and co-ordination between the client, development team and the end users.
- Setting up of the users by reconciliations, bulk load and bulk link in all the environments.
- Wrote requirements and detailed design documents, designed architecture for data collection.
- Developed OMSA UI using MVC architecture, Core Java, Java Collections, JSP, JDBC, Servlets, ANT and XML within a Windows and UNIX environment.
- Used Java Collection Classes like Array List, Vectors, Hash Map and Hash Table.
- Used Design Patterns MVC, Singleton, Factory, Abstract Factory.
- Developed algorithms and coded programs in Java.
- Co-ordinate with different IT groups and Customer.
- Performed all types of testing includes Unit testing, Integration and testing environments.
- Used Combination of client and server side validation using Struts validation framework.
Environment: JAVA, STL's, Design Patterns, Oracle, SQL/ PL SQL.
Confidential
Software Developer - I
Responsibilities:
- Involved in the design, coding, deployment and maintenance of the project.
- Involved in design and implementation of web tier using Servlets and JSP.
- Performed client side validations using Java Script.
- Used Apache POI for Excel files reading.
- Written build scripts with Ant for deploying war and ear applications.
- Configured connection pools and establish a connection with MySQL.
- Used technologies like JSP, JSTL, JavaScript and Tiles for Presentation tier
- Involved in JUnit testing of the application using JUnit framework.
Environment: Java, J2EE, Tomcat, MySQL, Eclipse, Apache POI, Java Script, CSS, HTML.