Sr. Hadoop Developer Resume
Princeton, NJ
SUMMARY:
- Over 9+ years of professional IT experience which includes 3+ years of experience in Have around 3 years of experience with Hadoop Map Reduce, HDFS and Hadoop Ecosystems like Bigdata, HDFS, MapReduce, Oozie, Cassandra, Hive, Sqoop, Pig, Flume, Hbase and Zookeeper and 5 years in Java and Oracle PLSQL development.
- 9+ years of experience in development of applications using Object Oriented Programming.
- In - depth knowledge of Hadoop architecture and its components like HDFS, Name Node, Data Node, Job Tracker, Application Master, Resource Manager, Task Tracker and Map Reduce programming paradigm.
- Experience in cluster planning, designing, deploying, performance tuning, administering and monitoring Hadoop ecosystem.
- Commendable knowledge / experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.
- Experience in developing Map/Reduce jobs to process large data sets utilizing the Map/Reduce programming paradigm.
- Good understanding of cloud configuration in Amazon web services (AWS).
- Experience in database design. Used PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle.
- Proficient in writing SQL, PL/SQL stored procedures, functions, constraints, packages and triggers.
- Good experience in Hive tables design, loading the data into hive tables.
- Good understanding of HDFS Designs, Daemons, federation and HDFS high availability (HA).
- Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
- Hadoop Shell commands, Writing Map reduce Programs, Verifying the Hadoop Log Files.
- Exposure on Query Programming Model of Hadoop.
- Expert on UML for Object Oriented Analysis & Design (OOAD) using MS Visio, IBM Rational.
- Expert on Core Java, multi-threading, debugging JVM and optimizing and profiling Java Applications.
- Experience on System Study, Analysis, of Business requirement, preparation of Technical design, UTP and UTC, Coding, Unit testing, Integration testing, System testing and Implementation.
- Experience in Object Oriented Analysis and Design (OOAD) and development of software using UML methodology.
- Hands on experience with Core Java with Multithreading, Concurrency, Exception Handling, File handling, IO, Generics and Java Collections.
- Implemented rich web applications such as HTML, XHTML, XML, XSLT, CSS, JavaScript, AJAX(DWR), jQuery, ExtJS, JSON, and DOJO.
- Excellent working knowledge of MVC architecture and Struts, Spring MVC and JSF Frameworks.
- Developed applications using Core Java, Servlets, JSP, JDBC, Struts, Spring, Hibernate.
- Good understanding of the SOA technologies like SOAP, WSDL Web Services.
- Knowledge of Software Development Methodologies like Agile (SCRUM), Waterfall.
- Proficient in using application servers like JBoss and Tomcat Servers.
- Configured and deployed applications on IBM Web sphere, BEA Web logic, Tomcat.
- Excellent working knowledge of Service Oriented Architecture(SOA), Messaging and Web Services.
- Experienced on developing, building and deploying applications on UNIX, Linux, Solaris and Windows platforms.
- Experienced in database design and development and JDBC connectivity for Oracle 11g/10g/9i/8i (SQL, PL/SQL, Stored procedures), MS SQL Server 2008/2005/2000, DB2 9.x/8.x and MySQL.
- Working knowledge of Java external applications like JUnit, Log4J, Apache Ant, Maven.
- Experienced in building and deploying applications on servers using Ant, Maven, and Perl.
- Worked with query tools like Toad, SQL*Plus, SQL Developer.
- Expert level skills in Designing and Implementing web servers solutions and deploying Java Application Servers like Websphere, Web Logic, configuring Apache Web Server and configuring various Servlet engines.
- Comprehensive knowledge of physical and logical data modeling, performance tuning.
- Resourceful and skilled in analyzing and solving problems.
- Extensive experience in writing and executing JUnit Test cases & debugging Java/J2ee applications.
- Hands on working experience with different version management software such as VSS, Win CVS, Subversion, Star Team and SVN.
- Excellent written, verbal communication, and customer service skills.
- Strong organizational, and interpersonal skills. And possess a high level of drive, initiative and self-motivation.
- A collaborative personality who enjoy working in a team-oriented environment.
- Excellent debugging skills. Able to debug complex technical issues including multiple system components.
- Highly creative and articulate. Can adapt quickly to rapidly changing conditions.
TECHNICAL SKILLS:
Big Data Technology: HDFS, Mapreduce, HBase, Pig, Hive, SOLR, Sqoop, Flume, MongoDB, Cassandra, Puppet, Oozie, Zookeeper, Spark, Kafka
Java/J2EE Technology: JSP, JSF, Servlets, EJB, JDBC, Struts, Spring, Spring MVC, Spring Portlet, Spring Web Flow, Hibernate, iBATIS, JMS, MQ, JCA, JNDI, Java Beans, JAX-RPC, JAX-WS, RMI, RMI-IIOP, EAD4J, Axis, Castor, SOAP, WSDL, UDDI, JiBX, JAXB, DOM, SAX, MyFaces(Tomahawk), Facelets, JPA, Portal, Portlet, JSR 168/286, LifeRay, WebLogic Portal, LDAP, JUnit.NET
Languages: Java (1.4/5/6), C/C++, Swing, SQL, HTML, CSS, i18n, l10n, DHTML, XML, XSD, XHTML, XSL, XSLT, XPath, XQuery, SQL, PL/SQL, UML, JavaScript, AJAX(DWR), jQuery, Dojo, ExtJS, Shell Scripts, Perl
Development Framework/IDE: RAD 8.x/7.x/6.0, IBM WebSphere Integration Developer 6.1, WSAD 5.x, Eclipse Galileo/Europa/3.x/2.x, MyEclipse 3.x/2.x, NetBeans 7.x/6.x, IntelliJ 7.x, Workshop 8.1/6.1, Adobe Photoshop, Adobe Dreamweaver, Adobe Flash, Ant, Maven, Rational Rose, RSA, MS Visio, OpenMake Meister
Web/Application Servers: WebSphere Application Server 8.x/ 7.0/6.1/5.1/5.0, WebSphere Portal Server 7.0/6.1, WebSphere Process Server 6.1, WebLogic Application Server 8.1/6.1, JBoss 5.x/3.x, Apache 2.x, Tomcat 7.x/6.x/5.x/4.x, MS IIS, IBM HTTP Server
Databases: NoSQL, Oracle 11g/10g/9i/8i, DB2 9.x/8.x, MS SQL Server 2008/2005/2000, MySQL
NoSQL: HBase, Cassandra, MongoDB, Accumulo
Operating Systems: Windows XP, 2K, MS-DOS, Linux (Red Hat), Unix (Solaris), HP UX, IBM AIX
Version Control: CVS, SourceSafe, ClearCase, Subversion, AllFusion Harvest Change Manager 7.1
Monitoring Tools: Embarcadero J Optimizer 2009, TPTP, IBM Heap Analyzer, Wily Introscope, JMeter
Other: JBoss Drools 4.x, REST, IBM Lotus WCM, MS ISA,CA SiteMinder, BMC WAM, Mingle
PROFESSIONAL EXPERIENCE:
Confidential, Princeton, NJ
Sr. Hadoop Developer
Environment: Hadoop (Cloudera HA cluster), HDFS, UNIX Shell Scripting, Python, Sqoop, FTP, Impala, Pig, Hive, Hbase, Yarn, Tableau.
Responsibilities:
- All the datasets are loaded from two different sources such as Oracle, MySQL to HDFS and Hive respectively on daily basis.
- Process 8 flat files all are delimitated by Comma.
- Responsible in creating Hive Tables to load the data which comes from MySQL and loading data from Oracle to HDFS using Sqoop
- Good hands on experience in writing core java level programming in order to perform cleaning, pre-processing and data validation.
- Involved in verifying cleaned data using Talend tool with other department.
- Experienced in creating Hive schema, external tables and managing views.
- Involved in developing Hive UDFs and reused in some other requirements.
- Worked on performing Join operations.
- Involved in creating partitioning on external tables.
- Good hands on experience in writing HQL statements as per the user requirements.
- Fetching the HQL results into CSV files and handover to reporting team.
- Work with hive complex datatypes and involved in Bucketing.
- Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Scala and have a good experience in using Spark-Shell and Spark Streaming.
- Develop Spark code using Scala and Spark-SQL for faster testing and data processing.
- Import millions of structured data from relational databases using Sqoop import to process using Spark and stored the data into HDFS in CSV format.
- Use Spark SQL to process the huge amount of structured data.
- Implement Spark RDD transformations, actions to migrate Map reduce algorithms.
- Assign name to each of the columns using case class option in Scala.
- Explore with the Spark for improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
- Use Data Frame API in Scala for converting the distributed collection of data organized into named columns.
- Register the datasets as Hive Table.
- Develop solutions to pre-process large sets of structured, with different file formats (Text file, Avro data files, Sequence files, Xml and JSON files, ORC and Parquet).
- Experience with batch processing of data sources using Apache Spark.
- Develop predictive analytic using Apache Spark Scala APIs.
- Expert in implementing advanced procedures like text analytics and processing using the in-memory computing capabilities like Apache Spark written in Scala.
- Expertise in running Hadoop streaming jobs to process terabytes data.
- Experience in importing the real time data to Hadoop using Kafka and implemented the Oozie job.
- Responsible in analysis, design, testing phases and responsible for documenting technical specifications.
- Along with the Infrastructure team, involved in design and developed Kafka and Storm based data pipeline.
- Develop storm-monitoring bolt for validating pump tag values against high-low and * Worked on Talend Administrator Console (TAC) for scheduling jobs and adding users
- Develop Kafka producer and consumers, Hbase clients, Spark and Hadoop MapReduce jobs along with components on HDFS, Hive.
- Good knowledge in partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
- Work on the core and Spark SQL modules of Spark extensively.
Confidential, New York, NY
Hadoop Developer
Environment: Hadoop (Cloudera), HDFS, Hive, Impala, Pig, UNIX Shell Scripting, Python, Flume, Oozie, NOSQL, SQOOP, Git
Responsibilities:
- Worked on analyzing Hadoop cluster and different Big Data analytic tools including Pig, Hive HBase database, Storm, Kafka, Elastic Search, Redis, Flume, Scoop and SQOOP.
- Installed Hadoop, Map Reduce, HDFS, and Developed multiple map reduce jobs in PIG and Hive for data cleaning and pre-processing.
- Coordinated with business customers to gather business requirements.
- Importing and exporting data into HDFS and Hive using SQOOP.
- Wrote Map Reduce code to convert semi Structured Data to Structured data.
- Developed a Framework that will create external and manageable tables in a batch processing based on the metadata files.
- Successfully designed and developed a solution for speeding up a SQL Job using Hadoop Map-Reduce framework. Processing time was reduced from 12 hours to 20 Minutes.
- Wrote extensive Hive queries and fine-tuned them for performance as part of the multiple step process to get the required results for Tableau to generate reports.
- Wrote Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
- Involved in creating Hive tables, loading with data and writing hive queries that will run internally in map reduce way.
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
- Experienced in managing and reviewing the Hadoop log files.
- Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS.
- Load and Transform large sets of structured and semi structured data.
- Responsible to manage data coming from different sources.
- Involved in creating Hive Tables, loading data and writing Hive queries in Apache Hadoop environment by Cloudera.
- Created Data model for Hive tables.
- Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
- Worked on Oozie workflow engine for job scheduling.
- Migrated the data from development cluster to QA cluster and from there to production cluster.
- Did extensive research on various NoSQL solutions and Algorithms.
- Created the developer Unit test plans and executed testing in the development cluster.
Confidential, Dallas, TX
Java Developer
Environment: Windows XP, Unix, Java5.0, Design Patterns, Web sphere, Apache Ant, J2EE (Servlets, JSP), HTML, JSON, JavaScript, Angular.js, CSS, Struts, Spring, Hibernate3.0, Eclipse, Oracle 10g, SQL Developer, WinSCP, Log4J and JUnit.
Responsibilities:
- Worked on multithreaded middleware using socket programming to introduce whole set of new business rules implementing OOPs design and principles.
- Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support.
- Applied OOAD principles for the analysis and design of the system.
- Involved in creating Json based UI backed by Web Service on RESTfull architecture.
- Used Web sphere Application Server to deploy the build.
- Developed front-end screens using JSP, HTML, JQuery, JavaScript and CSS.
- Used Spring Framework for developing business objects.
- Performed data validation in Struts Form beans and Action Classes.
- Designed dynamic and multi-browser compatible pages using HTML, CSS, JQuery, Angular.js and JavaScript.
- Good experience with Angular JS directives ng-app, ng-init, ng-model for initialization of Angular JS application data.
- Used Eclipse for the Development, Testing and Debugging of the application.
- Used DOM Parser to parse the xml files.
- Log4j framework has been used for logging debug, info & error data.
- SQL Developer was used as a database client.
- Used WinSCP to transfer file from local system to other system.
- Performed Test Driven Development (TDD) using JUnit.
- Used Ant script for build automation.
- Used Rational ClearQuest for defect logging and issue tracking.
Confidential
Java Developer
Environment: Unix (Shell Scripts), Eclipse, Java (jdk1.6), J2EE, JSP1.0, Servlets, Struts JavaScript, JDBC, Oracle 10g, UML, Rational Rose 2000, WebLogic Server, Apache Ivy, JUnit, SQL, PL/SQL, CSS, HTML, XML
Responsibilities:
- Involved in Analysis, design and coding on J2EE Environment.
- Implemented MVC architecture using Struts, JSP, and EJB's.
- Worked on Hibernate object/relational mapping according to database schema.
- Presentation layer design and programming on HTML, XML, XSL, JSP, JSTL and Ajax.
- Designed, developed and implemented the business logic required for Security presentation controller.
- Used JSP, Servlet coding under J2EE Environment.
- Designed XML files to implement most of the wiring need for Hibernate annotations and Struts configurations.
- Responsible for developing the forms, which contains the details of the employees, and generating the reports and bills.
- Involved in designing of class and dataflow diagrams using UML Rational Rose.
- Used CVS for maintaining the Source Code Designed, developed and deployed on Apache Tomcat Server.
- Created and modified Stored Procedures, Functions, Triggers and Complex SQL Commands using PL/SQL.
- Involved in the Design of ERD (Entity Relationship Diagrams) for Relational database.
- Developed Shell scripts in UNIX and procedures using SQL and PL/SQL to process the data from the input file and load into the database.
- Used Core Java concepts in application such as multithreaded programming, synchronization of threads used thread wait, notify, join methods etc.
- Creating cross-browser compatible and standards-compliant CSS-based page layouts.
- Involved in maintaining the records of the patients visited along with the prescriptions they were issued in the Database.
- Performed Unit Testing on the applications that are developed.