Sr Hadoop Developer Resume
Chicago, IL
SUMMARY
- 10.6 years of Software Development and System Engineering experience, wif a demonstrated ability to quickly learn and integrate new technologies in Retail, Telecom and supply chain domain using Java/J2EE technologies.
- 3+ Years of experience in Big data using Hadoop, Hive, Pig, Sqoop, Hbase, Impala, Airflow, SQL and MapReduce Programing.
- Strong noledge in using Mapreduce programming model for analyzing teh data stored in Hadoop.
- Hands - on experience on major components in Hadoop Ecosystem including Mapreduce, Hive, Sqoop and noledge of Mapper/Reduce/HDFS Frame work.
- Extending HIVE core functionality by using custom UDF's, UDAF and UDTF.
- Extensive experience in Data migration from existing data stores to Hadoop.
- Experience in JAVA/J2ee, Struts, JSF, Spring, Restful Web services, Hibernate and web services.
- Experience in working wif web/applications servers JBOSS, WebLogic, Apache Tomcat and glass fish.
- Experience in developing front-end using jQuery, JavaScript, JSF, HTML, XHTML and CSS.
- Experience in analyzing data using HIVEQL and custom MapReduce programs in JAVA.
- Good experience in NoSQL (HBase).
- Experience in SQL programming including complex SQL queries and dynamic queries.
- Experience in UNIX shells scripting.
- Supported ETL processes for large-scale data warehousing.
- Experience in designing highly transactional web-sites using J2EE technologies and handling design/implementation using NetBeans/Eclipse/My Eclipse.
- Good working noledge of XML (DOM, SAX), Multithreading, Collections and Exceptions.
- Supported QA team during teh QA phase.
- Experience of end-to-end implementation of teh SDLC; including finalization of specifications, design, development, implementation and quality.
- Experience in Agile, SDLC and waterfall methodologies.
- Experience in writing custom Exceptions.
- Good Experience in Data structures.
- Experience in carrying out teh JUnit Tests.
- Significant experience working wif customers, Project Managers and technical teams for securing and executing concurrent multi-technology projects.
TECHNICAL SKILLS
Big Data: Hadoop, HDFS, Map reduce, Hive, Pig, Sqoop, Impala, Airflow, oozie and Hbase
Languages: Java 1.4,1.5,1.6,1.7, Python, Scala, SQL, Shell Scripting
J2EE Technologies: JSP, Servlets, JDBC
Servers: Jboss, Web Logic and Tomcat
Frameworks: Struts, Spring, TSF, Hibernate, Spring boot and Hadoop.
Java IDEs: NetBeans, RAD, Eclipse.
Version Control / Tracking Tools: Subversion, CVS, Clear Case and GitHub
Databases: Oracle, SQL (DDL, DML, DCL) and PL/SQL.
Design Skills: J2EE design patterns, Object Oriented Analysis and Design (OOAD)
Operating Systems: Windows7, Windows XP, 2000, 2003, Unix and Linux
PROFESSIONAL EXPERIENCE
Confidential, Chicago, IL
Sr Hadoop Developer
Responsibilities:
- Developed dynamic complex impala queries in Java for Data Analysis on different data formats.
- Developed Shell scripts and schedule jobs on airflow to import data files from various sources to HDFS to archive.
- Develops and tests for extraction, transformation, and load processes.
- Develop airflow workflows by integrating all tasks relating to a project and schedule teh jobs as per requirements.
- Implemented Daily airflow jobs dat automate parallel tasks of loading teh data into HDFS and pre-processing data using Implala.
- Importing and exporting data into HDFS and Hive using Sqoop. Perform data analysis using Hive.
- Monitoring scheduled jobs on airflow web console.
- Planning and conducting code reviews for changes and enhancements dat ensure standards compliance.
- Experience maintaining code versions on Git and migrate code from dev to test and prod.
- Experience gathering requirements and data producers from non-technical end users.
- Translate business user requirements into technical data documentation such as data models, process flows and other required documentation.
- Develop data mapping and data loading processes.
- Preparing and packaging production-ready code modules for staging.
- Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement.
- Worked on teh proof-of-concept for Apache Spark framework initiation.
- Involved in configuring teh load balancer.
- Involved in performance tuning.
- Developed complex user defined functions in Hive and impala.
- Developed Restful API to register teh data in HDFS location.
- Involved in configuring teh SQS queue.
- Prepared mapping documents for ETL process.
- Developed build scripts to deploy in different environments.
- Developed restful API client to do load test on bulk calculations.
Environment: HDFS, MapReduce, Hive, Impala, Sqoop, Airflow, Zookeeper, Java, Python, Spring Boot, RESTful web services, Cloudera Manager Ubuntu, Jira, UNIX Shell Scripting.
Confidential, Louisville, KY
Sr Hadoop Developer
Responsibilities:
- Developed complex Map Reduce programs in Java for Data Analysis on different data formats.
- Developed Map Reduce programs dat filter bad and un-necessary records and find out unique records based on different criteria.
- Developed Secondary sorting implementation to get sorted values at reduce side to improve map reduce performance.
- Implemented custom Data Types, Input Format, Record Reader, Output Format, Record Writer for Map Reduce computationsto handle custom business requirements.
- Developed Shell scripts and schedule jobs on Cron to import data files from various sources to HDFS to archive.
- Develops and tests for extraction, transformation, and load processes.
- Develop Oozie workflows by integrating all tasks relating to a project and schedule teh jobs as per requirements.
- Implemented Daily Oozi jobs dat automate parallel tasks of loading teh data into HDFS and pre-processing wif Pig using Oozie co-coordinator jobs.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Perform data analysis using Hive and Pig.
- Monitoring scheduled jobs on Oozie web console.
- Planning and conducting code reviews for changes and enhancements dat ensure standards compliance.
- Experience maintaining code versions on SVN and migrate code from dev to test and prod.
- Experience gathering requirements and data producers from non-technical end users.
- Translate business user requirements into technical data documentation such as data models, process flows and other required documentation.
- Develop data mapping and data loading processes.
- Preparing and packaging production-ready code modules for staging.
- Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement.
- Worked on teh proof-of-concept for Apache Spark framework initiation.
Environment: HDFS, MapReduce, Hive, Pig, Sqoop, oozie, Zookeeper, Java, Cloudera Manager Ubuntu, UNIX Shell Scripting.
Confidential, Minneapolis, MN
Hadoop Consultant
Responsibilities:
- Involved in gathering teh business requirements and in turn converted teh same into IT standard requirements.
- Analyzed teh requirement to setup a 25 node cluster.
- Ensured NFS is configured for Name Node.
- Created two different users (hduser for performing hdfs operations and mapred user for performing map reduce operations only).
- Worked on setting password less Hadoop.
- Worked on setting up CRON job to delete Hadoop logs/local old job files/cluster temp files.
- Moved all crawl data flat files generated from various retailers to HDFS for further processing.
- Handled 7TB of data as a part of development.
- Written CLI commands using HDFS.
- Written teh script files for processing data and loading to HDFS
- Written teh Apache PIG scripts to process teh HDFS data.
- Configured Hive wif MySQL as a Remote Metastore
- Created Hive tables to store teh processed results in a tabular format.
- Created External Hive Table on top of parsed data.
- Developed teh SQOOP scripts in order to make teh interaction between Pig and MySQL Database.
- Moved all log/text files generated by various products into HDFS location
- Written Map Reduce code dat will take input as log files and parse teh logs and structure them in tabular format to facilitate effective querying on teh log data
- Preparation of status reports.
Environment: HDFS, MapReduce, Hive, Pig, Sqoop, oozie, Java, Cloudera Manager Ubuntu, UNIX Shell Scripting, HBase, Python.
Confidential, Denver, CO
Java/Hadoop Developer
Responsibilities:
- Archiving seven years claims data, which requires complex processing to get into a normalized format.
- Logging terabytes of data generated from transactional systems daily, and storing them in CDH for analytical purposes.
- Importing and exporting data into HDFS from Relational Database systems using Sqoop.
- Worked wif Hive to bulk load Data into HBase.
- Worked wif creating tables and Analyzing data using Hive queries.
- Scheduled Recurrent Jobs, Packaged Multiple Jobs, Sequenced jobs in Oozie.
- Experienced in running Hadoop streaming jobs to process terabytes of xml format.
- Load and transform large sets of structured, semi structured and unstructured data.
- Experienced in writing MapReduce java programs to process huge data.
- Experience in Data Integration using Microsoft BI tools.
- Hands on Experience wif installing, configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Hive, Sqoop, Pig.
- Experience in writing java programs to interact wif Hbase and Hive
Environment: Hadoop, MapReduce, HDFS, Hive, Java, jdbc, Pig, Cloudera, Ubuntu, UNIX Shell Scripting,HBase.
Confidential
Technical Lead
Responsibilities:
- Analyze existing application and finding out key areas dat required changes to in corporate new requirements.
- Design new workflow based on teh new requirements.
- Finalized requirements wif regular meetings and discussions wif users.
- Optimized many existing functionalities for better maintainability and quality of code.
- Provided mentoring and Technical training to new team members.
- Provided unit test cases for teh application using Junit3.0
- Getting clarifications from customer,
- Supporting our team members technically, functionally to finish their work in time,
- Supporting deployments in SIT, UAT and PROD environments.
- Used JSP, JavaScript, JSP Tag Libraries and AJAX YUI data table for web presentation layer.
- Developed application on MVC architecture using Struts 1.2 Framework, Action Classes and performed client side validations using JavaScript.
- Implemented Tiles framework to develop JSP templates wif headers and footer elements, and varying body contents.
- Build project using Apache Maven build scripts, and teh applications Version controlling was accomplished using SVN.
Environment: Java, Servelts, JSP, Spring, Hibernate, UNIX Shell Scripting, Tomcat, CVS, Toad, Eclipse, SQL, Oracle, Maven.
Confidential
Module Lead
Responsibilities:
- Involved in teh development of Functional Requirements Specifications, Technical Specifications, detailed design documents, user guides, test procedures, and test cases for teh application components
- Followed Object Oriented Design and Analysis by preparing Use Cases, business domain model, Sequence Diagrams and Class Diagrams and designed teh UML components for technical specification in Microsoft Visio.
- Implemented teh Software Development Guidelines based on Agile Methodologies.
- Developed front end of application on MVC architecture employing Struts Framework.
- Responsible for setting up configuration files- web.xml, struts- config.xml, tiles-defs.xml, and validation.xml; developed UI layer using JSP, Struts Tag Libraries, JavaScript, AJAX, HTML/DHTML, CSS
- Developed Action classes, Action Forms performed form validations using Struts Validation Frameworks, and used Tiles Frame work.
- Used Hibernate in DAO layer to access and update information in Oracle database, developed Hibernate configuration files (hbm.xml) for object relational mapping wif database, fine-tuned performance by optimizing query and data caching mechanisms.
- Developed HQL queries to retrieve data using Hibernate
- Developed SQL queries and Procedures using SQL and PL/SQL.
- Involved in Test data creation, Unit testing using JUNIT
- Used Subversion for code check-ins
- Used CSS for style sheets.
Environment: Java, J2EE, SQL Server, J2EE Web Services, Apache Tomcat Server, XML, Eclipse, Rational Rose, Ajax, HTML, JavaScript and WebLogic Application Server, subversion.
Confidential
Senior Java Developer
Responsibilities:
- Understanding business requirements and creating Detail Design Document.
- Extensively used Scrum Programming concept which is part of agile methodologies to plan work for every iteration and used continuous integration tool to make teh build passes before deploying teh code to other environments.
- Developed teh application using Struts Framework dat leverages teh classical Model-View-Controller (MVC) architecture.
- Used Hibernate Mapping Files to connect to database and spring framework as IOC container.
- Involved in creating all Java Server Pages, JavaBeans and invoking EJB Components through JNDI.
- Implemented Design Patterns like Business Delegates, Session Facades, Data Access Objects, and Value Objects etc.
- Created many stored procedures, PL/SQL transactions to interact wif oracle database and they in turn are called using EJB.
- Used Hibernate Mapping Files to connect to database
- Monitored teh error logs using Log4J and fixed teh problems.
- Involved in designing teh user interfaces using JSPs.
- Used JDBC to invoke Stored Procedures and database connectivity to Oracle.
- Developed JUnit test cases for all teh developed modules.
- Interact wif testing team for IST and UAT.
- Used CVS for version control across common source code used by developers.
Environment: Java, JSP, Struts, Spring, EJB, JDBC, HTML, XML, Hibernate, JNDI, Log4j, CVS, ANT, Web logic, Eclipse, Oracle.
Confidential
Java Developer
Responsibilities:
- Designed and developed an interactive module in Java
- Implemented Presentation layer using JSP, Servlets
- Developed teh application using Struts Framework dat leverages teh classical MVC architecture
- Worked on query handling, customer support, halpdesk
- Migrated a poorly performing and outdated application to Struts, Hibernate based system for Sprint.
- Maintained teh interface of Oracle using JDBC
- Wrote procedures and queries to extract data from database
- Tested teh flow of modules using JUNIT
- Monitored teh error logs using Log4J and fixed teh problems.
- Handled teh JDBC backend operations of teh respective modules also.
Environment: java, Struts, Hibernate, JSP, HTML, CSS, JDBC, JavaScript, xml, CVS, Log4J, JNDI, JUnit, Tomcat, My Eclipse.