We provide IT Staff Augmentation Services!

Hadoop Analyst/developer Resume

0/5 (Submit Your Rating)

IL

SUMMARY

  • Professional with 10 years of IT experience in requirement gathering, design, development, testing, and maintenance and progressive experience in all phases of SDLC
  • Proficient in working with various Hadoop ecosystems - Sqoop, Oozie, HBase, Pig, Hive, Flume
  • In depth working experience with Hadoop daemons - JobTracker, TaskTracker, MapReduce, Datanode, Namenode and Secondary Name node
  • Experience implementing big data projects using Cloudera and HDInsight
  • Involved in writing multiple MapReduce programs using agile methodology
  • Experience in Agile methodologies - SCRUM, XP and Lean
  • Experience in developing applications using JSP, Servlets, EJB, JMS, HIBERNATE, JavaScript, JDBC, SPRING Framework, Struts MVC
  • End-to end experience working with Oracle, SQL Server and MySQL
  • Extensive experience writing simple to complex SQL queries, PL/SQL, Stores procedures and Triggers
  • Involved in multiple full life cycle (SDLC) implementation from project planning to Go-live
  • Strong program experience in core Java, Java Script, Strut, J2EE
  • Experience setting up cluster of MPP through Amazon EC2
  • Experience using Application servers like WebSphere, TomCat and Oracle 10g
  • Experience in different version control systems like RCC (Rational Clear Case), CVS and worked on IDE’s for development project (Eclipse, Net Beans and Notepad plus)
  • Experience writing Linux Shell scripts
  • Ability to interact/work with business users, technical team and management professionals and can act as an effective resource with excellent communication skills
  • Experience working in various industries - finance, retail, e-commerce, web channel
  • Adept at supporting post-production issues by coordinating both onshore and offshore resources
  • Excellent document management skills - BRD, TDD (reports, conversion, interfaces, workflows), Unit Test Plan documents with ISO 9001 quality standards
  • Well versed in working with offshore and onshore model
  • Experience in providing post-go live support to customers
  • Extensive experience in preparing and providing end user documentation and training
  • Excellent team player with great communication and written skills
  • Knowledge of SCALA, SPARK, MAHOUT

TECHNICAL SKILLS

Big Data Technologies: Hadoop, HDFS, Hive, MapReduce, Pig, Sqoop, Flume, Zookeeper, ImpalaHBase, MongoDB, NoSQL, HDInsight, Cassandra, MRUnit, YARN

BI Technologies: SSIS, SSAS, Pentaho, SSRS 2008, Agile BITableau, SAP BW/BI, Visual Studio, MicroStatergy

RDBMS: DB2, SQL Server, Oracle 9i/11g, My SQL

Operating Systems: Linux, Windows 98/00/NT/XP, Sun Solaris, Unix

Languages: VB, JAVA, Python, Unix Shell Scripting, SQL, Java Script, PL/SQL, JDK 5/1.4, RMI, JDBC, JNDI, Java Mail, Swing

Web Technologies, IDE’s& Others: Ext JS, Ajax, Facelets, JavaScript, Dojo, CSS, jQuery, HTML, XML, WSDL, REST, SOAP

PROFESSIONAL EXPERIENCE

Confidential, IL

Hadoop Analyst/Developer

Responsibilities:

  • Assisted project managers in coming up with project plan in MS project
  • Involved in design and development phases of SDLC using SCRUM methodology
  • Assisted end users in coming up with user stories during product backlog meeting
  • Involved in daily scrum meeting, sprint review meeting, and agile release planning
  • Worked with SCRUM team in delivering agreed user stories on time for every sprint
  • Involved in setting up Sqoop framework to extract structured data (product price, stock levels, stocking costs and sales) from Oracle 11gi
  • Set up Flume system to periodically receive competitors product data in JSON format from webcrawler to HBase
  • Configured Flume to extract daily web log data in JSON format from web server to HBase
  • Worked with system administrator to come up with workflows in Oozie
  • Worked with business users, oracle database team and system administrators in designing data conversion to HBase
  • Developed custom InputFormat and Mapper program to read and parse files and return tab limited rows
  • Generated a Data Dictionary for category and sub categories of URL with MapReduce program
  • Simple to complex queries were written and executed in HIVE
  • Responsible for orchestrate hundreds of HIVE queries using Oozie workflows
  • Involved in importing data into Cloudera - Impala and ran several reports on it
  • Participated in HIVE optimization project to effectively handle all tables - archiving tables, deleting tables, converting tables to RC and ORC format, partitioning, and bucketing
  • Created UDF (User Defined Functions) to help create Pig Latin scripts
  • Worked with team in optimizing Pig Scripts following standard pig optimization techniques like split, schema and stream
  • Followed all performance enhancing techniques for executing Pig Latin scripts
  • Written various custom shell scripts to manage day to day operations
  • Interacted with Tableau team in designing and testing analytical reports
  • Worked with technical team (onshore and offshore) in preparing technical specs and unit testing development
  • Created and tested several java classes in JUnit4
  • Involved in pilot scale study of Spark with UC Berkley team

Environment: RedHat Linux, HDFS, CDH4, IMPALA, MapReduce2.0, Hive, Java JDK1.6, Pig, Sqoop, Flume, Zookeeper, Oozie, Oracle 11gi, HBase, Zookeeper, JUnit4

Confidential, CA

Hadoop Developer

Responsibilities:

  • Involved in analyzing business, system and data mapping requirements
  • Worked with Data scientists and architects to come up with SLA and HLD
  • Worked with team in coming up with LLD following HLD
  • Assisted administrator in setting up Hadoop cluster and created maintenance shell scripts
  • Worked on setting up 10 workstations to run Hadoop and provided POC to internal team
  • Sqoop was used to extract databases from MySQL database into HDFS, HIVE and HBase
  • Performed data validation on the data ingested using MapReduce by building a custom model to filter all the invalid data and cleansing the data
  • Assisted business users in ad-hoc queries and data metrics using Pig and Hive
  • Planned and designed HIVE for creating and maintaining tables - join tables, alter tables
  • Optimized Hive tables by partitioning (manual and dynamic), bucketing and compressing data
  • Created several Hive units tests using JUnit and executed these tests in Hive
  • Created Pig Latin scripts to clean up both structured and unstructured (social and chat) data for further analysis in HIVE
  • Developed custom load function for loading web server logs in Pig Latin
  • Extended existing UDF to create new UDF for further data analysis
  • Coordinates with test team lead in planning, creating and executing test scripts

Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase

Confidential, NC

Hadoop Developer

Responsibilities:

  • Prepared and conducted various meetings to finalize low level design documents
  • Assisted administrators in setting up Hadoop and its ecosystems
  • Worked with Oracle team in importing tables to HDFS using Sqoop
  • Flume was implemented to index email messages and load email data into HDFS
  • Created Pig Scripts to load unstructured data into structured data for further analysis
  • Planned and joined several relations using Pig JOIN commands and also used specialized joins for optimization
  • Developed custom filter function as UDF in Pig Script using java and registered respective Jar files with Pig Script
  • Tables were created using Hive scripts and tables were optimized by creating partitioning and bucketing of tables
  • DAG was designed to plan different workflows to execute several MapReduce, shell scripts, Hive and Pig Scripts in Oozie
  • Monitored daily all workflows scheduled in Oozie
  • Wrote shell script to verify the files availability in HDFS from various downstream processes
  • Developed MapReduce program to come up with inverted index for semi-structured and unstructured data files
  • Combiner program was developed to sort the data in Mapper phase and worked with admin in planning number of reducers for MapReduce programs
  • Worked with MicroStatergy vendor in data export from Hadoop system to MicroStatergy
  • Utilized MRUnit, HiveUnit and PigUnit to test MapReduce, Hive and Pig Scripts respectively
  • Coordinated with testing team in planning and executing test scripts
  • Involved in supporting (Level3) all Hadoop and its ecosystems related issues

Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase, JUnit3

Confidential, NJ

Sr. Java Developer

Responsibilities:

  • Followed Waterfall methodology throughout the life cycle of the project
  • Prepared program and system specifications for web project development
  • Build the prototype for the application and user end screens
  • Adopted SPRING MVC framework in web application development
  • Front end was developed using JavaScript, XHTML, JSP, CSS
  • Developed JavaScript to perform client side validations and client forms
  • Created web services and SOAP/WSDL was used as part web service implementation
  • Spring Framework was used to design and implement domain layer based on Aspect Oriented Programming (AOP) and inversion control(IOC)
  • JMS was used to handle transfer of asynchronous messages using publish/subscribe model
  • Involved in performing client side validation using JQuery and server side validation using struts 2 validation framework
  • Interacted with Oracle database through Hibernate
  • Retrieved data from Oracle database through SQL commands and stored procedures
  • Planned and created test plans and integrated testing in JUnit
  • Implemented MyBatis ORM tools to automate the mapping between SQL databases and java objects
  • Worked with testing team in fixing bugs and failed test scripts in QA and UAT

Environment: Java, JSF, Tomcat Web Server, J2EE, Hibernate, Oracle, JSON, HTML, SQL, Spring, Structs, JUnit, Eclipse, UNIX, Windows, PL/SQL, SOAP, WSDL

Confidential, MI

Sr. Java Developer

Responsibilities:

  • Reviewed functional specifications with analysts
  • Created LLD documents (sequence, class diagrams etc) using MS enterprise architecture tool
  • Used spring framework for application building based on MVC design paradigm
  • Developed several Web pages using JSP, JavaScript, HTML
  • Involved in design and development of log and transaction using SpringAOP
  • Developed Java classes and beans to interact with Oracle database using Hibernate
  • Developed DIS web services using jaxws and worked on consuming web services - JaxRs
  • Worked on creating DAO classes to get connection instances with Spring JDBC and configured JDBC connections with Oracle 10g
  • Designed and used Jmeter schemas for testing front tier of the application
  • Worked on Jersey Framework using REST API
  • Performed database development using PL/SQL, implemented search mechanism using stored procedures and sequences
  • Worked with project managers in planning change control management for the project
  • Identified and prepared Unit Test script for all the business scenarios
  • Assisted technical team in upgrade of CRM from 2007 to 7.0
  • Worked on completing cutover activities

Environment: Java 1.5, J2EE, Servlets, JSP, JUnit, Eclipse, WinCVS,Websphere5.1, Oracle 10gi, Windows XP, Eclipse 3.0, HTML, EJB 2.0

Confidential, TX

Java Developer

Responsibilities:

  • Documented functional/technical specifications based on business users requirements
  • Created UML diagrams such as Use Cases, Sequence Diagrams, Class Diagrams and flowcharts using Visio
  • Designed and developed interactive GUI screens using XML, JSPs, HTML, JavaScript, CSS
  • Worked on the JAVA collections API for handling the data objects between front end and business layers
  • Developed web service provider application using WSAD, XSLT and WSDL and implemented SoapUI for testing
  • Developed webpage layouts using Strut-tiles, JSF core and HTML tags
  • Persistence framework mapping the ORM objects to tables through Hibernate
  • Developed using Eclipse as IDE with Tomcat as the server
  • Developed various stored procedures and triggers with PL/SQL
  • Designed and developed Web Services to provide service to clients using WSDL and SOAP
  • Created extensive JUnit test cases in order to verify functionality of the methods and used log4j for logging messages
  • Worked with users and testing team in fixing defects during integration testing
  • Deployed WSS applications using Maven on to development and testing environments
  • Worked on integrating search system, metrics and email applications to websites using JSF framework
  • Created ant scripts and developed builds using Apache ANT

Environment: J2EE, Core Java, Struts 1.0, Tiles, JSP, JNDI, ANT, JavaScript, XML, Eclipse

Confidential, OH

Java Developer

Responsibilities:

  • Involved in the analysis, design and development of SDLC
  • Involved in writing multithreaded programs to measure system performance
  • Designed and documented functional/technical specifications based on business users requirements
  • Designed and developed UI view using Strut tags, JSP and HTML
  • Worked on JDBC APIs for database connectivity and data retrieval
  • Worked on bug tracking in JIRA and debugging using Eclipse 3.6
  • Analyzed and configured JCL to execute mainframe batch
  • Wrote stored procedures, view, user-defined functions, and triggers in SQL server database for report
  • Involved in automating nightly batch processing and job alerts using REXX
  • Worked on XML parser APIs like DOM for marshalling and unmarshalling of data
  • User Clear Case for version control and JUnit for unit testing of applications
  • Implemented stored procedures and PL/SQL for Oracle 9i database

Environment: Core Java, JUnit, JSP, Eclipse, JIRA, JDBC, Struts 2, Oracle 9i, MS SQL Server, REXX, Mainframe, JCL, TomCat, Clear Case

Confidential, AZ

Engineer II

Responsibilities:

  • Provided estimates of effort required to design and develop solutions and participated in design sessions with architects and developers
  • Prepared the functional/technical specification document following requirements from business users
  • Involved in creating HLD (High Level Design) and detailed design documents using Visio
  • Involved in development of web based component for RFP (Request For Proposal) processing
  • Developed web tier using Struts MVC framework
  • Created DAO (Data Access Object) classes to fetch and connect data from database
  • Involved in the development of both UI (JSP, Javascript, CSS) and backend flow (java, design patterns, Oracle)
  • Prepared test plans and test plan execution (unit, functional and integration testing, end to end testing)
  • Coordinated defect meeting and documented the same

Environment: Java, Oracle 9i, JDBC, javascript, Struts1.x, Hibernate, Log4j, Eclipse, CVS, HTML, JSP, Weblogic 8.1, CSS

We'd love your feedback!