Hadoop Analyst/developer Resume
IL
SUMMARY
- Professional with 10 years of IT experience in requirement gathering, design, development, testing, and maintenance and progressive experience in all phases of SDLC
- Proficient in working with various Hadoop ecosystems - Sqoop, Oozie, HBase, Pig, Hive, Flume
- In depth working experience with Hadoop daemons - JobTracker, TaskTracker, MapReduce, Datanode, Namenode and Secondary Name node
- Experience implementing big data projects using Cloudera and HDInsight
- Involved in writing multiple MapReduce programs using agile methodology
- Experience in Agile methodologies - SCRUM, XP and Lean
- Experience in developing applications using JSP, Servlets, EJB, JMS, HIBERNATE, JavaScript, JDBC, SPRING Framework, Struts MVC
- End-to end experience working with Oracle, SQL Server and MySQL
- Extensive experience writing simple to complex SQL queries, PL/SQL, Stores procedures and Triggers
- Involved in multiple full life cycle (SDLC) implementation from project planning to Go-live
- Strong program experience in core Java, Java Script, Strut, J2EE
- Experience setting up cluster of MPP through Amazon EC2
- Experience using Application servers like WebSphere, TomCat and Oracle 10g
- Experience in different version control systems like RCC (Rational Clear Case), CVS and worked on IDE’s for development project (Eclipse, Net Beans and Notepad plus)
- Experience writing Linux Shell scripts
- Ability to interact/work with business users, technical team and management professionals and can act as an effective resource with excellent communication skills
- Experience working in various industries - finance, retail, e-commerce, web channel
- Adept at supporting post-production issues by coordinating both onshore and offshore resources
- Excellent document management skills - BRD, TDD (reports, conversion, interfaces, workflows), Unit Test Plan documents with ISO 9001 quality standards
- Well versed in working with offshore and onshore model
- Experience in providing post-go live support to customers
- Extensive experience in preparing and providing end user documentation and training
- Excellent team player with great communication and written skills
- Knowledge of SCALA, SPARK, MAHOUT
TECHNICAL SKILLS
Big Data Technologies: Hadoop, HDFS, Hive, MapReduce, Pig, Sqoop, Flume, Zookeeper, ImpalaHBase, MongoDB, NoSQL, HDInsight, Cassandra, MRUnit, YARN
BI Technologies: SSIS, SSAS, Pentaho, SSRS 2008, Agile BITableau, SAP BW/BI, Visual Studio, MicroStatergy
RDBMS: DB2, SQL Server, Oracle 9i/11g, My SQL
Operating Systems: Linux, Windows 98/00/NT/XP, Sun Solaris, Unix
Languages: VB, JAVA, Python, Unix Shell Scripting, SQL, Java Script, PL/SQL, JDK 5/1.4, RMI, JDBC, JNDI, Java Mail, Swing
Web Technologies, IDE’s& Others: Ext JS, Ajax, Facelets, JavaScript, Dojo, CSS, jQuery, HTML, XML, WSDL, REST, SOAP
PROFESSIONAL EXPERIENCE
Confidential, IL
Hadoop Analyst/Developer
Responsibilities:
- Assisted project managers in coming up with project plan in MS project
- Involved in design and development phases of SDLC using SCRUM methodology
- Assisted end users in coming up with user stories during product backlog meeting
- Involved in daily scrum meeting, sprint review meeting, and agile release planning
- Worked with SCRUM team in delivering agreed user stories on time for every sprint
- Involved in setting up Sqoop framework to extract structured data (product price, stock levels, stocking costs and sales) from Oracle 11gi
- Set up Flume system to periodically receive competitors product data in JSON format from webcrawler to HBase
- Configured Flume to extract daily web log data in JSON format from web server to HBase
- Worked with system administrator to come up with workflows in Oozie
- Worked with business users, oracle database team and system administrators in designing data conversion to HBase
- Developed custom InputFormat and Mapper program to read and parse files and return tab limited rows
- Generated a Data Dictionary for category and sub categories of URL with MapReduce program
- Simple to complex queries were written and executed in HIVE
- Responsible for orchestrate hundreds of HIVE queries using Oozie workflows
- Involved in importing data into Cloudera - Impala and ran several reports on it
- Participated in HIVE optimization project to effectively handle all tables - archiving tables, deleting tables, converting tables to RC and ORC format, partitioning, and bucketing
- Created UDF (User Defined Functions) to help create Pig Latin scripts
- Worked with team in optimizing Pig Scripts following standard pig optimization techniques like split, schema and stream
- Followed all performance enhancing techniques for executing Pig Latin scripts
- Written various custom shell scripts to manage day to day operations
- Interacted with Tableau team in designing and testing analytical reports
- Worked with technical team (onshore and offshore) in preparing technical specs and unit testing development
- Created and tested several java classes in JUnit4
- Involved in pilot scale study of Spark with UC Berkley team
Environment: RedHat Linux, HDFS, CDH4, IMPALA, MapReduce2.0, Hive, Java JDK1.6, Pig, Sqoop, Flume, Zookeeper, Oozie, Oracle 11gi, HBase, Zookeeper, JUnit4
Confidential, CA
Hadoop Developer
Responsibilities:
- Involved in analyzing business, system and data mapping requirements
- Worked with Data scientists and architects to come up with SLA and HLD
- Worked with team in coming up with LLD following HLD
- Assisted administrator in setting up Hadoop cluster and created maintenance shell scripts
- Worked on setting up 10 workstations to run Hadoop and provided POC to internal team
- Sqoop was used to extract databases from MySQL database into HDFS, HIVE and HBase
- Performed data validation on the data ingested using MapReduce by building a custom model to filter all the invalid data and cleansing the data
- Assisted business users in ad-hoc queries and data metrics using Pig and Hive
- Planned and designed HIVE for creating and maintaining tables - join tables, alter tables
- Optimized Hive tables by partitioning (manual and dynamic), bucketing and compressing data
- Created several Hive units tests using JUnit and executed these tests in Hive
- Created Pig Latin scripts to clean up both structured and unstructured (social and chat) data for further analysis in HIVE
- Developed custom load function for loading web server logs in Pig Latin
- Extended existing UDF to create new UDF for further data analysis
- Coordinates with test team lead in planning, creating and executing test scripts
Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase
Confidential, NC
Hadoop Developer
Responsibilities:
- Prepared and conducted various meetings to finalize low level design documents
- Assisted administrators in setting up Hadoop and its ecosystems
- Worked with Oracle team in importing tables to HDFS using Sqoop
- Flume was implemented to index email messages and load email data into HDFS
- Created Pig Scripts to load unstructured data into structured data for further analysis
- Planned and joined several relations using Pig JOIN commands and also used specialized joins for optimization
- Developed custom filter function as UDF in Pig Script using java and registered respective Jar files with Pig Script
- Tables were created using Hive scripts and tables were optimized by creating partitioning and bucketing of tables
- DAG was designed to plan different workflows to execute several MapReduce, shell scripts, Hive and Pig Scripts in Oozie
- Monitored daily all workflows scheduled in Oozie
- Wrote shell script to verify the files availability in HDFS from various downstream processes
- Developed MapReduce program to come up with inverted index for semi-structured and unstructured data files
- Combiner program was developed to sort the data in Mapper phase and worked with admin in planning number of reducers for MapReduce programs
- Worked with MicroStatergy vendor in data export from Hadoop system to MicroStatergy
- Utilized MRUnit, HiveUnit and PigUnit to test MapReduce, Hive and Pig Scripts respectively
- Coordinated with testing team in planning and executing test scripts
- Involved in supporting (Level3) all Hadoop and its ecosystems related issues
Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase, JUnit3
Confidential, NJ
Sr. Java Developer
Responsibilities:
- Followed Waterfall methodology throughout the life cycle of the project
- Prepared program and system specifications for web project development
- Build the prototype for the application and user end screens
- Adopted SPRING MVC framework in web application development
- Front end was developed using JavaScript, XHTML, JSP, CSS
- Developed JavaScript to perform client side validations and client forms
- Created web services and SOAP/WSDL was used as part web service implementation
- Spring Framework was used to design and implement domain layer based on Aspect Oriented Programming (AOP) and inversion control(IOC)
- JMS was used to handle transfer of asynchronous messages using publish/subscribe model
- Involved in performing client side validation using JQuery and server side validation using struts 2 validation framework
- Interacted with Oracle database through Hibernate
- Retrieved data from Oracle database through SQL commands and stored procedures
- Planned and created test plans and integrated testing in JUnit
- Implemented MyBatis ORM tools to automate the mapping between SQL databases and java objects
- Worked with testing team in fixing bugs and failed test scripts in QA and UAT
Environment: Java, JSF, Tomcat Web Server, J2EE, Hibernate, Oracle, JSON, HTML, SQL, Spring, Structs, JUnit, Eclipse, UNIX, Windows, PL/SQL, SOAP, WSDL
Confidential, MI
Sr. Java Developer
Responsibilities:
- Reviewed functional specifications with analysts
- Created LLD documents (sequence, class diagrams etc) using MS enterprise architecture tool
- Used spring framework for application building based on MVC design paradigm
- Developed several Web pages using JSP, JavaScript, HTML
- Involved in design and development of log and transaction using SpringAOP
- Developed Java classes and beans to interact with Oracle database using Hibernate
- Developed DIS web services using jaxws and worked on consuming web services - JaxRs
- Worked on creating DAO classes to get connection instances with Spring JDBC and configured JDBC connections with Oracle 10g
- Designed and used Jmeter schemas for testing front tier of the application
- Worked on Jersey Framework using REST API
- Performed database development using PL/SQL, implemented search mechanism using stored procedures and sequences
- Worked with project managers in planning change control management for the project
- Identified and prepared Unit Test script for all the business scenarios
- Assisted technical team in upgrade of CRM from 2007 to 7.0
- Worked on completing cutover activities
Environment: Java 1.5, J2EE, Servlets, JSP, JUnit, Eclipse, WinCVS,Websphere5.1, Oracle 10gi, Windows XP, Eclipse 3.0, HTML, EJB 2.0
Confidential, TX
Java Developer
Responsibilities:
- Documented functional/technical specifications based on business users requirements
- Created UML diagrams such as Use Cases, Sequence Diagrams, Class Diagrams and flowcharts using Visio
- Designed and developed interactive GUI screens using XML, JSPs, HTML, JavaScript, CSS
- Worked on the JAVA collections API for handling the data objects between front end and business layers
- Developed web service provider application using WSAD, XSLT and WSDL and implemented SoapUI for testing
- Developed webpage layouts using Strut-tiles, JSF core and HTML tags
- Persistence framework mapping the ORM objects to tables through Hibernate
- Developed using Eclipse as IDE with Tomcat as the server
- Developed various stored procedures and triggers with PL/SQL
- Designed and developed Web Services to provide service to clients using WSDL and SOAP
- Created extensive JUnit test cases in order to verify functionality of the methods and used log4j for logging messages
- Worked with users and testing team in fixing defects during integration testing
- Deployed WSS applications using Maven on to development and testing environments
- Worked on integrating search system, metrics and email applications to websites using JSF framework
- Created ant scripts and developed builds using Apache ANT
Environment: J2EE, Core Java, Struts 1.0, Tiles, JSP, JNDI, ANT, JavaScript, XML, Eclipse
Confidential, OH
Java Developer
Responsibilities:
- Involved in the analysis, design and development of SDLC
- Involved in writing multithreaded programs to measure system performance
- Designed and documented functional/technical specifications based on business users requirements
- Designed and developed UI view using Strut tags, JSP and HTML
- Worked on JDBC APIs for database connectivity and data retrieval
- Worked on bug tracking in JIRA and debugging using Eclipse 3.6
- Analyzed and configured JCL to execute mainframe batch
- Wrote stored procedures, view, user-defined functions, and triggers in SQL server database for report
- Involved in automating nightly batch processing and job alerts using REXX
- Worked on XML parser APIs like DOM for marshalling and unmarshalling of data
- User Clear Case for version control and JUnit for unit testing of applications
- Implemented stored procedures and PL/SQL for Oracle 9i database
Environment: Core Java, JUnit, JSP, Eclipse, JIRA, JDBC, Struts 2, Oracle 9i, MS SQL Server, REXX, Mainframe, JCL, TomCat, Clear Case
Confidential, AZ
Engineer II
Responsibilities:
- Provided estimates of effort required to design and develop solutions and participated in design sessions with architects and developers
- Prepared the functional/technical specification document following requirements from business users
- Involved in creating HLD (High Level Design) and detailed design documents using Visio
- Involved in development of web based component for RFP (Request For Proposal) processing
- Developed web tier using Struts MVC framework
- Created DAO (Data Access Object) classes to fetch and connect data from database
- Involved in the development of both UI (JSP, Javascript, CSS) and backend flow (java, design patterns, Oracle)
- Prepared test plans and test plan execution (unit, functional and integration testing, end to end testing)
- Coordinated defect meeting and documented the same
Environment: Java, Oracle 9i, JDBC, javascript, Struts1.x, Hibernate, Log4j, Eclipse, CVS, HTML, JSP, Weblogic 8.1, CSS