We provide IT Staff Augmentation Services!

Java Developer Resume Profile

4.00/5 (Submit Your Rating)

PROFESSIONAL SUMMARY

  • Professional with 10 years of IT experience in requirement gathering, design, development, testing, and maintenance and progressive experience in all phases of SDLC
  • Proficient in working with various Hadoop ecosystems Sqoop, Oozie, HBase, Pig, Hive, Flume
  • In depth working experience with Hadoop daemons JobTracker, TaskTracker, MapReduce, Datanode, Namenode and Secondary Name node
  • Experience implementing big data projects using Cloudera and HDInsight
  • Involved in writing multiple MapReduce programs using agile methodology
  • Experience in Agile methodologies SCRUM, XP and Lean
  • Experience in developing applications using JSP, Servlets, EJB, JMS, HIBERNATE, JavaScript, JDBC, SPRING Framework, Struts MVC
  • End-to end experience working with Oracle, SQL Server and MySQL
  • Extensive experience writing simple to complex SQL queries, PL/SQL, Stores procedures and Triggers
  • Involved in multiple full life cycle SDLC implementation from project planning to Go-live
  • Strong program experience in core Java, Java Script, Strut, J2EE
  • Experience setting up cluster of MPP through Amazon EC2
  • Experience using Application servers like WebSphere, TomCat and Oracle 10g
  • Experience in different version control systems like RCC Rational Clear Case , CVS and worked on IDE's for development project Eclipse, Net Beans and Notepad plus
  • Experience writing Linux Shell scripts
  • Ability to interact/work with business users, technical team and management professionals and can act as an effective resource with excellent communication skills
  • Experience working in various industries - finance, retail, e-commerce, web channel
  • Adept at supporting post-production issues by coordinating both onshore and offshore resources
  • Excellent document management skills BRD, TDD reports, conversion, interfaces, workflows , Unit Test Plan documents with ISO 9001 quality standards
  • Well versed in working with offshore and onshore model
  • Experience in providing post-go live support to customers
  • Extensive experience in preparing and providing end user documentation and training
  • Excellent team player with great communication and written skills
  • Knowledge of SCALA, SPARK, MAHOUT

SKILL SET

Big Data Technologies

Hadoop, HDFS, Hive, MapReduce, Pig, Sqoop, Flume, Zookeeper, Impala

HBase, MongoDB, NoSQL, HDInsight, Cassandra, MRUnit, YARN

BI Technologies

SSIS, SSAS, Pentaho, SSRS 2008, Agile BI

Tableau, SAP BW/BI, Visual Studio, MicroStatergy

RDBMS

DB2, SQL Server, Oracle 9i/11g, My SQL

Operating Systems

Linux, Windows 98/00/NT/XP, Sun Solaris, Unix

Languages

VB, JAVA, Python, Unix Shell Scripting, SQL, Java Script, PL/SQL, JDK 5/1.4, RMI, JDBC, JNDI, Java Mail, Swing

Web Technologies, IDE's Others

Ext JS, Ajax, Facelets, JavaScript, Dojo, CSS, jQuery, HTML, XML, WSDL, REST, SOAP

PROJECTS

Confidential

Role: Hadoop Analyst/Developer

Responsibilities:

  • Assisted project managers in coming up with project plan in MS project
  • Involved in design and development phases of SDLC using SCRUM methodology
  • Assisted end users in coming up with user stories during product backlog meeting
  • Involved in daily scrum meeting, sprint review meeting, and agile release planning
  • Worked with SCRUM team in delivering agreed user stories on time for every sprint
  • Involved in setting up Sqoop framework to extract structured data product price, stock levels, stocking costs and sales from Oracle 11gi
  • Set up Flume system to periodically receive competitors product data in JSON format from webcrawler to HBase
  • Configured Flume to extract daily web log data in JSON format from web server to HBase
  • Worked with system administrator to come up with workflows in Oozie
  • Worked with business users, oracle database team and system administrators in designing data conversion to HBase
  • Developed custom InputFormat and Mapper program to read and parse files and return tab limited rows
  • Generated a Data Dictionary for category and sub categories of URL with MapReduce program
  • Simple to complex queries were written and executed in HIVE
  • Responsible for orchestrate hundreds of HIVE queries using Oozie workflows
  • Involved in importing data into Cloudera Impala and ran several reports on it
  • Participated in HIVE optimization project to effectively handle all tables archiving tables, deleting tables, converting tables to RC and ORC format, partitioning, and bucketing
  • Created UDF User Defined Functions to help create Pig Latin scripts
  • Worked with team in optimizing Pig Scripts following standard pig optimization techniques like split, schema and stream
  • Followed all performance enhancing techniques for executing Pig Latin scripts
  • Written various custom shell scripts to manage day to day operations
  • Interacted with Tableau team in designing and testing analytical reports
  • Worked with technical team onshore and offshore in preparing technical specs and unit testing development
  • Created and tested several java classes in JUnit4
  • Involved in pilot scale study of Spark with UC Berkley team

Environment: RedHat Linux, HDFS, CDH4, IMPALA, MapReduce2.0, Hive, Java JDK1.6, Pig, Sqoop, Flume, Zookeeper, Oozie, Oracle 11gi, HBase, Zookeeper, JUnit4

Confidential

Role: Hadoop Developer

Responsibilities:

  • Involved in analyzing business, system and data mapping requirements
  • Worked with Data scientists and architects to come up with SLA and HLD
  • Worked with team in coming up with LLD following HLD
  • Assisted administrator in setting up Hadoop cluster and created maintenance shell scripts
  • Worked on setting up 10 workstations to run Hadoop and provided POC to internal team
  • Sqoop was used to extract databases from MySQL database into HDFS, HIVE and HBase
  • Performed data validation on the data ingested using MapReduce by building a custom model to filter all the invalid data and cleansing the data
  • Assisted business users in ad-hoc queries and data metrics using Pig and Hive
  • Planned and designed HIVE for creating and maintaining tables join tables, alter tables
  • Optimized Hive tables by partitioning manual and dynamic , bucketing and compressing data
  • Created several Hive units tests using JUnit and executed these tests in Hive
  • Created Pig Latin scripts to clean up both structured and unstructured social and chat data for further analysis in HIVE
  • Developed custom load function for loading web server logs in Pig Latin
  • Extended existing UDF to create new UDF for further data analysis
  • Coordinates with test team lead in planning, creating and executing test scripts

Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase

Confidential

Role: Hadoop Developer

Responsibilities:

  • Prepared and conducted various meetings to finalize low level design documents
  • Assisted administrators in setting up Hadoop and its ecosystems
  • Worked with Oracle team in importing tables to HDFS using Sqoop
  • Flume was implemented to index email messages and load email data into HDFS
  • Created Pig Scripts to load unstructured data into structured data for further analysis
  • Planned and joined several relations using Pig JOIN commands and also used specialized joins for optimization
  • Developed custom filter function as UDF in Pig Script using java and registered respective Jar files with Pig Script
  • Tables were created using Hive scripts and tables were optimized by creating partitioning and bucketing of tables
  • DAG was designed to plan different workflows to execute several MapReduce, shell scripts, Hive and Pig Scripts in Oozie
  • Monitored daily all workflows scheduled in Oozie
  • Wrote shell script to verify the files availability in HDFS from various downstream processes
  • Developed MapReduce program to come up with inverted index for semi-structured and unstructured data files
  • Combiner program was developed to sort the data in Mapper phase and worked with admin in planning number of reducers for MapReduce programs
  • Worked with MicroStatergy vendor in data export from Hadoop system to MicroStatergy
  • Utilized MRUnit, HiveUnit and PigUnit to test MapReduce, Hive and Pig Scripts respectively
  • Coordinated with testing team in planning and executing test scripts
  • Involved in supporting Level3 all Hadoop and its ecosystems related issues

Environment: CDH3, J2EE, Pig Latin, HIVE, Oozie, JUnit, MRUnit, UML, JIRA, Linux, iLog, Maven, Flume, MapReduce, HDFS, HBase, JUnit3

Confidential

Role: Sr. Java Developer

Responsibilities:

  • Followed Waterfall methodology throughout the life cycle of the project
  • Prepared program and system specifications for web project development
  • Build the prototype for the application and user end screens
  • Adopted SPRING MVC framework in web application development
  • Front end was developed using JavaScript, XHTML, JSP, CSS
  • Developed JavaScript to perform client side validations and client forms
  • Created web services and SOAP/WSDL was used as part web service implementation
  • Spring Framework was used to design and implement domain layer based on Aspect Oriented Programming AOP and inversion control IOC
  • JMS was used to handle transfer of asynchronous messages using publish/subscribe model
  • Involved in performing client side validation using JQuery and server side validation using struts 2 validation framework
  • Interacted with Oracle database through Hibernate
  • Retrieved data from Oracle database through SQL commands and stored procedures
  • Planned and created test plans and integrated testing in JUnit
  • Implemented MyBatis ORM tools to automate the mapping between SQL databases and java objects
  • Worked with testing team in fixing bugs and failed test scripts in QA and UAT

Environment: Java, JSF, Tomcat Web Server, J2EE, Hibernate, Oracle, JSON, HTML, SQL, Spring, Structs, JUnit, Eclipse, UNIX, Windows, PL/SQL, SOAP, WSDL

Confidential

Role: Sr. Java Developer

Responsibilities:

  • Reviewed functional specifications with analysts
  • Created LLD documents sequence, class diagrams etc using MS enterprise architecture tool
  • Used spring framework for application building based on MVC design paradigm
  • Developed several Web pages using JSP, JavaScript, HTML
  • Involved in design and development of log and transaction using SpringAOP
  • Developed Java classes and beans to interact with Oracle database using Hibernate
  • Developed DIS web services using jaxws and worked on consuming web services JaxRs
  • Worked on creating DAO classes to get connection instances with Spring JDBC and configured JDBC connections with Oracle 10g
  • Designed and used Jmeter schemas for testing front tier of the application
  • Worked on Jersey Framework using REST API
  • Performed database development using PL/SQL, implemented search mechanism using stored procedures and sequences
  • Worked with project managers in planning change control management for the project
  • Identified and prepared Unit Test script for all the business scenarios
  • Assisted technical team in upgrade of CRM from 2007 to 7.0
  • Worked on completing cutover activities

Environment: Java 1.5, J2EE, Servlets, JSP, JUnit, Eclipse, WinCVS,Websphere5.1, Oracle 10gi, Windows XP, Eclipse 3.0, HTML, EJB 2.0

Confidential

Role: Java Developer

Responsibilities:

  • Documented functional/technical specifications based on business users requirements
  • Created UML diagrams such as Use Cases, Sequence Diagrams, Class Diagrams and flowcharts using Visio
  • Designed and developed interactive GUI screens using XML, JSPs, HTML, JavaScript, CSS
  • Worked on the JAVA collections API for handling the data objects between front end and business layers
  • Developed web service provider application using WSAD, XSLT and WSDL and implemented SoapUI for testing
  • Developed webpage layouts using Strut-tiles, JSF core and HTML tags
  • Persistence framework mapping the ORM objects to tables through Hibernate
  • Developed using Eclipse as IDE with Tomcat as the server
  • Developed various stored procedures and triggers with PL/SQL
  • Designed and developed Web Services to provide service to clients using WSDL and SOAP
  • Created extensive JUnit test cases in order to verify functionality of the methods and used log4j for logging messages
  • Worked with users and testing team in fixing defects during integration testing
  • Deployed WSS applications using Maven on to development and testing environments
  • Worked on integrating search system, metrics and email applications to websites using JSF framework
  • Created ant scripts and developed builds using Apache ANT

Environment: J2EE, Core Java, Struts 1.0, Tiles, JSP, JNDI, ANT, JavaScript, XML, Eclipse

Confidential

Role: Java Developer

Responsibilities:

  • Involved in the analysis, design and development of SDLC
  • Involved in writing multithreaded programs to measure system performance
  • Designed and documented functional/technical specifications based on business users requirements
  • Designed and developed UI view using Strut tags, JSP and HTML
  • Worked on JDBC APIs for database connectivity and data retrieval
  • Worked on bug tracking in JIRA and debugging using Eclipse 3.6
  • Analyzed and configured JCL to execute mainframe batch
  • Wrote stored procedures, view, user-defined functions, and triggers in SQL server database for report
  • Involved in automating nightly batch processing and job alerts using REXX
  • Worked on XML parser APIs like DOM for marshalling and unmarshalling of data
  • User Clear Case for version control and JUnit for unit testing of applications
  • Implemented stored procedures and PL/SQL for Oracle 9i database

Environment: Core Java, JUnit, JSP, Eclipse, JIRA, JDBC, Struts 2, Oracle 9i, MS SQL Server, REXX, Mainframe, JCL, TomCat, Clear Case

Confidential

Role: Engineer II

Roles and Responsibilities:

  • Provided estimates of effort required to design and develop solutions and participated in design sessions with architects and developers
  • Prepared the functional/technical specification document following requirements from business users
  • Involved in creating HLD High Level Design and detailed design documents using Visio
  • Involved in development of web based component for RFP Request For Proposal processing
  • Developed web tier using Struts MVC framework
  • Created DAO Data Access Object classes to fetch and connect data from database
  • Involved in the development of both UI JSP, Javascript, CSS and backend flow java, design patterns, Oracle
  • Prepared test plans and test plan execution unit, functional and integration testing, end to end testing
  • Coordinated defect meeting and documented the same

Environment: Java, Oracle 9i, JDBC, javascript, Struts1.x, Hibernate, Log4j, Eclipse, CVS, HTML, JSP, Weblogic 8.1, CSS

We'd love your feedback!