We provide IT Staff Augmentation Services!

Sr. Big Data/java Engineer Resume

Boston, MA


  • Above 9+ working experience as a Big Data/Java Engineer in designed and developed various applications like big data, Hadoop, Java/J2EE open - source technologies.
  • Strong development skills in Hadoop, HDFS, Map Reduce, Hive, Sqoop, HBase with solid understanding of Hadoop internals.
  • Experience in Programming and Development of java modules for an existing web portal based in Java using technologies like JSP, Servlets, JavaScript and HTML, SOA with MVC architecture.
  • Expertise in ingesting real time/near real time data using Flume, Kafka, Storm
  • Good knowledge of NO SQL databases like Mongo DB, Cassandra and HBase.
  • Excellent knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MRA and MRv2 (YARN).
  • Expertise in writing Hadoop Jobs to analyze data using MapReduce, Apache Crunch, Hive, Pig and SOLR, Splunk.
  • Hands on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, Pig, Hive, HBase, Apache Crunch, Zookeeper, Scoop, Hue, Scala, AVRO.
  • Strong Programming Skills in designing and implementing of multi-tier applications using Java, J2EE, JDBC, JSP, JSTL, HTML, CSS, JSF, Struts, JavaScript, Servlets, POJO, EJB, XSLT, JAXB.
  • Extensive experience in SOA-based solutions - Web Services, Web API, WCF, SOAP including Restful APIs services
  • Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services which provides fast and efficient processing of Teradata Big Data Analytics.
  • Experienced in collection of Log Data and JSON data into HDFS using Flume and processed the data using Hive/Pig.
  • Expertise in developing a simple web based application using J2EE technologies like JSP, Servlet, and JDBC.
  • Experience working on EC2 (Elastic Compute Cloud) cluster instances, setup data buckets on S3 (Simple Storage Service), set EMR (Elastic MapReduce).
  • Work Extensively in Core Java, Struts2, JSF2.2, Spring3.1, Hibernate, Servlets, JSP and Hands-on experience with PL/SQL, XML and SOAP.
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, DataNode.
  • Well versed working with Relational Database Management Systems as Oracle 9i/12c, MS SQL, MySQL Server
  • Hands on experience in working on XML suite of technologies like XML, XSL, XSLT, DTD, XML Schema, SAX, DOM, JAXB.
  • Hands on experience in advanced Big-Data technologies like Spark Ecosystem (Spark SQL, MLlib, SparkR and Spark Streaming), Kafka and Predictive analytics
  • Knowledge of the software Development Life Cycle (SDLC), Agile and Waterfall Methodologies.
  • Experienced on applications using Java, python and UNIX shell scripting
  • Experience in consuming Web services with Apache Axis using JAX-RS(REST) API's.
  • Experienced in building tool Maven, ANT and logging tool Log4J.
  • Experience in working with Web Servers like Apache Tomcat and Application Servers like IBM Web Sphere and JBOSS.
  • Good knowledge of NoSQL databases such as HBase, MongoDB and Cassandra.
  • Experience in working with Eclipse IDE, Net Beans, and Rational Application Developer.


Hadoop Ecosystem: Hadoop 2.7/2.5, MapReduce, Sqoop, Hive, Oozie, Pig, HDFS 1.2.4, Zookeeper, Flume, Impala, Spark 2.0/2.0.2, Storm, Hadoop (Cloudera), Hortonworks and Pivotal).

Big Data Platforms: Hortonworks, Cloudera, Amazon AWS, Apache.

Databases & NOSQL Databases: Oracle12c/11g, MYSQL, Microsoft SQL Server2016/2014, MongoDB, HBase and Cassandra.

Operating Systems: Linux, UNIX, Windows8/7.

Development Methodologies: Agile/Scrum, Waterfall.

IDEs: Eclipse, Net Beans, GitHub, Jenkins, Maven, IntelliJ, Ambari.

Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, Xpath, Spark

Java/J2EE Technologies: JDBC, Java Script, JSP, Servlets, JQuery

Web Technologies: HTML5/4, DHTML, XML, XHTML, JavaScript, CSS3/2, XSLT, AWS, Dynamo DB

Frameworks: Struts 1.2/2.0, spring 3.0, Hibernate 4.3.

Web/Application Servers: WebLogic (8.1), IBM WebSphere Application Server (6.0), Tomcat 5.x/6.x/7.x JBoss and Apache WebServer.


Confidential, Boston, MA

Sr. Big Data/Java Engineer


  • Analyzed client requirements by interacting with various business divisions and consolidated the SRS (Software Requirements Specification) Document.
  • Implemented UI Tier with JSP, Groovy, Servlet, JSF, XML, XML Schema, CSS, JavaScript, AJAX
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in developing JSP pages using Struts custom tags, JQuery and Tiles Framework.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Worked on customizing Map Reduce code in Amazon EMR using Hive, Pig, Impala frameworks
  • Implemented multiple Map Reduce Jobs in Java for data cleansing and pre-processing.
  • Worked on analyzing different big data analytic tools including Hive, Impala and Sqoop in importing data from RDBMS to HDFS.
  • Imported data from structured data source into HDFS using Sqoop incremental imports.
  • Implemented end-to-end systems for Data Analytics, Data Automation and integrated with custom visualization tools using R, Hadoop and MongoDB, Cassandra.
  • Developed job workflow in Oozie to automate the tasks of loading the data into HDFS.
  • Analyzed and prepared functional specifications for the business and system requirements.
  • Developed custom MapReduce use-cases using Java to log data of customer behavior and loaded it into the HDFS.
  • Fixed bugs and improved Java source code to support clusters.
  • Applied Sqoop to execute processes between the Oracle database to the HDFS.
  • Loaded and transformed large sets of structured, semi structured and unstructured data.
  • Migrated the dataset into Hive for ETL purposes and optimized Pig UDFs.
  • Wrote column-mapping scripts to generate ETL Queries in Hive.
  • Developed Hive Schema to help the business user extract data files.
  • Handled importing of data from various data sources, performed transformations using Pig and Hive.
  • Used Impala to query data stored in the HDFS.
  • Developed the presentation layer using CSS and HTML taken from Bootstrap to develop for browsers.
  • Participated in Mahout implementation for machine learning analysis.
  • Performed data analysis on large datasets and present results to risk, finance, accounting and pricing, sales, marketing, and compliance teams.
  • Imported data into excel and created pivot tables and statistical models.
  • Developed many modules & functionalities using JSP, Spring MVC, Spring DI & Hibernate.
  • Wrote SQL scripts to create database objects like tables, views triggers, sequences, PL/SQL stored procedures, functions packages.
  • Involved in developing forms using HTML and performing client side validations using Java Script.
  • Created JSP pages for the Customer module of the application.
  • Implemented logging and transaction manager using spring's Aspect Oriented Programming (AOP) concept.
  • Writing & executing JUnit & Soap UI test cases to test Java code & Web Services.
  • Designed and Developed the Data Repository for Report Dashboard using Oracle Business Intelligence.
  • Worked on developing UML diagrams like Use Cases and designing Class Diagrams, Sequence Diagrams.

Environment: Java, Python, Agile,HTML4, Hadoop, hive, HBase, HDFS, Sqoop, Spark, JSP, Servlets, Web sphere, Oracle 11g, Spring MVC, JQuery, JavaScript, CSS, Log4J, VSS, Unix, JSON

Confidential, Seattle, WA

Sr. Big Data/Java Engineer


  • Developed Pig Latin scripts for replacing the existing legacy process to the Hadoop and the data is fed to AWS S3.
  • Worked on MongoDB by using CRUD (Create, Read, Update and Delete), Indexing, Replication and Sharding features.
  • Created Talend jobs to read messages from Amazon AWS SQS queues & download files from AWS S3 buckets.
  • Worked on analyzing Hadoop cluster and different Big Data Components including Pig, Hive, Spark, HBase, Kafka, Elastic Search, database and SQOOP.
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Wrote MapReduce jobs to filter and parse inventory data which was stored in the HDFS.
  • Configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster for data pipelining.
  • Imported and exported data into the HDFS from the Oracle database using Sqoop.
  • Integrated MapReduce with Cassandra to import bulk amount of logged data.
  • Converted ETL operations to the Hadoop system using Hive transformations and functions.
  • Conducted streaming jobs with basic Python to process terabytes of formatted data for machine learning purposes.
  • Used Flume to collect, aggregate and store the web log data and loaded it into the HDFS.
  • Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
  • Developed custom and Pig UDFs for product specific needs.
  • Implemented and configured workflows using Oozie to automate jobs.
  • Performed Hadoop cluster management and configuration of multiple nodes on AWS.
  • Involved in creating buckets to store the data in AWS and stored the data repository for future needs and reusability.
  • Worked along with Tableau developers to help performance tune the visualizations graphs/analytics.
  • Involved in the cluster coordination services through Zookeeper.
  • Participated in the managing and reviewing of the Hadoop log files.
  • Used Elastic Search & MongoDB for storing and querying the offers and non-offers data.
  • Proficiency in developing Web applications using Servlets, JSP, JDBC, EJB2.0/3.0, web services using JAX-WS2.0 and JAX-RS APIS.
  • Import the data from different sources like HDFS/HBase into Spark RDD and developed a data pipeline using Kafka and Storm to store data into HDFS.
  • Used Spark streaming to receive real time data from the Kafka and store the stream data to HDFS using Scala and NoSQL databases such as HBase and Cassandra.
  • Worked with teams in setting up AWS EC2 instances by using different AWS services like S3, EBS, Elastic Load Balancer, and Auto scaling groups, VPC subnets and CloudWatch.
  • Utilized SDLC Methodology to help manage and organize a team of developers with regular code review sessions.
  • Developed Restful web services using JAX-RS and used DELETE, PUT, POST, GET HTTP methods
  • Created scalable and high-performance web services for data tracking and done High-speed querying.
  • Used Java Messaging Services (JMS) for reliable and asynchronous exchange of important information such as payment status report on IBM WebSphere MQ messaging system.
  • Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements.
  • Created and maintained various Shell and Python scripts for automating various processes and optimized MapReduce code, pig scripts and performance tuning and analysis.
  • Worked on Oozie workflow engine for job scheduling. Involved in Unit testing and delivered Unit test plans and results documents.
  • Involved with ingesting data received from various providers, on HDFS for big data operations.
  • Wrote MapReduce jobs to perform big data analytics on ingested data using Java API.
  • Wrote MapReduce in Ruby using Hadoop Streaming to implement various functionalities.
  • Performed transformations, cleaning and filtering on imported data using Hive, Map Reduce, and loaded final data into HDFS.
  • Conducted meetings with data analysts with basic Python and wrangled data for data repositories.

Environment: Hadoop, Java, MapReduce, AWS, HDFS, Scala, Python, MongoDB, Spark, Hive, Pig, Linux, XML, Cloudera, CDH4/5 Distribution, Oracle 12c, PL/SQL, EC2, Flume, Zookeeper, Cassandra, Hortonworks, Elastic search, IBM WebSphere

Confidential, Bloomington, IL

Sr. Java/J2EE Developer


  • Worked as part of the Agile Application Architecture (A3) development team responsible for setting up the architectural components for different layers of the application.
  • Developed the application using Spring Framework that leverages classical Model View Controller (MVC) architecture.
  • Involved in Software Development Life cycle starting from requirements gathering and performed OOA and OOD
  • Used Spring JDBC to execute database queries.
  • Created row mappers and query classes for DB operations.
  • Created a Transaction History Web Service using SOAP that is used for internal communication in the workflow process.
  • Designed and created components for company's object framework using best practices and design Patterns such as Model-View-Controller (MVC).
  • Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
  • Debugged the application using Firebug to traverse the documents.
  • Involved in writing SQL Queries, Stored Procedures and used JDBC for database connectivity with MySQL Server.
  • Developed the presentation layer using CSS and HTML taken from Bootstrap to develop for browsers.
  • Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clear case, and ANT.
  • Used Spring Core and Spring-web framework. Created a lot of classes for backend.
  • Involved in developing web pages using HTML and JSP.
  • Exposed business functionality to external systems (Interoperable clients) using Web Services (WSDL-SOAP) Apache Axis.
  • Used PL/SQL for queries and stored procedures in SQL as the backend RDBMS.
  • Involved in the Analysis and Design of the front-end and middle tier using JSP, Servlets and Ajax.
  • Implemented Spring IOC or Inversion of Control by way of Dependency Injection where a Factory class was written for creating and assembling the objects.
  • Implemented modules using Core Java APIs, Java collection, Threads, XML, and integrating the modules and used SOAP for Web Services by exchanging XML data between applications over HTTP.
  • Created EJB, JPA and Hibernate component for the application.
  • Implemented XML parsers with SAX, DOM, and JAXB XML Parser Libraries to Modify User view of Products and Product information in Customized view with XML, XSD, XSTL in HTML, XML, PDF formats.
  • Used Microsoft VISIO for developing Use Case Diagrams, Sequence Diagrams and Class Diagrams in the design phase.
  • Developed Restful Web services client to consume JSON messages using Spring JMS configuration. Developed the message listener code.
  • Providing production support which includes handling tickets & providing resolution. Used BMC Remedy Tool to add issues & update resolutions.
  • Create database objects like tables, sequences, views, triggers, stored procedures, functions packages.
  • Used Maven as the build tool and Tortoise SVN as the Source version controller.

Environment: Core Java, UNIX, J2EE, XML Schemas, XML, JavaScript, JSON, CSS3, HTML5, spring, Hibernate, Design Patterns, Servlets, JUnit, JMS, MySQL, REST, Python, Web Services, Apache Tomcat, Windows XP

Confidential, Richmond, VA

Sr. Java/J2EE Developer


  • Played a key role in discussing about the requirements, analysis of the entire system along with estimation, development and testing accordingly keeping BI requirements as a note.
  • Involved in the analysis, design and development of the application based on J2EE using Spring and Hibernate.
  • Involved actively in designing web page using HTML, Backbone, AngularJS, JQuery, JavaScript, Bootstrap and CSS.
  • Created Application Configuration tool using Web works MVC framework and HTML, CSS and JavaScript.
  • Developed Web applications using Spring Core, Spring MVC, IBatis, Apache, Tomcat, JSTL and Spring tag libraries.
  • User help tooltips implemented with Dojo Tooltip Widget with multiple custom colors
  • Used eclipse as IDE to write the code and debug application using separate log files.
  • Designed and developed frameworks for Payment Workflow System, Confirmations Workflow System, Collateral System using GWT, Core Java, Servlets, JavaScript, XML, AJAX, J2EE design patterns and OOPS/J2EE technologies.
  • Used Hibernate to manage Transactions (update, delete) along with writing complex SQL and HQL queries.
  • The business logic is developed using J2EE framework and deployed components on Application server where Eclipse was used for component building.
  • Established continuous integration with JIRA, Jenkins,
  • Developed the user interface screens using JavaScript and HTML and also conducted client side validations.
  • Used JDBC to connect to database and wrote SQL queries and stored procedures to fetch and insert/update to database tables.
  • Used Maven as the build tool and Tortoise SVN as the Source version controller.
  • Developed data mapping to create a communication bridge between various application interfaces using XML, and XSL.
  • Involved in developing JSP for client data presentation and, data validation on the client side with in the forms.
  • Involved in various phases of Software Development Life Cycle (SDLC) as design development and unit testing.
  • Excessive work in writing SQL Queries, Stored procedures, Triggers using TOAD.
  • Code development using core java concepts to provide service and persistence layers. Used JDBC to provide connectivity layer to the Oracle database for data transaction.
  • Implemented logging and transaction manager using spring's Aspect Oriented Programming (AOP) concept.
  • Created build scripts for compiling and creating war, jar using ANT tool kit.
  • Used Angular to connect the web application to back-end APIs, used RESTFUL methods to interact with several API's,
  • Developed POJO classes and writing Hibernate query language (HQL) queries.
  • Experience in using TIBCO Administrator for User Management, Resource Management and Application Management.
  • Developed user interface using JSP, JSP Tag libraries to simplify the complexities of the application.

Environment: Java 1.5/1.7, Core java, Swing, Struts Framework 2.0, Hibernate4.0, Eclipse 3.2, Junit 4.x, JSP 2.x, Oracle SQL Developer 2.1, Oracle WebLogic 12.1, Restful Web Services, SOAP, Tortoise SVN 1.5

Confidential, Hyderabad, IN

Java/J2EE Developer


  • Performed Requirements gathering, Analysis, Design, Code development, Testing using Agile methodologies.
  • Designed and implemented the User Interface using JavaScript, HTML, XHTML, XML, CSS, JSP, and AJAX.
  • Wrote web service client for tracking operations for the orders which is accessing web services API and utilizing in our web application.
  • Implemented data archiving and persistence of report generation meta-data using Hibernate by creating Mapping files, POJO classes and configuring hibernate to set up the data sources.
  • Developed Spring framework DAO Layer with JPA and EJB3 in Imaging Data model and Doc Import.
  • The business logic is developed using J2EE framework and deployed components on Application server where Eclipse was used for component building.
  • Actively involved in deployment EJB service jars, Application war files in WebLogic Application server.
  • Developed GUI screens for login, registration, edit account, forgot password and change password using Struts.
  • Used JUnit framework for unit testing of application and JUL logging to capture the log that includes runtime exceptions
  • Writing SQL queries for data access and manipulation using Oracle SQL Developer.
  • Developed Session Bean to encapsulate the business logic and Model and DAO classes using Hibernate
  • Designed and coded JAX-WS based Web Services used to access external financial information.
  • Implemented EJB Components using State less Session Bean and State full session beans.
  • Used spring framework with the help of Spring Configuration files to create the beans needed and injected dependency using Dependency Injection.
  • Utilized JPA for Object/Relational Mapping purposes for transparent persistence onto the Oracle database.
  • Involved in creation of Test Cases for JUnit Testing.
  • Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
  • Used SOAP as a XML-based protocol for web service operation invocation.
  • Packaged and deployed the application in IBM WebSphere Application server in different environments like Development, testing etc.
  • Used Log4J to validate functionalities and JUnit for unit testing.

Environment: Java, Servlets, JSP, Struts 1.0, Hibernate3.1, spring core, Spring JDBC, HTML, JavaScript, AJAX, XSL, XSLT, XSD schema, XML Beans, Web logic, Oracle9i

Hire Now