Sr. Big Data/hadoop Developer Resume
Atlanta, GA
PROFESSIONAL SUMMARY:
- Above 8+ working experience as a Big Data/Hadoop Developer in designed and developed various applications like big data, Hadoop, Java/J2EE open - source technologies.
- Strong development skills in Hadoop, HDFS, Map Reduce, Hive, Sqoop, HBase with solid understanding of Hadoop internals.
- Experience in Programming and Development of java modules for an existing web portal based in Java using technologies like JSP, Servlets, JavaScript and HTML, SOA with MVC architecture.
- Expertise in ingesting real time/near real time data using Flume, Kafka, Storm
- Good knowledge of NO SQL databases like Mongo DB, Cassandra and HBase.
- Excellent knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MRA and MRv2 (YARN).
- Expertise in writing Hadoop Jobs to analyze data using MapReduce, Apache Crunch, Hive, Pig and SOLR, Splunk.
- Hands on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, Pig, Hive, HBase, Apache Crunch, Zookeeper, Scoop, Hue, Scala, AVRO.
- Strong Programming Skills in designing and implementing of multi-tier applications using Java, J2EE, JDBC, JSP, JSTL, HTML, CSS, JSF, Struts, JavaScript, Servlets, POJO, EJB, XSLT, JAXB.
- Extensive experience in SOA-based solutions - Web Services, Web API, WCF, SOAP including Restful APIs services
- Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services which provides fast and efficient processing of Teradata Big Data Analytics.
- Experienced in collection of Log Data and JSON data into HDFS using Flume and processed the data using Hive/Pig.
- Expertise in developing a simple web based application using J2EE technologies like JSP, Servlet, and JDBC.
- Experience working on EC2 (Elastic Compute Cloud) cluster instances, setup data buckets on S3 (Simple Storage Service), set EMR (Elastic MapReduce).
- Work Extensively in Core Java, Struts2, JSF2.2, Spring3.1, Hibernate, Servlets, JSP and Hands-on experience with PL/SQL, XML and SOAP.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, DataNode.
- Well versed working with Relational Database Management Systems as Oracle 9i/12c, MS SQL, MySQL Server
- Hands on experience in working on XML suite of technologies like XML, XSL, XSLT, DTD, XML Schema, SAX, DOM, JAXB.
- Hands on experience in advanced Big-Data technologies like Spark Ecosystem (Spark SQL, MLlib, SparkR and Spark Streaming), Kafka and Predictive analytics
- Knowledge of the software Development Life Cycle (SDLC), Agile and Waterfall Methodologies.
- Experienced on applications using Java, python and UNIX shell scripting
- Experience in consuming Web services with Apache Axis using JAX-RS(REST) API's.
- Experienced in building tool Maven, ANT and logging tool Log4J.
- Experience in working with Web Servers like Apache Tomcat and Application Servers like IBM Web Sphere and JBOSS.
- Good knowledge of NoSQL databases such as HBase, MongoDB and Cassandra.
- Experience in working with Eclipse IDE, Net Beans, and Rational Application Developer.
- Experience in using PL/SQL to write Stored Procedures, Functions and Triggers.
- Experience includes Requirements Gathering, Design, Development, Integration, Documentation, Testing and Build.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, DataNode.
- Hands on experience on Hadoop /Big Data related technology experience in Storage, Querying, Processing and analysis of data.
TECHNICAL SKILLS
Hadoop Ecosystem: Hadoop 2.7/2.5, MapReduce, Sqoop, Hive, Oozie, Pig, HDFS 1.2.4, Zookeeper, Flume, Impala, Spark 2.0/2.0.2, Storm, Hadoop (Cloudera), Hortonworks and Pivotal).
Big Data Platforms: Hortonworks, Cloudera, Amazon AWS, Apache.
Databases & NOSQL Databases: Oracle12c/11g, MYSQL, Microsoft SQL Server2016/2014, MongoDB, HBase and Cassandra.
Operating Systems: Linux, UNIX, Windows8/7.
Development Methodologies: Agile/Scrum, Waterfall.
IDEs: Eclipse, Net Beans, GitHub, Jenkins, Maven, IntelliJ, Ambari.
Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, Xpath, Spark
Java/J2EE Technologies: JDBC, Java Script, JSP, Servlets, JQuery
Web Technologies: HTML5/4, DHTML, XML, XHTML, JavaScript, CSS3/2, XSLT, AWS, Dynamo DB
Frameworks: Struts 1.2, 2.0, spring 3.0, Hibernate 4.3.
Web/Application Servers: WebLogic (8.1), IBM WebSphere Application Server (6.0), Tomcat 5.x/6.x/7.x JBoss and Apache WebServer.
WORK EXPERIENCE:
Confidential, Atlanta, GA
Sr. Big Data/Hadoop Developer
Responsibilities:
- Worked closely with the business analysts to convert the Business Requirements into Technical Requirements and prepared low and high level documentation.
- Worked on business problems to develop and articulate solutions using Teradata's
- Worked on analyzing different big data analytic tools including Hive, Impala and Sqoop in importing data from RDBMS to HDFS.
- Configured Spark Streaming to receive real time data from the Kafka and store the stream data to HDFS.
- Developed MapReduce/Spark Python modules for machine learning & predictive analytics in Hadoop on AWS.
- Designed high level ETL architecture for overall data transfer from the OLTP to OLAP.
- Improving the performance and optimization of existing algorithms in Hadoop using Spark context, Spark-SQL and Spark YARN.
- Involved in creating Data Lake by extracting customer's Big Data from various data sources into Hadoop HDFS.
- Created various Documents such as Source-To-Target Data mapping Document, Unit Test, Cases and Data Migration Document.
- Imported data from structured data source into HDFS using Sqoop incremental imports.
- Performed data synchronization between EC2 and S3, Hive stand-up, and AWS profiling.
- Created Hive tables, partitions and implemented incremental imports to perform ad-hoc queries on structured data.
- Worked with NoSQL databases like HBase, Cassandra, DynamoDB (AWS) and MongoDB.
- Involved in loading data from UNIX tile system to HOPS using Flume and Kettle and HDFS API.
- Developed SQL scripts using Spark for handling different data sets and verifying the performance over Map Reduce jobs.
- Involved in converting Map Reduce programs into Spark transformations using Spark RDD's using Scala and Python.
- Supported MapReduce Programs those are running on the cluster and also wrote MapReduce jobs using Java API.
- Utilized Agile Scrum Methodology to help manage and organize a team of 4 developers with regular code review sessions.
- Wrote complex SQL and PL/SQL queries for stored procedures.
- Used S3 Bucket to store the jar's, input datasets and used Dynamo DB to store the processed output from the input data set.
- Created MapReduce running over HDFS for data mining and analysis using R and Loading & Storage data to Pig Script and R for MapReduce operations.
- Used Cloudera Manager for installation and management of Hadoop Cluster.
- Developing data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Worked on MongoDB, HBase (NoSQL) databases which differ from classic relational databases
- Involved in converting HiveQL into Spark transformations using Spark RDD and through Scala programming.
- Integrated Kafka-Spark streaming for high efficiency throughput and reliability
- Worked on Apache Flume for collecting and aggregating huge amount of log data and stored it on HDFS for doing further analysis.
- Worked in tuning Hive & Pig to improve performance and solved performance issues in both scripts
- Used Singleton, DAO, DTO, Session Facade, MVC design Patterns.
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
- Collaborating with business users/product owners/developers to contribute to the analysis of functional requirements.
Environment: Hadoop, HDFS, MapReduce, Unix, REST, Python, Pig, Hive, HBase, NoSQL, Flume, Zookeeper, Cloudera, AWS, Kafka, Cassandra, Informatica, Teradata, Scala
Confidential, PA
Big Data/Hadoop Developer
Responsibilities:
- Responsible for developing prototypes the selected solutions and implementing complex big data projects with a focus on collecting, parsing, managing, analyzing and visualizing large sets of data using multiple platforms.
- Developed Spark Applications by using Scala, Java and Implemented Apache Spark data processing project to handle data from various RDBMS and Streaming sources.
- Used Spark Streaming APIs to perform transformations and actions on the fly for building common learner data model which gets the data from Kafka in near real time and persist it to Cassandra.
- Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
- Developed Simple to complex MapReduce Jobs using Hive and Pig.
- Performed importing data from various sources to the Cassandra cluster using Sqoop.
- Worked on creating data models for Cassandra from Existing Oracle data model.
- Used Spark - Cassandra connector to load data to and from Cassandra.
- Handle ETL Framework in Spark for writing data from HDFS to Hive.
- Use Scala based written framework for ETL.
- Developed multiple spark streaming and core jobs with Kafka as a data pipe-line system
- Worked and learned a great deal from AWS Cloud services like EC2, S3, EBS.
- Migrated an existing on-premises application to AWS.
- Used AWS services like EC2 and S3 for small data sets processing and storage.
- Imported data from AWS S3 into Spark RDD, Performed transformations and actions on RDD's.
- Extensively use Zookeeper as job scheduler for Spark Jobs.
- Worked in migrating from Informatica Talend jobs.
- Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka and Zookeeper.
- Developed Kafka producer and consumer components for real time data processing.
- Worked on physical transformations of data model which involved in creating Tables, Indexes, Joins, Views and Partitions.
- Involved in Analysis, Design, System architectural design, Process interfaces design, design documentation.
- Understand how to apply technologies to solve big data problems and to develop innovative big data solutions.
Environment: Spark, HDFS, Kafka, MapReduce (MR1), Pig, Hive, Sqoop, Cassandra, AWS, Talend, Java, Linux Shell Scripting.
Confidential, Boston, MA
Sr. Java/Hadoop Developer
Responsibilities:
- Played a key role in discussing about the requirements, analysis of the entire system along with estimation, development and testing accordingly keeping BI requirements as a note.
- Involved in the analysis, design and development of the application based on J2EE using Spring and Hibernate.
- Designed and developed frameworks for Payment Workflow System, Confirmations Workflow System, Collateral System using GWT, Core Java, Servlets, JavaScript, XML, AJAX, J2EE design patterns and OOPS/J2EE technologies.
- Involved actively in designing web page using HTML, Backbone, AngularJS, JQuery, JavaScript, Bootstrap and CSS.
- Used Hibernate to manage Transactions (update, delete) along with writing complex SQL and HQL queries.
- The business logic is developed using J2EE framework and deployed components on Application server where Eclipse was used for component building.
- Established continuous integration (CI) practices and standards for PSCU with JIRA, Jenkins,
- Developed the user interface screens using JavaScript and HTML and also conducted client side validations.
- Used JDBC to connect to database and wrote SQL queries and stored procedures to fetch and insert/update to database tables.
- Conducted data analysis with basic Python and wrangled data for data repositories.
- Applied machine learning principles for studying market behavior for trading platform.
- Used Maven as the build tool and Tortoise SVN as the Source version controller.
- Developed data mapping to create a communication bridge between various application interfaces using XML, and XSL.
- Involved in developing JSP for client data presentation and, data validation on the client side with in the forms.
- Involved in various phases of Software Development Life Cycle (SDLC) as design development and unit testing.
- Excessive work in writing SQL Queries, Stored procedures, Triggers using TOAD.
- Code development using core java concepts to provide service and persistence layers. Used JDBC to provide connectivity layer to the Oracle database for data transaction.
- Implemented logging and transaction manager using spring's Aspect Oriented Programming (AOP) concept.
- Created build scripts for compiling and creating war, jar using ANT tool kit.
- Created Application Configuration tool using Web works MVC framework and HTML, CSS and JavaScript.
- Developed Web applications using Spring Core, Spring MVC, IBatis, Apache, Tomcat, JSTL and Spring tag libraries.
- User help tooltips implemented with Dojo Tooltip Widget with multiple custom colors
- Used eclipse as IDE to write the code and debug application using separate log files.
- Used Angular to connect the web application to back-end APIs, used RESTFUL methods to interact with several API's,
- Developed POJO classes and writing Hibernate query language (HQL) queries.
- Experience in using TIBCO Administrator for User Management, Resource Management and Application Management.
- Developed user interface using JSP, JSP Tag libraries to simplify the complexities of the application.
- Used Talend reusable components context variable and global Map variable.
- Assisted in tuning the performance of Hadoop eco system.
Environment: Java 1.5/1.7, Core java, Swing, Struts Framework 2.0, Hibernate4.0, Eclipse 3.2, Junit 4.x, JSP 2.x, Oracle SQL Developer 2.1, Oracle WebLogic 12.1, Restful Web Services, SOAP, HP Quality Center 10, Tortoise SVN 1.5
Confidential, New York, NY
Sr. Java/J2EE Developer
Responsibilities:
- Worked as part of the Agile Application Architecture (A3) development team responsible for setting up the architectural components for different layers of the application.
- Developed the application using Spring Framework that leverages classical Model View Controller (MVC) architecture.
- Involved in Software Development Life cycle starting from requirements gathering and performed OOA and OOD
- Used Spring JDBC to execute database queries.
- Created row mappers and query classes for DB operations.
- Created a Transaction History Web Service using SOAP that is used for internal communication in the workflow process.
- Designed and created components for company's object framework using best practices and design Patterns such as Model-View-Controller (MVC).
- Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
- Debugged the application using Firebug to traverse the documents.
- Involved in writing SQL Queries, Stored Procedures and used JDBC for database connectivity with MySQL Server.
- Developed the presentation layer using CSS and HTML taken from Bootstrap to develop for browsers.
- Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clear case, and ANT.
- Used Spring Core and Spring-web framework. Created a lot of classes for backend.
- Involved in developing web pages using HTML and JSP.
- Exposed business functionality to external systems (Interoperable clients) using Web Services (WSDL-SOAP) Apache Axis.
- Used PL/SQL for queries and stored procedures in SQL as the backend RDBMS.
- Involved in the Analysis and Design of the front-end and middle tier using JSP, Servlets and Ajax.
- Implemented Spring IOC or Inversion of Control by way of Dependency Injection where a Factory class was written for creating and assembling the objects.
- Implemented modules using Core Java APIs, Java collection, Threads, XML, and integrating the modules and used SOAP for Web Services by exchanging XML data between applications over HTTP.
- Created EJB, JPA and Hibernate component for the application.
- Implemented XML parsers with SAX, DOM, and JAXB XML Parser Libraries to Modify User view of Products and Product information in Customized view with XML, XSD, XSTL in HTML, XML, PDF formats.
- Used Microsoft VISIO for developing Use Case Diagrams, Sequence Diagrams and Class Diagrams in the design phase.
- Developed Restful Web services client to consume JSON messages using Spring JMS configuration. Developed the message listener code.
- Providing production support which includes handling tickets & providing resolution. Used BMC Remedy Tool to add issues & update resolutions.
- Create database objects like tables, sequences, views, triggers, stored procedures, functions packages.
- Used Maven as the build tool and Tortoise SVN as the Source version controller.
- Performed data manipulations using Talend.
Environment: Core Java, UNIX, J2EE, XML Schemas, XML, JavaScript, JSON, CSS3, HTML5, spring, Hibernate, Design Patterns, Servlets, JUnit, JMS, MySQL, REST, Python, Web Services, Apache Tomcat, Windows XP
Confidential
Java/J2EE Developer
Responsibilities:
- Performed Requirements gathering, Analysis, Design, Code development, Testing using Agile methodologies.
- Designed and implemented the User Interface using JavaScript, HTML, XHTML, XML, CSS, JSP, and AJAX.
- Wrote web service client for tracking operations for the orders which is accessing web services API and utilizing in our web application.
- Implemented data archiving and persistence of report generation meta-data using Hibernate by creating Mapping files, POJO classes and configuring hibernate to set up the data sources.
- Developed Spring framework DAO Layer with JPA and EJB3 in Imaging Data model and Doc Import.
- The business logic is developed using J2EE framework and deployed components on Application server where Eclipse was used for component building.
- Actively involved in deployment EJB service jars, Application war files in WebLogic Application server.
- Developed GUI screens for login, registration, edit account, forgot password and change password using Struts 2.
- Used JUnit framework for unit testing of application and JUL logging to capture the log that includes runtime exceptions
- Writing SQL queries for data access and manipulation using Oracle SQL Developer.
- Developed Session Bean to encapsulate the business logic and Model and DAO classes using Hibernate
- Designed and coded JAX-WS based Web Services used to access external financial information.
- Implemented EJB Components using State less Session Bean and State full session beans.
- Used spring framework with the help of Spring Configuration files to create the beans needed and injected dependency using Dependency Injection.
- Utilized JPA for Object/Relational Mapping purposes for transparent persistence onto the Oracle database.
- Involved in creation of Test Cases for JUnit Testing.
- Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
- Used SOAP as a XML-based protocol for web service operation invocation.
- Packaged and deployed the application in IBM WebSphere Application server in different environments like Development, testing etc.
- Used Log4J to validate functionalities and JUnit for unit testing.
Environment: Java, Servlets, JSP, Struts 1.0, Hibernate3.1, spring core, Spring JDBC, HTML, JavaScript, AJAX, XSL, XSLT, XSD schema, XML Beans, Web logic, Oracle9i