We provide IT Staff Augmentation Services!

Sr. Big Data Developer Resume

4.50/5 (Submit Your Rating)

West Point, PA

PROCESSIONAL SUMMARY:

  • Having 8+ years of experience in Analysis, Architecture, Design, Development, Testing, Maintenance, and User training of software application which includes around 3+ Years in Big Data, Spark & Hadoop environment including Java/J2EE Technologies.
  • Excellent understanding of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Excellent working experience with Hadoop distributions such as Hortonworks, Cloudera, and IBM Big Insights.
  • Strong hands on experience with Hadoop ecosystem components like Hadoop Map Reduce, YARN, HDFS, Hive, Pig, HBase, Storm, Sqoop, Impala, Oozie, Kafka, Spark, and Zookeeper.
  • Expertise in loading and transforming large sets of structured, semi - structured and unstructured data.
  • Good Knowledge in Amazon Web Service (AWS) concepts like EMR and EC2 web services.
  • Experienced in analyzing data with Hive Query Language (HQL) and Pig Latin Script.
  • Expertise in optimizing Map Reduce algorithms using Mappers, Reducers, and combiners to deliver the best results for the large datasets.
  • Hands on experience on building Data lakes using trending data lake tools Cask CDAP and Zaloni.
  • Worked on Data ingestion to Airbnb Superset for data visualization and data access.
  • Very good experience in writing Map Reduce jobs using Java native code, Pig, and Hive for various business use cases.
  • Excellent Hands on Experience in developing Hadoop Architecture within the project in Windows and Linux platforms.
  • Strong Experience in writing Pig scripts and Hive Queries and Spark SQL queries to analyze large datasets and troubleshooting errors.
  • Well versed in Relational Database Design/Development with Database Mapping, PL/SQL Queries, Stored Procedures and Packages using Oracle, DB2, Teradata and MySQL Databases.
  • Excellent working experience on designing and implementing complete end-to-end Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie, Flume and zookeeper.
  • Have extensive knowledge and working experience on Software Development Life Cycle (SDLC), Service-Oriented architecture (SOA), Object Oriented Analysis and Design (OOAD), UML and J2EE Architecture.
  • Extensive Experience in Applications using AJAX, Object Oriented (OO) JavaScript, JSON, JSONP, and XML.
  • Proficiency in developing and deploying applications using Application Servers such as Weblogic, WebSphere, Apache Tomcat and JBoss.
  • Experience in working on SOAP and Restful Web Services.
  • Extensive knowledge of OOPS, OOAD, UML concepts (Use Cases, Class Diagrams, Sequence Diagrams, Deployment Diagrams etc).
  • Proficiency in using frameworks and tools like Struts, Ant, JUnit, WebSphere Studio Application Developer (WSAD5.1), Jbuilder, Eclipse, IBM Rapid Application Developer (RAD)
  • Expertise in designing and coding Stored Procedures, Triggers, Cursers and Functions using PL/SQL.
  • Expertise in developing XML documents with XSD validations, SAX, DOM, JAXP parsers to parse the data held in XML documents.
  • Good in writing ANT scripts for development and deployment purposes.
  • Experienced in GUI/IDE Tool using Eclipse, Jbuilder and WSAD5.0.
  • Expertise in using java performance tuning tools like JMeter and Jprofiler and LOG4J for logging.
  • Extensive Experience in using MVC (Model View Controller) architecture for developing applications using JSP, JavaBeans, Servlets.
  • Knowledge in implementing enterprise Web Services, SOA, UDDI, SOAP, JAX-RPC, XSD, WSDL and AXIS.
  • Good Working Knowledge of Ant & Maven for project build/test/deployment, Log4j for logging and JUnit for unit and integration testing.
  • Highly Self-motivated and goal oriented team player with strong analytical, debugging and problem solving skills, Strong in object oriented analysis and design.

TECHNICAL SKILLS:

Big Data/Hadoop Ecosystem: Hadoop2.7/2.5, MapReduce, Sqoop, Hive, Oozie, Pig, HDFS1.2.4, Zookeeper, Flume, Impala, Spark2.0/2.0.2, Storm, Hadoop (Cloudera), Hortonworks and Pivotal).

Web Technologies: HTML5, CSS3, JavaScript, JQuery, Bootstrap, AngularJS, JSF, AJAX, NodeJS, JSON Restful.

XML Technologies: XML, XSL, DOM, SAX, DTD, XQuery, XSD and XSLT.

Cloud Technologies: AWS EC2, S3, Dynamo DB, ELB

Databases: Oracle, SQL-Server, MySQL server, MS SQL, IBM DB2, MongoDB, NoSQL.

Web/Application Servers: Apache Tomcat, IBM WebSphere, Glassfish, Web logic Application server, JBOSS.

Build Tools: ANT, Maven, Gradle, Vagrant, Docker and Jenkins

IDE / Tools: Eclipse, IntelliJ, Spring Tool Suite(STS), RAD

Testing Tools/ Others: JUnit, Mockito, Soap, UI, Log4j and JIRA

Platforms: Windows 8/7, Linux and UNIX.

Methodologies: Agile, Waterfall, Test Driven Development

PROFESSIONAL EXPERIENCE:

Confidential , West Point, PA

Sr. Big Data Developer

Responsibilities:

  • As a Sr. Big Data Developer worked on Hadoop eco-systems including Hive, MongoDB, Zookeeper, Spark Streaming with MapR distribution.
  • Developed Big Data solutions focused on pattern matching and predictive modeling.
  • Implemented Security in Web Applications using Azure and deployed Web Applications toAzure.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop.
  • Participated in all aspects of Software Development Life Cycle (SDLC) and Production troubleshooting, Software testing using Standard Test Tool.
  • Involved in Agile methodologies, daily scrum meetings, spring planning.
  • Involved in writing Spark applications using Scala to perform various data cleansing, validation, transformation and summarization activities according to the requirement.
  • Loaded the data into Spark RDD and Perform in-memory data computation to generate the output as per the requirements.
  • Developed Spark code and Spark-SQL/Streaming for faster testing and processing of data.
  • Used Java Persistence API (JPA) framework for object relational mapping which is based on POJO Classes.
  • Involved in identifying job dependencies to design workflow for Oozie and YARN resourcemanagement.
  • Designed solution for various system components using Microsoft Azure.
  • Worked on data using Sqoop from HDFS to Relational Database Systems and vice-versa. Maintaining and troubleshooting
  • Explored with Spark to improve the performance and optimization of the existing algorithms in Hadoop using Spark context, Spark-SQL, Data Frame, pair RDD.
  • Created Hive Tables, loaded claims data from Oracle using Sqoop and loaded the processed data into target database.
  • Exported data from HDFS to RDBMS via Sqoop for Business Intelligence, visualization and user report generation.
  • Developed ApacheNifi flows dealing with various kinds of data formats such as XML, JSON, and Avro.
  • Worked on importing data from HDFS to MYSQL database and vice-versa using SQOOP.
  • Configured Hive meta store with MySQL, which stores the metadata for Hive tables.
  • Performed data analytics in Hive and then exported those metrics back to Oracle Databaseusing Sqoop.
  • Upgraded the Hadoop Cluster from CDH3 to CDH4, setting up High Availability Cluster and integrating Hive with existing applications.
  • Worked on NoSQL support enterprise production and loading data into HBase using Impalaand Sqoop.
  • Performed multiple MapReduce jobs in Pig and Hive for data cleaning and pre-processing.
  • Build Hadoop solutions for big data problems using MR1 and MR2 in YARN.
  • Handled importing of data from various data sources, performed transformations using Hive,PIG, and loaded data into HDFS.
  • Proactively involved in ongoing maintenance, support and improvements in Hadoop cluster.
  • Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
  • Used Hive to analyze data ingested into HBase by using Hive-HBase integration and compute various metrics for reporting on the dashboard.
  • Worked on analyzing, writing Hadoop MapReduce jobs using JavaAPI, Pig and hive.
  • Analyzed large amounts of data sets using HBase to aggregate and report on it.
  • Developed reports, dashboards using Tableau for quick reviews to be presented to business.
  • Worked on configuring and managing disaster recovery and backup on Cassandra Data.
  • Developed many distributed, transactional, portable applications using Enterprise JavaBeans (EJB) architecture for Java 2 Enterprise Edition (J2EE) platform.
  • Used Cloudera Manager for installation and management of Hadoop Cluster.
  • Developed data pipeline using Flume, Sqoop, Pig and Java MapReduce to ingest customer behavioral data and financial histories into HDFS for analysis.
  • Worked on MongoDB, HBase databases which differ from classic relational databases
  • Involved in converting HiveQL into Spark transformations using Spark RDD and through Scala programming.
  • Used Hive to perform data validation on the data ingested using Sqoop and cleansed the data.
  • Developed several business services using Java RESTful Web Services using Spring MVC framework.

Environment: Hadoop 3.0 , Oozie 4.3, Zookeeper 3.4, Cassandra 3.0, Sqoop 1.4, Apache NiFi 1.4, ETL, Azure, Hive 2.3, HBase 1.4, Pig 0.17, HDFS 3.1, Flume 1.8, Tableau, GIT, Kafka 1.1, MapReduce, JSON, AVRO, Teradata, Maven, SOAP.

Confidential, Wilmington, DE

Sr. Hadoop Developer

Responsibilities:

  • Key member in design and development of decision engine to publish best suitable offers for marketing purposes on Confidential 360 website.
  • Worked closely with data scientists for data gathering enterprise level to predict consumer behavior, such as what products user has bought and made recommendations based on recognizing patterns.
  • Involved in end-to-end data processing like ingestion, processing, quality checks, and splitting.
  • Refined terabytes of data from different sources and created hive tables.
  • Developed MapReduce jobs for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and HIVE from an Oracle database using Sqoop
  • Responsible to manage data coming from different sources.
  • Monitoring the running MapReduce jobs on the cluster using Control-M.
  • Responsible for loading data from UNIX file systems into HDFS.
  • Installed and configured Hive and wrote Hive UDF.
  • Wrote Pig scripts to process unstructured data and create structure data for use with Hive.
  • Developed a spark pipeline to transfer data from lake to Cassandra in cloud to make the data available for decision engine to publish customized offers real time.
  • Implemented a POC on deploying cloud instances for hadoop 2.0, setting up network and security settings, connecting to cloud instances, and setting up configuration settings across hadoop clusters using parallel shell utility.

Environment: Hadoop 0.20.2 - PIG, Hive, Spark, Cassandra, Scala, AWS, 50 Node cluster with Linux-Ubuntu.

Confidential, Newport Beach, CA

Java/Hadoop Developer

Responsibilities:

  • Involved in analyzing data coming from various sources and creating Meta-files and control files to ingest the data in to the Data Lake.
  • Involved in configuring batch job to perform ingestion of the source files in to the Data Lake.
  • Used Pig to do the transformation of the data that were in the HDFS to fit the requirements.
  • Developed the Java code to transform the incoming files to the required file formats.
  • Created several Pig UDFs for the enrichment engine those were used to perform enrichment on the data.
  • Worked extensively on Hive to create, alter and drop tables and involved in writing hive queries.
  • Created and altered HBase tables on top of data residing in Data Lake.
  • Extracted and updated the data into HDFS using Sqoop import and export command line utility interface.
  • Expertise in designing and deployment of Hadoop cluster and different Big Data analytic tools including Spark, Pig, Hive, HBase, Oozie, Zookeeper, SQOOP, Flume, Impala, Cassandra with Hortonworks Distribution.
  • Involved in creating Hive tables, loading with data and writing hive queries.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Importing and exporting Data from MySQL/Oracle to HiveQL Using SQOOP.
  • Designed and Developed Reference table engine frameworks on Talend using Hadoop tools such as HDFS, Hive, HBase MapReduce.
  • Hands on experience with HTML, CSS, Java Script, JAVA, and AJAX.
  • Extensively used Pig scripts for data cleansing and optimization.
  • Used Pig to do transformations, event joins, filter boot traffic and some pre-aggregations before storing the data onto HDFS.
  • Created custom UDF java routines in Talend using core java.
  • Worked on HBase API to connect to HBase and perform filters on tables in Talend.
  • Involved in the process of data acquisition, data pre-processing and data exploration of telecommunication project in Scala.
  • Used Angular JavaScript for client side validation.
  • Performed benchmarks on the Reference table engine with various sizes of data sets.
  • Worked closely with System Analyst and Architects to design and develop Talend jobs to fit the business requirement.
  • Involved in loading data from UNIX file system to HDFS.
  • Used Sqoop to import data from Oracle database to HDFS cluster using custom scripts.
  • Worked closely with scrum master and various scrum teams to gather information and perform daily activities.

Environment: Hadoop, Map Reduce, Yarn, Hive, Pig, HBase, Sqoop, MapReduce, Core Java, Eclipse, SQL Server, MYSQL, Linux

Confidential, Erlanger, KY

Sr. Java/J2EE Developer

Responsibilities:

  • Worked on High-level and low-level system design, planning, estimation and implementation.
  • Developed presentation layer using JSP, JSF, HTML, CSS and client validation using JavaScript, DOM.
  • Worked on High-level and low-level system design, planning, estimation and implementation.
  • Used two-way data binding features of AngularJS to implement Data binding to bind various server controls to the data from database.
  • Created Node.js express Server combined with Socket.io to build MVC framework from front-end side AngularJS to back-end MYSQL.
  • Implemented Multithreading, Concurrency, Exception Handling and Collections whenever necessary.
  • Developed the project using industry standard design patterns like Singleton, Business Delegate Factory Pattern for better maintenance of code and re-usability.
  • Developed Micro services and Restful web services using JAX-WS.
  • Worked with Spring IOC for implementing the future of DI across different layers and implemented AOP for logging and auditing.
  • Involved in writing JSP and JSF components. Used JSTL Tag library to create standard dynamic web pages.
  • Worked on XML and JSON for transferring and retrieving data between different applications.
  • Worked on generating the web services classes by using Service Oriented Architecture SOA, WSDL, UDDI and SOAP.
  • Used Web Services for creating rate summary and used WSDL and SOAP messages for getting useful plans from different module.
  • Designed and integrated the full-scale Spring/Hibernate persistence solution with the application architecture.
  • Implemented Object Relational Persistence (Hibernate) for the domain model.
  • Involved in design and development of the Hibernate Domain Model for the services.
  • Worked on web services and associated business modules integration.
  • Worked on development and implementation of the MVC Architectural pattern using Spring Framework including JSP.
  • Involved in identifying and implementation of different J2EE design patterns like Service Locator, Business Delegate, and DAO.
  • Worked in wide-area network to maintain all the device related data onto MySQL database.
  • Developed the DAO layer for the application using Spring Hibernate template support.
  • Used JMS API for communication between the components.
  • Involved in writing complex PL/SQL queries using joins, Stored Procedures, Functions, Triggers, cursors, and indexes in Data Access Layer.
  • Collaborated with testers and developers and prepared test plans for producing high quality software applications.
  • Worked with JUnit and build tools such as Maven. Worked on Log4j for logging errors, messages and performance logs and SVN for version control.
  • Deployed and tested the application in UNIX on IBM Web Logic Application Server.

Environment: JSP, JSF, HTML, CSS, PL/SQL, XML, JSON, Web Logic, JUnit, Maven, Log4j, MySQL, AngularJS, Spring IOC, JavaScript, Spring, Hibernate, JMS.

Confidential

Java/J2EE Developer

Responsibilities:

  • Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support in Waterfall methodology.
  • Designed and developed dynamic Web Pages using JSP, HTML, CSS, JavaScript and JQuery.
  • Implemented the Struts framework based on MVC design pattern and Session Façade Pattern using Session and Entity Beans.
  • Used Struts for web tier development and created Struts Action Controllers to handle the requests.
  • Involved in writing the struts-config files and implemented the Struts Tag library.
  • Responsible for designing, coding and developed the application in J2EE using Struts MVC.
  • Implemented Struts framework (Action & Controller classes) for dispatching request to appropriate classes.
  • Used simple Struts Validation for validation of user input as per the business logic and initial data loading.
  • Developed the Enterprise Java Beans (Stateless Session beans) to handle different transactions such as online funds transfer, bill payments to the service providers.
  • Implemented Service Oriented Architecture (SOA) using JMS for sending and receiving messages while creating web services
  • Developed XML documents and generated XSL files for Payment Transaction and Reserve Transaction systems.
  • Developed SQL queries and stored procedures.
  • Developed Web Services for data transfer from client to server and vice versa using Apache Axis, SOAP and WSDL.
  • Worked on XML technologies like XML Parsers, JAXB for binding data to java objects.
  • Used Java Messaging Services (JMS) for reliable and asynchronous communication.
  • Implemented the persistence layer using Hibernate and JDBC Template and developed the DAL (Data Access Layer) to store and retrieve data from the database.
  • Responsible to writing JDBC programming to persist the data in My SQL database.
  • Written some SQL Queries and PL/SQL procedures to fetch data from the database.
  • Implemented various J2EE Design patterns like Singleton, Service Locator, DAO, and SOA.
  • Worked on AJAX to develop an interactive Web Application and JavaScript for Data Validations.
  • Developed the application under JEE architecture, developed Designed dynamic and browser compatible user interfaces using JSP, Custom Tags, HTML, CSS, and JavaScript.
  • Deployed & maintained the JSP, Servlets components on Web logic 8.0
  • Developed Application Servers persistence layer using, JDBC, SQL, Hibernate.
  • Used JDBC to connect the web applications to Data Bases.
  • Implemented Test First unit testing framework driven using JUnit.
  • Developed and utilized J2EE Services and JMS components for messaging communication in Web Logic.
  • Configured development environment using Web logic application server for developer's integration testing.
  • Tested Service and data access tier using JUnit.
  • Used Web Logic for application deployment and Log 4J used for Logging/debugging.
  • Used CVS version controlling tool and project build tool using ANT.
  • Worked with production support team in debugging and fixing various production issues.

Environment: Java, J2EE, JSP, HTML, CSS, JavaScript, jQuery, Struts, Restful Services, SOAP, WSDL, Hibernate, JDBC, JMS, My SQL, CVS, ANT, Log4j and Web Logic.

We'd love your feedback!