Big Data Developer Resume
CA
PROFESSIONAL SUMMARY:
- Over 8 years of IT industry experience with strong emphasis on Big Data/Hadoop, Apache Spark, Java/J2EE, Scala and Python.
- About 4.5 years of work experience in ingestion, storage, querying, processing and analysis of Big Data with hands on experience in Hadoop Ecosystem development including MapReduce, HDFS, Hive, Pig, Spark, SparkHql, Spark Streaming, YARN, Kafka, HBase, MongoDB, Cassandra, ZooKeeper, Sqoop, Flume and Oozie
- Experienced with Apache Spark by improving the performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark - SQL, Data Frame, Pair RDD's, Spark YARN.
- Experience and strong knowledge on implementation of SPARK core - Spark Streaming, Spark SQL, MLLib.
- Hands on experience in developing the Enterprise and Web applications on n-tier Architecture using Java/J2EE and java based technologies such as Servlets, JSP, Hibernate, Spring, EJBs, JNDI, Web services, Java Mail, XML, JPA, JMS and JDBC.
- Strong working experience in Software Development Life Cycle (SDLC) processes like Waterfall and Agile SCRUM Methodologies.
- Setting up, configuring, and programming on a Hadoop Framework with Strong knowledge in NoSQL databases like MongoDB, HBase, and Cassandra.
- Automated the process for extraction of data from warehouses and weblogs by developing work-flows and coordinator jobs in OOZIE.
- Good experience in implementing various J2EE design patterns like Singleton, Factory, Chain of responsibility, Session Facade, Command, Front Controller DTO and DAO.
- Experience in implementing Web Services based in Service Oriented Architecture (SOA) using SOAP, RESTful Web Services, JAX-WS, UDDI, WSDL.
- Experience in Object Relational Mapping Frameworks such as Hibernate.
- Integrated Spring DAO for data access using Hibernate, used HQL and SQL for querying databases.
- Expertise in Web Front End technologies Tag Libraries and Struts custom validation rules, Tiles, JSTL, Java Script, DHTML and XML/XSL.
- Generated Scripts for effective integration of JIRA applications with other tools.
- Good knowledge on creating the queries using RDBMS.
- Experience with Java compliant IDE’s like Eclipse, JDeveloper and NetBeans.
- Hands on experience in deploying applications using Application/Web servers like Jboss, Apache Tomcat and IBM Web Sphere.
- Developed testing plans and procedures for unit test, system test, acceptance test and automated tests.
- Experience in configuring build tools like Maven and ANT for the development and deployment with Jenkins.
- Experience with unit testing using JUnit.
- Experience in using version control tools like Rational Clear Case, Tortoise SVN, CVS (Concurrent Version System).
- Diligent and highly enthusiastic professional with strong technical, analytical and communicational skills.
- Excellent analytical and critical thinking, creative problem solving, quick learning, interpersonal skills.
TECHNICAL SKILLS:
Languages: Java, C, C++
BigData Technologies: Hadoop, HDFS, MapReduce, Hive, Pig, HBase, Impala, Sqoop, Oozie, ZooKeeper, Spark,Cassandra, Talend, Git, Maven
RDBMS: Oracle, MySQL, SQL Server.
No SQL: HBase, Cassandra
Scripting & Query Languages: Python, Shell, SQL & PL/SQL.
Web/Application Servers: IBM Web sphere, Tomcat and LDAP.
Middleware: RMI, EJB, JMS, Hibernate
Technologies: J2EE, JDBC, Multi-threading, JSP, Servlets, Struts, JSF, AJAX, SOAP, XSLT, DOM, CSS, DTD and Schema
PROFESSIONAL EXPERIENCE:
Confidential, CA
Big data Developer
Responsibilities:
- Expertise in designing and deployment of Hadoop cluster and different Big Data analytic tools including Pig, Hive, HBase, Oozie,Sqoop, flume, Apache Spark with Hortonworks Distribution.
- Involved in the high-level design of the Hadoop architecture for the existing data structure and Business process
- Migrated data from Teradata tables to Hadoop environment using Talend and Pig scripts.
- Created Hive tables with partitions and bucketing to store incremental data.
- Involved in code promotion and migration into various environments Non-Prod, Pre-Prod and Prod.
- Designed the code promotion flow from the Development environment to the Production via Quality Analysis environment Real time streaming of data using Spark with Kafka.
- Experienced with batch processing of data sources using Apache Spark and Elastic search.
- Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
- Written a plugin for Hudson integration with VersionOne as per business requirement.
- Conducted introductory and hands-on sessions of Hadoop HDFS architecture, Hive, Talend, Pig for other teams.
- Used Talend as a ETL tool to transform and load the data from different databases.
- Involved in importing the real time data to Hadoop using Kafka and implemented the Oozie job for daily imports.
- Automated the process for extraction of data from warehouses and weblogs by developing work-flows and coordinator jobs in Oozie.
- Optimized Map Reduce code, pig scripts and performance tuning and analysis.
- Designed the code promotion flow from the Development environment to the Production via Quality Analysis environment.
- Written QA test cases as a part of Agile unit testing process for validation of data.
- Developed the UI design and its connection to the Integration and deployment tools in Java using spring framework.
- Provided Technical support for Development environmentfor the automation process resolving the issues, analyzing the defects, providing solution for defects. Resolved more priority defects as per the schedule.
Environment: Apache Hadoop, HDFS, Hive, Pig, Pyspark, scala, Spark Streaming, Spark SQL, HBase, Kafka, Sqoop, Talend, Java, Scala, Git, Shell Scripting, Scala IDE.
Confidential, Quincy, MA
Sr. Hadoop Developer
Responsibilities:
- Analyzed large and critical datasets using HDFS, Map Reduce, Hive, Pig.
- Evaluated suitability of Hadoop and its ecosystem to the above project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
- Estimated the Software & Hardware requirements for the Namenode and Datanodes& planning the cluster.
- Extensive work with using log files and to copy them into HDFS using flume.
- Developed simple and complex Map Reduce programs in Java for Data Analysis on different data formats.
- Wrote extensive Map reduce jobs in java to train the cluster and developed Java map reduceprograms for the analysis of sample log files stored in cluster.
- Developed Pig scripts for data analysis and extended its functionality by developing custom UDF's written in Java or Python.
- Created an e-mail notification service upon completion of job for the particular team which requested for the data.
- Defined job work flows as per their dependencies in Oozie.
- Maintain System integrity of all sub-components related to Hadoop.
- Developed Shell and Python scripts to automate and provide Control flow to Pig scripts.
- Worked on designing NoSQL Schemas on HBase.
- Used Talend as a ETL tool to transform and load the data from different databases.
Environment: Hortonworks Hadoop, HDFS, Pig, Hive, Kafka, Python, HBase, Zookeeper, MapReduce, Java, Sqoop, Talend, Linux, UNIX Shell Scripting.
Confidential, Erie PA
Sr. Hadoop/ Java Developer
Responsibilities:
- Developing Managed, external and partition tables as per the requirement.
- Ingested structured data into appropriate schemas and tables to support the rule and analytics.
- Developing custom User Defined Functions (UDF’s) in Hive to transform the large volumes of data with respect to business requirement.
- Developing Pig Scripts, Pig UDF’s and Hive Scripts, Hive UDF’s to load data files.
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in loading data from edge node to HDFS using shell scripting
- Implemented scripts for loading data from UNIX file system to HDFS.
- Load and transform large sets of structured, semi structured and unstructured data.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Automated work flow using shell scripts.
- Participated in SDLC Requirements gathering, Analysis, Design, Development and Testing of application developed using AGILE methodology.
- Actively participated in Object Oriented Analysis Design sessions of the Project, which is based on MVC Architecture using Spring Framework.
- Participated in Daily Scrum meetings, Sprint planning and estimation of the tasks for the user stories, participated in retrospective and presenting Demo at end of the sprint.
- Developed the presentation layer using HTML, CSS, JSPs, BootStrap, and AngularJS.
- Adopted J2EE design patterns like DTO, DAO, Command and Singleton.
- Used SpringIOC/ORM, AOP and Spring Security.
- Published Web services - WSDL and SOAP for getting required information from the PostgreSQL.
- Implemented Object-relation mapping in the persistence layer using hibernate framework in conjunction with spring functionality.
- Generated POJO classes to map to the database table.
- Configured Hibernate’s second level cache using EHCache to reduce the number of hits to the configuration table data.
- ORM tool Hibernate to represent entities and fetching strategies for optimization.
- Implementing the transaction management in the application by applying Spring Transaction and Spring AOP methodologies.
- Written SQL queries and stored procedures for the application to communicate with Database (PostgreSQL).
- Used Junitframework for unit testing of application.
- Used Maven to build and deploy the application.
Environment: HDFS, Pig, Hive, UNIX Shell Scripting, JDK 1.7, PostgreSQL, BootStrap, AngularJS, JSTL, Hibernate 4.0, spring, JUnit, Maven, Spring IOC, Spring Security, Spring IOC, SOAP, Restful Web Services, HTML, JQuery, SVN, and WebLogic.
Confidential, Bloomington IL
Sr. Java Developer
Responsibilities:
- Used Scrum (Agile) as Development methodology
- Participated in all the phases of SDLC including Requirements Collection, Design & Analysis of the Customer Specifications, Development and Customization of the Application.
- Used JavaScript for client-side validations and Created UI with AJAX, HTML, CSS&JavaScript (JQuery) with AngularJS.
- Used Spring Model View Controller Framework (MVC) and Achieved Dependency Injection using Spring IOC
- Developed Spring DAOs (Data Access Object) using JDBC to interact with DBMS-Oracle 10g.
- Used Hibernate and Spring ORM for ORM Mapping and created mapping classes required.
- Created the Hibernate POJO Objects and mapped using Hibernate Annotations.
- Web Services implemented under RESTful Architecture and SOAP.
- Back end coding and development using Java Collections including Set, List, Exception Handling etc.
- Design Patterns like Facade, Singleton and Factory Pattern were used.
- Design and implementation of business logic and data processing routes using Apache Camel.
- Deployed the java application using Maven build script.
- Implemented test cases using JUNIT andUsed Log4j to implement logging facilities.
- Used CVS for version control.
Environment: JAVA, J2EE, Maven, JSTL, Spring MVC,Spring IOC, Hibernate Framework, Oracle Web logic Server, Oracle DB, Servlets, XML,ApacheCamel,JUnit, Ajax, AngularJS, RESTful Web Services, SOAP, PL/SQL, JSP, XSL, Java Script, HTML, CSS, Eclipse.
Confidential
Java Developer
Responsibilities:
- Dynamic web site design and programming using HTML, CSS, JavaScript, jQuery.
- Developed jQuery based calendar to schedule appointments.
- Responsible for creating efficient design and developing User Interaction screens using HTML, CSS, JavaScript, jQuery AJAX and JSON.
- Developed interface to display report data using jQuery, AJAX, JavaScript and JSON, XML
- Used various jQuery plugins like auto complete, date picker, color picker, document slider to enhance the usability, functionality and overall experience for website
- Implemented pagination and sorting functionalities in different modules using jQuery plugins
- Designed the front-end applications, user interactive (UI) web pages using web technologies like HTML, XHTML, and CSS.
- Using advanced level of JQUERY, AJAX, JavaScript, CSS and pure CSS layouts.
- Used Firebug and IE Developer Toolbar for debugging and browser compatibility, cross-browser / platform testing and compatibility assurance
- Using SVN for version control.
- Participated in requirement gathering, coordinated with various development teams in code integration.
Environment: HTML, CSS, JavaScript, jQuery, JSON, AJAX, XML, Adobe Dreamweaver, Eclipse, Apache HTTP Server, MySQL.