Sr. Java / Big Data Developer Resume
Arlington, VA
SUMMARY:
- Around 9+ years of experience in SDLC with key emphasis on the trending Big Data Technologies expertise in development of Hadoop/Big Data, web based technologies with different back end databases.
- Experience in Cloudera, HortonWorks, MapR and Amazon Web Services distributions of Hadoop.
- Experience in working with java for writing custom UDFs to extend Hive and Pig core functionality.
- Strong expertise in implementing End - to-End Big Data projects in Hadoop Ecosystem - Hortonworks Data Platform, Cloudera distribution, HDFS, Map-Reduce, Hive, Pig, Kafka, Spark, HBase, Oozie and EMR.
- Experience in using Hadoop ecosystem components like Map Reduce, HDFS, HBase, Zoo Keeper, Hive, Sqoop, Pig, Flume, Spark, Cloudera.
- Expertise in various Java/J2EE technologies like JSP, Servlets, Hibernate, Struts, spring.
- Strong expertise on Amazon AWS EC2, S3, Kinesis and other services
- Experience includes Requirements Gathering, Design, Development, Integration, Documentation, Testing and Build.
- Experience in working with Map Reduce programs, Pig scripts and Hive commands to deliver the best results.
- Diverse experience utilizing Java tools in business, Web, and client-server environments including Java Platform, Enterprise Edition (Java EE), Enterprise Java Bean (EJB), Java Server Pages (JSP), Java Servlets (including JNDI), Struts, and Java database Connectivity (JDBC) technologies.
- Hands on experience in Core Java, Servlets, JSP, JDBC, C#, JavaScript.
- Knowledge of Eclipse tool for the development of Java projects.
- Proficient in developing applications using Java/J2EE design patterns and industry's best design practices.
- Good middleware skills in J2EE, web services with application servers - Tomcat web server, BEA WebLogic, IBM WebSphere, JBoss with experience on heterogeneous operating systems.
- Extensive experiences in Log4j for creating logs of different categories.
- Good knowledge with web-based UI development using jQuery UI, jQuery, ExtJS, CSS3, HTML, HTML5, XHTML and JavaScript.
- Experience with unit testing, functional Testing, system Testing, Integration testing the applications using JUnit, Mockito, Jasmine and Cucumber, PowerMock & EasyMock.
- Knowledge in configuration and managing - Cloudera’s Hadoop platform along with CDH3&4 clusters.
- Knowledge and experience of architecture and functionality of NOSQL DB like Cassandra and Mongo DB.
- Experienced in developing web based GUIs using JavaScript, JSP, HTML, JQuery, XMLand CSS.
- Experienced in Collected logs data from various sources and integrated in to HDFS using Flume and experience in developing custom UDFs for Hive.
- Experienced in testing data in HDFS and Hive for each transaction of data.
- Experienced in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Excellent understanding of Hadoop architecture and underlying framework including storage management.
- Experienced in Worked on NoSQL databases - Hbase, Cassandra & MongoDB, database performance tuning & data modeling.
- Experience in using PL/SQL to write Stored Procedures, Functions and Triggers.
- Excellent technical and analytical skills with clear understanding of design goals of ER modeling for OLTP and dimension modeling for OLAP.
TECHNICAL SKILLS:
Hadoop/Big Data: MapReduce, HDFS, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, Flume, Scala, Akka, Kafka, Storm, MongoDB.
Languages: Java, J2EE, PL/SQL, Pig Latin, HQL, R, Python, XPath, Spark
Databases: Oracle 12c/11g/10g/9i, Microsoft Access, MS SQL
No SQL Databases: Cassandra, mongo DB
Web Technologies: HTML, DHTML, XML, XHTML, JavaScript, CSS, XSLT, EME AWS.
Web/Application servers: Apache Tomcat6.0/7.0/8.0, JBoss
Frameworks: MVC, Struts, Spring, Hibernate.
Operating Systems: UNIX, Ubuntu Linux and Windows, Centos, Sun Solaris.
Network protocols: TCP/IP fundamentals, LAN and WAN.
Java/J2EE Technologies: JDBC, Java Script, JSP, Servlets, JQuery
PROFESSIONAL EXPERIENCE:
Confidential,Bronx,NY
Sr. BigData / Hadoop Architect
Responsibilities:- Implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies such as Hadoop, Map Reduce Frameworks, HBase, Hive.
- Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including Pig, Hive, Sqoop, Oozie and Zookeeper.
- Used Sqoop to efficiently transfer data between databases and HDFS and used Flume to stream the log data from servers.
- Involved in loading and transforming large sets of structured, semi structured and unstructured data from relational databases into HDFS using Sqoop imports.
- Implement enterprise grade platform(mark logic) for ETL from mainframe to NOSQL(cassandra).
- Experience on BI reporting with At Scale OLAP for Big Data.
- Responsible for importing log files from various sources into HDFS using Flume
- Worked on tools Flume, Storm and Spark.
- Analyzed data using HiveQL to generate payer by reports for transmission to payer's form payment summaries.
- Design AWS architecture, Cloud migration, AWS EMR, Dynamo DB, Redshift and event processing using lambda function
- Imported millions of structured data from relational databases using Sqoop import to process using Spark and stored the data into HDFS in CSV format.
- Developed Spark streaming application to pull data from cloud to Hive table.
- Used Spark SQL to process the huge amount of structured data.
- Assigned name to each of the columns using case class option in Scala.
- Implemented Spark GraphX application to analyze guest behavior for data science segments.
- Utilize AWS services with focus on big data Architect /analytics / enterprise data warehouse and business intelligence solutions to ensure optimal architecture, scalability, flexibility, availability, performance, and to provide meaningful and valuable information for better decision-making.
- Enhancements to traditional data warehouse based on STAR schema, update data models, perform Data Analytics and Reporting using Tableau.
- Experienced in analyzing and Optimizing RDD's by controlling partitions for the given data.
- Struts (MVC) is used for implementation of business model logic.
- Worked with Struts MVC objects like Action Servlet, Controllers, validators, Web Application Context, Handler Mapping, Message Resource Bundles, and Form Controller.
- Expert in performing business analytical scripts using Hive SQL.
- Implemented continuous integration & deployment (CICD) through Jenkins for Hadoop jobs.
- Worked in writing Hadoop Jobs for analyzing data using Hive, Pig accessing Text format files, sequence files, Parquet files.
- Experience in different Hadoop distributions like Cloudera (CDH3 & CDH4) and Horton Works Distributions (HDP) and MapR.
- Designed and developed UI screens using Struts 2, DOJO, JavaScript, JSP, HTML, DOM, CSS, and AJAX.
- Implemented Database Level Sorting in Entire Application using DOJO Ajax Call.
- Implemented Dojo and JQuery JavaScript frame work in UI Layer.
- Developed proto type for Big Data analysis using Spark, RDD, Data Frames and Hadoop eco system with .csv, Json, parquet and hdfs files.
Environment: Big Data, SparkSpark, YARN, HIVE, Pig, Scala, Python, Hadoop, AWS, Dynamo DB, Kibana, Cloudera, EMR, JDBC, Redshift, NOSQL, Sqoop, MYSQL.
Confidential,Omaha,NB
Sr. BigData / Hadoop Developer
Responsibilities:- Worked as Big Data Developer in the team dealing with Firm's proprietary platform issues. Providing data analysis for the team as well as developing enhancements.
- Provisioning, installing, configuring, monitoring, and maintaining HDFS, Yarn, HBase, Flume, Sqoop, Oozie, Pig, Hive.
- Implemented in setting up standards and processes for Hadoop based application design and implementation.
- Created Hive External tables to stage data and then move the data from Staging to main tables .
- Responsible for creating Hive tables, loading the structured data resulted from MapReduce jobs into the tables and writing hive queries to further analyze the logs to identify issues and behavioral patterns.
- Working on extracting files from MongoDB through Sqoop and placed in HDFS and processed.
- All the data was loaded from our relational DBs to HIVE using Sqoop. We were getting four flat files from different vendors. These were all in different formats e.g. text, EDI and XML formats.
- Involved in developing Spark code using Scala and Spark-SQL for faster testing and processing of data and exploring of optimizing it using Spark Context, Spark-SQL, Pair RDD's, Spark YARN.
- Wrote programs in Scala using Spark and worked on migrating MapReduce programs into Spark using Scala
- Responsible for creation of Source to Target mapping document from source fields to destination fields mapping.
- Developed a shell script to create staging, landing and Semantic tables with the same schema like the source
- Developed HiveQL scripts for performing transformation logic and also loading the data from staging zone to landing zone and Semantic zone.
- Responsible for Debug, Optimization of Hive Scripts.
- Involved in installing EMR clusters on AWS.
- Swings being used to customized the specific tools for generic purpose.
- Documented application for its functionality and its enhanced features
- IRIS - Implemented Alteryx workflow to load IRIS data from Cassandra DB in MS Azure to our analytics cluster & used SPARK scripts for data clean up & formatting. Built Hive scripts to prepare data for analytics.
- Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.
- Develop DAO design pattern for hiding the access to data source objects.
- Worked directly with product owners to get requirement and implemented it.
- Actively participated in planning sessions for the Sprints, effort estimations, Backlog refinements and dividing features into User Stories and Tasks.
- Involve in implementation of REST and SOAP based web services.
- Worked on improvising the performance of the application.
- Developer in Big Data team, worked with Hadoop AWS cloud, and its ecosystem.
- Worked on writing Perl scripts covering data feed handling, implementingmark logic, communicating with web-services through SOAP Lite module and WSDL.
- Used Hive to analyze data ingested into HBase by using Hive-HBase integration and compute various metrics for reporting on the dashboard
- Involved in developing Map-reduce framework, writing queries scheduling map-reduce.
Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Cloudera,Pig, Sqoop, Kafka, Spark, Apache Cassandra, Oozie, Impala, Cloudera, Zookeeper, MySQL, Eclipse, PL/SQL and Python, AWS EMR.
Confidential,Arlington,VA
Sr. Java / Big Data Developer
Responsibilities:- Involved in Big Data Project Implementation and Support.
- Involved in the coding and integration of several business critical modules of CARE application using Java, spring, Hibernate and REST web services on Web Sphere application server.
- Deliver Big Data Products including re-platforming Legacy Global Risk Management System with Big Data Technologies such as Hadoop, Hive and HBase.
- Performed data extraction and migration, data cleaning, analysis, and visualization using SSIS, SSAS, Tableau, and Power BI to support OneCloud Data warehousing solution on AWS.
- Earned a reputation for designing AWS cloud deployment strategies that mitigate risk while meeting infrastructure, employee, customer and budgetary needs.
- Proficient in TDD (Test Driven Development) by writing test cases using Jasmine, J-Unit Framework Working as a Full Stack developer using JavaScript frameworks Angular.JS and Node.JS.
- Responsible for creating efficient design and developing User Interaction screens using HTML5, CSS3, JavaScript, JQuery, AJAX, Angular JS , and JSON used jQuery, Ajax for service calls on pages to interact with the servers.
- Implemented XML-based communications protocol using JAXB.
- Used Maven as the build tool, GIT for version control, Jenkins for Continuous Integration and JIRA as a defect tracking tool.
- Involved in using CVS, Git, for project management and version management.
- Extensively involved in the development of backend logic or data access logic using Hibernate and thus creating the object relational mapping with the Java beans in the application.
- Worked on performance tuning of Web Agents, Policy Servers, Policy Stores and User Stores to meet and maintain operational requirements (process, thread, connection, cache).
- Designed various tables required for the project in Oracle 11g database, Oracle web centre and used the stored procedures in the application and Wrote Stored Procedures, Triggers and Functions on SQL Server as well as PL/SQL on Oracle to support both databases for the application.
- Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
- Implemented partitioning, dynamic partitions and buckets in HIVE.
- Developed customized classes for serialization and Deserialization in Hadoop
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Implementing Spring MVC and IOC methodologies.
- Used the JNDI for Naming and directory services.
- Involved in the coding and integration of several business critical modules of application using Java, Spring, Hibernate and REST web services on WebSphere application server.
- Experience working with big data and real time/near real time analytics and big data platforms like Hadoop, Spark using programming languages like Scala and Java.
- Wrote programs in Scala using Spark and worked on migrating MapReduce programs into Spark using Scala
- Responsible for creation of Source to Target mapping document from source fields to destination fields mapping.
- Developed a shell script to create staging, landing and Semantic tables with the same schema like the source.
Environment: Java, J2EE, Java SE 6, UML, JSP 2.1, Hadoop 1x, Hive, Pig, HBASE, JSTL 1.2, Servlets 2.5, Spring MVC, Hibernate, JSON, Restful Web services, jQuery, AJAX, Angular Js, JAXB, IRAD Web sphere Integration Developer, Web Sphere 7.0, Eclipse Kepler-Maven, Serena Dimensions, Unix, JUnit, DB2, Oracle.
Confidential
Sr. Java / J2EE Developer
Responsibilities:
- Developed REST services to talk with adapter classes and exposed them to the angular JS front-end. Worked on Restful web services which enforced a stateless client server and support JSON few changes from SOAP to RESTFUL Technology Involved in detailed analysis based on the requirement documents
- Implemented Object-relation mapping in the persistence layer using hibernate frame work in conjunction with Spring Aspect Oriented Programming (AOP) functionality.
- Developed application framework using struts with design principles of the J2EE using Business Delegate, Service Locator, Session Facade, Domain object and DAO patterns and developed Stateless Session Bean to Achieve Session façade Design Pattern.
- Developed Stored Procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
- Developed SQL queries and Stored Procedures using PL/SQL to retrieve and insert into multiple database schemas.
- Developed the XML Schema and Web services for the data maintenance and structures Wrote test cases in JUnit for unit testing of classes.
- Used DOM and DOM Functions using Firefox and IE Developer Tool bar for IE.
- Used JSP, HTML, Java Script, Angular JS and CSS3 for content layout and presentation.
- Did core Java coding using JDK 1.3, Eclipse Integrated Development Environment (IDE), clear case, and ANT.
- Used Spring Core and Spring-web framework. Created a lot of classes for backend.
- Developing User Interface Screens using Spring MVC, to enable customers obtain auto finance. Extensive experience in developing various web based applications using Hibernate 3.0 and spring frameworks.
- Developed Spring REST Exception Mappers.
- Developed Stored Procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
- Responsible for deployment of application in the Integration/Functional Environment, providing necessary assistance to UA Ttester.
Environment: Java, XML, HTML, JavaScript, JDBC, CSS, SQL, PL/SQL, XML, Web MVC, Eclipse, Ajax, JQuery, spring with Hibernate, Active MQ, Jasper Reports, Ant as build tool and My SQL and Apache Tomcat.