Sr. Big Data Consultant Resume
New, YorK
SUMMARY:
- 10.5 years of profession experience as Software Engineer in Development of Business applications, big data applications and Product development.
- 3 plus years of experience in providing solutions for Big Data using HDFS, Map Reduce, Graph Databases,YARN, Spark, Kafka, PIG, Hive,Sqoop, Hbase, Zoo keeper, Oozie, Spark & HDP 2.x.
- Using MVEL framework, created dynamic transformation engine to work on huge volume of data.
- Expertise in NoSQL databases HBase performance tuning and resolving hot spot problems.
- Implemented spark graphx, to find the connected components in a graph.
- Good experience matching implementation and to search on huge data.
- Good hands on experience in implementing Elastic search for business & address search.
- Good experience in shell script, writing jobs to automate the scheduling process.
- Developed Apache Lucene indexing to perform fast searching on huge flat files.
- Implemented spark Hbase to scan and upload.
- Experience to execute KNN and GBDT models for prediction analysis.
- Good knowledge on AWS Neptune graph database.
- In depth knowledge of Spark concepts and experience with Spark in Data Transformation and Processing.
- Extensive hands on experience in writing complex Map reduce jobs and Hive data modeling.
- Implemented Partitioning, Dynamic Partitions and Bucketing in Hive for efficient data access.
- Worked on creating and optimizing Hive scripts for data analysts based on the requirements.
- Created Hive UDFs to encapsulate complex and reusable logic for the end users.
- Monitored workload, job performance and capacity planning using MAPR resource Manager.
- Coordinated with business customers to gather business requirements and interacted with other technical peers to derive Technical requirements.
- Good experience of Java web based applications that involve the technologies Restful Web services, Soap WebServices, JQuery, SPRING 4 - WEBFLOW MVC, ORM,HIBERNATE4,STRUTS1.x,SERVLETS,JSP,JDBC,JMS,JPA,XML,XSLT,SELENIUM,Arquillian,EJB3,DB Unit, JUnit, Maven.
- Experience using J2EE Design Patterns like Composite View, Value Object, DAO, MVC, Singleton, Facade, Decorator, Front Controller, MVC and Factory Pattern for reusing most efficient and effective strategies for new development.
- Extensive work experience using Web Services (REST, SOAP, WSDL) and Service Oriented Architecture (SOA).
- Expertise in developing MVC architecture for web based applications using Struts and Spring MVC frameworks.
- Hands on Experience in TDD (Test Driven Development).
TECHNICAL SKILLS:
Big data: Mapr 5.2, CDH 5.x,Hbase,Hive,Sqoop,Kafka,spark,Elastic search, Lucene,Zookeeper,MVEL & Map Reduce
Languages: Java,Scala, shell scripting,AWK,SQL,XML, Java script
No SQL: Hbase & Mongo DB.
Databases: Oracle,MySQL
Persistency: JDBC, JPA, Hibernate.
Methodologies: OOAD,MVC,Agile,TDD
Frameworks: Spark, Spring,EJB & struts.
Messaging: JMS,Weblogic MQ,Kafka
Information Exchange: XML,JAXB,JAXP, JSON
Web Services: SOAP, REST
Unit Testing: Junit
Business Experience: SCRUM, Test Driven Development (TDD)
Platforms: Linux, Windows
Application Server: Websphere, Tomcat, JBoss
Configuration Management: CVS, Subversion, Git
Tools: Eclipse, Jenkins, Sonar, Maven,TOAD,SQL developer
Database: Oracle and my SQL
Scripting: JavaScript, AJAX, JQuery, Node.js, Angular.js
Patterns: Enterprise Application Patterns, Core J2EE Patterns, GOF Patterns, Enterprise Integration Patterns.
WORK EXPERIENCE:
Confidential, New York
Sr. Big Data Consultant
Environment: MapR Hadoop, Map Reduce,Hive,HBase,Java,scala,oozie,sqoop,MySQL and JSON, Apache log4j, Maven,JUnit,JDBC and Eclipse Mars2
Roles and Responsibilities:
- Good exposure on graph database like AWS Neptune to find the relations between business and find connected components.
- Developed dynamic transformation engine to perform transformation logic on different source using MVEL transformation engine (rules engine).
- Developed spark application using Scala, to develop the connected components using sparkling graphx algorithm.
- Developed spark application to create row number for billions of records within 3 mins and provides hive table on top of data. Provided this as reusable component for the team.
- Worked on performance tuning of spark application, to resolve memory issues on huge data.
- Worked with the Spark for improving performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark-SQL, Spark MLlib, Data Frame, Pair RDD's, Spark YARN.
- Developed many hive UDFs as reusable components for business team for analysis and data transformation.
- Created hive tables on huge data using hive orc and compression techniques.
- Have worked on large volumes of structured, unstructured and semi structured data and have developed many MapReduce for data cleansing. i.e., filtering and projection of data.
- Hands on experience on Hue to import data on to the Graphical User Interface
- Executed different performance tuning techniques on HBase schema creation, region splitting and row key design.
- Implemented the index for fast and effective search for Business and address matching in Elastic search.
- Developed Apache Lucene indexing to perform fast searching on huge flat files.
- Extracted data from Hbase and stored as external tables in Hive using serde, to perform Data analytics.
- Written complex map reduce programs to perform data transformation.
- Written map reduce programs to perform bulk loading into HBase.
- Scheduled jobs using Maverick work flows.
- Created Hive external, internal tables and Implemented Partitioning, Dynamic Partitions and Bucketing in Hive for efficient data access.
Confidential
Big Data Developer/Tech Lead
Environment: Hadoop, Map Reduce, Hive, HBase, Java, scala, oozie, sqoop, MySQL and JSON, Apache log4j, Maven, JUnit, JDBC and Eclipse Mars2.
Roles & Responsibilities:
- Developed Map Reduce Input format to read specific data format
- Developing and maintaining Workflow Scheduling Jobs in Oozie.
- Developed efficient Map Reduce programs in java for filtering out the unstructured data.
- Developed Pig Latin Script for the log data and stored the cleansed data in the Apache Hadoop.
- Involved with the client and gathered the business requirements
- Developing Hive queries and UDF's as per requirement.
- Developed MapReduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis
- Implemented Partitioning, Bucketing in Hive for better organization of the data.
Confidential
Sr. Software Engineer
Environment: Java1.5, JSP, Spring-Web Flow, Hibernate, Tiles, JUnit, Maven, Oracle, Tomcat, soap Web services, log4j, UNIX, Oracle, Windows, Restful web services, UNIX.
Roles & Responsibilities:
- Involved in all phases of application (Analysis, design, development, unit testing, implementation, deployment and prod support).
- Analyzed the impact of new stories (Agile).
- Attended the grooming session with business analysts to gather/clarification the requirements.
- Automated the manual work with daily/monthly running jobs.
- Created web services for both provider and consumer for integration layer.
- Implementation of creating JPA entity for data layer.
- Implementation of creating Entity Beans (EJB) for service layer.
- Implementation of JSP for UI layer.
- Deployed application for DEV,DIT and SIT environments in both UNIX and Windows environments
- Handled daily scrum meeting, providing the status to client manager.
- Provided testing support (debug and fix) in testing phase.
- Providing technical assistance and lead team members to deliver the user stories in-time.
- Used spring-web services module to consume web services.
- Implemented Maven plugins to generate objects from WSDL and for project build.
- Implemented Log4j for development and run time debugging of the application.
- Implemented JUnit for test driven development and perform unit testing.
- Implemented selenium for front end unit testing.
- Implemented exposing web services using Restful API using spring integration
Confidential
Software Engineer
Environment: Java1.5, JSP, Java script, Struts, Hibernate, Tiles, Oracle, Tomcat, log4j, Oracle, Windows, UNIX, RESTful web services.
Responsibilities:
- Involved in the development of Action classes, Hibernate mapping files, HQL queries and back end validations.
- Involved in the development of user interface using JSP (Java Server Pages), HTML and JavaScript used for front end validation of the application.
- Preparing the estimates for CR’s and tickets.
- Involved in resolving critical, medium CR’s and developed the Enhancements.
- Managing the Build and Deployment process Activities.
- Involved in preparing process documents like SPP, SCMP, RMP and Training Plan.
- Involved in fixing the production issues.
Confidential
Software Engineer
Environment: Java1.5, JSP, Java script, Struts, Hibernate, Tiles.
Roles & Responsibilities:
- Involved in the Development of Document Search & Operational Reports Modules.
- Involved in the Development of Action Classes, Action Forms & Validations.
- Involved in the development of the Business Components.
- Involved in the preparing of UTCs.
Confidential
Software EngineerEnvironment: Flex, Action Script, Tomcat, Oracle, Eclipse 3.2
Roles & Responsibilities:
- Involved in the development of front end using Flex.
- Involved in the development validations using Action Script.
- Involved in the development of Business Components.
Confidential
Software Engineer
Environment: Springs-MVC, Hibernate, JSP, Oracle, Java Script, Log4J and Tomcat.
Roles & Responsibilities:
- Involved in the development of trade management and Cash& FX modules.
- Involved in the development of front end using JSPs.
- Involved in the development of Controller and Validator Classes Using Spring MVC framework.
- Involved in the development of the Business Components using Hibernate.
- Involved in preparing UTP’s.