Sr. Hadoop Developer Resume
New York City, NY
SUMMARY:
- Having 8+ years of experience in various IT sectors such as banking, health - care, and financial services, which includes good hands-on experience in Big Data technologies.
- Having 3+ years of experience as a Hadoop Developer in all phases of Hadoop and HDFS development.
- Hands on experience with HDFS, MapReduce and Hadoop Ecosystem (Pig, Hive, Oozie, Hbase, Zookeeper and Sqoop).
- Expertise in developing and implementing MapReduce jobs using Hadoop to work with Big Data.
- Expertise with Spark processing Framework such as Spark and Spark Sql.
- Expertise in NoSQL databases like HBase, MongoDB
- Good knowledge in cleansing and analyzing data using HiveQL, Pig Latin, and custom MapReduce programs in Java.
- Expertise in writing custom UDFs and UDAFs for extending Hive and Pig core functionalities.
- Expertise in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS), Teradata and vice versa.
- Expertise in creating workflows using Oozie for cron jobs.
- Worked extensively with Dimensional modeling, Data migration, Data cleansing, Data profiling, and ETL Processes features for data warehouses.
- Experience in fetching and loading data in Cassandra.
- Experience on AWS EC2, EMR, LAMBDA and Cloud Watch
- Equipped with good knowledge on Parsing, Indexing, and Searching concepts of Splunk.
- Expertise working with Distributions such as MAPR, Horton works and Cloudera.
- Having good knowledge on Hadoop Administration like Cluster configuration, Single Node Configuration, Multi Node Configuration, Data Node Commissioning and Decommissioning, Name Node Backup and Recovery, HBase, HDFS and Hive Configuration, Monitoring clusters, Access control List
- Expertise in all stages of SDLC (Agile, Waterfall), writing Technical Design document, Development, Testing and Implementation of Enterprise level Data mart and Data warehouses.
- Extensive experience in design & development of internet/intranet applications using Java, J2EE, JSP, Servlets, JMS, JDBC, EJB, HTML/DHTML, XML, JSON, jQuery, AJAX, CSS, JavaScript, Struts, Spring, JPA, Hibernate.
- Good Knowledge of different J2EE Design Patterns like MVC, Data Access Object, Data Transfer Object / Value Object, Business Delegate and Singleton.
- Strong Experience in Oracle 9i/10g including PL/SQL, SQL, Stored Procedures, Triggers, functions, packages and databases like DB2, MySQL and SQL Server.
- Ability to handle multiple tasks and to work independently as well as in a team, experienced in interacting with Business/Technology groups.
- Versatile team player and a quick learner with strong written, oral and inter-personal communication skills.
TECHNICAL SKILLS:
Languages: C, C++, Java, xml, spring, Javascript, Jquery, Webservices.
Technologies: Hadoop Ecosystem, Spring Boot, Microservices, AWS, J2SE, J2EE, ORACLE
Application/Web Servers: IBM Websphere, Oracle/BEA Weblogic, JBoss, Tomcat
Big Data Ecosystem: HDFS, Map Reduce, Oozie, Hive/Impala, Pig, Sqoop, Zookeeper and Hbase,Spark, Scala
Methodologies: Agile, Waterfall.
Databases/API: DB2, Oracle 8.0,9i,10g, JPA, Hibernate, MS Access, MySQL and SQL Server, SQL, PL SQL, Developer, Control Center.
Operating Systems: Windows 2003/2000/NT/XP, UNIX, LINUX
Data Modeling: Microsoft Visio, Rational Rose and UML Design
Version Control Systems: GIT, CVS,SVN
PROFESSIONAL EXPERIENCE:
Confidential, New York City, NY
Sr. Hadoop Developer
Responsibilities:- Followed AGILE Methodology and participated in SCRUM Meetings.
- Responsible for Installation and configuration of Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Developed Sqoop scripts to import export data from relational sources and handled incremental loading on the customer, transaction data by date.
- Generated dashboard from search and Scheduled searches of Inline search vs scheduled search in a dashboard.
- Created Splunk Dashboards, Visualizations as per the business end user requirements.
- Created Splunk Dashboards at UAT level before on-boarding them to production level.
- Created Dashboard for QA Testing team to monitor the results of different Testing results.
- Knowledge on integrating Cassandra with Hadoop.
- Developed a Scala API to download the required artifacts from Cassandra.
- Used KAFKA to load the data on Hadoop File system and move the same data to Cassandra NoSQL database.
- Develop strategy for integrating internal security model into new projects with Spring Security and Spring Boot.
- Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
- Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
- Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
- Import the data from different sources like HDFS/Hbase into Spark RDD.
- Experienced with batch processing of data sources using Apache Spark and Elastic search.
- Experienced in implementing Spark RDD transformations, actions to implement business analysis
- Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
- Installing, Upgrading and Managing Hadoop Clusters
- Used OOZIE Operational Services for batch processing and scheduling workflows dynamically.
Environment: Map Reduce, HDFS, Spark, Splunk 7.0.3, Cassandra, AWS, Hive, Pig, SQL, Sqoop, Oozie, Shell scripting, Cron Jobs, Apache Kafka, J2EE.
Confidential,Nashville, TN
Hadoop Developer
Responsibilities:- Involved in loading and transforming large sets of structured, semi structured and unstructured data from relational databases into HDFS using Sqoop imports.
- Developed Sqoop scripts to import export data from relational sources and handled incremental loading on the customer, transaction data by date.
- Migrated existing java application into microservices using spring boot and spring cloud.
- Working knowledge in different IDEs like Eclipse, IntelliJ, Spring Tool Suite.
- Working knowledge of using GIT, ANT/Maven for project dependency / build / deployment.
- Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
- Developed Spark code using Scala and Spark-SQL/Streaming for faster testing and processing of data.
- Worked as a part of AWS build team.
- Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.
- Create, configure and managing S3 bucket(storage).
- Developed customized application configurations in Splunk to parse, index multiple types of log format across all application environments.
- Worked on AWS EC2, EMR, LAMBDA and Cloud Watch.
- Import the data from different sources like HDFS/Hbase into Spark RDD.
- Worked on batch processing of data sources using Apache Spark and Elastic search.
- Implementing Spark RDD transformations, actions to implement business analysis
- Migrated Hive QL queries on structured into Spark QL to improve performance
- Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Worked on partitioning HIVE tables and running the scripts in parallel to reduce run-time of the scripts.
- Worked on Data Serialization formats for converting Complex objects into sequence bits by using AVRO, PARQUET, JSON, CSV formats.
- Responsible for analyzing and cleansing raw data by performing Hive/Impala queries and running Pig scripts on data.
- Administration, installing, upgrading and managing distributions of Hadoop, Hive, Hbase.
- Involved in performance of troubleshooting and tuning Hadoop clusters.
- Created Hive tables, loaded data and wrote Hive queries that run within the map.
- Implemented business logic by writing Hive UDFs in Java.
- Developed Shell scripts and some of Perl scripts based on the user requirement.
- Wrote XML scripts to build Oozie functionality.
- Used Oozie Operational Services for batch processing and scheduling workflows dynamically.
- Worked on creating End-End data pipeline orchestration using Oozie.
- Built Data set, Lens and visualization charts/graphs in the PLATFORA environment.
- Evaluated suitability of Hadoop and its ecosystem to the above project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
Environment: Map Reduce, HDFS, Spring Boot, Microservices, AWS, Hive, Pig, SQL, Sqoop, Oozie, Shell scripting, Cron Jobs, Apache Kafka, J2EE.
Confidential, Oaks, PA
Hadoop Developer
Responsibilities:- Involved in gathering and analyzing user requirements.
- Responsible for Installation and configuration of Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Developed Sqoop scripts to import export data from relational sources and handled incremental loading on the customer, transaction data by date.
- Build microservices for the delivery of software products across the enterprise.
- Develop strategy for integrating internal security model into new projects with Spring Security and Spring Boot.
- Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
- Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
- Import the data from different sources like HDFS/Hbase into Spark RDD.
- Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
- Installing, Upgrading and Managing Hadoop Clusters
- Used OOZIE Operational Services for batch processing and scheduling workflows dynamically.
- Administration, installing, upgrading and managing distributions of Hadoop, Hive, Hbase.
- Advanced knowledge in performance troubleshooting and tuning Hadoop clusters.
- Migrated Hive QL queries on structured into Spark QL to improve performance
- Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Worked on partitioning HIVE tables and running the scripts in parallel to reduce run-time of the scripts.
- Performed CRUD operations like Update, Insert and Delete data in MongoDB.
- Worked on MongoDB database design and indexing techniques.
- Worked on creating End-End data pipeline orchestration using Oozie.
- Worked on Amazon Web Services(AWS), Amazon Cloud Services like Elastic Compute Cloud(EC2), Simple Storage Service(S3), Elastic Map Reduce(EMR) Amazon Simple DB, Amazon Cloud Watch, SNS, SQS, LAMBDA.
- Evaluated suitability of Hadoop and its ecosystem to the above project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
- Participated in code reviews and weekly meetings.
Environment: Spring Boot, Microservices, AWS, Map Reduce, HDFS, Hive, Pig, SQL, Sqoop, Oozie, Shell scripting, Cron Jobs, Perl scripting, Apache Kafka, J2EE.
Confidential, Waterbury, CT
Java Developer
Responsibilities:- Developed the application based on MVC (Model-View-Controller) Architecture using Struts.
- Involved in end to end software development life cycle (SDLC).
- Involved in developing GUI using HTML, DHTML, JSP and AJAX for end users and loan processing persons.
- Used Struts Validation Frame Work for server-side data validations.
- Used WSAD for the entire application development.
- Used DAO using JDBC and Oracle 9i as the backend database.
- Set up relational database entities.
- Developed test cases using Junit.
- Wrote queries to retrieve data using SQL and PL/SQL.
- Developed producer and consumer web services component using Apache Axis.
- Used Apache camel, CXRF web services, Service Mix engine, Active MQ, etc. to build a custom ESB (Enterprise Service Bus).
- Used LDAP Authentication for single sign and authentication, logging and Validating User Credentials.
- Used JIRA for the issues reported by the testing team.
- Rational Clear Quest was used to log and track bugs.
- Used CVS repository for version control.
Environment: J2sdk1.5, WSAD 5.0, HTML, JSP, Struts 2.0, Apache camel, JDBC, SQL, WebSphere 5.X, Oracle 9i, UNIX (Solaris), CVS, JUnit, JWebUnit, EJB.
Confidential
Java/UI Programmer
Responsibilities:- Identifying the technical areas that contributed to the issues and documenting them.
- Involved in all the phases of SDLC including Requirements Collection, Design & Analysis of the Customer
- Specifications, Development and Customization of the Application.
- Responsible for writing detailed design specification document and implementing all business rules as per Design document.
- Wrote ANT Build for the Cruise Build Platform.
- Wrote JUnit Test Cases.
- Used JMESA, JSP, and HTML for Front View developments.
- Worked on Database batch updates, commits and roll back.
- Design Patterns like Façade, Singleton and Factory Pattern were used.
- Developed DAOs (Data Access Object) using JDBC to interact with DBMS - Oracle 10g
- Developed SQL Queries, Stored Procedures, Views.
- Developed application in SVN version control.
- Used Struts 1.2 for Model View Controller Framework
- Used JDBC for CRUD operations.
- Involved in fine tuning of Query.
Environment: Java 5, JSP, JMESA, ANT, CRUISE BUILD, Struts 1.2, WebLogic Server
Confidential
Web Developer/ Consultant
Responsibilities:- Involved in business requirements gathering and transforming to technical specifications.
- Created prototypes for approval from clients.
- Developed user interfaces using JSP, Servlet s, HTML, CSS, java scripts.
- Implemented logging using Log4j.
- Implemented DAO pattern for data access, used JDBC to interface with the database.
- Implemented test cases using jUnit.
- Set up the relational database with Oracle 8.
- Used Data modelling techniques. wrote several queries in SQL/PL SQL to test data consistency
Environment: Tomcat, Java, JSP, Java Script, SQL Server, Eclipse, Oracle 8.