Sr. Java Hadoop Developer Resume
New, JerseY
SUMMARY
- Over 7 Years of professional IT experience which includes experience in Big data ecosystem and Java/J2EE related technologies
- Hands on experience in development, installation, configuring, and using Hadoop & ecosystem components like Hadoop MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Flume, Kafka, Storm and Spark
- Implemented in setting up standards and processes for Hadoop based application design and implementation.
- Good Knowledge on Hadoop Cluster architecture and monitoring the cluster.
- Experience in managing and reviewing Hadoop log files.
- Specializing in ML/Big Data and Web architecture solutions using Scala2.11, Python, Spark, Kafka and Storm.
- Expertise in JVM (Java Virtual Machine) and Java based Middleware.
- Experienced with Cloudera, Horton networks and Map R
- In - depth understanding of Data Structure and Algorithms.
- Extensive experience in working with MS Excel, SQLServer and RDBMS databases
- Experience in developing some deliverable documentations including Data Flow, Use Cases, and Business rules.
- Responsible for developing efficient MapReduce on AWS cloud programs for more than 20 years' worth of claim data to detect and separate fraudulent claims.
- Uploaded and processed more than 30 terabytes of data from various structured and unstructured sources into HDFS (AWS cloud) using Sqoop and Flume.
- Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Extensive experience in developing applications using JSP, Servlets, Spring, Hibernate, Java Script, Angular, AJAX, CSS, JQuery, HTML, JDBC, JNDI, JMS, XML, and SQL across the platforms like Windows, Linux, and UNIX.
- Proven ability to investigate and customize large scale software like JVM, Web kit and open source projects.
- Extensive experience in developing and deploying RESTful and SOAP Web Services.
- Good experience on web/Application servers like WebSphere, Apache Tomcat, and JBoss.
- Experienced in a variety of scripting languages such as UNIX scripts and Java Scripts.
- Installing, configuring and managing of Hadoop Clusters and Data Science tools
- Managing the Hadoop distribution with Cloudera Manager, Cloudera Navigator, Hue.
- Setting up the High-Availability for Hadoop Clusters components and Edge nodes
- Experience in developing Shell scripts and Python Scripts for system management.
- Strong domain knowledge in Insurance, Finance & Health Care, Social Network
TECHNICAL SKILLS
Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig, Sqoop, Oozie, Flume, Spark.
Programming Languages: Java, C/C++, VB
Scripting Languages: JSP & Servlets, PHP, JavaScript, XML, HTML, Python.
Databases: Oracle, My SQL, MS SQL
Tools: Eclipse, CVS, Ant, MS Visual Studio, NetBeans
Platforms: Windows, Linux/Unix
Application Servers: Apache Tomcat 5.x 6.0, Jboss 4.0
Methodologies: Agile, UML, Design Patterns
PROFESSIONAL EXPERIENCE
Confidential, New Jersey
Sr. Java Hadoop Developer
Responsibilities:
- Loaded customer personal and transaction data into HDFS using Sqoop.
- Developed custom MapReduce programs for data validation to filter unwanted data and cleanse the data
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
- Developed complex Hive Scripts for processing the data. Created dynamic partitions and bucketing in hive to improve the query performance.
- Developed MapReduce applications using Hadoop Map-Reduce programming framework for processing. Used compression techniques to optimize MapReduce Jobs.
- Created HBase tables from Hive and Wrote HiveQL statements to access HBase table's data.
- Developed Spark programs using Scala for processing data in a faster way.
- Developed Pig UDF's to know the customer behavior and Pig Latin scripts for processing the data in Hadoop.
- Written code to read & create different file formats like ORC, Parquet.
- Scheduled automated tasks with Oozie for loading data into HDFS through Sqoop and pre-processing the data with Pig and Hive.
- Worked on Spring Security for user Authentication and Authorization using LDAP authentication provider.
- Written java code for file writing and reading, extensive usage of data structure ArrayList and HashMap.
- Implemented the MVC architecture using Spring MVC framework
- Composing the application classes as Spring Beans using Spring IOC/Dependency Injection.
- Designed and Developed server-side components using Java, REST, WSDL
Environment: MapReduce, Hive, Sqoop, Shell Scripting, LINUX, Cloudera CDH 5, Scala, Pig, HBase, Avro, Spark, Cassandra, spring, Java Collections, REST, WSDL, Zookeeper and MySQL.
Confidential, Denver, CO
Sr. Java Big Data / Hadoop Developer
Responsibilities:
- Configured a Spark streaming application to stream syslog’s and various application logs from 100+ nodes for monitoring and alerting as well as to feed the data to dynamic dashboards.
- Migrated traditional MR jobs to Spark MR Jobs.
- Worked on setup process Hadoop cluster on Amazon EMR/S3 for POC.
- Transformed the processing framework from Map-Reduce to Spark. This increase the overall speed of processing. Processed the data using Spark Context, Spark-SQL, Data Frame API and Pair RDDs using Scala.
- Worked on Spark SQL and Spark Streaming.
- Imported, exported file to the HDFS, Hive, Impala SQL.
- The processed results were consumed by HIVE, Scheduling applications and various other BI reports through data warehousing multi-dimensional models.
- Run Ad-Hoc query through PIG Latin language, Hive or Java MapReduce
- Worked on setup process of Hadoop cluster on Amazon EMR / S3 for poc.
- Big data analysis using Pig, Hive and User defined functions (UDF)
- Performed joins, group by and other operations in MapReduce using Java or PIG Latin
- Setting up Virtual Machines and managing storage devices
- Involved in managing and reviewing Hadoop log files
- Developed Scripts and Batch Jobs to schedule various Hadoop Program.
- Used Amazon EC2 as a instance from Amazon S3 web services on Databricks framework.
- I analyzed Apache webserver log files using Notebook.
- Notebook allow users to write and run arbitrary apache spark code, interactively visualize the results.
Environment: Hadoop 2.6.0-cdh5.4.2, Scala2.11, Spark, Hadoop, Cloudera Manager, Big Data, Redhat Linux, java, Perl, Cloudera Navigator, Amazon AWS, Elastic Search.
Confidential - San Jose-CA
Sr. Big Data/Hadoop Developer
Responsibilities:
- Imported logs from web servers with Flume to ingest the data into HDFS.
- Implemented custom interceptors for flume to filter data and defined channel selectors to multiplex the data into different sinks.
- Retrieved data from HDFS into relational databases with Sqoop.
- Parsed cleansed and mined useful and meaningful data in HDFS using Map-Reduce for further analysis
- Fine tuning hive jobs for optimized performance
- Implemented UDFS, UDAFS, UDTFS in java for hive to process the data that can't be performed using Hive inbuilt functions.
- Designed and implemented PIG UDFS for evaluation, filtering, loading and storing of data.
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
Environment: Hadoop, Big Data, Hive, Pig, Sqoop, Cloudera, Oracle10g, HDFS, Oozie, Flume.
Confidential -Baskin ridge-NJ
Java Developer
Responsibilities:
- Was a part of the architecture team for design and implementation of site components using J2EE framework?
- Involved in implementing the Spring Model-View-Controller Architecture for the site, which accomplishes a tight and neat co-ordination of JSP Pages, Java Beans.
- Used JQuery and JavaScript in conjunction to perform Front end validations and event handling.
- Improved the performance of the response page for the Survey responses which include loading, saving and validating the responses.
- Implemented the JQuery Pagination feature to drastically improve the response loading page which resulted in loading the page in 2 seconds (which was 8 seconds earlier).
- Used the JQuery Data Tables to render the data in the table format.
- Used AJAX to increase the web page’s interactivity, speed, functionality and usability.
- Used iText library to create and manipulate all the PDF reports.
- Used Apache POI API for exporting all the data in to Excel sheets.
- Used Scheduler to schedule and automate some back-end jobs.
- Used Dynatree JQuery view plugin to support different components of the system supported by multiple selection, Drag and Drop features.
- Used ClearCase for Code Repository, ClearQuest for defect tracking and Log4j for logging in the application.
Environment: Java, J2EE, RAD 8.0.4, WebSphere 8.0, JavaScript, JQUERY, AJAX, Hibernate, Windows/UNIX.
Confidential - Woodbridge, New Jersey
Java Developer
Responsibilities:
- Worked on designing the content and delivering the solutions based on understanding the requirements.
- Wrote web service client for tracking operations for the orders which is accessing web services API and utilizing in our web application.
- Developed User Interface using JavaScript, JQuery and HTML.
- Used AJAXAPI for intensive user operations and client-side validations.
- Worked with Java, J2EE, SQL, JDBC, XML, JavaScript, web servers.
- Utilized Servlet for the controller layer, JSP and JSP tags for the interface
- Worked on Model View Controller Pattern and various design patterns.
- Worked with designers, architects, developers for translating data requirements into the physical schema definitions for SQL sub-programs and modified the existing SQL program units.
- Designed and Developed SQL functions and stored procedures.
- Involved in debugging and bug fixing of application modules.
- Efficiently dealt with exceptions and flow control.
- Worked on Object Oriented Programming concepts.
- Added Log4j to log the errors.
- Installed and used MS SQL Server 2008 database.
- Spearheaded coding for site management which included change of requests for enhancing and fixing bugs pertaining to all parts of the website.
Environment: Java, JDK1.8, Apache Tomcat-7, JavaScript, JSP, JDBC, Servlets, MS SQL, XML, Windows XP, Ant, SQL Server database, Eclipse.
Confidential, Bensenville, IL
Java Developer
Responsibilities:
- Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
- Supported Java applications on 200 Red Hat, CentOS and Solaris servers in a NIS & NFS network
- Developed the modules based on struts MVC Architecture.
- Worked with several types of controllers like simple form controller, Abstract Controller and Controller Interface etc.
- Developed UI modules using HTML, JSP, JavaScript and CSS.
- Involved in writing and executing queries in MySQL.
- Build test cases and performed unit testing.
- Developed custom components using jQuery
- Implemented code for validating the input fields and displaying the error messages.
- Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects.
- Developed coded, tested, debugged and deployed JSPs and Servlets for the input and output forms on the web browsers.
- Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in Oracle9i.
- Experience in going through bug queue, analyzing and fixing bugs, escalation of bugs.
Environment: Java, MVC, HTML, CSS, JavaScript, JSP, MySQL, Oracle, JDBC, RHEL, Rational Rose, Tomcat Web Server.
