Sr. Java Hadoop Developer Resume
New, JerseY
PROFESSIONAL SUMMARY
- Over 7 Years of professional IT experience which includes experience in Big data ecosystem and Java/J2EE related technologies
- Hands on experience in development, installation, configuring, and using Hadoop & ecosystem components like Hadoop MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Flume, Kafka, Storm and Spark
- Implemented in setting up standards and processes for Hadoop based application design and implementation.
- Good Knowledge on Hadoop Cluster architecture and monitoring teh cluster.
- Experience in managing and reviewing Hadoop log files.
- Specializing in ML/Big Data and Web architecture solutions using Scala2.11, Python, Spark, Kafka and Storm.
- Expertise in JVM (Java Virtual Machine) and Java based Middleware.
- Experienced with Cloudera, Horton networks and Map R
- In - depth understanding of Data Structure and Algorithms.
- Extensive experience in working with MS Excel, SQLServer and RDBMS databases
- Experience in developing some deliverable documentations including Data Flow, Use Cases, and Business rules.
- Responsible for developing efficient MapReduce on AWS cloud programs for more TEMPthan 20 years' worth of claim data to detect and separate fraudulent claims.
- Uploaded and processed more TEMPthan 30 terabytes of data from various structured and unstructured sources into HDFS (AWS cloud) using Sqoop and Flume.
- Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Extensive experience in developing applications using JSP, Servlets, Spring, Hibernate, Java Script, Angular, AJAX, CSS, JQuery, HTML, JDBC, JNDI, JMS, XML, and SQL across teh platforms like Windows, Linux, and UNIX.
- Proven ability to investigate and customize large scale software like JVM, Web kit and open source projects.
- Extensive experience in developing and deploying RESTful and SOAP Web Services.
- Good experience on web/Application servers like WebSphere, Apache Tomcat, and JBoss.
- Experienced in a variety of scripting languages such as UNIX scripts and Java Scripts.
- Installing, configuring and managing of Hadoop Clusters and Data Science tools
- Managing teh Hadoop distribution with Cloudera Manager, Cloudera Navigator, Hue.
- Setting up teh High-Availability for Hadoop Clusters components and Edge nodes
- Experience in developing Shell scripts and Python Scripts for system management.
- Strong domain noledge in Insurance, Finance & Health Care, Social Network
TECHNICAL SKILLS
Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig, Sqoop, Oozie, Flume, Spark.
Programming Languages: Java, C/C++, VB
Scripting Languages: JSP & Servlets, PHP, JavaScript, XML, HTML, Python.
Databases: Oracle, My SQL, MS SQL
Tools: Eclipse, CVS, Ant, MS Visual Studio, NetBeans
Platforms: Windows, Linux/Unix
Application Servers: Apache Tomcat 5.x 6.0, Jboss 4.0
Methodologies: Agile, UML, Design Patterns
PROFESSION EXPERIENCE:
Confidential, New Jersey
Sr. Java Hadoop Developer
Roles and Responsibilities:
- Loaded customer personal and transaction data into HDFS using Sqoop.
- Developed custom MapReduce programs for data validation to filter unwanted data and cleanse teh data
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
- Developed complex Hive Scripts for processing teh data. Created dynamic partitions and bucketing in hive to improve teh query performance.
- Developed MapReduce applications using Hadoop Map-Reduce programming framework for processing. Used compression techniques to optimize MapReduce Jobs.
- Created HBase tables from Hive and Wrote HiveQL statements to access HBase table's data.
- Developed Spark programs using Scala for processing data in a faster way.
- Developed Pig UDF's to no teh customer behavior and Pig Latin scripts for processing teh data in Hadoop.
- Written code to read & create different file formats like ORC, Parquet.
- Scheduled automated tasks with Oozie for loading data into HDFS through Sqoop and pre-processing teh data with Pig and Hive.
- Worked on Spring Security for user Autantication and Authorization using LDAP autantication provider.
- Written java code for file writing and reading, extensive usage of data structure ArrayList and HashMap.
- Implemented teh MVC architecture using Spring MVC framework
- Composing teh application classes as Spring Beans using Spring IOC/Dependency Injection.
- Designed and Developed server-side components using Java, REST, WSDL
Environment: MapReduce, Hive, Sqoop, Shell Scripting, LINUX, Cloudera CDH 5, Scala, Pig, HBase, Avro, Spark, Cassandra, spring, Java Collections, REST, WSDL, Zookeeper and MySQL.
Confidential, Denver, CO
Sr. Java Big Data / Hadoop Developer
Roles and Responsibilities:
- Configured a Spark streaming application to stream syslog’s and various application logs from 100+ nodes for monitoring and alerting as well as to feed teh data to dynamic dashboards.
- Migrated traditional MR jobs to Spark MR Jobs.
- Worked on setup process Hadoop cluster on Amazon EMR/S3 for POC.
- Transformed teh processing framework from Map-Reduce to Spark. This increase teh overall speed of processing. Processed teh data using Spark Context, Spark-SQL, Data Frame API and Pair RDDs using Scala.
- Worked on Spark SQL and Spark Streaming.
- Imported, exported file to teh HDFS, Hive, Impala SQL.
- Teh processed results were consumed by HIVE, Scheduling applications and various other BI reports through data warehousing multi-dimensional models.
- Run Ad-Hoc query through PIG Latin language, Hive or Java MapReduce
- Worked on setup process of Hadoop cluster on Amazon EMR / S3 for poc.
- Big data analysis using Pig, Hive and User defined functions (UDF)
- Performed joins, group by and other operations in MapReduce using Java or PIG Latin
- Setting up Virtual Machines and managing storage devices
- Involved in managing and reviewing Hadoop log files
- Developed Scripts and Batch Jobs to schedule various Hadoop Program.
- Used Amazon EC2 as a instance from Amazon S3 web services on Databricks framework.
- I analyzed Apache webserver log files using Notebook.
- Notebook allow users to write and run arbitrary apache spark code, interactively visualize teh results.
Environment: Hadoop 2.6.0-cdh5.4.2, Scala2.11, Spark, Hadoop, Cloudera Manager, Big Data, Redhat Linux, java, Perl, Cloudera Navigator, Amazon AWS, Elastic Search.
Confidential, San Jose-CA
Sr. Big Data/Hadoop Developer
Roles and Responsibilities:
- Imported logs from web servers with Flume to ingest teh data into HDFS.
- Implemented custom interceptors for flume to filter data and defined channel selectors to multiplex teh data into different sinks.
- Retrieved data from HDFS into relational databases with Sqoop.
- Parsed cleansed and mined useful and meaningful data in HDFS using Map-Reduce for further analysis
- Fine tuning hive jobs for optimized performance
- Implemented UDFS, UDAFS, UDTFS in java for hive to process teh data that can't be performed using Hive inbuilt functions.
- Designed and implemented PIG UDFS for evaluation, filtering, loading and storing of data.
- Developed workflow in Oozie to automate teh tasks of loading teh data into HDFS and pre-processing with Pig.
Environment: Hadoop, Big Data, Hive, Pig, Sqoop, Cloudera, Oracle10g, HDFS, Oozie, Flume.
Confidential -NJ
Java Developer
Roles and Responsibilities:
- Was a part of teh architecture team for design and implementation of site components using J2EE framework?
- Involved in implementing teh Spring Model-View-Controller Architecture for teh site, which accomplishes a tight and neat co-ordination of JSP Pages, Java Beans.
- Used JQuery and JavaScript in conjunction to perform Front end validations and event handling.
- Improved teh performance of teh response page for teh Survey responses which include loading, saving and validating teh responses.
- Implemented teh JQuery Pagination feature to drastically improve teh response loading page which resulted in loading teh page in 2 seconds (which was 8 seconds earlier).
- Used teh JQuery Data Tables to render teh data in teh table format.
- Used AJAX to increase teh web page’s interactivity, speed, functionality and usability.
- Used iText library to create and manipulate all teh PDF reports.
- Used Apache POI API for exporting all teh data in to Excel sheets.
- Used Scheduler to schedule and automate some back-end jobs.
- Used Dynatree JQuery view plugin to support different components of teh system supported by multiple selection, Drag and Drop features.
- Used ClearCase for Code Repository, ClearQuest for defect tracking and Log4j for logging in teh application.
Environment: Java, J2EE, RAD 8.0.4, WebSphere 8.0, JavaScript, JQUERY, AJAX, Hibernate, Windows/UNIX.
Confidential, New Jersey
Java Developer
Roles and Responsibilities:
- Worked on designing teh content and delivering teh solutions based on understanding teh requirements.
- Wrote web service client for tracking operations for teh orders which is accessing web services API and utilizing in our web application.
- Developed User Interface using JavaScript, JQuery and HTML.
- Used AJAXAPI for intensive user operations and client-side validations.
- Worked with Java, J2EE, SQL, JDBC, XML, JavaScript, web servers.
- Utilized Servlet for teh controller layer, JSP and JSP tags for teh interface
- Worked on Model View Controller Pattern and various design patterns.
- Worked with designers, architects, developers for translating data requirements into teh physical schema definitions for SQL sub-programs and modified teh existing SQL program units.
- Designed and Developed SQL functions and stored procedures.
- Involved in debugging and bug fixing of application modules.
- Efficiently dealt with exceptions and flow control.
- Worked on Object Oriented Programming concepts.
- Added Log4j to log teh errors.
- Installed and used MS SQL Server 2008 database.
- Spearheaded coding for site management which included change of requests for enhancing and fixing bugs pertaining to all parts of teh website.
Environment: Java, JDK1.8, Apache Tomcat-7, JavaScript, JSP, JDBC, Servlets, MS SQL, XML, Windows XP, Ant, SQL Server database, Eclipse.
Confidential, Bensenville, IL
Java Developer
Roles and Responsibilities:
- Involved in teh complete SDLC software development life cycle of teh application from requirement analysis to testing.
- Supported Java applications on 200 Red Hat, CentOS and Solaris servers in a NIS & NFS network
- Developed teh modules based on struts MVC Architecture.
- Worked with several types of controllers like simple form controller, Abstract Controller and Controller Interface etc.
- Developed UI modules using HTML, JSP, JavaScript and CSS.
- Involved in writing and executing queries in MySQL.
- Build test cases and performed unit testing.
- Developed custom components using jQuery
- Implemented code for validating teh input fields and displaying teh error messages.
- Provided Technical support for production environments resolving teh issues, analyzing teh defects, providing and implementing teh solution defects.
- Developed coded, tested, debugged and deployed JSPs and Servlets for teh input and output forms on teh web browsers.
- Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in Oracle9i.
- Experience in going through bug queue, analyzing and fixing bugs, escalation of bugs.
Environment: Java, MVC, HTML, CSS, JavaScript, JSP, MySQL, Oracle, JDBC, RHEL, Rational Rose, Tomcat Web Server.
