Hadoop Developer Resume
Sunnyvale, CA
SUMMARY
- Over 8 +years of IT experience which includes 4+ years of experience with Hadoop, HDFS, MapReduce and Hadoop Ecosystem (Pig, Hive, HBase)
- Over 4 years of experience in JAVA programming with various technologies like Java, J2EE, JavaScript and Data Structures.
- Well versed in designing and implementing Map Reduce jobs using JAVA on Eclipse to solve real world scaling problems.
- Experience in deploying applications in heterogeneous Web/Application Servers TOMCAT, WebLogic and Oracle Application Server.
- Worked on different IDE tools like Oracle JDeveloper, WebLogic Workshop, NetBeans and Eclipse.
- Exposure in writing backend Stored Procedures, Triggers and Functions
- Experience in developing customized UDF’s in java to extend Hive and Pig Latin functionality.
- Experience in using Flume to load log files into HDFS.
- Expertise in using Oozie for configuring job flows.
- Experience in performance tuning for Mapreduce, Hive and Scoop
- Performed Importing and exporting data into HDFS and Hive using Sqoop.
- Experience in designing both time driven and data driven automated workflows using Oozie.
- Developed Map Reduce jobs based on the use cases using Java Map Reduce, Pig and Hive.
- Designed ingestion framework using flume for streaming logs and aggregated data into HDFS. Built data transform framework using MapReduce and Pig
- Hands - on experience with "productionalizing" Hadoop applications (e.g. administration, configuration management, monitoring, debugging, and performance tuning)
- Strong experience in RDBMS technologies like ORACLE, MY SQL.
- Experience in scripting for automation, and monitoring using Shell script.
- Excellent communication skills, interpersonal skills, problem solving skills a very good team player along with a can do attitude and ability to effectively communicate with all levels of the organization such as technical, management and customers.
TECHNICAL SKILLS
BIG Data & Ecosystem: Hadoop, HDFS, Map Reduce, Hive, Pig, Flume, Oozie, Zookeeper, HBASE and Sqoop, CasandraOperating System RHEL 5,6, Centos 6.4/6.5,Ubuntu, Windows 2000/2003/XP/Vista.
J2EE Technologies: JSP, JDBC, EJB, JMS, HTML. DHTML, JQuery and JavaSqcript.
Script Languages: Shell scripts, Java Scripts
Web Services: Restful, Soap and WSDL
RDBMS: MySQL, Postgre SQL, Oracle 10g/11g, SQL Server
Tools: /IDE: Eclipse 3.6, Net Beans 6.x, Oracle JDeveloper, TOAD 9.X/8.X
Programming Languages: C, SQL, HTML, Java, XML and Pig Latin
Web/App. Server: WebLogic 10.x/9.x, Apache Tomcat 6.x/5.x, Web Sphere 6.x/5.x, JBoss 5.x/4.x
Testing API: Junit
PROFESSIONAL EXPERIENCE
Confidential, Sunnyvale, CA
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Importing and exporting data into HDFS and Hive using Sqoop
- Experienced in defining job flows
- Experienced in managing and reviewing Hadoop log files
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data
- Load and transform large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Supported Map Reduce Programs those are running on the cluster
- Loaded data from Oracle database, MS SQL server into HDFS, HIVE, and Netezza using Informatica tool.
- Used Automation script to import Informatica mapping and workflow.
- Also used Informatica Developer for Incremental load and Python framework for data loading
- Involved in loading data from UNIX file system to HDFS.
- Extracted data from Teradata to HDFS using sqoop.
- Installed and configured Hive and also written Hive UDFs.
- Involved in running algorithms listed in HiBench paper after creating test data.
- Coding with Python, PHP and a little JavaScript.
- Wrote python scripts to parse XML documents and load the data in database.
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way
- Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process etc.
- Migrated high-volume OLTP transactions from Oracle to Cassandra in order to reduce Oracle licensing footprint.
Confidential, Kansas, MO
Big Data Consultant
Responsibilities:
- Implemented Cloudera Hadoop environment.
- Installed and configured Hadoop, Map Reduce, HDFS
- Installed and configured Hadoop Ecosystem components.
- Troubleshot performance and tunedHadoopclusters.
- Extracted data from Teradata to HDFS using sqoop
- Involved in extracting customer’s Big data from various data sources into Hadoop HDFS. This included data from mainframes, databases and also log data from servers.
- Used Sqoop to efficiently transfer data between databases and HDFS and used Flume to stream the log data from servers.
- Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
- The Hive tables created as per requirement were internal or external tables defined with appropriate static and dynamic partitions, intended for efficiency.
- Implemented partitioning, bucketing in Hive for better organization of the data.
- Created Python/MySQL back-end for data entry from Flash. I had to assist the Flash developer send the correct data via querystrings.
- Profiled python code for optimization and memory management.
- Developed UDFs in Pig and Hive sed Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as system specific jobs.
- Worked with BI teams in generating the reports on Tableau
- Used Informatica 9.6.1 for full refresh and incremental load
- Also used Informatica Developer for Incremental load and Python framework for data loading
- Experience in Developing the Hive queries for the transformations, aggregations and Mappings on the Customer Data.
- Worked on importing and exporting data into HDFS and Hive using Sqoop.
- Worked on analyzing/transforming the data with Hive and Pig.
- Developed map reduce programs for applying business rules on the data
- Developed and executed Hive Queries for denormalizing the data.
- Automated workflow using Shell Scripts.
- Performance Tuning on Hive Queries.
- Involved in migration of data from one Hadoop Cluster to the Hadoop Cluster.
- Worked on configuring multiple Map Reduce Pipelines, for the new Hadoop Cluster.
Confidential, Bloomington, IL
Java/BIG DATA Developer
Responsibilities:
- Involved to provide architect, design, develop and testing services to Confidential for sub-system components within the data aggregation infrastructure associated with the project.
- Developed Hive UDFs for rating aggregation.
- Importing and exporting data into HDFS and Hive using Sqoop
- Experienced in defining job flows
- Experienced in managing and reviewing Hadoop log files
- Developed Hbase java client API for CRUD Operations.
- Developed the Java client API for node provisioning, load balancing and artifact deployment.
- Responsible to manage data coming from different sources
- Used Oozie tool for job scheduling.
- Developed new enhancements for the system based on the user requirements
- Participated in JAD sessions with the developers
- Designed and developed user interface screens using JSP
- Developed stored procedures, triggers, functions for the application.
- Worked in a team of five developers to complete different modules .
- Implemented server-side business components using session beans.
- Added client side validations using JavaScript.
- Used Spring framework for wiring and maintained configuration file.
- Created various DAO components using Hibernate.
- Implemented SOAP based web services.
- Enhanced the resume search criteria by including additional parameters.
- Developed the automatic email alert feature for employers to see the overview of daily applications.
- Optimized existing reports for performance and generated new reports
- Worked on resolving service requests submitted by the management on a daily basis
- Used Oracle Java Developer and SQL Navigator as tools for Java and PL/SQL development
Confidential, Charlotte, NC
Java/J2EE Developer
Responsibilities:
- Responsible and active in the analysis, design, implementation and deployment of full software development life-cycle (SDLC) of the project.
- Designed and developed user interface using JSP, HTML and JavaScript.
- Developed struts action classes, action forms and performed action mapping using Struts Framework and performed data validation in form beans and action classes.
- Involved in multi-tiered J2EE design utilizing MVC architecture (Struts Framework) and Hibernate.
- Extensively used Struts Framework as the controller to handle subsequent client requests and invoke the model based upon user requests.
- Involved in system design and development in core java using Collections, multithreading.
- Defined the search criteria and pulled out the record of the customer from the database. Make the required changes to the record and save the updated information back to the database.
- Wrote JavaScript validations to validate the fields of the user registration screen and login screen.
- Developed build and deployment scripts using Apache ANT to customize WAR and EAR files.
- Used DAO and JDBC for database access.
- Developed applications with ANT based build scripts.
- Developed stored procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
- Design and develop XML processing components for dynamic menus on the application.
- Involved in postproduction support and maintenance of the application.
Environment: Oracle 11g, Java 1.5, Struts 1.2, Servlets, HTML, XML, MS SQL Server 2005, J2EE, JUnit, Tomcat 6.