Hadoop Developer Resume
LA
SUMMARY
- 5+ years of overall experience in IT Industry which includes experience in Java, Big data technologies (Hadoop).
- 2 years of comprehensive experience in Big Data Analytics.
- Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and Map Reduce concepts.
- Experience in using Pig, Hive, Sqoop and Cloudera Manager.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Hands on experience in RDBMS, and Linux shell scripting
- Extending Hive and Pig core functionality by writing custom UDFs.
- Experience in analyzing data using HiveQL, Pig Latin and Map Reduce.
- Expertise in Domain Retail and Finance.
- Knowledge in job work-flow scheduling and monitoring tools like Oozie.
- Knowledge of data warehousing and ETL tools like Informatica and Pentaho.
- Techno-functional responsibilities include interfacing with users, identifying functional and technical gaps, estimates, designing custom solutions, development, leading developers, producing documentation, and production support.
- Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.
- Trained in Cloudera for Apache Hadoop.
- Extensive experience spans with implementing SOA architecture using WebServices using both SOAP and Rest.
- Project delivery of various sizes, performance tuning, production support, trouble shooting and development.
- Extensive work experience as onsite coordinator attending the project review meetings with Clients, gathering requirements, preparing HLD, LLD, requirements transition and design to offshore for the Build activity.
- Strong problem solving skills and a thirst to learn the latest Technologies in the industry.
- Ability to handle multiple projects concurrently, meet strict deadlines, coordination with multiple teams, Systems and Applications.
- Trained in multiple Soft Skills like Time Management, Cross Culture, Client Interaction and Communications.
- Java developer with experience withHadoopand Kafka.
- Have knowledge in projects like Data Integration & Migration with functional knowledge in Banking & Finance, Retail and Healthcare & life science projects.
TECHNICAL SKILLS
Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Sqoop, Oozie
Build Tool & IDE: Eclipse
Languages: Java, JDBC, Sevlets, JSP, Java Script, Shell scripting, COBOL, JCL, PL/SQL, REXX, Ezytrieve
Mainframe Tools: Changeman, Endevor, Fila, File Manager, ISPF, Xpeditor, IBM Utilities, CA - 7.
Testing Tools: HP Quality Center
PROFESSIONAL EXPERIENCE
Confidential, LA
Hadoop Developer
Responsibilities:
- Work on analyzing and writing Map Reduce jobs based on the requirements.
- Involve in creating Hive tables, loading with data and writing hive queries that will internally run as Map Reduce job.
- Work on Flume to stream Twitter data in to HDFS to analyze Black Friday promotions.
- Work on creating Hive scripts for data analysts based on the Ad-hoc requirements.
- Involve in writing UDF for Hive.
- Schedule the workflow using Oozie workflow scheduler.
- Extract the data from external data sources into HDFS using Sqoop.
- Develop Map Reduce programs using combiner and custom partition.
- Create Sqoop job to populate Hive external tables.
- Involve in configuring Flume to import Sofi and DGW logs to HDFS.
- Using Cloudera Manager for monitoring purpose.
- Involve in loading data from Linux file system to HDFS
- Involve in all the phases of the SDLC using Agile Scrum methodology.
- Working in Agile environment with 3 weeks sprint, involving in grooming meeting, sprint planning meeting, retrospective meeting and daily stand up meeting.
- Using Version One tool for tracking all agile related activities like stories, tasks, estimated story points and actual burn hours.
Environment: Hadoop, Map Reduce, HDFS, Hive, Sqoop, Flume, Oozie, Cloudera ManagerHue UI, Java (JDK 1.6), Eclipse, Maven, JSON, MySQL, Ubuntu, Version One and Cloudera Distribution
Confidential, Washington D.C
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing
- Involved in data extraction from distributed RDBMS like Teradata and Oracle.
- Involved in loading data from UNIX file system to HDFS.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Used Map Reduce JUnit for unit testing.
- Troubleshooting the cluster by reviewing Hadoop LOG files. Involved in managing and reviewing Hadoop log files.
- Installed and configured Pig for ETL jobs.
- Used Oozie to manage the Hadoop jobs.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Used CDH3 and CDH4 distributions for development and deployment.
- Imported data using Sqoop from Tera data using Tera data connector.
- Implemented Partitioning, Dynamic Partitioning, and Bucketing in HIVE.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Used Zookeeper for various types of centralized configurations.
- Involved in maintaining various Unix Shell scripts.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
- Used SVN for version control.
- Maintain System integrity of all subcomponents (primarily HDFS, MR, HBase, and Flume).
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
Environment: Hadoop, Map Reduce, Hive, Pig, Sqoop,HBase(NoSQL database), Java 1.6,and UNIX Shell Scripting.
Confidential
System Analyst
Responsibilities:
- Involved in software development life cycle (SDLC) of the project (analysis, design, development, testing and implementation).
- Used MS-Visio for analysis & design flow and monitored the initial applications prototype development for the project.
- Leading a team of 4 developers and point of contact for onshore / offshore communication.
- Designed and developed web pages using JSP, JDBC and Servlets.
- Developed modules using Java.
- Used HTML, CSS, XML and JavaScript to design a page.
- Developed Web Services to get data from the external system in terms of .txt file to load into the database.
- Developed DTS/SSIS packages to load employee details into row Mony tables of the SQL server for further processing.
- Wrote Stored Procedures, functions and complex SQL queries for database operations.
- Used Java Scripts to perform client side validations.
- Worked on performance tuning of queries.
- Developed reports using Crystal Reports reporting tool.
- Have used LDAP and Active Directory Search Interface (ADSI) to authenticate and authorize user.
- Involved in unit testing and production support of the application.
- Defects were managed through the Remedy
Environment: JEE, JSP, Servlets, SQL Server 2005, T-SQL/PL-SQL, HTML, XML, Web Services, JavaScript, Tomcat, Visual Source Safe (VSS), Remedy and Crystal Reports
Confidential
Java Developer & Support
Responsibilities:
- Provide L3 application support as primary on call.
- Involved in the development of Report Generation module which includes volume statistics report, Sanctions Monitoring Metrics report, and TPS report.
- Implemented Online List Management (OLM) and FMM module using spring and Hibernate.
- Wrote various SQL, PL/SQL queries and stored procedures for data retrieval.
- Created Configuration files for the application using Spring framework.
Environment: JEE, JSP, Servlets, JQuery, JavaScript, CSS, HTML, SQL