Big Data Developer Resume
SUMMARY
- 11 plus years in IT industry encompassing a wide range of skill set and industry verticals.
- Of which 3 plus years of experience in processing and analysis of big data using Big data ecosystem such as Hadoop, HDFS, Map Reduce, Pig, Hive, HBase, Oozie, Kafka and Zookeeper.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce Framework concepts
- Analyzing text, streams, documents and big data with emerging Hadoop - based big data and NoSQL technologies and techniques.
- Experience in optimization of Map reduce algorithm using combiners and partitioners to deliver the best results
- Extensive experience with Query big data tools likePig Latin and Hive QL.
- Hands on experience in big data ingestion tools likeFlume and Sqoop.
- Familiarity with (No-SQL) big-data storage concepts
- Familiarity with Spark, Python, Impala and Talend.
- Worked extensively with Data migration, Data Cleansing and ETL Processes
- Experience in data management and implementation of Big Data applications using HADOOP frameworks.
- Proficient in analyzing and translating business requirements to technical requirements and comfortable with all the phases of SDLC
- Optimizing management of and deriving insights from non-structured non-relational data and provide business value from content through improved information management.
- Executed software projects for Pharmaceutical, Health Care, Banking, and Telecom.
- Good communication skills, interpersonal skills, self-motivated, quick learner, team player.
- Excellent knowledge of industry standard methodologies like Software Development Life Cycle (SDLC) for Waterfall and Agile methodologies.
TECHNICAL SKILLS
Big Data: Hadoop, HDFS, Map Reduce, Pig, Hive, HBase, Flume, Sqoop, Oozie, Spark
Programming Languages: Pig Latin, HiveQL, Core Java, SQL, C, MQ Series, RPG, CL.
Database: HBase, DB2/400, MySQL
O/S: Linux, iSeries, Win 7/XP.
Processes: Confidential ’s QMS (Quality Management System).
Tools: Eclipse, HP Quality Center, Clarity, Test Director.
Information Management: Confidential Content Manager OnDemand.
PROFESSIONAL EXPERIENCE
Confidential
Big Data developer
Responsibilities:
- Involved in source system analysis, data analysis, data modeling to ETL (Extract, Transform and Load)
- Experience in managing and reviewing Hadoop log files. Experience in running Hadoop streaming jobs to process data
- Designed a data warehouse using Hive
- Handling structured and unstructured data and applying ETL processes.
- Worked extensively with Sqoop for importing and exporting the data from Relational Database systems/mainframe to HDFS. Loading data into HDFS.
- Extensively used Pig for data cleansing.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
- Developed the Pig UDF’S to pre-process the data for analysis.
- Created partitioned tables in Hive. Develop Hive queries for the analysts
- Data is processed using MapReduceand the result is stored in Hbase and displayed as per the user requirement either Pie/Bar chart or both.
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Following Data Validation and Data transformation processes in Data lake.
- Prepare Developer (JUnit) Test cases and execute Developer Testing.
- Supports and assist QA Engineers in understanding, testing and troubleshooting. Used Agile methods.
- Production Rollout Support which includes monitoring the solution post go-live and resolving any issues that are discovered by the client and client services teams.
- Performed Tier 1 production support for supported production applications until handoff to support group.
- Deliver new and complex high quality solutions to clients in response to varying business requirements as per the organizational standards set in Organizational Process Asset Library.
- Submitting metrics for projects and providing managerial reports as needed to higher management.
Environment: Hadoop, Map Reduce, Pig, Hive, HDFS, Sqoop, Flume, Java (JDK1.6), Oozie, DB2, MySQL, Eclipse, HPQC
Confidential
Team Lead
Responsibilities:
- Worked on Maintenance and enhancement projects. Provided production support for many applications and handled the issues independently.
- Good at picking up new tool/software skill and got appreciation for learning and implementing OnDemand Content Manager project AMBER.
- Worked with SAP team for Business Transformation and got exposure to SAP environment. Learned a lot about SAP and BI (Business Intelligence) while working on development and integration testing part to communicate with SAP & BI system.
- Handling AS400 communications issues while receiving customer order from Desktop applications built on Visual Basic (VB), .Net & Java.
- Establish Quality Procedure for the new team members in project and continuously monitor to ensure they meet quality goals. Acquaint them with the business of the Client and various functions of the project.
- Provided technical & investigative support for functional users on interfaces, defects, change requests.
- Key projects handled include SAP Business Transformation and Amber (using Content Manager OnDemand).
Environment: DB2/400, iSeries, RPGLE, CLLE, SQL, MQ Series, Java, Confidential Content Manager OnDemand