Hadoop Developer Resume
0/5 (Submit Your Rating)
OH
SUMMARY
- Proficient Bigdata Hadoop Developer with 8 years of IT experience focusing on Big Data Technologies and Design and Development of Java based enterprise applications
- Hands - on experience designing and implementing complete big data solutions using Pig, Hive, Hbase, and Sqoop
- Involved in troubleshooting performance issuesin Hive and Pig scripts with understanding of Joins, Group and Aggregation and it’s translation to Map Reduce jobs
- Expertise in writing UNIX shells scripts
- Hands-on experience in Application Development using RDBMS and Python Scripting
- Worked on various performance optimizations using Distributed Cache for small datasets, Partitioning and Bucketing in Hive and Map-side joins
- Involved in creating Hive tables and using HQL for data validation
- Expert in creating Queries, Stored procedures, Functions and Triggers using SQL
- Proven skills in testing and verifying data in HDFS and Hive tables while data ingestion is performed with Sqoop from Hive to RDBMS tables
- Good knowledge and understanding of Hbase and basic knowledge of Java programming
- Supported development, testing, and operations teams during new system deployments
- Demonstrated knowledge of Software Development Life Cycle (Requirements Analysis, Design, Development, Testing, Deployment and Support)
- Strong analytical and problem solving skills with exceptional ability to learn and master new technologies and deliver outputs in short deadlines
- Good interpersonal skills and ability to work as part of the team
TECHNICAL SKILLS
Big Data Ecosystem: Map Reduce, Hbase, Pig, Hive, Sqoop, Zookeeper, HDFS
Programming Language: C, C++, VB, VC++, Java, Shell Programming, Python
Database: SQL Server, MySQL
IDE: Eclipse 3.2, Visual studio 2008, WINCE 6.0
PROFESSIONAL EXPERIENCE
Confidential, OH
Hadoop Developer
Responsibilities:
- Imported all the customer specific personal data to Hadoop using Sqoop component from various relational databases like SQL
- Created tasks for incremental load into staging tables and scheduled them to run
- Created reports for the BI team using Sqoop to export data into HDFS and Hive
- Wrote Map Reduce code that will take customer related flat file as input and parse the same data to extract the meaningful (domain specific) information for further processing
- Created Hive External tables with partitioning to store the processed data from Map Reduce
- Implemented Hive optimized joins to gather data from different sources and run ad-hoc queries on top of them
- Wrote Hive Generic UDF's to perform business logic operations at record level
- Used Pig to do data transformations, event joins, filter and some pre-aggregations before storing the data onto HDFS
- Created Hbase tables to load large sets of structured, semi-structured and unstructured data coming from HDFS and a variety of portfolios
- Experienced in using Sequence files, RCFile, and AVRO file formats
- Developed UDF sin Java as and when necessary to use in Pig and Hive queries
Confidential, NY
Hadoop Developer
Responsibilities:
- Worked on analyzing Hadoop cluster and different Big Data analytic tools including Pig, Hive,HBase and Sqoop
- Involved in importing and exporting data into Hive using Sqoop
- Solved performance issues in Hive and Pig scripts with apt understanding of joins, groups, and aggregation and it’s translation to Map Reduce jobs
- Wrote Hive jobs to parse the logs and structure them in a tabular format to facilitate effective querying on the log data
- Involved in creating Hive tables, loading with data and writing Hive queries that will run internally in Map Reduce way
- Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
- Experienced in managing and reviewing the Hadoop log files
- Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS
- Load and Transform large sets of structured and semi structured data
- Responsible to manage data coming from different sources
- Involved in creating Hive Tables, loading data and writing Hive queries
Confidential
Java Developer
Responsibilities:
- Involved in analysis and design phases of Software Development Life Cycle (SDLC)
- Created detailed design document, use cases and class diagrams using UML and Rational Rose
- Implemented/developed core J2EE patterns: MVC, DI/IOC, DAO, Interceptors, Business Delegate, Service Locator, Singleton for the enterprise applications
- Used Spring Framework to implement the MVC design pattern in the application
- Used My-Eclipse IDE as development environment, designed, developed and deployed EJB Components (EAR) on WebSphere
- Developed test cases using Junit and tested the application
- Used Log4j as the logging framework
- Involved in installation and setup of the class paths, maintenance and troubleshooting while deployment of software during development and QA phases
- Involved in unit integration, bug fixing, design reviews, code walk-through, client interaction and team support
Confidential
Java Developer
Responsibilities:
- Involved in Requirements gathering and analysis, design, development, integration and deployment
- Used JavaScript to perform checking and validations
- Designed and developed DAO layer using Spring and Hibernate apart from using Criteria API
- Created/generated Hibernate classes and configured XML apart from managing CRUD operations (insert, update, and delete)
- Used log4j for logging messages
- Developed the classes for Unit Testing using JUnit
- Developed Business components using Spring Framework and database connections using JDBC
Confidential
Software Engineer
Responsibilities:
- Developed serial protocol for interfacing GSM Modem/Printer or slave processor to the machine
- Developed firmware for reading and writing with RF tags, RF contactless smart card security encryption
- Wrote code for communication protocol such as TCP/IP
- Designed firmware for reading and writing different RAM (Static, NV etc.)
- Developed serial/Ethernet command for uploading and downloading data to machine
- Developed application to identify biometric user based on VC++ and SQL
- Developed software for reading writing smartcard tracks in VC++
