We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

OH

SUMMARY

  • Proficient Bigdata Hadoop Developer with 8 years of IT experience focusing on Big Data Technologies and Design and Development of Java based enterprise applications
  • Hands - on experience designing and implementing complete big data solutions using Pig, Hive, Hbase, and Sqoop
  • Involved in troubleshooting performance issuesin Hive and Pig scripts with understanding of Joins, Group and Aggregation and it’s translation to Map Reduce jobs
  • Expertise in writing UNIX shells scripts
  • Hands-on experience in Application Development using RDBMS and Python Scripting
  • Worked on various performance optimizations using Distributed Cache for small datasets, Partitioning and Bucketing in Hive and Map-side joins
  • Involved in creating Hive tables and using HQL for data validation
  • Expert in creating Queries, Stored procedures, Functions and Triggers using SQL
  • Proven skills in testing and verifying data in HDFS and Hive tables while data ingestion is performed with Sqoop from Hive to RDBMS tables
  • Good knowledge and understanding of Hbase and basic knowledge of Java programming
  • Supported development, testing, and operations teams during new system deployments
  • Demonstrated knowledge of Software Development Life Cycle (Requirements Analysis, Design, Development, Testing, Deployment and Support)
  • Strong analytical and problem solving skills with exceptional ability to learn and master new technologies and deliver outputs in short deadlines
  • Good interpersonal skills and ability to work as part of the team

TECHNICAL SKILLS

Big Data Ecosystem: Map Reduce, Hbase, Pig, Hive, Sqoop, Zookeeper, HDFS

Programming Language: C, C++, VB, VC++, Java, Shell Programming, Python

Database: SQL Server, MySQL

IDE: Eclipse 3.2, Visual studio 2008, WINCE 6.0

PROFESSIONAL EXPERIENCE

Confidential, OH

Hadoop Developer

Responsibilities:

  • Imported all the customer specific personal data to Hadoop using Sqoop component from various relational databases like SQL
  • Created tasks for incremental load into staging tables and scheduled them to run
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive
  • Wrote Map Reduce code that will take customer related flat file as input and parse the same data to extract the meaningful (domain specific) information for further processing
  • Created Hive External tables with partitioning to store the processed data from Map Reduce
  • Implemented Hive optimized joins to gather data from different sources and run ad-hoc queries on top of them
  • Wrote Hive Generic UDF's to perform business logic operations at record level
  • Used Pig to do data transformations, event joins, filter and some pre-aggregations before storing the data onto HDFS
  • Created Hbase tables to load large sets of structured, semi-structured and unstructured data coming from HDFS and a variety of portfolios
  • Experienced in using Sequence files, RCFile, and AVRO file formats
  • Developed UDF sin Java as and when necessary to use in Pig and Hive queries

Confidential, NY

Hadoop Developer

Responsibilities:

  • Worked on analyzing Hadoop cluster and different Big Data analytic tools including Pig, Hive,HBase and Sqoop
  • Involved in importing and exporting data into Hive using Sqoop
  • Solved performance issues in Hive and Pig scripts with apt understanding of joins, groups, and aggregation and it’s translation to Map Reduce jobs
  • Wrote Hive jobs to parse the logs and structure them in a tabular format to facilitate effective querying on the log data
  • Involved in creating Hive tables, loading with data and writing Hive queries that will run internally in Map Reduce way
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
  • Experienced in managing and reviewing the Hadoop log files
  • Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS
  • Load and Transform large sets of structured and semi structured data
  • Responsible to manage data coming from different sources
  • Involved in creating Hive Tables, loading data and writing Hive queries

Confidential

Java Developer

Responsibilities:

  • Involved in analysis and design phases of Software Development Life Cycle (SDLC)
  • Created detailed design document, use cases and class diagrams using UML and Rational Rose
  • Implemented/developed core J2EE patterns: MVC, DI/IOC, DAO, Interceptors, Business Delegate, Service Locator, Singleton for the enterprise applications
  • Used Spring Framework to implement the MVC design pattern in the application
  • Used My-Eclipse IDE as development environment, designed, developed and deployed EJB Components (EAR) on WebSphere
  • Developed test cases using Junit and tested the application
  • Used Log4j as the logging framework
  • Involved in installation and setup of the class paths, maintenance and troubleshooting while deployment of software during development and QA phases
  • Involved in unit integration, bug fixing, design reviews, code walk-through, client interaction and team support

Confidential

Java Developer

Responsibilities:

  • Involved in Requirements gathering and analysis, design, development, integration and deployment
  • Used JavaScript to perform checking and validations
  • Designed and developed DAO layer using Spring and Hibernate apart from using Criteria API
  • Created/generated Hibernate classes and configured XML apart from managing CRUD operations (insert, update, and delete)
  • Used log4j for logging messages
  • Developed the classes for Unit Testing using JUnit
  • Developed Business components using Spring Framework and database connections using JDBC

Confidential

Software Engineer

Responsibilities:

  • Developed serial protocol for interfacing GSM Modem/Printer or slave processor to the machine
  • Developed firmware for reading and writing with RF tags, RF contactless smart card security encryption
  • Wrote code for communication protocol such as TCP/IP
  • Designed firmware for reading and writing different RAM (Static, NV etc.)
  • Developed serial/Ethernet command for uploading and downloading data to machine
  • Developed application to identify biometric user based on VC++ and SQL
  • Developed software for reading writing smartcard tracks in VC++

We'd love your feedback!