We provide IT Staff Augmentation Services!

Hadoop Engineer Resume

2.00/5 (Submit Your Rating)

Warren, NJ

SUMMARY

  • 8 Years of experience in IT industry comprises of development, maintenance, and support projects in Big Data, Java, Data Warehouse and Mainframe Technologies.
  • Over 2.5 years of extensive experience in Bigdata analytics including Hadoop Map Reduce, HDFS, Hive, Sqoop, HBase, Pig, Flume, Impala, etc.
  • Hands on experience in installing, configuring, and using Apache Hadoop ecosystem components.
  • Good experience in Hbase NoSQL database using API
  • Good experience in UNIX/Linux and Shell Scripts.
  • Experienced in full Software Development Life Cycle (SDLC) Process.
  • Good Experience in web services.
  • Good understanding of HDFS Designs, Daemons, federation and HDFS high availability (HA).
  • Experienced in Hive Tables Design, Pig Latin, loading the data into Hive tables.
  • Good experience in core Java.
  • Good experience in Data Analysis, ETL Development, Data warehousing.
  • Strong Experience in analysis, design, development, implementation and troubleshooting of data warehouse application using ETL tools like Informatica power center 9.1/8.x/7.x, SQL Server Integration services (SSIS) and Data Transformation service (DTS).
  • Good experience on Agile and scrum methodologies.
  • Good knowledge on Financial Investments, Securities, and experienced in Money Movement applications (Fund Transfers - EFT, WIRE).
  • Excellent technical expertise on all mainframe technologies like COBOL, JCL, DB2, CICS, MQ Series and Data warehousing tool like Ab-Initio.
  • Good Knowledge on XML basics, XSD, XSL, and service oriented architecture.
  • Experienced in working Agile working environment.
  • Co-ordinate work requests among team members, sizing, work allocation, status reporting, defect tracking, change management, issues clarification.
  • Experienced in System study, analyzing of business requirements, technical design, coding, unit testing, integration testing, system testing and implementation.

TECHNICAL SKILLS

Big Data Eco System: HDFS, HBase, Hadoop MapReduce, Hive, Pig, Cassandra, Flume, Sqoop. SPARK, Oozie, NOSQL.

Languages: COBOL, JAVA, PL/SQL, C, C++

Methodologies: Agile, V-model (Verification & Validation Model).

Database: Oracle 10g, DB2, My SQL, No SQL.

IDE/Testing Tools: Eclipse

Operating System: Windows, UNIX, and Linux

Scripts: Shell scripting, Java Script

Others: VSAM, Change Man, TWS, Expediter, QC, MS-Office, JIRA, Share Point, Visio.

PROFESSIONAL EXPERIENCE

Confidential, Warren, NJ

Hadoop Engineer

Responsibilities:

  • Involved in review of functional and non-functional requirements.
  • Analyzed the source tables.
  • Designed and created the stage tables and main tables.
  • Implemented optimization and performance tuning in Hive and Pig.
  • Designed the Hive tables as per business requirements.
  • Created the various bulk load and incremental load scripts.
  • Imported data using Sqoop to load data from DB2 site to Hadoop on regular basis.
  • Wrote and implemented Pig UDF to preprocess the data and use it for analysis.
  • Installed and configured Pig for ETL jobs.
  • Created Hive tables and was working on them using Hive QL for data analysis in order to meet the business requirements.
  • Implemented partitioning and bucketing in Hive.
  • Extensively used Pig for data cleansing.
  • Run Ad-Hoc query through PIG Latin language, Hive.
  • Exported and analyzed data into relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Experienced with versioning, change control, and Problem management.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Hbase, Sqoop, flume, Java, Eclipse, Spark, Scala, SQL, DB2, Pig, Sqoop, Linux, Cloudera.

Confidential, Florida

Hadoop Developer

Responsibilities:

  • Involved in design and development phases of software Development Life Cycle (SDLC) using Scrum methodology.
  • Developed data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer data and purchase histories into HDFS into for analysis.
  • Implemented optimization and performance tuning in Hive and Pig.
  • Developed job flows in Oozie to automate the workflow for extraction of data from warehouse and weblogs.
  • Used Pig as ETL tool to do transformations, event joins, filters and some pre-aggregations before storing the data into HDFS.
  • Optimizing Map reduce code, pig scripts, user interface analysis, performance tuning and analysis.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting on the dashboard.
  • Loaded the aggregated data onto DB2 for reporting on the dashboard.
  • Implemented Partitioning and bucketing in Hive.
  • Experienced in managing and reviewing Hadoop log files.
  • Extensively used Pig for data cleansing.
  • Configured Flume to extract the data from web server output files to load into HDFS.
  • Experienced in working Agile working environment as team player.
  • Analysis, design, development, implementation and troubleshooting using ETL
  • Using Scrum call for reporting and updating reprocess of the project.

Environment: JDK1.6, Cent OS, HDFS, Map-Reduce, Java, Eclipse, Hive, Pig, Sqoop, Spark, Kafka, Flume, Zookeeper, Oozie, DB2, Mysql, ETL Datawarehouse and HBase.

Confidential, New York, NY

Programmer Analyst

Responsibilities:

  • Analyzed new business requirement documents and created Functional Specification documents and high level design documents.
  • Created process flow diagrams for the new enhancement request from business and reviewed with business, operations team, IT team and compliance team to make sure the requirements are correctly captured and in right direction.
  • Conducting code walkthroughs, reviewing the code changes done by offshore team and verify the results before implementations.
  • Development of Cobol/DB2/MQ series batch programs that call the stored procedures to get the result set data from other systems and build the message queue with messages.
  • Coordinating with the offshore team to get the work done as per the business requirements.
  • Providing support for the unit testing, Regression testing and User Acceptance testing in order to make sure the 100% bug free outcome of the new development.
  • Worked on web services to receive request onto mainframes and send back the response back to front end applications.
  • Tracking all the issues in Rational-Clear Quest with the estimates, efforts and resolution steps.
  • Creating implementation run book and plan document and distribute among the teams before the implementations and perform post deployment checkout after successful implementation.

Environment: IBM 3090, Z/OS, VS-COBOL II, COBOL, CICS, JCL, DB2, TSO/ISPF, INTERTEST, WSF, DB2 SPUFI, QMF, CHANGEMAN, MQ Series, VSAM, FILEAID and CICS Web Services.

Confidential

SQL Developer

Responsibilities:

  • Backing up, restoring system and others database as per requirements and also scheduled those backups.
  • Deploying and scheduling report subscriptions to generate daily, weekly, monthly and quarterly reports including current status.
  • Developing stored procedures, triggers, view and adding/changing tables for data load and transformation and data extraction.
  • Used SQL server profiler to trace the slow running queries and tried to optimize SQL queries for improved performance and availability.
  • Used SQL Profiler for troubleshooting monitoring and optimization of SQL Server and non-production database code as well as T-SQL code from developers and QA.
  • Involved in optimizing code and improving efficiency in database including re-indexing, recompiling stored procedures and performing other maintenance tasks.

Environment: SQL Server 2008, Access, SSIS SSRS, T-SQL, PL/SQL, BIDS, SQL Server Management Studio, Visual Studio 2012, OLAP, SVN, JIRA.

Confidential

Project Executive/ IT Analyst

Responsibilities:

  • Conception and development of an assessment service database reporting tool using Java and JSP.
  • Design of the Graphic User Interface of the Intracon Asia Pacific website.
  • Assistance in IT assessment projects in the area of “Imaging and Printing” for large enterprise customers from different markets segments.
  • Review of output environment recommendations with customers.
  • Generation of project reports (MS PowerPoint) statistical analysis.
  • Maintenance and management of the assessment services database and database reporting tool.
  • Assistance in project management.

Environment: MySQL, Java, Visual Basic Script, PHP, MS Visio, MS PowerPoint.

We'd love your feedback!