We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

NJ

SUMMARY

  • 6+ years of experience in various phases of projects in developing, enhancing and maintaining applications in industry verticals like Insurance, Banking and Retail using Informatica Power Center.
  • Around 3 years of Expertise in core Hadoop and Hadoop technology stack which includes HDFS, Sqoop, Hive, HBase, Oozie, Pig, Flume and Map Reduce programming.
  • 1 year of experience in Informatica Master data Management ( MDM HUB)
  • Configuring and maintaining various components of the MDM Hub including the schema, staging and landing tables, configuring base objects, Look ups, Hierarchies, display queries, put queries and query groups.
  • Migration of Informatica Mappings/Sessions/Workflows from Dev, QA to Prod environments.
  • Involved in INFA Admin repository upgrade Activities and INFA Admin support Activities.
  • Proficiency in developing SQL with various relational databases like Oracle, Teradata.
  • Knowledge in Full Life Cycle development of Data Warehousing.
  • Performed Informatica upgrade from V8.6.1 to 9.0.1 and 9.1.0.
  • Creation and maintenance of Informatica users and privileges.
  • Experience with dimensional modelling using star schema and snowflake models.
  • Understand the business rules completely based on High Level document specifications and implements the data transformation methodologies.
  • Created UNIX shell scripts to run the Informatica workflows and controlling the ETL flow.
  • Strong with relational database design concepts.
  • Extensively worked with Informatica performance tuning involving source level, target level and map level bottlenecks.
  • Vast experience in Designing and developing complex mappings from varied transformation logic like Unconnected and Connected lookups, Source Qualifier, Router, Filter, Expression, Aggregator, Joiner, Update Strategy etc
  • Ability to meet deadlines and handle multiple tasks, decisive with strong leadership qualities, flexible in work schedules and possess good communication skills.

TECHNICAL SKILLS

ETL Tools: Informatica PowerCenter 9.1/8.6,Informatica MDM Hadoop Ecosystem components

Database: Teradata 14.0, Oracle 11g/10g/9i

DB Tools: SQL*Plus, TOAD

Environment: Windows 2000/2008

Languages: SQL, PL/SQL

Reporting Tools: Actimize

Other Tools/Utilities: HP Quality center, Autosys, oozie

PROFESSIONAL EXPERIENCE

Confidential, NJ

Hadoop Developer

Responsibilities:

  • Understand business needs, analyze functional specifications and converting the prototype into Hadoop Ecosystem components.
  • Involved in Extracting, loading Data from RDBMS to Hive using Sqoop.
  • Tested raw data, executed performance scripts and also shared responsibility for administration of Hadoop, Hive and Pig
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Involved in troubleshooting the issues, errors reported by cluster monitoring software provided by Cloudera Manager
  • Created HBase tables to store variable data coming from different portfolios.
  • Implemented a script to transfer information from Teradata to HBase using Sqoop.
  • Involved in loading data from UNIX file system to HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Involved in loading data from LINUX file system to HDFS.
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.

Environment: HDFS, Hadoop MapReduce, Zookeeper, Hive, Pig, Sqoop, Oozie, Cloudera CDH - 4, HUE, Flume Impala, Toad.

Confidential, NJ

ETL Developer

Responsibilities:

  • Worked with Business Analyst and Analyzed specifications and identified source data needs to be moved to data warehouse, Participated in the Design Team and user requirement gathering meetings.
  • Worked on Informatica - Repository Manager, Designer, Workflow Manager & Workflow Monitor.
  • Involved in discussing Requirement Clarifications with multiple technical and Business teams.
  • Performed Informatica upgrade from V9.1 to 9.5.
  • Creation and maintenance of Informatica users and privileges.
  • Migration of Informatica Mappings/Sessions/Workflows from Dev, QA to Prod environments.
  • Documented the LDAP configuration process and worked closely with Informatica Technical support on some of the issues.
  • Fixing all the workflows failure in unit testing and system testing.
  • Scheduling all the ETL workflows for the parallel run comparison.
  • Involved in preparing the migration list inventory.
  • Involved in requirement gathering for redesign candidates
  • Worked along with the Informatica professional to resolved Informatica upgrade issue.
  • Monitoring the disk space issue and cleaning up the unwanted logs periodically.
  • Worked with BA in the QA phase of testing.
  • Worked on Informatica Schedulers to schedule the workflows.
  • Scheduled batch jobs using Autosys to run the workflows.
  • Extensively involved in ETL testing, Created Unit test plan and Integration test plan to test the mappings, created test data. Use of debugging tools to resolve problems.
  • Used workflow monitor to monitor the jobs, reviewed error logs that were generated for each session, and rectified any cause of failure.

Environment: Informatica Power Center 9.1/8.6, Oracle 11g, PL/SQL, Autosys, SQL, Teradata, SQL* LOADER, TOAD, Shell Scripting

Confidential, NJ

ETL Informatica developer

Responsibilities:

  • Involved in all phases of SDLC from requirement gathering, design, development, testing, Production, user training and support for production environment.
  • Create new mapping designs using various tools in Informatica Designer like Source Analyzer, Warehouse Designer, Mapplet Designer and Mapping Designer.
  • Develop the mappings using needed Transformations in Informatica tool according to technical specifications
  • Created complex mappings that involved implementation of Business Logic to load data in to staging area.
  • Used Informatica reusability at various levels of development.
  • Developed mappings/sessions using Informatica Power Center 8.6 for data loading.
  • Performed data manipulations using various Informatica Transformations like Filter, Expression, Lookup (Connected and Un-Connected), Aggregate, Update Strategy, Normalizer, Joiner, Router, Sorter and Union.
  • Developed Workflows using task developer, Worklet designer and workflow designer in Workflow manager and monitored the results using workflow monitor.
  • Building Reports according to user Requirement.
  • Extracted data from Oracle and SQL Server then used Teradata for data warehousing.
  • Implementedslowly changing dimensionmethodology for accessing the full history of accounts.
  • Write Shell script running workflows in UNIX environment.
  • Optimizing performance tuning at source, target,mapping and session level.
  • Participated inweeklystatus meetings, and conducting internal andexternal reviews as well as formal walk through among various teams and documenting the proceedings.

Environment: Informatica 8.6 .1,Oracle 11g, HP-UX.

Confidential

ETL Informatica developer

Responsibilities:

  • Involved in Business Requirements analysis and design, prepared and technical design documents.
  • Used Erwin for logical and Physical database modeling of the staging tables, worked with the Data Modeler and contributed to the Data Warehouse and Data Mart design and specifications.
  • Developed technical design specification to load the data into the data mart tables confirming to the business rules.
  • Involved in design and development of complex ETL mappings and stored procedures in an optimized manner.
  • Cleansed the source data, extracted and transformed data with business rules, and built reusable components such as Mapplets, Reusable transformations and sessions etc.
  • Involved in loading the data from Source Tables to ODS (Operational Data Source) Tables using Transformation and Cleansing Logic using Informatica.
  • Developed complex Informatica mappings to load the data from various sources using different transformations like source qualifier, connected and unconnected look up, update Strategy, expression, aggregator, joiner, filter, normalizer, rank and router
  • Developed mapplets and worklets for reusability.
  • Developed workflow tasks like reusable Email, Event wait, Timer, Command, Decision.
  • Implemented partitioning and bulk loads for loading large volume of data.
  • Used Informatica debugging techniques to debug the mappings and used session log files and bad files to trace errors occurred while loading
  • Involved in performance tuning of mappings, transformations and (workflow) sessions to optimize session performance.
  • Created Materialized views for summary tables for better query performance.
  • Implemented weekly error tracking and correction process using Informatica.
  • Developed Documentation for all the routines (Mappings, Sessions and Workflows).
  • Creating Test cases and detailed documentation for Unit Test, System, Integration Test and UAT to check the data quality.

Environment: Informatica Power Center 8.6/9.0.1, Oracle 10g, UNIX (AIX), WINSQL, Windows 7, Flat files, MS SQL Server 2008, MS-Access, Autosys, Ultra Edit.

Confidential

Support Analyst

Responsibilities:

  • Developed standard and re-usable mappings and mapplets with various transformations like expression, aggregator, joiner, source qualifier, filter, lookup, stored procedure and router.
  • Involved in enhancement for the COBOL coding and promoting the code in to the production systems.
  • Involved in scheduling jobs in scheduling tools.
  • Coding COBOL, JCL, DB2 programs
  • Handling Change Requests.
  • Handling Ad-hoc requests.
  • Involved in automated and manual FTP.
  • Review of Programs
  • Code walkthrough to ensure the programs are in accordance to ASA standards
  • Enhancement, Testing and Documentation of projects.
  • Handling failed jobs in production environment within SLA.

Environment: Mainframe, DB2,JCL,Oracle 10g, SQL, Informatica 8.6, Quality Center.

We'd love your feedback!