We provide IT Staff Augmentation Services!

Datastage Developer/hadoop Developer Resume

2.00/5 (Submit Your Rating)

Bentonville, AR

SUMMARY:

  • 10 years+ of experience in designing and managing polyglot persistence systems that deal with petabytes of data End - To-End Data Warehousing/ETL/ELT/Integration/Mapping solutions .
  • Worked for Confidential .
  • Extensive experience with Analysis, Design, Development, Implementation and Testing of Data Warehousing and Business Intelligent Projects for Retail, Banking, Insurance, Telecommunication and manufacturing industries for big data platforms and analytic solutions.
  • Extensive experience in developing strategies for Extraction, Transformation and Loading data from various sources into Data Warehouse and Data Marts using DataStage.
  • Over 5 years of Experience in writing SQL and PL/SQL programs which includes DDL, DML, Views, Materialized Views, Indexes, Packages, Procedures, Triggers and Functions.
  • Hands-on DWH designing and implementing best practices in Data Modeling, Data Architecture, Data Quality and Data Profiling with strong client needs assessment talents.
  • A leader with mentoring capabilities for DWBI development professionals, Data Modelers and the proven ability to function as a respected "thought leader" among peer groups inside and outside the enterprise.
  • Having extensive experience in Data Integration and Migration using IBM Infosphere DataStage(9.1), Quality stage, SSIS, Oracle, Teradata, DB2, SQL and Shell script along with technical certifications in ETL development from IBM and Cloudera .
  • Extensive experience on Agile unified process for the project deliveries and extensive work experience on data modeling for ETL projects.
  • Good experience in Designing, building and administering the Infosphere datastage servers like installation and Tuning and non root use configuration for environments.
  • Very good in customer facing and requirements gathering and experienced with high level and Low Level Designs (HLD & LLD) documents.
  • Extensive experience on Agile unified process for the project deliveries and extensive work experience on data modeling for ETL projects.
  • Well exposure with functional point analysis while Estimation, Planning and Design in DataStage platform with Implementation.

TECHNICAL SKILLS:

ETL Tools: IBM Infosphere Datastage 11.5, SSIS, Informatica, Talend.

Reporting tools: Microstrategy, Business objects, Cognos

Database: Oracle 10g, DB2, Teradata, MS SQL server, Cassandra

Languages: C, Java, Scala, SQL/PLSQL

Data Modeling tools: ER-Win, Visio, Information Analyzer, Business Glossary

Operating System: Sun Solaris, IBM - AIX, UNIX, Linux

Scripts: JavaScript, Shell Scripting

Big Data Ecosystem: Spark (Core, Spark SQL), Hadoop Ecosystem (Hive, Presto, Sqoop, Zookeeper, Kafka, Oozie etc.)

Scheduling tools: Autosys, CA7, Tivoli

PROFESSIONAL EXPERIENCE:

DataStage Developer/Hadoop developer

Confidential, Bentonville, AR

Responsibilities:

  • Involved in the data ingestion process through datastage to load data into HDFS from Mainframes, Greenplum,teradata, DB2.
  • Participated in discussions with the business users and documented the requirements and scope of the project.
  • Created ETL guidelines document which involves coding standards, naming conventions for development and production support log and root cause analysis documents for troubleshooting DataStage jobs.
  • Designed and Developed Data stage Jobs to Extract data from heterogeneous sources, Applied transform logics to extracted data and Loaded into Data Warehouse Databases.
  • Created Datastage jobs using different stages like Transformer, Aggregator, Sort, Join, Merge, Lookup, Data Set, Funnel, Remove Duplicates, Copy, Modify, Filter, Change Data Capture, Change Apply, Sample, Surrogate Key, Column Generator, Row Generator, Etc.
  • Extensively worked with BDFS, sequential file, Complex flat flie, dataset and fileset.
  • Worked with Json format files by using XML, Hierarchical Data stage stages.
  • Extensively used Parallel Stages like Row Generator, Column Generator, Head, and Peek for development and de-bugging purposes.
  • Used the Data Stage Director and its run-time engine to schedule running the solution, testing and debugging its components, and monitoring the resulting executable versions on ad hoc or scheduled basis.
  • Developed complex store procedures using input/output parameters, cursors, views, triggers and complex queries using temp tables and joins.
  • Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance.
  • Created the dataflow and erro handling techniques using job sequences.
  • Maintained Data Warehouse by loading dimensions and facts as part of project. Also worked for different enhancements in FACT tables.
  • Analyzed the job performance and monitored resources performance with capacity planning.
  • Performed performance tuning of the jobs by interpreting performance statistics of the jobs developed.
  • Documented ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing, prepared test data for testing, error handling and analysis.
  • Loaded the data into Greenplum and SAP HANA database by using GPconnector and
  • Coded generic reusable DataStage components for loading and unloading data to and from Teradata.
  • Have experience in migrating DataStage jobs from 8.1 and 8.5 versions to 9.1.
  • Experience in working with Remedy7 Tool for Change Controls and managing incident tickets created by customers.
  • Writing UNIX shell scripts by using SFTP to load data from external sources to UNIX box and then load into the HDFS
  • Write UNIX script to load data from Greenplum temp schema to production schemas
  • Worked on the Datastage 11.5 POC to load the data in to hadoop HDFS and hive.
  • Created best practices document for writing ETL jobs to load the data into teradata and Greenplum databases.

Environment: Hadoop, Hive, IBM Big integrate, Infosphere Data stage 11.5, Teradata, Greenplum, SAP Hana, Oracle, DB2and Linux CA7.

ETL Developer

Confidential, Chicago. IL

Responsibilities:
  • Involved in the business requirements gathering and to Analyze source system data to determine ETL solution that will meet business requirements by using Infosphere.
  • Involved in Technical Land scape review meetings. Participated in ETL frame work designing and Patterns, Conceptual data modeling, Physical data modeling and Physical design.
  • Worked extensively on Dataflow and control flow jobs while Converting SSIS ETL jobs to Datastage.
  • Created Logical and Physical data models using Erwin, Involved in architecture definition document, HLD and LLD design and mapping documents and Proof of concepts.
  • Design and Develop Error Framework, Audit Framework and ETL jobs to load data from sources to oracle target and sequential files.
  • Facilitated design workshops with key stake holders (including the in house Enterprise Architects Team and Solution Engineers) and created business specification documents (BSD) that were used by the MDM development team.
  • Implemented query extension, to override the address and contact method queries to append them with UR to improve the performance further
  • Define project scope, goals and deliverables that support business goals in collaboration with senior management and clients.
  • Designed and Developed Parallel jobs using various stages includes Join, Sort, Lookup, Modify, Aggregate, Transformer, Funnel, Copy, SCD, Dataset, Oracle Enterprise, Oracle Connector, Sequential and Peek stages.
  • Ensure that source to target transformations maintain data integrity and Code-Reviews & Testing.
  • Implemented oracle Exadata’s hybrid Columnar compression to save storage space.
  • Used Exadata Smart Flash cache and Logging techniques to effectively to speed up the retrieval of the data
  • Liaise with customers on an ongoing basis with excellent project management skills.
  • Involved in Performance tuning of DataStage jobs and Oracle queries.
  • Written UNIX shell scripts to SFTP files, wait for file activities and to run the jobs from shell script.

Environment: IBM Infosphere Datastage, SSIS,MDM, SAP, Oracle Exadata, Erwin, Clear-case, Jasper and Linux.

Datawarehouse Architect /Developer

Confidential, Mt. Laurel, NJ

Responsibilities:
  • Estimate the resources and participants needed to achieve project goals. Set and Work continually to manage project expectations with onshore-people.
  • Involved in the designing and implementing the data vault modeling.
  • Ensuring support for leading solution design and proposal presentations for Effective Module Design
  • Involved in scalability and performance related design aspects in Database, ETL and Reporting
  • Involved in the business requirements gathering and to analyze source system data to determine ETL solution that will meet business requirements by using Infosphere.
  • Define project scope, goals and deliverables that support business goals in collaboration with senior management and clients.
  • Ensure that source to target transformations maintain data integrity and Code-Reviews & Testing.
  • Involved in scalability and performance related design aspects in Database, ETL and Reporting
  • Liaise with customers on an ongoing basis with excelent project management skills. Estimate the resources and participants needed to achieve project goals. Set and Work continually to manage project expectations with onshore-people.
  • Ensuring support for pre-sales, leading solution design and proposal presentations for Effective Module Design

Environment: IBM Infosphere Datastage, Information Analyzer, Quality stage, MDM, Oracle, XML, JAVA webservices,VSS, QlikView, SSIS, ERWIN and Linux.

Datastage Developer/ETL Lead

Confidential, Charlotte, NC

Responsibilities:
  • As Technical lead/Designer, assisted in designing and developing the MLR data flow from oneclaim database and referencing with MLR data which they are getting from Business process server queue and loading in to ODS and then by applying the ETL business rules and loading into RDW. RDW data is standardized and integrated for smart decission making through COGNOS Reports.
  • Involved in architecture definition document and design documents.

Environment: Datastage, Informatica, visio, Oracle, COGNOS, Business Process server.

Datawarehouse Architect / Technical Leader

Confidential, Englewood, CO

Responsibilities:
  • As a Datawarehouse Architect, worked in Designing and developing the WELL Permits data flow into WIP25 Datawarehouse by applying the ETL strategy which is to maintain the WELL Data information for both permitted and non-permitted wells. WIP25dw Data is validated, standardized and integrated for smart decision making for Geological analysis
  • Liaise with customers on an ongoing basis. Estimate the resources and participants needed to achieve project goals. Set and continually manage project expectations with team members.
  • Involved in architecture definition document, Macro and Micro design documents.

Environment: IBM Infosphere Datastage, Business glossary, Meta broker, Fast track Oracle, Filenet, XML, Teradata, SQL server, GECS

BI/ETL Technical lead and Designer

Confidential, Detroit, MI

Responsibilities:
  • As a Technical Lead for CARDR, Involved in designing and coding through the DataStage jobs which is from CARDM data- CARDR database.

Environment: IBM DataStage 7.5.2, Microstrategy, DB2, Starteam and AIX

ETL Designer

Confidential, Colorado, Englewood

Responsibilities:
  • Designing the datastage processes for OMD which is for operation metrics for Confidential, Environment: Ascential DataStage 7.5.1A (PX), DB2, SAP BOXI and BODS, XML, PL/SQL, AIX, F-Secure SSH

ETL Developer

Confidential, Irving, TX

Responsibilities:
  • As an ETL Senior Designer, Involved in Network Operations reporting and analysis of installation, maintenance, Provisioning, Repair and customer experience information is currently supported through a series of legacy applications.

Environment: Ascential DataStage 7.5.2, DB2 UDB, PL/SQL, AIX, F-Secure SSH Client

We'd love your feedback!