We provide IT Staff Augmentation Services!

Senior Etl Datastage Developer / Data Analyst / Azure Resume

3.00/5 (Submit Your Rating)

SUMMARY

  • Having around 10 years of IT experience in Data warehousing and Data management which includes application development, Administration and Implementation of various BI and Data Management Tools like Datastage, Teradata with technical expertise in implementation, Designing, operations & support functions. Good experience in working with relational databases and programming languages like Mainframe and Knowledge on Azure cloud technologies.
  • Expertise in analysis, design, implementation, testing activities and production support of Enterprise Data warehouse and Business Intelligence applications.
  • Strong understanding of Data warehouse principles and methodologies - Ralph Kimball/Bill Inman. Experienced with designing and implementing star schemas.
  • Strong knowledge on DW and BI modelling concepts - Conceptual Data Modelling, Logical Data Modelling, Physical Data Modelling.
  • Domain knowledge in Telecommunications, Insurance, health care and Airline industries.
  • Extensive experience in working with large datasets and coherent with Data Ingestion, Acquisition, Integration, Transformation and Aggregation
  • Hands-on experience in working with relational databases like Oracle, MS SQL Server, DB2 and Extensive experience in working with MPP database like Teradata. Worked on Snowflake DB as part of Data migration
  • Strong hands on experience using Teradata utilities (SQL, Fast Load, Multiload, FastExport, Tpump)
  • Familiar with Big Data technologies Hadoop, Oozie, Pig, Hive.
  • Over 7+ years of experience in design & implementation of SQL queries.
  • Over 7+ years of hands on experience on UNIX/Linux shell scripting and knowledge on Perl scripting.
  • Efficient in creating source to target mappings (ETL Mapping) from various source systems into target EDW (Enterprise Data warehouse) and documenting design specifications.
  • Experienced in Datastage Administration, set up DS environment, Installing DS from scratch, configuring client (Windows) and servers (Unix).
  • Configuring databases, creating users and defining their roles, creating environmental variables at project level.
  • Strong knowledge in Azure Cloud environment, Implementing of Azure tools like Data Factory and related tools from Azure portal.
  • Implement and maintain storage and functions in Azure resource group which are used to build pipelines.
  • Implement, monitor and maintain Security (Key Vault), Azure Storage account (blob, File, Table) for IaaS and PaaS services based on project requirements.
  • Deploy, configure and maintain compute resources like VM’s, Container using ARM template or market place template.
  • Used post deployment scripts for azure VM’s using configuration management tools like Ansible.
  • Creating Azure resource groups and granting subscription access to users
  • Knowledge on how Azure Databricks, Data lake and Data analytical engine works.
  • Used version control tools like Git for maintaining multiple versions of application code and UCD to deploy the code in production.
  • Worked on UNIX-AIX (IBM), Solaris (Oracle), and LINUX (Red Hat) Servers
  • Built shell scripts to automate the day to day tasks like source file validations, data processing, Email notifications that reduce manual tasks.
  • Hands-on experience with Job scheduling tools Autosys & TWS.
  • Experienced in design and development of common re-usable ETL components to define audit processes, Job monitoring and error logging processes.
  • Experienced in using advance Datastage plug in stages like XML input and XML output.
  • Hands on Experience on DataStage Webservice hosting and consuming.
  • Worked and build data frameworks that allow data privacy by following company standard compliance and Data Governance like PCI, PII and SOX.
  • Hands-on experience in troubleshooting production issues and performance tuning of ETL jobs
  • Familiarity in using the service management and IT governance tools (HP Open View Service Desk, HPSM, HPQC and HP ITG).
  • Strong understanding of SDLC and Agile Methodologies.
  • Worked in Agile (using Rally) environment.
  • Strong Analytical and Problem-solving Skills. Can quickly adapt to a changing environment.

TECHNICAL SKILLS

ETL Tools: IBM Web Sphere Datastage v.7.5, IBM Info Sphere Information Server 8.5, 8.7, 9.1.2, 11.1, Teradata 13, Azure Data Factory (Concepts).

Cloud Technologies: Microsoft Azure technologies.

Databases: Teradata 12,13,14, DB2, MS SQL Server 2000, Oracle 12c, 11g/10g, Mainframes, Snowflake.

Environment: IBM AIX 5.3, 6.1, Linux 6.0, 7.1,7.5, Windows Server, Azure Platform.

DBMS: Oracle, MS SQL, TOAD, DB2 Visualizer, DB2 Control center, Teradata SQL loader.

Others: HPSD Open View, HPSM, Autosys 4.5, Tivoli Workload Scheduler, HPQC, ServiceNow, Jira, Rally.

PROFESSIONAL EXPERIENCE

Confidential

Senior ETL Datastage Developer / Data Analyst / Azure

Responsibilities:

  • As a Data Engineer responsible for building a large scale, complex, distributed computing, real-time streaming high throughput and low latency analytics application using Data warehousing technologies.
  • Understanding the Business requirements, coordinating with business analysts and Core team members to get specific requirements for the application development.
  • Strong knowledge in Azure Cloud environment, Implementing of Azure tools like Data Factory and related tools from Azure portal.
  • Implement and maintain storage and functions in resource group which are used to build pipelines.
  • Implement, monitor and maintain Security (Key Vault), Storage account (blob, File, Table) for IaaS and PaaS services based on project requirements.
  • Created metadata from the given mapping documents and mapped those columns to Database.
  • Developed jobs that read the data from sources as flat files and load into Tables like MS SQL, DB2, Oracle.
  • Working on data migration project to move data from legacy systems to Snowflake Database.
  • Designed and developed jobs that read/capture real time data using Hierarchal stages to make service calls between Datastage and other real time service environments.
  • Consistently used File stage, Database stage, Transformer, Copy, Join, Funnel, Aggregator, Lookup, Filter and other processing stages to develop parallel jobs.
  • Experienced in development and debugging of jobs using stages like Peek, Head & Tail Stage, Column generator stage.
  • Generated Surrogate Keys using Triggers in Oracle Database using unique primary function to distinguish the source records and capture them if there are any rejects while loading table.
  • Developed Job Sequencers for scheduling and job flow, used loop stages for continuous run of same job.
  • Created shell scripts for file watchers, Validating the received input files and auto generated Email Notification about process status.
  • Capture the rejected data and Send notification to source with rejected reason.
  • Created intermediate tables to maintain log info about the source files, processed counts and loaded counts for future any audit purpose and maintain Primary and foreign key indexes.
  • Built data frameworks and transfer of data by making sure it secure and maintained Data Governance by following company standard compliance like PCI, PII and SOX.
  • Worked on performance tuning of various jobs to reduce the run time of a job which saved processing time of the whole job.
  • Exported and imported Datastage components and Metadata tables in different environments.
  • Performed unit and system testing, Involved in design and code review meetings on behalf of team.
  • Involved in testing along with QA to help and meet Design and coding standards as per business requirements.
  • Setup EMFT (Secure File transfer) environment to send/received files from different sources.
  • Created job Environmental user variables, parameters and shared containers.
  • Worked in Agile (using Rally) environment.

Environment: IBM Information server DataStage 11.5, Azure, Snowflake, MS SQL, DB2, Oracle 12c, Linux 7.0, EMFT file transfer, ServiceNow, Rally

Confidential

Senior Datastage Developer / Data Analyst

Responsibilities:

  • Involved in understanding of business processes and coordinated with business analysts to get specific user requirements.
  • Designed and developed jobs that extract data from the source databases using DB connectors Oracle, DB2 and Teradata.
  • Designed and developed jobs that read the data from diverse sources such as flat files, XML files, and MQs.
  • Created job parameters, parameter sets and shared containers.
  • Consistently used Transformer, Modify, Copy, Join, Funnel, Aggregator, Lookup stages and other processing stages to develop parallel jobs.
  • Experienced in development and debugging of jobs using stages like Peek, Head & Tail Stage, Column generator stage, Sample Stage.
  • Generated Surrogate Keys for composite attributes while loading the data into the data warehouse.
  • Imported Metadata from the source database. Imported metadata definitions into the repository. Exported and imported Data Stage components.
  • Extracted data from multiple sources like Mainframes and from databases like Teradata, DB2, Oracle. Made the required transformations as per business and loaded the data into target which is DB2.
  • Experienced on working and loading data to Hadoop environment using IBM BigIntegrate.
  • Worked with Big Data technologies Hadoop, Oozie, Pig, Hive.
  • Used HDFS commands to view data in Unix environment as part of analytical purpose.
  • Used Hive to view HDFS data and to perform DDL and DML operations.
  • As part of ticket analysis or business requirements used Scoop to copy data from DB environment to other.
  • Developed Job Sequencers and batches, edited the job control to have jobs run in sequence.
  • Performed troubleshooting and tuning of Data Stage Jobs for better performance.
  • Involved in creating SQL queries, performance tuning and creation of indexes.
  • Extensively used materialized views for designing fact tables.
  • Ensured that operational and analytical data warehouses are able to support all business requirements for business reporting.
  • Developed Unix shell scripts and worked on Perl scripts for controlled execution of Datastage jobs
  • Followed and Used secure data transfer process to meet the PCI compliance.
  • Extensively designed jobs in TWS scheduler for execution of the jobs in sequence
  • Participated in DataStage Design and Code reviews.
  • Worked in Agile/Scrum environment.
  • Documented ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing, prepared test data for testing and error handling

Environment: IBM Information server DataStage 9.1.1, Teradata 13.1.1, AIX 5.3, AIX 6.1, Linux 2.6.18, Oracle 10g, TWS, DB2, TOAD, SQL Plus, HPSM, Jira.

Confidential

Datastage Developer / Data Analyst

Responsibilities:

  • Extensively used DataStage Designer and Teradata to build the ETL process which pulls the data from different sources like flat files, DB2, mainframes system and does the grouping techniques in job design.
  • Developed master jobs (Job Sequencing) for controlling flow of both parallel & server Jobs.
  • Good knowledge in parameterizing the variables rather than hardcoding directly, Used Director widely for monitoring the job flow and processing speed.
  • Based on the above analysis did performance tuning for improving job processing speed.
  • Developed Autosys jobs for scheduling the Jobs. These jobs include Box jobs, Command jobs, file watcher jobs and creating ITG requests.
  • Closely monitor schedules and look into the failures to complete all ETL/Load processes within the SLA.
  • Designed and developed SQL scripts and extensively used Teradata utilities like BTEQ scripts, FastLoad, Multiload to perform bulk database loads and updates.
  • After Completing ETL activities corresponding load file will be sent to Cube team for building Cubes.
  • Used Teradata Export utilities for reporting purpose
  • Created spec doc’s for automating the manual processes.
  • Closely work with On-shore people and Business people to resolve critical issues occurred during load process.

Environment: DataStage 8.5, Linux 2.5, Oracle 10g, Teradata 13.1.1, TWS maestro, TOAD, SQL*Loader, SQL Plus, SQL, HPSM, Mercury ITG

We'd love your feedback!