We provide IT Staff Augmentation Services!

Etl Architect Resume

5.00/5 (Submit Your Rating)

Vienna, VA

SUMMARY:

  • 6 years of IT experience includes Enterprise Data Warehousing, ETL, Data Modeling and Data Analysis.
  • Business Intelligence: Extensive experience in building ETL solutions using various ETL tools like IBM DataStage, Talend. Created and customized MicroStrategy reports.
  • Expertise in IBM CDC and IIDR 11.3.3. Extensive experience in DataStage including development, Administration, Upgrade and production support.
  • Worked on various versions of DataStage from 8.5 to 11.3. Also involved in upgrade process from version 8 to 9 and later version 9.7 to 11.3.
  • Performance enhancement: Fixed production issues and bugs. Improved the performance of the ETL process in production by 40%
  • Database expertise with Oracle, DB2, SQL Server, Teradata, MS Access. Built complex stored procs and triggers in various databases to perform ELT operations at Enterprise level.
  • Cloud integration: Designed and implemented the integration of Salesforce and IBM DataStage using a 3rd Party integration tool, Jitterbit. Also achieved near - real time data warehousing using JitterBit and Hierarchy Stage in DataStage.
  • Experience in providing trainings on DataStage and QualityStage
  • Test automation and task integration between SalesForce, ETL and MSTR.
  • Onsite, Offsite and Offshore project execution and management for clients based out of USA.
  • Team lead: Led a team of 12 developers (onsite and offsite)
  • Standard practices: Implemented standard practices for all Business and Technical solutions with proper documentation.
  • Industry Experience: Domain knowledge of Retail, Media and Ad Sales, Finance, Insurance and Education.

TECHNICAL SKILLS:

Data Warehousing Tools: IBM Datastage 11.x/9.x/, IBM InfoSphere Data Replication 11.x CDC, Quality Stage 8.0, Information Analyser 11.3.1, Quick Report, MicroStrategy 9.4.1

Databases: Normalized, De-normalized, Star, Dimension Modeling, SCD and Snowflake SchemaOracle 11g/10g/9i/8i, SQL Server 2000/2005/2008 , UDB DB2, MS Access 7.0, Teradata V2R6/R12

Internet Technologies: ASP, JSP, ASP.NET, XML/XSL, HTML

Programming Languages: Java, VBS, C/C++, Unix Shell scripting, AWK, SQL, SQL * Plus, PL/SQL, VSS, ERwin4.1, IBM Rational Rose Clear Case& Clear Quest, MS Visio, MS Project, FileZilla, Putty, Toad, SQL*Loader, WinSCP, JitterBit, Windows, LINUX, UNIX (Solaris, AIX, HP-UX)

PROFESSIONAL EXPERIENCE:

Confidential, Vienna, VA

ETL Architect

Responsibilities:

  • Increased the ETL performance by 40% by implementing performance enhancement techniques and adjusting the sequence of tasks.
  • Reduced number of production failures to 0 by improving the exception handling and upgrading the existing Data Model. Have a Zero-production-failure record for over a year so far.
  • Created the set-up plan for source version control on Unix servers.
  • Used IBM CDC for DataStage to decrease latency period between SAP and CRM data.
  • Assisted MicroStrategy team in SQL optimization.
  • Integrated SalesForce and other source systems to data warehouse.
  • Used JitterBit as a SalesForce connector. Received training and implemented complex solutions to achieve Near real-time data warehousing.
  • Successfully achieved automation for testing and validation of data. It resulted in saving 10 man hours of work on weekly basis.
  • Conducted performance analysis and source system analysis to understand the existing ETL logic.
  • Handling Salesforce and SAP dump files using Unix scripts to extract/load into DB tables.
  • Worked with Business team to develop and document complex Information Analyser (IA) rules.
  • Worked with DS admin to build a bug free production environment.
  • Handled real-time production failures by recompiled jobs and ensured data quality after failures.
  • Used Quality Stage and Information Analyzer for profiling data from multiple sources.
  • Data quality assessment (DQA) using Information Analyser (IA).
  • Worked on massive data migration from Oracle to HDFS using Sqoop.
  • Performed Sqoop-Oracle connectivity, incremental load, data formatting.
  • Performed year-over-year (YoY) Baseline profiling in IA.
  • Created data quality benchmark reports and impact analysis of clerical record in IA.
  • Efficient in performing key analysis, column analysis, cross table analysis, baseline analysis in IA.
  • Modified system wide Analysis settings for column, table and cross-table analysis based on business requirement.
  • Created subscriptions in CDC for data migration between ODS to local DB server.
  • Integrated SAP data and monitoring user prompts while reloading datastores in CDC.
  • Performed various activities while creating CDC subscriptions like filtering, mapping tables, export/import subscriptions, remapping source table and troubleshooting CDC subscriptions.
  • Worked with the DBA team to resolve issues and implementing the fix.
  • Created standard documentation to follow the migration of ETL code and maintain version control using IBM’s Information Server Manager Client.
  • Worked with Linux team to identify Phantom processes and finding the root cause and applying permanent fix.
  • Used various stages like XML input, XML output, XML transformer, WebSphere MQ connector, Surrogate key generator, Transformers, Filter, Pivot, funnel, join, merge, look-up, Change capture, SCD stage.
  • Involved in business meetings.
  • Have complete ownership of ETL system and have maintained data quality and performance.
  • Involved in CRM-to-SalesForce migration process using IBM Data Replication CDC and DataStage.
  • Installation, configuration and environment testing after upgrade of DataStage UNIX server from 9.1 to 11.3.

Environment: IBM WebSphere DataStage 9.1/11.3, IBM’s Multi-Client Manager, Information Analyser(IA), Quality Stage, IBM InfoShpere Data Replication 11.x, MicroStrategy 9.4.1, Junos Pulse VPN Client, Sqoop, Oracle 11g, XML, XSD, Putty, WinSCP, JitterBit, SalesForce.

Confidential, Northville, MI

ETL lead

Responsibilities:

  • Involved in the development and implementation of data warehouse of an Insurance company.
  • Extensive knowledge in integrating DataStage and SalesForce.com.
  • Worked on the development of reusable Delta jobs to provide incremental run.
  • Extensive development in Sequence jobs- Developed VBScripts to trigger Master sequence based on the status of the incoming files, Restart ability of the sequence, used Sort routines for removing duplicates, Exception Handling.
  • VBScripts included the operations like File Handling, migrating files from FTP to target folder. Renaming the files after every run to valid format for loading the files for cloud computing stage, specified according to the Design Spec Document.
  • Created .bat program for deleting all .ds(Data Sets) using ORCHADMIN command.
  • Extensive knowledge on BAPI, IDOC interfaces for SAP integration.
  • Communicating with the onshore team. Providing daily update and status reports.
  • Interacted with Project Managers and the Design manager on 1st hand basis.
  • Extensive efforts in the debugging and testing phase of the projects including modifying the logic building and jobs.
  • Provided facilities for reusability of Jobs.
  • Well versed with migration into different environments and maintaining the standards and best practices.
  • SCRUM methodology to track development status and performed design reviews at the end of each sprint.
  • The mappings were Unit tested to check for the expected results.
  • Using various Runtime parameters for incremental run purposes.
  • Used DataStage Director to schedule jobs.
  • Worked on Windows 7 and Server R2 2008 platform and different utilities throughout the project including Visual Basic Studio.
  • Creating regular backup as dsx/ isx and maintaining logs of daily progress reports and errors solved.
  • Created timesheets on weekly basis.

Environment: IBM DataStage 8.7, Cisco VPN Client, Visual Basic Studio 2012, Oracle 10g, MS SQL Server 2008, Windows Server R2 2008.

Confidential

Lead DataStage Developer

Responsibilities:

  • Actively participated in the Team meetings to gather business requirements and in developing job Specifications.
  • Involved in preparing functional, technical Specifications and test case documents.
  • Used the DataStageDesigner to develop processes for extracting, cleansing, transforming, integrating, and loading data into Data warehouse.
  • Used Information Analyzer and Data quality stage to access and profile the data.
  • Documented and created various data profiling reports using industry and company compliance.
  • Installed DS and importing metadata repository for different DS projects.
  • Setup of IADB (IA database) and creating various user profiles with appropriate access for various projects.
  • Used DataStage Director for monitoring performances and troubleshooting from LOG files.
  • Worked with DataStage Administrator to define the project properties, environment variables, buffer size, node configurations, etc.
  • Developed complex SQL queries and performed performance tuning on existing SQL queries.
  • Worked on Various stages with in DataStage 8.5 (ODBC, OCI, XML input, XML Output, Hashed File and Link Collector, Transformer etc.)
  • Performed unit testing of the developed Interfaces.
  • Extensively worked with DataStage Job Sequences to Control and Execute DataStage Jobs and Sequences using various Activities and Triggers (Command activity, Start Loop Activity, End loop activity, User parameters, Notification activity, Job activity stage, Exception handler etc).
  • Used DataStage Designer for developing various jobs to extract data from different sources and loading data into temporary tables for validation testing and Production run.
  • Created DataStage routines to help mitigate repetitive tasks, date conversions, data cleansing, etc.
Environment: IBM InfoSphere Information server 8.5 / 8.7, Oracle 11g, SQL Server 2005, XML, XSD, SQL Developer, Query Xpress, ORMS, JDA, JMS, ORIB, Toad, Flat Files, Windows Server 2008.

We'd love your feedback!