We provide IT Staff Augmentation Services!

Technical Lead / Bi Developer Resume

5.00/5 (Submit Your Rating)

San Antonio, TexaS

SUMMARY:

  • Over 10+ years of IT experience in Data Profiling and Data Analysis, Data Modeling, Data Migration, Development, Testing and Implementation of Data Warehouse/Data Mart Applications and Client/Server Applications on Windows, UNIX and Linux platforms.
  • Proficient in Hadoop technologies like Map Reduce, HDFS Architecture, Apache Spark, Scala, Pig, Sqoop, Hive, Impala, Oozie, HBase, etc.
  • Extensively used Datastage tools (Datastage Designer, Director, and Administrator & Infosphere Suites) and UNIX Shell Scripting.
  • Knowledge of SQL, PL/SQL functions, procedures and packages and experience in Database management (performance tuning, debugging) and Database design adhering to a mature software engineering process.
  • Developing ETL standards and documented for Data warehousing Application projects and Documented Best Practices for large complex Data warehousing Projects.
  • Expertise in Data Warehousing methodologies and Data modeling such as Star Schemas and Dimensional Model.
  • Experience in working on projects that follow agile methodologies and strong knowledge on Agile Scrum practices.
  • Experience in integration of various data sources (Oracle 8i/9i/11g, DB2, Sybase, MS SQL. Teradata and various File systems). Experience developing Data Models (Entity Relationship Diagram) (Logical and Physical Models) using CA ERWIN or Visio.
  • Prepared necessary technical documentation for projects with more complexity and involved in detailing architectural changes for the system
  • Configuring and scheduling of jobs through Autosys, IBM Tivoli Work Scheduler (TWS), Control M, UniCenter (UC4) and Zena.

TECHNICAL SKILLS:

Hadoop: IBM Big Insights, Hortonworks 2.6 (Hive, Pig, Kafka, Sqoop, Flume, Oozie, HBase, ZooKeeper, Apache Spark, MR, HDFS Architecture.

Tools: IBM InfoSphere DataStage and Quality Stage 7.x, 8.x, 9.x, 11.x, IBM InfoSphere Suite 8.1, 11.x, IBM InfoSphere CDC Data Mirroring Tool, IBM DB2 Replication Center Data Mirroring.

Databases: Netezza, ASE Sybase 13.5, MS SQLServer, Oracle 11i, Oracle 10G, IBM - DB2 Mainframe and UDB, Teradata

No SQL Databases: Mongo DB, Cassandra

Languages: C, C++ with oops, SQL, PL /SQL, Core Java 2.0, Microsoft Visual Basic, ASP, MQ Series, XML, XSL, HTML, Unix Shell Scripting

Software: Visual Studio.Net 2005, Toad, SQL Navigator, SQL Developer, Putty, Rapid SQL, SFTP, Winscp, IBM DB2 Data Server, IBM Message Explorer 7.5, IBM Rational Tools (Clear Case, Clear Quest), Rational Team Concert

Operating Systems: Windows 7/XP, UNIX, Linux

BPM Tool: Pega Rules Process Commander (PRPC)

Incident/Task Tracking: Service Now

Reporting Tool: Cognos, BO

PROFESSIONAL EXPERIENCE:

Confidential, San Antonio, Texas

Technical Lead / BI Developer

Responsibilities:

  • Involved in business requirement analysis.
  • Involved in creating micro level and macro level documents.
  • Perform data analysis and profiling across different source systems.
  • Generate data profiling and reviewed with business.
  • Imported data from RDBMS to HDFS using Sqoop Import/Export options.
  • Worked extensively with Hive DDLs and Hive Query language (HQLs)
  • Performed SCD Type I & SCD Type II within HDFS Hive (Beeline)
  • Used Datastage for data handshake between the Hadoop and other external tables.
  • Configured Oozie work flows to automate data flow, preprocess and cleaning tasks using Hadoop Actions. Used Control-M to automate the jobs.
  • Implemented common Spark Jobs (framework) to perform the Data validation
  • Implemented dashboards that internally use hive queries to perform analytics on structured data, ORC, Avro and JSON data to meet business requirements.
  • Involved in migrating the jobs from development to higher environments using GIT & Urban Code Deploy.
  • Involved in fixing the bugs and production support.
  • Performed Data Migration between the IBM BI & Hortonworks

Environment: Hortonworks Hadoop components such as Hive, Spark, Python, IBM InfoSphere Datastage 11.5, IBM Netezza, IBM Big Insights, BigSQL, IBM DB2 UDB, Oracle, SQL Server, Cisco ICM Data Model, Control-M, Rational Team Concert, SAP BO, Gitlab, GitHub & UCD (Urban Code Deploy)

Confidential

Responsibilities:

  • Work with Business team to understand the Business Requirement, Process Flow, Data structure and Hadoop ecosystem
  • Analyze & understand the existing legacy data model and worked on preparing the prototype design (low and high level) and walkthrough the same with Business and USAA Technical leads
  • Worked with multiple team and Data Science analyst to get the insights of the existing & upstream system
  • Develop the ETL data model based on Business Requirement also worked on customizing the existing data model
  • Design and develop Business Object Dashboard based on the semantic layer created in SAP BO Universe
  • Develop a replica of MCID data model into Hadoop environment and migrate the data into Hive from Netezza using Sqoop.
  • Configured Oozie work flows to automate data flow, preprocess and cleaning tasks using Hadoop Actions. Used Oozie for shell actions, java actions and ETL.
  • Implemented common spark jobs for each costumer to move data from traditional files to IBM JDM data ware house using HDFS.
  • Implemented HBase features such as compression and used to design, build Spark jobs
  • Implemented Device based business logic using Hive UDF to perform ad-hoc queries on structured data.
  • Design and develop Hive table and load the data into Hadoop Production environment
  • Developed the customized table in Hadoop for providing easy access to Site catalyst data to Business Analyst team
  • Create BIG SQL table and queries to access the Hive table and link the same into SAS library
  • Unit testing the code with all the possible scenarios and also performing validation testing for better support
  • Deploy the code in Production, support and monitor the jobs (using control-M) after Go-live and provide cognizant to the stake holders
  • Provide demo to Business Team and sponsors about the newly developed data model along with live data. Discuss about the impact of new model and changes to the downstream team

Environment: IBM InfoSphere Datastage 11.5, IBM Netezza, IBM Big Insights, Hadoop components such as Hive, Sqoop, BigSQL, IBM DB2 UDB, Oracle, SQL Server, Cisco ICM Data Model, Control-M, Rational Team Concert.

Confidential

Responsibilities:

  • Interacted with Business to get the high level requirements
  • Understand the requirements and prepare and create the documents
  • Coordinate with the Tech leads for review
  • Lead the datastage developer team and create some reusable components within Datastage.
  • Unit Testing the code with all the possible scenarios and also performing validation testing for better support
  • Deploy the code in Production and provide some cognizant to the stake holders

Environment: IBM InfoSphere DataStage 9.1, IBM DB2 UDB, Oracle, Control-M

Confidential, Austin, Texas

Consultant / ETL Lead / ETL Developer

Responsibilities:

  • Interacted with Architect team to get the High Level Design.
  • Preparation of Design & Mapping Documents
  • Coordinated with the Integration Architects for review the mapping and the design documents.
  • Leading the datastage developer team for the technical support as well as assisting them to complete the code in time.
  • Coordinating with IBM for initiating & supporting the PMR’s involved during the POC.
  • Developed the Datastage components using various commonly used stages and adapted new approach for handling the real time scenarios.
  • MQ Stages used to retrieve the message from Queue.
  • DTS (Distributed Transaction Stage) used to manage the multiple data source along with MQ.
  • DB2 & ODBC connectors were used as the reference table checks.
  • Shared Containers
  • And some commonly used stages: Transformers, Copy, Datasets, Lookup, and Joiner & Parallel Routines.

Environment: IBM InfoSphere DataStage 9.1, IBM DB2 10.5, IBM MQ Web Explorer 7.5, Toad 5.3, IBM Websphere Classic 9.5 for Adabas, ASG Zena is used for scheduling the jobs.

Confidential

Responsibilities:

  • Interacted with Architect team to capture the High Level Design.
  • Preparation of source to target mappings
  • Creating ETL jobs to extract and cleanse the data from source.
  • Creating stored procedures in DB2 to extract and load the data into New DB2 system.

Environment: IBM Infosphere Datastage 9.1, IBM DB2 10.5, Websphere Classic 9.5 for Adabas. Toad 5.3, ASG Zena is used for scheduling the jobs.

Confidential, Dublin, OH

Consultant / ETL Developer

Responsibilities:

  • Interacted with Architect team to get the High Level Design.
  • Translated with the HLD to DLD.
  • Managed compilation and review sessions of ETL detailed design document with necessary groups. Interacted with Data Modelers for the design of tables.
  • Developed the DataStage components using various DB and commonly used stages:
  • Teradata Connector Stage
  • DB2 and Oracle Connector Stages used as the staging database
  • Flat files access through sequential file stage
  • Commonly used DataStage stages: Transformer, Copy, datasets, lookup and join
  • CDC stages used for handling of SCD type.
  • AS400 used extensively throughout this project
  • IBM CDC tool used for replicating the data

Environment: IBM InfoSphere DataStage 9.1, IBM DB2, Teradata, Oracle 11G, AS400, IBM CDC, UC4 is used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Interacted with Architect team to get the High Level Design.
  • Translated with the HLD to DLD.
  • Managed compilation and review sessions of ETL detailed design document with necessary groups. Interacted with Data Modelers for the design of tables.
  • Developed the DataStage components using various DB and commonly used stages:
  • Teradata Connector Stage
  • DB2 and Oracle Connector Stages used as the staging database
  • Flat files access through sequential file stage
  • Commonly used DataStage stages: Transformer, Copy, datasets, lookup and join
  • CDC stages used for handling of SCD type.
  • Performance was the key factor since the data volume is very huge. We will be loading across 1.8 billion records every day.

Environment: IBM InfoSphere DataStage 9.1, IBM DB2, Teradata, Oracle 11G, UC4 is used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Interacted with IBM support and Administrator to migrate the code from 8.1 to 9.1
  • Documented the migration policy and the issues /resolutions.
  • Test run has been conducted and monitored.
  • Follow up approval with the Business SME’s and the end users.

Environment: IBM InfoSphere DataStage 9.1, IBM DB2, Teradata, Oracle 11G, AS400, IBM CDC, UC4 is used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Prepared design document for the sales and marketing data marts involved in the project.
  • Coordinated with the Integration Architect to translate the HLD to DLD.
  • Assisted with data modeling both in logical and physical.
  • Developed the DataStage components using various DB and commonly used stages:
  • Teradata Connector Stage
  • DB2 and Oracle Connector Stages used as the staging database
  • Flat files access through sequential file stage
  • Commonly used DataStage stages: Transformer, Copy, datasets, lookup and join
  • CDC stages used for handling of SCD type.
  • XML Input stage has used for pulling one of the source information
  • Business Object is used for reporting purpose.

Environment: InfoSphere DataStage 8.1, DB2 UDB, Oracle 11G, Axway for transferring files, Teradata UC4 is used for scheduling the jobs and its dependencies.

Confidential, Jersey City, NJ

Consultant / ETL Technical Lead / Senior Programmer / Programmer

Responsibilities:

  • Provided analysis of source system for availability of data elements for in scope data.
  • Gathered system specifications and requirements.
  • Designed fact tables and Dimension tables to fit in the columns through IAA and IIW data model.
  • Prepared the ETL design document. Managed the Development activities through DataStage.
  • Ensured XML was used as the source and target as DB2.
  • Ensured XML Input stage in DataStage is used to extract the data from the message queue.
  • Ensured surrogate key generator was used to generate the keys for loading both the claim and different data store in DataStage. Tracking and implementing Change requests.
  • Integrated the system with other applications. Analyzed root cause of defects and provide the best solutions.
  • Provided impact analysis and feasibility study of system to in corporate new functionalities in ETL.
  • Served as the single point of contact in offshore.
  • Analyzed root cause of defects and provide the best solutions.
  • Performed impact analysis and feasibility study of system to in corporate new functionalities in ETL and oversaw performance improvements.
  • Allocated task to team members and served as the single point of contact in offshore.

Environment: IBM InfoSphere DataStage 8.5, IBM MQ, DB2, Oracle and SQL Server, AutoSys & IBM TWS are used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Prepared prototype of the proposed design options.
  • Worked on this project to prepare the source to target mapping for 1100+ elements.
  • Provided analysis of the various source systems.
  • Provided Impact Analysis of the existing system. Prepared the profiling report of the existing system. Set up the environment for development and testing.
  • Prepared design document and reviewed with the business.
  • Prepared mapping documents. Monitored Construction and Testing.
  • Managed DataStage stages such as XML Output, Lookup, join, dataset, FileSet, Transformer has been extensively used.
  • Ensured xml was created with the help of DataStage XML output stages for the target team consumption. Ensured MQ was used to load the generated xmls. This was achieved through MQ Connector in DataStage.
  • Performed Unit testing was done thoroughly done for all the components developed.
  • Ensured Audit and Reconciliation report has been generated and loaded into the target DB2 for reporting purpose. Used IBM Cognos for Reporting.
  • Attended the weekly review meetings conducted by the project team members.
  • Oversaw coordinating with testers during system testing.
  • Managed effort estimation, planning and scheduling of the activities, preparation of Unit and System Test Plan. Reviewed code, technical trouble shooting and debugging the code.
  • Allocated work to team members. Provided progress reporting to Project Leader/Onsite Coordinator.

Environment: IBM DB2, XML, Message Queue, Oracle, SQL Server, AutoSys & IBM TWS are used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Did the Impact Analysis of the existing system.
  • Prepared the profiling report of the existing system for the business analysis.
  • Set up the environment for development and testing in DataStage 8.1.
  • Prepared mapping documents.
  • Ensured DataStage job was constructed to handle the master data for different modules.
  • Called stored procedures and loading the data into DB2 database has been extensively used in this project. Attended the weekly review meetings conducted by the project team members.
  • Oversaw coordinating with testers during system testing.
  • Managed effort estimation, planning and scheduling of the activities, preparation of unit and system test plan. Reviewed code, technical trouble shooting and debugging the code.
  • Allocated work to team members. Progress reporting to Project Leader/Onsite Coordinator

Environment: DataStage 8.1, IBM DB2, SQL Server, COBOL Stored Procedures AutoSys & IBM TWS are used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Provided analysis of the source system (Sybase) and Coordinated with the business team for generating the mapping spreadsheet.
  • This project covers most of the transaction data related to claims data.
  • Prepared Source to target mapping is really a challenge, since the number of target field was more than 1200 fields.
  • Prepared detailed design document. Reviewed and got it approved by business team.
  • Developed the code using DataStage by using across multiple processing stages lookup, join, and aggregator. Real time objects used such as XML and also Database stages such as DB2, Sybase, and ODBC and Oracle Connector stages in DataStage 8.1.
  • Ensured Unit testing was being done across multiple modules on a staging-in - Staging out basis in DataStage. Reconciliation process was created about the number of claims migrated.

Environment: DataStage 8.1, IBM DB2, SQL Server, COBOL Stored Procedures AutoSys & IBM TWS are used for scheduling the jobs and its dependencies.

Confidential

Responsibilities:

  • Managed source system Analysis. Oversaw approach and mapping document preparation.
  • Supported Construction and Testing. Supported SIT, UAT, Implementation and Production.
  • Provided analysis of the source system (Sybase) and Coordinated with the business team for generating the mapping spreadsheet.
  • Designed document preparation and reviewed with the business.
  • Generated the Contacts xml using XML Output stage and loaded into the Message Queue (MQ) MQ Connector for front end consumption in DataStage 8.1 Parallel.
  • Extracted the policy data from Sybase and loaded into the Target DB2 database as per the Source to target mapping.
  • Performed Unit testing is being done for all the components generated.
  • Worked on UAT and Production Implementation support.

Environment: DataStage 8.1 Server and Parallel, IBM Message queue, Sybase and IBM DB2. AutoSys is used for scheduling the jobs and its dependencies.

We'd love your feedback!