We provide IT Staff Augmentation Services!

Sr Principal Software Engineer- Solution Architect (data Science Group) Resume

3.00/5 (Submit Your Rating)

SUMMARY:

  • 12+ years of IT industry with substantial experience in providing information management consulting services and Solution Architect on Master Data Management, Data Science & Data Analytics, ETL, Data Integration, Data cleansing, Data profiling, Data Quality and Business Intelligence Applications.
  • Excellent experience as Enterprise solution architect, Technical solution architect, development, solution architectures/designs, Integrations in Master Data management.
  • Expertise in Building a MDM solutions and process/system by gathering/identifying master data items, architecting the solution, identifying/defining matching/merging/de - duplication rules and guiding the development team to implement.
  • Good experience on designs, data modeling & integrations.
  • Experience in Conduct source system analysis and data profiling for lifecycle management
  • Experience in Designing, Installing & Configuring and MDM upgrade, Informatica MDM Hub Console, Hub Store, Hub Server, Cleanse Match Server, Cleanse Adapter and IDD.
  • Master Data Management concepts, Methodologies and ability to apply this knowledge in building MDM solutions.
  • MDM Hub Configurations-Data modeling, Data Mappings, Data validation, Match and Merge rules, customizing/configuring Informatica Data Director.
  • Defining and configuring landing tables, staging tables, base objects, queries/custom queries and packages.
  • Informatica (Siperian) MDM processes including batch based and real-time processing.
  • Performing Data standardization of addresses using Address Doctor & Trillium and Informatica Data Quality and Data integration in real-time using SIF API’s.
  • Excellent in configuring and designing SIF API’s (Search, Match, Put, Task related, merge, Unmerge) integration with multiple systems with Informatica MDM for real time process.
  • Excellent in defining various data science technics to know more insights about the data.
  • Expertise in performing Data analysis using python/Jupyter notebook.
  • Experience in applying different NLP techniques to areas such as sentiment analysis, classification, data/knowledge extraction, disambiguation, Experience in NLP methods such as LSA, LDA.
  • Knowledge on docker container based applications.
  • Experience in Dev ops Ansible scripts , Puppet, Docker, Jenkins Pipeline CI/CD process , Kubernetes cluster.
  • Automated build and deployment using Jenkins and Maven to reduce human error and speed up production processes.
  • Managed GitHub repositories and permissions, including branching and tagging.
  • Experience in Groovy scripting to maintain Jenkins pipeline.
  • Expertise in AWS services (VPC, Subnets, Routing tables, Security groups, EC2, S3, ELB, RDS, IAM, CloudWatch, CloudFormation).
  • Configuring HUB User exists and IDD custom user exists functionality based on business need.
  • Informatica MDM Hub Match and Merge Rules, Batch Jobs and Batch Groups.
  • Configuring Informatica Data Director (IDD) to meet Data Governance objectives in many project life cycles.
  • Informatica Power Center tools such as Source Analyzer, Transformation Developer, Mapplet Designer, Mapping Designer, Workflow Manager, Workflow Monitor and Repository Manager.
  • Experience in transformations like Source Qualifier, Expression, Aggregator, Router, Filter, Update Strategy, Lookup, Sorter, Normalize, Union, Stored Procedure, and Sequence Generator, Slowly Changing Dimensions.
  • Scheduling and Creating Workflows sessions & involved in the Performance Tuning of Mappings & Sessions.
  • SQL, UNIX, Shell Scripting, Data extraction & Validation.
  • Worked with facts and dimensions and designed dimensional model utilizing star/snow flaked schemas for use in BI reporting.
  • Software development Life cycle (SDLC) phases including design, development and implementation and testing.
  • Exceptional communication and cross-group collaboration skills.
  • Created Business Requirement, functional, technical, and Solution design documents.

TECHNICAL SKILLS:

Tools: MDM, ETL, GitHub, JENKINS, CHEF, Cognos, ETL power center, IDQ, Confidential Data stage, Kafka.

Operating Systems: Windows, UNIX, Sun-Solaris, Linux.

AWS: EC2, S3, Redshift, Dynamo DB.

Enterprise applications /products: Big data, Hadoop, HIVE, PIG, HBASE, HDFS, MAPReduce, MAPR, CMDM, Data Lake.

Domain: Telecommunications, Retail, Pharma, Energy Utility.

Versioning Tools: CVS. Clear Case, SharePoint, RPM.

Data Warehousing Tools: Informatica Power center 9.x/8.x.Datastage, Ab inito, SSIS.

Databases Oracle 9.x, 10g/11g, DB2, Confidential Informix, Amazon RDS, No Sql databases (Arango DB, Mongo DB, Influx DB)

Business Intelligence Tool: Confidential Cognos, QlikView, Tableau

PROFESSIONAL EXPERIENCE:

Confidential

Sr Principal software Engineer- Solution Architect (Data Science group)

Responsibilities:

  • Involved in Proof of Concept.
  • Auto Support Mongo DB Source telemetry data analysis and understand it thoroughly.
  • Apply Various Data science technics to extract information and insights of the data.
  • Define and Design ETL(Miners) process from sources telemetry data for data mining.
  • Design Advisor processes for various ML algorithms.
  • Design Arango Database for documentary (JSON) data store and influx database (timeseries data) data store.
  • Run multiple ML Algorithms on various data sets to validate results.
  • Design and implement the Visualization for the output of ML as Alerts to business users.
  • Design and setup Dev opps process (Automate end to end deployment steps)
  • Architect & Implement AWS infra setup for pre-prod and production environments.
  • Implemented AWS VPC, Subnets, routing table, security groups for AWS EC2 instances
  • Design core platform components for CICD process using Jenkins.
  • Implement infrastructure as code in Jenkins
  • Monitoring Jenkins end to end Build process and managing GitHub repository.

Confidential

Integration Architect -MDM

Responsibilities:

  • Defined End to End Integration Architecture for MDM, SFDC, Oracle ERP, POS to create a single view of Customer master in MDM.
  • Requirement Analysis and JAD sessions with Business team to propose solutions.
  • Co-ordinate with informatica team on hosted solution (AWS) setup.
  • Define MDM data model and architecture for Informatica MDM
  • Design and implement Match and Merge rules and validate results.
  • Design Informatica SIF API calls to enable search in MDM for various applications.
  • Design process in cloud salesforce.com application to push customer data to MDM.
  • Design, Configuring and customizing IDD based on Business need.
  • Define data governance process how to handle data steward tasks.

Confidential

Solution Architect - Informatica MDM

Responsibilities:

  • Layout the overall Solution Architecture for MDM Platform (customer master).
  • Requirement Analysis and JAD sessions with Business team to propose solutions.
  • Understanding of business process and system processes to make recommendations and propose technical/non-technical solutions to meet business requirements
  • Creation of MDM Data Model and Creation of Design Artifacts, Key Design Decisions for MDM Hub.
  • Creation of Landing, Staging, and Base object tables in MDM and mappings for their data load for Initial load and daily batch process.
  • Define custom User exits process for merge, task creation and golden record publishing (IDD UE and Hub UE).
  • Key design of SIF API’s for Create, Update, Match, Search, search match, Merge, Unmerge, Task creations process in MDM.
  • Define Data governance and data standards and their process for cross business owners (Medicade, Child care, Child welfare programs)
  • Design Match and Merge rules as per business requirement.
  • Performance Tuning and Scaling of MDM processes for higher volume to meet State SLA.
  • Design, Configuring and customizing IDD based on requirement.
  • Train multiple business partner data steward’s usage of IDD application and approval process.

Confidential

MDM Architect

Responsibilities:

  • Define the overall Solution Architecture for MDM Platform.
  • Installed Informatica Data Quality 9.6.0 and MDM 9.7.1 including Address Doctor.
  • Upgraded Informatica MDM 9.7.1 to 10 HF1. (Hub store, Hub server, Process server)
  • Requirement Analysis and JAD sessions with Business and BA's.
  • Worked on MDM 10 version new futures smart search and entity 360 view.
  • Data Profiling, Standardization & data quality using IDQ and Address Doctor.
  • Modeling the Party Data Model for Sunovion.
  • Create and own Design Artifacts, Key Design Decisions for MDM Hub.
  • Configuration of Match (Fuzzy/Exact and Auto/Manual) rules, Merge/Trust settings to create the best version of truth (BVT).
  • Performance Tuning and Scaling of MDM processes for higher volume.
  • Configuring and customizing IDD based on needs of Business Data Steward.
  • Developed an architectural design for ETL flow and data quality integration with Informatica Power Center.
  • Developed the reusable code which is a combination of Informatica Power Center, Informatica data quality code and shell scripts which will be called from one platform i.e. Informatica Power Center.
  • Created mappings within Informatica MDM to move data from landing to staging tables.
  • Developed various standardizers, address validator mappings within IDQ (Informatica data quality).
  • Created the various base objects using Informatica 9.7.1 MDM Hub console.
  • Create batch groups in Informatica MDM Hub to run the staging, match and merge and load jobs as per the dependencies.
  • MDM Process Automation by integrating of Batch Group, UNIX wrappers, and Informatica Power center.
  • Used Informatica Analyst extensively to profile the source data and created Scorecards to keep an eye on the health of incoming source data from various sources.
  • Developed parameterized shell script for calling data quality mappings from Power center server.
  • Developed multiple partitions strategy for enabling the ETL code to run concurrently for better performance

Confidential

MDM Specialist

Responsibilities:

  • Implement and configure new customer master model and Customer Master MDM hub.
  • Enable Custer Master data model Front End for business to enable its function.
  • Leverage front end to validate and accept cleaned Customer master data.
  • One-time data load from CRM (Salesforce.com) to MDM Hub.
  • Installed & Configured MDM Hub in AWS Cloud environment, cleanse server and Trillium in QA.
  • Gathered requirements from business users by conducting brain storming sessions.
  • Played an important Key role to create data model according to the requirements.
  • Defined the Base objects, Staging tables, lookups, queries, packages and query groups.
  • Created Mappings to get the data loaded into the Staging tables during the Stage Process.
  • Defined Trust and validation rules before loading the data into the base tables.
  • Coordinated with Business Leads in making them understand Match & Merge and incorporated their requirements and ideas.
  • Created Match rule sets in for the base objects by defining the Match Path components, Match columns and rules.
  • Developed Hierarchies using Hierarchy Manager in the Hub as per the needs.
  • Created IDD application and Subject Areas, Subject Area Groups, Deploy and test IDD application, cleanse functions, utilizing timeline, export and import master data from flat file.
  • Worked on BDD config file to get the changes reflected on IDD.
  • Analyzed the data by running the queries and provided the stats after Initial data and incremental Loads.
  • Configured JMS message Queue and message triggers to publish MDM hub data.
  • High level review of SAM - discussed use of Roles, creation of users and assignment of user to Role.
  • Defined Roles and privileges for each environment according to the requirements.
  • Scheduled MDM Stage Jobs, Load jobs using batch group.
  • Helped UAT team testing the new platform and authoring test cases.
  • Involved with ETL team in getting the data loaded into the landing tables.

Confidential

MDM Specialist

Responsibilities:

  • Gathered requirements from business users as part of initial analysis.
  • Creating the Base objects, Staging tables and landing tables foreign key relationships, lookups, queries, packages and query groups.
  • Created Mappings to get the data loaded into the Staging tables during the Stage Process.
  • Defined Trust and validation rules for the base tables.
  • Coordinated with Business team and making them understand Match & Merge and incorporated their requirements.
  • Created Match rule sets in for the base objects by defining the Match Path components, Match columns and rules.
  • Created IDD application and Subject Areas, Subject Area Groups, Deploy and test IDD application, cleanse functions, utilizing timeline and export and import data.
  • Worked on BDD config file to get the changes reflected on IDD.
  • Developed and created queries which can be used for over and under Matching.
  • Analyzed the data by running the queries and provided the stats after Initial data and incremental Loads.
  • SAM - discussed use of Roles, creation of users and assignment of user to Role.
  • Defined Roles and privileges for each environment according to requirements.
  • Defined the security such that schema will be secured with access only granted for specific downstream integration uses, using users created for those specific integrations
  • Scheduled MDM Stage Jobs, Load jobs using Utilities workbench.

Confidential

Informatica ETL Consultant

Responsibilities:

  • Analyzing the Business requirement documents and Conceptual Model, Logical, Physical models Technical design document and data flow document against Business requirements.
  • Involve in review the Business Requirements with Business and Regulatory team.
  • Creation of ETL technical Design documents.
  • Extracted data from different sources like Oracle, flat files, XML loaded into DWH and Prepared Test SQL Scripts to Test ETL Transformation logic.
  • Validation of Data from Source systems to CAMP/C2C Database ETL Extraction.
  • Created mappings in Power Center Designer using Aggregate, Expression, Filter, and Sequence Generator, Update Strategy, Union, Lookup, Joiner, XML Source Qualifier and Stored procedure transformations.
  • Handle slowly changing dimensions of to populate current and historical data to Dimensions and Fact tables in the data warehouse.
  • Creation of Informatica maps, sessions, worklets, workflows for staging, EODS and HDS.
  • Involved in developing complex mappings, Mapplets, reusable transformations using Informatica 8.6 designer to integrate data from different sources like Oracle, Flat files, and SQL databases by following guidelines and loaded data successfully into target and used Mapping Variables & Parameters, Workflow Variables and Session Parameters.
  • Used Informatica components extensively Repository, Designer, Workflow Manager & Workflow Monitor .
  • For loading business logic successfully to database used different transformations like Router, Expression, Source Qualifier, Aggregator, Filter and Sequence Generator transformations extensively.
  • Used Informatica Power Center Workflow manager to create sessions, batches to run with the logic embedded in the mappings.
  • Created workflows and used various tasks like Email, Event-wait and raise, decision in workflow manager.
  • Used debugger to test the mappings and fixed bugs accordingly.
  • Helped the testing team for data integrity and consistency.
  • Created deployment groups, migrated the code into different environments.
  • Bug Tracking by using HP Quality Center and provide resolution.
  • Implemented audit process to ensure Data warehouse is matching with the source systems in all reporting.
  • Involved in scheduling the informatica workflows using CAWA scheduling tool.

Confidential

Informatica ETL Developer

Responsibilities:

  • Verifying the Business requirement document and Technical design document against requirements.
  • Extensively used Informatica Power center for extracting, transforming and loading data from relational sources and non-relational sources.
  • Extensively used various transformations such as Source Qualifier, Expression, Lookup, Sequence Generator, aggregator, Update Strategy, and Joiner while migrating data from various heterogeneous sources like Oracle, DB2, XML and Flat files to Oracle
  • Developed Mappings using Designer to extract, transform data according to the requirements and loaded into database and scheduling of jobs
  • Monitoring the ETL jobs and fixing the Bugs.
  • Involved in doing Unit Testing, Integration Testing and System Testing.
  • Prepared Daily Status and Weekly status report
  • Attending reviews, status meetings and customer interaction.

Confidential

ETL Developer

Responsibilities:

  • Verifying the Business requirement document and Technical design document against requirements.
  • Designing mapping templates to specify high level approach.
  • Developing the Informatica maps by using the functional documents with sources like flat files and XML etc.
  • Extensively used transformations such as Source Qualifier, Aggregator, Expression, Lookup, Router, Filter, Update Strategy, and Joiner.
  • Tuned sources, targets, mappings and sessions to improve the performance of data load
  • Experience in data migration of Informatica Mappings, Sessions, and Workflows to Data Integrator.
  • Verifying the generated product price codes in database with ATT billing components.
  • Worked cooperatively with the team members to identify and resolve various issues relating to Informatica and other database related issues.
  • Attending reviews, status meetings and customer interaction
  • Interacting with the on-Site Co-Coordinator and users, Understanding the issues and resolving it in a specified time frame, Interact with Client when and where required
  • Conducting Peer reviews.
  • Initiating calls with onsite co-coordinator and client.

We'd love your feedback!