Enterprise Architect Resume
SUMMARY
- Looking for a challenging role to utilize my Data Engineering, Cloud Engineering, Enterprise Architecture and Management skills along with strong experience in Financial and Fintech industries for the growth of the organization as well as to enhance my knowledge about new and emerging trends in the IT sector.
TECHNICAL SKILLS
Big Data and others: Hadoop, Spark, Apache NiFi, Kafka, Hive, Sqoop, Livy, Delta Lake, Databricks, DynamoDB, Cassandra, Informatica, Docker, Containers, Kubernetes, Domino Data Lab.
Cloud: AWS CLI, EC2, S3, VPC, IAM, EMR, Glue, ECS, Lambda, Azure, OCI, Terraform, CloudFormation.
Databases: Oracle, SQL Server, PostgreSQL, MySQL, Aurora, Exadata, Redshift, MongoDB.
Languages: Shell scripting, PowerShell, Java, Python.
Tools: Erwin, Golden Gate, Shareplex, DMS.
Blockchain: Ethereum, Hyperledger, Corda, Solidity, Go, Kotlin.
PROFESSIONAL EXPERIENCE
Confidential
Enterprise Architect
Responsibilities:
- Responsible for the engineering, operations and full life - cycle management of advanced analytics and machine learning platforms - Domino Data Lab leveraged by Data Scientists across Confidential .
- Worked closely with data scientists and other advanced analytical users to enable their development workflow and enable access to required datasets.
- Actively monitored and mitigated any performance, scalability or sustainability issues with platform and application components.
- Worked on reducing AWS costs by more than 45% - used Apptio Cloudability features like Rightsizing and reports. Educated users in selecting the right hardware tier for their workspaces in Domino Data Lab.
- Developed Python scripts for Auto shutdown of long running Workspaces(Jupyter, R Studio etc), and maintenance jobs.
- Good Knowledge of the Domino Data Lab Architecture; it’s underlying components like MongoDB, and Kubernetes
Confidential
Big Data Architect | AWS Solutions Architect
Responsibilities:
- Led a global team of Software and Data Engineers developing CPP Tier-1 Agile Big Data projects to achieve financial growth targets; evolve to scale and align Go-to-Market and Analytical resources.
- Managed and supported AI/ML underlying technology platforms (including Domino, Jupyter notebook, Spark etc.) and balanced short-term needs with long term strategies to deliver expected benefits.
- Designed a near real-time integrated analytics platform (data lake with 10 TB+ Actionable Data, 70k+ attributes, 3k+ tables, 41B+ Rows) of enterprise data assets for cross-domain analysis, data discovery, enterprise reporting, and new product innovation using AWS S3, EMR, Spark SQL, Hive, Scala, PySpark, Python, Sqoop, Kafka, Golden Gate and Autosys.
- Architected CreditSafe Firmographic and Financials data pipeline using Apache NiFi, Apache Spark, Apache Kafka, AWS S3, Airflow, Aurora Postgres to provide Index, Financials, CS Trade Payment, Equifax Trade Payment data to reduce workloads from application data stores like Credit Analytics for scoring PD and risk of default for 25+ million private companies globally.
- Worked on development of next-gen data ingestion platform, which will be used to build performant, scalable, and fault tolerant data ingestion pipelines using Apache NiFi, Apache Spark, Apache Kafka, Kubernetes on AWS and to drive next-gen Marketplace and Predictive Financial Analytics products.
- Strong Experience in writing Spark applications using Scala, Python and Spark SQL for big data processing and data analytics. Good knowledge of Snowflake, a modern data warehouse.
- Evaluated the features like Delta Lake of Databri cks as a Proof of Concept to analyze efficiency and complexity when compared to open source Apache Spark Grid Cluster management.
- Led EQR team to migrate from open source Spark grid onto Databricks to leverage delta lake features and performance of the platform.
- Worked with Application Architects to design RDSM(Ratings Dissemination Systems Modernization) program architecture, a key business initiative to disseminate an accurate, timely and compliant Ratings package that leverages technology to minimize the risk of selective Disclosure and associated legal costs.
- Led Telekurs replacement project by streamlining already available data sources that saved 4.5 Million USD annually.
- Architected CUSIP system in AWS Cloud by leveraging native cloud application principles to enhance security, and improve high availability and DR capabilities.
- Worked with other Architects and DBA teams to plan and migrate mission critical Oracle RAC databases from Exadata to AWS using FlashGrid SkyCluster for RAC DB nodes, combining local NVMe SSD and EBS GP2 storage. Worked with Application team, BA, and QA teams to test and fixed several performance issues found during load testing in staging env.
- Led CA team in designing Cassandra access patterns, data modeling and best practices to scale performance for MI Products and customers.
- Designed, developed, deployed and supported Informatica ETL workflows for different datasets including Ratings and Financials. Have Strong understanding of data warehouse methodologies, ETL processing and dimensional data modeling.
- Led PostgreSQL and Oracle database migration projects from on-prem to AWS cloud to improve reliability, stability, and performance of backend processing layers.
- Played a key role in reviewing architecture, designs and code reviews in Big Data, and Cloud communities.
- Established enterprise wide data architecture best practices and standards, maintaining enterprise Metadata Catalog.
- Executed the consolidation of multiple key Confidential amp;P Ratings databases on Oracle Exadata to improve reliability, stability, and performance of the platform, meet the business 99.99% SLA goals, and help reduce latency for information for Confidential amp;P Ratings customers.
Confidential
Senior Database Architect
Responsibilities:
- Worked on tuning applications with Hybrid Columnar Compression (HCC) Partitioning to improve the performance and set up data integration between heterogeneous databases with Golden Gate replication to reduce latency of data delivery to Products.
- Designed and implemented Oracle OLTP databases on AWS cloud servers in EU and APAC regions and set up data ingestion with Sqoop from internal databases to Hadoop platform so that data latency is reduced significantly to those region customers.
- Improved the uptime SLAs, configuration standards and 24x7 support for mission critical client facing products running on Oracle, SQL Server and Oracle 11g RAC Cluster environments to improve customer satisfaction.
- Designed Data security and protection solutions with Data Masking and encryption for compliance.
- Enabled bi-directional Shareplex replication with conflict resolution for applications running on multi master databases. Later migrated from Shareplex to Golden Gate to reduce the latency and provide performance of commit-only transactions; implemented the same conflict resolution in Golden Gate.
- Designed High availability architecture at remote data centers by combining Physical Data guard and Logical Database using Shareplex replication software to enable Load balancing for applications on two data centers.