Cloud And Data Architect Resume
Houston, TX
SUMMARY:
- Cloud and Data Architect for Confidential 500 companies in Aviation, eCommerce, Finance, Banking and Telecom domain.
- Expertise in dealing with Petabytes of data every hour using Big data solutions in Cloud and on prem infrastructure
- Experience in cloud computing - AWS / GCP
- Writing Pyspark and Map Reduce to convert data into information
- Performance Tuning Expert in OLTP, DW and DSS system
- Seasoned PL / SQL Developer, Handled development projects with huge databases of multi terabytes in real time environment
- Having strong interpersonal skills, Analytical Skills, Team Management Skills with ability to work effectively with management.
- Well versed and experienced in Agile development methodologies
- Strong experience in Onsite Offshore Model
TECHNOLOGY STACK:
Big Data platforms: Cloudera, Hortonworks, Pivotal and MapR
Big Data: PySpark, Hadoop, HBase, Hive, Sqoop, Spark Sql, Redis
Cloud: AWS / GCP, S3 / GCS, EC2 / Compute Engines, Infrastructure, Bigquery, Airflow
Scripting Languages: Python, Unix
Data Warehouse/ETL: Informatica, Talend
Database: Oracle, Netezza, Greenplum, postgres
NoSql Databases: Cassandra, MongoDB
Reporting: Oracle Forms, ApEx, Tableau, Business Objects
Oracle Performance tuning: ASH, AWR, ADDM, Trace files, StatsPack, Explain Plan, DBMS Profiler.
Version control: PVCS Version Manager, Visual source safe, Perforce, Clear Case
Email Suite: StrongView
WORK EXPERIENCE:
Confidential, Houston, TX
Cloud and Data ArchitectResponsibilities:
- I am working on creation and execution of cloud migration plan that includes detailed road map, risk assessment and mitigation approach to migrate SQL Server Legacy system to Bigquery. I am playing a key role in evaluating, establishing and conducting proof of concepts for achieving end results using GCP components like Google compute Engine, Composer / Airflow, Cloud Functions, Pub Sub and Service accounts.
- Working with Infrastructure team to create IAM access policies, User Roles and accesses, allowing accesses to certain nodes, ports and APIs.
- Working with compliance and information security teams to understand reported vulnerability and plan remediation.
Confidential, Mason, OH
Cloud and Data ArchitectResponsibilities:
- I am helping data scientists and field service engineers by providing the technology and easily consumable data structures for analytics.
- I have been instrumental in providing solutions for Data Architecture and Data fusion in RDF thereby enabling foundational framework for testing and deployment of data science algorithms and Analytics.
- I am responsible for overall On - Prem and cloud architecture. I am using Talend as primary tool for data Ingestion.
- Data is stored in HAWQ/Greenplum data lake and S3. Analytics run in Spark and Web applications Interact with Greenplum.
- I have been writing code in pySpark, postgresSQL and PL / Python to process massive engine IOT data.
- I have also written spark programs in Python for pattern analysis like phase shift detection.
- Cassandra NoSQL database is used to store flight level information and used for fast retrieval based on flight key.
- Spark is used for Advanced Analytics. Wrote numerous Zeppelin notebooks for advanced flight data Analytics.
Confidential, San Bruno, CA
Cloud and Data ArchitectResponsibilities:
- Creating consumer insight profile for millions of customers based on purchase behavior and browse history on Big Data platform.
- Fetching stores and site data from multiple sources to connect and create integrated single view of customers.
- Processing purchase data in Hadoop, Hive and HBase and determining Next Products to buy and generating relevant recommendations.
- Writing Map Reduce jobs in Java and Python to convert data into information.
- Using Email marketing platform to monetize the recommendations.
- Created multivariate platform to perform A/B testing of multiple variants at the same time and send optimized email to customers.
- I used Sqoop for Data Ingestion.
- Hive and HBase were used to store data.
- Store Data, Web access patterns, demographic information and purchase histories of 50 million users were processed every day to determine best email campaign and recommendations. Spark and Storm were used to implement more real time campaigns like abandoned Cart.
- Email and Customer information was stored in HBase and Kafka was used to determine based daily for each customer based on priority.
Confidential, Concord, CA
Cloud and Data ArchitectResponsibilities:
- Profitability is a computation intensive data warehouse to calculate performance of Bankers, Managers, Areas and Regions in Hierarchical form.
- Designing the Architecture of Database Schema to enable efficient Reporting
- Used Informatica power center to develop processes for extracting, transforming, integrating and loading data into data warehouse.
- Extensively worked with multiple data providers and linking these data providers based on the requirements.
- Developed various Charts and Graphs like Bar Chart (Side by Side, Stacked), Line graphs, Pie Charts.
- Developed computation intensive Code to prepare summary tables for financial data
Confidential, Tampa, FL
Cloud and Data ArchitectResponsibilities:
- Designed and developed highly configurable Data Movement Tool using Oracle Streams to store configuration details in tables such that addition of New Tables / Business Rules will not require any code change.
- Created physical and logical database design using data modeling tool ERWIN.
- Design and Implementation of Health Check Monitor Module to Monitor Health of system.
- Developed Notification system to send Alerts in Real Time for system component Failure.
- Developed Object Oriented code using core Java, providing data to front end using ref Cursors.
- Developing Servlets to provide health monitoring updates.