Principal Consultant/developer Resume
FL
SUMMARY:
- Principal Cloud and Enterprise/Hadoop Architect/Developer with 16+ years of experience building scalable, distributed, parallel computing data solutions with 2 PB+ of data and driving business improvements with innovative Hadoop, and Hadoop’s eco
- System components and using BI tools along with Cloud Computing: Azure, AWS, and OpenShift.
TECHNICAL SKILL
Hadoop Distribution: Cloudera, MapR, Hortonworks, EMR, EMR2, Azure HD Insights Java, J2EE, Scala, Python, Spring, Spring Boot, Microservices, Hibernate, Web Services, Solr.
RDBMS: Oracle, MySQL, DB2, Redshift
NoSQL: MongoDB, Cassandra, HBase
Cloud: Microsoft Azure, AWS, OpenShift, Pivotal Cloud. Azure Databricks, AWS S3 storage, Azure Data Lake
Source CI/CD: SVN, Git, CVS, Perforce, MKS, Jenkins, Hudson, JFrog, Nexus
Build Tools: Ant, Maven, SBT & IDE: Eclipse, MyEclipse, IntelliJ IDEA IBM s UDeploy, URelease automatic code deployment tools Java Integration with TIBCO, IBM MQ, MongoDB, Cassandra
Security tools: Kerberos, OAUTH2, HTTPS
Scripting: Shell. Python
Data Models: Star Schema, Snowflake Schema, Galaxy Schema, Dimensions, Facts, Measures, De - Normalization.
PROFESSIONAL EXPERIENCE:
Confidential, FL
Principal Consultant/Developer
Responsibilities:
- Project: Process Engineering Platform - Cybersecurity and Technology Controls (CTC)
- The firm is on a journey to transform the way we deliver technology while continuing to foster a strong controls environment. A key enabler of this transformation is a deep understanding of the critical processes and underlying technology responsible for delivering our core products and services.
- Process Engineering Platform (PEP) supports this transformation by driving firm wide understanding of the flows of connected business activities and related technology services. By leveraging a combination of discovery, assessment, and automation capabilities, the platform provides management and transparency of the firm’s critical business processes and aligned applications.
- PEP Application receives massive network data from multiple sources: JPMC CSOrion Data Lake (hosted on Cloudera Hadoop), SEAL, RDHP-BCT, REFODS, Verum, RADAR, Restful API Services, Flat Files, CSV, Excel, Kafka, RDBMS etc.
- Analyze the massive amount of network data (volume in TB to PB), identify network application unique connections, and application hosted infrastructure details like application id, hostname, ip address, port, app owners, app designers, app architects etc. every application has a recovery time objective (RTO), recovery point object (RPO), regulatory flags, Risk Assessment Structure Process (Kafka feed) etc.
- Business Functions operate on a Process, the process accommodates multiple applications to fulfil the lifecycle of a process.
- PEP dashboard app provides unified access to reports to admin users, end users, and the downstream systems connect via API calls. Which application connections, infra connections, and the connections which are unaware by owners for cyber scrutiny.
- I have designed and prototyped Spark Data Pipeline for Data Analytics.
- I have collected data from multiple sources in association with multiple data formats.
- I have implemented technically data process mechanism using software tools.
- I have cleansed, analyzed, processed data using Apache Spark Process Engine on Hadoop.
- Data mart data stored into the Datastax Cassandra DB for BI reports which implemented using Java, Spring Boot and microservices.
- Tools: Java 8, Cloudera CDH 5.16.x, Hadoop HDFS 2.6.0, Spark 2.3.0, Scala 2.11.8, Datastax-Cassandra 3.0.15, CQL 3.4.0, Kafka 2.2, Avro 1.7.6, RHEL 6.2, Moneta Spring Boot 2.1.2, Angular, GoJS, Apigee, OAuth 2.0, Cyber Ark(EPV), CI/CD - jules, nexus repository, Scala Code Coverage-SonarQube,Agile, Jira, Python 2.7.x, PySpark, Shell scripts.
- Data Models: Cassandra Schema, Partition Keys, Cluster Keys, UDT, Lists, Sets, Maps
Confidential, EI Segundo, Cali,
Principal Cloud/Hadoop Arch/Dev
Responsibilities:
- Vehicle Fleet Management - Fuel Overview and Fuel Level Change etc UC.
- Dashboard Web Application provided for Fleet Managers to manage 10-15 fleets per manager, maintain, total cost of ownership (TCO).
- Designed, Developed and Architected Confidential application by integrating various software systems in the Azure cloud platform.
- Install and Configured Azure Databricks Cluster setup for Confidential
- Databricks collaborative notebooks on Spark for development.
- Setup Data pipeline using Azure IOT Hubs, Event Hubs, Spark Stream from Event Hubs.
- Leverage Azure Cloud IOT Hub, Event Hub, IOT Hub Device Provisioning Service (DPS)
- Azure Storage v2 used to store and analyze massive amounts of fleet stream data for Vehicle Fleet dashboard reports where we found aggregations no support (For massive feed).
- Azure Storage doesn’t support aggregation, we switched from Azure Storage to SQL Warehouse with SQLDW 3 Units for storing data and aggregations
- We also received Confidential xml files for other Data predictive and prescriptive analysis we process and store in SQL DW.
- Tools: Java 8, J2EE, Spring Boot, Angular, Azure Databricks, Spark Stream-Scala 2.10, JFrog repo, Jenkins, Jira, Agile, Git, JaCoCO, Azure Cloud Platform, Azure Storage,, Shell scripts.
- Data Models: Star Schema, Dimensions, Facts, Measures, De-Norms
Confidential, SJ, California
Principal Hadoop Architect
Responsibilities:
- Global Supply chain Management (GSM) Self Service Analytics (SSA)
- Analyzed supply chain data using Hadoop big data analytic tool Hive
- Source Analytics for Confidential Products under BU PF, various suppliers, manufacture inventory organization, demand forecast for the products.
- Data mart data prepared and consumed by multiple product development teams to improve product engineering services.
- Tools: Hadoop HDFS, Hive, Sentry, Spark, MapR 5.1.0, Kafka 1.x, Spring Boot 1.x, Kerberos 5.x, microservices for data report consumption as a PaaS on openshift cloud, AWS cloud, rally tracking tool.
- Data Models: Star Schema, Dimensions, Facts, Measures, De-Norms
Confidential, California,
Principal Hadoop Architect
Responsibilities:
- Worked as a Technical Architect/Consultant
- Involved in the requirements gathered, analysis, and preparing the design document, architecture document.
- Data silos are from Hadoop Data Lake Hive Warehouse
- Hive objects designed on star schema data model.
- Kylin Cubes built on hive partition data sets for New products(NPI), components, calendar.
- Analyzed data using Hive tool and
- Hive Data indexed on Solr Engine
- Dashboard application implemented using Spring MVC
- Successfully implemented and deployed - promising response time from Solr for BI dashboard analytics.
- Code review with team.
- Application code Integration and testing.
- Preparing the deployment Plan Release Notes for Live Env.
- I successfully designed the logical, physical view of the Hadoop system
- Implemented Data Compare Tool Java Web application for
- PCAM and hosted on LAE environment (openshift customized for Confidential )
- Tools: Hadoop HDFS, Hive, Sentry, Spark, MapR 5.1.0, Kylin 1.x, Spring Boot 1.x, Kerberos 5.x, microservices for data report consumption as a PaaS on openshift cloud, AWS cloud.
- Data Models: Star Schema, Dimensions, Facts, Lookup,Measures, De-Norms
Confidential, SJ, California
Principal Hadoop Architect
Responsibilities:
- Worked as a Technical Architect.
- Involved in the requirements gathered, analysis, and prepared the design document.
- Hive Warehouse used for application.
- Hive objects designed on star schema data model.
- Kylin Cubes built on hive partition data sets for products, components and calendar.
- Mentor to our team in functional, technical aspects.
- CBOM implemented using Spark
- Code review with team.
- Application code Integration and testing completed
- Prepared the deployment Plan Release Notes for Live Env.
- I successfully designed the logical, physical view of the
- Hadoop system, Cloudera
- Implemented Data Compare Tool Java Web application for
- PCAM and hosted on LAE environment (openshift customized for Confidential )
- Tools: Hadoop HDFS, Hive, Sentry, Spark, MapR 5.1.0, Kylin 1.x, Spring Boot 1.x, Kerberos 5.x, microservices for data report consumption as a PaaS on openshift cloud, AWS cloud.
- Data Models: Star Schema, Dimensions, Facts, Lookup,Measures, De-Norms
Confidential
Technical Lead
Responsibilities:
- Worked as a Technical Lead.
- Involved in the requirements gathered, analysis, and prepared the design document.
- Mentor to our team in functional, technical aspects.
- Java Middleware application (EOP) for Telecom client from UK.
- Prepared the deployment Plan Release Notes for Live Env.
- Post release support during warranty period.
- Provided Solution for BT business workflow wholesale orders into retail orders
- Integrated various business workflow modules into the application
- Implemented various Solutions for Application and Co-Ordinator for the On-Offshore model.