We provide IT Staff Augmentation Services!

Big Data & Cloud Engineering & Java Full Stack Resume

3.00/5 (Submit Your Rating)

SUMMARY:

  • Data Cloud Solutions Engineering Architect/Developer having 15+ years of experience modernizing existing tech stacks and data platforms.
  • Driven the implementation of Cloud Native Microservices platform handling millions of customer’s real - time card authorization decisioning requests averaging at 450 TPS, running on 8 clusters with each cluster a 6-node c5.2xlarge virtual machines in a High Availability multi-region, multi-availability zone environment.
  • Architected and implemented big data pipelines using streaming and batch models processing 65 million customers & thousands of merchants’ history & real-time data on Apache Spark and PySpark clusters, Kafka clusters.
  • Leadership on the Capital Markets trading floor implementing Risk Management Systems. Worked in both Wall Street and Lasalle Street on Global Arbitrage and Trading Floors designing and implementing market risk, credit risk, operational risk and CCAR liquidity risk management systems for Exchange listed and OTC products.
  • Designed and implemented the Decisioning applications as Drools Rule Base engines where federated business teams develop and release the rules that are processed by spring boot microservices deployed as docker containers running in Kubernetes pods.
  • Delivered software releases through Sprints using Agile/Scrum methodology.

TECHNICAL SKILLS:

Big Data/Streaming/Messaging: AWS Kinesis, Apache Kafka, RabbitMQ, NATS Streaming, Apache Spark, PySpark, RDD, HIVE, gRPC, protobuf, Avro, Parquet, Confluent

Cloud: Kubernetes, Docker, AWS EC2, SNS, SQS, API Gateway, Lambda, Auto Scaling, Fargate, EKS, CloudWatch, CloudFormation, RDS, RedShift, S3, EBS, Load Balancers, Route53, Security Groups, VPC s, VPN Gateways, Elastic IP s, Subnets, Route tables, and NAT configuration, CloudTrail, Auditing, Compliance andGovernance, IAM roles and permissions, and cloud account consolidation, tagging, cost allocations, reporting, and budgeting.

Programming: Spring Boot, Spring Cloud, Angular JS, Netflix OSS, Python/Scala/JAVA, Spring Boot, OAuth 2.0, C#/ASP.NET Core3.x, UNIX Shell scripts, awk, sed, Perl

Market/Analytics Data: TSYS, Credit bureau, Bloomberg, FactSet, SWIFT, BARRA One, RiskMetrics.

Databases: Snowflake, Postgres, DynamoDB, PrestoDB, MySQL, Teradata, Neo4J

Infra and Monitoring: CloudFormation, Ansible, Terraform, Docker, Kubernetes, HashiCorpVault for secrets management, Prometheus, Grafana, Splunk, Datadog

ETL & BI Tools: Talend, Databricks, Tableau, Informatica IICS, Star & Snowflake design, QlikView, Microsoft Excel

BPMN: Red Hat Drools Rules, Camunda

PROFESSIONAL EXPERIENCE:

Confidential

Big Data & Cloud Engineering & Java Full Stack

Responsibilities:

  • Implemented Big Data Stack Data Pipelines using Python, Java, Amazon Glue, Spark, PySpark, Airflow, Hadoop, Hive, pandas, NumPy.
  • AWS serverless Lambdas receiving SNS notifications on arrival of Master and Delta files containing Peta Bytes of Credit Policy, Account History, Credit Bureau data, trigger Spark jobs producing aggregate datasets in Avro format for customers and merchants.
  • Aggregate data produced as Avro records are loaded into In-memory IMDG Hazelcast cluster as serialized protobuf objects.
  • Scan sensitive data in S3 data lake using Dataguise, Turing a set of Tokenization and Detokenization APIs built to secure NPI and PCI data.
  • Worked in an ecosystem publishing and consuming data using APIs(design as YAML document, Endpoint registration), Streams (Schema design, provisioning), Data Sets(publish to One Lake, Snowflake or Kafka or register datasets for Platform types: S3, Aurora, MySQL, Oracle, Postgres, Presto, RedShift, SQL Server, Cassandra, MongoDB, and Generic), and Features (well defined calculations and associated meta data for use in Machine Learning models).
  • Followed 12 factor App methodology developing spring boot microservies, unit tests, Git, Jenkins, component tests, Docker Containerization, JFrog Artifactory, Kubernetes deployments in QA, integration tests, on demand, pre-release Performance tests, ServiceNow production release/approval process.
  • Agile Scrum Methodology of 6-week PI planning, 2-week sprint planning, story refinement using Jira.
  • The implementation environment includes In-memory database Hazelcast, NoSQL DynamoDB, Kafka Event Processor, Drools, Maven, BDD, ATDD, CI/CD, Junit, Cucumber, Gherkin, Hashi Corp Vault for secrets, Centrify Security, Postgres, GO, API development, postman, API Gateway, OAUTH2.0
  • Application Monitoring environment includes NewRlic, Splunk, Dashboads, Alerts, DataDog, PagerDuty Consul Cluster, Resiliency engine Health Monitoring, suspend and resume traffic to the clusters
  • Apache Spark, DevSecOps, Container Security, Pod, IAM Roles, Security Groups, AWS Lambda, warm start, Route 53, ELB, JWT Json Web Tokens, PKI

Confidential

Big Data Cloud Engineering Architect & Java Full Stack

Responsibilities:

  • Worked closely with various trading desks, analysts, finance, Quants, business users and Risk Management to build out trading tools and risk/PnL reporting system.
  • Designed and developed the Hadoop/Apache Spark ecosystem handling quote and trade datasets, producing trade analytics and technical indicators.
  • Real-time streaming data for 2000 stock quote and trade data captured and ingested into Apache Spark, PySpark Big Data environment and processed data in memory using spark RDD transformations, queries using Impala. Apache Solr search setup Market data. Triggers are setup on services using Lambda Architecture to send out market orders. Neo4J queries for finding correlated stocks.
  • Setup and processing of Apache Kafka streams, Flink jobs for data transformations, Flume for collecting data from logs, jms, directories, Sqoop for bulk data transfer between Hadoop and relational databases.
  • Migrated analytics platform to Cloud Platform, redesigned asset management platform.
  • Extensively used AWS Lambda, micro-services, API Gateway, Amazon Aurora, RedShift, DynamoDB, MongoDB, Cassandra, JSON, ElastiCache.
  • Lead the Migration of infrastructure, data and applications out of legacy data centers into Cloud and hybrid environments (public and private). Identified the application stacks, classified, prioritized and analyzed applications for migration, integration, re-platforming and rehosting. Implemented configuration management tools puppet modules, CloudFormation templates using YAML to setup security groups, EC2 instances, Elastic IPs, S3 buckets, ELB.
  • Strong CLI and API scripting (Perl, Python, Power Shell, Bash) and automation skills (Terraform, Puppet, CHEF, DSC, ARM, CloudFormation)
  • Worked with MySQL, MS SQL Server, RDS, Azure Cosmos DB, DynamoDB, PostgreSQL, NoSQL MongoDB.
  • Strong experience with AWS services and Cloud Formation templates API gateway, AWS Lambda, and Docker.
  • Experienced with Cloud migration tools CloudEndure, RiverMeadow, CloudScape, ATAData
  • Developed Internal Securities Databases, Data warehousing historical data using Informatica PowerCenter.
  • High volume data ingestion of TBs of quote and trade data of nearly 5000 equities, bond, interest rates futures, options data using Apache Kafka and Apache Flink.
  • Acquired data from Market Data and Analytics vendors, designing ETL and Bigdata infrastructure, building analytical tools to provide actionable insight by designing large-scale Azure implementation for high-availability leveraging IaaS service offerings and build options, Azure PaaS (Web functions, SQL, Service/Event Hubs), Azure Logging, analytics and Azure Monitor, PowerShell, Active Directory, Group Policy, SCCM, SCOM
  • Produced data sets for Big Data projects using Hive, Pig, Sqoop, Impala, Spark, Map-reduce, HDFS, built analytical tools to provide actionable insight using Microsoft Power BI, QlikView, and Tableau.

Tools: AWS Cloud, Azure Cloud, Cloudera CDH, NO SQL, JAVA, J2EE, Spring Boot, SOA, RESTful API, MVC, Git, Maven, Jenkins, Drools, Apache Solr, R, Matlab, Excel, Python, JMS, TIBCO EMS, C++/C#, Econometrics, Statistics, Machine Learning, C++/C#, Eclipse IDE, Excel/VBA, Design Patterns, Informatica BDM, TIBCO Messaging

Confidential

Data Technical Lead

Responsibilities:

  • Designed and implemented Stat Arb (implied volatility, market beta) based High Frequency Market making application.
  • Responsible for Relational/Multi-Dimensional Data Warehouse design to consolidate more than 30 heterogeneous source systems, lead the development and implementation of management scorecard, operational dashboards and reporting.
  • Data Migration experience includes working with Deployment of PowerCenter on the AWS Cloud.
  • Broad and deep understanding of multiple technologies, architectures and design patterns in the areas of application and data integration (e.g. Cloud/SaaS integration, web services, REST services, messaging, ETL, SFTP), information (Hadoop, data warehousing, master data management), security (identity and access management, single sign on, data encryption, vulnerability), operations (monitoring) and infrastructure (network connectivity, high availability, disaster recovery)
  • Created an upstream Data warehouse using Informatica velocity best practices, consolidating source systems including Oracle, flat files, XML.
  • Designed and implemented complex mappings and workflows for multiple business units using Informatica and Oracle Technologies. Produced data sets for Big Data projects using Hive, Pig, Sqoop, Impala, Spark, Map-reduce, HDFS, built analytical tools to provide actionable insight.
  • Leadership in Business Requirements gathering, Analysis, System study, Preparing Functional & Technical specifications, Design (Logical and Physical model), Coding, Testing, Code migration, Implementation, System maintenance, Support, and Documentation.
  • Designed Dimensional Data model using Kimball bus Matrix to understand the business processes, identify Confirmed, degenerate, role-playing, junk dimensions, and creating the star schema using Slowly Changing Dimensions, surrogate keys, and snow-flake patterns.
  • Identified the granularity of fact tables and improved the fact table performance of the data warehouse that has around 4 billion rows using different types of design patterns such as Transaction fact, Periodic snapshot, Accumulating snapshot, fact-less fact, consolidated fact and Aggregate fact.
  • Implemented a customer Data Hub(MDM), to provide a unified and up-to-date, 360-degree view of each customer, used Informatica Data quality advanced data cleansing functions to identify and merge duplicate master data records also implemented a product hub for cross selling and up-selling of products.
  • Reduced the time to update warehouse data at night from nine hours to less than two hours, ensuring the information was up to-date when sales staff started work each day.
  • Replaced ad hoc budget controls developed using Excel and SQL with data warehouse-based, centralized, automated and accelerated budgetary control.
  • Supported concurrent queries from multiple internal users and authorized external parties.
  • Automated monitoring of everyday run of 600 workloads, with each workload containing between records.
  • Maintained 25GB PIM containing thousands of rows for each product, style and brand.
  • Developed PL/SQL Stored Programs (Procedures & Functions) to do data transformations and integrated them with Informatica programs.
  • Used Session parameters, Mapping variable/parameters and created Parameter files for imparting flexible runs of workflows based on changing variable values.
  • Created UNIX shell scripts for Automation and to invoke the stored procedures for the control process and parameter file generation.

Tools: AWS Cloud, Hadoop Cloudera, Hortonworks, HIVE, PIG, Informatica Power Center 8.6.1, IBM Data Stage, Oracle 10g/9i, SQL Server 2005, SQL*Plus, PL/SQL, TOAD, Win SCP, Unix (Sun Solaris), Ultra Edit, Maestro (Job scheduling tool), Reflection tool, C++/C#, JAVA, J2EE, Eclipse IDE, Excel/VBA, Design Patterns, Informatica PowerCenter, TIBCO Messaging.

Confidential

Central Funding Front Office

Responsibilities:

  • CFG comprises the global repo business, stock lending, securities finance desks and alternative asset group.
  • Created and supported the trading platform used by 100 front office personnel located across the Globe. Pricing and valuation of Equity Swaps, repo deals and daily PnL. Back of the envelop analysis, building daily interest rate curve and volatility curve.
  • Supported Prime brokerage operations, custody and clearance, corporate actions, dividend reinvestment trading, Stock loan plus Match Book.
  • Treasury funding to international banks and broker dealers against a range of securities collateral using tri-party repos.
  • Contributed and supported the Collateral Management System that calculates the haircuts, Initial Margin, and Variation Margins. Importing the collateral data from the custodians BoNY and JPM and querying the securities database using ISIN, CUSIP, SEDOL to verify the assets and Netting of Bilateral and Triparty transactions across various transits.
  • Providing the asset data to Risk management systems for pricing and VaR calculations, reporting of Market, Credit and Operational Risk.
  • Global Securities Database to capture CUSIP, ISIN, SEDOL, Calendar Data, Earnings and dividend data, Average Daily Volume, Market Value Capitalization, Turnover relative to float, VWAPGlobal Arbitrage and Trading:
  • RBCCM Equity Total Return Swaps expanded business to offer flow trading to Asset managers looking for perfect Index Replication, improved financing on hedging, exact/target exposure.
  • Designed and developed a flow delta one trading system to trade swaps on equity baskets using a Orchestration of various SOA services across multiple desks for Product setup, pricing process for financing, market making, collateral, FIX protocol Onboarding, Hedge execution, Swap booking and confirmations, settlements, valuations, PnL/product control, increase position and unwinds, rebalances and corporate actions.
  • Collaborated with the middle office on risk infrastructure development to calculate VaR, stress scenarios and sensitivity calculations. Historical market data and scenarios for Equity Implied Volatility, SwapZero curves FX Spot rates, FX Volatility surface, Equity Dividend estimates are sourced from Option Metrics, Bloomberg, SunGard.
  • Vetting the models, back testing

Tools: C++/C#, Java, J2EE web services, Oracle SOA 11g, Matlab, Excel, Informatica, FactSet, Reuters, RiskMetrics, DTCC, Informatica PowerCenter, Data Modeling, Oracle.

Confidential

Consultant

Responsibilities:

  • Biopharma company BMS implemented a global inventory Analytics and BW system for assessing and recommending appropriate inventory targets to ensure uninterrupted supply of medicines to the patients.
  • Responsible for data modeling, creating and executing ETL jobs to extract Product Data and provide inputs to inventory optimization for calculating business continuity inventory strategy and statistical safety stocks.
  • Collaborated with business users to understand business requirements, worked on setup and monitoring of import interface to improve and generate accurate Demand Plan.
  • Enabled the generation of multiple forecasts based on what-if analysis to allow planners to analyze situations based on exceptions.
  • Executed monthly processes for refreshing business continuity inventory strategy, statistically derived safety stocks and final safety stocks for drug products, drug substances, APIs, and key strategic raw materials.
  • Designed and developed KPIs and dashboards for monitoring compliance to target inventory levels across the supply chain.
  • Supported inventory management team providing ad hoc reports for inventory analysis, and data for budget and projection process and associated monthly reporting activities.
  • Enabled the company to manage 14 terabyte data warehouse and support up to 50% annual increase in the data volume.
  • Developed ETL processing tasks handle 40GB to 50GB of data and 100 analytics reports per day.
  • Implemented Oracle Database, allowing the organization to access secure, reliable, and accurate data (20TB) that drives critical business decisions daily.
  • Worked in the Performance tuning of the databases, ETL Procedures and processes.
  • Developed PL/SQL functions, procedures and packages to in corporate the complex transformation logic.
  • Worked closely with DBA to identify the performance bottlenecks in SQL using Oracle utilities like Explain Plain and improved the performance by creating index on appropriate indexes.
  • Wrote shell scripts and PL/SQL scripts, Stored Procedures for regular Maintenance and Production Support to load the warehouse in regular intervals and to perform Pre/Post Session Actions.
  • Created test plans, performed unit testing Used ClearCase to Check-in the developed code.
  • Created Drill-maps and implemented Drill Down functionality in the reports which are used by the end-users.

Tools: Informatica Power Center 8.6.1, Informatica Data Quality, Oracle 10g/9i, SQL Server 2005, SQL*Plus, PL/SQL, TOAD, Win SCP, Unix (Sun Solaris) Autosys Scheduler.

Confidential

Principle Consultant - Trading Systems Architect

Responsibilities:

  • Developed CME GLOBEX Equity and interest rates futures trading system.
  • Developed a FIX 4.2 gateway to CME GLOBEX iLink 2.0. Traders can hedge their equity portfolio with S&P futures contract E-mini S&P 500, E-mini NASDAQ 100. Real time risk limits control price, Volume and Time Risk. GLOBEX market data feed is distributed on TIBCO. Eurodollars (GE) interest rate futures spread (Calendar, Butterfly, Condor, packs and bundles) trading.
  • Designed and implemented e-CBOT, LIFFE Derivatives trading client and Gateway and got it certified.
  • Developed a .NET interop for LIFFE API that can be used by C#, VB.NET clients to access LIFFE CONNECT trading platform, order book and order depth. An Excel Add-in to receive LIFFE futures and options market data is developed as an Excel RTD Server a multicast client that joins the multicast group created by the Gateway a multicast server.
  • Implemented a strategy trading divergence in the correlation between S&P 500 and 10-year T-note futures, anticipating the mean reversion.
  • Developed Eurex, Eurex US derivatives trading client and Gateway using Values API.
  • Implemented market connectivity to Bloomberg, BrokerTec, ICAP, Reuters and TradeWeb.

Tools: .NET Threads, delegates, windows forms, J2EE WebSphere, JAVA/SWING, ADO.NET, .NET interoperability platform/invoke, VC++/MFC, ATL, COM, multicast sockets, Excel RTD Server, .NET assembly in Excel COM Client, GAC, XML/XSLT, SOAP, Web services, SQL Server 2000, Oracle

We'd love your feedback!