Solutions Architect - Cloud & Big Data Transformation Resume
0/5 (Submit Your Rating)
SUMMARY:
- Cloud & Big Data Solutions Architect with experience working within the Biomedical, Media/Advertisement,Telecom,Financial domains.
- AWS Certified Solution Architect(Associate): Development experience in cloud, Hadoop, C/C++, Java, .NET, Internet technologies, databases.
PROFESSIONAL EXPERIENCE:
Confidential
Solutions Architect - Cloud & Big Data Transformation
Responsibilities:
- Applying high-resolution/multi-dimensional projected public cloud pricing adjustment model using historical track record of AWS/Azure/HCP pricing, third party analytical reports about estimated profit margins of top cloud providers. Enhancing such model with structural and architectural differences of internal architecture of AWS vs Azure, estimated schedule of discounts and Azure Cloud Solution Partners incentives as well as projected future trends.
- Factoring in hardware performance/cost trends on commodity markets (Moore Law, Kryder’s Law) as well as correlation with AWS/Azure/HCP price schedules to predict ROI/TCO
- Highest resolution (+Intel vTune) bottleneck performance analysis as a consulting service to select optimal types of cloud compute/storage etc to maximize cloud TCO gains.
- Building OpEx/CapEx strategies for public cloud. Re-platforming and limited re-architecting as a consulting service to maximize gains from planned OpEx/CapEx spread (AWS Spot, RI).
- Reserved Instances trading to reduce risk of 1-yr, 3-yr commitment, re-classification;
- Utilization monitoring (Cloud Watch, Cloud Cruiser, AWS Trusted Advisor)
- Adjusting legacy architectures to utilize more Spot instances, balance RIs vs On-Demand dynamic deployments (AWS, Azure, OpenStack, containers, Cloudera Director)
- Containerization: containers conversion, CI/CD standards adoptions, multi-cloud portability:
- Docker CC/EE deployed on OpenStack; AWS ECS Docker;
- Drive Cloud partitioning(private vs public) strategy based on predicted benefits of richer public cloud PaaS/SaaS vs labor/time to match via Private Cloud; Security/Hardening advisory; DevOps adoption, driving the migration of continuous delivery and continuous integration solutions to Docker/multi-cloud environment
- Modeling existing On Prem Big Data deployments/projects: cost, performance, SLAs, usage patterns, licensing cost, life-cycle, growth patterns, hardware end-of-life.
- Enterprise Data Lake(EDL) creation from On-Prem data streams
- Modeling of Big Data stack technologies (Hadoop, HDFS/object stores/Glacier, Kafka, NoSQL, Hive, Spark/Spark SQL, Storm, Impala, RedShift) as well as distros(Cloudera, Hortonworks, MapR, Pivotal, Big Insight, HD Insight), Cloud providers (AWS, Azure, GCP, Softlayer): cost, performance, SLAs, usage patterns, licensing cost, life-cycle, growth patterns, hardware end-of-life, COTS vs FOSS, PaaS/SaaS to come up with initial Big Data Stack/provider/distro/cloud selection recommendation
- EDL: cost/performance/flexibility evaluation of alternative storage/processing tools vs hybrid cloud, multi-cloud, vendor-locking goals and considerations
- Guiding the initial roll-out and initial PoCs time-to-market reduction via planned/quantified PoC planning, technology/vendor modeling
- Developed offerings to assure better TCO outcomes for following use cases:
- Computational predictive model of MPP scalability for Cloudera Impala using TPC-DS benchmark as well as customer dataset/queries; estimation of non-linear scalability risks is even more important for elastic cloud ‘burst’ deployments;
- Highest possible resolution (Intel hardware assisted/vTune) of performance profiling for MPP/Big Data deployments (Cloudera HBase vs Cassandra, Spark, Kafka, Cloudera Impala) On-Prem; private cloud (OpenStack, Docker, microservices architecture, public cloud AWS ) to find precise bottleneck/performance root-cause analysis; optimize private cloud and public cloud infrastructure selection for optimal TCO;
- RI(+trading)/Spot/On-Demand mix Confidential offering to maximize AWS migration TCO gains
- Full detail/all costs accounting and optimization model for AWS S3-based data lakes
- S3 scalability limitations, partitioning/organization/granularity impact on cost/performance (Campanile, EC2 CPU/cost+S3 $ overhead model)
- EMR(Paas, HaaS) vs IaaS-level Big Data on Cloud approach(Cloudera, Hortonworks) cost, performance, scalability knowledgebase/model
- On-Premise traditional bare-metal deployment vs private cloud (OpenStack) vs public cloud (AWS, Azure, HCP) migration operation, scalability, performance adjustments needed to address virtualization overhead, changing networking vs storage vs CPU vs RAM availability/selection available/pricing/long-term TCO factors:
- Public cloud virtualization caveats; Private cloud PaaS, IaaS capability vs public clouds
- Support of re-platforming while migrating to cloud (Oracle, Microsoft SQL to AWS RDS PostgreSQL, MS SQL, Redshift) based on TCO, transformation schedules predictive model. Re-platforming applications/webservices to Spot/RI EC2 +RDS model & also to Docker while migrating to public cloud.
- Developed outcomes quantitative and qualitative prediction model:
- Provide a matrix of common architectural patterns for this class of cloud transformation task;
- Create TCO/risk models covering wide range of scalability/future changes;
- Propose quantitative as well as qualitative (feature sets, architecture patterns, software/infrastructure proposed to be aligned with Gartner, … recommendations/ranking)
- Enhancing Confidential strategic and business messaging for consulting engagements via linking Gartner/Forrester/… ontologies with vertical (i.e. Retail, CPG,..) needs and client stakeholders ps (micro-segmentation) for Big Data/Analytics/Cloud domains
- Big Data migration from MPP to Hadoop stack:
- Requirements, high-level design.
- Use Cases: MPP to Hadoop stack mapping.
- Big Data open platform adoption strategy.
- Infrastructure and Hadoop software selection and budgeting.
Confidential
Sr. Architect / Big Data Architect
Responsibilities:
- Big Data analytics (public/private/hybrid, multi-cloud; data masking, privacy-preserving computation):
- Requirements, high-level design, Hadoop deployment architecture, Pre-sales, PoC, RFP.
- Data Compliance: US/European deployments(Multi-tenancy, Safe Harbour)
Confidential
Sr. Solution Architect/Big Data Architect
Responsibilities:
- Requirements, high-level design; prototyping.
- Transaction shipping from Oracle to HDFS, Oracle Flashback simulation on HDFS, Hive; Tuning
- Hadoop performance research on Linux vs Windows platforms (Hortonworks, HDInsight).
- Create company-wide architectural roadmaps for Big Data and Big Data analytics within selected verticals.
- Direct and support Solution Architects and Associate Solution Architects within the team.
- Solution design and IP(Intellectual Property) creation for cloud multi-tenancy deployments with sensitive data for regulated industries (HIPAA, PCI, EU Directive 95/46/EC, EUGDPR, ITAR) to public (AWS+Azure) and Private Clouds.PII(Personally identifiable information) and other sensitive data isolation enforcement for multi-tenant cloud/Hadoop deployments, data masking and label security.
- Develop Big Data Analytics and business strategy across the enterprise qualitative and quantitative analysis of customer technology ps across all RFPs/engagements vs market trends harmonized with Gartner/Forrester ontologies/terminology and respective trend predictions. Building service expansion model for respective Big Cloud technologies/skillset needed/practices build-up . Liaising with major BI vendors(Microsoft Cosmos, EMC Greenplum/Pivotal) within Microsoft Atlas Solutions advertisement/marketing platform (Microsoft sold Atlas later to Facebook).
Confidential
Program Analyst/Infrastructure Architect & Business Analyst
Responsibilities:
- Project funded by Confidential contract with Department of Health and Human Services (HHS), Office of the National Coordinator for Health Information Technology (ONC):
- Specification review, requirements analysis, high-level design of implementation software to support the development, testing and adoption of future Nationwide Health Information Network capabilities, gateways to HIE.
- Engagement with HIOs and vendors for the development of real-world demonstrations and pilots for emergent Nationwide Health Information Network capabilities.
- Implementation (RI) national product owner of ONC “Transitions of Care” (ToC) and Lab Results Interface (LRI) projects (national effort to create inter-provider data transfer standard). ToC project defined Meaningful Use/MU (Stage 2) criteria for transitions of care scenarios HL7/CDA, IHE, Spring, Mulesoft, Java EE, SOA, Web Services, JSON, XML tools{Altova} . RI was a FOSS implementation and a blueprint for future Enterprise Architecture (EA) MU 2 vendor implementations.
- Used UML/SysML(Sparx Enterprise Architect), AWS/GovCloud, VMWare vCloud; Hadoop, Linux, Solaris)
- Working with CMS(Medicaid/Medicare) IT infrastructure/operations, analytical support of new CMS initiatives. Confidential serves as a main CMS infrastructure contractor.
- Technology lead of CMS next generation analytics (Hadoop, R, Datameer, migration of ETL and analytics from mainframe, Teradata MPP to Cloudera, Greenplum DB, Chorus. DR/BC).
Confidential
Systems Administrator/Manager/Software Projects Coordinator
Responsibilities:
- Dermatology EHR: Business continuance (BC) and Disaster Recovery (DR). Migration to virtualized cloud (Microsoft Hyper-V, VMWare ESX/vSphere, Citrix), Confidential t privacy data compliance enforcement
Confidential
Team Lead
Responsibilities:
- Directing the international team of developers to migrate Citrix Presentation Server to novel network API, develop new networking features (Vista, Windows Server 2008; IPv6. Win32/Win64 API, XP, XP x64, W2K3, W2K3 x64, ICA; VMWare ESX/vSphere, Xen; C/C++, Sharepoint, Caliber RM.
- Infrastructure: servers/workstations/ networking. Coordinating/reviewing/monitoring/establishing project goals, requirements, engineering response, coding, testing (automation, fuzzing), publications.
- Extensive OS and application virtualization infrastructure deployments (public, private, hybrid clouds). Working directly with cloud back-end (on source code level for Xen, XenServer, Presentation Server and Microsoft source code for Windows internals). Cloud internal architectures, performance caveats.
- Development of Citrix Presentation Server modules (Win32/Win64 API, multi-threaded, network API, troubleshooting, software localization on Windows NT4, 2000, XP, Windows XP x64, Windows 2003, Windows, Vista, Longhorn), core Citrix proxy features code owner.
- Languages and technologies used: C/C++, Java, C#, Visual Studio 6, 2002, 2003, 2005, VMWare ESX/Microsoft Virtual Server, ICA, tracing&debugging techniques, SCRUM/Agile/Waterfall, Sharepoint.
Confidential
Software Architect
Responsibilities:
- Designed Web GIS application and GIS raster and vector storage approach, which allows the storage of multi-Terabyte datasets on an array of x86 servers.
- This array sustained high rate of clients’ requests, and access nation-wide GIS data seamlessly in a Web browser (GUI Java, GIS image processing Java, C++,MapReduce ). This application had been used in TerraFly project. Collaborated with USGS.gov on Open GIS/OGC interoperability GML,KML (Geospatial Data Standards to promote data dissemination). Custom GIS application for verticals real-estate, intelligence/defense (ESRI: ArcGIS, FOSS GIS platforms). Real-estate/intelligence analyst GIS desktop fusing Data Science libraries with user-friendly visualization:
- Taught undergraduate course as sole lecturer, 2 Java undergraduate courses as a TA
- Taught 3 years as PhD level graduate course “Advanced topics in databases” principal lecturer