We provide IT Staff Augmentation Services!

Big Data & Analytics Architect Resume

4.00/5 (Submit Your Rating)

S UMMARY:

25+ years of high - impact Leadership in Corporate and Consulting. Ability to manage multiple, complex architecture and delivery engagements simultaneously. Oversight of large programs. Ability to present complex value propositions to executives. Fierce desire to be a critical team member in a rapidly growing business.

TECHNICAL SKILLS:

Python 2.7, Python 3.4, and R; NoSQL ( HBase, DSE Cassandra ), Real time streaming (Storm, Spark ), Hortonworks HDP 2.4 Platform ((Kafka, Storm, HBase, HDFS, Hive, Knox, Ranger, Oozie), Cloudera CDH 5.5, Solution Architect, Application, Data and Technical Architecture, Oracle Databases and Application, Cloud Computing ( EC2, Rackspace, Azure ), Search ( Lucene, Solr ), Data Science (Recommendation Engine, Predictive Analytics & Text Analytics). Git, mySQL, Unix/Linux. Experience in implementing projects using Waterfall, RUP, Agile and SCRUM project implementation

PROFESSIONAL EXPERIENCE:

Confidential

BIG DATA & ANALYTICS ARCHITECT

Responsibilities:

  • Architect and Implement a ROI-focused Big Data Solution in Workflow Performance Analytics. The solution helps customers reduce labor hours, decrease inventory adjustments in picking areas, boost accuracy rate when filling orders, and helps cut training time.
  • The Workflow Performance Analytics solution is built on a Common Data Analytics Platform leveraging Hortonworks HDP 2.4 (Kafka, Storm, HBase, HDFS, Hive, Knox, Ranger, Oozie).
  • Storm Topology is used for Real-time data ingestion from external device (Kafka Bridge), consumed in parallel to HBase and HDFS.
  • The stored data in HBase is processed on daily, weekly, monthly and yearly based on the business logic, and stored back into HBase.
  • Spark Analytics is used to run specific insight jobs. The output of Spark jobs are processed and exposed via Hive Thrift Server / Tableau / External Dashboard.
  • Extensive work on NOSQL data modeling, Column family design o Load data to and from HBase/Cassandra and Hadoop
  • Other Packages used: Python 2.7, HappyBase, Leiningen, Streamparse, Python Snappy, Avro, Pandas
  • Engagement Director: For a Confidential 50 Hi-Tech Manufacturing conglomerate:
  • Big Data Architecture and Strategy (real time, near real-time and batch processing).
  • Streaming Data Platform solution (ingest, transform, analyze, visualize and provide alerts). Provided comprehensive analytics and statistics on Operational efficiency
  • Key components: Cassandra, Kafka, Hadoop Cluster, Spark Cluster, Tabeau, mySQL
  • Data Analysis using Python
  • The Project Team comprised of Lead Architects, BI (Tableau, Visualization Tools, Custom Java applications developers) and Data Engineers (Java, Python, Hive, Pig).
  • For a Gaming company, defined the Big Data Architecture (real time, near real-time and batch processing) and led the development of Streaming Data Platform solution (ingest, transform, analyze, visualize and provide alerts). Provided comprehensive statistics on user acquisition, player engagement, monetization, betting recommendation, player progression, in-game economies and error tracking.
  • Kafka, S3, Hadoop Cluster, Spark Cluster, HBase, Tabeau, mySQL
  • Hands - on Data Engineering on Amazon Cloud o Data Analysis using R and Python
  • The Project Team consisted of BI (Tableau, Visualization Tools, Custom Java applications developers) and Data Engineers (Java, Python, Hive, Pig).
  • Project provided comprehensive statistics in user engagement, app usage, gaming behavior and betting recommendations via real time prediction engine.

Technology Lead and Engineer

Confidential

Responsibilities:

  • Responsible for Cassandra Database Design, Data Access (Pig, Hive, Python), and Data Integration (Storm, Spark Streaming, Kafka).
  • The overall platform included a comprehensive set of Big Data ecosystem components:
  • Governance and Integration (Sqoop, Flume, Falcon)
  • Data Operations (Ambari, Oozie, Zookeeper)
  • Key Metrics, Automated Workflow and Data Visualization

Technical Program Manager

Confidential

Responsibilities:

  • Defined and executed structured PMO approach.
  • Managed stakeholder communications, and scope to ensure commitments are achieved within agreed on time, cost, and quality parameters.
  • Directly managed all aspects of the project lifecycle and worked with matrix IT and Business partners to oversee all SDLC phases of the projects.
  • Led project teams in day-to day execution, project charter development, project initiation and facilitation, risk and issues management, quality planning, project Schedule development, configuration management, status report
  • Technical Lead for the Big Data and Analytics Architecture engagement to help unlock business value from large data sets of structured and unstructured data
  • Defined the logical and physical architecture for a Hadoop platform (HBase, MapReduce, Hive, Pig, Sqoop, Flume and Zookeper), to process 25-30% more data (install base, service contracts)
  • Accelerated processing time for identifying service sales opportunities by over 50%.
  • Created the foundation for other big data analytics projects.
  • Defined the roadmap to enable end-to-end Sales Planning automation through enterprise standard data and analytics. Program structure encompassed 3 tracks:
  • Business Capabilities Track; Data Solution, Analytics Track; Architecture Track
  • Architecture components included EDW, Reporting and Sales Planning Automation tools

Confidential

Technical Lead

Responsibilities:

  • Big Data Architecture engagement to assemble and integrate existing and new Data sources (SaaS, Enterprise and custom) from SFDC, Oracle, Eloqua, D&B, Model N, WorkDay, Omniture, MS SQL, flat files, and product logs.
  • Technologies included amongst others Hadoop, Oracle OBIEE, Greenplum EDW, Tableau
  • Advanced Analytics engagement to help improve Revenue Management with propensity use cases, pricing analytics and log data mining (0.5-0.75 PB of data) across multiple sources and regions.
  • Technology landscape included Hadoop, Splunk, Model N and others
  • Led a Data Integration Strategy and Architecture Initiative to integrate on premise and cloud Applications in Finance (SAP, Concur, ADP), Sales (SFDC, Hoovers), Marketing (Eloqua, Jive, Google, Adobe), HR(Workday, Jobvite LinkedIn) and Technology (Box,Tableau, Microsoft). Used a heat map as a technical reference model (TRM) to assess project alignment to the architecture and surface recurring architectural challenges.
  • Led the data and analytics strategy engagement to develop architectural vision and blueprint And identified pertinent data and technology gaps.
  • Conducted Business Process & Data Strategy Executive workshops / interviews with business and technology leadership.
  • Defined use cases by business unit (Consumer Lending, Mortgage, Investments, and Cards) and a created a multi-channel reference architecture, prioritized sequence of opportunities and roadmap.
  • Responsible for Technology Operations, Budgets, Skills & Training, Recruitment & Retention Managed 24x7 complex Web Operations infrastructure for over 50+ websites, 8M users on multi-platform, multi-location with high availability and security
  • For a Confidential 50 Hi-Tech Conglomerate:
  • Program managed the Business Intelligence and Data Services (BIDS) Transformation
  • Led and managed Corporate Quality (Hardware, Software, Customer Experience Metrics)
  • Led Program Office, Governance and Data Stewardship. Led Quality Data Migration
  • Initiative (Teradata) - Roadmap and Value Case, Interim and Target Architecture
  • Developed BI & IM Reference Architecture frameworks to accelerate formulation of BI Vision
  • Business Capabilities Assessment: Identified and prioritized Data & Technology gaps
  • Between current vs. target state BI capabilities.
  • Sequenced projects across initiatives on value And constrained by data and technology gaps

We'd love your feedback!