We provide IT Staff Augmentation Services!

Data Architect Resume

3.00/5 (Submit Your Rating)

SUMMARY

  • Highly analytical, motivated, and results - driven professional with well-rounded experience in the IT field, focusing on business intelligence, Big Data, data architect, and project development.
  • Armed with comprehensive knowledge of Business Intelligence, Big Data, Cloud products , including AWS,Google Cloud, Tableau, ETL, Spark,Scala,Hive, Snowflake NOSQL DBs, as well as of Agile methodologies, along with Atlassian tools, such as JIRA and Confluence.
  • Expert in big data ecosystem using Hadoop, Spark, Kafka with column-oriented big data systems on cloud platforms such as Amazon Cloud AWS and Google Cloud Platform.
  • Expert in setting up Cloud infrastructure for Data orchestrations, batch processing,querying,Machine Learning and Data streaming use cases
  • Proven leader with 12+ years of experience in implementation and delivery for Data Reporting, Data Warehouse ETL, Data Design, Data Analysis, Data governance, Data integration, Data quality, Application/cost Optimization and Security.
  • Fluent in architecture and engineering of the Hadoop ecosystem, ETL & reporting solutions
  • Expertise in fixing engineering & support issues includes cloud, platform technologies, Big Data, global IT systems
  • Conducted routine price comparisons with competing middleware, Business Intelligence, Big data products
  • Focused to deliver compelling customer value proposition, automated infrastructure systems, delivering visionary services, enterprise solutions.
  • Known as an efficient technical leader; skilled at managing high-quality solutions, influencing senior leadership teams by addressing complex initiatives, involved in handling highly technical complex projects
  • Forward-thinker and analytic problem solver; expert at assessing risks and pitfalls and driving key initiatives and timelines with minimal supervision.
  • Recognized for integrity, unparalleled work ethic, and organizational aptitudes; effective at handling multiple projects which includes analyzing trends, preparing analyses, while working within fast-paced settings.
  • Recognized as a Strong Team Influencer: effective in building synergy across engineering teams, non-technical teams, building strategic partnership between teams and playing a key role in generating revenue and minimizing the operational cost.

TECHNICAL SKILLS

Big Data Technologies: Apache Hadoop; Hadoop Clusters, MapR 5.5, Hadoop Common, Hadoop Distributed File System HDFS ; Google Cloud Platform, Data Proc, Composer, Apache Airflow, AWS EMR,AWS EC2,Glue,Athena Replication; Snowflake; Redshift; Hadoop Pig; Scala, Kafka, Strom, Big Query, Apache Beam, Data flow, HBase, Apache Airflow, Flume, Relational, hierarchical and graph databases, Python, Hive, distributed data file systems, data federation and query optimization

RDBMS: Oracle 12c,11g/10g, DB2 8.0/7.0, MS-Server 2005, MYSQL5 + & Teradata

Data Modeling: Dimensional Data Modeling, Snow Flake Modeling, FACT and Dimensions Tables, Physical and Logical Data Modeling, Erwin 3.5.2/3.x & Toad

Programming Languages: UNIX Shell Scripting, SQL, PL/SQL, Python 3. (pandas,Numpy,scipy,matplotlib)

ETL Tools: ODI, Alteryx, Informatica,SSIS, & Data Stage, Oracle 12c, Redshift, Alation,Snaplogic

Reporting tools: OBIEE, SAP Business Objects, Tableau, Operating Systems Windows 2000, UNIX AIX.RHEL

PROFESSIONAL EXPERIENCE

Confidential

Data Architect

Responsibilities:

  • Implemented key initiatives across multiple workshares includes Business Intelligence, Big Data Advanced Analytics & MDM domains
  • Involved in creating warehouses, loading legacy data to Snowflake warehouses, capacity planning
  • Taken care of complete ETL migration project to Snowflake warehousing platform.
  • Played a key role in setting up complete Data analytics platform for HBO MAX release which includes products like Alation,Snaplogic,Snowflake,AWS EC2, EMR instances, Tiger graph, Looker
  • Involved in integrating multiple scheduling tools like Tidal & Airflow through advances API calls.
  • Proficient in writing DAGs in python for complex ETL orchestrations.
  • Involved in Advanced Analytics projects includes configuring and setting high scale EMR clusters & EC2 instances in AWS
  • Proficient in Ansible, Terraform, Jenkins and involved in implementing CI-CD pipelines.
  • Hands on experience Hadoop/Hive, Snowflake, Scala, Spark & Python
  • Hands on experience with working Agile framework and fixing the key issues discovered during Delivery phase
  • Influencing teams and building synergy across, business and technical teams

Confidential

Big Data Architect

Responsibilities:

  • Worked with the Fulfillment Business Intelligence Team (FBIT) responsible for capturing and configuring the overall Confidential ’s store daily operational data into fully operational data warehouse, such as shipments, reservations, orders gathered and processed through the TIBCO and Oracle Data Integrator (ODI)
  • Implemented the migration of the entire Confidential ’s legacy ETL framework into Google Cloud Platform
  • Well versed in most of google cloud framework like gsutil, Google File Storage, Composer,
  • Offered strategic roadmap for data migration projects and initiated ETL solutions in Google Cloud Platform (Big query)
  • Part of Various initiatives which includes but not limited to the following:
  • Involved in setting up Data Proc (Managed Hadoop & Spark High Availability Cluster through GCP console}
  • Involved in setting up Dataflow (Kafka equivalent service in GCP)
  • Setting up GCP composer using Python - Apache Airflow pipeline
  • Migration of on-premise to cloud applications; and
  • Tuning GCP Big queries
  • Converting 1000s of lines of legacy Oracle SQL queries to GCP Big query equivalent

Confidential

Architect

Responsibilities:

  • Initiated MDM project for School Staff Data Management (SSDM) integrating entire teacher information throughout the Confidential employees from multiple sources, such as flat files, web-services, and CSVs
  • Generated documents used in tracking changes in conjunction with the end users and business partners
  • Used Erwin in performing Data Modeling as well as Oracle Data Integrator ODI in developing processes to extract, cleans, transform, integrate, and load data into data warehouse database and Oracle EBS interface tables
  • Partook in devising Data Lake and canonical models while working with different concepts, such as pivoting and parallel job execution in ETL product
  • Rendered active participation with performance tuning and load balancing for the jobs in ETL tools as well as in Unit testing, User Acceptance Testing, and System Testing to validate the accuracy and thoroughness of the ETL process
  • Exemplified expertise in developing ETL framework.
  • Evaluating big data products for new Big data Initiatives.

Confidential

Technical Lead

Responsibilities:

  • Held responsibility for the implementation of review and Data Governance processes, while providing adequate support to internal customers
  • Played a key role in managing the following major solution releases:
  • Re-establishment of data load strategies;
  • Developing new teams
  • Development of data model;
  • Driving new BI initiatives from scratch;
  • Initiation of new data reconciliation strategies;
  • Automation of manual jobs through shell scripting;
  • Implementation of macros instead of manual reports

We'd love your feedback!