We provide IT Staff Augmentation Services!

Data Architecture/dw/bigdata/hadoop/data Integration

4.00/5 (Submit Your Rating)

SUMMARY

  • Very high energy technical solution architect, developer and team lead enterprise data architect offering 20+ years of expertise in architecture definition of large distributed systems, technical consulting, project management and technology implementation in Cloud, Big Data, Hadoop, Database, Business Intelligence, Enterprise Information Management, Data Management, Product Management and Application integration.
  • Proficient in Data Architecture/DW/Bigdata/Hadoop/Data Integration/Master Data Management, Data Migration and Operational Data Store, BI Reporting projects with a deep focus in design, development and deployment of BI and data solutions using custom, open source and off the shelf BI tools. Experience with implementing Modern Data Warehouse / Data Lake
  • Work with product owners, business SME and data ingestion and reporting architects to identify requirements and consolidate enterprise data model consistent with business processes.
  • Prioritize and scale architecture efforts in close coordination with business teams, Data Lake operational team, and other stakeholders.
  • Lead key business critical projects in the capacity of a Data Architect and Advisor
  • Review business requirements and technical design documents to develop effective data and database solutions
  • Support and Influence projects/initiatives and drive decisions related to data acquisition, processing and utilization through Big Data Platform
  • Strong aptitude to learn business processes/products and the role of data within business domain
  • Experience with Scala, Python and Spark
  • Hands on experience with Hadoop distributions like Cloudera, Horton works, EMR, Hadoop architecture and technology stack (Hive, HBase, Map Reduce, Sqoop, HDFS, Oozie and zoo keeper, Kafka, Spark, Kinesis and Lambda).
  • Hands on experience in AWS Cloud technologies using AWS EC2, EMR, VPC, S3, Redshift, Aurora, Data pipeline, AWS IAM, SQS, Dynamo DB, RDS and DMS.
  • Hands on experience with AWS Redshift, DynamoDB NoSQL databases.

TECHNICAL SKILLS

Data Warehousing: RelationalMySQL, Oracle, SQL Server, IBM DB2, Hive, Redshift

Analytical/Columnar: HBase

NoSQL: Elastic search, Dynamo DB, Redshift

Distros: Apache, Cloudera Distribution, Hortonworks Distribution

Big Data Frameworks: HadoopHDFS, MapReduce, Hive, Pig, Sqoop, Flume, Ooziee, Zoo keeper

Stacks: Hadoop, Spark

Data Analysis: Hive, Python

ETL/ELT: Talend, SSIS, SSRS, Amazon Data pipeline

Data Modelling: Erwin, Visio, MDM

Data Collection: API's, Web services, SQL, Amazon S3

Cloud/OS: Amazon Web Services, Google Cloud

Virtualization: VMWare, Virtual Box

Scheduling: Autosys, Control M

Languages: C#, Python, XML, SQL/PL - SQL

Version Controlling: Git

Others: Oracle SQL Developer, TOAD, SQL Work bench, DBeaver, Cygwin64, Subline Text3

We'd love your feedback!