We provide IT Staff Augmentation Services!

Data Architect Resume Profile

5.00/5 (Submit Your Rating)

Eden Prairie, MN

Experience Summary

  • More than 13 years of experience in the Information technology
  • The open group architecture framework TOGAF 8 certified
  • Played roles of Big data, ETL, Solution and Data Architect
  • Scored more than 90 in MongoDB for DBAs course from MongoDB University.
  • DataStage Enterprise Edition Certified
  • American Healthcare Management AHM-250 certified
  • Electronic Data Interchange HL7 Health level seven Certified
  • Worked extensively in Healthcare and Insurance domains
  • Design the entire process flow for the Data Warehouse projects
  • ETL Jobs Design, Development, Testing and Implementation
  • SQL performance tuning
  • Data modeling
  • Worked extensively on HDFS, Hive, DataStage, Oracle SQL loader, SQL, PL/SQL, and UNIX Scripts.
  • Liaison with client teams, business analysts and architecture boards on DW, Oracle and SQL server Projects, played active role in technical consulting, architecture, ETL Design, Data modelling and Performance tuning.
  • In consulting assignments analyzed client requirements and came up with solutions to support both tactical and strategic requirements. Analyzed the current state, documented the findings and envisioned target state, provided options and recommendations to achieve the goals.
  • Won the first prize at Optum Garage Innovation day challenge for excellent working solution to a given problem.
  • Awarded certification of appreciation from senior management SVP's and senior director's for outstanding contribution for successful delivery of important project
  • As a technology architect did POV's and POC's on Big data. Wrote whitepaper on Pentaho integration with Big data. Provided solutions for some customer engagements
  • As a solution architect analyzed existing environment of a leading UK based Insurance Company and came up with solutions to consolidate multiple databases, ETL and reporting tools which were result of acquisitions and Individual LOB's growth.
  • Technical manager at onsite/offshore for implementations of various data warehouse projects for leading US Health Insurance Company. Designed the ETL process for some of the strategic projects with high service level agreements.
  • Worked as senior technical member at onsite for a leading disability management Insurance company in USA. Created ETL standards designed and developed DataStage jobs for various projects. Did code reviews of the team members, performance tuning of ETL jobs for other business units.
  • Managed a team to develop EMR Electronic Medical Records and research data warehouse for a leading healthcare organization in India.
  • Maintained strong working relationship with Business users, senior management and other teams.

Software

  • Big Data HDFS, Hive, Pig, MongoDB, HBase
  • DataStage Enterprise and Server edition
  • Connect Direct, Microsoft Visual Source Safe VSS
  • ERWIN, Visio, TOAD
  • Pentaho Kettle ETL

Databases

  • Oracle 11G, DB2, SQL Server, MS-Access

Languages

  • C, C , SQL,PL/SQL, UNIX shell scripting, DataStage Basic

PROFESSIONAL EXPERIENCE

Confidential

Data architect

Roles and Responsibilities

  • Analyze the functional and non-functional requirements and provider overall solution working with application architect.
  • Prepare high level process flow from data ingestion to reporting
  • Create logical and Physical data model for application database and Analytics data mart
  • Work on the deployment strategy and topology
  • Work with other teams to process data via Hadoop. Cleanse, Validate, extract and load data into Database
  • Provide ETL load strategies
  • Used Sqoop to dump data from relational database into HDFS for processing.
  • Configured flume to capture the news from various sources for testing the classifier.
  • Wrote extensive MapReduce Jobs in Java to train the classifier.
  • Wrote MR jobs using various Input and Output formats. Also used custom formats whenever necessary.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, analyzing and training the classifier using MapReduce jobs, Pig jobs and Hive jobs.
  • Used OpenNLP in the removal of stop words and Stemming of the words.
  • Used Pig and Hive in the analysis of data.
  • Created/modified UDF and UDAFs for Hive whenever necessary.
  • Loaded and transformed large sets of structured, semi structured and unstructured data.
  • Supported Map Reduce Programs those are running on the cluster.
  • Managed and reviewed Hadoop log files to identify issues when job fails.
  • Developed Pig UDFs for preprocessing the data for analysis.
  • Involved in writing shell scripts in scheduling and automation of tasks.
  • Worked on Hive for further analysis and for generating transforming files from different analytical formats to text files.
  • Recently started using Mahout for machine learning in identifying a more subtle classifier.

Environment: Big data technologies HDFS, Hive, HBase etc. , ERWIN 9.5, DataStage 8.5, Visio, Hadoop Gen 1 , HDFS, MapReduce MR1 , Pig, Hive, Sqoop, Oozie, Mahout, Java, Linux Shell Scripting.

Confidential

Hadoop Architect

Activities

  • Installed and configured Hadoop and Hadoop stack on a 7 node cluster.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Involved in data ingestion into HDFS using Sqoop and Flume from variety of sources.
  • Responsible for managing data from various sources.
  • Good experience with NoSQL database Hbase.
  • Designed and implemented MapReduce-based large-scale parallel relation-learning system.
  • Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
  • Evaluated the use of Zookeeper in cluster co-ordination services.
  • Installed and configured Hive and also wrote Hive UDAFs that helped spot market trends.
  • Used Hadoop streaming to process terabytes data in XML format.
  • Involved in loading data from UNIX file system to HDFS.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Involved in creating Hive tables, loading the data using it and in writing Hive queries to analyze the data.
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process etc.

Environment: CDH4 with Hadoop 1.x, HDFS, Pig, Hive, Hbase, MapReduce, Java, Sqoop, Flume, Oozie, Linux, UNIX Shell Scripting and Big Data.

Confidential

Technical Architect

Activities

  • Setup the hadoop cluster with Cloudera distribution of hadoop and Pentaho Kettle
  • Create template for evaluating various ETL tools from big data perspective.
  • Build hadoop capabilities in the team, headed the panel to review POC on Informatica, Pentaho and Talend for big data.
  • Provided technical help to pull data from twitter for analysis
  • Helped team to work on Windows Azure Cloud big data, move data to and from cloud

Environment: CentOS Linux , CDH 3.0, Pentaho Kettle, Windows Azure Cloud

Confidential

Technical Architect/ DW Projects and Architecture Analysis

Activities

  • Analyze the existing BI environment of the client
  • Check the management, governance and quality aspects of data, come up with strategy to enhance data quality, availability to meet the business and technical requirements.
  • Analyze the environment non-functional aspects such as scalability, performance and security and suggest the required changes.
  • Come up with a strategy to change some of the weekly full volume loads into incremental daily loads.
  • Study the security, high availability and disaster recovery of the environment and suggest necessary actions.

Environment: Sun Solaris, AIX, DataStage, SQL Server, Visio, MS Office

Confidential

Technical Manager/ Architect -IM Data Warehouse Projects

Activities

  • Requirement Study Reviewing the requirement specifications to understand how it improves or impacts the business users
  • Client liaison Coordinating with client and Business analysts and Risk mitigation Assessing the risk factors and ways to overcome those risks
  • Design the ETL process flow of the projects as per the requirement.
  • Develop the DataStage jobs and review the jobs created by other team members to make sure they follow the Standards and best possible approach. Develop Oracle packages and queries. Performance tune queries for other Client teams.

Environment: Sun Solaris, AIX, DataStage, Oracle 10g/11g,VSS, TOAD, Connect Direct, SQL, PL/SQL, UNIX Shell scripts

Confidential

Project Leader / Specialty Data Warehouse Projects

Activities

  • Functional Requirements Study Reviewing the functional specification document with Business Analysts
  • System Analysis
  • Design the process flow, prepare the logical data model and send to data modeler to create physical data model.
  • Design and develop the DataStage jobs and sequences. Review of deliverables Review the task released by the individual team members . Documentation of changes To ensure proper documentation in the project
  • Client Communication Communicate with client for business clarifications

Environment: Sun Solaris, AIX, DataStage, Oracle 9i/10g,VSS, TOAD, Connect Direct, SQL, PL/SQL, VMS, UNIX Shell scripts

Confidential

Project Leader / Electronic Medical records

Activities

  • Effort Estimation
  • Requirement Study Reviewing the requirement specifications to understand the business functionalities
  • Database design, objects creation, writing packages and performance tuning.
  • Knowledge sharing Sharing the business logic among the team members in seminar sessions or meetings
  • Co-ordination of project activities and Review of Deliverables

Environment: Sun Solaris, Oracle, Visual Basic 6.0, VSS, SQL, PL/SQL

Confidential

Module Leader / Hospital Management System

Activities

  • Prepare the High level and low level design of the database objects.
  • Coding of Visual Basic applications and Oracle procedures for the modules.
  • Review of design Reviewing the low level and high level design documents prepared by the team members
  • Status Reporting

Environment: Oracle 8i, VSS, VB 6.0, SQL, PL/SQL, VB Script

Confidential

Module Leader

Activities

  • System Analysis
  • Prepare the application flow diagram
  • Interacted with Head of departments, doctors, billing clerks etc to create the GUI which is easy to use and still robust.
  • Involved in the DB design, object creation and performance tuning with the DBA. Developed PL/SQL Packages, procedures and functions
  • Review of test plan test script
  • Review of deliverables

Environment: Oracle 8i, VSS, VB 6.0, SQL, PL/SQL, VB Script

Confidential

Team Member / Hospital Management System

Activities

  • Requirement Study Reviewing the requirement specifications with team leader to understand the business functionalities
  • Prepared the process flow diagram of the modules assigned to me.
  • Development of application and PL/SQL packages.
  • Create unit test cases and do unit testing. Worked with QA team on system testing.
  • Code walkthrough Performing the code walkthrough with other modules team as suggested by the team Leader

Environment: SQl Server 2000, VSS, VB 6.0, T-SQL, VB Script

Confidential

Activities

  • Requirements Study
  • System Analysis
  • Coding
  • Testing and Release

Environment: Windows NT, SQl Server 7.0, VSS, VB 6.0, T-SQL, VB Script

We'd love your feedback!