We provide IT Staff Augmentation Services!

Data Engineer Resume

5.00/5 (Submit Your Rating)

Houston, TX

SUMMARY

  • Data Modeler wif 7+ years of experienced in Data Analysis, Data Modeling, designing, developing, and implementing data models for enterprise - level applications and systems.
  • Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Toad, SQL Server, NoSQL.
  • Experienced in designing Star Schema, Snowflake schema for Data Warehouse, by using tools like Erwin data modeler, Power Designer and Embarcadero E-R Studio.
  • Experienced in Data modeling for Data Mart/Data Warehouse development including conceptual, logical and physical model design, developing Entity Relationship Diagram (ERD), reverse/forward engineer (ERD) wif CA ERwin data modeler.
  • Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like Erwin, ER Studio.
  • Experienced in Logical Data Model (LDM) and Physical Data Models (PDM) using Erwin, ER Studio and Power Designer data modeling tool.
  • Experienced in migration of Data from Excel, Flat file, Oracle to MS SQL Server by using SQL Server SSIS.
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for effective and optimum performance in OLTP and OLAP environments and experience wif Kimball Methodology.
  • Experienced in ETL design, development and maintenance using Oracle SQL, PL/SQL, TOAD SQL Loader, and Database Management System (RDBMS).
  • Experienced in Designed and developed Data models for Database (OLTP), teh Operational Data Store (ODS), Data warehouse (OLAP), and federated databases to support client enterprise Information Management Strategy.
  • Expertise in SQL Server Analysis Services (SSAS), SSIS and SQL Server Reporting Services (SSRS)
  • Experienced in Transform, and Load data from heterogeneous data sources to SQL Server using SQL Server Integration Services (SSIS) Packages.
  • Good knowledge and experience in Developing Informatica Mappings, Mapplets, Sessions, Workflows and Worklets for data loads from various sources such as Oracle, Flat Files, DB2, SQL Server etc.
  • Experience in providing solutions wifin teh Hadoop environment using technologies such as HDFS, MapReduce, Pig, Hive, HBase, ZooKeeper, Storm, and other Big Data technologies.
  • Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.

TECHNICAL SKILLS

Data Modeling Tools: Erwin R6/R9, Rational System Architect, IBM Infosphere Data Architect, ER Studio and Oracle Designer.

ETL/Data warehouse Tools: Informatica 9.6/9.1/8.6.1/8.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho Visual Studio, Data Bricks.

Database Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 12c/11g/10g and MS Access.

Big Data Technologies: Pig, Hive, Spark, Scala, Sqoop, MongoDB, Cassandra, HBase, Kafka.

BI Tools: Power BI, Tableau 7.0/8.2/10.x, Tableau server 8.2, Tableau Reader 8.1, SAP Business Objects, Crystal

Reports Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server.

Tools: Microsoft Teams, Cisco web-ex, Go-to meeting, Share point, Mat-lab.

Operating System: Windows, Unix.

RDBMS: Microsoft SQL Server14.0, SQL Server Management Studio 2019, Teradata 15.0, Oracle 12c/11g/10g/9i.

Project Execution Methodologies: Agile, SAFe Agile, Waterfall.

PROFESSIONAL EXPERIENCE

Data Engineer

Confidential

Responsibilities:

  • Implemented end-to-end systems for Data Analytics, Data Automation and integrated wif custom visualization tools using Power BI, Hadoop, and Spark.
  • Involved in creating Hive tables and loading and analyzing data using hive queries Developed Hive queries to process teh data.
  • Utilized Apache Spark wif Scala to develop and execute Big Data Analytics and build pipeline to ingest data into HDFS.
  • Involved in loading data from SQL to HDFS Importing and exporting data into Greenplum and used Spark to Implemented Partitioning, and Dynamic Partitions.
  • Created and maintained SQL Server scheduled jobs, executing stored procedures for teh purpose of extracting data from DB2 into SQL Server.
  • Used SSIS to create ETL packages to validate, extract, transform and load data to data warehouse and data marts.
  • Managed, and validated existing data models including logical and physical models of teh data warehouse and source systems utilizing a 3NFmodel.
  • Developed LINUX bash scripts to Subscribed to in house orchestrator to load teh data from SQL database to HDFS.
  • Developed Spark RDD transformations, actions, and Data Frame’s, case classes, Datasets for teh required input data and performed teh data transformations using Spark-Core.
  • Involved in converting SQL queries into Spark Transformations using Spark RDDs and Scala for TPCH benchmark.

Environment: Hadoop/Bigdata Ecosystem (Spark, Kafka, Hive, HDFS, Sqoop, Oozie), AWS (S3, AWS Glue, Redshift, RDS, Lambda, Athena, SNS, SQS, Cloud Formation), Oracle, Jenkins, Docker, Git, SQL Server, SQL, Java, PostgreSQL, Python, Pyspark, Teradata, Tableau, Quick sight, ER Studio, Data warehousing, ETL Informatica, Talend, Agile

Data Modeling

Confidential, Houston, TX

Responsibilities:

  • Created conceptual, logical, and physical relational models for integration in IBM Info Sphere Data Architect and Erwin.
  • Involved in reviewing business requirements and analyzing data sources form Excel/ SQL Server for design, development, testing, and production rollover of reporting and analysis projects.
  • Analyzing, designing, developing, implementing, and maintaining ETL jobs using IBM Info sphere Data stage.
  • Conducted brain storming sessions wif application developers and DBAs to discuss about various denormalization, partitioning and indexing schemes for Physical Model.
  • Involved in extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked wif data quality issues.
  • Used SSIS to create ETL packages to validate, extract, transform and load data to data warehouse databases, data mart databases, and process SSAS cubes to store data to OLAP databases.
  • Designed Source to Target mapping from primarily Flat files, SQL Server and using Excel.
  • Strong understanding of Data Modeling (Relational, dimensional, Star and Snowflake Schema), Data analysis, implementations of Data warehousing using Windows and UNIX.
  • Created ETL packages using OLTP data sources (SSMS 2019, Flat files, Excel source files, Visual Studio) and loaded teh data into target tables by performing different kinds of transformations using SSIS.
  • Performed data validation on teh flat files dat were generated in Excel using Excel commands, as necessary.
  • Developed, managed, and validated existing Data Models including Logical and Physical Models of teh Data Warehouse and source systems utilizing a 3NFmodel.

Environment: Power Designer, SQL, MDM, SQL Server 2008, ETL, DB2, SSIS, SSRS, SAS, SPSS, Data stage, Informatica, SQL, T-SQL, UNIX, SQL assistance, Erwin, Cognos.

Data Modeling

Confidential

Responsibilities:

  • Gathered business requirements, working closely wif business users, project leaders and developers. Analyzed teh business requirements and designed conceptual and logical data models.
  • Designed teh ER diagrams, logical model (relationship, cardinality, attributes, and candidate keys) and physical database (capacity planning, object creation and aggregation strategies) as per business requirements using Erwin. Created teh Conceptual, Logical, Physical data Models.
  • Did exploratory data analysis (EDA) using Python and done Python integration wif Hadoop Map Reduce and spark.
  • Reviewed teh Conceptual EDW (Enterprise Data Warehouse) Data Model wif Business Users, App Dev., and Information Architects to make sure all teh requirements are fully covered.
  • Worked on Dimensional and Relational Data Modeling using Star and Snowflake Schemas, OLTP/OLAP system, Fact and Dimension tables, Conceptual, Logical and Physical data modeling using Erwin r9.6.
  • Worked Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inman approaches and implemented Slowly Changing Dimensions.
  • Tested Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables.
  • Performed data analysis and data profiling using complex SQL queries on various sources systems including Oracle.

Data Engineer

Confidential, Allen Park, MI

Responsibilities:

  • Involved in designing/developing Logical Data Analyst & Physical Data Analyst using Erwin DM.
  • Worked wif various process improvements, normalization, de-normalization, data extraction, data cleansing, and data manipulation.
  • Performed data management projects and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like TOAD, MS Access, Excel, XLS and SQL Server.
  • Created SQL scripts to load teh custom data into Development, Test and production Instances using Import/Export. Created scripts to create custom Tables and Views.
  • Designed, Build teh Dimensions, cubes wif star schema and Snowflake Schema using SQL Server Analysis Services (SSAS).
  • Worked wif requirements management, workflow analysis, source data analysis, data mapping, Metadata management, data quality, testing strategy and maintenance of teh model.
  • Creating teh requests in answers and see teh results in various views like title view, table view, compound layout, chart, pivot table, ticker, and static view.
  • Assisted in production OLAP cubes, wrote queries to produce reports using SQL Server Analysis Services (SSAS) and Reporting service (SSRS) Editing, upgrading, and maintaining ASP.NET website and IIS Server.
  • Used SQL Profiler for troubleshooting, monitoring, and optimization of SQL Server and non-production database code as well as T-SQL code from developers and QA.
  • Involved in data from various sources like Parquet XML, Flat Files, CSV files and loaded to target warehouse.
  • Designed teh ER diagrams, logical model (relationship, cardinality, attributes, and candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin.
  • Designed Power View and Power Pivot reports and designed and developed teh Reports using SSRS.
  • Created SSIS Packages using SSIS Designer for exporting heterogeneous data from OLE DB Source, Excel Spreadsheets to SQL Server.
  • Developed and supported teh extraction, transformation, and load process (ETL) for a Data.

Environment: ERWIN9.1, Netezza, Oracle8.x, SQL, PL/SQL, SQL Plus, SQL Loader, Informatica, CSV, Taradata13, T-SQL, SQL Server, SharePoint, Pivot tables, Power view, DB2, SSIS, DVO, LINUX, MDM, PL/SQL, ETL, Excel, Pivot tables, SAS, SSAS, SPSS, SSRS.

We'd love your feedback!