We provide IT Staff Augmentation Services!

Senior Data Engineer/data Migration Resume

0/5 (Submit Your Rating)

Vienna, VA

OBJECTIVE

  • A challenging Data Engineer that will utilize my extensive technical skills and will lead me to innovative work environment.

SUMMARY

  • Over 15 experience in IT as Data Architect/Database Administration/Data Engineer experience which includes Three years of experience in Big Data Ecosystem and Data Engineering technologies.
  • Excellent knowledge in Hadoop Architecture and its major component like Hadoop Map Reduce, HDFS Frame work, HIVE, PIG, Impala, Sqoop, Flume, Hbase, Oozie and Zookeeper.
  • Extensive experience in Data Modeling, Database Design, Data Architecture, Involved in several Enterprise wide Technology Architecting and Implementation.
  • Experience in importing and exporting data using Sqoop and Flume from HDFS to Relational Database vs vice - verse.
  • Extensive Experience in Development of Big Data projects using Hadoop, HIVE, Pig, Impala and other open source tools/technologies.
  • Well versed in Installation, Configuration, Supporting and Managing of Big Data and Underlying Infrastructure of Hadoop Cluster.
  • Managed the data management team of ETL Architects, database architects, middleware architects by providing architectural solutions for databases.
  • Experience with version control tools like Git, SVN and TFS
  • Designed Erwin logical and physical data model of the data warehouse.
  • Responsible for RFC & RFP document and attending CAB, PAB for Approval
  • Experience in high availability such as mirroring, log shipping, replication and clustering
  • Extensive experience in Replication(Oracle Golden Gate and Double take)
  • Skilled and Experienced in Server performance tuning using SQL Server 2005’s Dynamic Management View and performance counters in SQL SERVER 2012/2008/2005
  • Good experience in Coding and Improving Efficiency of Rebuild index, Re-indexing, Updating Statistics, Recompiling Stored Procedures and performing other maintenance tasks.
  • Experience in maintaining the security for databases (creating logins for users, roles, assigning permissions to them) and Encrypt the Database
  • Good Experience in Database Backup, Recovery and Restoring as daily activities.
  • Good Knowledge languages such as Scala and hand-on Experience with Python.
  • Expertise in ETL tools (SSIS, Pentaho, Attunity), SQL Server reporting Services(SSRS), SQL Server Analysis Services(SSAS), SAS, Confidential VA, Confidential EG.

TECHNICAL SKILLS

Databases: MS SQL Server, Oracle, HBase, Informix, Refshift

Programming: Python, Scala, C, C#, C++, Java, COBOL

Database Administrator: MS SQL Server 2014/ 2012/2008/2005 , Oracle

Data Warehousing: DMX-h, SSIS/SSAS, Informatics, PIG, HIVE, Impala

Data Analytics: SAS, Confidential VA, Confidential EG, Tableau, Crystal Report

Operating System: Windows, Linux

Other Software: IIS, Microsoft Transaction Server, VBScripts, HTML

Tools: SQL DB Access, Performance Monitor, Backup Utility, SQL Lite- Speed

Database Modeling: Erwin 9.5, Visio, UML

Virtual Tools Server: Microsoft Hyper-V, VMWARE

Network Topologies: TCP/IP, DNS, DHCP, FTP, SMTP

PROFESSIONAL EXPERIENCE

Confidential, Vienna, VA

Senior Data Engineer/Data Migration

Responsibilities:

  • Understand the Business Requirements and develop designs both at Low Level and high level for the ETL solutions Proposed.
  • Worked on design and developing Hadoop solution using Syncsort DMX-h, Hive, Impala and HDFS on Cloudera Framework in Agile-based environment.
  • Designed and implemented ETL processes using Syncsort DMX transformation tool
  • Created Sqoop jobs to replicate data between Hive and Oracle databases.
  • Analyzed and processed different file formats: EBCDIC files, fixed width text files, XML files Performed required actions on Hadoop file system.
  • Defined and developed batch schedule using CA Workload scheduling tool.
  • Created UNIX scripts to perform files manipulation, archival, batch execution orchestration and other command line level activities.
  • Design and develop Hive/Oracle SQL (DDL and DML) statements, stored procedures, functions, views and triggers to be used during the ETL processes
  • Created several Hive tables, implementing dynamic and static partitioning, and bucketing for efficient data access.
  • Worked on creating and maintaining documentation of the ETL processes low level designs, flow diagrams
  • Collaborated with business solution analysts, other developers, solution architects and other team members to deliver sound solutions
  • Performed Unit testing to ensure that the code is working as per requirements and efficiently

Confidential

Senior Data Engineer

Responsibilities:

  • Work closely with the business and Advanced Analytics team in gathering the system requirements.
  • Understand the Business Requirements and develop designs both at Low Level and high level for the ETL solutions Proposed
  • Worked on design and developing Hadoop solution using Syncsort DMX-h, Hive, and HDFS on Hortonworks Framework in Agile-based environment
  • Designed and implemented ETL processes using Syncsort DMX/SSIS transformation tool
  • Created Sqoop jobs to replicate data between Hive and Oracle databases.
  • Responsible for establishing, growing and maturing data integration development and support capabilities to deliver measurable business value.
  • Involved in development, building, testing, and deploy to hadoop cluster in distributed mode.
  • Participate in POC/POT efforts to integrate new Big Data Management technologies, Software engineering tools and new patterns into existing structures
  • Responsible for storing data in S3 and created table in AWS Athens for Adhoc queries.
  • Responsible for create Data warehouse solution in AWS Redshift
  • Handled Importing of data from various data sources, performed transformation using Hive, loaded data into HDFS and extracted the data from SQL Server, PostgreSQL, and oracle using sqoop
  • Create Sqoop Scripts for loading data into Hadoop and Schedule into Python.
  • Create Custom ETL Process to move data from Mainframe System to Enterprise Data Lake using DMX-h Syncsort
  • Handle Multivariable VSAM File in the Syncsort & Talend and loading Data to HIVE
  • Created the layout for the ASCII file in the Syncsort and Transform data to EDL
  • Experienced in parsing the VSAM copy book and apply copy to EBCDIC file format
  • Experienced in Converting File from EBCDIC to UTF8 and loading Data in EDL
  • Handle heavy data volume on the daily basis and optimize query and ETL Process.
  • Analyzed the data by performing Hive Queries and Running Pig scripts for fraud analysis
  • Experience in Oozie and workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph(DAG) of action with control flows
  • Maintain involvement in continuous improvement of Big Data Solution processes, tools and templates.

Confidential

Senior Database Architect

Responsibilities:

  • Involved in Installation, Configuration, supporting and managing Hadoop Cluster using Cloudera Manager 5.3 and CDH 5.2 and Hortonworks
  • Installing and configuring Hadoop eco system like sqoop, pig, hive, Impala.
  • Data Requirement Analysis, Data Modeling and Data Architecture for several projects.
  • Involved in Designing the architecture for Enterprise data warehouse and Business Intelligence Environment
  • Responsible for Driving Data Migration project as part of the Labyrinth project for the Confidential division of the Ministry of Health.
  • Involved in all phases of SDLC including analysis, designing, developing, testing and maintenance.
  • Participated in development of request for proposal (RFP) documents.
  • Managed the data management team of ETL architects, database architects, middleware architects by providing architectural solutions for databases
  • Involved in participating the design reviews, Regular and ad-hoc architectural reviews, and ensure correct interpretation of the Data Architecture, and applied strategies.
  • Prototyped a Master Data Services (MDS) Solution and presented it to Management.
  • Experienced in loading and transforming large sets of structured, semi-structured and unstructured data.
  • Responsible for building scalable distributed data solutions using Cloudera Distribution on Premise and Deploying Hadoop with Cloudera CDH to AWS.
  • Responsible for Cluster maintenance, managing cluster nodes.
  • Analyzed data using Hadoop components Hive and Pig and Impala.
  • Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism for Hadoop Cloudera.
  • Experienced in loading and transforming large sets of structured, semi-structured and unstructured data Hadoop concepts
  • Involved in creating Hive tables, and loading and analyzing data using hive queries and Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Hands on Experience Python Panda Frame to provide data analysis.
  • Involved in importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS using SQOOP and Flume
  • Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 11ig database.
  • Experience in migrating the data using Sqoop from HDFS to Relational Database System PostgreSQL and vice-versa according to client's requirement.

Confidential

Senior Database Administrator/Analyst

Responsibilities:

  • Involved in installation and configuration of SQL Server 2012/2008R2/2008/2005 in Windows environment.
  • Involved in data modeling for the SQL Server and for individual databases.
  • Involved in scheduling backups of databases and logs and restoring them when necessary.
  • Managing the use of disk space memory and connections.
  • Responsible for monitoring and making recommendations for performance improvement in hosted databases. This involved index creation, index removal, index modification, file group modifications, and adding scheduled jobs to re-index and update statistics in databases.
  • Developed and optimized database structures, stored procedures, Dynamic Management views, DDL triggers and user-defined functions.
  • Installation of replication server configuration and monitoring the replication between two sites implemented Transactional Replication.
  • Managing the clustering environment.
  • Implemented replication (transactional and snapshot) between Main production server and offsite (disaster recovery) server.
  • Involved in Source Data Analysis, analysis and designing mappings for data extraction also responsible for Design and Development of SSIS Packages to load the Data from various Databases and Files.
  • Performed daily tasks including backup and restore by using SQL Server 2012 tools like SQL Server Management Studio, SQL Server Profiler, SQL Server Agent, and Database Engine Tuning Advisor
  • Transferred data (ETL) to Data ware house Environment using SSIS from Oracle, PostgreSQL.
  • To analyze long running slow queries and tune the same to optimize application and system performance.
  • Expertise in creating and supporting in OLAP Cubes.
  • Expertise in SQL Server Analysis Services.
  • Involved in development activities included hardware and network specification, installation and configuration, and participation in database design, architecture and SQL scripting.
  • Did trouble shooting of Client connections, viewing error log and applications.
  • Designed coded and tested several T-SQL Programs.
  • Created SSIS Packages for uploading various formats of files and databases to MS SQL.
  • Monitoring Oracle RAC and SQL Server cluster environment with Foglight performance analysis
  • Performed backup and recovery using RMAN and Flask Recovery Area
  • Created and configure new oracle database as required using DBCA
  • Perform database tuning and assist developing during testing and implementation of the application.
  • Managed database security by creating and assigned appropriate roles and privileges to user.
  • Quest SQL Lite Speed used for backup speed and compression
  • SharePoint Database administrator for Enterprise Services department Consisting of production support team and Enterprise Services Management
  • Developed and introduced a set of basic SQL Standards for Implemented Centralized DBA Management Server.
  • Expertise in Hadoop ecosystems HDFS, Map-Reduce, Hbase, Pig, Sqoop and Hive for scalability, distributed computing and high performance computing

Confidential

SQL Server BI/DBA

Responsibilities:

  • Designed SSIS Packages for importing and Exporting data migration and conversion
  • Experience in Creating and maintaining reports in SSRS using RB 2.0 and Visual Studio
  • Develop reports based on pre-defined requirements
  • Develop stored procedures using MSSQL Server 2008
  • Create solutions to facilitate data analysis and ad hoc reporting
  • Creating data cubes for data analysis in SSAS
  • Create dashboard functionality for executives, directors, and managers with drill-down capabilities
  • Responsible for Backup and Recovery. Monitored the server for tuning and memory Management
  • Created an automated SSIS package which will call many packages based upon several conditions
  • Day to day administration of the SQL Servers (Backup and Restores, Scheduled jobs, Database Tuning, Query Tuning, Performance Monitoring).
  • Responsible for creating Databases, Tables, Cluster/Non-Cluster Index, Unique/Check Constraints, Views, Stored procedures, Triggers
  • Responsible for creating and updating report in MS-ACCESS and MS-EXCEL

Confidential, Irvine, CA

SQL Server Database Administrator/Business Intelligence

Responsibilities:

  • Installed and maintained MS SQL Server 2005 on Windows 2003 Server Worked on SQL Server 2005 failover clustering which is built on top of a Windows 2003 server cluster
  • Used Erwin and UML for physical and logical data modeling.
  • Configured and maintained database mirroring/log shipping, clustering, replication for high availability
  • Extensively worked on Dynamic Management Views and Dynamic Management Functions
  • Utilized Expertise in creating and supporting in OLAP Cubes in SSAS
  • Generated and managed reports using SQL Server 2005 Reporting Services
  • Transferred data (ETL) to Data warehouse Environment using SSIS
  • Dropped and recreated indexes on tables for performance improvements. Used Index Tuning Wizard for tuning of indexes
  • Scheduled the backups for Databases and Transaction log, performed restoration operations and suggested backup strategies
  • Traced long running queries and deadlocks using SQL Profiler
  • Managed use of disk space, memory and connections
  • Created and managed schema objects such as tables, views, indexes, procedures, and triggers & maintained Referential Integrity
  • Developed, deployed and monitored SSIS Packages including upgrading DTS to SSIS
  • Maintained the database consistency with DBCC at regular intervals

Confidential, Glenview, IL

SQL Server Database Administrator/Business Intelligence

Responsibilities:

  • Maintained logins and roles and managed Day to Day user maintenance
  • Managed the migration of SQL Server 2000 database to SQL Server 2005
  • Ensured availability and performance of databases that supported the system
  • Scheduled and monitored all maintenance activities including database consistency checks and index de-fragmentation
  • Created Jobs to Automate and scheduled database administrative tasks
  • Analyzed and tuned long running slow queries to optimize application and system performance
  • Developed OLAP cubes using SSAS and Calculate the metrics that were revealed in the functional Analysis
  • Developed Report using SQL Server Reporting Services(SSRS)

Confidential

Database Administrator

Responsibilities:

  • Prepared and delivered lecture to undergraduate students, average class size: 60-70 students
  • Maintained a teaching assignment for each module
  • Taught various IT course: Data Structures, Programming languages, RDBMS, Operating Systems, and Software Engineering etc. for Bachelor Degree students
  • Maintained the student data in the college.
  • Handled day to day activities using SQL Server and tools such as backup and Recovery.
  • Wrote T-SQL Queries for retrieval of data.

We'd love your feedback!