Senior Data Engineer/data Migration Resume
Vienna, VA
OBJECTIVE
- A challenging Data Engineer that will utilize my extensive technical skills and will lead me to innovative work environment.
SUMMARY
- Over 15 experience in IT as Data Architect/Database Administration/Data Engineer experience which includes Three years of experience in Big Data Ecosystem and Data Engineering technologies.
- Excellent knowledge in Hadoop Architecture and its major component like Hadoop Map Reduce, HDFS Frame work, HIVE, PIG, Impala, Sqoop, Flume, Hbase, Oozie and Zookeeper.
- Extensive experience in Data Modeling, Database Design, Data Architecture, Involved in several Enterprise wide Technology Architecting and Implementation.
- Experience in importing and exporting data using Sqoop and Flume from HDFS to Relational Database vs vice - verse.
- Extensive Experience in Development of Big Data projects using Hadoop, HIVE, Pig, Impala and other open source tools/technologies.
- Well versed in Installation, Configuration, Supporting and Managing of Big Data and Underlying Infrastructure of Hadoop Cluster.
- Managed the data management team of ETL Architects, database architects, middleware architects by providing architectural solutions for databases.
- Experience with version control tools like Git, SVN and TFS
- Designed Erwin logical and physical data model of the data warehouse.
- Responsible for RFC & RFP document and attending CAB, PAB for Approval
- Experience in high availability such as mirroring, log shipping, replication and clustering
- Extensive experience in Replication(Oracle Golden Gate and Double take)
- Skilled and Experienced in Server performance tuning using SQL Server 2005’s Dynamic Management View and performance counters in SQL SERVER 2012/2008/2005
- Good experience in Coding and Improving Efficiency of Rebuild index, Re-indexing, Updating Statistics, Recompiling Stored Procedures and performing other maintenance tasks.
- Experience in maintaining the security for databases (creating logins for users, roles, assigning permissions to them) and Encrypt the Database
- Good Experience in Database Backup, Recovery and Restoring as daily activities.
- Good Knowledge languages such as Scala and hand-on Experience with Python.
- Expertise in ETL tools (SSIS, Pentaho, Attunity), SQL Server reporting Services(SSRS), SQL Server Analysis Services(SSAS), SAS, Confidential VA, Confidential EG.
TECHNICAL SKILLS
Databases: MS SQL Server, Oracle, HBase, Informix, Refshift
Programming: Python, Scala, C, C#, C++, Java, COBOL
Database Administrator: MS SQL Server 2014/ 2012/2008/2005 , Oracle
Data Warehousing: DMX-h, SSIS/SSAS, Informatics, PIG, HIVE, Impala
Data Analytics: SAS, Confidential VA, Confidential EG, Tableau, Crystal Report
Operating System: Windows, Linux
Other Software: IIS, Microsoft Transaction Server, VBScripts, HTML
Tools: SQL DB Access, Performance Monitor, Backup Utility, SQL Lite- Speed
Database Modeling: Erwin 9.5, Visio, UML
Virtual Tools Server: Microsoft Hyper-V, VMWARE
Network Topologies: TCP/IP, DNS, DHCP, FTP, SMTP
PROFESSIONAL EXPERIENCE
Confidential, Vienna, VA
Senior Data Engineer/Data Migration
Responsibilities:
- Understand the Business Requirements and develop designs both at Low Level and high level for the ETL solutions Proposed.
- Worked on design and developing Hadoop solution using Syncsort DMX-h, Hive, Impala and HDFS on Cloudera Framework in Agile-based environment.
- Designed and implemented ETL processes using Syncsort DMX transformation tool
- Created Sqoop jobs to replicate data between Hive and Oracle databases.
- Analyzed and processed different file formats: EBCDIC files, fixed width text files, XML files Performed required actions on Hadoop file system.
- Defined and developed batch schedule using CA Workload scheduling tool.
- Created UNIX scripts to perform files manipulation, archival, batch execution orchestration and other command line level activities.
- Design and develop Hive/Oracle SQL (DDL and DML) statements, stored procedures, functions, views and triggers to be used during the ETL processes
- Created several Hive tables, implementing dynamic and static partitioning, and bucketing for efficient data access.
- Worked on creating and maintaining documentation of the ETL processes low level designs, flow diagrams
- Collaborated with business solution analysts, other developers, solution architects and other team members to deliver sound solutions
- Performed Unit testing to ensure that the code is working as per requirements and efficiently
Confidential
Senior Data Engineer
Responsibilities:
- Work closely with the business and Advanced Analytics team in gathering the system requirements.
- Understand the Business Requirements and develop designs both at Low Level and high level for the ETL solutions Proposed
- Worked on design and developing Hadoop solution using Syncsort DMX-h, Hive, and HDFS on Hortonworks Framework in Agile-based environment
- Designed and implemented ETL processes using Syncsort DMX/SSIS transformation tool
- Created Sqoop jobs to replicate data between Hive and Oracle databases.
- Responsible for establishing, growing and maturing data integration development and support capabilities to deliver measurable business value.
- Involved in development, building, testing, and deploy to hadoop cluster in distributed mode.
- Participate in POC/POT efforts to integrate new Big Data Management technologies, Software engineering tools and new patterns into existing structures
- Responsible for storing data in S3 and created table in AWS Athens for Adhoc queries.
- Responsible for create Data warehouse solution in AWS Redshift
- Handled Importing of data from various data sources, performed transformation using Hive, loaded data into HDFS and extracted the data from SQL Server, PostgreSQL, and oracle using sqoop
- Create Sqoop Scripts for loading data into Hadoop and Schedule into Python.
- Create Custom ETL Process to move data from Mainframe System to Enterprise Data Lake using DMX-h Syncsort
- Handle Multivariable VSAM File in the Syncsort & Talend and loading Data to HIVE
- Created the layout for the ASCII file in the Syncsort and Transform data to EDL
- Experienced in parsing the VSAM copy book and apply copy to EBCDIC file format
- Experienced in Converting File from EBCDIC to UTF8 and loading Data in EDL
- Handle heavy data volume on the daily basis and optimize query and ETL Process.
- Analyzed the data by performing Hive Queries and Running Pig scripts for fraud analysis
- Experience in Oozie and workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph(DAG) of action with control flows
- Maintain involvement in continuous improvement of Big Data Solution processes, tools and templates.
Confidential
Senior Database Architect
Responsibilities:
- Involved in Installation, Configuration, supporting and managing Hadoop Cluster using Cloudera Manager 5.3 and CDH 5.2 and Hortonworks
- Installing and configuring Hadoop eco system like sqoop, pig, hive, Impala.
- Data Requirement Analysis, Data Modeling and Data Architecture for several projects.
- Involved in Designing the architecture for Enterprise data warehouse and Business Intelligence Environment
- Responsible for Driving Data Migration project as part of the Labyrinth project for the Confidential division of the Ministry of Health.
- Involved in all phases of SDLC including analysis, designing, developing, testing and maintenance.
- Participated in development of request for proposal (RFP) documents.
- Managed the data management team of ETL architects, database architects, middleware architects by providing architectural solutions for databases
- Involved in participating the design reviews, Regular and ad-hoc architectural reviews, and ensure correct interpretation of the Data Architecture, and applied strategies.
- Prototyped a Master Data Services (MDS) Solution and presented it to Management.
- Experienced in loading and transforming large sets of structured, semi-structured and unstructured data.
- Responsible for building scalable distributed data solutions using Cloudera Distribution on Premise and Deploying Hadoop with Cloudera CDH to AWS.
- Responsible for Cluster maintenance, managing cluster nodes.
- Analyzed data using Hadoop components Hive and Pig and Impala.
- Experienced in working with Amazon Web Services (AWS) using EC2 for computing and S3 as storage mechanism for Hadoop Cloudera.
- Experienced in loading and transforming large sets of structured, semi-structured and unstructured data Hadoop concepts
- Involved in creating Hive tables, and loading and analyzing data using hive queries and Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Hands on Experience Python Panda Frame to provide data analysis.
- Involved in importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS using SQOOP and Flume
- Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 11ig database.
- Experience in migrating the data using Sqoop from HDFS to Relational Database System PostgreSQL and vice-versa according to client's requirement.
Confidential
Senior Database Administrator/Analyst
Responsibilities:
- Involved in installation and configuration of SQL Server 2012/2008R2/2008/2005 in Windows environment.
- Involved in data modeling for the SQL Server and for individual databases.
- Involved in scheduling backups of databases and logs and restoring them when necessary.
- Managing the use of disk space memory and connections.
- Responsible for monitoring and making recommendations for performance improvement in hosted databases. This involved index creation, index removal, index modification, file group modifications, and adding scheduled jobs to re-index and update statistics in databases.
- Developed and optimized database structures, stored procedures, Dynamic Management views, DDL triggers and user-defined functions.
- Installation of replication server configuration and monitoring the replication between two sites implemented Transactional Replication.
- Managing the clustering environment.
- Implemented replication (transactional and snapshot) between Main production server and offsite (disaster recovery) server.
- Involved in Source Data Analysis, analysis and designing mappings for data extraction also responsible for Design and Development of SSIS Packages to load the Data from various Databases and Files.
- Performed daily tasks including backup and restore by using SQL Server 2012 tools like SQL Server Management Studio, SQL Server Profiler, SQL Server Agent, and Database Engine Tuning Advisor
- Transferred data (ETL) to Data ware house Environment using SSIS from Oracle, PostgreSQL.
- To analyze long running slow queries and tune the same to optimize application and system performance.
- Expertise in creating and supporting in OLAP Cubes.
- Expertise in SQL Server Analysis Services.
- Involved in development activities included hardware and network specification, installation and configuration, and participation in database design, architecture and SQL scripting.
- Did trouble shooting of Client connections, viewing error log and applications.
- Designed coded and tested several T-SQL Programs.
- Created SSIS Packages for uploading various formats of files and databases to MS SQL.
- Monitoring Oracle RAC and SQL Server cluster environment with Foglight performance analysis
- Performed backup and recovery using RMAN and Flask Recovery Area
- Created and configure new oracle database as required using DBCA
- Perform database tuning and assist developing during testing and implementation of the application.
- Managed database security by creating and assigned appropriate roles and privileges to user.
- Quest SQL Lite Speed used for backup speed and compression
- SharePoint Database administrator for Enterprise Services department Consisting of production support team and Enterprise Services Management
- Developed and introduced a set of basic SQL Standards for Implemented Centralized DBA Management Server.
- Expertise in Hadoop ecosystems HDFS, Map-Reduce, Hbase, Pig, Sqoop and Hive for scalability, distributed computing and high performance computing
Confidential
SQL Server BI/DBA
Responsibilities:
- Designed SSIS Packages for importing and Exporting data migration and conversion
- Experience in Creating and maintaining reports in SSRS using RB 2.0 and Visual Studio
- Develop reports based on pre-defined requirements
- Develop stored procedures using MSSQL Server 2008
- Create solutions to facilitate data analysis and ad hoc reporting
- Creating data cubes for data analysis in SSAS
- Create dashboard functionality for executives, directors, and managers with drill-down capabilities
- Responsible for Backup and Recovery. Monitored the server for tuning and memory Management
- Created an automated SSIS package which will call many packages based upon several conditions
- Day to day administration of the SQL Servers (Backup and Restores, Scheduled jobs, Database Tuning, Query Tuning, Performance Monitoring).
- Responsible for creating Databases, Tables, Cluster/Non-Cluster Index, Unique/Check Constraints, Views, Stored procedures, Triggers
- Responsible for creating and updating report in MS-ACCESS and MS-EXCEL
Confidential, Irvine, CA
SQL Server Database Administrator/Business Intelligence
Responsibilities:
- Installed and maintained MS SQL Server 2005 on Windows 2003 Server Worked on SQL Server 2005 failover clustering which is built on top of a Windows 2003 server cluster
- Used Erwin and UML for physical and logical data modeling.
- Configured and maintained database mirroring/log shipping, clustering, replication for high availability
- Extensively worked on Dynamic Management Views and Dynamic Management Functions
- Utilized Expertise in creating and supporting in OLAP Cubes in SSAS
- Generated and managed reports using SQL Server 2005 Reporting Services
- Transferred data (ETL) to Data warehouse Environment using SSIS
- Dropped and recreated indexes on tables for performance improvements. Used Index Tuning Wizard for tuning of indexes
- Scheduled the backups for Databases and Transaction log, performed restoration operations and suggested backup strategies
- Traced long running queries and deadlocks using SQL Profiler
- Managed use of disk space, memory and connections
- Created and managed schema objects such as tables, views, indexes, procedures, and triggers & maintained Referential Integrity
- Developed, deployed and monitored SSIS Packages including upgrading DTS to SSIS
- Maintained the database consistency with DBCC at regular intervals
Confidential, Glenview, IL
SQL Server Database Administrator/Business Intelligence
Responsibilities:
- Maintained logins and roles and managed Day to Day user maintenance
- Managed the migration of SQL Server 2000 database to SQL Server 2005
- Ensured availability and performance of databases that supported the system
- Scheduled and monitored all maintenance activities including database consistency checks and index de-fragmentation
- Created Jobs to Automate and scheduled database administrative tasks
- Analyzed and tuned long running slow queries to optimize application and system performance
- Developed OLAP cubes using SSAS and Calculate the metrics that were revealed in the functional Analysis
- Developed Report using SQL Server Reporting Services(SSRS)
Confidential
Database Administrator
Responsibilities:
- Prepared and delivered lecture to undergraduate students, average class size: 60-70 students
- Maintained a teaching assignment for each module
- Taught various IT course: Data Structures, Programming languages, RDBMS, Operating Systems, and Software Engineering etc. for Bachelor Degree students
- Maintained the student data in the college.
- Handled day to day activities using SQL Server and tools such as backup and Recovery.
- Wrote T-SQL Queries for retrieval of data.