Big Data Engineer Resume
NY
EXPERIENCE SUMMARY:
- 9+ years of experience in the field of Software Design, Development and Implementation life cycle (SDLC) including analysis, design, architecture, development, testing, implementation, maintenance and production support of application software.
- Extensively working on Apache Spark, Scala, Java, Hive, SBT, HDFS, YARN, Impala
- Proficient on Scala functional programing language.
- Knowledge of NoSQL database like Hive, MySQL, MongoDB etc.
- Knowledge of Data ingestion using Apache Sqoop, Apache Kafka and Flume
- Expensively working on structured, unstructured data with various file formats such as Avro data files, XML files, JSON, sequence files, ORC and Parquet
- Strong knowledge in SQL Server 2005/2008 database programming using T - SQL (Stored Procedures, Functions, Triggers, Views, Cursors, CTE’s) and standard Built-in SQL Server packages
- Expert in RDBMS logical and physical Database Designs with excellent Normalization and De-normalization techniques for efficient data storage and retrieval
- Extensive experience in Microsoft BI Technologies like SQL Server Integration Services (SSIS), SSMS, SQL Profiler, DTS packages.
- Experienced in working with utilities like BCP, External Tables, Import, and Export to extract and load large volume of data. Extensively working on handling XML/JSON Data to Import and Export.
- Experienced in Query Optimization, Debugging, Performance Tuning and Improvising performance of applications using SQL Profiler and Execution Plan to analyze queries for efficient data retrieval.
- SQL Server database Migration.
- Providing assistance in logical/physical data modeling using ERWIN.
- Performance tuning using Execution Plan, SQL-Trace, Profilers SQL Server utilities and creating relevant system and user’s documentation
- Experience in Sybase ASE 12.5, Sybase ASE 15.7, Also have experience in Sybase Database migration
- Thorough in Unit testing and Integration Testing
- Efficient in Interacting with users and help analyze business processes and operations
- Great Team player with Strong interpersonal skills, communication skills and documentation skills
TECHNICAL SKILLS:
Big Data Ecosystem: Apache Spark, Sqoop, Flume, Kafka, SBT, HBase, HDFS, Avro, Parquet, YARN, Scala, Python, Hive, Impala
Hadoop Distribution: Cloudera CDH 5.8.2, CDH 5.4.0, Hortonworks
Scheduling Tools: Autosys, TWS, Oozie
NO SQL: MySQL, Apache Hive, MongoDB
Programming Language: Scala, Python, C#, Java, UNIX Shell Scripting, Perl Scripting, SQL
RDBMS: Microsoft SQL Server 2000/2005/2008 , Sybase ASE
Database Tools: Microsoft SQL Server 2005/2008 SSMS, Rapid SQL, DB Artisen 9
Operating Systems: UNIX, Windows NT/2000/2003/XP/7
Data Modeling: Erwin, Microsoft Visio 2003
Web Methods: Integration Server: 4.6, 6.x, Developer 3.5, 6.x
ETL Tools: Microsoft SQL Server Integration Services (SSIS)
PROFESSIONAL WORK EXPERIENCE:
Confidential, NY
Big Data Engineer
Responsibilities:
- Coordinate with the stakeholders to identify their requirements
- Work on the requirements and align them with the offshore team.
- Develop Spark jobs using Scala to process the data into Data Lake (Hadoop Cluster)
- Configure Kafka to get the LaaS entitlement data using Kafka topics.
- Create Hive tables, loaded data and wrote Hive queries that run within the Map.
- Worked on partitioning Hive tables and running the scripts in parallel to reduce run time of the scripts.
- Integrate Hive tables with Impala and manage the regular refresh and sync.
- Import the data from different sources like Kafka/HDFS into Spark RDD.
- Develop Kafka topic in Java to get the data from producer.
- Configure the OOZIE jobs and schedule.
Environment: Cloudera CDH 5.8.2, Hadoop 2, HDFS, Spark 2.2, Scala, Java, Kafka, Hive, HiveQL, Oozie, Sqoop, Impala, Tradmill, Git, HBase.
Confidential
Big Data - Hadoop Developer
Responsibilities:
- Research existing Financial Advisor data from IM Silo.
- Develop Spark jobs using Scala to process the data.
- Import the data from various relational database SQL Server, Sybase into Hadoop Cluster using Sqoop.
- Working on PoC for Confidential using SQOOP, Hive, HDFS, Spark, HiveQL, Java and Scala
Environment: Cloudera CDH 5.4.0, HDFS, Spark 1.6, Scala, Sqoop, Tradmill, GitHub
Confidential
SQL Database Developer
Responsibilities:
- System Study of existing application and Requirement analysis from Business Users.
- Database Design, development and maintenance.
- Writing database functions, scripts, stored procedures, views, triggers etc using T-SQL
- Creating and maintenance of Indexes on the database tables.
- Develop, Maintain and enhance the SSIS Packages for data ETL.
- Writing BCP IN/OUT data load jobs, Bulk Data update scripts (XML, Java, Perl, Unix Scripting and Autosys).
- Write T-SQL complex stored procedures to load the data from different sources.
- Database maintenance and clean-up.
- Writing and optimizing database indexes.
- Strong experience in Performance Tuning and working with the optimization of query plans using Execution Plan.
- UAT and Production deployment of the application, push all the development stuff to the UAT and Production Parallel environment.
- Documenting the processes and storing them in Perforce as well as on wiki site.
- Experience with CRUD (Create, Read. Update & Delete) data cycles.
- Co-ordinate with the support team and help them on issues.
- Extensively used SQL Scripts for developing reports
- SQL Server 2005/2008 Database Migration
- Sybase 12.5 to 15.7 database server migration.
Environment: MS SQL Server 2005/2008, SQL Profiler, T-SQL, Java, SSMS, SSBI - Development Studio, UNIX Shell and Perl Scripting, BCP’s, DB Artisan for Sybase, SYTS, Putty, SSIS, CA ERwin Data Modeler, Autosys and Open Text Exceed, XML
Confidential
SQL Database Developer
Responsibilities:
- System Study of existing application and Requirement analysis from Business Users.
- Database Design, development and maintenance.
- Writing database Functions, SQL - scripts, Stored Procedures, Views, Triggers etc using T-SQL
- Develop, Maintain and enhance the SSIS Packages for data ETL.
- Writing BCP IN/OUT data load jobs, Bulk Data update scripts (Java, Unix Scripting and Autosys).
- Write T-SQL complex stored procedures to load the data from different sources.
- Database maintenance and clean-up. Performance Tuning and optimization of query plans. iPRAM monthly data load implementation
- Experience with CRUD (Create, Read. Update & Delete) data cycles.
- Extensively used SQL Scripts for developing reports
Environment: MS SQL Server 2005/2008, T-SQL, UNIX Shell and Perl Scripting, DB Artisan for Sybase, SYTS, Putty, CA ERwin Data Modeler, XML
Confidential
SQL Database Developer
Responsibilities:
- System Study of existing application and Requirement analysis from Business Users.
- Writing database functions, scripts, stored procedures, views, triggers etc using T-SQL
- Creating and maintenance of Indexes on the database tables.
- Develop, Maintain and enhance the SSIS Packages for data ETL.
- Writing BCP IN/OUT data load jobs, Bulk Data update scripts (XML, Perl, Unix Scripting and Autosys).
- Write T-SQL complex stored procedures to load the data from different sources.
- Strong experience in Performance Tuning and working with the optimization of query plans.
- Experience with CRUD (Create, Read. Update & Delete) data cycles.
- Extensively used SQL Scripts for developing reports
Environment: MS SQL Server 2005/2008, SQL Profiler, T-SQL, SSMS, SSBI - Development Studio, UNIX Shell and Perl Scripting, BCP’s, DB Artisan for Sybase, SYTS, Putty, SSIS, CA ERwin Data Modeler, Autosys and Open text Exceed, XML
Confidential
SQL Database Developer
Responsibilities:
- Writing database functions, scripts, stored procedures, views, triggers etc using T-SQL
- Develop, Maintain and enhance the SSIS Packages for data ETL.
- Writing BCP IN/OUT data load jobs, Bulk Data update scripts (XML, Perl, Unix Scripting and Autosys).
- Write T-SQL complex stored procedures to load the data from different sources.
- Strong experience in Performance Tuning and working with the optimization of query plans.
- Extensively used SQL Scripts for developing reports
Environment: T-SQL, UNIX Shell and Perl Scripting, MS SQL Server 2005/2008, DB Artisan for Sybase, SYTS, Putty, SSIS, CA ERWIN, XML.
Confidential
SQL Database Developer
Responsibilities:
- System Study of existing application and Requirement analysis from Business Users.
- Writing database functions, scripts, stored procedures, views, triggers etc using T-SQL
- Write T-SQL complex stored procedures to load the data from different sources.
- Strong experience in Performance Tuning and working with the optimization of query plans.
- Experience with CRUD (Create, Read. Update & Delete) data cycles.
- Writing Data marts and categories in ESP (State Street Data Warehouse Tool)
- Extensively used SQL Scripts for developing reports
Environment: T-SQL, UNIX Shell and Perl Scripting, MS SQL Server 2005/2008, DB Artisan for Sybase, SYTS, Putty, Visual Studio 2005, CA ERWIN
Confidential
Database and UI Developer
Responsibilities:
- Writing Database Schema Design, Stored Procedure, Triggers, Views, Functions and Indexes.
- Design UI and code using Ext JS code.
- Performing Impact Analysis of the requirements for existing applications and review the estimations and do re-estimation, if needed.
Environment: T-SQL for Sybase Adaptive Servers, Ext. JS as a functional language, MSDE Eclipse 3.7prod, DB Artisan (Sybase), MS SQL Server 2005, Visual Studio 2005, SYTS
Confidential
Database and UI Developer
Responsibilities:
- Writing Database Schema Design, Stored Procedure, Triggers, Views and SQL CLR Functions.
- Develop new modules based on the new requirement provided by Business user/Client.
- UAT and Production deployment of the application, push all the development stuff to the UAT and Production.
Environment: T-SQL for SQL Server 2005, SQL Programming, C#.Net, Ext. JS, Java Script, Web Services
