Technical Lead Resume
Rochester, MN
SUMMARY
- Expertise in data warehousing wif 5.5 years in OLTP environments. Expert in writing SQL, SSIS packages, SSRS reporting. Proficient in analysing healthcare domain data from systems like Epic, Cerner and McKesson.
- Experienced analysis of healthcare data involving patient encounter, appointments, provider schedules, patient charges and transactions, hospital supplies, diagnosis and procedures codes, ATB accounts.
- Worked in agile and scrum processes to analysis, design, deploy and support in data warehousing. Collaborating wif Data Architect and DBA for database modelling, database capacity planning, gap analysis, data retention policy and performing code review.
- Experience in data conversion and data migration using SSIS packages to transfer data between Oracle, DB2, flat files, and reports to SQL Server using SSIS.
- Experience in logical and physical Dimensional Modeling, Star Schema / Snowflake Schema, Fact and Dimensional tables and Operational Data Store (ODS).
- Thorough noledge of Machine Learning and writing algorithms like linear, logistic regression, neural network, k - means, SVM, PCA anomaly detection and collaborative-filtering algorithms.
- Hands on experience on spark libraries like pySpark to create machine learning models and evaluate on test set. Theoretical understanding of analysing error in model using confusion matrix, ROC Curve or regression evaluation method.
- Familiarity wif Hadoop architecture such as HDFS, Name Node, Data Node, Job Tracker, Task Tracker.
- Strong experience in client interaction and understanding business applications, Data Modeling, Data Mapping, Table Normalization, Optimization and Tuning.
- Well-organized, goal-oriented, highly motivated TEMPeffective team member wif excellent analytical, troubleshooting, and problem solving skills.
- Practical hands on creating spark sessions and creating data frames from loading files.
- Massaging the data to handle missing data, grouping by, using spark.sql or data columns to filter data.
- Used python and pySpark to preparing data pipeline for data model predictions like Linear and Logistic regression. Used various classes like assembler, StringIndexer and OneHotIndexer to prepare pipeline features from raw data.
- Used Neural Networks using forward and backward propagations to predict handwritten numbers. And tested the algorithms wif gradient checking to make sure the cost function is working correctly.
- Build a spam email classifier using support vector algorithm wif simple RBF (radial-bias function) which uses Gaussian kernel.
- Used K-means algorithm to compress images by reducing the number of color that occurs in the image. Applied regression by randomly selecting the cluster centroid points and minimizing the cost function by calculating the mean distance. Improved the performance of the model by optimizing number of clusters involved and the mean of the cluster.
- Used anomaly detection algorithm to detect anomalous ping from server computers data. Optimized the algorithm by estimating the regularized and sigma parameters based on the F1 score.
- Build PCA algorithm to reduce the dimension of face images dataset from 1024 to 100 dimension as such that it retained 95% of the variation in dataset.
- Completed basic python programs using strings, functions, lists, dictionaries, and series and dataframes data structures.
- Used Pandas and Numpy library for math functions; import and manipulate data files; apply indexing and querying; apply merge and join functions; and slicing and dicing on dataframes.
- Analyse data wif group-by, and applied aggregate functions to dataframes.
- Interpret data to evaluate hypothesis tests like Null Hypothesis.
TECHNICAL SKILLS
Big Data Tools: Hadoop, Spark, Linux, Octave, R, Python, Juypter, (libraries like) spark.sql, Numpy, Pandas, Pyspark
Languages: T-SQL, Linux, Unix Scripting, C, C++, Core Java, C#
Tools & Database: SQL Server 2014/2012, BIDS 2010, Oracle 9i/10g, DB2 10.5, SQL Performance Dashboard, SQL Execution Plans, SQL Server Profiler, RapidSQL, BIDS, R-Studio, Winscp API, Informatica 10.1, FastTrack
Management Tools: IBM IDA (InfoSphere Data Architect), JIRA, GIT Repository, MS Excel, FastTrack
PROFESSIONAL EXPERIENCE
Confidential, Rochester, MN
Technical Lead
Responsibilities:
- My role involves monitoring and improving data quality for hospital supplies and access management for epic. Analysing and validating data, metrics, and finding solutions to outstanding quality and data validating issues.
- Perform data validation and resolving epic data discrepancies for Hospital Supplies. Validated data from old SIMS (supply inventory management systems) in the database to replicate similar grain form data source.
- Developing source queries that involves supplies, manufactures, supply types/charges, location, departments, sites, specialty and providers and procedures from Epic database.
- Created complex Stored Procedures, Triggers, Functions (UDF), Indexes, Tables, Views and other SQL code and SQL joins for the applications and following best practices.
- Coding complex transformations to clean data from source before loading data into staging tables. Resolving data discrepancy at all layers of data models that is at staging, mart model and in reports.
- Coordinating wif client to analyze the gaps in the reports and wif find the source and wif ETL developer to fix the gaps.
- Analyzing access management data which involves patient appointments, encounters, provider care teams, sites and departments. Analyses provider data that involves monitoring of staff on floor time, specialty / procedure time, inpatient hospital time, offsite care time, and personal time.
- Monitoring data quality of the providers visit types and unavailable reason free text of providers coming from epic templates and Cerner schedules block. Also coding the missing visit types to their correct time slots.
- Also monitoring data for providers who use specialty visit types or blocks in primary care departments. Provided monthly reports to management which helped in training operational staff on using correct visit types.
- Coordinated wif SME’s to created source-to-target mapping and database design using UDMH and IDA. Data analyzes included identifying correct business Keys, lookup’s, business logics, and expressions to map correct data using UDMH framework.
- Providing walkthrough sessions to handover model design and requirements to ETL and QA team. Design included entities like anchor, detail, bridge and array tables.
Confidential
MS SQL 2014 Standard Edition
Responsibilities:
- Worked as Lead modeler in the project to create transactional data model for CCDA (consolidated clinical document architecture) files which has patient history like allergies, medication, problem list, procedure, encounters, Payers, Family history etc.. And to store logs, messages, and subscription notification involving 150 tables covering 2500 attributes.
- Performing database and data migration (23 million rows of 75 GB which includes 36 GB BLOB data) using scripts to move from phase-2 to phase-3 of project into production. Base script and delta scripts were versioned in sync wif every application bugs fixing releases.
- Optimizing the database performance by implementing table partitioning, tuning archiving stored procedures, UI queries in order to handle up to 33GB of data daily.
- Improving performance of application wif indexing, setting fill-factor, data-page compression which reduced storage by 17%, tuned application queries to do index seek for fast data retrievals, and avoiding datatype conversions.
- Creating SSIS packages to extract flat-files and sent it over FTPS; created POC model to load delimited files into database using SSIS packages.
- Monitoring database performance during volume testing of application high peak load of 35000 mgs/hour; archiving; and archiving performance of peak load of 18000 messages/hour.
- Created ETL Packages using SSIS to extract data from staging tables to partitioned tables wif incremental load.
- Created SSIS packages wif error handling and worked wif different methods of logging in SSIS.
- Created SSIS Packages using Pivot Transformation, Fuzzy Lookup, Derived Columns, Conditional Split, Term extraction, Aggregate, Execute SQL Task, Data Flow Task, and Execute Package Task etc. to generate underlying data for the reports and to export cleaned data from Excel Spreadsheets, Text file, MS Access and CSV files to data warehouse/OLTP database.
- Implemented complex business logic wif user-defined-functions, indexes, views and clustered & non-clustered indexes.
Confidential, Parsippany, N
Database Developer II, Database Developer me
Responsibilities:
- Performed reconciliation of charges and transaction of patient accounts, aged trial balances. Developing code to extract data elements from data source according to data model. Developing complex stored procedure from source data to populate data elements of data model.
- Performing data validation of patient accounts, healthcare data codes from client’s patient systems and writing logic to resolve them.
- Developed end-to-end ETL solutions using SSIS packages, performed full data validation, changed during test phase and data analysis phase, scheduled the packages into production and provide support.
- Built efficient SSIS packages for processing fact and dimension tables wif complex transformations and type 1, type 2 and type 3 changes.
- Optimized stored procedures using indexes, reducing complex joins to temporary or derived tables, using CTE, partition by functions, and merge statement.
- Used staging tables, table variable and temporary tables to improve the query performance.
- Developed new mappings using SSIS Look Up/Aggregation transformations to match the needs of the business case.
- Designed and developed dynamic SSIS packages to process dynamic queries and dynamically produce flat files at runtime.
- Involved in error and event handling: precedence constraints, break points, check points and logging.
- Implemented TEMPeffective error handling to notify the support person about any errors raised in SSIS package during the extraction process by sending emails wif complete error details.
- Defined the logging process for all the packages, populated it wif entries like inserted record, updated record, error record and run-time records.
- Scheduled and maintained packages by implementing execute SQL server agent job task, history clean up, maintenance cleanup task, shrink database, rebuild index tasks for metadata and data quality
- Involved in maintenance and administration of SSIS by creating jobs, alerts, SQL mail agent, and schedule SSIS Packages.
- Developed error handling logics in all the SSIS packages to handle bad data and avoid package failures.
- Implemented logging tables which maintained time taken for control tasks and procedures for all the packages and stored procedures. dis helped tremendously in checking time taken by procedure and optimize them.
- Used CTE that improved the performance of complex logic by over 30% in time. Developed complex queries that validated the anomalies in charges or transactions.
- Implemented error handling logics in all the SSIS packages to handle bad data to error flow and avoid package failures.
Confidential, Malvern, PA
SQL Developer
Responsibilities:
- Fetched data from Teradata using Microsoft attunity connector and parse report files (non-delimited files). Perform data quality in staging tables to ensure correct data loaded for analytics.
- Created C# script to iterate over flat files and parse date, time, and decimal data values, change to ascii characters from report files provider by client.
- Creating ad-hoc queries and analyzing data quality wif respect to master codes, number of records, duplicates records, maximum-minimum values, syncing number of accounts across tables, null percentage in columns.
- Scheduling .dtsx packages using SQL job agent, monitoring and troubleshooting failed packages. Debugging and handling error in failed stored procedure in production server.
Confidential - Pittsburgh, PA
SQL Developer
Responsibilities:
- Performed data extraction and migration activities for populating test (or non-production) and customer databases using import export wizard.
- Exported/imported data from text files and Excel to SQL Server database using bulk insert & BCP utility.
- Creating TSQL queries and SQL Scripts to perform efficient data load based on complex business rules and produced results for users in SQL Server 2005.
- Worked on complex join, temporary table, derived table and retrieve data from multiple tables.
- Optimized the performance of queries wif modification in T-SQL queries, normalized tables, established joins and created indexes wherever necessary.
- Finding bottle necks in queries and stored procedures and fine tuning.