We provide IT Staff Augmentation Services!

Senior Data Scientist Resume

Detroit, MI


  • I am best known for my ability to rapidly deliver solutions that deliver significant business value.
  • Over the course of my career, I have been instrumental in delivering numerous cost and process improvement actions using targeted Confidential to drive business value.
  • I am passionate about building diverse teams with deep connections while finding innovative solutions to real world business challenges.
  • Strong in design and integration problem solving skills Interested in a challenging technical along with managerial track career in a Management and Confidential & Information Management.


BI Tools: Business Objects (Supervisor, Designer, Reporter, Desktop Intelligence, Web Intelligence, Central Management Console, Info View), SSRS

Confidential Tools: Alteryx, QlikView, Tableau, Sqoop, Pig and Hive, Machine Learning Tensor flow, Informatica BDM, NumPy, Pandas, Pytorch, PostgresSQL, MSSQL, MySQL

Languages, Software and Tools: Scala, Python, Java, C, C++, Microsoft, IBM compilers Rational Rose Enterprise Edition, Rational Clear Case

Live Feed/Stream Source and Logs Monitoring: Apache Kafka Kafka RDD, Kafka Streams, Attunity Spark library Apache Spark, Flask, Angular, Spring - Hibernate, Visual Studio Code.

Cloud Confidential: AWS stack - ECS-S3-RDS-EBS-EB-Autoscaling and ELoad balancers, PIvotal

Automation /Container Confidential: Docker Expert, Jenkins, Git

Databases: Hadoop, DB2, SQL server, Teradata, VSAM, IMS DB DC, Oracle

Scripting Languages: AIX, Sun Solaris, HP Unix Korn shell, Bash shells, Perl, Sed, Awk, Dos batch scripts. Html, JavaScript.

Operating Systems: Aix, HP Unix, Sun Solaris, RHEL, RT Linux, IBM, Windows

Mainframe Tools: CA- File Master, File-Aid, Endevor, CA7, Xpeditor, SMART test, DB2 Visual Studio, DB2 Connect, IBM Fault Analyzer, Move for DB2

Project Management: Project Planning, Monitoring and Management, Quality Assurance, Client Liaison, Proposal and RFE, Rally


Confidential, Detroit, MI

Senior Data Scientist

Responsibilities .

  • Responsible for modeling complex Institute problems, discovering insights and identifying opportunities with statistical, algorithmic, mining and visualization techniques.
  • Proficient at integrating and preparing large, varied datasets, designing specialized database and computing environments, and communicating results.
  • Work closely with clients, data stewards, project/program managers, and other IT teams to turn data into critical information and knowledge that can be used to make sound organizational decisions
  • Provide data that is congruent and reliable. Propose innovative ways to look at problems by using data mining method approaches on the set of information available.
  • Lead the design and deployment of enhancements and fixes to systems as needed and come up with solutions to loosely defined business problems by leveraging pattern detection over potentially large datasets using strong programming skills (such as Hadoop Map Reduce or other big data frameworks, Java), statistical modeling (like SAS or R,Informatica and Alteryx).
  • Consistently foster collaboration and respect among team members by addressing elements of the group process that impedes, or could impede the group from reaching its goal. Engage the “right people,” despite location or functional specialty, in the team by matching individual capabilities and skills to the team’s goals.
  • Tracking and synthesizes systemic benchmarking trends. Evaluates composite information in relation to its impact on decision-making and strategic implications. Sets expectations for and reviews management and key stakeholder reports. Assesses validity of business strategy recommendations against trend data. Steers senior leadership toward making informed, sound strategic decisions.
  • Lead the Informatica development team for data profiling and master data management using Informatica BDM. Also, key player in building the self-serve utility in processing the data from Hadoop and writing to business zone for further processing. Parse complex multi-structured, hierarchical, and unstructured data automatically with Informatica Intelligent Structure Discovery. Easily handle schema drifts.
  • Developed the new Spark jobs using Scala, to run on the HDP clusters, which provided significant gains on the completion times. Design development of Spark SQL Scripts based on Functional Specifications
  • Responsible for Spark Streaming configuration based on type of Input Source
  • Apache Kafka Streaming API for ingesting the data to Spark Streams and also publish to Kafka Topics for publishing the anomaly
  • Importing and exporting data into HDFS and HIVE, PIG using Spark and  cnfidential and Sqoop
  • Used Dynamic partition for Hive when loading data. Implemented External and Hive managed Tables for significant performance gains.
  • Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Spark jobs in the backend.
  • Writing Spark (Hadoop) programs to convert text files into AVRO and loading into Hive (Hadoop) tables
  • Implemented the workflows using Apache Oozie Spark library to automate tasks.
  • Worked with NoSQL databases like HBase, MongoDB in creating HBase tables to load large sets of semi structured data coming from various sources.

Environment: Hadoop, Alteryx, Tableau, R programming, SAS, PIG, Hive, Sqoop, QlikView, Scala, Spark, Kafka and Machine Learning Tensor flow, Informatica BDM,Hbase HortonWorks HDP, Kerberos, Hive, Splunk, Zeppelin, Oozie,Gitlab, Ambari\Tez Views, HBase, Linux, SQL Developer, SQL Server, Tableau,Qlikview,, NumPy, Pandas, Pytorch, PostgresSQL, MSSQL, MySQL

Confidential, Detroit, MI

Senior Technical Big Data Analyst


  • Reviewed the Architectural design, Conceptual Design Summary, UI Mock-Ups with the Solution Design Team. Communicating major milestones and providing direction to the project teams. Establishing status control procedures and conducting regular status meetings tracking and monitoring the project's financial health across the program.
  • Created design and technical user stories on various release using Rally. Also provides technical solution to improve the performance of the application and coordinate, train both offshore and onsite team. Used SAFe agile methodology and Kanban for story development and mapping
  • Subject matter expert to manage the entire development process as the development lead by ensuring both the onshore and the offshore developers are coding to standard while also keeping them on schedule. Collaborate with the Quality Assurance Analysts to develop test data gathering and validation strategy.
  • Leveraged Informatica Big Data Management tools to access, analyze, prepare, transform, and stream data faster than traditional data processing environments. ETL to read from and write to diverse big data sources and targets. Performed data replication and data discovery on a Hadoop cluster. Performed data lineage on big data sources from source to end-user data transformation.
  • Created and key integral player in creating data models using normalization for Relational Database and Schema based design for Warehouse data.
  • Create logical and physical data models using best practices to ensure high data quality and reduced redundancy. Optimize and update logical and physical data models to support new and existing projects.
  • Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop.

Environment: Data warehousing, SSRS, ETL, Teradata, Mainframe, Quality Center, Visio, Data stage 8.5, Rally, Oracle, Informatica BDM, SSIS, SQL Server, Scala, Horton Works, Kerberos, Hive, Ambari\Tez Views, HBase, Linux, SQL Developer 4.1.5, SQL Server, Tableau, QlikView


Data Analyst


  • Responsible for walk through the documents with customer during the end user review, maintain versions and take the user sign-off on the documents and changes. Organize Gate reviews at end of each phase of SDLC
  • Involved in implementation of software solutions ranging from Mainframe based legacy applications
  • Interacted with business owners and users to identify business system needs, evaluate solutions for business problems and proposed alternate methods.
  • Interacted with developer team, operations support team and end customer to understand the current system and new system functionalities.
  • Developed Business Process Flows, Functional Requirements Specifications, and System Design documents.
  • Worked with the development team to convert the functional specification into program specification and reviewed the low-level design and specifications.
  • Worked with Project Manager to Create and maintain work break down and schedule for the onsite implementation team.
  • Assisted in the development of quality assurance (QA) documentation including the testing strategy, user acceptance testing (UAT) scripts and business scenarios.
  • Determined the efforts for the development and implementation of the functional gaps.
  • Evaluated user request for new or modified program to determine feasibility, cost and time required, compatibility with current system and computer abilities.
  • Developed the business requirements, high-level design, Business Overview Documents, Requirement Traceability matrix using the UML and OOAD techniques. Prepare the details of the dataflow for the new scope based on the system architecture and create the DFD, Use Case, Swim Lane and Context Diagrams
  • Identify any database changes by interacting with the DBAs and creating the logical model diagrams and get necessary sign-offs.
  • Recommend opportunities for reuse of data models in new environments. Perform reverse engineering of physical data models from databases and SQL scripts.
  • Evaluate data models and physical databases for variances and discrepancies. Develop data models according to company standards.
  • Employed superior analytical abilities, data mining skills, and diverse supply chain experience to create inventory targets and develop strategies for optimizing PCB stock levels and reducing worldwide scrap.
  • Spearheaded and launched inventory rotation process to align demands to supplies on transition products and reduced worldwide material requirements by 10%.
  • Produced databases, tools, queries, and reports for analyzing, summarizing, and root causing board failure data. Versed in finding patterns and trends in complex, multivariable data sets.
  • Created Inventory Model and BOM tool, identified supply chain excess, and sold back $400,000 of excess Testers thru revenue stream with no disruptions to customer service levels.
  • Worked closely with IT, Document Control, and purchasing and created last time buy tool/database and automated the last time buy process within Oracle, leading to the elimination of two full time roles.
  • Worked across Logistics, Engineering, Finance, Customer Service, Sales, and Manufacturing organizations to improve product reliability, reduce board repair time, automate reporting, and minimize waste.
  • Presented tools and recommendations to directors and executive staff.

Environment: COBOL, DB2, SQL SERVER, IMS DB DC, DB2, VSAM, JCL, ENDEVOR, FILE-AID, DB2 Stored proc, CA7, Xpeditor, DB2 Visual Studio, DB2 Connect, IBM Fault Analyzer, Data warehousing, SSRS, ETL, Teradata, Mainframe, Quality Center, Visio, Data stage 8.5, Rally, Oracle, Informatica BDM, SSIS, SQL Server


Software Engineer


  • Translated logical design into physical program specifications.
  • Developed and integrates portions of IT solutions by using appropriate development methodologies and standards, following standard architectures/patterns, and working within the physical application framework with guidance
  • Developed and executed unit test plans and scripts to validate that assigned modules meet technical specifications
  • Participated in technical inspections at appropriate stages of development
  • Creates production support documentation
  • Worked with Operations Analysts, System Designer/Analysts and other developers to prepare and stage production builds
  • Facilitates and contributes to project chartering sessions
  • Coordinates and provides design leadership and guidance (mentoring) to systems designers at all levels
  • Consults with technical peer experts to determine and decide on technical environment and database architecture for strategic applications
  • Collaboration with external vendors and consultants to strategize emerging design methodologies and exchange of experience
  • Establishes direction for design related work for major programs
  • Establishes, reviews, approves and publishes guidelines and standards for methodology deployment across global IT • Responsible for resolving design issues on major programs
  • Responsible for system designs for major and enterprise wide programs

Environment: COBOL, DB2, SQL SERVER, IMS DB DC, DB2, VSAM, JCL, ENDEVOR, FILE-AID, DB2 Stored proc, CA7, Xpeditor, DB2 Visual Studio, DB2 Connect, IBM Fault Analyzer, Data warehousing, SSRS, ETL, Teradata, Mainframe, Quality Center, Visio, Data stage 8.5, Rally, Oracle, Informatica BDM, SSIS, SQL Server


Project Engineer


  • Leads translation of logical design into physical program design/specifications
  • Develops and integrates IT solutions by using appropriate development methodologies and standards, adopting/applying standard architectures/patterns, taking into account critical performance characteristics, security measures and customer requirements
  • Ensures the development and execution of unit test and integration plans and scripts to validate that application modules meet technical specifications
  • Ensures technical inspections are performed at appropriate stages of development
  • Leads creation of production support documentation
  • Worked with others (e.g. Operations, System Designer/Analysts, Developers) to prepare and stage production builds, and verify production releases meet technical and business expectations
  • Leads data migration/conversion activities
  • Provides technical leadership and mentoring to developers, and leads the technical application development/integration of IT solutions

Environment: VS COBOL II, JCL, VSAM, CICS & DB2, Endevor, Xpediter & File-Aid

Hire Now