We provide IT Staff Augmentation Services!

Data Scientist Resume

5.00/5 (Submit Your Rating)

Denver, CO

PROFESSIONAL SUMMARY:

  • Experience in develop and enhance advanced data mining algorithms and machine learning, which will help drive data analysis for a wide range of information and product features.Experience in Data management, Data Engineering, or Data Science.
  • Experience with R or similar statistical languages 1 - 2 years of experience using SQL to query databases (Oracle, SQL Server) Preferred Qualifications: 1-2 years of experience with web development: HTML, CSS and JavaScript Background in development and production of electronic and RF components, or a degree in a related field Experience with Business Objects, Tableau, or other charting tools.
  • Experience with shop floor systems, experience with the Software Development lifecycle, and both Agile and Waterfall projects, experience with electrical and mechanical data management systems Data analytics and visualization, statistics Database query, experience in Excel Macros, experience with object-oriented programming, including Java, Python, Scala, or JavaScript, experience with source control and dependency management software, including Git or Maven, experience with Linux administration, experience with Big Data, Cloud computing technologies, NoSQL systems, or Lambda architectures, experience with building complex data extraction, transformation, and loading, including ETL pipelines into structured databases, data warehouses, or data processing systems.
  • Experience with developing solutions with and integrating and extending free and open source software, experience with Semantic Web, including RDF or OWL and SPARQL
  • Experience with Navy mission systems, experience with designing and implementing machine learning, data mining, statistics, or graph algorithms-Experience with Amazon Web Services, including
  • AWS or Microsoft Azure Cloud services-Experience with the Hadoop ecosystem, including HDFS, Hive, HBase, Accumulo, or Spark, experience with Storm, Kafka, NiFi, or similar ETL frameworks
  • Experience with writing analytics using MapReduce, Pig, SQL, or SparkSQL-MS degree in CS, Software Engineering, or Informatics.

TECHNICAL SKILLS:

  • Data Science, Data Scientist
  • Engineer, Big Data, R, SQL
  • Python, Insight, Analytics, Data
  • Statistics, Modeling, Machine Learning
  • Algorithms, Bayesian, Deep Learning Relational databases
  • (MS SQL Server, RedShift, Postgres), R, Numpy, Scipy, Scikit-learn
  • Pandas, Spring Technology stack, Cluster Computing frameworks
  • (Spark, Hadoop, EMR),Dynamic languages (Groovy, Python)
  • Static languages (Java, Scala), Cloud based development
  • (AWS, Azure, Google, PCF), Log aggregators (Splunk, Data Dog)
  • Source control systems (Git, SVN), CD/CI tools
  • (GoCD, Jenkins, Artifactory, Terraform, Ansible), CD/CI techniques (TDD, BDD
  • Infrastructure as Code), ORM (JPA/Hibernate
  • Messaging systems (JMS, RabbitMQ, SQS), Enterprise Integration Patterns (Mule ESB, Camel, Tibco)
  • Web application containers (Tomcat, Resin, Jetty, IIS)
  • Web services (REST, SOAP), API specification tools (RAML, Swagger).

PROFESSIONAL EXPERIENCE:

Confidential, Denver, CO

Data Scientist

Responsibilities:

  • As an Architect design conceptual, logical and physical models using Erwin and build data marts using hybrid Inmon and Kimball DW methodologies.
  • Worked closely with business, data governance, SMEs and vendors to define data requirements.
  • Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
  • Designed the prototype of theDatamart and documented possible outcome from it for end-user.
  • Involved in business process modeling using UML
  • Developed and maintaineddatadictionary to create metadata reports for technical and business purpose.
  • Created SQL tables with referential integrity and developed queries using SQL, SQL*PLUS and PL/SQL
  • Experience in maintaining database architecture and metadata that support the EnterpriseData-warehouse.
  • Design, coding, unit testing of ETL package source marts and subject marts using Informatica ETL processes for Oracle database.
  • Developed various Qlik-View Data Models by extracting and using the data from various sources files, DB2, Excel, Flat Files and Big data.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
  • Interaction with Business Analyst, SMEs and other Data Architects to understand Business needs and functionality for various project solutions
  • Researched, evaluated, architected, and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for the clients
  • Identifying and executing process improvements, hands-on in various technologies such as Oracle, Informatica, and Business Objects.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow flake Schemas.

Confidential, Stamford, CT

Data Scientist

Responsibilities:

  • Worked as aDataScientist to generateDataModels using Erwin and developed relational database system.
  • Analyzed the business requirements of the project by studying the Business Requirement Specification document.
  • Designed mapping to process the incremental changes that exists in the source table. Whenever source data elements were missing in source tables, these were modified/added in consistency with third normal form based OLTP source database.
  • Designed tables and implemented the naming conventions for Logical and Physical Data Models in Erwin 7.0.
  • Provide expertise and recommendations for physical database design, architecture, testing, performance tuning and implementation.
  • Designed logical and physical data models for multiple OLTP and Analytic applications.
  • Extensively used the Erwin design tool & Erwin model manager to create and maintain the Data Mart.
  • Designed the physical model for implementing the model into oracle 9i physicaldatabase.
  • Involved with Data Analysis Primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats
  • Performance tuning of the database, which includes indexes, and optimizing SQL statements, monitoring the server.
  • Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
  • Collaborated thedatamapping document from source to target and thedataquality assessments for the sourcedata.
  • Used Expert level understanding of different databases in combinations for Data extraction and loading, joining data extracted from different databases and loading to a specific database.
  • Co-ordinate with various business users, stakeholders and SME to get Functional expertise, design and business test scenarios review, UAT participation and validation of financial data.
  • Worked very close withDataArchitectsand DBA team to implementdatamodel changes in database in all environments.
  • Created PL/SQL packages and Database Triggers and developed user procedures and prepared user manuals for the new programs.
  • Performed performance improvement of the existingDatawarehouse applications to increase efficiency of the existing system.
  • Designed and developed Use Case, Activity Diagrams, Sequence Diagrams, OOD (Object oriented Design) using UML and Visio.

Confidential

Data Architect/Data Modeler

Responsibilities:

  • Deployed GUI pages by using JSP, JSTL, HTML, DHTML, XHTML, CSS, JavaScript, and AJAX.
  • Configured the project on Web-Sphere 6.1 application servers
  • Implemented the online application by using Core Java, Jdbc, JSP, Servlets and EJB 1.1, Web-Services, SOAP, WSDL
  • Communicated with other Health Care info by using Web-Services with the help of SOAP, WSDL JAX-RPC.
  • Used Singleton, Factory Design Pattern (FDP), DAO Design Patterns based on the application requirements
  • Used SAX and DOM parsers to parse the raw XML documents.
  • Used RAD as Development IDE for web applications.
  • Preparing and executing Unit test cases
  • Used Log4J logging framework to write Log messages with various levels.
  • Involved in fixing bugs and minor enhancements for the front-end modules.
  • Implemented Microsoft Visio and Rational Rose for designing the Use Case Diagrams, Class model, Sequence diagrams, and Activity diagrams for SDLC process of the application
  • Doing functional and technical reviews
  • Maintenance in the testing team for System testing/Integration/UAT
  • Guaranteeing quality in the deliverables.
  • Conducted Design reviews and Technical reviews with other project stakeholders.
  • Was a part of the complete life cycle of the project from the requirements to the production support
  • Created test plan documents for all back-end database modules
  • Implemented the project in Linux environment.

Confidential

Data Analyst/Data Modeler

Responsibilities:

  • Worked with project team representatives to ensure that logical and physical ER/Studio data models were developed in line with corporate standards and guidelines.
  • Involved in defining the source to target data mappings, business rules, data definitions.
  • Worked with BTEQ to submitSQLstatements, import and export data, and generate reports in Teradata.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Responsible for defining the functional requirement documents for each source to target interface.
  • Document, clarify, and communicate requests for change requests with the requestor and coordinate with the development and testing team.
  • Work with users to identify the most appropriate source of record and profile the data required for sales and service.
  • Document the complete process flow to describe program development, logic, testing, and implementation, application integration, coding.
  • Involved in defining the business/transformation rules applied for sales and service data.
  • Define the list codes and code conversions between the source systems and the data mart.
  • Worked with internal architects and, assisting in the development of current and target state data architectures.
  • Coordinate with the business users in providing appropriate, effective and efficient way to design the new reporting needs based on the user with the existing functionality.
  • Remain knowledgeable in all areas of business operations in order to identify systems needs and requirements.
  • Responsible for defining the key identifiers for each mapping/interface.
  • Implementation of Metadata Repository, Maintaining Data Quality, Data Cleanup Procedures, Transformations, Data Standards, Data Governance Program, Scripts, Stored Procedures, Triggers and execution of test plans
  • Performed data quality in Talend Open Studio.
  • Coordinated meetings with vendors to define requirements and system interaction agreement documentation between client and vendor system.
  • Enterprise Metadata Library with any changes or updates.
  • Document data quality and traceability documents for each source interface.
  • Establish standards of procedures.
  • Generate weekly and monthly asset inventory reports.

We'd love your feedback!