We provide IT Staff Augmentation Services!

Data Science / Big Data Developer (consultant) Resume

3.00/5 (Submit Your Rating)

Jersey City, NJ

SUMMARY:

Data Science / Big Data Developer with 12 + yr including 9 yr in Data Analysis, 2 yr of Data Analytics using R, Python, Machine Learning algorithms like Recommendation, 3 yr Big Data development on Spark, Hive, on Hortonworks, Cloudera, Linux VMs.

PROFESSIONAL EXPERIENCE:

Confidential, Jersey City, NJ

Data Science / Big Data Developer (Consultant)

Responsibilities:

  • For Client Analytics, Use R programming, Microsoft Revolution Open (MRO), SAS, SAS R API.
  • Analyze source data, Exploratory Data Analysis, Data Extraction, Data Munging, Data Transformation, Data Visualisation, Statistical Data Analysis, Use R Big Data packages like SparkR, PySpark to process large volume data, DPLYR for run SQL, Data Analysis, Caret for Clean Data ETL, RODBC, RJDBC for connections with JVM, Data Visualization using ggplot2.
  • Used packages like FF, Snow, Microsoft Revolution Open(MRO) for better performance, kknn, class for K Nearest Neighbor, pvclust, hclust, mclust for Clustering, Hierarchical Clustering, Reccomenderlab, Market Basket Analysis, Classification, Regression, histograms.
  • With team members, create Database, Tables using Hive, importing data into Hive tables from local, Confidential, running Scoop to import data from NoSQL databases, RDBMS to Confidential . Use R to connect to Oracle to import data from RDBMS, Qlikview, process data using SparkR.
  • Presentation on R connection to Back & Front end, brainstorm Data Lake, R Big Data, Analytics.
  • Build Windows Server & RedHat Unix Servers & install SparkR.

Confidential, Herndon, VA

Big Data Architect

Responsibilities:

  • Analyze multiple sources of Data sources from the customer using Python, R. Studio, dataframes. Write Python code for Apache Spark, scripts for Hive on Linux system.
  • R Big Data packages like SparkR, PySpark for large volume data, DPLYR for run SQL, Data Analysis, Caret for Clean Data ETL, Data Visualization using ggplot2
  • Use R, Python on top of Spark to build histograms, Pi charts, correlation, linear and multiple regression and work with Solution Architect for building Machine Learning algorithms for Classification, Regression for network security data. Import Data from Google Analytics.
  • Creating Database, Tables using Hive, importing data into Hive tables from local, Confidential, running Scoop to import data from RDBMS to Confidential .
  • Clean, merge large volume data using unix, R & Python & initiate integration processes for data storage and analysis in fully virtualized linux environments.
  • Build Servers for Big Data applications using on Ubuntu, provision of Servers using Vargrant.

Confidential, Herndon, VA

Lead Data Consultant

Responsibilities:

  • Analyze Confidential Data using SQL, Unix scripts. Based on Data Analysis discuss on Data migration plan with Business team, Architect, Developers in Informatica, Master Data Management, Reporting teams.
  • Analyze multiple Data sources from the customer using Python, R. Write Hive, Pig scripts, Spark code using Python.
  • Test Run ETL code from RDBMS to check clean, transformed data load on server and if ETL logic is as per requirements.
  • Run SQL & Unix scripts to perform Data validation from Source, Staging, Data Mart, Business Objects Reports, data Projected vs Actual from Source to Target is as per the Business rules for data mapping, lineage in Informatica, MDM.
  • Work with Architect for proof of Concept for Big Data migration from RDBMS to Confidential using Hive scripts, Creating Database, Tables using Hive, importing data into Hive tables.
  • Cloudera, Hortonworks 2.2 HDP for running Spark, Hive, Pig, H2O. Data Cleaning, Data Ingestion and Data Analytics
  • Data Analysis using Python, Data migration from relational databases to Big Data using Sqoop, Hive, Pig, Spark.

Confidential, Washington, DC

Data Consultant

Responsibilities:

  • Work to with development team to migrate data from PL SQL to Unix based Oracle system.
  • Run SQL scripts in TOAD to perform check on Source Data, timing of running of SQL scripts, timing of Unix scripts. Check root cause of variance in Cognos Reporting system, Unix system, Source database and document the causes of variance in data and errors.
  • Based on source data, data timing from JPMC others, source to target mapping, Data Modeling based on feedback from Data Architect, Business, Technical team.
  • Analyze and discuss Reporting data document with ETL (Informatica team), Reporting (Cognos team), DBA & Architect for production ready Source to Target mapping, Master Data document. SQL scripts to check data per business rules.
  • Proof of Concept for Big Data implementation for Pipeline re - Engineering Reporting Data. Run Hive scripts to test data between Source and Confidential, transfer of data into and out of RDBMS to Hadoop using Scoop scripts, import export of data logs between RDBMS warehouse & Confidential .

Confidential, Pittsburgh, PA3

Data Consultant

Responsibilities:

  • For Confidential banking applications for commercial and retail banking - Procure data from source database (Mainframe developer), run Unix scripts to migrate data from Mainframe system to Unix based Oracle system.
  • Run SQL scripts in TOAD to perform check on Source Data and upload dummy data in the Data Staging area for Testing Data and working with ETL (Informatica) developers.
  • Work with business to finalize business rules. Work with Cognos and Business Objects developers to build dummy reports for the Confidential Bank, Confidential business. Based on dummy reports, develop SQL scripts.
  • Tests Reports using Cognos, Business Objects, data quality test, report frequency, Confidential vs Standard reports run timely and partial vs full data load as per Business requirement.
  • Perform Regression test, run SQL, Unix scripts from source to target data, and post approval of Testing team, conduct walk through, get feedback / Document Sign Off in Clarity tool with consent of all stakeholders.
  • Proof of Concept document for Big Data implementation of Confidential Data.
  • Develop scripts for Hadoop Hive, Cloudera VM, import export of data between RDBMS Data warehouse & Confidential .

Confidential, Boston, MA

Data Analyst

Responsibilities:

  • Run SQL scripts to Analyze Data during stages including Trade submission, release, execution, matching, trade posting.
  • Run Unix scripts, Unix logs in Virtual environment to check Trade Data, Price margin errors in Unix logs vs Database, check XML, for Compliance rule, Trade Surveillance logs, Portfolio rules in logs of XML in various Virtual Machines.
  • Analyze trade issues with FIX log for Trade execution, XML logs for Trade n Confidential ’s Central Trade Manager (CTM).
  • Liaison with Confidential teams in Business, PMs.

Confidential, MD

Functional Business / Data Analyst

Responsibilities:

  • Discuss & Document Blue print for System Integration - from fragmented asset data, compliance tools for watch list filtering, fraud case management, plan for Trade Surveillance, Transaction monitoring, positions, building risk profiles.
  • Procure Anti Model Laundering data Procure from source database (Mainframe, Oracle, Access, Sybase), run Unix scripts to migrate data from Unix to Oracle based system. Run SQL, TOAD to perform check on Source vs Test Data.
  • Analyze data in multiple systems from Bridger, Charles River, AWD, Pershing for migration.
  • Document change management discussion in SCRUM meetings, weekly SPRINT, for new screens, prepare wireframes.
  • Run SQL scripts to update the Master Data Dictionary, Analyze Source to Target data.

Confidential, Baltimore, MD

Senior Data System Analyst

Responsibilities:

  • Work with Data Architect to document database design, develop data mapping, flow diagram, System migration guideline for Data Reporting.
  • Run Oracle & Unix jobs on Linux VM to import data from Server to Client Machine.
  • Testing & analysis using Oracle, TOAD, reporting error.

Confidential, New York

Business Systems Analyst

Responsibilities:

  • Worked on Global Environment for Accounting and Reporting system for GL & bookkeeping.
  • Discussion, feedback from multiple stakeholders for requirement gathering with Business, Legal, Data, As-Is, Confidential for Actimize implementation, in corporate the feedback.
  • Functional requirement for Suspicious Activity report, Transaction monitoring, Watch List filtering using Actimize. Performed data reporting using Business Objects (from sources to data warehouse), WebI (Web Intelligence Actimize). checking reconciliation of Client positions.
  • System Test on the customized Actimize based reports, analyze errors in report, data feeds.
  • Web based applications: Compliance, Trade Processing & Settlement, Client positions, BASEL II.

We'd love your feedback!