Data Scientist Resume
NY
SUMMARY
- Over 7 years of experience in the IT industry, including machine learning, data mining with large amounts of structured and unstructured data, data collection, data validation, predictive modeling, and data visualization.
- Extensive experience in text analysis, development of various statistical machine learning, data mining solutions for various business issues and data visualization using Python and Tableau.
- Experience with Machine learning techniques and algorithm (such as K - NN, Naive Bayes, etc.)
- Tagging of experience in foundational machine learning models and concepts of regression, random forest, deep learning.
- Familiar with Deep learning projects for image identification CNN, RNN for stock price prediction auto-encoders for Recommender System (PyTorch), Image captioning (CNN-RNN auto-encoder architecture).
- Exposure to AI and Deep learning platforms/methodologies like Tensorflow, RNN, LSTM
- Experience in training Artificial Intelligence ChatBots.
- Developed agreed analytical solution by applying suitable statistical & machine learning techniques (e.g., A/B testing, prototype solutions, mathematical models, algorithms, machine learning, deep learning, artificial intelligence) to test, verify, refine hypotheses.
- Extracted data to create value-added datasets using Python and SQL to analyze behavior to target specific customers to gain hidden insights into data to effectively implement project objectives.
- Worked extensively on statistical analysis tools and adept at writing code in Advanced Excel , Python .
- Experienced with AWS SageMaker .
- Experienced with working on Linux .
- Familiar with Unix environment.
- Experienced with Git and Git-Hub
- Worked collaboratively on Amazon Web Services ( AWS ) environment.
- Familiar with the EC2 Instances, Route 53 , Load balancers , S3 buckets
- Implemented numerical Computation and deep learning models with the help of data flow graphs using Tensor Flow Machine Learning.
- Experienced with the full software lifecycle in SDLC, Agile and Scrum methodologies.
- Hands on experience on implementing LDA, Naive Bayes and skilled in Random Forests, Decision Trees, Linear and Logistic Regression, SVM, Clustering, neural networks, Principle Component Analysis and excellent knowledge of Recommender Systems.
- Experienced with machine learning algorithms such as logistic regression, random forest, KNN, SVM, neural network, linear regression, Lasso regression and K-means.
- Used Slack for real-time business communications.
- Updated and followed up on project progress with Trello.
TECHNICAL SKILLS
Programming Languages & Machine Learning Algorithms: Python, Polynomial Regression, Random Forest, Logistic Regression, Classification, Clustering, Association, Simple/Multiple linear, Kernel SVM, K-Nearest Neighbors (K-NN)
Databases: SQL, SQL Databases SQL-Server, MySQL
Reporting Tools: MS Office (Word/Excel/Power Point/ Visio), Tableau, Business Intelligence
Version Control Tools: Git, GitHub.
BI Tools: Tableau, Amazon Redshift, Amazon Web Services (AWS): SageMaker
Operating System: Windows, Linux, Unix, Ubuntu
Visualization Tools: Tableau, Kibana (less than year)
PROFESSIONAL EXPERIENCE
Confidential, NY
Data Scientist
Responsibilities:
- As a backend developer, I am working on implementing the ChatBot platform for Confidential to provide users with an easier, more reliable, and more personalized experience.
- Working on a POC (proof of concept) for delivering a foreign-exchange trade based ChatBot.
- Researched different types of NLP libraries that help the bot to perform well.
- ChatBot is designed to interact with people as a customer service representative, who book sales, answer questions such as spot and subsequent exchange rates, and use an effective sales system.
- Main technologies utilized include Python and NLP libraries.
- Responsible for shaping the personality and conservational flow of chat to ensure customers receive a highly personalized AI experience.
- Best practices for ChatBot were determined to have the most optimal result for the user’s request.
- Using technical and analytical expertise to explore and test data from multiple data sources to find previously hidden models and insights that will provide a competitive advantage or solve an urgent mission problem.
- Working on Linux environment.
- Stored data and running applications on Hadoop.
- Checked building conversational interfaces of ChatBot for application of voice and text recognition on AWS Lex Console service.
- Implemented deep learning ASR functionalities for conversion of speech to text.
- Integrated Amazon CloudWatch with Amazon Lex for better security, monitoring, business logic, storage and mobile app development.
- Worked on GCP to empower the application deployment on to the web.
- Working on Hybrid.Chat to connect SQL database with the ChatBot.
- Responsible for data warehousing, relational databases, ETL, data mining.
- Working on NLP (natural language processing), NLU (natural language understanding), and unstructured data analysis - text, image and graph/network analysis using GraphFrames, Igraph GraphTools, NetworkX
- Build, train and deploy ML models using Amazon SageMaker.
- Building intentional stories using Rasa-X for more personalized interaction with the end user.
- Work closely with technical, consulting and knowledge teams to develop solutions that exceed customer expectations.
- Extensive experience solving analytical problems using quantitative approaches.
- Experience with machine learning algorithms for building ML models, their accuracy, cleanliness, and reliability.
- Continually test, analyze and iterate to train the AI and expand the range of responses.
Environment: Python, NLP libraries- Rasa, Spacy, Hadoop, Deep Learning (TensorFlow), PyTorch, AWS: SageMaker, Jenkins, Kibana, Linux, SVM, JSON, GCP.
Confidential
Data Analyst
Responsibilities:
- Provided the architectural leadership in shaping strategic, business technology projects, with an emphasis on application architecture.
- Utilized domain knowledge and application portfolio knowledge to play a key role in defining the future state of large, business technology programs.
- Participated in all phases of data mining, data collection, data cleaning, developing models, validation, and visualization and performed Gap analysis.
- Accumulated strong knowledge of machine learning algorithms, metric-driven data set analysis and NLP techniques from work and self-study.
- Deployed Machine Learning models for item-item similarity on Amazon SageMaker (AWS)
- Conducted studies, rapid plots and using advance data mining and statistical modeling techniques to build a solution that optimize the quality and performance of data.
- Demonstrated experience in design and implementation of Statistical models, Predictive models, enterprise data model, metadata solution and data life cycle management in both RDBMS, Big Data environments.
- Analyzed large data sets apply machine learning techniques and develop predictive models, statistical models and developing and enhancing statistical models by leveraging best-in-class modeling techniques.
- Development of Tableau visualizations and stories using tableau desktop. Documenting Business requirements and plans for creating dashboards. Implementation of Python, Hadoop and Tableau to extract and import data.
- Developed MapReduce/Spark Python modules for machine learning & predictive analytics in Hadoop on AWS.
- Worked on customer segmentation using an unsupervised learning technique - clustering.
- Utilized Spark, Scala, Hadoop, HBase, Kafka, Spark Streaming, MLlib, Python, a broad variety of machine learning methods including classifications, regressions, dimensionally reduction etc.
- Designed and implemented system architecture for Amazon EC2 based cloud-hosted solution for the client.
- Tested Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables.
Environment: Python, SQL, Oracle 12c, SQL Server, PL/SQL, MLlib, regression, Cluster analysis, Spark, logistic regression, PySpark, random forest, Azure, NLP, JSON, Tableau, XML, MapReduce, AWS SageMaker, Tableau
Confidential
Python Developer
Responsibilities:
- Worked on the project from gathering requirements to developing the entire application.
- Worked Anaconda Python Environment for predictive analysis.
- Experience with environment.yml for reproducible codes within the specific packages.
- Created, activated and programmed in Anaconda environment.
- Wrote programs for performance calculations using NumPy and SQL Alchemy.
- Wrote python routines to log into the websites and fetch data for selected options.
- Used python modules of urllib, urllib2, Requests for web crawling.
- Involved in development of Web Services using SOAP for sending and getting data from the external interface in the XML format. Used with other packages such as Beautiful Soup for data parsing
- Worked on development of SQL and stored procedures on MYSQL.
- Analyzed the code completely and have reduced the code redundancy to the optimal level.
- Design and build a text classification application using different text classification models.
- Used Jira for defect tracking and project management.
- Worked on writing and as well as read data from CSV and excel file formats.
- Involved in Sprint planning sessions and participated in the daily Agile SCRUM meetings.
- Conducted daily scrum as part of the SCRUM Master role.
- Developed the project in Linux environment.
- Worked on resulting reports of the application.
- Performed QA testing on the application.
- Held meetings with client and worked for the entire project with limited help from the client.
Environment: Python, Anaconda, Spyder (IDE), Windows 7, Beautiful Soup, Tableau, python libraries such as NumPy, SQLAlchemy, MySQL
Confidential
Software Developer
Responsibilities:
- Responsible for the design of web applications right from the conceptualization stage to its implementation and maintenance.
- Designed and developed the user interface of the project with HTML, CSS and PHP
- Worked on fetching documents from the local web server and provided the documents into the multimedia web pages.
- Created structured documents into multimedia on the webpage using HTML scripting.
- Collaborated with the team of web designers to match and deploy the visual design intention.
- Conducted sprint meetings every other week to discuss the progress of the project.
- Designed and developed the data management systems using MySQL databases.
- Integrated the database with the standard CSS and HTML practices.
- Created arrays of the product.
- Worked to create dashboard for the project which contains many charts and drag and drop options using JQuery UI library
- Gathered needful and specified technical requirements for the team from the client-side.
- Created layouts of the user interface for the website.
- Developed components that involve closures, event handling and Object-Oriented Java script.
- Developed the custom grids upon jQuery framework to deliver the business data involved in daily defect triage calls, fixed various serious defects based on their impact on the application.
- Written efficient and testable codes for the website on HTML
- Experienced in using various IDE’s line MS Visual Studio, Notepad++ and sublime text.
- Worked extensively with the back-end team to integrate the HTML code into JSPs.
- Used jQuery plug-in to implement features as a light box, sliders and other intuitive features.
- Fabricated style sheets to manage the look-and-feel of the site using CSS, bootstrap and XML.
- Participated in meetings with the end-clients to develop layout, color scheme for the web page and implement them.
- Client-side validations and manipulations were done using JavaScript and JQuery
- Experienced in writing indexes, views, constraints, stored procedures, triggers, cursors and user defined functions or subroutines in MySQL.
- Responsible for Debugging and troubleshooting the application.
- Utilized Subversion control tool to coordinate teamwork.
Environment: UNIX, HTML, PHP, CSS, JavaScript, Object-Oriented JavaScript, Notepad++, MS Visual Studio, JSP, UI library, XML and MySQL.