Python/django Developer Resume
Bloomington, IL
SUMMARY
- Result - driven IT Professional with referable & strong experience in using various packages in Python like ElementTree XML API, lxml, subprocess, Pandas DataFrame, Matplotlib, Bokeh, Seaborn, NumPy, NLTK, librosa, regex, Beautiful Soup, SciPy, SQLAlchemy, PyTest, urllib2, MySQL Connector, PyMySQL for database connectivity).
- Expertise in BigData using Hadoop framework and Analysis, Design, Development, Documentation, Deployment and Integration using SQL and Big Data technologies.
- Experience in Text Analytics, developing different Data Cleansing, Data Wrangling, Data Mining solutions to various business problems and generating data visualizations using Python and R.
- Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing Data Mining and reporting solutions that scales across massive volume of structured and unstructured Data.
- Experience with Unix/Linux systems with scripting experience and building data pipelines.
- Experience in developing web applications and coding with an analytical programming using Python, Django, Flask, PHP, MySQL, HTML, CSS, Bootstrap, JavaScript and jQuery.
- Experience in importing and exporting data using Sqoop from Relational Database Systems to HDFS and vice-versa.
- Hands-on experience in writing and reviewing requirements, architecture documents, test plans, design documents and quality analysis.
- Highly skilled in System Analysis, E-R/Dimensional Data Modeling, Database Design and implementing RDBMS specific features.
- Versatile team player with excellent analytical, inter-personnel and presentation skill with ability to quickly adapt to new technologies & project environments.
TECHNICAL SKILLS
Programming Languages: Python, Shell Script, PL/SQL, HiveQL, Spark SQL
Web Technologies: HTML, CSS, JavaScript, jQuery, PHP
Version Control: GIT, GitHub, GitLab
Databases: MySQL, PostgreSQL, RDBMS
IDE’s/Development Tools: PyCharm, Jupyter, Anaconda, Sublime Text, Notepad++, Visual Studio Code, Atom.
Hadoop Ecosystem: HDFS, Hive, Impala, HBase, Spark, Sqoop
Cloud Platform: AWS, GCP
AWS Services: EMR, S3, EC2, RDS
Python Framework: Django, Flask
BI Tools: Tableau
Modeling Languages: Unified Modeling Language (UML)
Methodologies: Agile, Scrum and Kanban
Operating Systems: Windows, Linux, UNIX
PROFESSIONAL EXPERIENCE
Confidential - Bloomington, IL
Python/Django Developer, BigData Engineer
Responsibilities:
- Developed pipelines using Python, Spark to ingest customer behavioral data and purchase histories into HDFS for analysis .
- Created External Hive tables to stage data and then move the data from staging to main tables.
- Developed interactive dashboard using Python Bokeh package with the help of Flask framework which provides easy access to data tables, reports, process maps, and insights.
- Developed ETL applications on large volumes of data using different tools: Python, MapReduce and PySpark.
- Used librosa, scipy.io to analyze audio signals and to generate two individual mono wav files from a two-channel stereo file.
- Created indexes on the tables for faster retrieval of the data to enhance database performance.
- Worked with PCI team to build the pipeline for ingesting agent customer call data to perform the Sentiment Analysis.
- Developed BI dashboards to showcase Key Performance Indicators (KPIs) to the mangers to help them make the current process more efficient.
- Implement software enhancements to port legacy software systems to Hadoop ecosystems.
- Developed ETL batch automation using shell scripting and scheduled the Cronjob.
- Performed data collection, data cleansing and data visualization using Python and extracted key statistical findings to develop business strategies.
- Involved in various pre-processing phases of text-data like Tokenization, Stemming, Redaction and converting the raw text data to structured data.
- Used adobe analytics tool to generate key insights reports which includes fallout funnel, page flow and tables.
- Developed the UNIX shell scripting to run SQL files based on initial load instruction.
- Suggested different A/B testing opportunities to improve online user experience and drive key changes to developer backlogs.
Environment: Python, R, Hive, Flask, XML, JSON, Parquet, MySQL, Apache, Splunk, Adobe Analytics, Optimizely, Tealeaf, Spark, HDFS, Linux, GitLab, EC2, S3, Lambda, Dynamo DB
Confidential, Memphis, TN
Python/Django Developer, BigData Engineer
Responsibilities:
- Managed datasets using Pandas data frames and MySQL, queried MySQL relational database (RDBMS) queries from python using MySQL Connector, PyMySQL packages to retrieve information.
- Involved in designing database Model, API’s, Views using python to build an interactive web-based solution.
- Worked on ETL objects that were already in production and successfully modified to improve the performance.
- Implemented AWS instances and worked with EBS and S3 storage.
- Used AWS Beanstalk for deploying and scaling web applications and services developed with Python.
- Developed and reviewed custom SQL queries with use of joins clauses (inner, left, right, full) in Tableau desktop to validate and reconcile data.
- Designed and Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
- Worked on different data formats such as JSON, XML and performed machine learning algorithms in Python.
- Re-write some Hive queries to Spark SQL to reduce the overall batch time.
Environment: Python, Django, ERS, S3, MySQL, Apache, Hadoop, Hive, Tableau, Spark, HBase, HDFS, EC2, S3, SQL*Plus, Linux, UNIX
Confidential - Teaneck, NJ
Application Developer
Responsibilities:
- Used ETL for handling of the raw data, organizing and structuring the data.
- Involved in creating Hive tables, loading with data and writing hive queries that will run internally in MapReduce way.
- Unit Test Python library was used for testing many programs on python and other codes.
- Performed different testing methodologies like unit testing, integration testing, web application testing was performed.
- Ensured high quality data collection and maintaining the integrity of the data.
- Used Pandas API to put the data as time series and tabular format for east timestamp data manipulation and retrieval.
- Analyzed the data by performing Hive queries to study customer behavior.
- Worked on Big data analytics like Apache Hive, Apache HBase.
- Collaborated with a team of instructors and programmers to develop the curriculum and guidelines for workshops to teach the logic of programming.
Environment: Python, Django, Hive, Spark, Apache, HBase, Tableau, MySQL, Hadoop
Confidential
Data Engineer
Responsibilities:
- Managed datasets using data frames and MySQL, queried MySQL database queries from python using Python -MySQL connector package to retrieve information.
- Designed and developed components using Python with Django framework.
- Implemented code in python to retrieve and manipulate data.
- Carried out various mathematical operations for calculation purpose using Python library NumPy.
- Designed back end and part of the front end of the application using python on Django web framework.
- Used Test driven approach for developing the application and Implemented the unit tests using Python Unit test framework.
- Responsible for gathering requirements, system analysis, design, development, testing and deployment.
- Designed and developed various analytical reports from multiple data sources by blending data on a single worksheet in Tableau Desktop.
- Wrote Python normalizations scripts to find duplicate data in different environments.
- Engaged in understanding Java Code and re-build the code in Python .
Environment: Python, Django, MySQL, LAMP, WAMP, Tableau, Windows, Linux