We provide IT Staff Augmentation Services!

Senior Data Engineer Resume

3.00/5 (Submit Your Rating)

Richmond, VirginiA

SUMMARY

  • Senior Data Engineer wif 7+ years of experience as Data Engineer, Data Analyst and Python Developer. Proficient in designing, documenting, development, and implementation of data models for enterprise - level applications. Background in Data Lake, Data Warehousing, Data Mart, Data modelling, ETL Data pipeline & Data Visualization.
  • Proficient in Big data storage, processing, analysis, and reporting on all major Cloud vendors- AWS, Azure and GCP.
  • Experience in Big Data ecosystems using Hadoop, MapReduce, YARN, HDFS, HBase, HIVE, PIG, Sqoop, Storm, Spark, Scala, Airflow, Flume, Kafka, Oozie, Impala, HBase and Zookeeper.
  • Experience developing Spark applications using Spark Core, Streaming, SQL, DataFrames, Datasets & Spark-ML. Developed Spark Streaming jobs by developing RDD’s using Scala, PySpark and Spark-Shell.
  • In-depth understanding/knowledge of Hadoop Architecture and components including HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce. Worked in Design, Implementation, Deployment and Maintenance of end-to-end Hadoop based analytical solutions.
  • Experienced in HIVE Queries to process large sets of structured, semi-structured & unstructured data. Experience in loading data into HDFS using Sqoop as well as saving data in Hive tables.
  • Involved in end-to-end implementation of Enterprise Data Warehouse, Data Lakes & Data Mart wif Batch and Real-time processing using Spark streaming, Kafka, Flume and Sqoop.
  • Experience using Pig scripts for data transformations, event joins, filters, and pre-aggregations.
  • Experience in setting up workflow using Apache Airflow and Oozie to manage & schedule Hadoop jobs.
  • Strong database skills in IBM- DB2, Oracle and proficient in database development, including Constraints, Indices, Views, Stored Procedures, Triggers and Cursors.
  • Experience in ETL using Informatica DEI/BDM, Power Center, DataStage & IICS tools.
  • Experience in configuration, deployment & automation of Data warehousing pipelines in all major Cloud environments (AWS, Azure & GCP).
  • Experience in deployment and testing infrastructure wifin AWS, using tools like Jenkins, Puppet and Docker.
  • Involved in setting up Jenkins Master and multiple slaves as a CI tool as part of Continuous Deployment.
  • Worked wif AWS EC2 cloud instance. Used EMR, Redshift, and Glue for data processing.
  • Worked wif AWS storage, OLTP, NoSQL & data warehouse- S3, RDS, DynamoDB & RedShift.
  • Proficient in AWS CodePipeline and worked wif CodeCommit, CodeBuild & CodeDeploy.
  • Worked on creating IAM policies for delegated administration wifin AWS and Configured IAM Users / Roles / Policies to grant fine - grained access to AWS resources.
  • Hands on experience wif Microsoft Azure Cloud services, Storage Accounts and Virtual Networks. Worked on Security in Web Applications using Azure and deployed Web Applications to Azure .
  • Experience in migration from On-Premises Instances & Azure Classic Instances to Azure. Designed and implemented Migration plan dat involved Rehosting & Replatforming.
  • Experience in GCP platform- compute engine, cloud load balancing, cloud storage, database (Cloud SQL, Bigtable, Cloud Datastore), stack driver monitoring and cloud deployment manager.
  • Hands on experience in Big Query, GCS bucket, G-Cloud function, cloud Dataflow and Data Fusion, Pub/Sub cloud shell, G-Suite, BQ command line utilities, Data Proc, Stack driver.
  • Experience in data preprocessing (data cleaning, data integration, data reduction, and data transformation) using Python libraries including NumPy, SciPy and Pandas for data analysis and numerical computations.
  • Experience working on various file formats including delimited text files, clickstream log files, Apache log files, Parquet files, Avro files, JSON files, XML files and others.
  • Experience in working wif UNIX/LINUX environments & writing Shell scripts on various UNIX distributions.
  • Experienced in working wif Spark eco system using SCALA and HIVE Queries on various data formats like Text file, XML files, Parquet & others.
  • Experience in generating on-demand and scheduled reports for business analysis or management decision using SSRS, Tableau, Qlikview, POWER BI as well as Python libraries like Seaborn, matplotlib, plotly, ggplot, etc.
  • Worked wif Tool Nagios for Resource Monitoring/Network Monitoring/Log Trace Monitoring.
  • Experience working wif DNS, HTTP, Tomcat, NFS, Proxy servers (Squid), NAT, Apache Web Server, DNS Server (BIND), FTP & SFTP Server wif health monitoring tools (Cloud Watch, Solar Winds, Logic Monitor).
  • Good understanding of data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
  • Experience in MS SQL Database Analysis, Design, Development & Support (2008R2, 2012, 2014, 2016).
  • Worked on SQL Server Log shipping, Database Mirroring, snapshot/transactional/Merge, SSIS & SSRS.
  • Experience in Oracle Database Administration, System Analysis, Design, Development, Maintenance and Support (8i/9i/10g/11g) . Performed Query Optimization, Performance Tuning, and Trouble Shooting.
  • Experience working on MySQL on Linux and Windows environments. Performed MySQL Replication setup and administration on Master-Slave and Master-Master.
  • Proficient wif all major PostgreSQL procedural languages (PL/PgSQL, PL/Perl, PL/PgPython, PL/Tcl).
  • Worked on PostgreSQL Streaming Replication and Pgpool for load balancing & monitoring tools for better performance like PgBadger, Kibana, Graphana, and Nagios.
  • Experience in NoSQL database. Worked wif Snowflake, HBase, Couchbase, Cassandra and MongoDB.
  • Worked wif Version Control Systems - CodeCommit, GIT/GitHub, Cloud Source Repository & BitBucket.
  • Worked on installing & configuring Jfrog. Integrated wif environments of Artifactory instances.
  • Experience in using Docker and setting up ELK wif Docker and Docker-Compose. Knowledge on various Docker components like Docker Engine, Hub, Machine, Compose and Docker Registry.
  • Experience on Docker & Kubernetes based container deployments to create environments for development teams as well as managing delivery for releases wif continuous load management.
  • Experience in Docker Installation wif Docker toolbox. Created Docker images, tagging and pushing images.
  • Worked in TERADATA Database design, implementation & maintenance in large scale Data Warehouse. Proficient in TERADATA SQL, Stored Procedures, Macros, Views, Indexes Primary, PPI & Join indexes.
  • Understanding of both traditional statistical modeling and Machine Learning techniques and algorithms like Linear & Logistic Regression, Naïve Bayes, kNN, SVM, clustering, ensembling(random forest, gradient boosting), deep learning (neural networks), etc.

TECHNICAL SKILLS

ETL Tools: AWS Glue, Azure Data Factory, GCP Data Fusion & DataFlow, Airflow, Spark, Sqoop, Flume, Apache Kafka, Spark Streaming, Apache NiFi, Microsoft SSIS, Informatica PowerCenter & IICS, IBM DataStage

NoSQL Databases: MongoDB, Cassandra, Amazon DynamoDB, HBase, GCP DataStore

Data Warehouse: AWS RedShift, Google Cloud Storage, SnowFlake, Teradata, Azure Synapse

SQL Databases: Oracle DB, Microsoft SQL Server, IBM DB2, PostgreSQL, Teradata, Azure SQL Database, Amazon RDS, GCP Cloud SQL, GCP Cloud Spanner

Hadoop Distribution: Cloudera, Hortonworks, MapR, AWS EMR, Azure HDInsight, GCP DataProc

Hadoop Tools: HDFS, Hbase, Hive, YARN, MarReduce, Pig, HIVE, Apache Storm, Sqoop, Oozie, Zookeeper, Spark, SOLR, Atlas

Programming & Scripting: Spark Scala, Python, MySQL, PostGreSQL, Shell Scripting, Pig Latin, HiveQL

Visualization: Tableau, Looker, QuickSight, Qlikview, PowerBI, Grafana, Python Libraries

AWS: EC2, S3, Redshift, RDS, EMR, Lambda, Glue, CloudWatch, Rekognition, Kinesis, CloudFront, Route53, DynamoDB, CodePipeline, EKS, Athena, QuickSight

Azure: DevOps, Synapse Analytics, Data Lake Analytics, Databricks, Blob Storage, Azure Data Factory, SQL Database, SQL Data Warehouse, CosmosDB

Google Cloud Platform: Compute Engine, Cloud Storage, Cloud SQL, Cloud Data Store, BigQuery, Pub/Sub, DataFlow, DataProc, Data Fusion, Data Catalog, Cloud Spanner, AutoML

Web Development: HTML, XML, JSON, CSS, JQUERY, JavaScript

Monitoring Tools: Splunk, Chef, Nagios, ELK

Source Code Management: JFrog Artifactory, Nexus, GitHub, CodeCommit

Containerization: Docker & Docker Hub, Kubernetes, OpenShift

Build & Development Tools: Jenkins, Maven, Gradle, Bamboo

Methodologies: Agile/Scrum, Waterfall

PROFESSIONAL EXPERIENCE

Senior Data Engineer

Confidential | Richmond, Virginia

Responsibilities:

  • Designed, and build scalable distributed data solutions using wif AWS & planned migration plan for existing on-premises Cloudera Hadoop distribution to AWS based on business requirement.
  • Worked wif legacy on-premises VMs based on UNIX distributions. Worked wif batch data as well as 3rd Party data through FTP. Configured traditional ETL tools- Informatica, SAP Data Service & IBM Data Stage.
  • Worked wif HDFS dat stores distributed data. Configured Oozie along wif Sqoop to ingest relational data.
  • Integrated & configured NiFi wif Sqoop/ Flume pipelines for graphical interpretation, encryption, scalability & easy organization. Implemented data transfer to APIs & other servers using NiFi.
  • Wrote YML files for Kafka Producers for ingesting streaming data. Assigned partitions to customers.
  • Imported data from to Cassandra cluster using Java API. Used DataStax OpsCenter and NodeTool to monitor clusters. Worked to deploy and maintain MongoDB clusters. Wrote scripts to create and configure Shard sets.
  • Used Snowflake for Data Warehousing. Developed Talend jobs to load data into Snowflake after extraction. Used Hive metastore connector to transmit events to keep external tables synchronized wif Hive metastore.
  • Worked on Hive queries and Python Spark SQL to create HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL databases, and a variety of portfolios.
  • Developed Scala scripts using both Data frames/SQL/Data sets and RDD/MapReduce in Spark for Data Aggregation, queries and writing data back into OLTP system through Sqoop.
  • Set up static and dynamic resource pools using YARN in Cloudera Manager for job scheduling & cluster resource management. Used Zookeeper for configuration management, synchronization & other services.
  • Perform data profiling, modeling and Meta-Data Management tasks on complex data integration scenarios adhering to Enterprise Data governance and Data Integration standards using Apache Atlas.
  • Developed the core search module using Apache Solr and customized the Apache Solr for handling fallback searching and to provide custom functions. Worked wif big data tools to integrate Apache Solr search.
  • Worked wif Snowflake to integrate PowerBI, Tableau & Quicksight for dashboards & visualizations.
  • Utilized Apache Spark ML & Mllib wif Python to develop and execute Big Data Analytics & ML.
  • Part of On-premises Hadoop infrastructure to AWS EMR Migration (Refactoring) Team. Participated in defining strategy (understanding motivations, business justification & expected outcomes), planning (drive incremental digital estate plan, skill readiness & cloud adoption plan), readying AWS Cloud Architecture (train on AWS Setup Guide, design landing zones, define best practices) & adoption (operate migration & innovation).
  • Worked wif S3DistCp tool to copy data from relational HDFS data to S3 buckets. Created a custom utility tool in Hive dat targets and deletes backed up folders dat are manually flagged.
  • Implemented auto-migration from MongoDB server in JSON format to DynamoDB wif AWS Data Migration Service Document Mode. Created Replication server, defined endpoints and wrote replication scripts.
  • Created S3 & EMR endpoints using Private Link & used AWS Private Subnet Network for fast transfers.
  • Used Pyspark scripts implemented on EMR to automate, compare & validate S3 files to the original HDFS files.
  • Developed Spark custom framework to load the data from AWS S3 to Redshift for data warehousing.
  • Experience in Jenkins for deployment of project. Help deploy projects on Jenkins using GIT. Used Docker to achieve delivery goal on scalable environment & used Kubernetes for orchestration & automation.
  • Integrated Apache Airflow and wrote scripts to automate workflows in AWS data pipeline.
  • Created Restful API using Flask to integrate functionalities & communicate wif other applications.
  • Integrated Teradata Warehouse into EMR cluster. Developed BTEQ scripts to load data from Teradata Staging area to Teradata data mart. Handled Error & tuned performance in Teradata queries and utilities.
  • Used AWS CloudFormation to defines all settings- provision hardware, define security, & set up elements for an EMR cluster. Wrote Infrastructure management as code, checked in and managed wif source control.
  • Helped develop a CodePipeline based on CodeCommit, CodeBuild and CodeDeploy wif CloudFormation.
  • Implemented AWS Athena as a replacement for Hive query engine. Migrated Analysis, end-users, and other processes (automated Tableau and other dashboard tools) to query S3 directly.
  • Worked on IAM, KMS, Secrets, Config, Systems Manager and others for security and access management.
  • Implemented AWS RDS to store relational data & integrated it along wif ElastiCache for load balancing.
  • Implemented Kinesis on EMR for streaming analysis. Implemented existing clickstream Spark jobs to Kinesis.
  • Used AWS Glue as the new ETL tool. Used Glue to catalog wif crawler to get the data from S3 and perform SQL query operations. Implemented ETL Scripts on Glue for data transformation, validation, and data cleansing.
  • Used DNS management in Route53 & configured CloudFront for access to media files (images). Worked wif AWS Rekognition for real-time content filtering. Wrote Lambda functions to resize/scale images .
  • Worked wif Data Analytics Services such as Athena, Glue Data Catalog. Used QuickSight for dashboards.

Environment: Hadoop(HDFS, Hbase, Hive, YARN, MarReduce, Pig, HIVE, Apache Storm, Sqoop, Oozie, Zookeeper, Spark, SOLR, Atlas), AWS (EC2, S3, Redshift, RDS, EMR, Lambda, Glue, CloudWatch, Rekognition, Kinesis, CloudFront, Route53, DynamoDB, CodePipeline, Athena, QuickSight), Python, MongoDB, Cassandra, Snowflake, Airflow, Tableau

Data Engineer

Confidential | Richmond, Virginia

Responsibilities:

  • Handled importing of data from various data sources, performed transformations using Hive & MapReduce. Loaded data into HDFS and extracted the data from MySQL into HDFS using Sqoop.
  • Used YARN for scheduling and managing resources. Worked to finetune Resource Manager (using Reservation System and RDL) & monitored Node Managers for optimum performance of on-premises servers. Also worked wif YARN Federation to integrate multiple YARN network for smooth communication wif multiple teams.
  • Wrote MapReduce scripts for applications dat handled data storage across server clusters. Used Hadoop Streaming to create jobs using UNIX Shell Scripts & integrated Hadoop Pipes wif custom Hadoop Applications. Performed MapReduce chains wif dependencies using JobControl, ChainMapper & ChainReducer.
  • Used Tez to transform MapReduce chained jobs to simplify workflow & improve resource management.
  • Integrated codes written using Scala, Java & Python on Spark for workloads like text analysis & querying.
  • Used SparkSQL to load JSON data and create Schema RDD and loaded it into Hive Tables & Cassandra.
  • Worked to design tables in Hive using SQOOP. Implemented data processing on large datasets of different forms including structured, semi-structured and unstructured data and loaded the data in HDFS.
  • Managed data from different data sources, performed transformation using Hive, Pig and MapReduce and loaded data in HDFS. Integrated Java codes wif Jython module in Pig to perform text preprocessing.
  • Used DAGs to design workflow of Hive, MapReduce and Pig jobs & implemented them wif Oozie. Monitored recurring as well as stream Oozie jobs and resolved issues wif resource availability.
  • Exported the analyzed data to database using Sqoop for visualization and to generate reports for the BI team.
  • Implemented Avro to serialize data stream and reduce processing time. Used Flume to collect, aggregate, and store the web log data from sources like web servers, mobile and network devices and pushed to HDFS.
  • Used Zookeeper to track cluster health, and handle coordination, leader election & fault tolerance. Used Ambari for health metrics & upgrading clusters using Rolling Upgrade(Ambari) and Manual Upgrade(Command line).
  • Used Apache Storm to build reliable real-time data pipelines to process unbounded streams of data generated from OLTP applications (mobile & web apps) and used HBase to save & process streaming data.
  • Build Cassandra queries for performing CRUD operations. Used Bootstrap to organize HTML page layout.
  • Created multiple data processing tasks using Spark dat included reading data from external sources, merge data, perform data enrichment and load in to target data destinations.
  • Wrote Kafka producers to stream data from external REST APIs to Kafka topics. Wrote Spark-Streaming applications to consume the data from Kafka topics and write the processed streams to HBase.
  • Worked wif batch ETL tools (Informatica PowerCenter, Microsoft SSIS) to extract & transform data and store the data in Oracle DB2. Understanding of other tools like IBM DataStage, Oracle Data Integrator.
  • Involved in installation, configuration, and deployment of Oracle DB2. Proficient in managing objects, users, privileges, and roles and experienced in creating, configuring, and utilizing Recovery Manager Catalog.
  • Designed & implemented on-Premises Hadoop distribution migration to Azure (Rehosting & Replatforming). Participated in designing strategy & expected outcomes, planning skill readiness and adoption plan, working wif AWS to determine optimal strategy & appropriate tools, and finally moving to the cloud.
  • Processes and moves data between different compute and storage services, as well as on-premises data sources at specified intervals. Coordinated to create, schedule, orchestrate, and manage data pipelines. Used Azure Data Factory to process and move Azure & scheduled on-premises data between compute and storage.
  • Used Azure Data Lake Gen2 for storage on large scale & implemented Data Lake Analytics for processing data.
  • Implemented Hadoop on-premises applications to Azure using Azure HDInsight. Also deployed Spark applications (Python, Scala, Java & SQL) using Spark environment implemented in DataBricks.
  • Automated an ELT workflow in Azure using Azure Data Factory wif Azure Synapse Analytics using Synapse Pipeline. Used Cosmos DB to store & replicate data for web applications across different datacenters in the US.
  • Worked PowerBI for custom visuals, presentations, and real-time dashboards for stream data.
  • Worked wif GCP and its various services- PubSub, Cloud Storage, Cloud Storage, Big Query, among others.
  • Helped deploy DataProc Service dat can run Apache Spark jobs & integrated cloud storage to the clusters.
  • Used Google Cron Service to build a task orchestrator on GCP dat can schedule jobs in data pipeline.
  • Used GCP Pub/Sub for ingestion from Streaming sources & replicating data to servers wif DataFlow.
  • Used DataFlow to ingest batch data, worked on Dataflow wif Apache Beam Unified Batch for validation.
  • Worked wif Cloud SQL for relational database services and extensively worked wif MySQL scripts. Also configured GCP Datastore and connected Pub/Sub streaming wif DataFlow for ingestion from other servers.
  • Wrote a program to download SQL dump and load to GCP Cloud Storage for Data Lake dat pulled information from servers to perform Big Query tasks. Built Spark Scala based configurable framework to connect common Data sources like MYSQL, Oracle, Postgres, SQL Server and load it in Bigquery.
  • Built dashboards in Tableau and Looker wif ODBC connections from like Big Query engine.
  • Monitored Bigquery, Dataproc and cloud Data flow jobs via Stackdriver for all environments.
  • Coordinated wif team and developed framework to generate Daily adhoc reports and extract data from various enterprise servers using BigQuery.

Environment: Hadoop(HDFS, Hbase, Hive, YARN, MarReduce, Pig, HIVE, Storm, Sqoop, Oozie, Zookeeper, Spark), Azure (VMs, Synapse Analytics, Blob, Databricks, Data Factory, SQL Database, SQL DataWare house, Cosmos DB), GCP (ComputeEngine, Cloud Storage, Cloud SQL, Data Store, BigQuery, Pub/Sub, DataFlow, Data Fusion, DataProc)

Data Engineer

Confidential

Responsibilities:

  • Analyzed, designed, and build scalable distributed data solutions using wif Hadoop, AWS & GCP.
  • Worked on multi-tier applications using AWS services (EC2, Route53, S3, RDS, Dynamo DB, SNS, SQS, IAM) focusing on high-availability, fault tolerance, and auto-scaling in AWS Cloud Formation.
  • Participated in documenting Data Migration & Pipeline for smooth transfer of project from development to testing environment and then moving the code to production.
  • Worked to implement persistent storage in AWS using Elastic Block Storage, S3, Glacier. Created Volumes and configured Snapshots for EC2 instances. Also, built, and managed Hadoop EMR clusters on AWS.
  • Used Data Frame API in Scala to convert distributed data into named columns & halped develop Predictive Analytics using Apache Spark Scala APIs.
  • Developed Scala scripts using both Data frames/SQL/Data sets and RDD/MapReduce in Spark for Data Aggregation, queries and writing data back into OLTP system through Sqoop.
  • Developed Hive queries to pre-process the data required for running business processes.
  • Implemented multiple generalized solution model using AWS SageMaker.
  • Extensive expertise using the core Spark APIs and processing data on an EMR cluster.
  • Worked on Hive queries and Python Spark SQL to create HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL databases, and a variety of portfolios.
  • Worked on ETL Migration services by developing and deploying AWS Lambda functions for generating a serverless data pipeline which can be written to Glue Catalog and can be queried from Athena.
  • Programmed in Hive, Spark SQL, Java, and Python to streamline & orchestrate the incoming data and build data pipelines to get the useful insights.
  • Loaded data into Spark RDD and in-memory data computation to generate the output response stored datasets into HDFS/ Amazon S3 storage/ relational databases.
  • Migrated Legacy Informatica batch/real time ETL logical code to Hadoop using Python, Spark Context, Spark-SQL, Data Frames and Pair RDD’s in Data Bricks.
  • Experienced in handling large datasets using partitions, Spark in-memory capabilities, Broadcasts in Spark, effective & efficient Joins, Transformation and other during ingestion process itself.
  • Worked on tuning Spark applications to set Batch Interval time, level of Parallelism and memory tuning.
  • Implemented near-real time data processing using Stream Sets and Spark/Databricks framework.
  • Stored Spark Datasets into Snowflake relational databases & used data for Analytics reports.
  • Migrated SQL Server Database into multi cluster Snowflake environment and created data sharing multiple applications and created snowflake virtual warehouses based on data volume/ Jobs.
  • Developed Apache Spark jobs using Python in the test environment for faster data processing and used Spark SQL for querying.
  • Used Hadoop Spark Docker container for validating data load for test/ dev-environments.
  • Worked on ETL pipeline to source tables and to deliver calculated ratio data from AWS to Datamart (SQL Server) & Credit Edge server.
  • Worked in tuning relational databases (Microsoft SQL Server, Oracle, MySQL, PostgreSQL) and columnar databases (Amazon Redshift, Microsoft SQL Data Warehouse).
  • Hands-on experience in Amazon EC2, Amazon S3, Amazon RedShift, Amazon EMR, Amazon RDS, Amazon ELB, Amazon CloudFormation, and other services of the AWS family.
  • Developed job processing scripts using Oozie workflow. Experienced in scheduling & job management.
  • Wrote Jenkins Groovy script for automating workflow including, ingesting, ETL and reporting to dashboards.
  • Worked in development of scheduled jobs using wif commands/BASH Shell in UNIX

Environment: Hadoop (Hive, Sqoop, Pig), AWS (EC2, S3, RedShift, EMR, EBS), Java, Django, Python, Flask, XML, MySQL, MSSQL Server, Shell Scripting, MongoDB, Python 3.3, Django, Cassandra, Docker, Jenkins, JIRA, jQuery

Python Developer

Confidential

Responsibilities:

  • Created APIs, Database Model and Views Utilization using Python to build responsive web page application.
  • Worked on a fully automated continuous integration system using Git, Gerrit, Jenkins, MySQL and in-house tools developed in Python and Bash.
  • Participated in the SDLC of a project including Design, Development, Deployment, Testing and Support.
  • Deployed & troubleshoot applications used as a data source for both customers and internal service team.
  • Wrote and executed MySQL queries from Python using Python-MySQL connector and MySQL dB package.
  • Implemented UI standards & guidelines in website development using CSS, HTML, JavaScript and jQuery.
  • Worked on a Python/Django based web application wif PostgreSQL DB and integrated wif third party email, messaging & storage services.
  • Developed GUI using webapp2 for dynamically displaying the test block documentation and other features of Python code using a web browser.
  • Involved in design, implementation and modifying back-end Python code and MySQL database schema.
  • Developed user friendly graphical representation of item catalogue configured for specific equipment.
  • Used BeautifulSoup for web scrapping to extract data & generated various capacity planning reports (graphical) using Python packages like NumPy, matplotlib.
  • Automated different workflows, which are initiated manually wif Python scripts and UNIX shell scripting.
  • Fetched Twitter feeds for certain important keyword using Twitter Python API.
  • Used Shell Scripting for UNIX Jobs which included Job scheduling, batch-job scheduling, process control, forking and cloning and checking status.
  • Monitored Python scripts dat are run as daemons on UNIX to collect trigger and feed arrival information.
  • Used JIRA for bug & issue tracking and added algorithms to application for data and address generation.

Environment: Python 2.7 (BeautifulSoup, NumPy, matplotlib), Web Development (CSS, HTML, JavaScript, JQuery), Database (MySQL, PostgreSQL), UNIX/Linux Shell Script, JIRA, Jenkins, GIT.

We'd love your feedback!