We provide IT Staff Augmentation Services!

Software Engineer Resume

2.00/5 (Submit Your Rating)

West Chester, PA

SUMMARY:

  • 15+ Years of experience in IT industry.
  • Recent years Bigdata, NoSQL and Science experience mentioned below.
  • Kafka - spark streaming-AWS EMR- HBase(Opentsdb)-Grafana (Realtime analysis of Timeseries events via Dashboard)
  • Data lake(s3)-Greenplum-Tableau/Grafana (Troubleshooting and reporting use case)
  • Application logs-kafka-spark streaming-cloudera Hadoop-Spark ML/DL Keras/TensorFlow (Customer chunk prediction use case)
  • Datastax/apache Cassandra-spark streaming-Apache Hadoop/hive (Event as a platform use case)
  • Oracle RAC-Goldengate-Oracle Exadata (CDC Events analysis/archive use case)
  • Couchbase-Elasticsearch-Kibana (Realtime Customer event analysis)
  • Experience on setting up and supporting large dataset multinode Cloudera Hadoop cluster for DEV, STG & PROD Environment and used tools like Fluentd, Spark streaming, Java APIs for data ingestion from different sources to Hadoop system and Hive/Impala/Sparksql to query and Tableau for reporting.
  • Experience on setting up AWS EMR Cluster along with Hbase, Spark, Opentsdb(Time Series), EC2,S3(HDFS). Integrating with XSP Billing data service info via kafka platform using spark streaming to AWS EMR cluster.
  • 2+ experience on support/implementing complete workflow for multinode Elastic search clusters for Dev, QA & PROD for Application event log analysis using Fluentd as data ingestion tool.
  • 1.5 Years on Spark ML, Scikit-learn, NLP, Numpy and Tensorflow/Keras DL models using Jupiter notepad and Data cleansing.
  • Experience working on End to end model/flow for understanding different customer data patterns. for use cases i.e., Customer pattern detection:
  • Predict disconnect/dissatisfied customer across regions.
  • Anomaly detection for higher number of disconnections/more change of services, etc.
  • Working on POC for usecase i.e., Log analytics and Chatbot (to understand customer sentiments)
  • Using text analysis using structure extraction, TF-IDF, word2vec, stemming, tokenization, etc
  • Working closely with Deep Learning Engineer to develop RNN model using Tensorflow/Keras for the above use cases.
  • Day to day support to analyse/troubleshoot workflow performance issues on training/real set data.
  • 3+ years of experience on setting up multiple Kafka clusters datacenters. Worked on writing Producers/Consumers for different services via Kafka brokers. Used Confluent to setup Topic replication across multi Datacenters. Setup Grafana Dashboard/Kafka monitoring Tool on different metrics for Kafka producers/consumers and zookeeper services. Setup Kafka-Spark streaming for loading event data to Analytics platform.
  • 5+ years of experience on Oracle CDC bi/unidirectional replication from OLTP - Warehouse servers using Goldengate streaming.
  • 1.5 years of spark streaming experience from kafka.
  • 3+ year of experience on Couchbase technology i.e., Multiple Datacenter/cluster configuration, data migration from oracle to couchbase, High availability bidirectional replication for multi Datacenters, Redesign of Couchbase Buckets/views and automation of various tasks.
  • 4+ years of strong experience on Datastax/Apache Cassandra experience, migration from oracle to Cassandra & from physical servers to cloud, Upgradation/Data modelling/Dataload for new Cassandra clusters, Day to day support of Cassandra/ servers using Opscenter, Ganglia & Splunk for monitoring.
  • 9+ years of Experience on managing/working on Oracle database with RAC along with PL/SQL programming, Disaster Recovery and performance tuning, etc.
  • 1+ year of experience supporting Warehouse environment of size i.e., 500+ terabytes on Oracle Exadata X2 & X3.
  • 1+ year of experience on setting up Greenplum, loading event data to Greenplum and integrating with Tableau reporting tool.
  • 2+ years of experience on Timeseries databases i.e., OPENTSDB and INFLUXDB to provide Customer payload info in different aggregate intervals.
  • Cloud -> 5+ years on AWS (EMR, S3, RDS, SQS/SNS, VPC, Elasticsearch service,etc) and openstack steelcloud
  • 5+ years python scripting and programming experience for Cassandra/Couchbase/Hadoop. Scripts includes i.e., data comparison across multiple couchbase datacenters, couchbase/Cassandra data comparison with Oracle, Data loading, Map reduce/spark programming for agreegation/analytics requirement. Develop python APIs for bigdata platform services to internal customer.
  • 3+ Years of experience on Python coding experience for data filtering, text processing/analysis using NLP, Spark ML/Tensorflow/keras code for RNN model.
  • 2+ year experience of core Java programming/code work through /unit testing, implementing Rest APIs, support using spring boot/mvc framework. Support project activities include Log event workflow, DAO implementation for couchbase/Cassandra, implementing utils, etc.
  • 9+ year shell scripting experience for automation, monitoring, deployment & alerting for different environment i.e., Oracle/Couchbase/Cassandra/Hadoop, etc.
  • 6 months of experience on scala development project.
  • Agile/CI -> 5+ Years’ experience following Devops model, managing scrum teams & CI using Git, SVN, Jenkin.

WORK EXPERIENCE:

Confidential, West Chester, PA

Software Engineer

Responsibilities:

  • Lead to develop/design Bigdata platform for customer pattern analysis and use ML/AI pattern.
  • Retrieve/Archive Customer events from In-memory datagrid(Coherene/couchbase/appl logs) to stream to bigdata platform(using kafka-spark streaming) to hadoop/hdfs.
  • Build up analysis query /Dataframes on spark/hive and integrate with Grafana/tableau report.
  • Convert RAW datasets to a concise/flatten format to find the right features for prediction using python, scikit,spark ML.
  • Worked with Data scientist team to build Scikit/Spark ML and Deep Learning/RNN models to implement in production.
  • Build/support Rest APIs/microservices for customer info to integrate with Confidential internal dependent applications.
  • Worked on day to day support of clusters i.e., Couchbase, Cassandra, coherence clusters performance issues.
  • Automation of kafka/couchbase/Cassandra cluster/scaling deployment using ansible script and workflow management using Jenkins via git.
  • Worked on troubleshooting issues related to spark/Hadoop performance/scaling/query issues.
  • Follow agile process, played role of Scrum master to deliver Rally features on time.
  • DB level Support for legacy old system i.e., on Oracle RAC 11g and Goldengate to stream logs to standby databases to make sure few API services always up and running.
  • Initially started with 5+ Architect team focused on Single Enterprise analytics platform solutions for multiple applications REST services.
  • Have presented on various topics related to Nosql and BigData Analytics across IT platforms and also provided trainings to juniors/peers on various technology.
  • Conducted Workshop related to AI/Deep Learning with Data scientist across Confidential Applications.
  • Continue support for Teletraan project, mainly focused providing solutions for 3 usecases i.e.,
  • Category 1 Use cases: Telemetry and Metrics: Gather usage and performance metrics for all service operations as well as invocations of external systems for profiling and reporting purposes.
  • Develop and support kafka-spark java streaming to hbase-opentsdb and integrate report to Grafana.
  • AWS integration with VPC, automation the deployment and scalling using terraform script.
  • Involve on troubleshooting performance issues on hbase partition level, optimize kafka - spark partitioning, tune opentsdb queries for improving read latencies.
  • Support for recent migration system from opentsdb time series model to Influxdb model to use the new required features i.e., agreegation, indexing, granular downsampling.
  • Category 2 Use Cases - Troubleshooting and Historical reporting: Provide a tool for use by the DevOps teams in troubleshooting production and integration issues related to the services layers
  • For reporting, build new streaming services i.e., Spark streaming to s3 (data lake) and integrate with Greenplum stacks with Tableau for historical reporting and troubleshooting.
  • Integrate Greenplum with AWS VPC and implement incremental loading to Greenplum using python script.
  • Working on feature engineering tasks for NLP chatbot usecase i.e., loading and integrating data for customers, converting data to support model format using both scikit-learn and RNN sequence model way.
  • Creating different spark dataframes queries for analysis purpose of data.
  • Heavily used Jupiter notebook for developing ML/DL algorithm.

Confidential

Senior Analyst/Consultant

Responsibilities:

  • Worked as a Solution Architect to provide DBaaS solutions for Cassandra, couchbase along with automation of messaging platform.
  • Worked on Micro service/Rest APIs for multiple consumers.

Technology/Environment: Couchbase, Cassandra, Elasticsearch, AWS SQS/SNS,Kafka, Spark streaming, Spring Boot,, Java, Python,etc.

Confidential, New York, NY

Solution Architect

Responsibilities:

  • Along with automation of various jobs related to backup/recovery,implmenration,scaling,etc.
  • Wrote multiple Rest APIs as per consumer demand.

Technology/Environment: Oracle 10g/11g R1/R2, Casandra DSE 3.0.8/3.2.7 , Mongodb 2.2.4, Apache Hadoop v2, Spark, CDH4, shell/python, Java, Hive/Impala/Hbase/Flume/Sqoop, Java, APIS, etc.

Confidential

Senior Technical Specialist

Responsibilities:

  • My role is to provide DB Engineer support of multiple OLTP & Warehouse(Exadata/Hadoop) platform along with ETL operations i.e, Goldengate.
  • Worked on writing code/script to integrate various platforms.
  • Involved on lots of initiative to automate tasks and also lead a team of 6 members from Offshore team while working as an Onsite coordinator.
  • Also worked on migrating from oracle to Cassandra.

Technology/Environment: Hadoop/Hive, MySQL, pl/sql, Exadata X2/X3, python, Oracle RAC 10g/11g, Goldengate Director/Monitor/Veridata, Casandra 2.1.2, etc.

Confidential

Senior Technical Specialist

Responsibilities:

  • My role is supporting their Database /Warehouse/ETL platforms (Oracle, RAC), Goldengate, ETL Tools.
  • Also taken initiative for automation of various day to day jobs/reports, etc along with various Disaster recovery plan, backup strategy.

Technology/Environment: Oracle, Goldengate,RAC,Shell scripting,Python,Cognos, etc.

We'd love your feedback!