We provide IT Staff Augmentation Services!

Lead Bigdata Developer Resume

0/5 (Submit Your Rating)

Plano, TX

SUMMARY

  • Big Data and AWS Development enthusiast with around 14 years of experience in Managing, Analysis, Design, Development and Implementation of Multi - Tier Enterprise Applications using Apache Spark, Java, JEE, Spring framework, Hibernate and other client-server technologies.
  • Overall around 14 years of industry experience in Java/J2ee with around 2 years of experience in Big data development and Cloud based technologies like Apache Spark and AWS
  • Experience in designing and development of data processing applications using Apache Spark, Spark streaming in Java and Scala to load data to cloud services such as S3 and MPP databases like Redshift and Snowflake and deploy applications in EMR and spark clusters
  • Create cloud formation templates to deploy EC2s and databases on AWS.
  • Programming using AWS Lambda to create server-less applications.
  • Experience in automating deployment of Cloud services using Jenkins
  • Learning and minimal working knowledge with Hive and Pig
  • Hands on experience in all areas of Software Development Life Cycle (SDLC) Analysis, Design, Development, Documentation, Production support, Testing and Implementation.
  • Experience in implementing projects using Agile methodologies.
  • Extensive experience in software configuration management (SCM) tools like Github, Borland Starteam, Rational Clearcase and CVS.
  • Experience in using change control and defect tracking tools like Service Center, HPSM etc..
  • Experience in estimating effort, iteration planning, mentoring new developers and monitoring and reporting progress.
  • Extensive experience in Java, Jsp, Java Script, Struts, WebServices and Hibernate, UNIX, Oracle, PL/SQL.
  • Extensive experience working on large scale, complex systems and integration projects.
  • Assisted business partners in resolving IT issues and facilitate working with other IT teams.
  • Project Management, Project Schedule, Project Tracking and Team Management
  • Rich experience in translating Business requirements to understandable IT systems requirements and Implementing CMM Methods & Procedures.

TECHNICAL SKILLS

  • Java 1.7/8
  • Apache Spark
  • Scala
  • Kafka
  • Snowflake
  • Redshift
  • AWS S3
  • AWS Lambda
  • EC2
  • AWS EMR
  • AWS Cloud Formation
  • Kafka-Connect
  • MemSQL
  • UNIX
  • Snaplogic
  • Hadoop
  • Pig
  • Hive
  • NoSQL DBs
  • Maven
  • Spring Framework
  • Jenkins
  • Struts
  • Hibernate/JPA
  • Project Management
  • APIs
  • MVC
  • Tomcat/JBOSS
  • Web-Services
  • JMeter
  • Java Performance
  • JAX-WS
  • EJB
  • Weblogic/WebSphere
  • CMS
  • JUnit
  • JAX-RPC
  • RAD/Eclipse
  • Oracle/MySQL/DB2
  • SOA
  • JQuery
  • Agile Methodologies

PROFESSIONAL EXPERIENCE

Confidential, Plano, TX

Lead BigData Developer

Responsibilities:

  • Working as a Lead Big Data developer, design and develop applications to migrate data from Hadoop data-lake databases and other sources to AWS cloud (s3). Create frameworks for automatic data quality checks, refine datasets and move data to Warehousing databases like Redshift and Snowflake using Apache spark using scala and java
  • Design and develop spark streaming applications to consume data from kafka, process and transform data
  • Create frameworks in Apache spark using scala and java for automatic data quality check of datasets, perform tokenization of sensitive data
  • Design and develop applications to load validated and refined datasets to Warehousing databases like Snowflake and Redshift
  • Developed a test suite for performance testing MPP databases like Redshift, snowflake and MemSQL
  • Cloud formation templates for automatically deploy applications, database clusters using auto-scaling functionalities
  • Designing pipelines to automate data processing, transformations using Snaplogic tool
  • CreateAWS lambda functions to create server-less applications to process the datasets in S3 using Java 8
  • Configure and deploy spark applications on AWS EMR and spark clusters
  • Pipelines using kafka connect connecting different database as sources to load data to S3

We'd love your feedback!