Sr. Hadoop & Spark Developer Resume
Henderson Las, VegaS
SUMMARY
- I am a Cloudera certified Hadoop and Spark developer, having 9+ Years of extensive IT experience which includes 4+ years of experience in Big Data within Banking, Insurance and Retail.
- I built my career with variety of technologies, where I am not just the developer but also an admin and the best team player.
- Worked extensively with Hadoop Distributions - Cloudera and Hortonworks. I have very good experience in Hadoop & Spark ecosystem components to handle data ingestion, data storage, data processing and data visualization layers effectively.
- I am a quick learner and have thirst for learning new things, pursuing knowledge and ability to adapt the changes to new environment at fast pace.
TECHNICAL SKILLS
Programming / Web tech: C, C++, Core Java, Scala, Python, JavaScript J2EE- Servlets, JSP, JSF, Maven, Tomcat, JBoss
RDBMS: MySQL, IBM DB2, Oracle, MS SQL Server
Cloud Services: Amazon Web Services (AWS)
IDEs / Version Control: Eclipse, IntelIJ, SVN, GIT
SDLC - Methodologies: Agile Methodology, Waterfall model
PROFESSIONAL EXPERIENCE
Confidential, Henderson, Las Vegas
Sr. Hadoop & Spark Developer
Skill Set: Hadoop - HDFS, Spark - SQL & streaming, Kafka, Sqoop, Hive, Core Java, Scala, Unix Shell Scripting, Oozie workflows, Ambari - Hortonworks, Informatica Power center.
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop and Spark.
- Developing data ingestion pipelines using Sqoop and Kafka to ingest the database tables and streaming data into HDFS for analysis.
- Developing spark streaming application to receive the data streams from Kafka and process the continuous data streams and trigger actions based on fixed events.
- Teamed up with architects to design Spark streaming model for the existing Map Reduce model and migrating the Map Reduce models to Spark using Scala.
- Using Hive to analyze the partitioned and bucketed data and compute various metrics for creating dashboards in Tableau.
Confidential
Hadoop & Spark Developer
Skill Set: Hadoop - HDFS, HBase, Spark core - RDDs, Spark Streaming, Kafka, Sqoop, Hive, Core Java, Scala, Amazon Web Services - EC2, Lambda services, S3 storage.
Responsibilities:
- Worked on ingesting huge amount of Click Stream and custom application data into Hadoop as various file formats like raw text files, CSV, ORC, Parquet files with different compression codecs.
- Involved in designing and development of various custom data processing modules in Spark core, Spark SQL and Hive.
- Worked extensively on integrating Kafka (Data Ingestion) with Spark streaming to achieve high performance real time processing system.
- Worked on improving the in-memory computing performance of Spark applications by optimizing the Spark core RDD transformations based on requirement.
Confidential
Hadoop Developer
Skill Set: Hadoop - HDFS, Sqoop, Hive, Pig, MySQL, Core Java, Unix Shell Scripting, J2EE - JPA, JSF, JDBC, Python.
Responsibilities:
- Involved in loading large sets of structured, semi structured and unstructured data from various sources and transforming into required formats.
- Worked on setting up environment and updating configurations for executing Pig-Latin scripts and Sqoop commands via Unix shell scripting.
- Involved in writing PIG and Hive User Defined Functions (UDFs) based on the custom and frequent analysis requirements.
- Involved in writing Hive scripts with various source formats of same data to identify the performance and query execution time for Hive optimization.
Confidential, Los Angeles, California
Big Data Developer
Skill Set: Hadoop - HDFS, Apache Hive, Apache Pig, Apache Sqoop, Apache Oozie, MSBI - SSIS, MS SQL Server.
Responsibilities:
- Worked on ingesting data from MS SQL server to HDFS using Sqoop commands.
- Extensively worked on optimizing Sqoop data ingestion by enhancing to delta and Incremental data extraction from source systems.
- Involved in writing Pig Latin scripts to implement the cleaning operations and data transformations based on the requirements.
- Involved in writing custom Hive - HQL queries for order and sales analysis based on adhoc requirements.
- Involved in monitoring several running batch jobs.
Confidential
Oracle PL/SQL Developer
Skill Set: Oracle 10g Express, Toad for Oracle, PL/SQL, MS SQL Server, Microsoft VSS, Microsoft VB.Net, SQL Loader, UNIX.
Responsibilities:
- Involved in Production Support and Operations teams to resolve production issues in a timely and efficient manner.
- Worked on troubleshooting the existing PL/SQL procedures, functions, triggers
- Used PL/SQL, Ref Cursors to process the data and used bulk collect and bulk bind for mass update as performance improvement process.
- Involved in SQL Tuning by creation of indexes, rebuilding Indexes using Explain Plan, SQL Trace and TKPROF Tools.
Confidential
JAVA Developer
Skill Set: Java, J2EE, JSP, Struts, JNDI, DB2, HTML, XML, DOM, SAX, ANT, AJAX, Eclipse, SOAP, Apache Tomcat, Oracle 10g Express, Log4J, SVN.
Responsibilities:
- Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
- Involved in Documentation and Use case design using UML modeling include development of Class diagrams, Sequence diagrams, and Use Case Transaction diagrams.
- Developed User Interface using AJAX in JSP and performed client-side validation.