We provide IT Staff Augmentation Services!

Hadoop/spark Developer Resume

4.00/5 (Submit Your Rating)

SUMMARY

  • Strong in Big Data tools and cluster computing framework. And hands on experience in creating teh RDD's, DF's for teh required input data and performed teh data transformations using Spark Scala.
  • As an Analyst worked in a high performance team within a professional environment and implementing practical business solutions under multiple deadlines
  • Good knowledge of functional programming Scala, and also proficient in software engineering principals and relational database design principals.
  • Proven problem solving, self - starter, co-operative and worked individually as well as adapting rapid changes in team environments and priorities.
  • Able to collaborate with clients, identify engagement follow-on opportunities, and has a strong desire to excel

TECHNICAL SKILLS

Big Data/Hadoop Technologies: HDFS, YARN, MapReduce, Hive, Pig, Sqoop, Flume, Spark, Kafka, Zookeeper, Talend and OOZie

NOSQL Databases: HBase, Cassandra

Languages: Scala, SQL, Pig Latin, HiveQL, Java, Unix, Shell Scripting

Databases: Microsoft SQL Server, MySQL, Oracle

Build Tools: Maven, SBT

Development Tools: Microsoft SQL Studio, Eclipse

Development Methodologies: Agile/Scrum, Waterfall

PROFESSIONAL EXPERIENCE

Confidential

Hadoop/Spark Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions with Apache Spark using Scala.
  • Optimization of existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frames and Pair RDD's.
  • Worked with large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins and Transformations.
  • Implemented Partitioning, Dynamic Partition, Buckets in HIVE.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Load and transform large sets of structured, semi-structured data.
  • Developed OOZie workflow on actions Spark, Linux script, HIVE scripts and HBase loads.
  • Worked on Talend open studio for designing ETL Jobs for Processing of data.
  • Used Reporting tool SpotFire to connect with Hive and HBase for generating Financial data reports.
  • Worked in creating HBase tables to load large sets of data from reporting layer
  • Experienced with performing CURD operations in HBase.
  • Actively involved in code review and bug fixing for improving teh performance.

Confidential

Hadoop/Spark Developer

Responsibilities:

  • Designed, developed and did maintenance of data integration programs in a Hadoop and RDBMS environment with both traditional and non-traditional source systems as well as RDBMS and NoSQL data stores for data access and analysis.
  • Used Spark-Streaming APIs to perform necessary transformations and data model which gets teh data from Kafka in near real time and Persists into Cassandra.
  • Configured deployed and maintained multi-node Dev and Test Kafka Clusters.
  • Developed Spark scripts by using Scala shell commands as per teh requirement.
  • Used Spark API over Hadoop YARN to perform analytics on data in Hive.
  • Developed Scala scripts, UDFFs using both Data frames/SQL/Data sets and RDD in Spark 1.6 for Data Aggregation, queries and writing data back into HIVE.
  • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
  • Worked on POC's with Apache Spark using scala to implement spark in project.
  • Experience in Job management using Capacity scheduler and Developed job processing scripts using Oozie workflow.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Implemented schema extraction for Parquet and Avro file Formats in Hive.
  • Developed Hive queries to process teh data and generate teh data cubes for visualizing
  • Understanding in Core Java in Multithreading, Collections, Concurrency, and Exception handling concepts, Object-oriented analysis, design, and development.
  • Collaborated with teh infrastructure, network, database, application and BI teams to ensure data quality and availability.

Environment: Hadoop YARN, Spark Core, Spark Streaming, Spark SQL, Scala, Java, Kafka, Hive, Sqoop, Cassandra, Oozie, Hortonworks, Linux.

Confidential

PeopleSoft

Responsibilities:

  • Created business requirements, process and analyzed to understand, defined, documented, analyzed and demonstrate functional and technical solution requirements related to Human Resources, Time & Labor, Finance /SCM and Security Role reviews.
  • Conducted Fit-Gap sessions with teh user group and teh business team.
  • Single handedly Project Managed Portal Upgrade with leadership and guidance for development and implementation of teh Upgrade

Confidential, Delaware

Technical Lead / Upgrade Specialist/ Functional Analyst

Responsibilities:

  • PeopleSoft integration using app engine programs & peoplecode to external system.
  • Created SQL views to optimize performance and used views as search records.
  • Upgraded PeopleSoft HRMS 7.5 to 8.0 SP1 using Upgrade Assistant.
  • As a lead responsibility include collecting pre-upgrade planning documents and fit gap analysis.

Confidential, Delaware

PeopleSoft 7.5

Responsibilities:

  • Lead conversion project for loading dependent data from BCBS (Current and Historical) into PeopleSoft
  • Rewrote SQR processes in daily jobs - Increased performance by tuning SQL statements and programming logic, which reduced total time from 12 hours to 5 hours.

We'd love your feedback!