We provide IT Staff Augmentation Services!

Spark Developer Resume

0/5 (Submit Your Rating)

Beaverton, OregoN

SUMMARY

  • Around 6 years of experience in IT industry, including Java, SQL, Big data environment, Hadoop ecosystem and Design, Developing, Maintenance of various applications in the domains like HealthCare and E - Commerce applications.
  • Good in YARN, Oozie MapReduce, Spark, Hive 1.2.2, Sqoop 1.4.6, HBase, Spark 2, Kafka 0.10.2 ecosystem components.
  • Excellent knowledge of Hadoop Architecture and its related components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Have a hands-on experience on fetching the live stream data from web applications using Spark Streaming and Apache Kafka.
  • Experienced in Developing Spark application using Spark Core, Spark SQL and Spark Streaming API's.
  • Good in Spark framework for batch and real-time data processing.
  • Good Knowledge of Spark and Scala programming.
  • Experience in handline messaging services using Apache Kafka.
  • Good Knowledge in apache Nifi.
  • Data Ingestion in to Hadoop (HDFS): Ingested data into Hadoop from various data sources like Oracle, MySQL using Sqoop tool. Created Sqoop job with incremental load to populate Hive External tables.
  • Good in writing Map-Reduce Jobs in Java for processing large sets of structured semi-structured and unstructured data sets and stores them in HDFS.
  • Expertise experience in writing Hive scripts for processing and analyzing large volumes of structured data.
  • Experience in analyzing data using HiveQL.
  • Good knowledge in NoSQL databases like HBase and Cassandra.
  • Managing and scheduling Jobs to remove the duplicate log data files in HDFS using Oozie.
  • Experience in Oozie and workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph (DAG) of actions with control flows.
  • Extensive knowledge on file formats like AVRO, Sequence files, Parquet, ORC.
  • Experience in managing Hadoop clusters using Cloudera Manager Tool and Ambari.
  • Experience in working in Hadoop eco-system integrated to Cloud platform provided by
  • AWS with several services like Amazon EMR instances, S3 bucket.
  • Worked with version control systems like GIT for providing common platform for all the developers.
  • Experience in working with web services like RESTful and SOAP in JAVA for data transfer.
  • Strong programming skills in designing and implementation of applications using Core Java, HTML, WebLogic, J2EE includes JDBC, JSP, Servlets.
  • Knowledge of java virtual machines (JVM) and multithreaded processing.
  • Used Jira to monitor and track the project updates.
  • Good experience in Software development life cycle.

TECHNICAL SKILLS

Hadoop Technology \Databases: \: Apache Hadoop (MRv2), Hive1.2.2, Sqoop \Oracle 9g/11g/12c, MySQL 5.7.18, HBase \1.4.6, HBase, Flume, Kafka 0.10.2, Spark \1.2.6, MS SQL server 2016, Cassandra 3.1.0\2.1.1, Zookeeper, Oozie\

IDE’s \Web Services: \: Eclipse, Visual Studio\SOAP & RESTful Web Services\

Operating Systems \Build Tools: \: Windows, Ubuntu\Maven, Sbt\

Programming Languages \Web Technologies: \: Java, Scala, SQL/PLSQL.\HTML, JSP, CSS, \

Cloud Technologies \Scripting languages: \: Amazon Web Services (AWS) \Shell\

PROFESSIONAL EXPERIENCE

Confidential - Beaverton, Oregon

Spark Developer

Responsibilities:

  • Used Spark Scala APIs to perform necessary transformations and actions on the data which gets from Kafka in real time.
  • Configured Kafka with Spark Streaming to collect the data from the Kafka.
  • Experience in handling more than 2 TB of daily data.
  • Performed various Parsing technique's using spark API'S to cleanse the data.
  • Used Spark API over Horton works Hadoop YARN to perform analytics on data in HDFS.
  • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
  • Analyzed the data using SparkSQL Queries.
  • Experienced in working with Spark SQL on different file formats like Avro and Parquet and created data frames for processing.
  • Used Spark data frame API for data processing.
  • Very good understanding of Partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Implemented Incremental Imports of analyzed data into MYSQL tables using Sqoop.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Moved Relational Database data using Sqoop into Hive Dynamic partition tables using staging tables.
  • Implemented the workflows using Apache Oozie framework to automate tasks.

Environment: Spark streaming/Scala, Spark SQL, Hive, Kafka, Sqoop, Hortonworks Distribution of Hadoop, Oozie, Java (JDK 1.6), Eclipse, HDFS

Confidential, basking ridge, NJ

Hadoop developer

Responsibilities:

  • Used Hive and MapReduce to perform the ETL on the data received from the various resource.
  • Used Nifi to schedule the data incremental data loads.
  • Created Hive queries for extracting data and sending them to clients.
  • Very good understanding of Partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Performed different types of joins on Hive tables and implementing Hive SerDe.
  • Hands on experience in HiveQL.
  • Used AWS services like S3 for storage and EMR for data processing.
  • Worked on ingesting data from different sources.
  • Implemented the workflows using Apache Oozie framework to automate tasks.
  • Moved Relational Database data using Sqoop into Hive Dynamic partition tables.
  • Exported the analyzed data to relational databases using Sqoop for visualization and to generate reports.
  • Followed agile methodology during project delivery.
  • Used Git and Jira for project integration and monitoring.

Environment: Hive, Nifi, MapReduce, Sqoop, EMR, S3

Confidential

Java/J2EE Programmer

Responsibilities:

  • Code development and debugging with Eclipse.
  • Development of UI using Java Server Pages.
  • Developed Java code and Session Beans and Deployed the Application in Web Logic 8.1
  • Worked extensively with JSP's and Servlets to accommodate all presentation customizations on the front end.
  • Preparation of Unit Test cases.
  • Performed Unit Testing, Integration Testing.
  • Involved in bug fixing and enhancements

Environment: Core Java, JSP, JDBC, Servlets, ANT, XML, Oracle, DOM, Web Logic, Junit, log4j, Eclipse

Confidential

Java/J2EE Developer

Responsibilities:

  • Worked with functional leads to discuss business specifications and transformed them in to technical requirements.
  • Developed environment set up document and assisted development team to install development environment.
  • Assisted in developing the technology proposal document required for the system.
  • Designed the application workflow using Struts 2.0 and authored struts configuration
  • Developed core business components to automate the process of screening applicants for the requirement.
  • Developed REST web services for internal consumers.
  • Created user interfaces for the presentation layer using JSP, HTML and JavaScript.
  • Designed front-end, data driven GUI using JSP, HTML, JavaScript and CSS.
  • Developed business components in the rules engine using business components. XML based configuration is provided to configure rules engine dynamically on the run time.
  • Worked with the database team to analyze the data from development perspective.
  • Written JUNIT test cases for unit

Environment: Java, JDBC, Servlets, JSP, Struts 2.0, Web Logic 10.3, Eclipse, Web Services, Oracle 10g, JDBC, HTML, JavaScript, ANT, Soap UI, CVS

Confidential

Java/J2EE Programmer

Responsibilities:

  • Gathered specifications for the Library site from different departments and users of the services.
  • Assisted in proposing suitable UML class diagrams for the project.
  • Wrote SQL scripts to create and maintain the database, roles, users, tables, views, procedures and triggers in Oracle
  • Designed and implemented the UI using HTML, JSP, JavaScript and Java.
  • Implemented Multi-threading functionality using Java Threading API
  • Extensively worked on IBM Web Sphere 6.0 while implementing the project.
  • Involved in Analysis, Design, Coding and Development of custom Interfaces.
  • Involved in the feasibility study of the project.
  • Gathered requirements from the client for designing the Web Pages.
  • Participated in designing the user interface for the application using HTML, DHTML, and Java Server Pages (JSP).
  • Involved in writing Client-side Scripts using Java Scripts and Server-Side scripts using Java Beans and used Servlets for handling the business.
  • Developed the Form Beans and Data Access Layer classes.
  • XML was used to transfer the data between different layers.
  • Involved in writing complex sub-queries and used Oracle for generating on-screen reports.
  • Worked on database interaction layer for insertions, updating and retrieval operations on data.
  • Deployed EJB Components on Web Logic.
  • Involved in deploying the application in test environment using Tomcat.

Environment: Java, Servlets, JDBC, HTML, DHTML, SQL Server, Web Logic

We'd love your feedback!