We provide IT Staff Augmentation Services!

Hadoop/spark Developer Resume

3.00/5 (Submit Your Rating)

DallaS

PROFESSIONAL SUMMARY:

  • Certified AWS Developer with 3 years of experience as JAVA and Hadoop/Spark Developer with strong skills in providing solutions to business problems using Big Data analytics.
  • Extensive experience in implementing, consulting and managing Hadoop Clusters & eco system components like HDFS, MapReduce, Spark, Hive, Sqoop, Flume, Oozie, Confidential & HBase.
  • Hands on experience with Scala programming with good knowledge on Spark Architecture and its In - memory Processing.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.
  • Experience in analyzing data using HIVEQL and custom MapReduce programs in JAVA.
  • Expertized in Implementing Spark using Scala and Spark SQL for faster testing and processing of data.
  • Experience in NoSQL databases such as HBase and Cassandra.
  • Worked on different job workflow scheduling and monitoring tools like Confidential and Confidential .
  • Involved in converting Hive/SQL queries into Spark transformations using RDDs and Scala.
  • Very good understanding of Static and Dynamic Partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Exploring with teh Spark for improving teh performance and optimization of teh existing algorithms in Hadoop using Spark Context, Spark SQL, Data Frame, Pair RDD's, YARN.
  • TEMPHas good understanding on Python.
  • In depth understanding of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce.
  • Experienced with Cloudera and MapR Distributions.
  • Expert in Java based technologies in design, development and testing of web application and integration projects using Object Oriented technologies such as Core Java, Java Collection Framework, Spring MVC, AOP, J2EE, JDBC, Hibernate, Servlets, JavaScript, AngularJS, XML, JSON.
  • Web Services experience using JAXRS, SOAP, WSDL, Rest API with Jersey, Spring Rest, Rest Easy and SOAP Services.
  • Experience of working with Oracle, SQL/MySQL and developing applications by using JDBC, JSP Servlets, Hibernate and Web Services.
  • Experience with ANT and Maven for teh build and deployment of applications.
  • Expertise in version control systems like GIT, SVN subversion.
  • Working knowledge about using Agile and SDLC methodologies.
  • Hands-on experience in AWS services like EC2, Docker, IAM, S3, VPC, Kinesis, Dynamo DB, RDS, Cloud Watch, SNS, SQS, Security Group, API-Gateway, Cloud Formation, Auto Scaling, Glacier, Lambda and Elastic Load Balancing.

PROFESSIONAL EXPERIENCE

Confidential, Dallas

Hadoop/Spark Developer

Job Responsibilities:

  • Used Spark -Streaming APIs to perform necessary transformations and actions on teh fly for building teh common learner data model which gets teh data from Kafka in near real time and Persists into Cassandraand HIVE.
  • Used Spark API over Cloudera HadoopYARN to perform analytics on data in Hive.
  • Loaded teh data into SparkRDD and do in memory data Computation to generate teh Output response.
  • Optimizing of existing algorithms in Hadoop using SparkContext, Spark -SQL, Data Frames and Pair RDD's.
  • Performed advanced procedures like text analytics and processing, using teh in-memory computing capabilities of Spark using Scala.
  • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
  • Experienced in handling large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, TEMPEffective & efficient Joins, Transformations and other during ingestion process itself.
  • Have an experience to load and transform large sets of structured, semi structured and unstructured data, using SQOOP from Hadoop Distributed File Systems to Relational Database Systems Also Relational Database Systems to Hadoop Distributed File Systems.
  • Written teh Apache PIG scripts to process teh HDFS data.
  • Created HIVE tables to store teh processed results in a tabular format.
  • Used Hive QL to analyze teh partitioned and bucketed data and compute various metrics for reporting.
  • Performed data transformations by writing MapReduce as per business requirements.
  • Implemented schema extraction for Parquet and Avro file Formats in Hive.
  • Active involvement in Scrum meetings and Followed Agile Methodology for implementation.
  • Implemented Kafka Storm topologies, which are capable of handling and channelizing high stream. of data and integrating teh storm topologies with Esper to filter and process dat data across multiple clusters for complex event processing
  • Involved in various NOSQL databases like HBase, Cassandra in implementing and integration.
  • Queried and analyzed data from Cassandra for quick searching, sorting and grouping through CQL.Responsible for developing data pipeline using flume, sqoop and pig to extract teh data from weblogs and store in HDFS Designed and implemented various metrics dat can statistically signify teh success of teh experiment.
  • Extensively worked with SQOOP for importing and exporting data into HDFS and Hive.

Environment: Sqoop, Hive, HDFS, YARN, HBase, Spark, Scala, Java, Informatica, Avro, EMR, Kafka, Cassandra, Oozie, Cloudera distribution, Oracle, Confidential, NiFi

Confidential

JAVA Developer

Responsibilities:

  • Used Agile SCRUM Methodology to develop teh application.
  • Developed RESTful application using Spring MVC, Spring Boot Framework with Java and Cassandra as a Back end.
  • Developed Java API to interact with teh Amazon SQS used for sending bulk emails.
  • Used Hibernate framework to persist teh data into teh database.
  • Worked with various spring components like Spring IOC, Spring AOP, Spring MVC, Spring JDBC, Spring JPA, Spring Security, Spring boot.
  • Developed RESTful web services for exposing teh functionalities, generated JSON response and exposed teh REST API using Swagger UI.
  • Developed REST services using Spring RestTemplate and JSON parsing using Jackson fasterxml.
  • Configured Spring Boot Application by adding dependencies in Gradle.
  • Configured Spring to inject DAO objects into business layer and Hibernate to map POJOs with Oracle Database tables using XML configuration file.
  • Used Mockito to write Junit test cases, worked with different annotations such as @Mock, @Spy, @MockBean, @InjectMocks.
  • Focused primarily on teh spring components such as Dispatcher Servlet, Controllers, Model and View Objects and changed teh controllers based on new JSON architecture.
  • Implemented RESTful clients using Spring RestTemplate where Basic Autantication was resolved.
  • Installed Security Plugins SonarQube and FindBugs to detect Security issues for teh application and fixed teh security issues.
  • Extensively used log4j for logging teh log files.
  • Used GIT repository for software configuration management and version control.

Environment: IntelliJ IDEA, GIT, Maven, Gradle, Java EE, Spring (MVC, AOP), Spring Boot, Swagger, Mockito, Gradle, JPA, Cassandra, Tomcat, Mac OS X, Oauth2, log4J, Cassandra MySQL.

TECHNICAL SKILLS

Programming Languages: JAVA, Scala

Hadoop Ecosystem: MapReduce, Spark, Kafka, Hive, HBase, Sqoop, Oozie, Scala, Kafka, Confidential, YARN

Data Visualization: Tableau

Databases: HBase, Cassandra, Oracle, MySQL

Web Technologies: JavaScript, Web Services

Servers: Tomcat and WebSphere

IDE’s: Eclipse, IntelliJ

We'd love your feedback!