We provide IT Staff Augmentation Services!

Hadoop/scala Developer Resume

3.00/5 (Submit Your Rating)

Dallas, TX

SUMMARY

  • Over 8 years of IT experience, with around 4+years of experience in Hadoop and Hadoop Ecosystem.
  • Hadoop/Scala developer having end to end experience in developing applications in Hadoop ecosystem.
  • Hands on experience in installing, configuring and using ecosystem components like Hadoop Map Reduce, Hive, Sqoop, HDFS.
  • Knowledge in using Map Reduce programming model for analyzing the data stored in HDFS and experience in writing Map Reduce codes in Java as per business requirements.
  • Good experience working with Scala and Python.
  • Experience in extending Hive functionalities by writing custom UDFs.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Good understanding/knowledge of Hadoop architecture and various components such as HDFS, Resource manager, Node manager, Name Node, Data Node and Map Reduce programming paradigm.
  • Experience in installation, configuration, supporting and managing of Hadoop clusters along with Hive, Shark and Spark.
  • Hands on experience in installing and configuring Shark Server, Spark cluster, Spark SQL.
  • Familiar with Spark programming using RDD's
  • Familiar with Data Migration between Hadoop clusters using DISTCP.
  • Exposure on NOSQL Data bases such as Cassandra, MongoDB.
  • Good Experience in managing and reviewing Hadoop log files.
  • Used Zookeeper to provide coordination services to the cluster.
  • Rich Object Oriented Programming knowledge and Good Experience on various technologies and frameworks.
  • Experience with ORM tools like Slick, Hibernate to connectivity with RDBMS like Oracle and MySQL.
  • Experience with DI module of Spring.
  • Good experience with Agile Scrum Methodology as Software life cycle methodology.
  • Good working knowledge in IDE’s tool like Eclipse with configuration tools like Maven.
  • Development experience with Web/Application servers like Apache Tomcat.
  • Good Experience in front-end development with GXT, GWT.
  • Knowledge of version control using Git and Stash.
  • Good exposure to application development in coding, Integrating and reporting.
  • Good Analytical & Communication skills, Team player and highly self-motivated.
  • Flexible and versatile to adapt to any new environment and work on any project.
  • Self-starter, fast learner and a team player with good communication skills and personality traits.

TECHNICAL SKILLS

Big Data Technologies: Hadoop, HDFS, MapReduce, Hive, Sqoop

Hadoop Distributions: Cloudera CDH4

Languages: Java, Scala, Python and SQL

Databases: Oracle, MySQL, Cassandra, MongoDB

Source control: GIT

App/Web servers: Tomcat

Operating Systems: Windows Family, Linux, CentOS

Frameworks: Spring, Hibernate, Spark, Slick, Spray

IDEs: Eclipse, Microsoft Visual Studio

Scripting Languages: HTML, VBScript

PROFESSIONAL EXPERIENCE

Confidential, Dallas TX

Hadoop/Scala Developer

Responsibilities:

  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
  • Implemented several Akka Actors which are responsible for loading of data into hive.
  • Planning for production cluster hardware and software installation on production cluster and communicating with multiple teams to get it done.
  • Implemented entire dao layer in the application using slick for inserting and retrieving metadata into oracle tables.
  • Designed, configured and managed the backup and disaster recovery for HDFS data
  • Developed a Folder Watcher mechanism to watch several input folders.
  • Migrated data across clusters using DISTCP.
  • Implemented a generic Housekeeping mechanism to control the retention of data in Hadoop and hive.
  • Integration of JMS Queue, Publisher, Listener mechanisms.

Environment: Scala, Hadoop, Slick, Spray, Spark, Hive, Java, Ext Js

Confidential, Richmond VA

Hadoop/Java Developer

Responsibilities:

  • Implemented entire validation framework which is used for validating various source files incoming into the application.
  • Created and configured Sqoop jobs with incremental load to populate Hive External tables.
  • Involved in writing code for several DAO’s in the application
  • Implemented UDTF’s and UDF’s in the application to perform the necessary calculations and enrichment of data.
  • Involved in creating Hive tables, loading with data and writing hive queries.
  • Created generic frameworks for creating the UI screens, writing the hive loaders for loading mechanism.
  • Used Spark to enhance the performance of the project.
  • Assisted in creating and maintaining Technical documentation to launching HADOOP Clusters and even for executing Hive queries.
  • Performed Unit testing for all the modules in the application.

Environment: Core java, Hadoop, Hive, Spark, Shark, Spring, Hibernate, MySQL, GXT

Confidential, New York NY

Hadoop/Scala Developer

Responsibilities:

  • Implemented various roots for the application using spray.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Worked on the backend using Scala and Spark to perform several aggregation logics.
  • Developed MapReduce(YARN) jobs for cleaning, accessing and validating the data.
  • Written multiple MapReduce programs in java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.
  • Performed Data Migration between Hadoop clusters using DISTCP.
  • Worked with application teams (Onsite and Offshore) to install Hadoop updates, patches, version upgrades as required.
  • Provided extensive support to L2 team regarding the environment setup, monitoring and solving of several issues.
  • Used Spark to enhance the performance of the project.

Environment: Scala, Hadoop, Slick, Spray, Spark, Hive, Java.

Confidential, New York NY

Java/J2EE Developer

Responsibilities:

  • Developed many user interface screens for GPC2 using GWT.
  • Worked on core Java concepts like Collections and Exception Handling for writing the backend API's.
  • Used Oracle Coherence API for querying and transforming data in parallel across the cluster.
  • Used JDBC, with stored procedures to accomplish business logic implementation on server side.
  • Worked in developing application using Spring Core, Spring Jdbc and DAO layers to offer abstraction for the business from the database related code (CRUD).
  • Developed POJO’s, Data Access Object (DAO) which handles all database operations using HQL.
  • Used Hibernate as ORM to map Java classes to database tables.
  • Developed unit/integration test cases using JUnit and JMock.
  • Involved in regular meetings with client and business analysts for requirements

Environment: Core java, Oracle Coherence, Spring, Hibernate, MySQL, GWT

Confidential

Java Developer

Responsibilities:

  • Used Spring (MVC) Framework to develop the application. Used DAO and DTO Design patterns.
  • Used Rational Rose to draw UML diagrams and to develop the Use cases, Domain model and Design Model.
  • Implemented the functionalities using Java, J2EE, JSP, and AJAX, Servlets and spring framework.
  • Used Hibernate for Object Relational Mapping.
  • Developed Data Access Objects (DAOs) for easy data access.
  • Involved in the development of the presentation layer in Ext JS using spring framework.
  • Client pages are built using AJAX, HTML, JSP, XML and CSS.
  • Applied J2EE design patterns like Business Delegate, Service Locator, Session Façade, DAO etc.
  • Utilized JSP Standard Tag Libraries in the design and development of Views.
  • Involved Database programming in Oracle 9i.
  • Created the Stored Procedures using PL/SQL.
  • Developed Web-Services module for integration using SOAP, WSDL.

Environment: Java, J2EE, JSP, MVC, Spring 2.0, DAO, Hibernate 2.8, JNDI, Tomcat, Eclipse, Oracle 9i, PL/SQL, EXT JS, JMS, HTML, XML

We'd love your feedback!