Hadoop/scala Developer Resume
Dallas, TX
SUMMARY
- Over 8 years of IT experience, with around 4+years of experience in Hadoop and Hadoop Ecosystem.
- Hadoop/Scala developer having end to end experience in developing applications in Hadoop ecosystem.
- Hands on experience in installing, configuring and using ecosystem components like Hadoop Map Reduce, Hive, Sqoop, HDFS.
- Knowledge in using Map Reduce programming model for analyzing the data stored in HDFS and experience in writing Map Reduce codes in Java as per business requirements.
- Good experience working with Scala and Python.
- Experience in extending Hive functionalities by writing custom UDFs.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Good understanding/knowledge of Hadoop architecture and various components such as HDFS, Resource manager, Node manager, Name Node, Data Node and Map Reduce programming paradigm.
- Experience in installation, configuration, supporting and managing of Hadoop clusters along with Hive, Shark and Spark.
- Hands on experience in installing and configuring Shark Server, Spark cluster, Spark SQL.
- Familiar with Spark programming using RDD's
- Familiar with Data Migration between Hadoop clusters using DISTCP.
- Exposure on NOSQL Data bases such as Cassandra, MongoDB.
- Good Experience in managing and reviewing Hadoop log files.
- Used Zookeeper to provide coordination services to the cluster.
- Rich Object Oriented Programming knowledge and Good Experience on various technologies and frameworks.
- Experience with ORM tools like Slick, Hibernate to connectivity with RDBMS like Oracle and MySQL.
- Experience with DI module of Spring.
- Good experience with Agile Scrum Methodology as Software life cycle methodology.
- Good working knowledge in IDE’s tool like Eclipse with configuration tools like Maven.
- Development experience with Web/Application servers like Apache Tomcat.
- Good Experience in front-end development with GXT, GWT.
- Knowledge of version control using Git and Stash.
- Good exposure to application development in coding, Integrating and reporting.
- Good Analytical & Communication skills, Team player and highly self-motivated.
- Flexible and versatile to adapt to any new environment and work on any project.
- Self-starter, fast learner and a team player with good communication skills and personality traits.
TECHNICAL SKILLS
Big Data Technologies: Hadoop, HDFS, MapReduce, Hive, Sqoop
Hadoop Distributions: Cloudera CDH4
Languages: Java, Scala, Python and SQL
Databases: Oracle, MySQL, Cassandra, MongoDB
Source control: GIT
App/Web servers: Tomcat
Operating Systems: Windows Family, Linux, CentOS
Frameworks: Spring, Hibernate, Spark, Slick, Spray
IDEs: Eclipse, Microsoft Visual Studio
Scripting Languages: HTML, VBScript
PROFESSIONAL EXPERIENCE
Confidential, Dallas TX
Hadoop/Scala Developer
Responsibilities:
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Implemented several Akka Actors which are responsible for loading of data into hive.
- Planning for production cluster hardware and software installation on production cluster and communicating with multiple teams to get it done.
- Implemented entire dao layer in the application using slick for inserting and retrieving metadata into oracle tables.
- Designed, configured and managed the backup and disaster recovery for HDFS data
- Developed a Folder Watcher mechanism to watch several input folders.
- Migrated data across clusters using DISTCP.
- Implemented a generic Housekeeping mechanism to control the retention of data in Hadoop and hive.
- Integration of JMS Queue, Publisher, Listener mechanisms.
Environment: Scala, Hadoop, Slick, Spray, Spark, Hive, Java, Ext Js
Confidential, Richmond VA
Hadoop/Java Developer
Responsibilities:
- Implemented entire validation framework which is used for validating various source files incoming into the application.
- Created and configured Sqoop jobs with incremental load to populate Hive External tables.
- Involved in writing code for several DAO’s in the application
- Implemented UDTF’s and UDF’s in the application to perform the necessary calculations and enrichment of data.
- Involved in creating Hive tables, loading with data and writing hive queries.
- Created generic frameworks for creating the UI screens, writing the hive loaders for loading mechanism.
- Used Spark to enhance the performance of the project.
- Assisted in creating and maintaining Technical documentation to launching HADOOP Clusters and even for executing Hive queries.
- Performed Unit testing for all the modules in the application.
Environment: Core java, Hadoop, Hive, Spark, Shark, Spring, Hibernate, MySQL, GXT
Confidential, New York NY
Hadoop/Scala Developer
Responsibilities:
- Implemented various roots for the application using spray.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Worked on the backend using Scala and Spark to perform several aggregation logics.
- Developed MapReduce(YARN) jobs for cleaning, accessing and validating the data.
- Written multiple MapReduce programs in java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.
- Performed Data Migration between Hadoop clusters using DISTCP.
- Worked with application teams (Onsite and Offshore) to install Hadoop updates, patches, version upgrades as required.
- Provided extensive support to L2 team regarding the environment setup, monitoring and solving of several issues.
- Used Spark to enhance the performance of the project.
Environment: Scala, Hadoop, Slick, Spray, Spark, Hive, Java.
Confidential, New York NY
Java/J2EE Developer
Responsibilities:
- Developed many user interface screens for GPC2 using GWT.
- Worked on core Java concepts like Collections and Exception Handling for writing the backend API's.
- Used Oracle Coherence API for querying and transforming data in parallel across the cluster.
- Used JDBC, with stored procedures to accomplish business logic implementation on server side.
- Worked in developing application using Spring Core, Spring Jdbc and DAO layers to offer abstraction for the business from the database related code (CRUD).
- Developed POJO’s, Data Access Object (DAO) which handles all database operations using HQL.
- Used Hibernate as ORM to map Java classes to database tables.
- Developed unit/integration test cases using JUnit and JMock.
- Involved in regular meetings with client and business analysts for requirements
Environment: Core java, Oracle Coherence, Spring, Hibernate, MySQL, GWT
Confidential
Java Developer
Responsibilities:
- Used Spring (MVC) Framework to develop the application. Used DAO and DTO Design patterns.
- Used Rational Rose to draw UML diagrams and to develop the Use cases, Domain model and Design Model.
- Implemented the functionalities using Java, J2EE, JSP, and AJAX, Servlets and spring framework.
- Used Hibernate for Object Relational Mapping.
- Developed Data Access Objects (DAOs) for easy data access.
- Involved in the development of the presentation layer in Ext JS using spring framework.
- Client pages are built using AJAX, HTML, JSP, XML and CSS.
- Applied J2EE design patterns like Business Delegate, Service Locator, Session Façade, DAO etc.
- Utilized JSP Standard Tag Libraries in the design and development of Views.
- Involved Database programming in Oracle 9i.
- Created the Stored Procedures using PL/SQL.
- Developed Web-Services module for integration using SOAP, WSDL.
Environment: Java, J2EE, JSP, MVC, Spring 2.0, DAO, Hibernate 2.8, JNDI, Tomcat, Eclipse, Oracle 9i, PL/SQL, EXT JS, JMS, HTML, XML