We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Pittsburg, PA

SUMMARY

  • Over 8 years of IT experience as a Developer, Designer & quality reviewer with cross platform integration experience using Hadoop, Java, J2EE and SOA.
  • 5 years of experience as Java Developer with hands - on in advanced J2EE technologies such as Spring, Struts, Hibernate, Servlets, JSP and JDBC.
  • Experience in all the phases of Data warehouse life cycle involving Requirement Analysis, Design, Coding, Testing, and Deployment.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Experience in analyzing data using HiveQL, PigLatin, and custom MapReduce programs in Java.
  • Experience in developing PIG Latin Scripts and using Hive Query Language.
  • Experience in installation, configuration and management of Hadoop Clusters.
  • Developed core modules in large cross-platform applications using JAVA, J2EE, Spring, Hibernate, and JAX-WS Web Services.
  • Extensive experience in architect the Hadoop cluster.
  • Practical knowledge on functionalities of every Hadoop daemon, interaction between them, resource utilizations and dynamic tuning to make cluster available and efficient.
  • Experience in understanding and managing Hadoop Log Files.
  • Experience in managing the Hadoop infrastructure with Cloudera Manager.
  • Experience in writing custom scripts to check the impact of any changes to cluster.
  • Experience in setting up data gathering tools such as Flume and Sqoop
  • Good understanding of NoSQL databases such as Hbase.
  • Experience in analyzing data in HDFS through MapReduce, Hive and Pig
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce concepts.
  • Experience on UNIX commands.
  • Extensively worked on the ETL mappings, analysis and documentation of OLAP reports requirements. Solid understanding of OLAP concepts and challenges, especially with large data sets.
  • Proficient in Oracle 9i/10g/11g, SQL and PL/SQL.
  • Experience in integration of various data sources like Oracle, DB2, Sybase, SQL server and MS access and non-relational sources like flat files into staging area.
  • Strong work ethic with desire to succeed and make significant contributions to the organization
  • Strong problem solving skills, good communication, interpersonal skills and a good team player
  • Have the motivation to take independent responsibility as well as ability to contribute and be a productive team member

TECHNICAL SKILLS

Hadoop/BigData Technologies: HDFS, Map Reduce, Hbase, Pig, Hive, Sqoop, Flume and Sqoop

Programming Languages: C, Java, SQL, PL/SQL, Shell Scripting

Frameworks: MVC, Spring, Struts, Hibernate

Web Technologies: HTML, XML, JavaScript, Ajax, SOAP and WSDL

Databases: Oracle 9i/10g/11g, SQL Server, MySQL

Database Tools: TOAD, Chordiant CRM tool, Kenan-Fx 2.0 Billing tool, Oracle Warehouse Builder (OWB).

Operating Systems: Linux, Unix, Windows, Mac, CentOS

Other Concepts: OOPS, Data Structures, Algorithms, Software Engineering, ETL

PROFESSIONAL EXPERIENCE

Confidential, Pittsburg, PA

Hadoop Developer

Responsibilities:

  • Developed framework to import the data from database to HDFS using sqoop.
  • Handled importing of data from various data sources, performed transformations using Hive,
  • MapReduce, loaded data into HDFS and extracted the data from MySQL into HDFS using Sqoop
  • Developed Pig Latin script for delta generation and filter relevant data from the web server output files to load into HDFS.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Worked on Accumulo, Pig, Kafka, Storm and UC4 workflow.
  • Installed UC4 workflow engine to run multiple Sqoop, Hive and Pig jobs
  • Experience with Agile - Scrum Model and Rally.
  • Worked on Teradata, Oracle and SQL server.
  • Played a major role in choosing the right configurations for Hadoop.
  • Continuous monitoring and managing the Hadoop cluster using Hortonworks.
  • Involved in start to end process of Hadoop cluster installation, configuration and monitoring.
  • Responsible for building scalable distributed data solutions using Hadoop
  • Setup and benchmarked Hadoop/HBase clusters for internal use
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
  • Used UDF's to implement business logic in Hadoop.
  • Coordinated with the QA team and participated in testing.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as Required

Confidential, Indianapolis, IN

Hadoop Developer

Responsibilities:

  • Developed MapReduce programs in Java for Data Analysis.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Developed HQL for the analysis of semi structured data.
  • Handle the installation and configuration of a Hadoop cluster.
  • Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive, and HBase.
  • Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.
  • Used Sqoop to import data into HDFS and Hive from other data systems.
  • Handle the data exchange between HDFS and different web sources using Flume and Sqoop
  • Close monitoring and analysis of the MapReduce job executions on cluster at task level.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks
  • Set up automated processes to analyze the System and Hadoop log files for predefined errors and send alerts to appropriate groups.
  • Commission or decommission the datanodes from cluster in case of problems.
  • Set up automated processes to archive/clean the unwanted data on the cluster, in particular on Namenode and Secondary namenode.
  • Set up and manage HA namenode and Namenode federation using Apache 2.0 to avoid single point of failures in large clusters.
  • Set up the checkpoints to gathering the system statistics for critical set ups.
  • Discussions with other technical teams on regular basis regarding upgrades, Process changes, any Special processing and feedback.

Confidential, Foster City, CA

Hadoop Developer/Admin

Responsibilities:

  • Installed and configured Hadoop and responsible for maintaining cluster and managing and reviewing Hadoop log files.
  • Developed MapReduce programs in Java for Data Analysis.
  • Load data from various data sources into HDFS using Flume.
  • Worked on Cloudera to analyze data present on top of HDFS.
  • Worked extensively on Hive and PIG.
  • Worked on large sets of structured, semi-structured and unstructured data.
  • Use of Sqoop to import and export data from HDFS to RDBMS and vice-versa.
  • Developed PIG Latin scripts to play with the data.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Good knowledge on reading data from Cassandra and also writing to it.
  • Participated in design and development of scalable and custom Hadoop solutions as per dynamic data needs.
  • Coordinated with technical team for production deployment of software applications for maintenance.
  • Provided operational support services relating to Hadoop infrastructure and application installation.
  • Handled the imports and exports of data onto HDFS using Flume and Sqoop.
  • Supported technical team members in management and review of Hadoop log files and data backups.
  • Formulated procedures for installation of Hadoop patches, updates and version upgrades.
  • Automated processes for troubleshooting, resolution and tuning of Hadoop clusters.
  • Set up automated processes to send alerts in case of predefined system and application level issues.
  • Set up automated processes to send notifications in case of any deviations from the predefined resource utilization.

Confidential, Atlanta, GA

Hadoop/Big Data Analyst

Responsibilities:

  • Installed and configured Hadoop and Hadoop stack on a 16 node cluster.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Involved in data ingestion into HDFS using Sqoop from variety of sources using the connectors like jdbc and import parameters.
  • Responsible for managing data from various sources and their metadata.
  • Worked with NoSQL database Hbase to create tables and store data.
  • Designed and implemented MapReduce-based large-scale parallel relation-learning system.
  • Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
  • Installed and configured Hive and also wrote Hive UDF’s that helped spot market trends.
  • Involved in loading data from UNIX file system to HDFS.
  • Implemented Fair schedulers on the Job tracker with appropriate parameters to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Involved in creating Hive tables, loading the data using it and in writing Hive queries to analyze the data.
  • Gained very good business knowledge on different category of products and designs within.

Environment: CDH4 with Hadoop 1.x, HDFS, Pig, Hive, Hbase, zookeeper, MapReduce, Java, Sqoop, Oozie, Linux, UNIX Shell Scripting and Big Data.

Confidential, Bloomfield, CT

Responsibilities:

  • Responsible for understanding the scope of the project and requirement gathering.
  • Review and analyze the design and implementation of software components/applications and outline the development process strategies
  • Coordinate with Project managers, Development and QA teams during the course of the project.
  • Used Spring JDBC to write some DAO classes to interact with the database to access account information.
  • Used Tomcat web server for development purpose.
  • Involved in creation of Test Cases for JUnit Testing.
  • Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
  • Used CVS, Perforce as configuration management tool for code versioning and release.
  • Developed application using Eclipse and used build and deploy tool as Maven.
  • Used Log4J to print the logging, debugging, warning, info on the server console.
  • Extensively used Core Java, Servlets, JSP and XML

Confidential, Dallas, TX

Java/J2EE Developer

Responsibilities:

  • Involved in requirements gathering and analysis from the existing system. Captured requirements using Use Cases and Sequence Diagrams.
  • Designed web screens using HTML & used Java script, AJAX.
  • Web tier is implemented using Spring MVC module
  • Used Spring IOC for dependency injection and Spring AOP for cross cutting concerns like logging, security, transaction management.
  • Integrated Spring JDBC for the persistence layer
  • Developed DAO Classes and written SQL for accessing Data from the database
  • Used XML for the data exchange and developed SOAP Web Services.
  • Deployment of the application into WebSphere Application Server.
  • Implemented Ant and Maven build tools to build jar and war files and deployed war files to target servers.
  • Implemented test cases with Junit.
  • Used RAD for developing and debugging the application
  • Utilized Rational Clear Case as a version control system and for code management
  • Coordinated with the QA team and participated in testing.
  • Involved in Bug Fixing of the application.

Confidential

Java Developer

Responsibilities:

  • Individually worked on all the stages of a Software Development Life Cycle (SDLC).
  • Used JavaScript code, HTML and CSS style declarations to enrich websites.
  • Implemented the application using SpringMVC Framework which is based on MVC design pattern.
  • Implemented the design patterns like DAO Factory, Singleton, Service Locator and Business Delegate
  • Developed application service components and configured beans using (applicationContext.xml) Spring IoC
  • Designed User Interface and the business logic for customer registration and maintenance.
  • Worked on Integrating the Customer Service software with the Website version.
  • Integrating Web services and working with data in different servers.
  • Involved in designing and Development of SOA services using Web Services.
  • Understanding the requirements from business users and end users.
  • Working with XML/XSLT files.
  • Experience creating UML class and sequence diagram.
  • Experience in Creating Tables, Views, Triggers, Indexes, Constraints and functions in SQL Server2005.
  • Worked on Integration of UI interface with SQL Server 2005 database.
  • Worked in content management for versioning and notifications.
  • Used SVN for version controlling.

We'd love your feedback!