We provide IT Staff Augmentation Services!

Hadoop Developer Resume

HoustoN

SUMMARY

  • 9+ years of experience in Enterprise Application Development, Data Warehousing and Big Data technologies.
  • 5+ years of experience as Hadoop Developer and hands on experience in Hadoop Ecosystem.
  • Experience in performing Big Data Analytics using MapReduce for Apache Hadoop.
  • In depth knowledge of HDFS file system.
  • Writing custom data types, input and output formats.
  • Good understanding of differences in different releases of Apache Hadoop such as new MapReduce API, old MapReduce API, classic runtime and yarn runtime.
  • Exporting data to Relational Databases using SQOOP and vice versa.
  • Using flume to load weblog data into HDFS.
  • Analyzing data using pig scripts and hive queries. Writing custom UDF’s for analysis.
  • Chaining jobs and implement workflows using OOZIE.
  • Installing and Administering the Hadoop Cluster. Using Cloudera Manager.
  • Working knowledge of Hadoop federation and high - availability features in new release.
  • Knowledge of NoSql and hands on experience in Hbase and MongoDB.
  • Knowledge of Sql and hands on experience in mySQl and Microsoft SQL Server.
  • Performing analysis using high level languages such as Ruby and Python. Extending Apache pig functionality using Python.
  • Experience as a Java Developer in Web/intranet, client/server technologies using Java, J2EE, Servlets, JSP, JSF, EJB, JDBC and SQL.
  • Experience with Application Servers and Web Servers such as BEA Web Logic Server, JBoss Server, IBM Web Sphere and Apache Tomcat.
  • Good understanding of XML methodologies (XML,XSL,XSD) including Web Services and SOAP
  • Extensive experience in different IDE’s like Net Beans, Eclipse - Indigo/Galileo & Helios.
  • Good interpersonal and communication skills. Team player with strong problem solving skills.
  • Ability to learn and master new technologies and to deliver outputs in short deadlines with excellent communication and inter personnel skills.

TECHNICAL SKILLS

Hadoop/Big Data: HDFS, Mapreduce, HBase, Pig, Hive, Sqoop, Flume, MongoDB, HBase, Oozie, Zookeeper, spark, storm, & Kafka

Java & J2EE Technologies: Core Java

IDE’s: Eclipse, Net beans

Big data Analytics: Datameer 2.0.5

Frameworks: MVC, Struts, Hibernate, Spring

Programming languages: C, C++, Java, Python, Ruby, Ant scripts, Linux shell scripts

Databases: Oracle 11g/10g/9i, MySQL, DB2, MS-SQL Server

Web Servers: Web Logic, Web Sphere, Apache Tomcat

Web Technologies: HTML, XML, JavaScript, AJAX, SOAP, WSDL

Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP, FTP

ETL Tools: Informatica, Pentaho, SSRS, SSIS, BO, Crystal reports, Cognos.

Testing: Win Runner, Load Runner, QTP

WORK EXPERIENCE:

Confidential, Houston,

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Worked extensively with Flume for importing social media data
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager
  • Upgraded the Hadoop Cluster from CDH3 to CDH4, setting up High availability Cluster and integrating HIVE with existing applications
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs
  • Handled importing of data from various data sources using sqoop, performed transformations using Hive, MapReduce, loaded data into HDFS
  • Configured Sqoop and developed scripts to extract data from MySQL into HDFS
  • Hands-on experience with productionalizing Hadoop applications viz. administration, configuration management, monitoring, debugging and performance tuning
  • Created Hbase tables to store various data formats of PII data coming from different portfolios
  • Data processing using SPARK.
  • Cluster co-ordination services through ZooKeeper
  • Partitioning data streams using KAFKA.

Environment: Hadoop, HDFS, MapReduce, Pig, Hive, Sqoop, HBase, Oozie, Flume, Storm, java, Python, SQL Scripting, Storm, Cassandra, Ignite.

Confidential, Cincinnati OH

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, managing and reviewing data backups and Hadoop log files
  • Worked extensively with Flume for importing social media data
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs
  • Handled importing of data from various data sources using sqoop, performed transformations using Hive, MapReduce, loaded data into HDFS
  • Configured Sqoop and developed scripts to extract data from MySQL into HDFS
  • Hands-on experience with productionalizing Hadoop applications viz. administration, configuration management, monitoring, debugging and performance tuning
  • Created Hbase tables to store various data formats of PII data coming from different portfolios
  • Processing of streaming data using STORM.
  • Cluster co-ordination services through ZooKeeper

Environment: Hadoop, MapReduce, HDFS, Hive, Java, SQL, Cloudera Manager, Pig, Sqoop, Oozie, ZooKeeper, PL/SQL, MySQL, Windows, Oozie, HBase, STORM

Confidential, Seattle, WA

Hadoop developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Used Multithreading, synchronization, caching and memory management
  • Used JAVA, J2EE application development skills with Object Oriented Analysis and extensively involved throughout Software Development Life Cycle (SDLC)
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures
  • Extracted files from MongoDB through Sqoop and placed in HDFS and processed
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS
  • Load and transform large sets of structured, semi structured and unstructured data
  • Supported Map Reduce Programs those are running on the cluster
  • Wrote shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
  • Utilized Java and MySQL from day to day to debug and fix issues with client processes
  • Managed and reviewed log files
  • Implemented partitioning, dynamic partitions and buckets in HIVE

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, CouchDB, Flume, HTML, XML, SQL, MySQL J2EE, Eclipse

Confidential, Detroit, MI

Hadoop Developer

Responsibilities:

  • Responsible for complete SDLC management using different methodologies like Agile, Incremental, Waterfall, etc
  • Installed and configured Hadoop throughAmazon Web Services in cloud.
  • Developed MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Design dynamic Responsive User Interface screens using, HTML5, CSS3, JavaScript, Angular js, Bootstrap 3.1, media queries
  • Develop HTML5 interactive workflows
  • Responsible for Design Assets, UI Design Specifications and 508 Accessibility Guidelines
  • Perform Expert Review before UAT
  • Coordinate with offshore development team on the client implementations
  • Participate in daily scrum UX/UI meetings to track the progress and issues in the project. Responsible to manage data coming from different sources.
  • Supported Map Reduce Programs those are running on the cluster.
  • Jobs management using Fair scheduler.
  • Managed works including indexing data, tuning relevance, developing custom tokenizers and filters, adding functionality includes playlist, custom sorting and regionalization with Solr Search Engine.
  • Cluster coordination services through Zoo Keeper.
  • Involved in loading data from UNIX file system to HDFS
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.
  • Automated all the jobs for pulling data from FTP server to load data into Hive tables using Oozie workflows.
  • Data scrubbing and processing with Oozie.

Environment: Java, CDH 4, JMS, HDFS, Hadoop, Flume, MapReduce, Hive, PIG, Sqoop, Oozie, Teradata, Ganglia, Cassandra, python, Zookeeper, AWS

Confidential, McLean, VA

JAVA Developer

Responsibilities:

  • Developed the application using Struts Framework that leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used
  • Worked in an Agile work environment with Content Management system for workflow management and content versioning
  • Involved in designing user screens and validations using HTML, jQuery, Ext JS and JSP as per user requirements
  • Responsible for validation of Client interface JSP pages using Struts form validations
  • Integrating Struts with Spring IOC
  • Used Spring Dependency Injection properties to provide loose-coupling between layers
  • Implemented the Web Service client for the login authentication, credit reports and applicant information using Apache Axis 2 Web Service
  • Used Hibernate ORM framework with Spring framework for data persistence and transaction management
  • Used Hibernate 3.0 object relational data mapping framework to persist and retrieve the data from database
  • Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
  • Developed ANT Scripts to do compilation, packaging and deployment in the WebSphere server
  • Implemented the logging mechanism using Log4j framework
  • Wrote test cases in JUnit for unit testing of classes

Environment: JDK 1.5, J2EE 1.4,Agile Development Process, Struts 1.3, Spring 2.0, Web Services (JAX-WS, Axis 2) Hibernate 3.0, RSA, JMS, JSP, Servlets 2.5, WebSphere 6.1, SQL Server 2005, Windows XP, HTML, XML, IBM Rational Application Developer (RAD), ANT 1.6, Log4J, XML, XSLT, XSD, jQuery, JavaScript, Ext JS, JUnit 3.8, SVN

Confidential

JAVA Developer

Responsibilities:

  • Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
  • Reviewed the functional, design, source code and test specifications
  • Involved in developing the complete front end development using Java Script and CSS
  • Author for Functional, Design and Test Specifications
  • Implemented Backend, Configuration DAO, XML generation modules of DIS
  • Analyzed, designed and developed the component
  • Used JDBC for database access
  • Used Data Transfer Object (DTO) design patterns
  • Unit testing and rigorous integration testing of the whole application
  • Written and executed the Test Scripts using JUNIT
  • Actively involved in system testing
  • Developed XML parsing tool for regression testing
  • Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product

Environment: Java, JavaScript, HTML, CSS, JDK 1.5.1, JDBC, Oracle10g, XML, XSL, Solaris and UML

Hire Now