We provide IT Staff Augmentation Services!

Hadoop Developer/ Administrator Resume

0/5 (Submit Your Rating)

Oklahoma, OK

SUMMARY

  • Over 7+ years of experience in IT and 4+ years of experience Hadoop ecosystem's implementation, maintenance, ETL and Big Data analysis operations.
  • Excellent understanding / knowledge of BigData and Hadoop ecosystems.
  • Experience with configuration of Hadoop ecosystem components such as MapReduce, HDFS, HIVE, HBase, PIG, Sqoop, Zookeeper, Oozie and Flume.
  • Experienced in using NFS (network file systems) for Name node metadata backup.
  • In depth knowledge of Hadoop architecture and various components such as HDFS, JobTracker, NameNode, DataNode and MapReduce concepts.
  • Good Exposure to Map Reduce programming, PIG Scripting and Distributed Application and NoSQL Databases like HBase, Mongo DB.
  • Experience in maintaining multiple Hadoop clusters (prod, dev etc.).
  • Experience in performance tuning of the Hadoop Cluster by gathering and analyzing the existing infrastructure.
  • Experience in automating the Hadoop installation, configuration and maintaining the cluster by using the tools like puppet.
  • Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
  • Knowledge on designing both time driven and data driven automated workflows using Oozie.
  • Excellent analytical, multi - tasking, problem-solving, time-management & communication skills with particular emphasis on clearly communicating & documenting detailed Test Requirements & Tests.
  • Experience in support analysts by administering and configuring Hive.
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Developed MapReduce programs to perform analysis.
  • Performed importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in writing shell scripts to dump the Shared Data from MySQL server to HDFS.
  • Experience in Core Java, Java Virtual Machine (JVM) and multi-thread processing.
  • Over 3 years of extensive experience as a Netcool Management Service (NMS) Consultant, involved in developing architectural solutions for customers in the areas of Fault Management, Performance Management, Operational Support Systems (OSS) and other related disciplines of included IBM Netcool.
  • Responsible for monitoring hadoop cluster using ambari by configuring various alerts for troubleshooting, setting parameters for enhancing performance.
  • Experience in Web based portal system such using Edge tool (JARVIS) that displays and consolidates web enabled network management applications into a single view.
  • Experience in Unix Shell Scripting, SQL, Reporting and validating complex Stored Procedures, Triggers.
  • Worked on UNIX platform and got a clear understanding of the VI editor and UNIX commands.
  • Excellent communication and inter-personal skills, self-motivated and detail oriented.

TECHNICAL SKILLS

Operating Systems: Linux, Windows XP, Server 2003, Server 2008

Databases: Oracle, MySQL, MS SQL Server 2000, DB2

Hadoop Ecosystem Development: HDFS, HBase, MapReduce, YARN, Hive, Pig, Sqoop, Flume, Oozie, Zookeeper

Fault Management Tool: IBM Tivoli suite of products (OMNIbus, ITNM, Web-GUI, TBSM, TCR, Impact, ITM

Languages: C, Java, SQL, Pig Latin, Unix Shell Scripting

IDE/ Tools: Eclipse

Others: JARVIS, HP-BSM

PROFESSIONAL EXPERIENCE

Confidential, PA

Hadoop Developer

Responsibilities:

  • Analyzing the requirement to setup a cluster.
  • Installed and configured Hadoop, MapReduce, HDFS (Hadoop Distributed File System), developed multiple MapReduce jobs in java.
  • Worked with the infrastructure and admin team in designing, modelling, sizing and configuring Hadoop cluster of 15 nodes.
  • Developed Map Reduce programs in Java for parsing the raw data and populating staging Tables.
  • Created Hive queries to compare the raw data with EDW reference tables and performing aggregates
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced in analyzing data with Hive and Pig.
  • Writing Pig scripts to process the data.
  • Developed PIG Latin scripts to extract the data from the web server output files to load into HDFS.
  • Integrating bulk data into Cassandra file system using MapReduce programs.
  • Got good experience with NOSQL database.
  • Involved in HBASE setup and storing data into HBASE, which will be used for further analysis.
  • Experienced in managing and reviewing Hadoop log files.
  • Experienced in defining job flows.
  • Experienced in managing and reviewing Hadoop log files.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries using the HiveQL which will run internally in map reduce way.
  • Extracted the data from MySQL into HDFS using Sqoop.
  • Used HiveQL to analyse the partitioned and bucketed data and compute various metrics for reporting.
  • Deployed Hadoop Cluster in Fully Distributed and Pseudo-distributed modes.
  • Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig, Hive and Sqoop.
  • Unit tested a sample of raw data and improved performance and turned over to production.

Environment: Java (JDK1.7), Java 7, Hadoop 2.6.0, MapReduce, HDFS, Hive 0.13.0, Sqoop 1.4.4, HBase, Pig 0.12.0, Oozie, Kerberos, Linux, Shell Scripting, Oracle 11g, PL/SQL, SQL*PLUS, HDInsight

Confidential

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing.
  • Installed and Configured Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase and Sqoop.
  • Good understanding and related experience with Hadoop stack - internals, Hive, Pig and Map/Reduce.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Involved in defining job flows.
  • Involved in database connection by using SQOOP
  • Involved in creating Hive tables, loading with data and writing hive queries using the HiveQL.
  • Experienced in SQL DB Migration to HDFS
  • Got good experience with NOSQL database.
  • Used Oozie to orchestrate the workflow.
  • Installed and configured Pig for ETL jobs.
  • Used big data technologies with hands on experience on Hadoop Distributed File System.
  • Worked on Map Reduce programming and Pig latin scripts.
  • Experience with Core Distributed computing and Data Mining Library using Apache Spark.
  • Integrating bulk data into Cassandra file system using MapReduce programs.
  • Involved in creating data-models for customer data using Cassandra Query Language.
  • The data is collected from distributed sources into Avro models. Applied transformations and standardizations and loaded into HBase for further data processing.

Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Hive, Python, Pig, Sqoop, Cloudera, Cassandra, Spark, Shell Scripting, Java (JDK 1.6), Java 6, Oracle 10g, PL/SQL, SQL*PLUS

Confidential, Oklahoma, OK

Hadoop Developer/ Administrator

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter experts.
  • Involved in installing Hadoop ecosystem components.
  • Responsible to manage data coming from different sources.
  • Monitoring the jobs to analyze performance statistics.
  • Performing Unit Testing of completed jobs.
  • Involved in database connection by using SQOOP.
  • Applying optimization techniques at both Hadoop and Database level.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries.
  • Developed Simple to complex MapReduce Jobs using Hive and Pig.
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required .
  • Experienced in defining job flows.
  • Experienced in managing and reviewing Hadoop log files.
  • Extracted files from MongoDB through Sqoop and placed in HDFS and processed.
  • Experienced in running Hadoop streaming jobs to process terabytes of XML format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Automation script to monitor HDFS and HBase through cronjobs.
  • Plan, design, and implement processing massive amounts of marketing information, complete with information enrichment, text analytics, and natural language processing.
  • Prepare multi-cluster test harness to exercise the system for performance and failover.

Environment: HadoopHDFS, Map Reduce, Hive, Pig, Sqoop, Oozie, HBase, Linux, Java, Xml, MongoDB.

Confidential, CA

Java Developer

Responsibilities:

  • Actively participated in the Analysis, Design, Development and Testing phases of various modules.
  • Scrupulously involved in Agile Standup meetings throughout the project with team members on day to day basis for the update on project status using Rally Software.
  • Estimated Level1 and Level2 and Designed Class, Sequence Diagrams.
  • Worked on J2EEdesign patternslikeDAO, Front Controller, Session Façade, Factory Pattern and Singleton.
  • Created mobile web applications using HTML5, CSS3, JSON, and JavaScript.
  • Worked on OOPS concepts, Collections Framework using Core Java.
  • Extensively used Spring MVC architecture for the application development.
  • Developed application service components and configured beans using Spring IOC .
  • Used JMS for Asynchronous communication by consuming messages in the Message queue.
  • Used Hibernate as data persistence layer and configured service and persistence object using Hibernate mapping files.
  • Involved in frameworks integration of Spring and Hibernate.
  • Produced and Consumed REST based webservices, did unit testing using SOAPUI tool and generated json.
  • Worked on Database queries using Oracle instance.
  • Used JAXB to convert XML content into a Java Object.
  • Involved in unit testing, test case reviews and code reviews in accordance with applicable development standards.
  • Involved in integration system testing and user acceptance testing (UAT)
  • Worked on Web deployments in DEV, QA and UAT environments.
  • Support the Application whenever encountered Production issues.
  • Deployed application in Weblogic Application server.
  • SVN is used as Code Versioned control.
  • Firebug is used as debugger.
  • Used Maven script to build application.

Environment: Core Java1.6, Oracle10g, Toad, JSP, Spring3.0, Hibernate4.0, JavaScript, HTML5, jQuery, Ajax, Apache CXF Framework, REST Webservices, JAXB, Linux, Shell scripting, WinSCP, log4J, Maven, SVN, Weblogic Server11g, Eclipse, Microsoft Visio.

Confidential

Java Developer

Responsibilities:

  • Participates in the fullSDLC including collaborative review meetings, estimations, design sessions, development and testing.
  • Developed the User Interface using JSP/AJAX/ HTML / CSS/ Java Script.
  • Designed and developedUse Cases, Class diagramsandSequence Diagram.
  • Implementedvarious J2EEdesign patternslikeDAO, DTO, Factory Pattern and Singleton.
  • Developed the application using Java Server Faces (JSF) and Spring MVC.
  • Implement GET and POST operations using Restful web services.
  • Totally involved in building components using DOM and JaxB which helps to parse the XML.
  • Written JDBC programs for callable statements, prepared statements written in Oracle andused TOAD toolto interact with database.
  • UsedCVSfor version controlling.
  • Used log4j for logging statements.

Environment: Core Java 1.5, JSF1.1, Spring 2.0, Restful webservices, Rest Client, XML, XSD, JAXB, DOM, Tomcat, JSP, Oracle 10g, SQL/PL SQL, Hibernate2.0, Java script, CSS, HTML, Ajax, log4j, My Eclipse, Toad, Ant1.7, Microsoft Visio.

We'd love your feedback!