We provide IT Staff Augmentation Services!

Hadoop Developer/admin Resume

2.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Over Seven (7+) years of referable experience in Systems Administration and Enterprise Application Development in diverse industries which includes hands on experience in Big data ecosystem related technologies
  • Including 2.5 years of comprehensive experience as a Hadoop, Big Data & Analytics Developer and Admin
  • Experienced in processing Big data on teh Hadoop framework using MapReduce programs
  • Hands on experience wif Big Data application testing
  • Experienced in installation, configuration, testing and monitoring Hadoop clusters.
  • Experience using Cloudera distributions including CDH4 and CDH5, Hortonworks and Amazon Web Services(AWS)
  • Experienced in using Pig, Hive, Sqoop, Oozie, Tivoli, ZooKeeper, HBase, Impala and Cloudera Manager
  • Imported and exported data using Sqoop from HDFS to RDBMS
  • Application development using Java, RDBMS, and Linux shell scripting
  • Extended Hive and Pig core functionality by writing custom UDFs
  • Experienced in analyzing data using HiveQL, Pig Latin, and custom Map Reduce programs in Java
  • Worked on Hortonworks on teh development cluster
  • Familiar wif systems like GIT, SVN
  • Familiar wif Java virtual machine (JVM) and multi - threaded processing
  • Worked on NoSQL databases including HBase, Cassandra and MongoDB
  • Worked on Hadoop cluster running on MapReduce.
  • Experienced in job workflow scheduling and monitoring tools like Oozie and Zookeeper
  • Very good understanding of Data Modelling concepts
  • Experienced in designing, developing and implementing connectivity products dat allow efficient exchange of data between teh core database engine and teh Hadoop ecosystem
  • Experienced in Data warehousing and using ETL tools like Informatica, Pentaho and Talend
  • Expert level skills in developing intranet/internet application using JAVA/J2EE technologies which includes Struts framework, MVC design Patterns, Servlets, JSP, JSLT, XML/XLST, Java Script, AJAX, EJB, JDBC, JMS, JNDI, RDMS, SOAP, Hibernate and custom tag Libraries
  • Experience using XML, XSD and XSLT
  • Experienced in creating PoC’s for new data sets
  • Used Falcon for defining data pipelines to auto-generate workflows and Hue User Administration
  • Experience wif web-based UI development using jQuery UI, jQuery, ExtJS, CSS, HTML, HTML5, XHTML and JavaScript
  • Worked on scheduling for maximizing CPU time utilization and performing backup and restore if different components
  • Extensive experience in middle-tier development using J2EE technologies like JDBC, JNDI, JSP, Servlets, JSP, JSF, Struts, Spring, Hibernate, JDBC, EJB
  • Created Shell scripts for running Datameer jobs
  • Possess excellent technical skills, consistently outperformed schedules and acquired interpersonal and communication skills.

TECHNICAL SKILLS

Hadoop/Big Data: HDFS, MapReduce, HBase, Pig, Hive, Sqoop, Flume, Kafka, MongoDB, Cassandra, Yarn, Puppet, Spark, Oozie, Zookeeper

Java & J2EE Technologies: Core Java, Servlets, JSP, JDBC, JNDI, Java Beans

IDE’s: Eclipse, Net beans

Big data Analytics: Datameer 2.0.5

Frameworks: MVC, Struts, Hibernate, Spring

Programming languages: C, C++, Java, Python, Ant scripts, Linux shell scripts

Databases: Oracle 11g/10g/9i, MySQL, DB2, MS-SQL Server

Web Servers: Web Logic, Web Sphere, Apache Tomcat

Web Technologies: HTML, XML, JavaScript, AJAX, SOAP, REST, WSDL

Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP

ETL Tools: Informatica, Pentaho, Teradata

Testing: Win Runner, Load Runner, QTP

PROFESSIONAL EXPERIENCE

Confidential, Atlanta, GA

Hadoop Developer/Admin

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, managing and reviewing data backups and Hadoop log files
  • Continuous monitoring and managing teh Hadoop cluster through Cloudera Manager
  • Upgrading teh Hadoop Cluster from CDH3 to CDH4, setting up High availability Cluster and integrating HIVE wif existing applications
  • Performed Big Data application testing
  • Analyzed teh data by performing Hive queries and running Pig scripts to know user behavior
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs
  • Handled importing of data from various data sources, performed transformations using Hive, SerDe, MapReduce, loaded data into HDFS and extracted data from Teradata into HDFS using Sqoop
  • Worked extensively wif Sqoop for importing metadata from Oracle and migrating teh ETL jobs into HDFS performing aggregations.
  • Configured Sqoop and developed scripts to extract data from MySQL into HDFS
  • Hands-on experience wif productionalizing Hadoop applications viz. administration. configuration management, monitoring, debugging and performance tuning
  • Created HBase tables to store various data formats of PII data coming from different portfolios
  • Cluster co-ordination services through ZooKeeper
  • Created Hive queries using SerDe, performing data analysis and improving performance using tuning parameters
  • Experienced in working on Python for Hadoop Streaming and Python for MapReduce
  • Trained and mentored analyst and test team on Hadoop framework, HDFS, Map Reduce concepts, Sharding and Hadoop Ecosystem
  • Experienced working on Network logs and Syslogs and performing integrations
  • Worked on documenting DataGuise for internal users and disaster recovery
  • Responsible for architecting Hadoop clusters and installing OS, performing Hadoop updates, patches and Hadoop version upgrades.
  • Assist wif teh addition of Hadoop processing to teh IT infrastructure
  • Perform data analysis using Hive and Pig and managing jobs

Environment: Hadoop, MapReduce, Yarn, HDFS, Hive, Java, SQL, Cloudera Manager, CDH4, CDH5, Pig, Sqoop, Hortonworks, Oozie, ZooKeeper, Teradata, PL/SQL, Python, MySQL, Impala, Cassandra, HBase

Confidential, Schaumburg, IL

Hadoop & Java Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup and disaster recovery systems and procedures
  • Extracted files from CouchDB, MongoDB through Sqoop and placed in HDFS for processed
  • Used Flume to collect, aggregate, and store teh web log data from different sources like web servers, mobile and network devices and pushed to HDFS
  • Developed Puppet scripts to install Hive, Sqoop, etc. on teh nodes
  • Data back up and synchronization using Amazon Web Services
  • Worked on Amazon Web Services as teh primary cloud platform
  • Load and transform large sets of structured, semi structured and unstructured data
  • Supported Map Reduce Programs those are running on teh cluster
  • Load log data into HDFS using Flume, Kafka and performing ETL integrations
  • Designed and implemented DR and OR procedures
  • Wrote shell scripts to monitor teh health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
  • Utilized Java and MySQL from day to day to debug and fix issues wif client processes
  • Used JAVA, J2EE application development skills wif Object Oriented Analysis and extensively involved throughout Software Development Life Cycle (SDLC)
  • Hands-on experience of Sun One Application Server, Web logic Application Server, Web Sphere Application Server, Web Sphere Portal Server, and J2EE application deployment technology
  • Monitoring Hadoop cluster using tools like Nagios, Ganglia, Ambari and Cloudera Manager
  • Automation script to monitor HDFS and HBase through Cron jobs
  • Used MRUnit for debugging MapReduce dat uses sequence files containing key value pairs.
  • Develop high-performance cache, making teh site stable and improving its performance
  • Create a complete processing engine, based on Cloudera's distribution
  • Proficient wif SQL languages and good understanding of Informatica and Talend
  • Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster

Environment: Hadoop, MapReduce, HDFS, Hive, Apache Spark, Kafka, CouchDB, Flume, AWS, Cassandra, Oracle 11g, Java, Struts, Servlets, HTML, XML, SQL, J2EE, MRUnit, Informatica, JUnit, Tomcat 6. Java, JDBC, JNDI, Struts, Maven, SQL language, Oracle, XML, Eclipse

Confidential, Warren, NJ

Sr. Java Developer

Responsibilities:

  • Developed teh application using Struts Framework dat leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams (sequence and collaboration) and activity diagrams were used
  • Worked in an Agile work environment wif Content Management system for workflow management and content versioning
  • Involved in designing user screens and validations using HTML, jQuery, Ext JS and JSP as per user requirements
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process, etc.
  • Familiar wif SQL Server Integration SSIS and SSRS
  • Responsible for validation of Client interface JSP pages using Struts form validations
  • Integrating Struts wif Spring IOC
  • Used Spring Dependency Injection properties to provide loose-coupling between layers
  • Implemented teh Web Service client for teh login authentication, credit reports and applicant information using Apache Axis 2 Web Service
  • UsedHibernateORM framework wifSpringframework for data persistence and transaction management
  • Used Hibernate 3.0 object relational data mapping framework to persist and retrieve teh data from database
  • Created physical and logical data model, design and implementation
  • Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
  • Developed ANT Scripts to do compilation, packaging and deployment in teh WebSphere server
  • Implemented teh logging mechanism using Log4j framework
  • Designing and implementing algorithms
  • Wrote test cases in JUnit for unit testing of classes

Environment: JDK 1.5, J2EE 1.4, Agile Development Process, Struts 1.3, Spring 2.0, Web Services (JAX-WS, Axis 2) Hibernate 3.0, RSA, JMS, JSP, Servlets 2.5, WebSphere 6.1, SQL Server 2005, DB2, Windows XP, HTML, XML, IBM Rational Application Developer (RAD), ANT 1.6, Log4J, XML, XSLT, XSD, jQuery, JavaScript, Ext JS, JSON, JUnit 3.8, SVN

Confidential

Java Developer

Responsibilities:

  • Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
  • Reviewed teh functional, design, source code and test specifications
  • Involved in developing teh complete front end development using Java Script and CSS
  • Author for Functional, Design and Test Specifications
  • Implemented Backend, Configuration DAO, XML generation modules of DIS
  • Analyzed, designed and developed teh component
  • Used JDBC for database access
  • Used Data Transfer Object (DTO) design patterns
  • Unit testing and rigorous integration testing of teh whole application
  • Written and executed teh Test Scripts using JUNIT
  • Actively involved in system testing
  • Developed XML parsing tool for regression testing
  • Prepared teh Installation, Customer guide and Configuration document which were delivered to teh customer along wif teh product

Environment: Java, JavaScript, HTML, CSS, JDK 1.5.1, JDBC, Oracle10g, XML, XSL, Solaris and UML

We'd love your feedback!