We provide IT Staff Augmentation Services!

Hadoop Team Lead / Architect Resume

5.00/5 (Submit Your Rating)

El Segundo, CA

PROFESSIONAL SUMMARY:

  • Over 8+ years of experience in software development, 3+ years of experience in Development of big data Hadoop Framework.
  • Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
  • Experience with Data Torrent.
  • Experienced on major Hadoop ecosystem’s projects such as PIG, HIVE and HBASE.
  • Good working experience using Sqoop to import data into HDFS from RDBMS and vice - versa
  • Good knowledge in using job scheduling and monitoring tools like Oozie and ZooKeeper
  • Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera
  • Experience in developing solutions to analyze large data sets efficiently
  • Knowledge of administrative tasks such as installing Hadoop and its ecosystem components such as Hive and Pig
  • Knowledge of NoSQL databases such as HBase, and MongoDB
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce concepts
  • Extending Hive and Pig core functionality by writing customUDFs
  • Good understanding of Data Mining and Machine Learning techniques
  • Experience in analyzing data using HiveQL, Pig Latin, and custom MapReduce programs in Java.
  • Strong work ethic with desire to succeed and make significant contributions to the organization
  • Strong problem solving skills, good communication, interpersonal skills and a good team player
  • Have the motivation to take independent responsibility as well as ability to contribute and be a productive team member

TECHNICAL SKILLS:

Hadoop/Big Data Technologies: HDFS, MapReduce, Hive, Pig, Sqoop, Flume, Hbase, Oozie, Zookeeper, Kerberos, Data torrent.

Programming Languages: Java JDK1.4/1.5/1.6 (JDK 5/JDK 6), C/C++, HTML, SQL, PL/SQL, AVS & JVS

Frameworks: Hibernate 2.x/3.x, Spring 2.x/3.x,Struts 1.x/2.x

Web Services: WSDL, SOAP, Apache CXF/XFire, Apache Axis, REST, Jersey

Client Technologies: JQUERY, Java Script, AJAX, CSS, HTML 5, XHTML

Operating Systems: UNIX, Windows, LINUX, Redhat, Ubuntu, XP

Application Servers: IBM Web sphere, Tomcat, Web Logic, Web Sphere

Web technologies: JSP, Servlets, JNDI, JDBC, Java Beans, JavaScript, Web Services(JAX-WS)

Databases: Oracle 8i/9i/10g & MySQL 4.x/5.x, SQL server Teradata

Java IDE: Eclipse 3.x, IBM Web Sphere Application Developer, IBM RAD 7.0

Tools: TOAD, SQL Developer, SOAP UI, ANT, Maven, Visio, Rational Rose, Endur 8.x/10.x/11.x, Data Torrent

WORK EXPERIENCE:

Confidential, El Segundo, CA

Hadoop Team Lead / Architect

Responsibilities:

  • Involved in Architecture of the Proof of Concept Involved in initial meetings with Cloudera Architect and BI teams for requirement gathering.
  • Created Hive external tables for append only tables and managed tables for Reload tables.
  • Installed various Hadoop ecosystems and Hadoop Daemons.
  • Managed commissioning & decommissioning of datanodes.
  • Balanced cluster after adding/removing nodes or major data cleanup.
  • Compiled new kernels, created and modified scripts (mainly bash) to accommodate the administration of daily duties.
  • Configured Fair scheduler to ensure proper resources usage of the Cluster for Map Reduce jobs submitted by the users.
  • Administered users login account, quotas, acl.
  • Implemented NameNode backup using NFS for High availability
  • Worked collaboratively to manage build outs of large data clusters and involved n hadoop hardware clustering and Pivotal Hadoop.
  • Helped design big data clusters and administered them.
  • Worked both independently and as an integral part of the development team.
  • Communicated all issues and participated in weekly strategy meetings.
  • Administered back end services and databases in the virtual environment.
  • Implemented system wide monitoring and alerts.
  • Worked on Append, Insert and Updates tables using Hive windowing functions.
  • Worked on Spark and proved that run the programs run faster than Map reduce.
  • Written map reduce programs within hive for loading data from temp tables to external tables.
  • Extensively worked on java the operators of data torrent live streamed data to HDFS.
  • Working in POC project to improve performance using Data Torrent on Hadoop YARN 2.0
  • Worked on Sqoop Importing JDBC to HDFS. Worked in Agile environment and used pivotal tracker for tracking stories.
  • Documented Low Level Design Document, Migration Document and updated existing Interface document with changes made.
  • Participated in weekly status meetings, and conducting internal and external reviews as well as formal walkthroughs among various teams, and documenting the proceedings.

Environment: Hadoop 2.0, Hive, Sqoop, Hue, Data Torrent, Rapid Miner, Core Java, Cloudera Manager, Oracle, MySQL, UNIX, Oozie, Cloudera Distribution

Confidential, San Francisco, CA

Hadoop Admin/ Developer

Responsibilities:

  • Involved in writing MapReduce programs and tested using MRUnit.
  • Managing and scheduling Jobs on a Hadoop cluster using Oozie.
  • Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
  • Involved in loading data from UNIX file system to HDFS.
  • Worked on Hue interface for querying the data.
  • Created Hive tables to store the processed results in a tabular format.
  • Created HBase tables to store variable data formats of data coming from different portfolios.
  • Involved in transforming data from Mainframe tables to HDFS, and HBASE tables using Sqoop and Pentaho Kettle.
  • Implemented best income logic using Pig scripts.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
  • Have deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Hue, Oozie, Core Java, Eclipse, Hbase, Flume, Cloudera Manager, Oracle 10g, DB2, IDMS, VSAM, SQL*PLUS, Toad, Putty, Windows NT, UNIX Shell Scripting, Pentaho Kettle, Pentaho Bigdata, YARN

Confidential, Boston, MA

Big Data/ Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced indefining jobflowson Jive Software.
  • Experienced in managing andreviewingHadooplog files.
  • Extracted files from CouchDB through Sqoop and placed in HDFS and processed.
  • Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Got good experience with NoSQL database.
  • Supported Map Reduce Programs those are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.

Environment: Java 6, Eclipse, Linux, Hadoop, HBase, Sqoop, Pig, Hive, Flume.

Confidential, MD

Java/J2EE Developer

Responsibilities:

  • Responsible for gathering and analyzing requirements and converting them into technical specifications
  • Used Rational Rose for creating sequence and class diagrams
  • Developed presentation layer using JSP, Java, HTML and JavaScript
  • Used Spring Core Annotations for Dependency Injection
  • Designed and developed a ‘Convention Based Coding’ utilizing Hibernate’s persistence framework and O-R mapping capability to enable dynamic fetching and displaying of various table data with JSF tag libraries
  • Designed and developed Hibernate configuration and session-per-request design pattern for making database connectivity and accessing the session for database transactions respectively. Used HQL and SQL for fetching and storing data in databases
  • Participated in the design and development of database schema and Entity-Relationship diagrams of the backend Oracle database tables for the application
  • Implemented web services with Apache Axis
  • Designed and Developed Stored Procedures, Triggers in Oracle to cater the needs for the entire application. Developed complex SQL queries for extracting data from the database
  • Designed and built SOAP web service interfaces implemented in Java
  • Used Apache Ant for the build process
  • Used ClearCase for version control and ClearQuest for bug tracking

Environment: Java, JDK 1.5, Servlets, Hibernate, Ajax, Oracle 10g, Eclipse, Apache Ant, Web Services (SOAP), Apache Axis, Apache Ant, Web Logic Server, JavaScript, HTML, CSS, XML

Confidential, IL

Java/ J2EE Developer

Responsibilities:

  • Involved in creation of UML diagrams like Class, Activity, and Sequence Diagrams using modeling tools of IBM Rational Rose
  • Involved in the development of JSPs and Servlets for different User Interfaces
  • Used Struts action forms and developed Action Classes, which act as the navigation controller in Struts framework
  • Implemented the template-based categorization of presentation content using Struts-Tiles. MVC implementation using Struts framework
  • Involved in Unit Testing of Various Modules based on the Test Cases
  • Involved in Bug fixing of various modules that were raised by the Testing teams in the application during the Integration testing phase
  • Involved and participated in Code reviews
  • Used Log4J logging framework for logging messages
  • Used Rational ClearCase for version control
  • Used Rational Clear Quest for bug tracking
  • Involved in deployment of application on IBM Websphere Application Server

Environment: Java, J2EE, XML, XML Schemas, JSP, HTML, CSS, PL/SQL, Junit, Log4j, IBM Web sphere Application Server.

Confidential, NC

JAVA/ J2EE Application Developer

Responsibilities:

  • Responsible for design, document, implement, and unit test robust webservices framework to support templated payments and group payments in Java.
  • Worked in an onsite-offshore model. Lead offshore developers, assisted them in understanding requirements and provided code reviews.
  • Extensively used Spring Framework for Business Layer - accessing enterprise services like JNDI, JMS, and Job Scheduling. Also used spring for transaction management and dependency injection.
  • Created Database model, domain objects and DAO’s that interact with the database and store the template related data and events.
  • Used JAX-WS and spring webservices to create and consume SOAP based webservices.
  • Used Hibernate as an ORM tool for database operations.
  • Created MDB’s to consume messages on various user events.
  • Worked in Agile software methodology with Scrum type development.

Environment: Java 1.6, Spring, Hibernate, Maven, Apache MQ, JUnit, JAXB, Oracle 10g, Oracle Coherence, Log4j, Shell Script, Soap UI, TOAD, SQL Developer, Quality Center, Linux, Windows

Confidential, OH

Junior JAVA Developer

Responsibilities:

  • Responsible for understanding the business requirement.
  • Worked with Business Analyst and helped representing the business domain details in technical specifications.
  • Was also actively involved in setting coding standards and writing related documentation.
  • Developed the Java Code using Eclipse as IDE.
  • Developed JSPs and Servlets to dynamically generate HTML and display the data to the client side
  • Developed application on Struts MVC architecture utilizing Action Classes, Action Forms and validations.
  • Tiles were used as an implementation of Composite View pattern
  • Was responsible in implementing various J2EE Design Patterns like Service Locator, Business Delegate, Session Facade and Factory Pattern.
  • Code Review & Debugging using Eclipse Debugger.
  • Was responsible for developing and deploying the EJB (Session & MDB).
  • Configured Queues in Web Logic server where the messages, using JMS API, were published.
  • Consumed Web Services (WSDL, SOAP, and UDDI) from third party for authorizing payments to/from customers.
  • Writing/Manipulating the database queries.
  • Build web application using MAVEN as build tool.
  • Used CVS for Version control,
  • Responsible for understanding the business requirement.
  • Performed unit testing using JUnit Testing Framework and Log4J to monitor the error log.

Environment: Java,/J2EE, Eclipse, Web Logic Application Server, Oracle, JSP, HTML, JavaScript, JMS, Servlets, UML, XML, Eclipse, Struts, Web Services, WSDL, SOAP, UDDI.

We'd love your feedback!