Senior Hadoop Admin Resume
Cleveland, OH
SUMMARY
- 7+ years of experience in Admin, Analysis, Design, Development, Testing, Implementation, Maintenance and Enhancements on various IT Projects. Around 3+ Years of experience in Big Data in implementing complete Hadoop solutions.
- Working experience in Administration, designing and implementing Big Data projects using Hadoop ecosystem components like Hadoop Distributed File System (HDFS), Map Reduce, PIG, HIVE, Zookeeper, Sqoop, Kafka, Storm Hue, Maven and JSON.
- Expertise in installing, configuring and using Hadoop components like Hadoop MapReduce, HDFS, HBase, Hive, Oozie, Pig, Sqoop, Yarn, Flume and Zookeeper.
- Experience in installing Cloudera Manager and extensively can use Cloudera Manager to create a Hadoop Cluster, monitor and upgrade the version of the cluster.
- Good Experience in writing UDFs, PIG scripts and Hive Queries for processing and analyzing large volumes of data.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems(RDBMS) and vice - versa
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
- Worked on Cascading API for Hadoop application development and work flows
- Experience in analyzing data using Hive QL, Pig Latin, and custom MapReduce programs in Java.
- Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper
- Experience in optimization of MapReduce algorithm using combiners and partitioners to deliver the best results.
- Strong command and experience over ETL tool Informatica 8/9.
- Good exposure on usage of NoSQL databases like HBase, Cassandra, MongoDB.
- Good understanding of Data Mining and Machine Learning techniques
- Expertise in core Java, J2EE, Multithreading, JDBC,Web Services Shell Scripting and proficient6 in using Java API’s for application development
- Solid background in Core Java concepts like Threads, Collections Framework, Java Reflection and also have handsome experience in building Class diagrams, activity diagrams, sequence diagrams, activity diagrams and flow charts using Rational Rose and Visio.
- Strong database connectivity skills which includes Oracle, MYSQL, and DB2 and in programming with SQL, PL/SQL, and Stored Procedures, Triggers, Functions and Packages besides writing DDL, DML and Transaction queries with development tools like sql developer.
- Strong hands-on knowledge with DW platforms and databases like MS SQL Servers 2012 and 2008, Oracle 11g/10g/9i, MySQL, DB2 and Teradata.
- Good proficiency with modeling tools like SAS, R.
- Experience in working with Python and Hadoop Streaming Command options
- Proficient in Working with Various IDE tools including Eclipse Galileo, IBMRational Application Developer (RAD) and IntelliJ IDEA.
- Worked on different operating systems like UNIX/Linux, Windows XP and Windows 2K
- Excellent working experience in Scrum / Agile framework and Waterfall project execution methodologies.
- Very good experience in customer specification study, requirements gathering, system architectural design and turning the requirements into final product.
- Excellent Analytical ability, Strong technical background along with good debugging and communication skills.
- Quick learner and adaptive to new and challenging technological environments and highly goal oriented.
TECHNICAL SKILLS
Hadoop Eco Systems: HDFS, Map Reduce, Apache Crunch, Hive, Pig, HBase, Sqoop, HBase, Cassandra, Hadoop Streaming, ZooKeeper, Oozie, Kafka, Spark, Storm and Flume.
Architecture & Framework: Client-Server, MVC, J2EE, Struts, Spring, Hibernate.
Database: Cassandra, HBase, Oracle 11g, SQL server 2014, 2012,2008R2,MySQL
IDE: Eclipse, WSAD, Net Beans, IBM RAD, JBuilder.
Design Methodology: UML, Water Fall, Perl, Agile
Operating Systems: Windows 9x/NT/XP/2003/VISTA/7/8, Linux, Unix
GUI: HTML, XML, XSLT, JSF, AJAX, JavaScript, CSS, JQuery, YUI.
Query Languages: SQL, PL/SQL.
Programming Language: Core Java APIs (I/O, Thread, Collection), C, C++, AJAX, Python, JavaScript.
Design patterns: Business Object, Business Delegate, Value Object, Front Controller, Database Access Object, Factory, Singleton, Session Facade.
Tools: BEA WebLogic 8.1, JBOSS, IBM Websphere Application Server 6.1, Tomcat 6.0, Informatica 8/9, SAS, Tableau,Teradata Studio, SSMS,SSIS, JUnit 4.0, ANT, Log4j, Mercury Quality Centre, Rational Clear Quest. ANT, Maven, SVN, Toad.
Design & Control: UML, Rational Rose, CVS, Clear Case V 7.0.
PROFESSIONAL EXPERIENCE
Senior Hadoop Admin
Confidential, Cleveland, OH
Responsibilities:
- Install, maintain and tune 35 node Hortonworks HDP2.2 hadoop cluster
- Monitor Hadoop cluster though Ambari, implemented capacity scheduler to share the resources of the cluster for the mapreduce job s given by users.
- Manage and review Hadoop log files, file system management and monitoring Hadoop cluster capacity planning.
- Expertise in support activities including installation, configuration and successful deployment of changes across all environments
- Hands on experience in working with ecosystem like hive, pig, sqoop, mapreduce, yarn, flume zookeeper, hue and impala. Strong knowledge of Hadoop and hive’s analytical functions.
- Importing and exporting data in hdfs using Sqoop.
- Involved in complete SDLC of project from Design, Analysis,Logical and Physical Architecture Modeling,Development,Implementation and testing.
- Good understanding and related experience with Hadoop stack-internals, Hive, Pig and MapReduce.
- Responsible for Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Involved in loading data from UNIX file system to HDFS
- Designed workflows by scheduling Hive processes for Log file data.
- Used Flume to load the log data into HDFS.
- Optimized HIVE analytics SQL queries and achieve job performance.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Importing and exporting data into RDBMS and Hive using Sqoop.
- Developed Pig scripts in the areas where extensive coding needs to be reduced.
- Managing and scheduling Jobs on a Hadoop cluster using Oozie.
- Implement both major and minor version upgrades to the existing cluster and also rolling back to the previous version.
- Responsible to manage data coming from different sources.
- Worked with the Linux administration team to prepare and configure the systems to support Hadoop deployment.
- Implemented test scripts to support test driven development and continuous integration.
- Participated in requirement gathering form Business Partners and converting the requirements into technical specifications.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment.
- Generated reports using the Tableau report designer.
Environment: Hadoop, Map Reduce, Hive QL, Hive, HBase, Sqoop, Kafka, Storm, Solr, Cassandra, Flume, JasperSoft, Impala, Oozie, Informatica, MYSQL, Oracle SQL, Java, Unix Shell, YARN, Pig Latin,SAS,Teradata,Sql server.
Senior Big-Data Engineer
Confidential, Atlanta,GA
Responsibilities:
- Installed and configured Hadoop Map Reduce, HDFS, Developed multiple maps reduce jobs in java for data cleaning and preprocessing.
- Analysed Hadoop stack and different big data analytic tools including Pig and Hive, Hbase & Cassandra databases and Sqoop.
- Involved in requirement gathering of the enhancement for the project
- Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms.
- Involved in loading data from UNIX file system to HDFS.
- Involved in writing MapReduce jobs to discover trends in data usage by users.
- Involved in managing and reviewing Hadoop log files and running Hadoop streaming jobs to process terabytes of data.
- Extensively used Pig for data cleansing.
- Load and transform large sets of structured, semi structured and unstructured data.
- Developed Hive queries and PigUDF’s for analysis.
- Monitoring and tuning Map Reduce Programs running on the cluster.
- Involved in HDFS maintenance and loading of structured and unstructured data.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Used Zookeeper for various types of centralized configurations.
- Involved in maintaining various Unix Shell scripts.
- Worked on debugging, performance tuning of Hive & Pig Jobs.
- Involved in scheduling Oozie workflow engine to run multiple Hive and Pig jobs
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
- Used SVN for version control.
- Maintain System integrity of all sub-components (primarily HDFS, MR, HBase, and Flume).
- Monitor System status and logs and respond accordingly to any warning or failure conditions.
Environment: Hadoop, Map Reduce, Hive QL, Hive, HBase, Sqoop, Kafka, Storm, Solr, Cassandra, Flume, Tableau, Impala, Oozie, Informatica, MYSQL, Oracle SQL, Java, Unix Shell, YARN, Pig Latin,Teradata,SAS.
Hadoop Developer
Confidential, Warren, NJ
Responsibilities:
- Worked on analyzing Hadoop cluster using different big data analytic tools including Pig, Hive, and MapReduce
- Involved in installing Hadoop Ecosystem components on 50 nodes production.
- Installed and configured the Hadoop name node ha service using Zookeeper.
- Installed and configured Hadoop security and access controls using Kerberos, Active Directory
- Responsible for writing Hive Queries for analyzing tera bytes of customer data from Hbase and put the results in CSV file.
- Used data to import from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
- Supported MapReduce Programs those are running on the cluster
- Implemented test scripts to support test driven development and continuous integration
- Worked on tuning the performance Pig queries
- Involved in loading data from LINUX file system to HDFS using Kettle
- Importing and exporting data into HDFS and Hive using Sqoop
- Experience working on processing unstructured data using Pig and Hive
- Gained experience in managing and reviewing Hadoop log files.
- Wrote Hive UDFS to format the data.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, Oozie, LINUX, and Big Data.
Senior Java/J2EE Developer
Confidential, Sacramento, CA
Responsibilities:
- Actively participated in requirements gathering, analysis, design, and testing phases.
- Designed use case diagrams, class diagrams, and sequence diagrams as a part of Design Phase using Rational Rose.
- Responsible for coordinating on-site and off-shore development teams in various phases of the project.
- Developed the entire application implementing MVC Architecture integrating JSF with Hibernate and spring frameworks.
- Designed User Interface(UI) using Java Server Faces (JSF),Cascading Style Sheets (CSS), and XML.
- Developed the Enterprise Java Beans (Stateless Session beans) to handle different Sessions.
- Developed deployment descriptors for the EJB have to deploy on Web Sphere Application Server.
- Implemented Service Oriented Architecture (SOA) using JMS for sending and receiving messages while creating web services.
- Developed Web Services for data transfer from client to server and vice versa using Apache Axis, SOAP, WSDL, and UDDI.
- Developed the applications on Linux Environment by connecting to Database and implementing Programs.
- Extensively worked on MQ Series using point-point, publisher/subscriber messaging Domains for implementing Exchange of information through Messages.
- Developed XML documents and generated XSL files for Payment Transaction and Reserve Transaction systems.
- Implemented various J2EE Design patterns like Singleton, Service Locator, Business Delegate, DAO, Transfer Object, and SOA.
- Worked on AJAX to develop an interactive Web Application and JavaScript for Data Validations.
- Used Subversion to implement version control System.
- Build ANT Script for the application and used Log4J for debugging.
- Used JUnit Framework for the unit testing of all the java classes.
Environment: JDK 1.5, J2EE, JSF 1.2, EJB 2.0, JNDI 1.2, Hibernate 2.1, Spring 2.0, HTML, JavaScript, XML, CSS, JUnit, UML, Ireport 2.0 and 4.0, Web Services, SOAP, WSDL, UDDI, AXIS 2, Ajax, Ant, Eclipse 3.3, IBM Web Sphere 6.1, DB2, subversion, Linux.
Java/J2EE Developer
Confidential, Hartford, CT
Responsibilities:
- Actively Participated in JAD (Joint application development) sessions for requirements gathering and documenting business process.
- Contributed to the Design, Created Class diagrams, Sequence diagrams and Activity Diagrams.
- Worked with cross-browser issues related to the front end that is created using JavaScript, CSS, and HTML5.
- Worked on how to produce the graphs using JavaScript.
- Created the scripting code to validate the data.
- Worked on different J2EE Design Patterns such as Front controller, Session Facade, Service Locator, Singleton and DAO.
- Developed/modified the bean components to in corporate new business level validations.
- Implemented SAX parsing of XML production data.
- Developed the Core Services which consume the partners Web Services and exposed them as Web services using Axis 1.2.
- Deployed Web Services in Web Sphere.
- Wrote queries, stored procedures and functions using SQL, PL/SQL
- Developed the Business logic Java components.
- Developed the email component using Java Mail to send confirmation emails to users.
- Developed the Search Widget using JSP, Struts, Tiles, JavaScript and AJAX.
- Was responsible for deploying and testing entire Web Services layer developed as a part of Core, using SoapUI.
- Involved in coding Data Layer, which is wrapper class over database.
- Generated build file using Ant 1.6.
- Responsible for performing end-to-end system testing of application writing JUnit test cases
- Developed test environment for testing all the Web Service exposed as part of the core module and their integration with partner services in Integration test.
- As part of the development team Contributed for Application Support in Soft launch and UAT phase.
Environment: Java, Servlets, JSP, JDBC, Struts, Tiles, AJAX, JAXP, XML, XSL, XSLT, SOAP,JSON, Web Services, Java Mail, Eclipse, Web Sphere, Axis, JUnit, Oracle 8, PL/SQL, Ant, Rational Rose 2000, TOAD, Edit plus, HTML, Windows NT, Unix, Java / J2EE Patterns.
Java Developer
Confidential
Responsibilities:
- Involved in design and development of web front end using HTML, Java Script, CSS and JSP’s for Administration, Efficiency Management and Self-Assessment modules and also part of Data Warehousing development team using Informatica.
- Developed and tested the Efficiency Management module using EJB, Servlets, and JSP&Core Java components in WebLogic Application Server
- Developed Struts framework, providing access to system functions of a server’s business layer
- Developed Workflows using Informatica and automated them using Unix Scripts.
- Used the Struts Validation framework for validating the forms.
- Implemented business components as persistent object model as EJBCMP and BMP Entity Beans for storing and retrieving data objects from Resources
- Deployed web components, presentation components and business components in WebLogic Application Server.
- Developed business objects and business object helpers which interact with middleware stubs.
- Implemented the application MVC Architecture using Struts framework
- Involved in stored procedures using PL/SQL to interact with the Oracle database required by the Efficiency Module, Informatica tool.
- Developed Informatica transformations using Informatica power center designer.
Environment: Java, J2EE (Servlets, JDBC, EJB, JSP, JMS), HTML, CSS, Java Script, eclipse, Struts Frame Work 1.1, ANT, XML, CVS, Oracle 8i,PL/SQL, Log4j, Windows XP.