Senior Hadoop Administrator Resume
Wilmington, DE
PROFESSIONAL SUMMARY:
- Having 10 years of experience in IT Industry with strong experience in Big Data Hadoop and Java
- 4+ years of experience in BIG DATA Hadoop Administration and Development.
- 6 years of experience in JAVA Development.
- Hands - on experience in designing and implementing solutions using Apache Hadoop 2.4.0, HDFS 2.7,MapReduce2,Hbase 1.1,Hive 1.2,Oozie 4.2.0,Tez 0.7.0,Yarn 2.7.0,Sqoop 1.4.6,MongoDB
- Having Knowledge to implement Hortonworks (HDP 2.3 and HDP 2.1 ), Cloudera ( CDH3, CDH4, CDH5) on Linux.
- Configuring Name-node High availability and Name-node Federation
- Disaster recovery and Backup activities
- Multi-node setup of Hadoop cluster
- Performance tuning and benchmarking of Hadoop Cluster
- Monitoring, maintenance and troubleshooting of Hadoop cluster.
- Security integration of Hadoop Cluster.
- Maintained the cluster by adding and removal of nodes.
- Good knowledge on Kerberos Security.
- Setting up and integrating Hadoop eco system tools - Hbase, Hive, Pig, Sqoop etc.
- Familiar with writing Oozie workflows and Job Controllers for job automation - Hive automation.
- Familiar with writing Hive Queries and Hive query optimization by setting different queues.
- Experience in Importing and exporting data from different databases like MySQL, RDBMS into HDFS and HBASE using Sqoop.
- Strong knowledge in configuring High Availability for Name Node, Hbase, AWS, Hive and ResourceManager.
- Experience in deploying and managing the multi-node development and production Hadoop cluster with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, HBASE, ZOOKEEPER) using Hortonworks Ambari.
- Gaining optimum performance with data compression, region splits and by manually managing compaction in Hbase
- Upgrading from HDP 2.1 to HPD 2.2 and then to HDP 2.3.
- Good knowledge on cluster monitoring tools like Ganglia and Nagios.
- Working experience in Map Reduce programming model and Hadoop Distributed File System.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
- Hands on experience on Unix/Linux environments, which included software installations/upgrades, shell scripting for job automation and other maintenance activities.
- Work on AWS configuration optimization for Hadoop.
- Thorough knowledge and experience in SQL and PL/SQL concepts.
- Sound knowledge of ORACLE 9i, Core Java, jsp, servlet.
- Dedication: Willingness to walk extra mile to achieve excellence.
- Good Knowledge on database stored procedures, functions and Triggers.
- Enthusiasm: High level of motivation.
- Scheduling: Time sense.
- Self-starter and team player, capable of working independently and motivating a team of professionals.
TECHNICAL SKILLS:
Programming Languages: Core Java, C++
Distribution Frameworks: Hadoop
Hadoop Distributions: Hortonworks (HDP 2.1, 2.3 and HDP 2.3), Cloudera (CDH 4.7,5.4)
Hadoop Technologies: MapReduce,HBase0.98,Hive0.13,Sqoop1.4.4,Pig0.12.1,Oozie Oozie 4.0.0
J2EE Components: Servlets, JSP.
Frame work: Hibernate.
Operating Systems: Windows 2000/XP, Linux & Unix
RDBMS: Oracle 9i, 10g, MySQL
Scripting Languages: JavaScript
Markup Languages: HTML
Web/Application Servers: Tomcat 6.0, Weblogic
IDE: Eclipse
PROFESSIONAL EXPERIENCE:
Confidential, Wilmington, DE
Senior Hadoop Administrator
Responsibilities:
- Currently working as admin on Cloudera (CDH 5.5.1) distribution for 4 clusters ranges from POC to PROD.
- Responsible for Cluster maintenance, Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
- Day to day responsibilities includes solving developer issues, deployments moving code from one environment to other environment, providing access to new users and providing instant solutions to reduce the impact and documenting the same and preventing future issues.
- Adding/installation of new components and removal of them through Cloudera Manager.
- Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades.
- Monitored workload, job performance and capacity planning using Cloudera Manager.
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
- Interacting with Cloudera support and log the issues in Cloudera portal and fixing them as per the recommendations.
- Imported logs from web servers with Flume to ingest the data into HDFS.
- Using Flume and Spool directory loading the data from local system to hdfs
- Retrieved data from HDFS into relational databases with Sqoop. Parsed cleansed and mined useful and meaningful data in HDFS using Map-Reduce for further analysis
- Fine tuning hive jobs for optimized performance.
- Implemented custom interceptors for flume to filter data and defined channel selectors to multiplex the data into different sinks.
- Partitioned and queried the data in Hive for further analysis by the BI team.
- Extending the functionality of Hive and Pig with custom UDF s and UDAF’s.
- Involved in extracting the data from various sources into Hadoop HDFS for processing.
- Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.
- Monitoring Solr transparency’s, stat’s, cle’s dashboard and review the solr servers.
- Creating and deploying a corresponding Solr Cloud collection.
- Creating collections and configurations, Register a Lily HBase Indexer configuration with the Lily HBase Indexer Service.
- Configuring, Managing permissions for the users in hue.
- Commissioned and Decommissioned nodes on CDH5 Hadoop cluster on Red hat LINUX.
- Involved in loading data from LINUX file system to HDFS.
- Creating and managing the Cron jobs.
- Worked on tuning the performance Pig queries.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Experience in configuring the Storm in loading the data from MYSQL to HBASE using jms
- Responsible to manage data coming from different sources.
- Involved in loading data from UNIX file system to HDFS.
- Experience in managing and reviewing Hadoop log files.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Installed Oozie workflow engine to run multiple Hive and pig jobs.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
ENVIRONMENT: HDFS, MAPREDUCE, HIVE 1.1.0,HUE 3.9.0, AWS, PIG, FLUME, OOZIE, SQOOP, CDH5, APACHE HADOOP 2.6, SPARK, KAFKA, SOLR, STORM, CLOUDERA MANAGER, REDHAT, MYSQL and ORACLE.
Confidential
Senior Java Developer
Responsibilities:
- Involving in collecting requirements for the enhancement of new functionalities.
- Coding, Unit testing and SIT.
- Involving code reviews
- Coded the business methods according to the IBM Rational Rose UML model.
- Used Apache log 4j Logging framework for logging of trace and Auditing.
- Extensively used Core Java, Servlets, JSP and XML.
- Used Struts 1.2 in presentation tier.
- Used IBM Web-SphereApplication Server.
- Generated the Hibernate XML and Java Mappings for the schemas
- Used DB2 Database to store the system data
- Used IBM Rational Clearcase as the version controller.
- Used Asynchronous JavaScript and XML (AJAX) for better and faster interactive Front-End.
- Used Rational Application Developer (RAD) as Integrated Development Environment (IDE).
- Used unit testing for all the components using JUnit.
ENVIRONMENT: Apache log 4j, Core java, Servlets, JSP, XML, Struts, Hibernate XML, Java Mappings, IBM, JavaScript, XMLAJAX, RAD, IDE, Junit.
Confidential
Senior Java Developer
Responsibilities:
- Designed Entegrate Screens with Java Swings for displaying the transactions.
- Involved in the development of code for connecting to database using JDBC with the help of Oracle JDevelper 9i.
- Involved in the development of database coding including Procedures, Triggers in Oracle.
- Worked as Research Assistant and a Development Team Member
- Coordinated with Business Analysts to gather the requirement and prepare data flow diagrams and technical documents.
- Identified Use Cases and generated Class, Sequence and State diagrams using UML.
- Used JMS for the asynchronous exchange of critical business data and events among J2EE components and legacy system.
- Worked in Designing, coding and maintaining of Entity Beans and Session Beans using EJB 2.1 Specification
- Worked in the development of Web Interface using MVC Struts Framework.
- User Interface was developed using JSP and tags, CSS, HTML and Java Script.
- Database connection was made using properties files.
- Used Session Filter for implementing timeout for ideal users.
- Used Stored Procedure to interact with database.
- Development of Persistence was done using DAO and Hibernate Framework.
ENVIRONMENT: JSP, tags, CSS, HTML, Java Script, MVC Struts Framework, UML, EJB.
Confidential
Java Developer
Responsibilities:
- Designed Entegrate Screens with Java Swings for displaying the transactions.
- Involved in the development of code for connecting to database using JDBC with the help of Oracle JDevelper 9i.
- Involved in the development of database coding including Procedures, Triggers in Oracle.
- Worked as Research Assistant and a Development Team Member
- Coordinated with Business Analysts to gather the requirement and prepare data flow diagrams and technical documents.
- Identified Use Cases and generated Class, Sequence and State diagrams using UML.
- Used JMS for the asynchronous exchange of critical business data and events among J2EE components and legacy system.
Confidential
Software Engineer
Responsibilities:
- Understand the requirements and preparing the Query Register and updating the Query Register as per the design changes if any query on the new design.
- Raise Queries and Involve in discussions with development team for resolving the queries
- Proficient participation in Test Case reviews and Traceability matrix (peer reviews)
- Prepared the Test Cases and Executed those Test Cases
- Raising the defects into the Quality Centre and taking the defect report and distributing the reports to the Development Team
- Understand the change requests suggested by the client during demo and maintaining those changes in the General Issue Register
- Prepared the CR Test cases for the suggested changes
- Analyzing the logs for the issues and reporting the defect with the issue description for better and easy and directly developers can dig into the issue with the provided log
