We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

SUMMARY

  • Total 6+ years of professional experience in IT industry.
  • Having 3.3 years of hands on experience as a Hadoop Administrator in MapR and Hortonworks Distribution.
  • Hands on experience on ecosystem components Hive, Sqoop, Pig, Hbase, Oozie, Zookeeper, Ranger, Kerberos and MapReduce.
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters.
  • Decommissioning and commissioning the Node on running Hadoop cluster.
  • Expertise in HDFS Architecture and Cluster Concepts.
  • Installation of various Hadoop Ecosystems and Hadoop Daemons.
  • Enabling High Availability (HA) for various components in Hadoop Cluster Production Environment.
  • Rebalancing the Hadoop Cluster.
  • Hands on experience on Hadoop Security in Ranger and Kerberos.
  • Hands on experience on data transfer/migration across the clusters in Hortonworks.
  • Hands on experience on mirroring a volume in MapR.
  • Hands on experience on hive and hbase data migration.
  • Expertise in Cluster Installation for POC, Dev, Staging and Production environment
  • Troubleshooting, diagnosing, tuning and solving the Hadoop issues.
  • Worked on importing and exporting data from MySQL databases into HDFS and Hive using Sqoop
  • Involved in Hive table creation, partitioning and bucketing of tables
  • Written Hive queries (HQL) for data analysis to meet the business requirements
  • Sound knowledge of Relational Database Management System (RDBMS)
  • Hands on experience in Reporting and Dashboard tool like Pentaho BI Tool
  • Good Knowledge in Amazon AWS concepts like EC2 web services which provides fast and efficient processing
  • Adequate knowledge and working experience in agile methodologies
  • Ability to play a key role in the team and communicates across the team.

TECHNICAL SKILLS

Operating Systems: Centos And RHEL

Relational DBMS: MySQL

Hadoop Distributions: MAPR, Hortonwork.

Cloud Platforms: Amazon Web Services (AWS).

MapReduce/Hadoop Ecosystem: Hadoop, HDFS, Sqoop, Hive, Hbase, Oozie, Pig, Ranger, Kerberos

Ticketing Tools: Service Now, JIRA,HPSM

Change Management Tool: HPSM

Reporting Tools: Pentaho BI Tool

PROFESSIONAL EXPERIENCE

Confidential

Hadoop Administrator

Responsibilities:

  • Installed Hadoop Clusters for PROD, ITG, DEV and POC in MapR and Horton works.
  • Collaborated with multiple teams for design and implementation of Hadoop clusters.
  • Responsible for commissioning & decommissioning of nodes from Clusters.
  • Maintaining cluster health and HDFS space for better performance.
  • Responsible for disk repairing in cluster.
  • Rebalancing the Hadoop Cluster in Hortonworks.
  • Working on Name node high availability.
  • Allocating the name and space Quotas to the users in case of space problems as per the analysis in Grafana.
  • Installation of various Hadoop Ecosystems and Hadoop Daemons.
  • Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons.
  • Implemented Kerberos integration with LDAP.
  • Implementing and troubleshooting SSH key based password less authentication.
  • Implemented scripts for Kerberos keytab generation.
  • Able to transfer the data across the cluster in Hortonworks.
  • Configured the mirroring setup in MapR cluster for data transfer.
  • Good knowledge on hive, hbase and oozie installation and DB configuration for the same.
  • Installing and upgradation of Packages and patches according to the client requirement.
  • Changing file permissions as per the client request.
  • Good exposure in coordinating with vendor related issues for all kinds of hardware failures.
  • Processes administration and Management like monitoring, start/stop/kill various process

Confidential

Hadoop Cluster Design/Develop

Responsibilities:

  • Cluster maintenance, commissioning & decommissioning data nodes.
  • Installation and configuration of MapR Hadoop cluster, Design & develop MapR DR setup, and manage data on MapR cluster
  • End - to-end performance tuning of MapR clusters and Hadoop Map/Reduce routines against very large data sets, working with MapR cluster along with MapR-Table(creation, import, export, scan, list)
  • Managing & monitoring cluster.
  • Performed data balancing on clusters
  • Applications PROD Support as roaster and Hadoop Platform Support.
  • Managing MFS cluster users (MAPR), permissions and Application users access.
  • Working on Name Node high availability customizing zookeeper services.
  • Improve speed, efficiency and scalability of the continuous integration environment
  • Managing quotas to MapR File System.
  • Recovering from node failure and troubleshooting common Hadoop cluster issues.
  • Responsible for MapR File system data rebalancing.
  • Responsible for performing the backup and Restoration of data from MFS to SAN and Tapes as per Retention Policy.
  • Coordinating with team members for proper resolution of tickets
  • Checking daily jobs and space alerts
  • Manage and review Hadoop log files.
  • Troubleshooting day-to-day issues, such as login problems, network issues, permission issues.

Confidential

Hadoop Cluster Enginieer

Responsibilities:

  • Load the processed data from each upstream application to HDFS via SQOOP
  • Experience in analyzing structured data using HIVE, PIG
  • Involving in developing the Hive Reports.
  • Monitoring and managing the Hadoop cluster.
  • Working on setting up Hadoop multi node clusters, pig, Hive and Hbase.
  • Using Sqoop extensively to import data from RDMS sources into HDFS.
  • Performed transformations, cleaning and filtering on imported data using Hive and loaded final data into HDFS.
  • Involving in creating external table, partitioning, bucketing of table.
  • Responsible for creation and setting up of environment and re-configuration activities
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive.

Confidential

Hadoop Cluster Enginieer

Responsibilities:

  • Working on data analysis in HDFS using Hive, PIG and Map Reduce jobs.
  • Involved in start to end process of Hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
  • Experienced in managing and reviewing Hadoop log files.
  • Involving in developing the Hive Reports.
  • Working together with infrastructure, network, database, application and business
  • Working on setting up Hadoop multi node clusters, pig, Hive using Ambari.
  • Using Sqoop extensively to import data from RDMS sources into HDFS. Performed transformations, cleaning and filtering on imported data using Hive and loaded final data into HDFS.
  • Involving in creating external table, partitioning, bucketing of table.
  • Extracting Hive query o/p to local file system in text/csv format.
  • Responsible for creation and setting up of environment and re-configuration activities
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive.
  • Analyzed the applications issues which are reported by production support team/business users and provide the solution for bug fix.

We'd love your feedback!