Hadoop Administrator Resume
2.00/5 (Submit Your Rating)
SUMMARY
- Total 6+ years of professional experience in IT industry.
- Having 3.3 years of hands on experience as a Hadoop Administrator in MapR and Hortonworks Distribution.
- Hands on experience on ecosystem components Hive, Sqoop, Pig, Hbase, Oozie, Zookeeper, Ranger, Kerberos and MapReduce.
- Hands on experience in installation, configuration, supporting and managing Hadoop Clusters.
- Decommissioning and commissioning the Node on running Hadoop cluster.
- Expertise in HDFS Architecture and Cluster Concepts.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Enabling High Availability (HA) for various components in Hadoop Cluster Production Environment.
- Rebalancing the Hadoop Cluster.
- Hands on experience on Hadoop Security in Ranger and Kerberos.
- Hands on experience on data transfer/migration across the clusters in Hortonworks.
- Hands on experience on mirroring a volume in MapR.
- Hands on experience on hive and hbase data migration.
- Expertise in Cluster Installation for POC, Dev, Staging and Production environment
- Troubleshooting, diagnosing, tuning and solving the Hadoop issues.
- Worked on importing and exporting data from MySQL databases into HDFS and Hive using Sqoop
- Involved in Hive table creation, partitioning and bucketing of tables
- Written Hive queries (HQL) for data analysis to meet the business requirements
- Sound knowledge of Relational Database Management System (RDBMS)
- Hands on experience in Reporting and Dashboard tool like Pentaho BI Tool
- Good Knowledge in Amazon AWS concepts like EC2 web services which provides fast and efficient processing
- Adequate knowledge and working experience in agile methodologies
- Ability to play a key role in the team and communicates across the team.
TECHNICAL SKILLS
Operating Systems: Centos And RHEL
Relational DBMS: MySQL
Hadoop Distributions: MAPR, Hortonwork.
Cloud Platforms: Amazon Web Services (AWS).
MapReduce/Hadoop Ecosystem: Hadoop, HDFS, Sqoop, Hive, Hbase, Oozie, Pig, Ranger, Kerberos
Ticketing Tools: Service Now, JIRA,HPSM
Change Management Tool: HPSM
Reporting Tools: Pentaho BI Tool
PROFESSIONAL EXPERIENCE
Confidential
Hadoop Administrator
Responsibilities:
- Installed Hadoop Clusters for PROD, ITG, DEV and POC in MapR and Horton works.
- Collaborated with multiple teams for design and implementation of Hadoop clusters.
- Responsible for commissioning & decommissioning of nodes from Clusters.
- Maintaining cluster health and HDFS space for better performance.
- Responsible for disk repairing in cluster.
- Rebalancing the Hadoop Cluster in Hortonworks.
- Working on Name node high availability.
- Allocating the name and space Quotas to the users in case of space problems as per the analysis in Grafana.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons.
- Implemented Kerberos integration with LDAP.
- Implementing and troubleshooting SSH key based password less authentication.
- Implemented scripts for Kerberos keytab generation.
- Able to transfer the data across the cluster in Hortonworks.
- Configured the mirroring setup in MapR cluster for data transfer.
- Good knowledge on hive, hbase and oozie installation and DB configuration for the same.
- Installing and upgradation of Packages and patches according to the client requirement.
- Changing file permissions as per the client request.
- Good exposure in coordinating with vendor related issues for all kinds of hardware failures.
- Processes administration and Management like monitoring, start/stop/kill various process
Confidential
Hadoop Cluster Design/Develop
Responsibilities:
- Cluster maintenance, commissioning & decommissioning data nodes.
- Installation and configuration of MapR Hadoop cluster, Design & develop MapR DR setup, and manage data on MapR cluster
- End - to-end performance tuning of MapR clusters and Hadoop Map/Reduce routines against very large data sets, working with MapR cluster along with MapR-Table(creation, import, export, scan, list)
- Managing & monitoring cluster.
- Performed data balancing on clusters
- Applications PROD Support as roaster and Hadoop Platform Support.
- Managing MFS cluster users (MAPR), permissions and Application users access.
- Working on Name Node high availability customizing zookeeper services.
- Improve speed, efficiency and scalability of the continuous integration environment
- Managing quotas to MapR File System.
- Recovering from node failure and troubleshooting common Hadoop cluster issues.
- Responsible for MapR File system data rebalancing.
- Responsible for performing the backup and Restoration of data from MFS to SAN and Tapes as per Retention Policy.
- Coordinating with team members for proper resolution of tickets
- Checking daily jobs and space alerts
- Manage and review Hadoop log files.
- Troubleshooting day-to-day issues, such as login problems, network issues, permission issues.
Confidential
Hadoop Cluster Enginieer
Responsibilities:
- Load the processed data from each upstream application to HDFS via SQOOP
- Experience in analyzing structured data using HIVE, PIG
- Involving in developing the Hive Reports.
- Monitoring and managing the Hadoop cluster.
- Working on setting up Hadoop multi node clusters, pig, Hive and Hbase.
- Using Sqoop extensively to import data from RDMS sources into HDFS.
- Performed transformations, cleaning and filtering on imported data using Hive and loaded final data into HDFS.
- Involving in creating external table, partitioning, bucketing of table.
- Responsible for creation and setting up of environment and re-configuration activities
- Created reports for the BI team using Sqoop to export data into HDFS and Hive.
Confidential
Hadoop Cluster Enginieer
Responsibilities:
- Working on data analysis in HDFS using Hive, PIG and Map Reduce jobs.
- Involved in start to end process of Hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
- Experienced in managing and reviewing Hadoop log files.
- Involving in developing the Hive Reports.
- Working together with infrastructure, network, database, application and business
- Working on setting up Hadoop multi node clusters, pig, Hive using Ambari.
- Using Sqoop extensively to import data from RDMS sources into HDFS. Performed transformations, cleaning and filtering on imported data using Hive and loaded final data into HDFS.
- Involving in creating external table, partitioning, bucketing of table.
- Extracting Hive query o/p to local file system in text/csv format.
- Responsible for creation and setting up of environment and re-configuration activities
- Created reports for the BI team using Sqoop to export data into HDFS and Hive.
- Analyzed the applications issues which are reported by production support team/business users and provide the solution for bug fix.
