We provide IT Staff Augmentation Services!

Sr Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

Alpharetta, GA

SUMMARY:

  • Around 8+ years of Information Technology experience. Extensive experience in design Experience in Hadoop administration activities such as installation and configuration of clusters using Apache, Cloudera, Hortonworks, AWS ECS and ISILON Able to understand business and technical requirements quickly; Excellent communications skills and work ethics; Able to work independently; Experience working with clients of all sizes.
  • Over 5+ years of experience in Hadoop Administration.
  • Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop Confidential, Yarn, MapReduce, Spark, HBase, Oozie, Hive, Sqoop, Pig, Flume, SmartSense, Storm, Kafka, Ranger, Falcon and Knox.
  • Worked in Agile Development Methodology.
  • Experience in deploying Hadoop cluster on Public and Private Cloud Environment like Cloudera, Hortonworks, Amazon AWS, ECS & ISILON.
  • Setting up automated 24x7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia.
  • Experience in managing and reviewing Hadoop log files.
  • Experience in setting up the High - Availability Hadoop Clusters.
  • Ability to prepare documents including Technical Design, testing strategy, and supporting documents.
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4), Yarn distributions (CDH 5.X).
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Hortonworks (HDP 2.2, HDP2.3).
  • Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Good experience on Design, configure and manage the backup and disaster recovery for Hadoop data.
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Experience in understanding the security requirements for Hadoop and integrating with Kerberos authentication infrastructure- KDC server setup, creating realm /domain, managing.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • As a admin involved in Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
  • Experience in Confidential data storage and support for running map-reduce jobs.
  • Installing and configuring hadoop eco system like sqoop, pig, hive.
  • Knowledge on HBase and zookeeper.
  • Experience in importing and exporting the data using Sqoop from Confidential to Relational Database systems and vice-versa.
  • Knowledge on architecture and implementation of Confluent Kafka
  • Hands on experience on Nagios and Ganglia tool.
  • Scheduling all Hadoop/hive/sqoop/HBase jobs using Oozie.
  • Rack awareness configuration for quick availability and processing of data.
  • Handsome experience in Linux admin activities
  • Experience in configuration management tools chef.
  • Good understanding in Deployment of Hadoop Clusters Using Automated Puppet scripts
  • Experience in hardware recommendations, performance tuning and benchmarking
  • Experience in IP Management (IP Addressing, Sub-netting, Ethernet Bonding, Static IP)
  • Flexible with Unix/Linux and Windows Environments working with Operating Systems like Centos 5/6, Ubuntu 10/11 and Sun.
  • Experience in Linux Storage Management. Configuring RAID Levels, Logical Volumes.

TECHNICAL KNOWLEDGE:

Hadoop Framework: Confidential, Map Reduce, Pig, Hive, HBase, sqoop, zookeeper, Ranger, Storm, Kafka, Oozie, flume, Hue, Knox, Spark

Databases: Oracle 9i/10g, DB2, SQL Server, MYSQL

Cloud Environment: AWS, Azure and ISILON

Operating Systems: Linux RHEL/Ubuntu/CentOS, Windows (XP/7/8)

Scripting Languages: Shell scripting

Network Security: Kerberos

Monitoring Tools: Cloudera Manager, Ambari, Nagios, Ganglia, New Relic

Configuration management: Chef

Containers: Docker and Mesosphere

PROFESSIONAL EXPERIENCE:

Confidential, Alpharetta, GA

Sr Hadoop Administrator

Responsibilities:

  • Manage several Hadoop clusters in production, development, Disaster Recovery environments.
  • Responsible for building a cluster on CDH 5.9
  • Work with engineering software developers to investigate problems and make changes to the Hadoop environment and associated applications.
  • Expertise in recommending hardware configuration for Hadoop cluster
  • Installing, Upgrading and Managing Hadoop Cluster on Cloudera
  • Trouble shooting many cloud related issues such as Data Node down, Network failure and data block missing.
  • Major Upgrade from CDH 5.2 to CDH 5.9.
  • Involved in Map-Reduce programs to cleanse the data in Confidential obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Managing and reviewing Hadoop and HBase log files
  • Proven results-oriented person with a focus on delivery
  • Built and configured log data loading into Confidential using Flume.
  • Performed Importing and exporting data into Confidential and Hive using Sqoop.
  • Managed cluster coordination services through Zoo Keeper.
  • Provisioning, installing, configuring, monitoring, and maintaining Confidential, Yarn, HBase, Flume, Sqoop, Oozie, Pig, Hive, Falcon, Smartsense, Storm, Kafka and Spark
  • Recovering from node failures and troubleshooting common Hadoop cluster issues.
  • Scripting Hadoop package installation and configuration to support fully-automated deployments.
  • Supporting Hadoop developers and assisting in optimization of map reduce jobs, Pig Latin scripts, Hive Scripts, and HBase ingest required.
  • Installed Kafka cluster with separate nodes for brokers.
  • Performed Kafka operations on regular basis.
  • Implemented Kerberos for authenticating all the services in Hadoop Cluster.
  • System/cluster configuration and health check-up.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera manager.
  • Created user accounts and given users the access to the Hadoop cluster.
  • Used Git to version control custom shell scripts.
  • Resolving tickets submitted by users, troubleshoot the error documenting, resolving the errors.
  • Performed Confidential cluster support and maintenance tasks like Adding and Removing Nodes without any effect to running jobs and data.
  • Experience in using Chef and Docker

Environment: Hadoop Confidential, Map Reduce, Hive 10.0, Pig, Puppet, Zookeeper, HBase, Flume, Ganglia, Sqoop, Linux, CentOS, Cloudera, Git, chef, Ansible, Docker

Confidential, Charlotte, NC

Hadoop Administrator

Responsibilities:

  • Responsible for implementation and ongoing administration of Hadoop infrastructure.
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
  • Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing Confidential , Hive, Pig and MapReduce access for the new users.
  • Cluster maintenance as well as creation and removal of nodes using Cloudera Manager Enterprise
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
  • Screen Hadoop cluster job performances and capacity planning.
  • Monitor Hadoop cluster connectivity and security
  • Configuring of Hive, PIG, Impala, Sqoop, Flume, Spark and Oozie in CDH 5
  • Manage and review Hadoop log files.
  • File system management and monitoring.
  • Major Upgrade from CDH 4 to CDH 5.2
  • Confidential support and maintenance.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
  • Scheduling and Managing Oozie Jobs to automate sequence of rotational activity.
  • Deciding on security and access control model for cluster and data protection.
  • Testing plan of production cluster before and after hadoop installation for HA's and performance.
  • Planning on requirements for migrating users to production beforehand to avoid last minute access issues.
  • Planning on data topology, rack topology and resources availability for users to share as required.
  • Planning and implementation of data migration from existing staging to production cluster.
  • Installed and configured Hadoop ecosystem components like MapReduce, Hive, Pig, Sqoop, HBase, ZooKeeper, fuse, Spark and Oozie.
  • Supported MapReduce Programs and distributed applications running on the Hadoop cluster.
  • Prepared multi-cluster test harness to exercise the system for performance, failover and upgrades.
  • Ensured data integrity using 'fsck' and another Hadoop system admin tools for block corruption.
  • Performed a POC on cluster back using distcp, Cloudera manager BDR and parallel ingestion.
  • Implemented commissioning and decommissioning of data nodes, killing the unresponsive task tracker and dealing with blacklisted task trackers.
  • Performed various configurations which include networking and IPTables, resolving hostnames, SSH key less login.
  • Implemented Kerberos Security Authentication protocol for existing cluster.
  • Configured Ganglia which include installing gmond and gmetad daemons which collects all the metrics running on the distributed cluster and presents them in real-time dynamic web pages which would further help in debugging and maintenance.
  • Continuous monitoring and managing the Hadoop cluster through Ganglia and Nagios.
  • Dumped the data from Confidential to MYSQL database and vice-versa using SQOOP.

Environment: Hadoop Confidential, Map reduce, Yarn, Hive, Pig, Flume, Oozie, Sqoop, Cloudera Manager.

Confidential, Pittsburgh, PA

Hadoop Administrator

Responsibilities:

  • Solid Understanding of Hadoop Confidential , Map-Reduce and other Eco-System Projects.
  • Installation and Configuration of Hadoop Cluster
  • Working with Cloudera Support Team to Fine Tune Cluster
  • Experienced in managing and reviewing Hadoop log files.
  • Created user accounts and given users the access to the Hadoop cluster.
  • Worked with application teams to install operating system and Hadoop updates, patches, version upgrades as required.
  • Advised file system team on optimizing IO for Hadoop / analytics workloads.
  • Importing the data from the MySQL and Oracle into the Confidential using Sqoop.
  • Importing the unstructured data into the Confidential using Flume.
  • Written Map Reduce java programs to analyze the log data for large-scale data sets.
  • Experienced in running Hadoop streaming jobs to process terabytes of XML format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Supported Map Reduce Programs those are running on the cluster.
  • Assisted with data capacity planning and node forecasting.
  • Upgraded the Hadoop cluster from CDH3 to CDH4.
  • Jobs management using Fair scheduler.
  • Cluster coordination services through Zoo Keeper.
  • Involved in loading data from UNIX file system to Confidential .
  • Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID configurations.
  • Automate administration tasks using scripting and Job Scheduling using CRON.
  • Manage the day to day operations of the cluster for backup and support.
  • Creating and managing Logical volumes. Using Java JDBC to load data into MySQL.

Environment: Hadoop Confidential, MapReduce, Hive, Pig, Flume, Oozie, Sqoop, spark, Cloudera Manager.

Confidential, Phoenix, AZ

Linux Administrator

Responsibilities:

  • Installation and configuration of Linux for new build environment.
  • Created Virtual server on Citrix Xen Server based host and installed operating system on Guest Servers.
  • Installed Pre-Execution environment boot and Kick start method on multiple servers, remote installation of Linux using PXE boot.
  • Software installation, disk partitioning, file system creation, user id creation and configuration of Linux.
  • Configuring and Managing Yum Repository
  • Disk space Management, Disk quota management, Maintenances of password & shadow file, NIS master and client configuration, NFS file system configuration.
  • Working on Logical Volume Manager (LVM).
  • Installed and Configured 5 node Hadoop cluster
  • LVM configuration and increasing the size of Logical Volume and updating the file system
  • Restricted file and directory access permissions securely by set UID & GID. Set as per project requirements and data security.
  • IP tables configuration and Maintenance.
  • Performed various configurations which include networking and IP Tables, resolving hostnames, SSH key less login.
  • User & Group Management:
  • User Creation, Modification & Deletion as per requirements.
  • Group id creation, deletion, and addition of a group for a user.
  • Utilization of dump and restore for file system backup and restoration.
  • Log management using cron jobs
  • Automate administration tasks through use of scripting and Job Scheduling using CRON.
  • Performance tuning of MySQL engines like MYIASM and Innodb.
  • MySQL application using MySQL workbench, Toad for MySQL, MySQL Administrator.
  • Setting up MySQL Cluster on 2 node servers.
  • Performance tuning for high transaction and volumes data in mission critical environment.

Environment: MySQL, PHP, Shell Script, Apache, Linux.

Confidential

Technical Analyst

Responsibilities:

  • Experience in Provisioning/Installing Linux Operating system (CentOS, Redhat, and Ubuntu) on bare metal and on cloud.
  • Experienced in Linux network Administration tasks like IP Management (IP Addressing, Ethernet Bonding, Static IP)
  • Experience in Linux User management tasks like configuring users, groups, permissions and access control
  • Experience in Linux storage management like LVM, Partitioning, RAID 0/5/6/10
  • &NFS server. Experience in writing shell scripts and cron jobs for automation.
  • Identifying & troubleshooting the defects in Storage component of Windows. Administering and troubleshooting Shared drive access issues for users.
  • Optimizing & troubleshooting on regular basis for server management tasks like Disk Space
  • Monitoring (Root partition only), Disk Defragmentation & Event Log Monitoring, File & Share
  • Permissions / Print Management.
  • Creation of inventory using ARIS tool.
  • Providing Active Directory support which Includes: Transferring FSMO roles to ADC’s during ServerPatching. Troubleshooting Replication Issues.
  • Group Policy Maintenance and Updates Establishing Trust between external domains and fixing Trust Related issues. Managing and Maintaining Groups and User Accounts.
  • Restoring Deleted objects in Active Directory using LDP.exe
  • Managing and Maintaining DNS, DHCP Servers which Includes Scope Creation, Modification, Reservations, Backup and restore DHCP Scopes, Creating External and Internal DNS records, Creating Zones.
  • Extending Certificates and Issuing new certificates to all Servers from root CA.
  • Administration of Citrix Presentation server 4.0, 4.5 on Windows 2000 & 2003 platforms Building Citrix servers and adding them in to the Farms.
  • Performing tasks for publishing the Customer Applications on Citrix farm servers.
  • Managing production server’s accordance to Customer standards.
  • Troubleshooting to resolve issues which arias in Citrix infra Servers and from end users Hands on experience in different stages of Remote Support (Knowledge Transfer, Shadowing, Steady-state, Take-over).

Environment: AD, Citrix, RAID 0/5/6/10,

We'd love your feedback!