We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

San Jose, CA

SUMMARY:

  • Over 7 years of professional experience including around and 3 plus years in Big Data analytics as Hadoop Administrator, 4 years of Linux Administrator
  • 2 plus years’ experience in configuring, installing, benchmarking and managing Apache Cloudera distribution of Hadoop and Map R distribution
  • Progressive experience with large scale Hadoop clusters deployment and support, which includes cluster design, capacity planning, installation and configuration, performance tuning, monitoring and troubleshooting.
  • Involved in the Mapr5.1 upgrade installation and configuration of a Hadoop cluster.
  • Worked on analyzing job logs errors using different big data analytical tools including Hive, Spark, Sqoop, Drill, MapR DB, Impala, pig, Hue.
  • Gained experience in managing and reviewing log files.
  • Monitoring the MCS alerts and DB Event Dashboard Alerts
  • Strong understanding of Hadoop eco system components such as HDFS, MapReduce, Sqoop, Spark, Oozie, Pig, Hive, HBase, and Zookeeper
  • Close monitoring and analysis of the Map Reduce job executions on cluster at task level.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
  • Excellent knowledge of Resource Manager's scheduling component Fair Scheduler, FIFO & Capacity Scheduler
  • Good Knowledge of recommended settings of Yarn, HDFS & Data nodes by Cloudera to optimized a Hadoop cluster.
  • Experience in understanding Hadoop multiple data processing engines such as interactive SQL, real time streaming, data science and batch processing to handle data stored in a single platform in Yarn.
  • Experience in analyzing logs using HIVEQL, PIG Latin and custom MapReduce programs in JAVA.
  • Good understanding of NoSQL databases like HBase, Cassandra and MongoDB.
  • Proficient in administering Linux operating system like Cent OS 6xStrong Object Oriented Programming concepts in Java (core)
  • Familiar with Java virtual machine (JVM) and multi - threaded processing.
  • Experience on UNIX commands and Shell Scripting.
  • Possessing Knowledge on the Core Java.

TECHNICAL SKILLS:

Operating System: Windows, Mac OS-X, Unix, Linux.

Big Data Ecosystems: HDFS, YARN, Spark, Hive, Pig, Oozie, HBASE, Impala, Sqoop Drill MapR DB and MapReduce

Project Management / Tools: MS Project, MS Office.

Databases: DB2, SQL server, MySQL

PROFESSIONAL EXPERIENCE:

Confidential, San Jose, CA

Hadoop Administrator

Responsibilities:

  • Involved in the Mapr5.1 upgrade installation and configuration of a Hadoop cluster.
  • Worked on analyzing job logs errors using different big data analytical tools including Hive, Spark,
  • Sqoop, Drill, MapR DB, Impala, pig, Hue.
  • Gained experience in managing and reviewing log files.
  • Monitoring the MCS alerts and DB Event Dashboard Alerts.
  • Creating volumes and allocating yarn resources based on the requirements.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
  • Changes to the configuration properties of the cluster based on volume of the data being processed and performance of the cluster.
  • Handle the node failures.
  • Maintaining Cluster in order to remain healthy and in optimal working condition.
  • Handle the upgrades and Patch updates.

Environment: MapR4.1/5.1, Hadoop 2.7.0, Hive, Pig, Sqoop, Hue, Spark, Impala, Drill, MapR DB, Shell Scripting, Red Hat Linux 6.2.

Confidential, Raleigh NC

Hadoop Administrator

Responsibilities:

  • Handle the installation and configuration of a Hadoop cluster.
  • Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase.
  • Handle the data exchange between HDFS and different Web Applications and databases using Flume and Sqoop.
  • Monitor the data streaming between web sources and HDFS.
  • Close monitoring and analysis of the Map Reduce job executions on cluster at task level.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
  • Changes to the configuration properties of the cluster based on volume of the data being processed and performance of the cluster.
  • Setting up Identity, Authentication, and Authorization.
  • Worked on Automation tools Puppet
  • Maintaining Cluster in order to remain healthy and in optimal working condition.
  • Handle the upgrades and Patch updates.
  • Set up automated processes to analyze the System and Hadoop log files for predefined errors and send alerts to appropriate groups.
  • Adding and removing a node and Data rebalancing and Tuning Map reduce Jobs.
  • Maintaining Backup for name node.
  • Performed Name node recoveries from previous backups.

Environment: Cloudera, Hadoop, HDFS, Hive, Pig, Sqoop, Hbase, Spark, Storm, Zookeeper, Shell Scripting, Linux Red Hat.

Confidential, Columbus, OH

Linux Administrator

Responsibilities:

  • Installing and upgrading OE & Red hat Linux and Solaris 8/9/10 x86 & SPARC on Servers like HP DL 380 G3, 4 and 5 & Dell Power Edge servers.
  • Experience in LDOM's and Creating sparse root and whole root zones and administered the zones for Web, Application and Database servers and worked on SMF on Solaris 10.
  • Experience working in AWS Cloud Environment like EC2 & EBS.
  • Implemented and administered VMware ESX 3.5, 4.x for running the Windows, Centos, SUSE and Red hat Linux Servers on development and test servers.
  • Creating and restoring the backups in Linux and AIX.
  • Monitoring the system performance and doing the kernel tuning to enhance the system performance.
  • Backup management through regular ufs dump, ufs restore, fbackup, recovery, DLT tape drives.
  • Installation, configuring all the hardware, peripheral, operating system and software to effectively deliver system up time.
  • Administering the software packages and patches on servers as well as workstations. • Configuration and administration of NFS environment.
  • Configuration of system initialization files.
  • Bourne and K-shell scripts were developed for regular system maintenance.
  • Troubleshooting and fixing the issues at User level, System level and Network level by using various tools and utilities.
  • Maintaining updated documentation of all the Servers and the installed applications and patches
  • Setting up labs from scratch, testing hardware, installing and configuring various hardware devices like printers, scanners, modems, network and communication devices.
  • Configuration of DHCP, DNS, NFS and auto mounter.
  • Creating, troubleshooting and mounting NFS File systems on different OS platforms.
  • Installing, Configuring and Troubleshooting various software's like Wind, Citrix - Clarify, Rave, VPN, SSH Gate, Visio 2000, Star Application, Lotus Notes, Mail clients, Business Objects, Oracle, Microsoft Project.
  • Troubleshooting and solving problems related to users, applications, hardware etc.
  • Working 24/7 on call for application and system support.
  • Experience in working and supported SIBES database running on Linux Servers.

Environment: HP ProLiant Servers, SUN Servers (6500, 4500, 420, Ultra 2 Servers), Solaris 7/8, VERITAS Net Backup, VERITAS Volume Manager, Samba, NFS, NIS, LVM, Linux, Shel l Programming

Confidential

Linux Administrator

Responsibilities:

  • Worked as Unix Engineer / Administrator in IT infrastructure environment providing server administration, application administration and automation solutions to support business objectives.
  • Responsible for Installation, Configuration, Support and Troubleshooting existing systems in AIX 4.3.3 / 4.5/5.1 L, Red Hat Enterprise Linux operating environment
  • Responsible for the operation, maintenance and integrity of a distributed, networked AIX based environment.
  • Provided 24X7 on call production support.
  • Worked with one of the largest VMware and RHES V3 installations in Texas.
  • Instituted centrally managed Linux configurations for application deployments, disaster recovery.
  • Responsible for customer and advisor facing web services with a volume of 4 million hits per day.
  • Build LPAR and DLPAR system and Installed AIX 5.2, 5.3 on LPAR using NIM.
  • Successfully tested and documented the alternate disk install procedure to clone O/S
  • Performed process automation, scheduling of processes using CRON jobs.
  • Worked in the project of configuring NIS Servers (Master and Slave) and NIS clients and maintained system and custom configuration files in NIS maps.
  • Maintained intranet DNS environment by configuring Primary and Secondary Domain Name Servers and DNS clients.
  • Reconfigured IBM Total Storage SSA Raid arrays … periodically updated microcodes for array components, performed problem determination and advanced diagnostic procedures using diag and SSA service aids.
  • Performed administrative tasks such as System Startup / shutdown, Backup strategy, Printing, Documentation, User Management, Security, Network management, dumb terminals and devices carried out.

Environment: Solaris 8/9/10, Red Hat Linux, SUSE, Sar, Vmstat, weblogic, jboss, Veritas Foundation Suite, Jumpstart, kickstart, ESM and HIDS, korn and perlRem

We'd love your feedback!