Hadoop Administrator Resume
San Jose, CA
SUMMARY:
- Over 7 years of professional experience including around and 3 plus years in Big Data analytics as Hadoop Administrator, 4 years of Linux Administrator
- 2 plus years’ experience in configuring, installing, benchmarking and managing Apache Cloudera distribution of Hadoop and Map R distribution
- Progressive experience with large scale Hadoop clusters deployment and support, which includes cluster design, capacity planning, installation and configuration, performance tuning, monitoring and troubleshooting.
- Involved in the Mapr5.1 upgrade installation and configuration of a Hadoop cluster.
- Worked on analyzing job logs errors using different big data analytical tools including Hive, Spark, Sqoop, Drill, MapR DB, Impala, pig, Hue.
- Gained experience in managing and reviewing log files.
- Monitoring the MCS alerts and DB Event Dashboard Alerts
- Strong understanding of Hadoop eco system components such as HDFS, MapReduce, Sqoop, Spark, Oozie, Pig, Hive, HBase, and Zookeeper
- Close monitoring and analysis of the Map Reduce job executions on cluster at task level.
- Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
- Excellent knowledge of Resource Manager's scheduling component Fair Scheduler, FIFO & Capacity Scheduler
- Good Knowledge of recommended settings of Yarn, HDFS & Data nodes by Cloudera to optimized a Hadoop cluster.
- Experience in understanding Hadoop multiple data processing engines such as interactive SQL, real time streaming, data science and batch processing to handle data stored in a single platform in Yarn.
- Experience in analyzing logs using HIVEQL, PIG Latin and custom MapReduce programs in JAVA.
- Good understanding of NoSQL databases like HBase, Cassandra and MongoDB.
- Proficient in administering Linux operating system like Cent OS 6xStrong Object Oriented Programming concepts in Java (core)
- Familiar with Java virtual machine (JVM) and multi - threaded processing.
- Experience on UNIX commands and Shell Scripting.
- Possessing Knowledge on the Core Java.
TECHNICAL SKILLS:
Operating System: Windows, Mac OS-X, Unix, Linux.
Big Data Ecosystems: HDFS, YARN, Spark, Hive, Pig, Oozie, HBASE, Impala, Sqoop Drill MapR DB and MapReduce
Project Management / Tools: MS Project, MS Office.
Databases: DB2, SQL server, MySQL
PROFESSIONAL EXPERIENCE:
Confidential, San Jose, CA
Hadoop Administrator
Responsibilities:
- Involved in the Mapr5.1 upgrade installation and configuration of a Hadoop cluster.
- Worked on analyzing job logs errors using different big data analytical tools including Hive, Spark,
- Sqoop, Drill, MapR DB, Impala, pig, Hue.
- Gained experience in managing and reviewing log files.
- Monitoring the MCS alerts and DB Event Dashboard Alerts.
- Creating volumes and allocating yarn resources based on the requirements.
- Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
- Changes to the configuration properties of the cluster based on volume of the data being processed and performance of the cluster.
- Handle the node failures.
- Maintaining Cluster in order to remain healthy and in optimal working condition.
- Handle the upgrades and Patch updates.
Environment: MapR4.1/5.1, Hadoop 2.7.0, Hive, Pig, Sqoop, Hue, Spark, Impala, Drill, MapR DB, Shell Scripting, Red Hat Linux 6.2.
Confidential, Raleigh NCHadoop Administrator
Responsibilities:
- Handle the installation and configuration of a Hadoop cluster.
- Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase.
- Handle the data exchange between HDFS and different Web Applications and databases using Flume and Sqoop.
- Monitor the data streaming between web sources and HDFS.
- Close monitoring and analysis of the Map Reduce job executions on cluster at task level.
- Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
- Changes to the configuration properties of the cluster based on volume of the data being processed and performance of the cluster.
- Setting up Identity, Authentication, and Authorization.
- Worked on Automation tools Puppet
- Maintaining Cluster in order to remain healthy and in optimal working condition.
- Handle the upgrades and Patch updates.
- Set up automated processes to analyze the System and Hadoop log files for predefined errors and send alerts to appropriate groups.
- Adding and removing a node and Data rebalancing and Tuning Map reduce Jobs.
- Maintaining Backup for name node.
- Performed Name node recoveries from previous backups.
Environment: Cloudera, Hadoop, HDFS, Hive, Pig, Sqoop, Hbase, Spark, Storm, Zookeeper, Shell Scripting, Linux Red Hat.
Confidential, Columbus, OHLinux Administrator
Responsibilities:
- Installing and upgrading OE & Red hat Linux and Solaris 8/9/10 x86 & SPARC on Servers like HP DL 380 G3, 4 and 5 & Dell Power Edge servers.
- Experience in LDOM's and Creating sparse root and whole root zones and administered the zones for Web, Application and Database servers and worked on SMF on Solaris 10.
- Experience working in AWS Cloud Environment like EC2 & EBS.
- Implemented and administered VMware ESX 3.5, 4.x for running the Windows, Centos, SUSE and Red hat Linux Servers on development and test servers.
- Creating and restoring the backups in Linux and AIX.
- Monitoring the system performance and doing the kernel tuning to enhance the system performance.
- Backup management through regular ufs dump, ufs restore, fbackup, recovery, DLT tape drives.
- Installation, configuring all the hardware, peripheral, operating system and software to effectively deliver system up time.
- Administering the software packages and patches on servers as well as workstations. • Configuration and administration of NFS environment.
- Configuration of system initialization files.
- Bourne and K-shell scripts were developed for regular system maintenance.
- Troubleshooting and fixing the issues at User level, System level and Network level by using various tools and utilities.
- Maintaining updated documentation of all the Servers and the installed applications and patches
- Setting up labs from scratch, testing hardware, installing and configuring various hardware devices like printers, scanners, modems, network and communication devices.
- Configuration of DHCP, DNS, NFS and auto mounter.
- Creating, troubleshooting and mounting NFS File systems on different OS platforms.
- Installing, Configuring and Troubleshooting various software's like Wind, Citrix - Clarify, Rave, VPN, SSH Gate, Visio 2000, Star Application, Lotus Notes, Mail clients, Business Objects, Oracle, Microsoft Project.
- Troubleshooting and solving problems related to users, applications, hardware etc.
- Working 24/7 on call for application and system support.
- Experience in working and supported SIBES database running on Linux Servers.
Environment: HP ProLiant Servers, SUN Servers (6500, 4500, 420, Ultra 2 Servers), Solaris 7/8, VERITAS Net Backup, VERITAS Volume Manager, Samba, NFS, NIS, LVM, Linux, Shel l Programming
ConfidentialLinux Administrator
Responsibilities:
- Worked as Unix Engineer / Administrator in IT infrastructure environment providing server administration, application administration and automation solutions to support business objectives.
- Responsible for Installation, Configuration, Support and Troubleshooting existing systems in AIX 4.3.3 / 4.5/5.1 L, Red Hat Enterprise Linux operating environment
- Responsible for the operation, maintenance and integrity of a distributed, networked AIX based environment.
- Provided 24X7 on call production support.
- Worked with one of the largest VMware and RHES V3 installations in Texas.
- Instituted centrally managed Linux configurations for application deployments, disaster recovery.
- Responsible for customer and advisor facing web services with a volume of 4 million hits per day.
- Build LPAR and DLPAR system and Installed AIX 5.2, 5.3 on LPAR using NIM.
- Successfully tested and documented the alternate disk install procedure to clone O/S
- Performed process automation, scheduling of processes using CRON jobs.
- Worked in the project of configuring NIS Servers (Master and Slave) and NIS clients and maintained system and custom configuration files in NIS maps.
- Maintained intranet DNS environment by configuring Primary and Secondary Domain Name Servers and DNS clients.
- Reconfigured IBM Total Storage SSA Raid arrays … periodically updated microcodes for array components, performed problem determination and advanced diagnostic procedures using diag and SSA service aids.
- Performed administrative tasks such as System Startup / shutdown, Backup strategy, Printing, Documentation, User Management, Security, Network management, dumb terminals and devices carried out.
Environment: Solaris 8/9/10, Red Hat Linux, SUSE, Sar, Vmstat, weblogic, jboss, Veritas Foundation Suite, Jumpstart, kickstart, ESM and HIDS, korn and perlRem