We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

3.00/5 (Submit Your Rating)

Tempe, AZ

PROFESSIONAL SUMMARY:

  • Around 8 years of extensive experience including 2+ years of Hadoop Administration and 5 years of Linux/Unix Administration
  • Experienced Installation/Configuration/Managing/Upgrading the Hortonworks Hadoop Cluster on RedHat Linux servers in multiple environments
  • Responsible for upgrading Hortonworks Hadoop HDP2.2 and Mapreduce 2.0 with YARN in Multi Clustered Node environment
  • Good knowledge on Cloudera Distributes Hadoop(CDH) Cluster
  • Experience upgrading the clusters using Rolling Upgrade(Ambari) and Manual Upgrade(Command line)
  • Performed major upgrade in production environment from HDP 1.3 to HDP 2.2
  • Implemented Kerberos authentication integrated with enterprise Active Directory, creating realm domain, managing principles, generating key tab files for each and every service and managing key tab using key tab tools
  • Involved in the new projects on boarding into the cluster and setting required space quota and permissions for the users to restrict access to respective projects/use cases
  • Proactively working with Hortonworks Support team for resolving any cluster related issues
  • Debugging failed MapReduce jobs and finding the root cause and providing the solution
  • Supporting development team for their questions on Hadoop cluster environment and resolving and helping in connecting to the Hadoop cluster
  • Experience on Hadoop cluster maintenance including data and metadata backups, file system checks, commissioning and decommissioning nodes
  • As an admin followed standard Back up policies to make sure the high availability of cluster
  • Monitor multiple Hadoop clusters environments using Ganglia and Nagios
  • Monitor workload, job performance and capacity planning using Ambari
  • Install and configure Hortonworks on single node clusters for POCs
  • Configured Name node High Availability and Resource Manager High Availability in HDP 2.3.
  • Expertise in Cluster Upgrade experience(HDP 2.2.4.2, Ambari 2.0 to HDP 2.3, Ambari 2.1)
  • Configured Kerberos through Ambari admin panel
  • Created Jenkins server to deploy the code from GitHub onto Hadoop cluster
  • Involved in implementing security on Hortonworks Hadoop Cluster with Ranger and Apache Knox
  • Created HIVE databases and granted appropriate permissions through Ranger policies
  • Experience in working with Ranger in enabling metadata management, governance and audit
  • Configured NiFi in Kerberized Cluster
  • Implemented search on HDFS using operational metadata stored in Elastic Search/Solr
  • Implemented ELK (Elastic Search, Log stash, Kibana) stack to collect and analyze the logs
  • Working with Linux Admin for quarterly OS and other security patching and any hardware issues
  • Preparing the documentation and run books for supporting/upgrading the cluster
  • Monitoring the health of the cluster and making sure the cluster is up, running and healthy
  • Hands - on programming experience in various technologies like Shell Scripting, Perl and Python
  • Having Working Knowledge on Sqoop and Flume for Data Processing
  • Loading the data from the different Data sources like Teradata, MySQL and Oracle into HDFS using Sqoop and load into partitioned Hive tables
  • Excellent knowledge of in NOSQL databases like HBase,MongoDB
  • Formulated procedures for installation of Hadoop patches, updates and version upgrades and automated processes for troubleshooting, resolution and tuning of Hadoop clusters
  • Working with other team members in supporting data ingestion into the cluster
  • Involved in log file management where the logs greater than 7 days old were removed from log folder and loaded into HDFS and stored for 3 months
  • Managing users and application quota on the cluster
  • Installing apache HTTP for hosting internal repository for all the packages
  • Installing other third party tools based on the user requirements
  • Installed Analytics tool R on Linux servers and debugged connectivity issues from R to Hadoop cluster
  • Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modelling and data mining, machine learning and advanced data processing. Experience optimizing ETL workflows
  • Supported Web Sphere Application Server WPS, IBM HTTP/ Apache Web Servers in Linux environment for various projects
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive
  • Supported geographically diverse customers and teams in a 24/7 environments
  • Team player with strong analytical, technical negotiation and client relationship management skills
  • Developed Oozie workflows and sub workflows to orchestrate the Sqoop scripts, pig scripts, hive queries and the Oozie workflows are scheduled through Autosys
  • Conducted detailed analysis of system and application architecture components as per functional requirements
  • Ability to work effectively in cross-functional team environments and experience of providing training to business user

TECHNICAL SKILLS:

Operating Systems: Windows, Red hat Linux, IBM AIX, and Mac

Big Data Technologies: Hortonworks Hadoop Cluster, Map-Reduce, HDFS, Sqoop, Pig, Hive, HBase, Zookeeper,Ranger, Spark, Ambari, YARN, Flume, Oozie, Cassandra, Hcatalog, Mahout, Avro, Storm

Hadoop Cluster: Hortonworks, Cloudera

Frameworks: Hadoop MapReduce, Elastic Search, MVC, Struts 2.x/1.x, Hibernate, Springs

JEE Technologies: JSP, Servlets, JDBC, EJB, spring

Programming Languages: Core Java, J2EE, SQL, UNIX Shell Scripting, Perl, Java Script, Python

IDE’s: Eclipse, Net beans

Web Technologies: JSP, JDBC, ODBC, HTML, CSS3, XML, JavaScript, Servlets, Java Beans, Junit

Web/Application servers: Tomcat 7/8, Web Logic Server, Web Sphere, JBoss, Apache

Databases: Oracle 11G, MySQL, Teradata, postgress, MongoDB, Cassandra

Version Control: SVN, GitHub

Tools: ATOM, Eclipse3.5, R, Ganglia, Nagios, Puppet, Kerberos, Remedy, JIRA, Tableau, Active Directory, Jenkins

PROFESSIONAL EXPERIENCE:

Confidential - Tempe, AZ

Hadoop Administrator

Environment: - Hortonworks, MySQL, R, Postgress, Shell Scripting,Elastic Search, Java, Python, RedHat Linux, Autosys, Tableau, JDK1.8, Teradata, Oracle, Nagios, Ganglia

Responsibilities:

  • Installing/Configuring/Managing Hortonworks Hadoop cluster in multiple environments
  • Integrated Hadoop with Active Directory and enabled Kerberos for Authentication
  • Configured NiFi in Kerberized Cluster
  • Created roles based on the project requirements to restrict the access for users using Ranger
  • Implementation of Ranger, Ranger plug-ins and Knox security tools
  • Experience in Ranger, Knox configuration to provide the security for Hadoop services (hive, base, hdfs etc.)
  • Worked with Ambari Rest API and built custom scripts to automate the process
  • Worked with Ranger, Knox configuration to provide centralized security to Hadoop services
  • Coordinated with technical team for production deployment of software applications for maintenance
  • Implemented search on HDFS using operational metadata stored in Elastic Search/Solr
  • Implemented ELK (Elastic Search, Log stash, Kibana) stack to collect and analyze the logs
  • Planning for cluster capacity and submitting the requests for respective hardware and software
  • Designed, configured and managed the backup and disaster recovery for HDFS data
  • Commissioned Data Nodes when data grew and decommissioned when the hardware degraded
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters
  • Provided operational support services relating to Hadoop infrastructure and application installation
  • Involved in implementing High Availability and automatic failover infrastructure to overcome single point of failure for Name node utilizing zookeeper services
  • Worked with cluster users, designers and scientists in troubleshooting map reduce job failures and issues with Hive and YARN
  • Supporting Tableau connectivity to Hadoop Cluster using ODBC/JDBC

Confidential - Phoenix, AZ

Hadoop Administrator

Environment: -Hortonworks, Map Reduce, Hive, Pig, Java JDK 1.6, AWS, Cent OS 6.4, Shell Scripting, Flume, Apache, Sqoop, base, Kerberos, Active Directory, Red Hat Linux 6.4, My SQL 5.5

Responsibilities-

  • Worked closely with data analysts to construct creative solutions for their analysis tasks.
  • Installed, Configured and managed Flume Infrastructure
  • Performed data analytics in Hive and then exported this metrics back to Oracle Database using Sqoop
  • Created HIVE databases and granted appropriate permissions through Ranger policies.
  • Experienced in managing and reviewing Hadoop log files
  • Installation of various Hadoop Ecosystems and Hadoop Daemons
  • Provided ad-hoc queries and data metrics to the Business Users using Hive, Pig
  • Conducting root cause analysis and resolve production problems and data issues
  • Proactively involved in ongoing maintenance, support and improvements in Hadoop cluster
  • Executed tasks for upgrading cluster on the staging platform before doing it on production cluster
  • Monitor cluster stability, use tools to gather statistics and improve performance
  • Help to plan for future upgrades and improvements to both processes and infrastructure
  • Keep current with latest technologies to help automate tasks and implement tools and processes to manage the environment

Confidential - Dallas, TX

UNIX/LINUX Administrator

Environment: - RedHat Linux, Windows 2008/2010, Tomcat, WebLogic Server administration, Apache, IIS 6.0, Autosys, VERITAS Volume Manager 4.x/5.0

Responsibilities-

  • Managed server on VMware provided test environments on virtual machines
  • Provided application support to large users groups
  • Installed hardware, installed Linux OS, and configured required network on 25 Node HPC cluster
  • Configured and Manage Apache web server
  • Manage software and hardware RAID systems
  • Manage user accounts, user disk usage by setting up quotas
  • Manage System Firewall utilizing IP Chains, IP Tables
  • Administer System logs, Security logs
  • Update software packages and apply security patches
  • Write documentation for internal use about System changes, Systems Administration etc.
  • Maintained 50 Linux and UNIX servers running high availability application and database for client
  • Monitored and maintained all storage devices, created reports, creating vdisks and allocated them to different hosts under the storage network
  • Performed storage allocation from EVA storage to match client needs for production and staging environment. Configured LUNs on server end and created file systems as per application/database requirements
  • Created CIFS and NFS shares on Linux and HPUX hosts as per client request
  • Monitored all environments through Site scope, and other monitoring systems
  • Installed application and database packages requested by clients
  • Managing RPM and YUM package installations, patch and other servers
  • Made changes required by Database and Application team for Linux application servers
  • Ran monthly security checks through UNIX and Linux environment and installed security patches required to maintain high security level for our clients
  • Monitored and maintained backup and disaster recovery environment created for clients.
  • Performed backup and restores for client for production and staging environments
  • Copied/moved file systems from one server to another as per client requests
  • Configuring the NFS servers, setting up servers in network environment and configuring FTP/NTP/NIS servers, clients for various departments and clients

Confidential

System Administration RedHat-Linux

Environment: -RedHat Linux, Tomcat, WebLogic Server administration, Apache HTTP, VERITAS Volume Manager 4.x/5.0, Oracle 10, MySQL, PHP, Shell Scripting

Responsibilities-

  • Worked as a System Administrator on Linux- UNIX platforms
  • Involved in maintenance of systems including patching, setting up Print servers, Configuring File systems, using LVM/VERITAS on HPUX/RedHat Linux
  • Setting up the Backup solution using native Unix/Linux tools and also worked with Omni back/Data protector solutions
  • Installation, configuration and administration of Linux Red Hat, Centos
  • Administered Linux servers Linux Red hat - RHEL 6, for several functions including managing Apache/Tomcat server, mail server, MySQL database, and firewalls in both development and production
  • Responsibilities as a Unix System Administrator include monitor and tune system to make sure of the optimum level of the performance
  • Extensive experience in the concept of LVM, User System Resource Management and Job Scheduling
  • Experience in Hardware and Software refreshes on the server
  • Performed common system administration task including adding users, creating file systems, configuring volumes weekly mksysb
  • Adding Oracle ASM disk to the server. Creating and managing LVM
  • Experience in installation and configuration of HBAs and associated firmware/drivers and scanning and configuration of LUNs/Volumes
  • Create users with limited and full root privileges. Create and manage sudoers
  • Develop Linux shell scripts to accomplish redundant tasks to simplify in distributed administration
  • Responsible for resolving network issues using network tools like ping, tcptraceroute, traceroute, tcp dump
  • Install and configure various services like DHCP, NFS, DNS, Apache Web Server, NIS, LDAP, Samba, SSH, FTP/SFTP, Sendmail/SMTP, Telnet, RPM Package Management, File System Management
  • Perform standard system administration tasks, such as patching, software and hardware upgrades, troubleshooting and problem resolution
  • Monitored the servers and Linux scripts regularly and performed troubleshooting steps tested and installed the latest software on server for end-users
  • Responsible for Patching Linux Servers
  • Perform day to day Linux administration such as user accounts, logon scripts, directory services, file system shares, permissions. Support Oracle database
  • Experience in Binding Luns, Naming, and resizing, Unbinding LUNs through Navisphere
  • Develop and maintain the documents, library and procedural documents of the system
  • Provided 24x7 on call support on a rotation basis

We'd love your feedback!