We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

0/5 (Submit Your Rating)

OK

SUMMARY

  • 8+ years of professional experience in analysis, design, development, implementation, integration and testing of Client - Server applications using Object Oriented Analysis Design (OOAD) with 3+ years of experience in deploying, maintaining, monitoring and upgradingHadoopClusters. (ApacheHadoop, Cloudera, Hortonworks).
  • Experience with managing, troubleshooting and security networks.
  • Experience working with, extending, and enhancing monitoring systems like Nagios.
  • Storage experience with JBOD, NFS and RAID.
  • Hands on experience usingHadoopecosystem components likeHadoopMap Reduce, HDFS, Zoo Keeper, Oozie, Hive, Sqoop, Pig.
  • Worked on Multi Clustered environment and setting up Hortonworks Hadoopecho -System.
  • In-depth understanding of Data Structure and Algorithms.
  • Experience in setting up monitoring tools like Nagios and Ganglia forHadoop.
  • Experience in Importing and exporting data from different databases like MySQL, SQL Server, Oracle into HDFS and Hive using Sqoop.
  • Experience in configuring Zookeeper to provide Cluster coordination services.
  • Experience in providing security forHadoopCluster with Kerberos.
  • Troubleshooting and Transform data from RDBMS to HDFS.
  • Setup and manage HA on nodes to avoid single point of failures in large clusters.
  • Experience in setting up the High-AvailabilityHadoopClusters.
  • Experience in writing UDFs for Hive and Pig.
  • Experience in upgrading the existingHadoopcluster to latest releases.
  • Worked on large scale data migration in Cloud (AWS)
  • Experience in developing Shell Scripts and automated them using scheduling tool for cluster management.
  • Ability to diagnose network problems
  • Installing and Administration Namenode and Job tracker High Availability
  • Understanding of TCP/IP networking and its security considerations.
  • Excellent in communicating with clients, customers, managers, and other teams in the enterprise at all levels.
  • Effective problem solving skills and outstanding interpersonal skills. Ability to work independently as well as within a team environment. Driven to meet deadlines.
  • Motivated to produce robust, high-performance software.
  • Benchmarking, Monitoring Metrics and Stress Testing on Hadoop Cluster.
  • Ability to learn and use new technologies quickly.

TECHNICAL SKILLS

Bigdata Technologies: MapReduce, Hive, Pig, Zookeeper, Sqoop, Oozie, Flume, HBase, AWS

Bigdata Frameworks: HDFS, YARN, Storm, Kafka

Hadoop Distributions: Cloudera(CDH3, CDH4, CDH5), Hortonworks 2x

Programming Languages: Core Java, Shell Scripting, PowerShell

Operating Systems: Windows 98/2000/XP/Vista/NT/ 8.1, Redhat Linux/Centos 4, 5, Unix

Resource Management: SailPoint

Database: Oracle 10g/11g, T-SQL, MySQL, PL/SQL

ETL Stack: Informatica 8x/9x, SSIS, SSRS, SSAS

Business Modeling tools: UML, MS office, Remedy, Service Now, MS-Visio

PROFESSIONAL EXPERIENCE

Confidential, OK

Hadoop Administrator

Responsibilities:

  • Responsible for designing & configuration of Hadoop clusters Translation of functional and technical requirements into detailed architecture and design.
  • Installed and configured multi-nodes fully distributed Hortonworks Hadoop cluster of large number of nodes.
  • Provided Hadoop, OS, Hardware optimizations.
  • Setting up the machines with Network Control, Static IP, Swap memory and mount points.
  • Installed and configured Hadoop ecosystem components like MapReduce, Hive, Pig, Sqoop, HBase, Zookeeper and Oozie.
  • Involved in testing HDFS, Hive, Pig and MapReduce access for the new users.
  • Cluster maintenance as well as creation and removal of nodes using Cloudera and Hortonworks Manager Enterprise.
  • Debugging the issues after upgrades, by escalating them to vendor.
  • Worked on Apache Storm in combination with Kafka for Website Activity, Tracking Metrics Collection & Monitoring.
  • Implemented best income logic using Pig scripts and UDFs.
  • Researched and shared the best practices for business analysis in Hive.
  • Added users to different LOB’s depending on the requests raised through sailpoint.
  • Configured capacity scheduler and allocated resources to various pool through YARN queue manager.
  • Worked on setting up High availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
  • Performed operating system installation, Hadoop version updates using automation tools.
  • Configured Oozie for workflow automation and coordination.
  • Implemented Fair scheduler on the job tracker to allocate fair amount of resources to small jobs.
  • Set up automated processes to archive/clean the unwanted data on the cluster, in particular on Name node and Secondary name node.
  • Implemented rack aware topology on the Hadoop cluster.
  • Importing and exporting structured data from different relational databases like MySQL into HDFS and Hive using Sqoop.
  • Configured Zookeeper to implement node coordination, in clustering support.
  • Configured Flume for efficiently collecting, aggregating and moving large amounts of log data from many different sources to HDFS.
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Implemented Knox gateway on web services node for providing LDAP based security.
  • Implemented Kerberos Security Authentication protocol for existing cluster.
  • Good experience in troubleshooting production level issues in the cluster and its functionality.
  • Backed up data on regular basis to a remote cluster using DistCP.
  • Regular Commissioning and Decommissioning of nodes depending upon the amount of data.
  • Monitored and configured a test cluster on amazon web services for further testing process and gradual migration.

Environment: Horton Works 2.2, MapReduce, Hive, Pig, Oozie, Sqoop, HDFS, YARN.

Confidential, Des Moines, IA

Hadoop Administrator

Responsibilities:

  • Provided and administered from the ground up - Apache Hadoop & Cloudera clusters for BI and other product/systems development.
  • Worked in the administration activities in providing installation, upgrades, patching and configuration for all hardware and software Hadoop components.
  • Work closely with other administrators (database, storage, Windows) and developers to keep the Hadoop environment running efficiently and effectively.
  • Configured and implemented Apache Hadoop technologies i.e., Hadoop distributed file system (HDFS), MapReduce framework, Pig, Hive, Sqoop, Flume.
  • Helped in setting up Rack topology in the cluster.
  • Helped in the day-to-day support for operation.
  • Worked on performing minor upgrade from CDH3-u4 to CDH3-u6
  • Upgraded the Hadoop cluster from CDH3 to CDH4.
  • Implemented Fair scheduler on the job tracker to allocate the fair amount of resources to small jobs.
  • Implemented Kerberos for authenticating all the services in Hadoop Cluster.
  • Deployed Network file system for Name Node Metadata backup.
  • Dumped the data from MYSQL database to HDFS and vice-versa using SQOOP.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitoring workload, job performance and capacity planning using Apache Hadoop.
  • Worked with application teams to install OS level updates, patches and version upgrades required for Hadoop cluster environments.
  • Created a local YUM repository for installing and updating packages.
  • Dumped the data from one cluster to other cluster by using DISTCP, and automated the dumping procedure using shell scripts.
  • Configured and deployed hive metastore using MySQL and thrift server.
  • Worked with the Linux administration team to prepare and configure the systems to support Hadoop deployment.
  • Created volume groups, logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
  • Designed and allocated HDFS quotas for multiple groups.
  • Performed various configurations, which includes, networking and IPTable, resolving hostnames, user accounts and file permissions, http, ftp, SSH keyless login.
  • Designed machines and worked with installation of NoSQL databases such as MongoDB, Cassandra.

Environment: Linux, CDH 3/4, Pig, Hive, Sqoop, Nagios, Ganglia, Oozie, Sqoop, Apache Hadoop, Cloudera

Confidential - Charlotte, NC

UNIX & Hadoop Administrator

Responsibilities:

  • Worked on Administration, monitoring and fine-tuning on an existing Cloudera Hadoop Cluster used by internal and external users as a Data and Analytics as a Service Platform.
  • Worked on Cloudera Cluster backup and recovery, performance monitoring, load balancing, rebalancing and tuning, capacity planning, and disk space management
  • Assisted in designing, development and architecture of HADOOP systems.
  • Coordinated with technical teams for installation of HADOOP and third related applications on systems.
  • Formulated procedures for planning and execution of system upgrades for all existing HADOOP clusters.
  • Supported daily operations and helped to develop strategies in-order to improve availability and utilization of UNIX environments.
  • Worked on system administration, user creation, file/directory permissions, LVM, loading of software and system patches.
  • Supported technical team members for automation, installation and configuration tasks.
  • Involved in troubleshooting problems and issues related to the efficient, secure operation of the Linux operating system.
  • Worked closely in designing and optimizing the configuration of Linux to meet the Service Level Agreements of our applications and services.
  • Worked in the development and maintenance of UNIX shell scripts for automation and suggested improvement processes for all process automation scripts and tasks.
  • Worked on Linux Administration LVM (Logical Volume Manager) administration, user management, exposure to yum, zypper, rpm and security hardening.
  • Alllocated space quotas for users and other directories in HDFS.
  • Monitored server and application performance & tuning via various commands (vmstat, top, iostat, ping, uptime etc.).
  • Implemented Firewall with (iptables) rules for new servers to enable communication with application servers.
  • Extensive UNIX system administration experience, user creation, file/directory permissions, LVM, loading of software and system patches in a large-scale server environment.
  • Worked upon concepts of tools, technologies and methodologies to collaborate with other technical specialists when carrying out assigned duties.
  • Effectively negotiated with technical peers and customers to implement technical solutions.
  • Worked in performing firmware upgrade in Linux Platform.

Environment: Cloudera 4.X, Java, HDFS, Hive, and HBase, Redhat Linux 4, LVM, Firewall

Confidential

LINUX Administrator

Responsibilities:

  • Installing and upgrading OE & Red hat Linux and Solaris 8/ & SPARC on Servers like HP DL 380 G3, 4 and 5 & Dell Power Edge servers.
  • Experience in LDOM's and Creating sparse root and whole root zones, administered the zones for Web, Application and Database servers, and worked on SMF on Solaris 5.10.
  • Implemented and administered VMware ESX 3.0, for running the Windows, Centos, SUSE and Red hat Linux Servers on development and test servers.
  • Installed and configured Apache on Linux and Solaris and configured Virtual hosts and applied SSL s.
  • Implemented Jumpstart on Solaris and Kick Start for Red hat environments.
  • Experience working with HP LVM and Red hat LVM.
  • Experience in implementing P2P and P2V migrations.
  • Involved in Installing and configuring Centos & SUSE 11 & 12 servers on HP x86 servers.
  • Implemented HA using Redhat Cluster and VERITAS Cluster Server 4.0 for Web Logic agent.
  • Managing DNS, NIS servers and troubleshooting the servers.
  • Troubleshooting application issues on Apache web servers and database servers running on Linux and Solaris.
  • Experience in migrating Oracle, MYSQL data using Double take products.
  • Used Sun Volume Manager for Solaris and LVM on Linux & Solaris to create volumes with layouts like RAID 1, 5, 10, 15.
  • Re-compiling Linux kernel to remove services and applications that are not required.
  • Performed performance analysis using tools like prstat, mpstat, iostat, SAR, vmstat, truss and Dtrace.
  • Experience working on LDAP user accounts and configuring LDAP on client machines.
  • Worked on patch management tools like Sun Update Manager.
  • Experience supporting middle ware servers running Apache, Tomcat and Java applications.
  • Worked on day-to-day administration tasks and resolve tickets using Remedy.
  • Used HP Service center ILO and change management system for ticketing.

Environment: Redhat Linux/CentOS 4, 5, Logical Volume Manager, Hadoop, VMware ESX 3.0, Apache and Tomcat Web Server, Oracle 9g, Oracle RAC, HPSM, HPSA

Confidential

System’s Administrator

Responsibilities:

  • Directed the implementation and performance tuning of Windows 2003 Server environment for client’s global operations. Delivered a major improvement over old VPN system that catapulted productivity of remote sales force.
  • Led in-house and consultant team in large-scale Linux server upgrade for multinational consulting firm, which significantly enhanced system performance.
  • Resolved a history of problematic network failure by troubleshooting LAN/WAN connections between multiple stores for a regional retail chain.
  • Drove a 25% decrease in Web site downtime by migrating RaQ4 server sites and online stores to Linux server. Corrected server configuration problems, wrote scripts to ensure smooth server/reboot functioning, and assisted with a variety of Linux-based inquiries and concerns.
  • Replaced major manufacturer’s vulnerable network with robust security through joint architecture of firewall and DHCP.
  • Stabilized, expanded and protected client network and PC environment. Built new file servers to maximize Web hosting, terminal server, file/print sharing and domain control performance.
  • Evaluated and reconfigured company’s Unix/Linux/Oracle setup, including reallocating SAN disk space to engineer a robust, scalable solution.
  • Improved system security through setup and ongoing maintenance of NetScreen-Security Manager. Delivered a cohesive solution managing Juniper firewalls and containing all IDP software.
  • Researched, recommended and implemented network enhancements that improved system reliability and performance.
  • Revamped NetBackup schedule and upgraded 15 Oracle servers. Optimized system availability (averaging 99.9% uptime) and equipped company with scalable systems to support fast business growth.
  • Trained employees across multiple departments on network operations including log-in procedures, network management software, permissions, printing issues, security and use of software.

Environment: Network Admins administer/monitor network related stuff. Firewall rules, DNS etc.

We'd love your feedback!