Hadoop Administrator Resume
OK
SUMMARY
- 8+ years of professional experience in analysis, design, development, implementation, integration and testing of Client - Server applications using Object Oriented Analysis Design (OOAD) with 3+ years of experience in deploying, maintaining, monitoring and upgradingHadoopClusters. (ApacheHadoop, Cloudera, Hortonworks).
- Experience with managing, troubleshooting and security networks.
- Experience working with, extending, and enhancing monitoring systems like Nagios.
- Storage experience with JBOD, NFS and RAID.
- Hands on experience usingHadoopecosystem components likeHadoopMap Reduce, HDFS, Zoo Keeper, Oozie, Hive, Sqoop, Pig.
- Worked on Multi Clustered environment and setting up Hortonworks Hadoopecho -System.
- In-depth understanding of Data Structure and Algorithms.
- Experience in setting up monitoring tools like Nagios and Ganglia forHadoop.
- Experience in Importing and exporting data from different databases like MySQL, SQL Server, Oracle into HDFS and Hive using Sqoop.
- Experience in configuring Zookeeper to provide Cluster coordination services.
- Experience in providing security forHadoopCluster with Kerberos.
- Troubleshooting and Transform data from RDBMS to HDFS.
- Setup and manage HA on nodes to avoid single point of failures in large clusters.
- Experience in setting up the High-AvailabilityHadoopClusters.
- Experience in writing UDFs for Hive and Pig.
- Experience in upgrading the existingHadoopcluster to latest releases.
- Worked on large scale data migration in Cloud (AWS)
- Experience in developing Shell Scripts and automated them using scheduling tool for cluster management.
- Ability to diagnose network problems
- Installing and Administration Namenode and Job tracker High Availability
- Understanding of TCP/IP networking and its security considerations.
- Excellent in communicating with clients, customers, managers, and other teams in the enterprise at all levels.
- Effective problem solving skills and outstanding interpersonal skills. Ability to work independently as well as within a team environment. Driven to meet deadlines.
- Motivated to produce robust, high-performance software.
- Benchmarking, Monitoring Metrics and Stress Testing on Hadoop Cluster.
- Ability to learn and use new technologies quickly.
TECHNICAL SKILLS
Bigdata Technologies: MapReduce, Hive, Pig, Zookeeper, Sqoop, Oozie, Flume, HBase, AWS
Bigdata Frameworks: HDFS, YARN, Storm, Kafka
Hadoop Distributions: Cloudera(CDH3, CDH4, CDH5), Hortonworks 2x
Programming Languages: Core Java, Shell Scripting, PowerShell
Operating Systems: Windows 98/2000/XP/Vista/NT/ 8.1, Redhat Linux/Centos 4, 5, Unix
Resource Management: SailPoint
Database: Oracle 10g/11g, T-SQL, MySQL, PL/SQL
ETL Stack: Informatica 8x/9x, SSIS, SSRS, SSAS
Business Modeling tools: UML, MS office, Remedy, Service Now, MS-Visio
PROFESSIONAL EXPERIENCE
Confidential, OK
Hadoop Administrator
Responsibilities:
- Responsible for designing & configuration of Hadoop clusters Translation of functional and technical requirements into detailed architecture and design.
- Installed and configured multi-nodes fully distributed Hortonworks Hadoop cluster of large number of nodes.
- Provided Hadoop, OS, Hardware optimizations.
- Setting up the machines with Network Control, Static IP, Swap memory and mount points.
- Installed and configured Hadoop ecosystem components like MapReduce, Hive, Pig, Sqoop, HBase, Zookeeper and Oozie.
- Involved in testing HDFS, Hive, Pig and MapReduce access for the new users.
- Cluster maintenance as well as creation and removal of nodes using Cloudera and Hortonworks Manager Enterprise.
- Debugging the issues after upgrades, by escalating them to vendor.
- Worked on Apache Storm in combination with Kafka for Website Activity, Tracking Metrics Collection & Monitoring.
- Implemented best income logic using Pig scripts and UDFs.
- Researched and shared the best practices for business analysis in Hive.
- Added users to different LOB’s depending on the requests raised through sailpoint.
- Configured capacity scheduler and allocated resources to various pool through YARN queue manager.
- Worked on setting up High availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
- Performed operating system installation, Hadoop version updates using automation tools.
- Configured Oozie for workflow automation and coordination.
- Implemented Fair scheduler on the job tracker to allocate fair amount of resources to small jobs.
- Set up automated processes to archive/clean the unwanted data on the cluster, in particular on Name node and Secondary name node.
- Implemented rack aware topology on the Hadoop cluster.
- Importing and exporting structured data from different relational databases like MySQL into HDFS and Hive using Sqoop.
- Configured Zookeeper to implement node coordination, in clustering support.
- Configured Flume for efficiently collecting, aggregating and moving large amounts of log data from many different sources to HDFS.
- Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
- Implemented Knox gateway on web services node for providing LDAP based security.
- Implemented Kerberos Security Authentication protocol for existing cluster.
- Good experience in troubleshooting production level issues in the cluster and its functionality.
- Backed up data on regular basis to a remote cluster using DistCP.
- Regular Commissioning and Decommissioning of nodes depending upon the amount of data.
- Monitored and configured a test cluster on amazon web services for further testing process and gradual migration.
Environment: Horton Works 2.2, MapReduce, Hive, Pig, Oozie, Sqoop, HDFS, YARN.
Confidential, Des Moines, IA
Hadoop Administrator
Responsibilities:
- Provided and administered from the ground up - Apache Hadoop & Cloudera clusters for BI and other product/systems development.
- Worked in the administration activities in providing installation, upgrades, patching and configuration for all hardware and software Hadoop components.
- Work closely with other administrators (database, storage, Windows) and developers to keep the Hadoop environment running efficiently and effectively.
- Configured and implemented Apache Hadoop technologies i.e., Hadoop distributed file system (HDFS), MapReduce framework, Pig, Hive, Sqoop, Flume.
- Helped in setting up Rack topology in the cluster.
- Helped in the day-to-day support for operation.
- Worked on performing minor upgrade from CDH3-u4 to CDH3-u6
- Upgraded the Hadoop cluster from CDH3 to CDH4.
- Implemented Fair scheduler on the job tracker to allocate the fair amount of resources to small jobs.
- Implemented Kerberos for authenticating all the services in Hadoop Cluster.
- Deployed Network file system for Name Node Metadata backup.
- Dumped the data from MYSQL database to HDFS and vice-versa using SQOOP.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitoring workload, job performance and capacity planning using Apache Hadoop.
- Worked with application teams to install OS level updates, patches and version upgrades required for Hadoop cluster environments.
- Created a local YUM repository for installing and updating packages.
- Dumped the data from one cluster to other cluster by using DISTCP, and automated the dumping procedure using shell scripts.
- Configured and deployed hive metastore using MySQL and thrift server.
- Worked with the Linux administration team to prepare and configure the systems to support Hadoop deployment.
- Created volume groups, logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
- Designed and allocated HDFS quotas for multiple groups.
- Performed various configurations, which includes, networking and IPTable, resolving hostnames, user accounts and file permissions, http, ftp, SSH keyless login.
- Designed machines and worked with installation of NoSQL databases such as MongoDB, Cassandra.
Environment: Linux, CDH 3/4, Pig, Hive, Sqoop, Nagios, Ganglia, Oozie, Sqoop, Apache Hadoop, Cloudera
Confidential - Charlotte, NC
UNIX & Hadoop Administrator
Responsibilities:
- Worked on Administration, monitoring and fine-tuning on an existing Cloudera Hadoop Cluster used by internal and external users as a Data and Analytics as a Service Platform.
- Worked on Cloudera Cluster backup and recovery, performance monitoring, load balancing, rebalancing and tuning, capacity planning, and disk space management
- Assisted in designing, development and architecture of HADOOP systems.
- Coordinated with technical teams for installation of HADOOP and third related applications on systems.
- Formulated procedures for planning and execution of system upgrades for all existing HADOOP clusters.
- Supported daily operations and helped to develop strategies in-order to improve availability and utilization of UNIX environments.
- Worked on system administration, user creation, file/directory permissions, LVM, loading of software and system patches.
- Supported technical team members for automation, installation and configuration tasks.
- Involved in troubleshooting problems and issues related to the efficient, secure operation of the Linux operating system.
- Worked closely in designing and optimizing the configuration of Linux to meet the Service Level Agreements of our applications and services.
- Worked in the development and maintenance of UNIX shell scripts for automation and suggested improvement processes for all process automation scripts and tasks.
- Worked on Linux Administration LVM (Logical Volume Manager) administration, user management, exposure to yum, zypper, rpm and security hardening.
- Alllocated space quotas for users and other directories in HDFS.
- Monitored server and application performance & tuning via various commands (vmstat, top, iostat, ping, uptime etc.).
- Implemented Firewall with (iptables) rules for new servers to enable communication with application servers.
- Extensive UNIX system administration experience, user creation, file/directory permissions, LVM, loading of software and system patches in a large-scale server environment.
- Worked upon concepts of tools, technologies and methodologies to collaborate with other technical specialists when carrying out assigned duties.
- Effectively negotiated with technical peers and customers to implement technical solutions.
- Worked in performing firmware upgrade in Linux Platform.
Environment: Cloudera 4.X, Java, HDFS, Hive, and HBase, Redhat Linux 4, LVM, Firewall
Confidential
LINUX Administrator
Responsibilities:
- Installing and upgrading OE & Red hat Linux and Solaris 8/ & SPARC on Servers like HP DL 380 G3, 4 and 5 & Dell Power Edge servers.
- Experience in LDOM's and Creating sparse root and whole root zones, administered the zones for Web, Application and Database servers, and worked on SMF on Solaris 5.10.
- Implemented and administered VMware ESX 3.0, for running the Windows, Centos, SUSE and Red hat Linux Servers on development and test servers.
- Installed and configured Apache on Linux and Solaris and configured Virtual hosts and applied SSL s.
- Implemented Jumpstart on Solaris and Kick Start for Red hat environments.
- Experience working with HP LVM and Red hat LVM.
- Experience in implementing P2P and P2V migrations.
- Involved in Installing and configuring Centos & SUSE 11 & 12 servers on HP x86 servers.
- Implemented HA using Redhat Cluster and VERITAS Cluster Server 4.0 for Web Logic agent.
- Managing DNS, NIS servers and troubleshooting the servers.
- Troubleshooting application issues on Apache web servers and database servers running on Linux and Solaris.
- Experience in migrating Oracle, MYSQL data using Double take products.
- Used Sun Volume Manager for Solaris and LVM on Linux & Solaris to create volumes with layouts like RAID 1, 5, 10, 15.
- Re-compiling Linux kernel to remove services and applications that are not required.
- Performed performance analysis using tools like prstat, mpstat, iostat, SAR, vmstat, truss and Dtrace.
- Experience working on LDAP user accounts and configuring LDAP on client machines.
- Worked on patch management tools like Sun Update Manager.
- Experience supporting middle ware servers running Apache, Tomcat and Java applications.
- Worked on day-to-day administration tasks and resolve tickets using Remedy.
- Used HP Service center ILO and change management system for ticketing.
Environment: Redhat Linux/CentOS 4, 5, Logical Volume Manager, Hadoop, VMware ESX 3.0, Apache and Tomcat Web Server, Oracle 9g, Oracle RAC, HPSM, HPSA
Confidential
System’s Administrator
Responsibilities:
- Directed the implementation and performance tuning of Windows 2003 Server environment for client’s global operations. Delivered a major improvement over old VPN system that catapulted productivity of remote sales force.
- Led in-house and consultant team in large-scale Linux server upgrade for multinational consulting firm, which significantly enhanced system performance.
- Resolved a history of problematic network failure by troubleshooting LAN/WAN connections between multiple stores for a regional retail chain.
- Drove a 25% decrease in Web site downtime by migrating RaQ4 server sites and online stores to Linux server. Corrected server configuration problems, wrote scripts to ensure smooth server/reboot functioning, and assisted with a variety of Linux-based inquiries and concerns.
- Replaced major manufacturer’s vulnerable network with robust security through joint architecture of firewall and DHCP.
- Stabilized, expanded and protected client network and PC environment. Built new file servers to maximize Web hosting, terminal server, file/print sharing and domain control performance.
- Evaluated and reconfigured company’s Unix/Linux/Oracle setup, including reallocating SAN disk space to engineer a robust, scalable solution.
- Improved system security through setup and ongoing maintenance of NetScreen-Security Manager. Delivered a cohesive solution managing Juniper firewalls and containing all IDP software.
- Researched, recommended and implemented network enhancements that improved system reliability and performance.
- Revamped NetBackup schedule and upgraded 15 Oracle servers. Optimized system availability (averaging 99.9% uptime) and equipped company with scalable systems to support fast business growth.
- Trained employees across multiple departments on network operations including log-in procedures, network management software, permissions, printing issues, security and use of software.
Environment: Network Admins administer/monitor network related stuff. Firewall rules, DNS etc.