Sr. Hadoop Administrator Resume
Milwaukee, WI
SUMMARY:
- Around 8 years of professional IT experience including Big data ecosystem related technologies.
- Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Proficient in handling HDFS and Yarn Framework as per business application
- Proficient in Setting up Hadoop cluster with different Hadoop components
- Proficient in integration of Kerberos security for HADOOP Cluster
- Sounds good knowledge with HBASE in provision Cluster on requirement
- Implementation of flume streaming data process.
- Experience in understanding the security requirements for Hadoop and integrating with Kerberos authentication infrastructure - KDC server setup, creating realm /domain, managing principles, generating key tab file for each and every service and managing key tab using key tab tools.
- Exposure to automation Chef
- Worked on setting up Name Node high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
- Experience in Importing and exporting data from different databases like MySQL into HDFS using Sqoop.
- Capable in setting up Cloud infrastructure for HADOOP running Application & Vmware Virtualization
- Building automation using Puppet
- Experience in benchmarking with YSCB for load optimization, performing backup and disaster recovery of Namenode metadata and important sensitive data residing on cluster.
- Experience in installing and administering PXE Server with kick start, setting up FTP, DHCP, DNS servers and Logical Volume Management.
- Familiar with Java virtual machine (JVM) and multi-threaded processing.
- Analyzing the client's existing Hadoop infrastructure and understand the performance bottlenecks and provide the performance tuning accordingly.
- Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
- Have flair to adapt to new software applications and products, self-starter, have excellent communication skills and good understanding of business work flow.
- Working knowledge in SQL, Stored Procedures, Functions,Packages, DB Triggers and Indexes.
- Hands on Experience in using IDE tools like Eclipse.
- Experience working on installing and configuration of windows activedirectory.
- Well versed in designing and implementing MapReduce jobs using JAVA on Eclipse to solve real world scaling problems.
TECHNICAL SKILLS:
Languages: C, C++, Java, Shell Scripting
Big Data Technologies: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Flume, Zookeeper, SparkMongoDB
Security: Kerberos, LDAP, AD
Monitoring tools: Nagios,Ganglia
Databases: Oracle, MySQL and SQL Server, Microsoft SQL Server 2000, MS Access
DB Languages: SQL
Operating Systems: Linux, Windows XP, Server 2003, Server 2008
Development Tools: Eclipse 3.3
Other programming Languages: HTML5,CSS 3,JAVASCRIPT,AJAX,JQUERY,.NET,Visual Studio 2010
Network protocols: TCP/IP, UDP, HTTP, DNS, DHCP, OSPF, RIP
Frameworks: MVC
PROFESSIONAL EXPERIENCE:
Confidential, Milwaukee, WI
Sr. Hadoop Administrator
Responsibilities:
- Capable to automation HADOOP Environments using Automation
- Installed and Setup Hadoop CDH clusters manual intervention & Automation
- Performed a Major upgrade HADOOP Cluster running 100 Nodes from CDH 4 to CDH 5.
- Performed various configurations which includes, networking and iptables, resolving hostnames, user accounts and file permissions, http, ftp, SSH key less login.
- Implemented Fair scheduler on the Job tracker to share the resources of the cluster for the mapreduces jobs given by the users.
- Tuned the cluster parameters w.r.t application and run optimal.
- Installed, Configured and maintained HBASE.
- Enabled HA on the Hadoop cluster for making system fault tolerant.
- Implemented Fair scheduler on the job tracker to allocate the fair amount of resources to small jobs.
- Configured Oozie for workflow automation and coordination.
- Exposure on streaming data using Kafka & Spark.
- Proficient in integration of Kerberos security for HADOOP Cluster.
- Implemented Fair schedulers to share the resources of the cluster for the map.
- Installing and updating packages using YUM.
- Installed Amazon web services (AWS) to check whether Hadoop is a feasible solution or not.
- Used Puppet Enterprise edition for configuration management.
- Wrote shell scripting to automate task all across servers through Puppet.
- Actively involved in configuring puppet open source with several third parties applications capable to transfer data across clusters using distcp
- Created volume groups logical volumes and partitions on the Linux servers and mounted file systems on the created partitions.
- Deep understanding of monitoring and troubleshooting mission critical Linux machines.
- Improve system performance by working with the development team to analyze, identify and resolve issues quickly.
- Ensured data recover by implementing system and application level backups.
- Monitoring connection, CPU usage, Memory usage and preparing reports.
- Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID
- Solving hardware related IssuesTicket assessment on daily basis.
- Automate administration tasks through the use of scripting and Job Scheduling using CRON.
Environment: HDFS, HBASE,YARN, MR, HIVE, PIG, FLUME, OOZIE, SQOOP,KAFKA,SPARK
Confidential, Richfield, MN
Sr. Hadoop Administrator
Responsiblities:
- Built & Deployed Hadoop clusters with different Hadoop components (HDFS, YARN, HBASE, ZOOKEEPER).
- Orchestrating Hadoop cluster using Ambari & maintaining house hold tasks for system to sustain.
- Configuring Scheduler on the Resource Manager to provide a way to share large cluster resources.
- Deployed Name Node high availability for Hadoop cluster where to handle automatic failover control
- Successful implementation of using zookeeper service and quorum journal nodes for HA Environment
- Implementing rack aware topology on the Hadoop cluster ensure data integrity
- Integrating Hadoop cluster with Kerberos authentication for secured authentication & authorization of Hadoop cluster
- Making Work flow automation using OOZIE to maintain jobs flows
- Troubleshooting issues with Hadoop components and fine tuning cluster to run smoothly
- Regular Ad-Hoc execution of Hive and Pig queries depending upon the use cases.
- Capable in setting up Nagios and Ganglia for Infrastructure monitoring to tuning application
- Implementation rack aware topology on the Hadoop cluster.
- Good experience in troubleshoot production level issues in the cluster and its functionality.
- Backed up metadata & db information on regular basis using linux utility scripts
- Regular Ad-Hoc execution of Hive and Pig queries depending upon the use cases.
- Commissioning and Decommissioning of nodes depending upon the amount of data.
Environment: Hive, Pig, Map reduce, Flume, Oracle, Sqoop, Flume,Casssandra,YARN, Hadoop, Hbase
Confidential, Woodland Hills, CA
Hadoop Administrator
Responsibilities:
- Experience in architecting, designing, installation, configuration and management of Cloudera Hadoop Distribution.
- Installation /Configuring /Maintaining Apache Hadoop clusters for application development & platform sub projects like Hive, Pig, HBase, Zookeeper and Sqoop.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Managing and scheduling Jobs on a Hadoop cluster using Ambari & Ambari views
- Deployed Hadoop Cluster w.r.t application needs & enable client and server to communicate with ease
- Implemented Namenode backup using NFS. This was done for High availability.
- Worked on in tuning large / complex SQL queries and manage alerts from PDW and Hadoop.
- Automated environment using Ansible . chef . puppet for day-to-day processes
- Log aggregation from web servers and integrated into HDFS using Flume.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Mapreduce jobs given by the users.
- Environment: HADOOP HDFS, HIVE, HBASE,PIG, FLUME, OOZI, SQOOP, CLOUDERA MANAGER, SQL
Confidential, Tampa, FL
Hadoop Administrator
Responsibilities:
- Benchmarking Hadoop cluster to ensure that application runs smoothly like Tera sort & Teagen.
- Configured Fair scheduler to share the resources of the cluster for resource management
- Managing filesystem check for blocks & directories on Hadoop cluster to ensure no block are missing.
- Manage the day-to-day operations of the cluster for backup and support.
- Performed minor version and major patch upgrades on Hadoop cluster
- Visualizing data using Tableau, in such a way the data is directly accessed from HDFS and reports were populated in Tableau.
- Restructuring platform for application such as Installing and maintaining
- Ensure HDFS user & block permissions to make system smoother
- Installing and updating packages using YUM.
Environment: JDK, Ubuntu Linux, HDFS, Map-Reduce, Hive, Pig, Sqoop, Flume, Zookeeper, HBase
Confidential, River Woods, IL
Linux Administrator
Responsibilities:
- Setting up os Installation and configuration new rack servers
- Created Virtual server on Citrix Xen Server based host and installed operating system on Guest Servers.
- Installed Pre-Execution environment boot and Kick start method on multiple servers, remote installation of Linux using PXE boot.
- Deep understanding of monitoring and troubleshooting mission critical Linux machines.
- Experience with Linux internals, virtual machines, and open source tools/platforms.
- Monitoring the System activity, Performance, Resource utilization.
- Extensive use of LVM, creating Volume Groups, Logical volumes where we can add dynamically on requirement.
- Performed RPM and YUM package installations, patch and other server management.
- Performed scheduled backup and necessary restoration.
- Configured Domain Name System (DNS) for hostname to IP resolution
- Troubleshooting and fixing the issues at User level, System level and Network level by using various tools and utilities. Schedule backup jobs by implementing cron job schedule during non business hours.
- Developed an automation script for Replication failover, means if somehow database fails in replication process, this script will bring your system up to date within 5 minutes without manual intervention.
- Implementing file sharing on the network by configuring NFS on the system to share essential resources.
- Performed reorganization of disk partitions, file systems, hard disk addition, and memory upgrade.
ENVIRONMENT: MYSQL 5.1.6, PHP 4.X, 5.X, TOAD, SHELL SCRIPT, LINUX .
Confidential
Linux/System Administrator
Responsibilities:
- Master in Linux internals, virtual machines, and open source tools/platforms.
- Improve system performance by working with the development team to analyze, identify and resolve issues quickly.
- Ensure data recoverability by implementing system and application level backups.
- Performed various configurations which include networking and IPTables, resolving hostnames, SSH key less login.
- Managed CRONTAB jobs, batch processing and job scheduling.
- Software installation and maintenance.
- Security, users and groups administration.
- Networking service, performance, and resource monitoring..
- Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID configurations.
- Support pre-production and production support teams in the analysis of critical services and assists with maintenance operations.
- Automate administration tasks through use of scripting and Job Scheduling using CRON.
- Performance tuning for high transaction and volumes data in mission critical environment..
- Setting up alert and level for MySQL (uptime, Users, Replication information, Alert based on different query).
- Estimate MySQL database capacities; develop methods for monitoring database capacity and usage.
- Develop and optimize physical design of MySQL database systems..
- Support in development and testing environment to measure the performance before deploying to the production.
ENVIRONMENT: MYSQL 5.1.4, PHP, SHELL SCRIPT, APACHE, MYSQL WORKBENCH, TOAD, LINUX 5.0, 5.1.
