We provide IT Staff Augmentation Services!

Sr. Hadoop Administrator Resume

4.00/5 (Submit Your Rating)

Bethlehem, PA

SUMMARY:

  • Cloudera Certified Apache Hadoop Administrator (CCAH).
  • 8+ years of professional IT experience which includes around 4+ years of proven experience in Hadoop Administration in deploying, maintaining, monitoring and upgradingHadoopClusters using Apache, Cloudera (CDH), Hortonworks (HDP) Distributions and MapR.
  • Experience in using Ambari, Cloudera Manager for installation and management of Hadoop Cluster and good knowledge on MapR File system.
  • Experience with managing, troubleshooting and security networks.
  • Experience working with, extending, and enhancing monitoring systems like Nagios & Ganglia.
  • Storage experience with JBOD, NFS, SAN and RAID.
  • Hands on experience usingHadoopecosystem components likeHadoopMap Reduce, HDFS, Spark, Impala, Sentry, HDFS encryption, Zookeeper, Oozie, Hive, Tez, Sqoop, Pig.
  • Worked on Multi Clustered environment and setting up Cloudera Hadoopecho - System.
  • Experience in benchmarking, performing backup and disaster recovery of Name Node metadata and important sensitive data residing on cluster.
  • In-depth understanding of Data Structure and Algorithms.
  • Experience in setting up monitoring tools like Nagios, Ganglia & Cloudera Manager forHadoop.
  • Experience in Importing and exporting data from different databases like SQL Server, Oracle into HDFS and Hive using Sqoop & Kafka.
  • Experience in configuring Zookeeper to provide Cluster coordination services.
  • Worked on setting up Name Node high availability for major production cluster and designed automatic failover control using Zookeeper and quorum journal nodes.
  • Experience in providing security forHadoopCluster with Kerberos.
  • Troubleshooting and Transform data from RDBMS to HDFS.
  • Setup and manage HA on nodes to avoid single point of failures in large clusters.
  • Experience in setting up teh High-AvailabilityHadoopClusters.
  • Experience in upgrading teh existingHadoopcluster to latest releases.
  • Hands on experience on configuring a Hadoop cluster in a professional environment and on VMware and Amazon Web Service (AWS) using an EC2 instance.
  • Experience in developing Shell Scripts for system management.
  • Setting up and maintaining NoSQL Databases like Cassandra & MongoDB.
  • Experience in deploying Cassandra cluster (Apache Cassandra, Datastax).
  • Benchmarking and Stress Testing on Hadoop Cluster.
  • Ability to diagnose network problems.
  • Understanding of TCP/IP networking and its security considerations.
  • Excellent in communicating with clients, customers, managers, and other teams in teh enterprise at all levels.
  • TEMPEffective problem solving skills and outstanding interpersonal skills. Ability to work independently as well as within a team environment. Driven to meet deadlines.
  • Motivated to produce robust, high-performance software.
  • Ability to learn and use new technologies quickly.
  • 24x7 production support.

TECHNICAL SKILLS:

Bigdata Technologies: HDFS, MapReduce (Mrv1 & Mrv2), Spark, Impala, Hive, Pig, Zookeeper, Sqoop, Oozie, Flume, Kafka, Hbase, Cassandra.

Hadoop Distributions: Cloudera (CDH4, CDH5), HDP (2.2, 2.3), MapR.

Programming Languages: Shell Scripting, HQL, Python.

Operating Systems: Windows 98/2000/XP/Vista/NT/ 8.1, Red hat Linux/Centos 4, 5, UNIX.

Database: Oracle 10g/11g, T-SQL, Mongo DB, PL/SQL

Monitoring Tools: Cloudera Manager, Ambari, Nagios, and Ganglia.

PROFESSIONAL EXPERIENCE:

Confidential, Bethlehem, PA

Sr. Hadoop Administrator

Responsibilities:

  • Experienced in Administration, installing, upgrading and managing distributions of Hadoopwith Apache & Cloudera (CDH4, CDH5) on aCluster of 200+ Nodes in different environments such as Development, Test and Production (Operational & Analytics) environments.
  • Extensively involved in Cluster Capacity planning, Hardware planning, and Performance Tuning of teh HadoopCluster.
  • Installation/configuration and troubleshooting to Hortonworks Hadoop cluster (30+nodes) for application development.
  • Installed and configured Cloudera Manager (CM 5.4.10) for easy management of existing Hadoop cluster.
  • Administered and supported distribution of Hortonworks.
  • Worked on HadoopCDH upgradefrom CDH4.x to CDH5.x.
  • Demonstrate and understanding of concepts, best practices and functions to implement a Big Data solution in a corporate environment.
  • Help design of scalable Big Data clusters and solutions.
  • Design and Configure teh Cluster with teh services required (Sentry, Hive server2, Kerberos, HDFS, Hue, Spark, Hive, Hbase, Zookeeper).
  • Troubleshooting teh cluster by reviewing Hadoop LOG files to find RCA.
  • Implemented automatic failover Zookeeper and Zookeeper failover controller.
  • Deployed high availability on teh Hadoop cluster quorum journal nodes.
  • Implemented Rack Awareness for data locality optimization.
  • Worked on developing scripts for performing benchmarking in Hadoop clusters.
  • Involved in collecting and aggregating large amounts of streaming data and every node logs into HDFS using Flume.
  • Monitoring and controlling local file system disk space usage, log files, cleaning log files with automated scripts.
  • Possess good familiarity with open source configuration management and deployment tools such as Puppet.
  • Responsible in creating shell scripts and python scripts for back up of metadata for all Hadoop components and backing up configuration xml files..
  • Implemented Capacity Scheduler to share teh resources of teh cluster for teh MapReduce jobs given by teh users.
  • Experience in using Sqoop to migrate data to and fro from HDFS and RDBMS (MySQL, ETL, Oracle & Teradata) and deployed Hive and HBaseintegration to perform OLAP operations on Hbase data.
  • As a part of POC, used teh Amazon AWSS3 as an underlying file system for theHadoop and implemented teh elastic Map-Reduce jobs on teh data in S3 buckets.
  • Supporting Hadoop developers and assisting in optimization of HQL, Pig, Impala and Spark.
  • Experience in troubleshooting errors inHbase Shell/API, Pig, Hive and MapReduce.
  • Experienced in creating shell scripts and python scripts dat gets teh jobs info running every hour and puts into hive tables to generate teh cluster usage reports.
  • Experience in setting up perimeter level for Kerberos with ADintegration and Sentry role based authorization for HDFS, HIVE.
  • Implementation of Hue for HTTPS, HueSSLClient configurations.
  • Installing and upgrading VMware tools on client machines.
  • Worked on user onboarding to setup newHadoop users. This includes setting up Linux users, HUE users & adding to appropriate groups in AD to get data access.
  • Worked with Kafka for teh proof of concept for carrying out log processing on distributed system.
  • Involved in defining job flows using Oozie for scheduling jobs.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into teh Hadoop Distributed File System and PIG to pre-process teh data.
  • Automated backing up data on regular basis to a DR cluster using distcp scripts.
  • Work with network and Linux system engineers/admin to define optimum network configurations, server hardware and operating system.
  • Evaluate and propose new tools and technologies to meet teh needs of teh organization
  • 24x7 support

Environment: Cloudera, Flume, Kafka, Pig, Oozie, Hive, Sqoop, Impala, Kerberos, UNIX Shell Scripts, Python, Zoo Keeper, SQL, Map Reduce, AWS, Cassandra, VMware, SQL,ETL, Oracle,Taradeta .

Confidential, Birmingham, AL

Sr. Hadoop Consultant

Responsibilities:

  • Hands on experience in Installing, Upgrade and maintain Hadoop clusters with Apache & Hortonworks Hadoop Ecosystem components such as Sqoop, Hbase and MapReduce.
  • Deployed 100+ node Hortonworks Hadoop Cluster (HDP 2.1) using Ambari server 1.5.1.
  • Experienced in managingHadoopinfrastructure like adding capacity, load balancing.
  • Good understanding and related experience with Hadoop stack - internals, Hive, Pig and Map/Reduce.
  • Implemented network topology for rack awareness.
  • Experience in managing teh cluster resources by implementing fair and capacity scheduler with ACL enabled.
  • Developed automated scripts using Unix Shell for running Balancer, file system health check and User/Group creation on HDFS.
  • Involved in Monitoring and support through Nagios and Ganglia.
  • Enabled Name node HA with auto Fail over.
  • Managing teh configuration of teh clusters to meet teh needs of analysis whether me/O bound or CPU bound.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig, Hive and Sqoop.
  • Loaded data into teh cluster from dynamically generated files using Flume and from relational databases management systems using Sqoop.
  • Flume configuration for teh transfer of data from teh webservers to teh HDFS.
  • Performing benchmark test on Hadoop clusters and tweak teh solution based on test results.
  • Supported users in running Pig and Hive queries and with teh debugging.
  • Responsible for troubleshooting issues in teh execution of Map Reduce jobs by inspecting and reviewing log files.
  • Performed Data scrubbing and processing with Oozie.
  • Installed security on Hortonworks Hadoop Cluster using with Kerberos by working along with operations team to move non secured cluster to secured cluster.
  • Responsible to manage data coming from different sources.
  • Involved in loading data from UNIX file system to HDFS.
  • Monitor System health and respond accordingly to any warning or failure conditions.
  • Writing automation scripts for loading data to cluster and deployment with installation of services using scripts.
  • Create Execute and Debug SQL queries to perform data completeness, correctness, data transformation and data quality testing.
  • Sentry configuration for appropriate user permissions accessing Hive server2/beeline.
  • Cluster maintenance as well as creation and removal of nodes using Hortonworks.
  • Created Data migration plan from one cluster to another using BDR.
  • Monitoring teh cluster on a daily basis and to check teh error logs and debugging them.
  • Constantly learning various Big Data tools and providing strategic direction as per development requirement.

Environment: RHEL, CentOS, Ubuntu, CDH3, Apache Hadoop, Hortonworks, HDFS, Map, Reduce, Hbase, Shell Scripts. Nagios, Ganglia

Confidential, Dearborn, MI

Sr. Linux System Administrator

Responsibilities:

  • Worked on Administration, monitoring and fine tuning on an existing Cloudera Hadoop Cluster used by internal and external users as a Data and Analytics as a Service Platform.
  • Worked on Cloudera Cluster backup and recovery, performance monitoring, load balancing, rebalancing, tuning, capacity planning and disk space management.
  • Assisted in designing, development and architecture of Hadoop and HBase systems.
  • Coordinated with technical teams for installation of Hadoop and third related applications on systems.
  • Formulated procedures for planning and execution of system upgrades for all existing Hadoop clusters.
  • Supported daily operations and helped to develop strategies in-order to improve availability and utilization of UNIX environments.
  • Worked on system administration, user creation, file/directory permissions, LVM, loading of software and system patches.
  • Supported technical team members for automation, installation and configuration tasks.
  • Involved in troubleshooting problems and issues related to teh efficient, secure operation of teh Linux operating system.
  • Worked closely in designing and optimizing teh configuration of Linux to meet teh Service Level Agreements of our applications and services.
  • Worked in teh development and maintenance of UNIX shell scripts for automation and suggested improvement processes for all process automation scripts and tasks.
  • Worked on Linux Administration LVM (Logical Volume Manager) administration, user management, exposure to yum, zypper, rpm and security hardening.
  • Monitored server and application performance & tuning via various commands (vmstat, top, iostat, ping, vmstat, free etc.).
  • Implemented Firewall with (iptables) rules for new servers to enable communication with application servers.
  • Extensive UNIX system administration experience, user creation, file/directory permissions, LVM, loading of software and system patches in a large scale server environment.
  • Worked upon concepts of tools, technologies and methodologies to collaborate with other technical specialists when carrying out assigned duties.
  • TEMPEffectively negotiated with technical peers and customers to implement technical solutions.
  • Worked in performing firmware upgrade in Linux Platform.

Environment: Hadoop, Cloudera, Linux, LVM, Redhat Linux 4, Firewall, Shell Scripting, HBase

Confidential

Linux System Administrator

Responsibilities:

  • Installing and upgrading OE & Redhat Linux and Solaris 8/9/10 x86 & SPARC on Servers like
  • HP DL 380 G3, 4 and 5 & Confidential Power Edge servers.
  • Experience in LDOM's and Creating sparse root and whole root zones and administered teh zones for Web, Application and Database servers and worked on SMF on Solaris 5.10.
  • Implemented and administered VMware ESX 3.0, for running teh Windows, Centos, SUSE andRed Hat Linux Servers on development and test servers.
  • Installed and configured Apache on Linux and Solaris and configured Virtual hosts and applied SSL Certificates.
  • Implemented Jumpstart on Solaris and Kick Start for Red hat environments.
  • Experience working with HP LVM and Red hat LVM.
  • Experience in implementing P2P and P2V migrations.
  • Involved in Installing and configuring Centos & SUSE 11 & 12 servers on HP x86 servers.
  • Implemented HA using Red hat Cluster and VERITAS Cluster Server 4.0 for Web Logic agent.
  • Managing DNS, NIS servers and troubleshooting teh servers.
  • Troubleshooting application issues on Apache web servers and also database servers running onLinux and Solaris.
  • Experience in migrating Oracle, MYSQL data using Double take products.
  • Used Sun Volume Manager for Solaris and LVM on Linux & Solaris to create volumes with layoutslike RAID 1, 5.
  • Re-compiling Linux kernel to remove services and applications dat are not required.
  • Performed performance analysis using tools like prstat, mpstat, iostat, sar, vmstat, truss and Dtrace.
  • Creation of VMs, cloning and migrations of teh VMs on VMware vSphere 4.0/4.1
  • Experience working on LDAP user accounts and configuring LDAP on client machines.
  • Worked on patch management tools like Sun Update Manager.
  • Experience supporting middle ware servers running Apache, Tomcat and Java applications.
  • Worked on day to day administration tasks and resolve tickets using Remedy.
  • Used HP Service center and change management system for ticketing.

Environment: Redhat Linux/CentOS 4, 5, Logical Volume Manager, Hadoop, VMware ESX 3.0, Apache and Tomcat Web Server, Oracle 9g, Oracle RAC, HPSM, HPSA.

Confidential

System Administrator

Responsibilities:

  • Installing, configuring, upgrading and administrating Sun Solaris, RedHat Linux.
  • User account management and support.
  • Jumpstart & Kick-start OS integration, DDNS, DHCP, SMTP, Samba, NFS, FTP, SSH, LDAP integration.
  • Network traffic control, IPSec, Quos, VLAN, Proxy, Radius integration on Cisco Hardware via RedHat Linux Software.
  • Responsible for configuring and managing Squid server in Linux.
  • Configuration and Administration of NIS environment.
  • Managing file systems and disk management using Solstice Disk suite.
  • Involved in Installing and configuring of NFS.
  • Package and patch management on Sun servers.
  • Kernel tuning, Writing Shell scripts for system maintenance and file management.
  • Worked on Solaris volume manager to create file systems as per user and database requirements.
  • Trouble shooting teh system and end user issues.
  • Responsible for configuring real time backup of web servers.
  • Log file was managed for troubleshooting and probable errors.
  • Responsible for reviewing all open tickets, resolve and close any existing tickets.
  • Document solutions for any issues dat has not been discovered previously.

Environment: Sun Solaris 2.6/7, SUN Ultra Enterprise 6000/450, SUN Ultra 10/5/2/1, Windows NT 4.0, RHEL 3.x

We'd love your feedback!