We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

3.00/5 (Submit Your Rating)

Schaumburg, IL

PROFESSIONAL SUMMARY:

  • I have 4+ years of experience in IT field including experience in Hadoop Administration in diverse industries which includes hands on experience in Big data ecosystem related technologies.
  • Extensive knowledge and experience in Big Data with Map - Reduce, HDFS, Hive, Pig, Impala, Spark, Kafka, Zookeeper, Oozie, Sentry, HUE and Sqoop.
  • Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Resource Manager, Name Node, Data Node, and MapReduce(MRV1 and YARN) concepts.
  • Experience in performing various major and minor Hadoop upgraded on large environments.
  • Experience with Securing Hadoop clusters using Kerberos.
  • Hands on experience in installation, configuration, management and development of big data solutions using Cloudera(CDH4, CDH5) and Hortonworks(HDP2.3, HDP2.4 ) distributions.
  • Good experience with design, management, configuration and troubleshooting of distributed production environments based on Apache Hadoop/ Hbase etc.
  • Working experience on designing and implementing complete end to end Hadoop Infrastructure.
  • Good experience on Design, configure and manage the backup and disaster recovery for Hadoop data.
  • In-depth knowledge of modifications required in static IP (interfaces), hosts, setting up password-less SSH and Hadoop configuration for Cluster setup and maintenance.
  • Have experience in AWS EC2 management console, S3 buckets, EBS, Glacier, VPC, ELB, IAM, Cloud Watch, System manager. Creating, launching and upscaling servers using AWS managent console.
  • Have basic knowledge on AWS EMR, GIT, Python.
  • Managed Users and Groups using IAM
  • Have experience in creating and deploying SSL, CSR and CA certs.
  • Experienced using Sqoop to import data into HDFS from RDBMS and vice-versa.
  • Experience in understanding the security requirements for Hadoop and integrating with Kerberos authentication infrastructure- KDC server setup, creating realm /domain.
  • Experience in Administering, Installation, configuration, troubleshooting, Security, Backup, Performance Monitoring and Fine-tuning of Redhat Linux and CentOS.
  • Have Experience in installing and configuring Rstudio, Jupyterhub, SAS Foundation Server, H2O.
  • Experience in upgrading kernel of the servers.
  • Experience in supporting systems with 24X7 availability and monitoring.

TECHNOLOGY SKILLS:

Big Data Technologies: HDFS, Hive, Map Reduce, Kafka, Pig, Spark, HUE, Scoop, Flume, Zookeeper, Oozie, HBase, Hive Server2, Storm, CDH 5.3, CDH 5.4, CDH 5.12.1

Scripting Languages: Shell Scripting, Python, Bash.

Databases: Oracle 11g, MySQL, Hbase, Cassandra, MongoDB

Networks: HTTP, HTTPS, FTP, UDP, TCP/TP, SNMP, SMTP

Monitoring Tools: Cloudera Manager, Solr, Ambari, Nagios, Ganglia

Application Servers: Apache Tomcat, Weblogic Server, Websphere

Security: Kerberos

Analytic Tools: RStudio, Jupyterhub, SAS, H2O, Tableau

EXPERIENCE:

Confidential, Schaumburg, IL

Hadoop Administrator

Responsibilities:

  • Administering Hadoop Cloudera production cluster (CDH 5.10.2) with 120+ nodes.
  • Creating and Launching EC2 instances in AWS.
  • Used AWS S3 buckets to store files.
  • Scaling up the servers in AWS.
  • Creating and Deploying CSR, SSLS certs.
  • Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • Kernel and OS patching of the servers.
  • Created 30+ Hadoop nodes in production cluster to scale up the cluster storage capacity and performance.
  • Managing Data in Hadoop.
  • Monitoring workload, job performance and capacity planning and debugged failed jobs using Cloudera Manager.
  • Installed and configured Hiveserver2, Hue Server, Spark 2 on the Hadoop Cloudera cluster.
  • Installed and configured Applications like SAS Foundation, Rstudio, Tableau, H2O, Jupyterhub Server on Hadoop Cluster.
  • Examining job fails and trouble shooting.
  • Adding and Removing users from SAS Metadata.
  • Installed multiple versions of R.
  • Upgraded RStudio and Tableau Servers.
  • Installed R and Jupyterhub pacakages.
  • Working on Sevicenow incident tickets.
  • Troubleshooting issues related to SAS Foundation, Rstudio, Tableau, H2o, Jupyterhub Server.

Environment: Cloudera Hadoop, Rstudio, H2O, Jupyterhub, Confidential, Tableau, Windows 2012R2, Sqoop, Hive, Oozie, SAS, Hive Server2(HS2), Spark, Unix Shell Scripts, ZooKeeper, SQL, Map Reduce, Pig, AWS S3, EC2, EBS, Glacier, AWS system manager.

Confidential, Tampa, Florida

Hadoop Administrator

Responsibilities:

  • Involved in upgradation process of the Hadoop cluster from CDH4 to CDH5.
  • Worked on installing cluster, commissioning& decommissioning of datanode, namenode recovery, capacity planning, and slots configuration.
  • Installed and configured Flume&Oozie on the Hadoop cluster and Managed, Defined and Scheduled Jobs on a Hadoop cluster.
  • Responsible for Cluster maintenance, Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
  • Adding & installation of new components and removal of them through Cloudera Manager.
  • Responsible for implementation and ongoing administration of Hadoop infrastructure.
  • Installed, Configured & Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
  • Involved in implementing security on Hortonworks Hadoop Cluster using with Kerberos by working along with operations team to move non secured cluster to secured cluster.
  • Bootstrapping instances using Chef and integrating with auto scaling.
  • Monitored workload, job performance and capacity planning.
  • Managing and reviewing Hadoop log files and debugging failed jobs.
  • Managed the Hadoop infrastructure with Cloudera Manager and Ambari.
  • Supported cluster maintenance, Backup and recovery for production cluster.
  • Backed up data on regular basis to a remote cluster using distcp.
  • Setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
  • Monitored Hadoop cluster connectivity and security and also involved in management and monitoringHadoop log files.
  • Assembled Puppet Master, Agent and Database servers on Red Hat Enterprise Linux Platforms.

Environment: Hortonworks Hadoop, Cassandra, Flat files, Oracle 11g/10g, mySQL, Windows NT,Sqoop, Hive, Oozie, Cloudera, SAS, SPSS, Unix Shell Scripts, Zoo Keeper, SQL, Map Reduce, Pig.

Confidential, Santa Clara, CA

Hadoop Administrator

Responsibilities:

  • Installed and configured multi-node fully distributed Hadoop Clusters on Cloudera Platform.
  • Collaborate with Cloudera Professional Services and Premier Support Engineering to troubleshootand debug Cloudera Data Platform (CDH) issues.
  • Extensive experience in Capacity Planning, Backup and recovery methods, setting up DR cluster, growthanalysis, commissioning and decommissioning data nodes based on growth estimation, cluster maintenance on daily basis.
  • Implemented Capacity Scheduler for multiple tenants, setup Kerberos security, applied data retention and data governance rules.
  • Involved in installing Hadoop Eco System Components.
  • Configured NameNode High Availability and Resource Manager High Availability in CDH.
  • Involved in Hadoop Cluster Administration that includes adding and removing cluster nodes, clusterCapacity planning, performance tuning.
  • Written Unix Shell scripts and used Resource Manager REST API’s to monitor the jobs which are running on the cluster for long time to alert the support & development teams.
  • Configured Apache Ranger to secure Access to HDFS & Hive Databases and implemented various
  • Configured MySQL databases for storing Hive Metastore data and Oozie.

Environment: Cloudera Hadoop, HDFS, MapReduce, HIVE, PIG, Sqoop, HBase, Oozie, PL/SQL & ZookeeperSQL plus, Oracle Enterprise Manager (OEM), Shell Scripting, Golden Gate, Redhat.

Confidential

Hadoop Administrator

Responsibilities:

  • Involved in start to end process of hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
  • Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring, Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
  • Monitoring systems and services, architecture design and implementation of hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Importing and exporting data into HDFS using Sqoop.
  • Experienced in define being job flows with Oozie.
  • Loading log data directly into HDFS using Flume.
  • Experienced in managing and reviewing Hadoop log files.
  • Installation of various Hadoop Ecosystems and Hadoop Daemons.
  • Installation and configuration of Sqoop and Flume, Hbase
  • Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes.
  • Communicate and escalate issues appropriately.
  • As admin followed standard Back up policies to make sure the high availability of cluster.
  • Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
  • Documented the systems processes and procedures for future s.
  • Worked with systems engineering team to plan and deploy new hadoop environments and expand existinghadoop clusters.
  • Monitored multiple hadoop clusters environments using Ganglia and Nagios. Monitored workload, jobperformance and capacity planning using Cloudera Manager.
  • Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the hadoop cluster.
  • Involved in Installing and configuring Kerberos for the authentication of users and hadoop daemons. .

Environment: Hadoop, HDFS, Hive, Sqoop, Flume, Zookeeper and Hbase, Big Data Cloudera CDH Apache Hadoop, Toad, SQL plus, Shell Scripting, Redhat/SuseLinix, EM Cloud Control.

We'd love your feedback!