We provide IT Staff Augmentation Services!

Hadoop Admin Resume

3.00/5 (Submit Your Rating)

Mountain View, CA

PROFESSIONAL SUMMARY:

  • Overall 5+ Years of professional IT experience which includes around 3 years of hands on experience in Hadoop Administration using Cloud era (CDH) and Horton works (HDP) Distributions on large distributed clusters.
  • Hands on Experience in Installing, Configuring and using Hadoop Eco System Components like HDFS, Hadoop Map Reduce, Yarn, Zookeeper, Sentry, Sqoop, Flume, Hive, HBase, Pig, Oozie.
  • Hands on experience on configuring a Hadoop cluster in a professional environment and Amazon Web Service (AWS) using an EC2 instance.
  • Good working experience on Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Experience in Importing and Exporting Data between different Database Tables like MySQL, Oracle and HDFS using Sqoop.
  • Had good working experience on Hadoop architecture, HDFS, Map Reduce and other components in the Cloudera - Hadoop eco system.
  • Experience in writing scripts for Automation.
  • Experience in Benchmarking, Backup and Disaster Recovery of Name node Metadata.
  • Experience in performing minor and major Upgrades of Hadoop Cluster (Hortonworks Data Platform 2.0 to 2.1.)
  • Experience with multiple Hadoop distribution s like Apache, Cloudera and Hortonworks.
  • Experience in securing Hadoop clusters using Kerberos and Sentry.
  • Experience with distributed computation tools such as Apache Spark Hadoop.
  • Experience as Deployment Engineer and System Administrator on Linux (Centos, Ubuntu, Red Hat).
  • Experience working with Deployment tools such as Puppet/Ansible.
  • Well versed in installing, configuring and tuning Hadoop distributions: Cloudera, Hortonworks on Linux systems.
  • Experience with Red hat Packet Manager packaging and RPM deployments.
  • Experience with Nagios and writing plugins for Nagios to monitor Hadoop clusters.
  • Experience in supporting users to debug their job failures.
  • Experience with Source Code Management tools and proficient in GIT, SVN.

TECHNICAL SKILLS:

Hadoop Ecosystem: Hive, Pig, Sqoop, Flume, Impala, Oozie, Sentry, Spark, Zookeeper, Apache Kafka, Storm

Hadoop Management: Ambari, Cloudera Manager

Hadoop Paradigms: Map Reduce, Yarn, High Availability

Other Relevant Tools: SVN, Tableau, JIRA, QC, MS Office Suite

RDBMS: Oracle 10g/11g/12c, MS SQL Server 2000/2003/2008 R2/2012, DB2, Teradata, Netezza, MySQL.,Mongo DB

Programming Languages: Linux, Unix Shell scripting, JAVA, SQL

Monitoring and Alerting: Nagios, Ganglia, Ambari Metrics

Operating Systems: Centos 5,6, Red hat 6, Ubuntu Server 14.04(Trusty), Windows Server 2012

PROFESSIONAL EXPERIENCE:

Confidential, Mountain View, CA

Hadoop Admin

Responsibilities:

  • Responsible for running, managing and operating scalable,highly available and falut tolerant systems on AWS.
  • Experience in enabling & administering AWS cloud infrastructure for various applications including (Analytics, Enterprise Applications, etc.)
  • Responsible for day-to-day activities which includes HDFS support and maintenance, Cluster maintenance, creation/removal of nodes, Cluster Monitoring/ Troubleshooting, Manage and review Hadoop log files, Backup and restoring, capacity planning
  • Worked with Hadoop developers and operating system admins in designing scalable supportable infrastructure for Hadoop
  • Responsible for Operating system and Hadoop Cluster monitoring using tools like Nagios, Ganglia, Ambari Metrics.
  • HA Implementation of Namenode replication to avoid single point of failure.
  • Involved in troubleshooting issues on the Hadoop ecosystem, understanding of systems capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks
  • Involved in setup, configuration and management of security for Hadoop clusters using Kerberos and integration with LDAP/AD at an Enterprise level
  • Developed design for data migration form one cluster to another cluster using DISTCP.
  • Responsible for scheduling jobs in Hadoop using FIFO, Fair scheduler and Capacity scheduler
  • Possess good Linux and Hadoop System Administration skills, networking, shell scripting and familiarity with open source configuration management and deployment tools such as Puppet or Ansible.
  • Built data platforms, pipelines, storage systmes using the Apache Kafka, Apache Storm and search technologies such as Elastic search.

Environment: Hadoop distributions (HDP 2.2), Red hat Linux 6.x,7.x Solaris 11, AWS (EC2) Shell Scripts, Nagios, Ganglia monitoring, Kerberos, Shell scripting, Puppet, Ansible, Python Scripting, Java,Hive, Pig, Sqoop, Flume, HBase, Zookeeper, Oozie, YARN, Ambari(2.1) etc.

Confidential, Jersey City, NJ

Hadoop Admin

Responsibilities:

  • Install, Upgrade and maintain the Hadoop Clusters using Cloudera Manager.
  • Design and Configure the Cluster with the services required (Sentry, Hive server2, Kerberos, HDFS, Hue, Hive, Zookeeper).
  • Flume configuration for data transfer from Webservers to Hadoop cluster.
  • Sentry configuration for appropriate user permissions accessing Hive server2/beeline.
  • Design and maintain the Name node and Data nodes with appropriate processing capacity and disk space.
  • Developed design for data migration form one cluster to another cluster using Distcp.
  • Performing benchmark test on Hadoop clusters and tweak the solution based on test results.
  • Integrating Hadoop clusters to use AD/LDAP for access management.
  • Monitoring and supporting clusters and resolving any issues to maintain uptime of the clusters.
  • Writing scripts to automate the data loading to cluster.
  • Automated Cluster installations, Validations using scripts.
  • Responsible for troubleshooting issues in the execution of jobs by inspecting and reviewing log files.
  • Worked collaboratively with all levels of business stakeholders to architect, implement and test Big Data based analytical solution from disparate sources.
  • Monitored multiple hadoop clusters environments using Ganglia and Nagios, monitored workload, job performance and capacity planning using Ambari.
  • Experience with professional software engineering practices and best practices for the full software development life cycle including coding standards, code reviews, source control management and build processes

Environment: Hadoop, Hive, HBase, Map Reduce, HDFS, Pig, Nagios, Ganglia, Java (JDK 1.6), Hadoop Distribution of Cloudera (CDH 5, CM 5.2),Cassandra, IBM Data Stage 8.1,PL/SQL, SQL*PLUS, Toad 9.6, UNIX Shell Scripting

Confidential

System Administrator

Responsibilities:

  • Installed, configured and administered RHEL 5/6 servers.
  • Managed file space and created logical volumes, extended file systems using LVM.
  • Performed daily maintenance of servers and tuned system for optimum performance by turning off unwanted peripheral and vulnerable service.
  • Managed RPM Package for Linux distributions
  • Monitored system performance using TOP, FREE, VMSTAT & IOSTAT.
  • Set up user and group login ID's, password, ACL file permissions, and assigned user and group quota
  • Configured networking including TCP/IP and troubleshooting.
  • Monitored scheduled jobs, workflows, and related to day-to-day system administration.
  • Respond to tickets through ticketing systems.
  • Installing and maintaining the Linux servers.
  • Installed Cent OS using Pre-Execution environment boot and Kick-start method on multiple servers.
  • Monitoring System Metrics and logs for any problems.
  • Running cron-tab to back up data.
  • Applied Operating System updates, patches and configuration changes.
  • Adding, removing, or updating user account information, resetting passwords, etc.
  • Using Java JDBC to load data into MySQL.
  • Maintaining the MySQL server and Authentication to required users for databases.

We'd love your feedback!