We provide IT Staff Augmentation Services!

Junior Intern/hadoop Administrator Resume

2.00/5 (Submit Your Rating)

TX

SUMMARY

  • Around 5+ years of professional experience spanning around Big Data analytics as Hadoop/Big Data Administrator.
  • Experience in all the phases of Data warehouse life cycle involving Requirement Analysis, Design, Coding, Testing, and Deployment.
  • Experienced in installing, configuring, and administrating Hadoop cluster of major distributions.
  • Working experience on Hortonworks (HDP) and Cloudera distribution.
  • Excellent experience in supporting production clusters, troubleshooting and handling critical issues.
  • Coordinated with technical teams for installation of Hadoop and third party related applications on systems.
  • Performed upgrades, patches and bug fixes in HDP and CDH clusters.
  • Experience on building dashboards for operations from FS Image to project existing and forecasted data growth.
  • Working experience with large scale Hadoop environments build and support including design, configuration, installation, performance tuning and monitoring.
  • Excellent experience in Pig, Hive, Sqoop, Hbase, Yarn, Oozie and Map/Reduce jobs to support distributed data processing and process large data sets utilizing the Hadoop cluster.
  • Experience in architecting, designing, installation, configuration and management of Apache Hadoop Clusters, Mapr, Hortonworks & Cloudera Hadoop Distribution.
  • Experience in managing the Hadoop infrastructure with Cloudera Manager.
  • Good Understanding in Kerberos and how it interacts with Hadoop and LDAP.
  • Practical knowledge on functionalities of every Hadoop daemon, interaction between them, resource utilizations and dynamic tuning to make cluster available and efficient.
  • Experience in understanding and managing Hadoop Log Files.
  • Upgrading Hortonworks clusters with latest release packs.
  • Experience in Adding and removing the nodes in Hadoop Cluster.
  • Experience in extracting the data from RDBMS into HDFS Sqoop.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including HDFS, Hive, HBase, Flume, Oozie and Sqoop.
  • Good understanding of No SQL databases such as HBase
  • Experience in analyzing data in HDFS through Map Reduce, Hive and Pig.
  • Experience in Data Analysis, Data Cleansing (Scrubbing), Data Validation and Verification, Data Conversion, Data Migrations and Data Mining.
  • Excellent interpersonal, communication, documentation and presentation skills.

TECHNICAL SKILLS

Hadoop Ecosystem: HDFS, Yarn, Map - Reduce, Hive, Hue, Impala, Sqoop, Oozie, Flume, Zookeeper.

NoSql Databases: HBase, PIG and Bash

Querying Languages: PIG, Hive, SQL, PL/SQL

Database: Oracle 11g/8i/9i, MY SQL, MS SQL server. Web Dev. Technologies: HTML, XML, TeraData.

Scripting Languages: Shell Scripting.

Operating Systems: Linux, Unix, Windows, Mac

Protocols: Protocols TCP/IP, UDP, HTTP and HTTPS.

Cluster Management Tools: Cloudera Manager and HDP Ambari, HortonWorks,Apache Hadoop

Virtualization technologies: VMware, VMbox, VSphere, Citrix Xen-Server.

PROFESSIONAL EXPERIENCE

Confidential, TX

Junior Intern/Hadoop Administrator

Responsibilities:

  • Successfully upgraded the Hadoop cluster from CDH 4.7 to CDH 5.0.0.
  • Maintained Hortonworks cluster with HDP Stack 2.4.2 managed by Ambari 2.2.Client wanted to migrate from In-Premise cluster to the Amazon Web services Cloud (AWS), hence
  • Built a Production and QA Cluster with the latest distribution of Hortonworks - HDP stack 2.6.1 managed by Ambari 2.5.1 on AWS Cloud.
  • The Production and QA AWS Cluster both are 8 node cluster.
  • Various patch upgrades happen on the Cluster for different services.
  • Providing immediate support to users for various Hadoop related issues.
  • User management, involving user creation, granting permission for the user to various tables and database, giving group permissions.
  • Performed HDP stack upgrade.
  • Disabled save in Hive views.
  • Ranger security enabled on all the Clusters.
  • Working closely with the Development team, providing support, fine tuning the cluster for various use cases, and resolving day to day issues in the cluster, with respect to the services health.
  • Working with the Development team, optimizing the hive queries using bucketing, partitions and Joins concept.

Environment: Hadoop, HDFS, Yarn, Pig, Hive, Sqoop, Oozie, Control-M, HBase, Shell Scripting, AWS, Ubuntu, Linux Red Hat.

Confidential

Software Engineering Analyst/Hadoop Administrator

Responsibilities:

  • Collaborated with different teams for Cluster Planning, Hardware requirement, server configurations, network equipment’s to implement CDH 3.6.
  • Responsible for setting up and configuring MYSQL database for the cluster.
  • Responsible for ensuring adherence to security policies, including regular security updates to UNIX and Linux systems.
  • Responsible for creating new users to the Hadoop cluster and providing access to the datasets
  • Administration and maintenance of the cluster using OpsCenter, Devcenter, Linux, Node tool etc.
  • Responsible for Cluster maintenance, Commissioning and Decommissioning cluster nodes, Cluster Monitoring and Troubleshooting, Manage and review data backups, Manage and review Hadoop log files.
  • Resolving tickets submitted by users, troubleshoot the documented errors , resolving the errors.
  • Designed and developed an API for rider's preferences with all CRUD capabilities.
  • Created views for ambary for different components in Hortonworks
  • File system management and monitoring and responsible for giving presentations about new ecosystems to be implemented in the cluster with different teams and managers.
  • Managed Hortonworks Ranger Policies to restrict access for multiple users.
  • Worked on developing scripts for performing Benchmarking Techniques.
  • Adding new data nodes when needed and running balancer.
  • Implemented Kerberos for authenticating all the services in Hadoop Cluster.
  • Dumped the data from MySQL database to HDFS and vice-versa using Sqoop.
  • Dumped the data from one cluster to another cluster by using Distcp.
  • Continuous monitoring and managing the Hadoop cluster through Ganglia and Nagios.
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs, which run independently with time and data availability.
  • Upgraded the Hadoop cluster from cdh3 to cdh4 using Cloudera distribution packages.
  • Upgraded the Done stress and performance testing, benchmark for the cluster.
  • Deployed Network file system (NFS) for NameNode Metadata backup.
  • Debug and solve the major issues with Cloudera manager by interacting with the Infrastructure team from Cloudera.
  • Provide weekly status reports to project managers describing accomplishments.
  • Follow department quality assurance guidelines and processes.
  • Monitoring Red Hat Enterprise 5X in a 100+ server farm using monitoring tools like Nagios and virtualization tools like VMbox, VSphere and VCenter Server.
  • Worked with network and system engineers to define optimum network configurations.
  • Performed Enterprise level Hadoop upgrade on existing cluster, applied patches and did version upgrade.
  • Troubleshoot the Hortonworks Data Platform in multiple types of environments and take ownership of problem isolation and resolution, and bug reporting.
  • Provided POC for test and QA cluster using HDP 1.7 using NOSQL DB using HBASE.
  • Built lab Hadoop cluster based on multiple virtual machines for multiple testing reasons.

Environment: HortonWork HDP 2.2, 2.4; Cloudera 3, 4; Hive, Hue, Ambari, Sqoop, RHEL 5.4, 5.5; Linux, Redhat, CDH.

Confidential

Junior Linux/ Systems Administrator

Responsibilities:

  • Installing, configuring and updating Red Hat Enterprise Linux 5.0.
  • Creating System Disk Partition, mirroring root disk drive, configuring device groups in UNIX and Linux environment
  • Performed scheduled backup and necessary restoration.
  • Installation of Nagios monitoring tool, setting of different alert level to different parameters (System, Memory and Hard disk).
  • Installation of MYSQL (4.0.30) databases on Red hat Linux.
  • Working with VERITAS Volume Manager 3.5 and Logical Volume Manager for file system management, data backup and recovery.
  • Apache Server Administration with virtual hosting.
  • Proxy server configuration.
  • User administration which included creating backup account for new users and deleting account for the retired or deleted users.
  • Installed and Configured SSH Gate for Remote and Secured Connection.
  • Automate administration tasks through use of scripting and Job Scheduling using CRON
  • Installed and managed packages using YUM and RPM.

Environment: Redhat 5.0; Linux, Unix, VERITAS, MySQL, YUM, CRON, RPM

We'd love your feedback!