We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

3.00/5 (Submit Your Rating)

Bloomington, IL

SUMMARY:

  • Over all 7+ years of professional Information Technology experience in Hadoop and Linux Administration activities such as installation, configuration and maintenance of systems/clusters.
  • Having extensive experience in Linux Administration & Big Data Technologies as a Hadoop Administration.
  • Hands on experience in Hadoop Clusters using Hortonworks (HDP), Cloudera (CDH3, CDH4), oracle big data and Yarn distributions platforms.
  • Possessing skills in Apache Hadoop, Map - Reduce, Pig, Impala, Hive, Hbase, Zookeeper, Sqoop, Flume, OOZIE, Kafka, storm, Spark, Java Script, and J2EE.
  • Experience in deploying and managing the multi-node development and production Hadoop cluster with different Hadoop components (Hive, Pig, Sqoop, Oozie, Flume, HCatalog, Hbase, Zookeeper) using Hortonworks Ambari.
  • Good experience in creating various database objects like tables, stored procedures, functions, and triggers using SQL, PL/SQL and DB2.
  • Used Apache Falcon to support Data Retention policies for HIVE/HDFS.
  • Experience in Configuring Name-node High availability and Name-node Federation and depth knowledge on Zookeeper for cluster coordination services.
  • Experience on Design, configure and manage the backup and disaster recovery for Hadoop data.
  • Experience in administering Tableau and Green Plum databases instances in various environments.
  • Experience in administration of Kafka and Flume streaming using Cloudera Distribution.
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Extensive knowledge in Tableau on Enterprise Environment and Tableau administration experience including technical support, troubleshooting, reporting and monitoring of system usage.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
  • Worked on NoSQL databases including Hbase, Cassandra and MongoDB.
  • Designing and implementing security for Hadoop cluster with Kerberos secure authentication.
  • Hands on experience on Nagios and Ganglia tool for cluster monitoring system.
  • Experience in scheduling all Hadoop/Hive/Sqoop/Hbase jobs using Oozie.
  • Knowledge of Data Ware Housing concepts and Cognos 8 BI Suit and Business Objects.
  • Experience in HDFS data storage and support for running map-reduce jobs.
  • Experience in Installing Firmware Upgrades, kernel patches, systems configuration, performance tuning on Unix/Linux systems.
  • Expert in Linux Performance monitoring, kernel tuning, Load balancing, health checks and maintaining compliance with specifications.
  • Hands on experience in Zookeeper and ZKFC in managing and configuring in Name Node failure scenarios.
  • Team Player with good communication and interpersonal skills and goal oriented approach to problem solving issues.

TECHNICAL SKILLS:

Operating System: RedHat, CentOS, Ubuntu, Solaris, Windows 2008/08R2

Hardware: Sun Ultra Enterprise Servers (E3500, E4500), SPARC server 1000, SPARC server 20 Enterprise Servers

Languages: C++, Core Java and JDK 7/8

Web Languages: HTML, CSS, and XML

Hadoop Distribution: Cloudera and HortonWorks

Ecosystem Hadoop: MapReduce, YARN, HDFS, Sqoop, Hive, Pig, Hbase, Sqoop, Flume, and Oozie.

Tools: JIRA, PuTTy, WinSCP, FileZilla.

Database: HBase, RDBMS Sybase, Oracle 7.x/8.0/9i, MySQL, SQL.

Protocols: TCP/IP, FTP, SSH, SFTP, SCP, SSL, ARP, DHCP, TFTP, RARP, PPP and POP3

Scripting: Bash Shell scripts, Python, Perl

Cloud Technologies: AWS, Azure

PROFESSIONAL EXPERIENCE:

Confidential, Bloomington, IL

Hadoop Administrator

Responsibilities:

  • Configuring hosts as edge nodes with the desired file systems and adding them to the cluster.
  • Involving in the weekly releases and maintaining the sync between different environments such as Production, Disaster Recovery and Pre-Production
  • Deleting the users or adding the users as of the request in the Hue, Data Robot and Trifacta.
  • Actively involved in the OS Patching activities, Cloudera Upgrade and other maintenance activities.
  • Actively involved in the palling and implementation of the Rack Awareness in the different environments.
  • Involved in migrating the MySQL database to Oracle database and PSQL database to Oracle database.
  • Performed Requirement Analysis, Planning, Architecture Design and Installation of the Hadoop cluster
  • Acted as a point of contact between the vendor and my team on different issue,
  • Actively involved in the planning and implementation on the Load Balancer with a single GTM and multiple LTM’s
  • Involved in writing an automation script for different applications and different purposes such as installing the applications.
  • Involved in configuring the LDAP on different application for a secure login.
  • Actively involved in the trouble shooting the users issue on 24/7 basis.
  • Implemented strategy to upgrade entire cluster nodes OS from RHEL5 to RHEL6 and ensured cluster remains up and running
  • Involved in Cluster Level Security, Security of perimeter (Authentication- Cloudera Manager, Active directory and Kerberoes) Access (Authorization and permissions- Sentry) Visibility (Audit and Lineage - Navigator) Data ( Data Encryption at Rest)
  • Worked on YARN capacity scheduler by creating queues to allocate resource guarantee to specific groups
  • Worked on installing production cluster, commissioning & decommissioning of Data Nodes, Name Node recovery, capacity planning, and slots configuration.

Environment: Hadoop, MapReduce, Hbase, Tez, Hive, Pig, Sqoop, HDP 2.6, HDFS, Talend.

Confidential, Saint Louis, MO

Hadoop Administrator

Responsibilities:

  • Experience in supporting and managing Hadoop Clusters using Hortonworks distributions.
  • Interacting with Hortonworks support and log the issues in Hortonworks portal and fixing them as per the recommendations
  • Scheduled several time based Oozie workflow by developing Python scripts.
  • Implemented custom interceptors for flume to filter data and defined channel selectors to multiplex the data into different sinks
  • Created instances in AWS as well as migrated data to AWS from data Center using snowball and AWS migration service
  • Extending the functionality of Hive and Pig with custom UDF s and UDAF's on Java
  • Involved in extracting the data from various sources into Hadoop HDFS for processing
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop
  • Creating and truncating HBase tables in hue and taking backup of submitter ID(s)
  • Responsible for building scalable distributed data solutions using Hadoop
  • Commissioned and Decommissioned nodes on Hortonworks Hadoop cluster on Red hat LINUX
  • Worked with BI teams in generating the reports and designing ETL workflows on Tableau
  • Configured, supported and maintained all network, firewall, storage, load balancers, operating systems, and software in AWS EC2 and created detailed AWS Security groups which behaved as virtual firewalls that controlled the traffic allowed reaching one or more AWS EC2 instances.
  • Hands on experience in Hadoop administration and support activities for installations and configuring Apache Big Data Tools and Hadoop clusters using Ambari.
  • Strong capability to utilize Unix shell programming methods, able to diagnose and resolve complex configuration issues, ability to adapt Unix domain for Hadoop Tools.
  • Maintains the EC2 (Elastic Computing Cloud) and RDS (Relational Database Services) in amazon web services.
  • Created Hive External tables and loaded the data in to tables and query data using HQL.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Ambari, Hortonworks, Sqoop, Flume, Oozie, CDH3, MongoDB, Cassandra, HBase, Java, Eclipse, Oracle and Unix/Linux.

Confidential, Palo Alto, CA

Hadoop/Big data Administrator

Responsibilities:

  • Handle the installation and configuration of a Hadoop cluster using Hortonworks Distribution.
  • Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase.
  • Handle the data exchange between HDFS and different Web Applications and databases using Flume and Sqoop.
  • Monitor the data streaming between web sources and HDFS.
  • Worked in Kerberos and how it interacts with Hadoop and LDAP.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
  • Experience in Continuous Integration and expertise in Jenkins and Hudson tools.
  • Changes to the configuration properties of the cluster based on volume of the data being processed and performance of the cluster.
  • Set up automated processes to analyze the System and Hadoop log files for predefined errors and send alerts to appropriate groups.
  • Experience in architecting, designing, installation, configuration and management of Apache Hadoop, Hortonworks Distribution.
  • Responsible for doing capacity planning based on the data size requirements provided by end-clients.
  • Worked in Unix commands and Shell Scripting.
  • Experience in doing performance tuning based on the inputs received from the currently running jobs.
  • Used Apache Oozie for scheduling and managing the Hadoop Jobs. Knowledge on HCatalog for Hadoop based storage management.
  • Worked in core competencies in Java, HTTP, XML and JSON.
  • Worked on spark it's a fast and general -purpose clustering computing system.
  • Worked on Storm its distributed real-time computation system provides a set of general primitives for Commission and decommissions the Data nodes from cluster in case of problems.
  • Experience in a Web-based Git repository hosting service, which offers all the distributed revision control and source code management (SCM) functionality of Git as well as adding its own features in Git Hub.
  • Experience in Hortonworks Distribution Platform (HDP) cluster installation and configuration.
  • Worked in statistics collection and table maintenance on MPP platforms.
  • Worked on large sets of structured, semi-structured and unstructured data.
  • Use of Sqoop to import and export data from HDFS to RDBMS and vice-versa.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.

Environment: Hadoop, MapReduce, Hive, HDFS, PIG, Sqoop, Oozie, Hortonworks, Flume, HBase, ZooKeeper, Oracle, NoSQL and Unix/Linux.

Confidential

Linux Administrator

Responsibilities:

  • Installing and upgradingRed hat Linux and Solaris 8/ & SPARC on Servers like HP DL 380 G3, 4 and 5 & Dell Power Edge servers.
  • Experience in LDOM's and Creating sparse root and whole root zones and administered the zones for Web, Application and Database servers and worked on SMF on Solaris 10.
  • Experience working with HP LVM and Red hat LVM.
  • Experience in implementing P2P and P2V migrations.
  • Involved in Installing and configuring Centos & SUSE 11 & 12 servers on HP x86 servers.
  • Implemented HA using Red Hat Cluster and VERITAS Cluster Server 5.0 for Web Logic agent.
  • Used Sun Volume Manager for Solaris and LVM on Linux & Solaris to create volumes with layouts like RAID 1, 5, 10, 51.
  • Performed performance analysis using tools like prstat, mpstat, iostat, sar, vmstat, truss, Dtrace.
  • Experience working on LDAP user accounts and configuring Ldap on client machines.
  • Upgraded Clear-Case from 4.2 to 6.x running on Linux (Centos &Red hat)
  • Worked on patch management tools like Sun Update Manager.
  • Expertises in Build, Install, load and configure boxes.
  • Worked with the team members to create, execute and implement the plans.
  • Experience in Installation, Configuration and Troubleshooting of Tivoli Storage Manager (TSM).
  • Remediating failed backups, Take manual incremental backups of failing servers.
  • Upgrading TSM from 5.1.x to 5.3.x.Worked on HMC Configuration and management of HMC Console which included up gradation, micro partitioning
  • Provide redundancy with HBA card, Ether channel configuration and network devices.
  • Coordinating with application and database team for troubleshooting the application or Database outages.
  • Used HP Service center and change management system for ticketing.
  • Worked on the administration of the Web Logic 9, JBoss 4.2.2 servers including installation and deployments.
  • Worked on F5 load balancers to load balance and reverse proxy Web Logic Servers.
  • Installed, configured, troubleshoot and maintain Linux Servers and Apache Web server, configuration and maintenance of security and scheduling backups, submitting various types of cron jobs.
  • Installations of HP Open view, monitoring tool, in more than 300 servers and worked with monitoring tools such as Nagios and HP Open view.
  • Creation of VMs, cloning and migrations of the VMs on VMware vSphere 4.0/4.1

Environment: Solaris 8/9/10, Veritas Volume Manager, web servers, LDAP directory, Active Directory, BEA Web logic servers, SAN Switches, Apache, Tomcat servers, Web Sphere application server.

We'd love your feedback!