Sr.hadoop Administrator Resume
Indianapolis, IN
PROFESSIONAL SUMMARY:
- Over 8+ years of professional Information Technology experience in Hadoop and Linux Administration activities such as installation, configuration and maintenance of systems/clusters.
- Having 3 years of experience in Linux Administration & Big Data Technologies and 5+ years of experience into Hadoop administration.
- Hands on experience on Hadoop Clusters using Hortonworks (HDP), Cloudera (CDH3, CDH4), oracle big data and Yarn distributions platforms.
- Possessing skills in Apache Hadoop , Map - Reduce , Pig , Impala, Hive , Platfora, Hbase , Zookeeper , Sqoop , Flume, OOZIE , Kafka, storm, Spark, Datameer, Java Script, and J2EE.
- Experience in deploying and managing the multi-node development and production Hadoop cluster with different Hadoop components ( Hive , Pig , Sqoop, Oozie , Flume, HCatalog, Hbase , Zookeeper ) using Hortonworks Ambari.
- Good experience in creating various database objects like tables, stored procedures, functions, and triggers using SQL, PL/SQL, and DB2.
- Experience in Configuring Name-node High availability and Name-node Federation and depth knowledge on Zookeeper for cluster coordination services.
- Experience on Design, configure and manage the backup and disaster recovery for Hadoop data.
- Experience in administering Tableau and Green Plum databases instances in various environments.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Extensive knowledge in Tableau on Enterprise Environment and Tableau administration experience including technical support, troubleshooting, reporting and monitoring of system usage.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
- Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
- Worked on NoSQL databases including Hbase , Cassandra and MongoDB .
- Designing and implementing security for Hadoop cluster with Kerberos secure authentication.
- Hands on experience on Nagios and Ganglia tool for cluster monitoring system.
- Experience in scheduling all Hadoop / Hive / Sqoop / Hbase jobs using Oozie .
- Knowledge of Data Ware Housing concepts and Cognos 8 BI Suit and Business Objects.
- Experience in HDFS data storage and support for running map-reduce jobs .
- Experience in Installing Firmware Upgrades, kernel patches, systems configuration, performance tuning on Unix/Linux systems .
- Expert in Linux Performance monitoring, kernel tuning, Load balancing, health checks and maintaining compliance with specifications.
- Hands on experience in Zookeeper and ZKFC in managing and configuring in Namenode failure scenarios.
- Team Player with good communication and interpersonal skills and also goal oriented approach to problem solving issues.
TECHNICAL SKILLS:
Big DataTechnologies: Hadoop, HDFS, MapReduce, Yarn, Hive, Pig, Sqoop, Hbase, Flume, Oozie,Spark,Zookeeper.
Hadoop Platforms: Hortonworks and Cloudera, Apache Hadoop
Networking Concepts: OSI Model, TCP/IP, UDP, IPV4, Subnetting, DHCP & DNS
Programming Languages: PIG LATIN, UNIX shell scripting and Bash.
Operating Systems: Linux (CentOS, Ubuntu, Red Hat), Windows, UNIX and Mac OS-X
Database/ETL: Oracle, Cassandra, DB2, MS-SQL Server, MySQL, MS-Access, Hbase, MongoDB,Informatica, Teradata.
XML Languages: XML, DTD, XML Schema, XPath.
Monitoring and Alerting: Nagios, Ganglia, Cloudera Manager, Ambari.
PROFESSIONAL EXPERIENCE:
Confidential - Indianapolis, IN
Sr.Hadoop Administrator
Responsibilities:
- As an admin worked in Huge Cluster on maintaining over 100 nodes with High availability environment using HortonworksAmbarimanager.
- Involved in Installation and configuration, Hadoop Cluster and Maintenance, Cluster Monitoring, Troubleshooting and Transform data from RDBMS to HDFS and followed proper backup& Recovery strategies.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS.
- Provide Business Intelligence support using Tableau for implementing effective Business dashboards & visualizations of data.
- Configuring, implementing and supporting High Availability (Replication) with Load balancing (Sharding) cluster of Mongo DB having Terabytes of data.
- Hadoop cluster monitoring and troubleshooting Hive, datameer, Platfora and flume.
- Experience with securing Hadoop clusters including Kerberos KDC installation, OpenLDAP installation, data transport encryption with TLS, and data-at-rest encryption with Hortonworks Navigator Encrypt.
- Used Cassandra in multiple virtual and physical data centers to ensure the system was highly redundant and scalable.
- Exported the analyzed data from MySQL to the HDFS using Sqoop for visualization and to generate reports for the BI team.
- Importing of data from various data sources such as Oracle and Comptel server into HDFS using transformations such as Sqoop, Map Reduce.
- Developed various dashboards in Tableau, used context filters, sets while dealing with huge volume of data.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like frequency of calls, top calling customers and designed and implemented service layer over Hbase Database.
- Designed and developed scalable and custom Hadoop solutions as per dynamic data needs and coordinated with technical team for production deployment of software applications for maintenance.
- Data is loaded back to the Teradata for the BASEL reporting and for the business users to analyze and visualize the data using Datameer. Real time streaming data using Spark with Kafka.
- Implemented a different use case to run applications in YARN containers as a long running jobs.
- Experience in providing support to data analyst in running Pig and Hive queries.
- Experience in administering the Linux systems to deploy Hadoop cluster and monitoring the cluster using Nagios and Ganglia. Reviewing the log files and error solving.
- Data Standardization using Talend components like tmap, tloqateaddress, tmatchgrouping, trecordmatch, tnormalize, tdenormalize and loading the data into MDM Server by creating Data models, SQL Server Database Table and Archive Tables.
- Developed custom Process chains to support master data and transaction data loads from BI to BPC.
- Involved in various POC activity using technology like Map reduce, Hive, Pig, and Oozie.
Environment: Hadoop, HDFS, Hive, Sqoop, Flume, Hortonworks, Cassandra, Java, Impala, Talend, Tableau, kafka, storm, Zookeeper and Hbase,YARN,Oracle 9i/10g/11g RAC with Solaris/redhat, MongoDB, Kerberos, SQL plus, PHP, Shell Scripting, Redhat/Suse Linux, EM Cloud Control.
Confidential - Northbrook, IL
Hadoop Administrator
Responsibilities:
- Installed, configured and maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, Hbase, Zookeeper and Sqoop.
- Managing and scheduling Jobs onHadoop Clusters using Apache, Cloudera (CDH3, CDH4) distributions.
- Worked on importing and exporting data from Oracle and DB2 into HDFS using Sqoop.
- Developed PIG Latin scripts to extract the data from the web server output files to load into HDFS.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Deployed Hadoop Cluster in Pseudo-distributed and Fully Distributed.
- Implemented NameNode backup using NFS. This was done for High availability.
- Created Hive External tables and loaded the data in to tables and query data using HQL.
- Collected the logs data from web servers and integrated in to HDFS using Flume.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Configured custom interceptors in Flume agents for replicating and multiplexing data into multiple sinks.
- Worked on NoSQL databases including Hbase and MongoDB.
- Good experience in analysis using PIG and HIVE and understanding of SQOOP and Puppet.
- Setting up automated 24*7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia.
- Hands-on experience on major components in Hadoop Ecosystem including Hive, Hbase, Hbase-Hive Integration, PIG, Sqoop, Flume & knowledge of Mapper/Reduce/HDFS Framework.
- Extensive experience in data analysis using tools like Syncsort and HZ along with Shell Scripting and UNIX.
- Handle the data exchange between HDFS& Web Applications and databases using Flume and Sqoop.
- Experienced in developing MapReduce programs using Apache Hadoop for working with Big Data.
- Good understanding of XML methodologies (XML, XSL, XSD) including Web Services and SOAP.
- Expertise in working with different databases like Oracle, MS-SQL Server, Postgresql and MS Access 2012 along with exposure to Hibernate for mapping an object-oriented domain model to a traditional relational database.
- Familiarity and experience with data warehousing and ETL tools.
- Good understanding of Scrum methodologies, Test Driven Development and continuous integration.
- Involved in log file management where the logs greater than 7 days old were removed from log folder and loaded into HDFS and stored for 3 months.
Environment: Cloudera, CDH 4.4, and CDH 3, Cloudera manager, Sqoop, Flume, Hive, HQL, Pig, RHEL, Cent OS, Oracle, MS-SQL, Zookeeper, Oozie, MapReduce, Apache Hadoop 1.x, Postgresql, Ganglia and Nagios.
Confidential - Houston, TX
Hadoop Administrator
Responsibilities :
- Installed/Configured/Maintained Apache Hadoop and Cloudera Hadoop clusters for application development and Hadoop tools like Hive, Pig, Hbase, Zookeeper and Sqoop.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Managing and scheduling Jobs on a Hadoop cluster.
- Worked on Providing User support and application support through remedy ticket management system on Hadoop Infrastructure.
- Installed and configured Hadoop cluster in Development, Testing and Production environments.
- Performed both major and minor upgrades to the existing CDH cluster.
- Responsible for monitoring and supporting Development activities.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Installed and configured flume agents with well-defined sources, channels and sinks.
- Configured safety valve to create active directory filters to sync the LDAP directory for Hue.
- Developed scripts to delete the empty Hive tables existing in the Hadoop file system.
- Understanding the existing Enterprise data warehouse set up and provided design and architecture suggestion converting to Hadoop using MapReduce, HIVE, SQOOP and Pig Latin.
- Implemented Name Node backup using NFS. This was done for High availability.
- Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop.
- Wrote shell scripts for rolling day-to-day processes and it automated using crontab.
- Collected the logs data from web servers and integrated in to HDFS using Flume.
- Implemented FIFOschedulers on the Job tracker to share the resources of the Cluster for the MapReduce jobs given by the users.
- Involved in Data model sessions to develop models for HIVEtables.
- Involved in gathering business requirements and analysis of business use cases.
Environment: Apache Hadoop, CDH4, Hive, Hue, Pig, Hbase,MapReduce, Sqoop, RedHat, CentOS and Flume.
Confidential - Reston, VA
Hadoop Administrator
Responsibilities:
- Responsible for installation and configuration of RedHat 4.x, 5.x and SUSE Linux Enterprise Server (SLES) 10.x, 11.0 on IBM and HP hardware.
- Installed Confidential director client on various servers and maintained the Confidential director for Hardware monitoring.
- Manage Virtual servers on VMwareESXServer3.0/3.5.
- Experienced in deploying, cloning and creating template of virtual machines using VMware.
- Installed and configured Tivoli End point client on servers for OS level monitoring.
- Implemented Relax and Recover (ReaR) for disaster recovery.
- Learned and managed systems with the Puppet configuration management tool.
- Built 3 and 4 node Oracle RAC Servers Configured SUDO to provide SUDO access to various users and group for elevated access.
- Worked on tools like Yum, YaST, RPM, Zypper, rug for Package management.
- Deployed latest patches and performed RedHatLinux Kernel Tuning for Oracle.
- Installed LAMP server (RHEL 5.4, Apache 2.2, MySQL 5.0 server and Perl).
- Used the LVM extensively for Creating/Modifying/Deleting Physical Volume, Volume Groups, Logical Volumes and file systems.
- Troubleshooting and analysis of hardware and software failures using core dump and log file analysis.
- Experience with IBM Power7 Virtualization and XEN Virtualization, VMware.
- Creating Cron jobs and giving Cron eligibility for administrative accounts.
- Performance Monitoring and Tuning using iostat, vmstat, netstat,sar, ksar, top.
- Provided 24x7 on call support for production environment.
- Resolving issues as they arise with JIRA and Confluence in coordination with the IT department.
- Completed a full migration of UNIX Server (AIX and Linux), including all network services.
Environment: RedHat Linux, Suse Linux Enterprise server, UNIX, puppet, VMWare, Oracle, Tivoli, Kernel Tuning, LVM
Confidential
Linux Administrator
Responsibilities:
- Installing, configuring and upgrading Linux (Primarily REDHAT and UBUNTU) and Windows Servers.
- Co-coordinating with the customers vendors for any system up gradation and giving the exact procedure to follow up.
- Contact various systems administration works under CentOS, Red Hat Linux environments.
- Configured Domain Name System (DNS) for hostname to IP resolution.
- Creating users, assigning groups and home directories, setting quota and permissions; administering file systems and recognizing file access problems.
- Installed Cent OS using Pre-Execution environment boot and Kick start method on multiple servers, remote installation of Linux using PXE boot.
- Performed various configurations which include networking and IPTable, resolving host names and SSH keyless login.
- Maintained and monitored all servers' operating system, managing Disk File systems, server performance, application patch level, disk space and memory usage, user activities on daily basis, also RAID configurations.
- Installed and verified that all AIX/Linux patches are applied to the servers.
- Responsible for maintenance Raid-Groups, LUN Assignments as per agreed design documents.
- Maintenance and installation of RPM and YUM package installations and other server management.
- Performed various configurations which include networking and IPTables, resolving hostnames, SSH key less login.
- Developed and optimize physical design of MySQL database systems.
- Implemented new releases to add more functionality as per the requirements.
- Supported LANs, WANs, network segments, Internet, and intranet systems and maintaining the integrity of the network.
- Schedule backup jobs by implementing cron job schedule during non-business hour.
- Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID configurations.
- Troubleshooting and fixing the issues at User level, System level and Network level by using various tools and utilities.
Environment: YUM, RAID, MYSQL 5.1.4, PHP, SHELL SCRIPT, MYSQL, WORKBENCH, LINUX 5.0, 5.1, CentOS, AIX
Cerner Corporation
Linux Administrator
Responsibilities:
- Installed, configured, maintained and administrated the network servers DNS, NFS and application servers Apache and Samba server.
- Installation and Configuration of SSH, TELNET, FTP, DHCP, DNS.
- Worked on UNIX shell scripting for system/application in automating server task, installing and monitoring applications and data feeding file transfer and log files.
- Maintained UNIX (Red Hat Enterprise Linux4, 5, CentOS4, 5, VMware) on Sun Enterprise servers & Dell Servers.
- Implemented the Jumpstart servers and Kickstart Servers to automate the server builds for multiple profiles.
- Installed and deployed RPM Packages.
- APACHE Server Administration with Virtual Hosting.
- Worked extensively in using VI editor to edit necessary files writing shell script.
- Worked on adding new Users and +groups and give sudo access and central file synchronization via sudoers, authorized keys, password, shadow, and group.
- Coordinated with application team in installation, configuration and troubleshoot issues with Apache, Web logic on Linux servers.
- Local and Remote administering of servers, routers and networks using Telnet and SSH.
- Monitored client disk quotas &disk space usage.
- Worked on backup technologies like VeritasNetbackup4.x, 5.0, 6.x and Tivoli Storage Manager 5.5.
- Involved in back up, firewall rules, LVM configuration, monitoring servers and on call support.
- Created BASH shell scripts to automate cron jobs and system maintenance. Scheduled cron jobs for job automation.
Environment: Redhat 4/5, Solaris 8/9/10, CentOS 4/5, SUSE Linux 10.1/10.3, VMware