We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

AZ

SUMMARY

  • Around 8 years of System Administrator experience in IT background which includes around 5 years of experience with emphasis on Hadoop Technologies.
  • Strong Knowledge on Multi Clustered environment and setting up Cloudera Hadoop and Hortonworks Eco - System.
  • Experience in installation, configuration and management of Hadoop Clusters.
  • Experience with Hbase, Cassandra, NoSQL and MongoDB. Experience writing Map Reduce Jobs, HIVEQL.
  • Set up and configure Hadoop System on Amazon AWS for processing massive volumes of Data.
  • Worked with Enterprise Analytics team and transform analytics requirements into Hadoop centric technologies.
  • Performance tuning the Hadoop Cluster by gathering and analyzing the existing infrastructure.
  • Automating the Hadoop Installation, configuration and maintaining the Cluster by using the tools like puppet.
  • Familiar with open source configuration management and deployment tools such as Puppet or Chef and Linux scripting.
  • Setting up monitoring infrastructure for Hadoop Cluster using Nagios and Ganglia.
  • Experience in managing and installing scalable and secureHadoop(CDH) clusters including role base authorization using Sentry as well as authentication using Kerberos.
  • Good experience in implementing Kerberos & Ranger inHortonworks HadoopEcosystem.
  • Experience in Ranger, Knox configuration to provide the security forHadoopservices (Hive, Hbase, Hdfs etc.)
  • Experience in using Knox to connect toHadoopcluster from outside the network in secured way.
  • In depth knowledge of Hortonworks, Cloudera Hadoop ecosystem tools, Kerberos, Ranger, Hadoop Authentication/Authorization, NoSql databases, Nosql data modeling techniques and mapreduce design patterns.
  • Experience in integrating AD/LDAP users with Ambari and Ranger.
  • Working with Flume to load the log Data from multiple sources directly into HDFS.
  • Experience in setting up replication betweenHadoopclusters using Falcon and Oozie for disaster recovery.
  • Configuring the Zookeeper to coordinate the servers in Clusters and to maintain the Data consistency.
  • Data migration from existing Data stores to Hadoop using Sqoop.
  • Upgrade and maintenance of Apache product stack.
  • Experience in managing Hadoop infrastructure like commissioning, decommissioning, log rotation, rack topology implementation.
  • Designing both Time driven and Data driven automated workflows using Oozie.
  • Supporting analysts by Administering and Configuring HIVE. Providing support to Data analyst in running PIG and HIVE queries.
  • Experience with Configuring AWS EC2 instances, S3 buckets, Auto-Scaling groups and CloudWatch.
  • Hands on experience in installing, configuring, and usingHadoopecosystem components like Hive, Pig, Impala, Hbase, Spark, Kafka, sqoop on Cloudera and Hortonworks distributions.
  • Having good understanding of Hortonworks (HDP) and Ambari tool.
  • Experience in deployingHadoopcluster on Public and Private Cloud Environment like Amazon AWS.
  • Writing shell scripts to dump the Shared Data from RDBMS Mysql servers to HDFS.
  • Familiar with Java virtual machine (JVM) and multi-threaded processing.
  • Experience in Agile methodology.

TECHNICAL SKILLS

Big Data Ecosystem: HDFS, HBase, Hadoop MapReduce, Zookeeper, Oozie, Hive, Pig, Sqoop, Flume, Oozie, Hue, Ranger, Sentry, Knox, Spark, Kafka.

RDBMS/ Database: SQL Server 2000/2005/2008 R2, MS-Access XP/2007/2008, ORACLE 10g/9i, Mysql, Nosql.

Scripting Languages: Shell scripting, Java Scripting, Unix shell scripting, Python, SQL, PIG Latin

Operating Systems: Unix, Linux, AIX, Windows XP, Server 2000, 03, Server 2008.

Storages Arrays: EMC Clariion 400/500/700, CX Series

PROFESSIONAL EXPERIENCE

Confidential, AZ

Hadoop Administrator

Responsibilities:

  • Worked on performing major upgrade of cluster from 4.0.x to 5.0.x.
  • Implemented Namenode High Availability on the Hadoop cluster to overcome single point of failure.
  • IntegratingHadoopcluster with Kerberos authentication for secured authentication & authorization ofHadoopcluster and monitored the connectivity.
  • Involved in efficiently collecting and aggregating large amounts of streaming log data into Hadoop Cluster using Apache Flume.
  • User behavior and their patterns were studied by performing analysis on the data stored in HDFS using Hive.
  • Used HiveQL to write Hive queries from the existing SQL queries.
  • The analyzed data mined from huge volumes of data was exported to Mysql using Sqoop.
  • Involved in installing and configuring Kerberos to implement security to the Hadoop cluster and providing authentication for users.
  • Worked with Big Data Analysts, Designers and Scientists in troubleshooting map reduce job failures and issues with Hive, Pig, and Flume etc.
  • Assisted in designing, development and architecture of Hadoop and Hbase systems.
  • Answering trouble tickets around theHadoopecosystem.
  • Integration support of tools that need to connect to like Tableau, Talend, generic ODBC/JDBC, etc.,
  • Coordinated with technical teams for installation of Hadoop and third related applications on systems.
  • Provided infrastructure recommendations, capacity planning and develop utilities to monitor cluster better.
  • Experience in rebalancing of hadoop cluster.
  • Experience on administration of HDFS, HIVE and HA MR2 / YARN.
  • Provided technical assistance for configuration, administration and monitoring of Hadoop clusters.
  • Setting up and validating Disaster Recovery replication of data from Production cluster
  • Formulated procedures for planning and execution of system upgrades for all existing Hadoop clusters.
  • Installing and monitoring the Hadoop cluster resources using Nagios.
  • Experience in designing and implementation of secure Hadoop cluster using Kerberos.
  • Experience in managing Hadoop infrastructure like commissioning, decommissioning, log rotation, rack topology implementation.
  • InstallingHadoopUpdates, patches and version upgrades. Automating the same through scripts.
  • Performance tuning ofHadoopclusters andHadoopMap Reduce/Spark routines.
  • Experience in using Zookeeper for coordinating the distributed applications.
  • Configuring, installing, managing and administrating Nosql Hbase clusters and scheduling jobs using Oozie workflow.
  • Worked with systems engineering team to plan and deploy newHadoopenvironments and expand existingHadoopclusters.
  • Managing the cluster resources by implementing fair scheduler and capacity scheduler.
  • Supported technical team members for automation, installation and configuration tasks.

Environment: Hadoop, MapReduce, Hive, Oozie, Sqoop, RDBMS, Flume, Cloudera Manager, Shell Script, HDFS, Yarn, Spark, Pig, RDBMS, HBase, Sqoop, kafka, Zookeeper, Java, Linux/Unix, Namenode Federation, Datanodes.

Confidential, TN

Hadoop Administrator

Responsibilities:

  • Installed and configured various components of Hortonworks Hadoop ecosystem and maintained their integrity.
  • Planned for production cluster hardware and software installation on production cluster and communicating with multiple teams to get it done.
  • Designed, configured and managed the backup and disaster recovery for HDFS data.
  • Commissioned Data Nodes when data grew and decommissioned when the hardware degraded.
  • Migrated data across clusters using DISTCP.
  • Experience in collecting metrics for Hortonworks Hadoop clusters using Nagios and Ambari.
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
  • Evaluate the administration and operational practices, and evolve automation procedures (Using scripting languages such as Shell, Python, Chef, Puppet,Rubyetc.)
  • Used python scripts to automate tasks.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios.
  • Monitored workload, job performance and capacity planning using Ambari.
  • Installed and configured Hive, Pig, Sqoop and Oozie on the HDP cluster.
  • Upgrading HortonworksHadoopcluster from current version to minor version upgrade as well as to major versions.
  • Enabled Ranger security for Hive in the cluster.
  • Implemented HDFS snapshot feature.
  • Working with data delivery teams to setup newHadoopusers. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
  • Migrated date from RDBMS to hadoop using sqoop.
  • Deployed cluster in AWS cloud. Experience in setting up EC2 instances in AWS cloud.
  • Worked with big data developers, designers and scientists in troubleshooting map reduce job failures and issues with Hive, Pig and Flume
  • Worked on Tableau for generating reports on HDFS data

Environment: Hadoop, HDFS, Ambari, HDP/HDF, Mapreduce, RDBMS, Yarn, Oracle 11g/10g, BigData, SQL, Hbase, Shell Scripting, Ranger, Python, Kafka, AWS, Zookeeper, Hortonworks, Oozie, Red Hat Linux, UNIX, Federation, Nosql.

Confidential, CA

Hadoop Administrator

Responsibilities:

  • Responsible for implementation and ongoingadministrationofHadoopinfrastructure.
  • Helped in setting up Rack topology in the cluster.
  • Good experience with Hadoop Ecosystem components such as Hive, Sqoop and Oozie.
  • Worked on pulling the data from oracle databases into the Hadoop cluster using Sqoop import.
  • Experience working on processing data using Pig and Hive. Involved in creating Hive tables, data loading and writing hive queries.
  • Install, configure and manage HBase clusters to store huge datasets & random data access.
  • Implemented IMPALA for data processing on top of HIVE.
  • Developed custom MapReduce programs and custom User Defined Functions (UDFs) in Hive to transform the large volumes of data with respect to business requirement.
  • Upgrading Hadoop cluster from current version to minor version upgrade as well as to major versions.
  • Worked extensively on automating task using bash script.
  • Experience with capacity planning and performance tuning on cloudera cluster.
  • Installing and Configuring Cloudera Manager and install CDH, managed service, and Cloudera Manager Agent software on the hosts.
  • Created Oozie workflows to automate data ingestion using Sqoop and process incremental log data ingested by Flume using Pig.
  • Good experience on scheduling CRON jobs in Linux.
  • Responsible for designing and implementing ETL process to load data from different sources, perform data mining and analyze data using visualization/reporting tools to leverage the performance of Open Stack.

Environment: Hadoop, MapReduce, Hive, Oozie, Sqoop, Flume, CDH, Cloudera Manager, Shell Script, HDFS, Yarn, AWS, Pig, HBase, Sqoop, Nosql, Ruby, Python, Zookeeper, Java.

Confidential

System Administrator

Responsibilities:

  • Involved in testing of products and documentation of necessary changes required in this environment.
  • Installation, configuration and Operating System upgrade on Sun Solaris 8,9,10, on sun Servers
  • Monitored the filesystemsand CPU load for better performance
  • Performed daily backups, restoration of data as per requests from developers using Veritas Netbackup.
  • Performance & Process Management with using Linux base commands.
  • Network trouble shooting using ‘ndd’, 'traceroute', 'netstat', 'ifconfig' and 'snoop' etc
  • Installed of all the required Software, patches, documentation, licensing trouble shooting.
  • Set up user and group login ID's, printing parameters, network configuration, password, resolving permissions issues, and user and group quota.
  • Monitored overall system performance, performed user management, system updates and disk & storage management.
  • Created users, manage user permissions, maintain User & File System quota on Redhat Linux and Solaris.
  • Setup of full networking services and protocols on UNIX including NIS/NFS, DNS, SSH, DHCP, NIDS, TCP/IP, ARP, applications, kick start and print servers to insure optimal networking, application, and printing functionality.
  • Installed and configured Sudo for users to access the root privileges.
  • Wrote Technical documentation for various projects and worked with Technical writer to finalize the documents.

Environment: Solaris, Linux, JBOSS, Ssh, Oracle.

Confidential

System Administrator

Responsibilities:

  • Hand on Experience in Linux server administration tasks, which include - IP Addressing, Sub-netting, Ethernet Bonding, Static IP, debugging and resolvingsystem/application issues
  • System audit report like no. of logins, success & failures, running cron jobs.
  • Remotely coping files using sftp, ftp, scp, winscp, and filezilla.
  • Set up user and group login ID's, printing parameters, network configuration, password, resolving permissions issues, and user and group quota.
  • Taking backup using tar and recovering during the data loss.
  • Experience in writing bash scripts for job automation.
  • Documenting the installation of third-party software’s.
  • Configuring printers to the Solaris and Linux servers and also installing third party software’s.
  • Maintaining relations with project managers, DBA’s, Developers, Application support teams and operational support teams to facilitate effective project deployment.
  • Manage system installation, troubleshooting, maintenance, performance tuning, managing storage resources, network configuration to fit application and database requirements.
  • Responsible for modifying and optimizing backup schedules and developing shell scripts for it.

Environment: Linux (Red Hat Enterprise, CentOS), Windows, HP, IBM, Solaris, Oracle 8i, Cisco routers/switches, Dell 6400, 1250, Sun E450, E250.

We'd love your feedback!