Hadoop Administrator Resume
2.00/5 (Submit Your Rating)
Dallas, TX
SUMMARY PROFILE:
- Over 10+ years of professional IT experience this includes 3 plus years of experience in Big Data ecosystem & Administration.
- Excellent Experience in Hadoop architecture and various components such as HDFS, YARN and MapReduce programming paradigm.
- Hands on experience in Installing, Configuring, Troubleshooting and Developing using Hadoop & ecosystem components like MapReduce, HDFS, Hive, Pig, Sqoop, Spark, Flume, Kafka, Storm & Elastic Search.
- Experience working on various distributions such as Azure HDInsights, Hortonworks, Cloudera and EMR.
- Experience in Hadoop Cluster monitoring the troubleshooting using tools such as Ganglia & Nagios.
- Experienced in designing, implementing and managing Secure Authentication mechanism to Hadoop Cluster with Kerberos.
- Experienced in Ranger, Apache Centri in Access Control and Authorizations.
- Experience with Azure, AWS Cloud (EC2, S3 & EMR).
- Experienced in installation, configuration, troubleshooting and maintenance of Kafka & Spark clusters.
- Experience in sizing, setting up and maintaining Kafka cluster.
- Worked on setting up Apache NiFi in orchestrating data pipeline actions.
- Experienced in installation, configuration and maintenance of Elastic Search cluster.
- Worked on ingesting log data into Hadoop using Flume.
- Experience in managing and reviewing Hadoop log files.
- Extensive experience in writing Map Reduce, Hive, PIG Scripting.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Excellent understanding and knowledge of NOSQL databases like MongoDB, HBase, and Cassandra.
- Implemented setting up standards and processes for Hadoop based application design and implementation.
- Experience in managing Hadoop clusters using Ambari, Cloudera Manager.
- Experience in Administering, Installation, configuration, troubleshooting, Security, Backup, Performance Monitoring and Fine-tuning of Redhat Linux.
- Good UNIX/Linux knowledge including the ability to understand the interaction between applications, HW, and operating system and to provide recommendations and suggestions related to troubleshooting and performance improvement.
- Create and implement shell and PL/SQL scripts to automate BAU tasks, configure scheduler jobs using CA tools.
- Ability to adapt to evolving technology, strong sense of responsibility and .
TECHNOLOGY EXPERIENCE:
- Hadoop, HDFS, Yarn, Kerberos, Centri, Ranger, Map Reduce, Spark, Shark, Hive, Pig, Sqoop, Flume, Kafka, NiFi, Storm, Oozie, ZooKeeper.
- HBase, Cassandra, MongoDB
- Ambari, Cloudera Manager, Nagios, Zabbix.
- JAVA, J2EE.
- Oracle 8i, 9i, 10g, 11g,MS Sql Server, Teradata, Sybase, SAS, Informatica, Datastage.
- Net backup, Oracle Grid, RMAN, BMC Patrol, Novell e-Directory, Novell streamline server, Novell workbench.
- Eclipse, NetBeans.
- Linux, Confidential AIX 5.3, Solaris10, Windows
EXPERIENCE:
Confidential, Dallas, TX
HADOOP ADMINISTRATOR
Responsibilities:
- Worked on setting up cluster for various services such as Hadoop, Spark, HBase, Kafka etc on Azure HDInsights.
- Installation, configuration and Administration of HDP 2.x Hadoop cluster
- Worked on migration strategy from on-premise to Azure HDInsights.
- Run the benchmark tools to test the cluster performance
- Configure the Hadoop properties based on the benchmark result
- Tuning the cluster based on the benchmark results
- Monitoring System Metrics and logs for any problems
- Installation and configuration of Hadoop ecosystem components like HBase, Hive, Pig, Sqoop, Spark, Zookeeper etc. as per requirement.
- Provided support to users for diagnosing, reproducing and fixing Hadoop related issues.
- Work on HDP environments and took ownership of problem isolation and resolution, and whenever case arises do the bug reporting.
- Ensure that critical user issues are addressed quickly and effectively.
- Apply troubleshooting techniques to provide solutions to our user’s individual needs.
- Troubleshoot, diagnose and potentially escalate user inquiries during their engineering and operations efforts.
- Investigate product related issues both for individual customers and for common trends that may arise.
- Setting up new Hadoop users with HDFS maintenance and support. Keeping a track of Hadoop Cluster connectivity and security.
- Worked on setting up Apache NiFi in orchestrating data pipeline actions.
Technology: Azure HDInsights, Hortonworks, Hadoop, Yarn, Spark, Kafka, Hive, Pig, Sqoop, Kerberos, Ranger, NiFi, Oracle, Netezza, Tableau, Python, Java 8.0, Log4J, GIT, Azure, JIRA.
Confidential, Austin, TX
HADOOP ADMINISTRATOR
Responsibilities:
- Implemented multiple CDH5 & HDP Hadoop clusters on Redhat Enterprise Linux.
- Responsible for building scalable distributed data solutions using Hadoop
- Resource management of Hadoop Cluster in configuring the cluster with optimal parameter.
- Performing day to day activities such as upgrades, applying patches, adding/removing nodes from the cluster for maintenance and capacity needs.
- Responsible for monitoring the Hadoop cluster using Nagios.
- Involved in upgrading all the Hadoop components such as HDFS, MapReduce, Yarn, Impala, Hive, Pig, Sqoop, Flume, Oozie, HBase, Spark etc to the latest versions.
- Worked on implementing NOSQL database Cassandra cluster.
- Worked on setting up of Hadoop ecosystem & Kafka Cluster on AWS EC2 Instances.
- Performed benchmarking of Kafka cluster to measure the performance and resource considerations and tuning the cluster for optimal performance.
- Worked with tuning and configuring various parameters to maintain High Availability and consistency targets of the cluster.
- Experience in enabling security to Kafka cluster using Kerberos authentication.
- Implemented Apache Centri in Access Control and Authorizations.
- Extensively worked on managing Kafka logs for traceability and debugging.
- Worked on designing, implementing and managing Secure Authentication mechanism to Hadoop Cluster with Kerberos.
- Working on Centri in enabling metadata management, governance and audit.
- Performed backup of metadata at regular intervals and other maintenance activities such as balancing the cluster, and HDFS health check.
- Responsible for maintaining the clusters in different environments.
- Involved in upgradation process of the Hadoop cluster from CDH4 to CDH5.
- Installed and configured Flume, Oozie on the Hadoop cluster.
- Managing, defining and scheduling Jobs on a Hadoop cluster.
- Worked on installing cluster, commissioning & decommissioning of datanode, namenode recovery, capacity planning, and slots configuration.
- Worked with different file formats such as Text, Sequence files, Avro, ORC and Parquette.
- Installed and configured Spark on Yarn.
- Implemented indexing for logs from Oozie to Elastic Search.
- Analysis on integrating Kibana with Elastic Search.
- Monitoring the log flow from LM Proxy to Elastic Search-Head
- Responsible to manage data coming from different sources.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Experience in managing and reviewing Hadoop log files.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Supported in setting up QA environment and updating configurations for implementing scripts.
Technology: Cloudera Manager, Ambari, Hadoop, Nagios, Zabbix, Spark, Kafka, Storm, Shark, Hive, Pig, Sqoop, MapReduce, Kerberos, Centri, Ranger, Salt, Kibana, Talend, Oracle, Teradata, SAS, Tableau, Java 7.0, Log4J, Junit, MRUnit, SVN, JIRA.