We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

5.00/5 (Submit Your Rating)

Louisville, KY

SUMMARY

  • Hands on experience in installing, configuring, and using Hadoop ecosystem components like HadoopMapReduce, HDFS, Hbase, Oozie, Hive, Sqoop, Pig, Zookeeper and Flume.
  • Good Understanding of the Hadoop Distributed File System and Eco System (MapReduce, Pig, Hive, Sqoop and Hbase)
  • Technical expertise in Big data/Hadoop HDFS, Map Reduce, Spark, HIVE, PIG, Sqoop, Flume, Oozie, NoSQL Data bases HBase, SQL, Unix Scripting.
  • Experience with leveraging Hadoop ecosystem components including Pig and Hive for data analysis, Sqoop for data migration, Oozie for scheduling and HBase as a NoSQL data store.
  • Experience in importing and exporting data using Sqoopfrom HDFS to Relational Database Systems and vice - versa.
  • Experience in installing Hadoop cluster using different distributions of Apache Hadoop, Cloudera, and Hortonworks.
  • Around 4years of working experience in setting, configuring and monitoring of Hadoop cluster of Cloudera, Hortonworks distribution.
  • Monitor Hadoop cluster using tools like Nagios, Ganglia, Ambari and Cloudera Manager.
  • Experience in Hadoop Shell commands, writing MapReduce Programs, verifying managing and reviewing Hadoop Log files.
  • Installation, Configuration, and Administration of Hadoop cluster of major Hadoop distributions such as Cloudera Enterprise (CDH3 and CDH4) and Hortonworks Data Platform (HDP1 and HDP2)
  • Good Knowledge on Hadoop Cluster architecture and monitoring the cluster.
  • Exceptional organizational, multi-tasking, problem solving and leadership skills with result-oriented attitude.
  • Effectively work independently and in team with users, project managers, business analysts and developers.
  • Ability to interact with developers and product analysts regarding issues raised and following up with them closely.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result- oriented with problem solving and leadership skills.

TECHNICAL SKILLS

Hadoop Ecosystems: HDFS, Hive, Yarn, HBase, Sqoop, Flume, Oozie, Pig, Kafka

Operating Tools: Linux (RHEL/CENT OS), Windows (XP/7/8)

Monitoring Tools: Nagios Databases Oracle, MySQL Network TCP/IP, HTTP/HTTPS, SSH, FTP Security Kerberos security.

Programming and Scripting: C & shell.

PROFESSIONAL EXPERIENCE

Hadoop Administrator

Confidential - Louisville, KY

Responsibilities:

  • Installed and configured CDH5.0.0 cluster, using Cloud era manager.
  • Implemented automatic failover zookeeper and zookeeper failover controller.
  • Developed scripts for benchmarking with Terasort / Teragen. worked on commission and decommission of data node.
  • Monitored workload, job performance and capacity planning.
  • Managing and reviewing Hadoop log files and debugging failed jobs.
  • Tuned the cluster by Commissioning and decommissioning the Data Nodes.
  • Supported cluster maintenance, Backup and recovery for production cluster.
  • Backed up data on regular basis to a remote cluster using distcp
  • Knowledge on supporting data analysis projects using Elastic Map Reduce on the Confidential Web Services (Confidential) cloud.
  • Fine tuning of Hive jobs for better performance. worked on data processing on AWS EC2 Cluster.
  • Automated all the jobs for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Collected and aggregated large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.

Hadoop Administrator

Confidential - Louisville, KY

Responsibilities:

  • Worked on implementation of SSL /TLS implementation.
  • Configuration of SSL and trouble shooting in Hue.
  • Responsible for building scalable distributed data solutions using Hadoop Cloudera works
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanism
  • Enabled Kerberos for authorization and authentication.
  • Enabled HA for NameNode, Resource Manager, Yarn Configuration and Hive Megastore.
  • Configured Journal nodes and Zookeeper Services for the cluster using Cloudera.
  • Monitored Hadoop cluster job performance and capacity planning.
  • Monitored and reviewed Hadoop log files.
  • Performed Cloudera Manager and CDH upgrades
  • Taking backup of Critical data, Hive data and creating snapshots.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster
  • Monitoring and troubleshooting, and review Hadoop log files.
  • Handled importing of data from various data sources, performed transformations using Hive, Ma and loaded data into HDFS.
  • Extraction data using Flume. Import/Export to HDFS/RDMS using Sqoop.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Good Knowledge of NoSQL database like HBase.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
  • Installed Oozie workflow engine to run multiple Hive.
  • Performance tuning of Impala jobs and resource management in cluster.

Environment: MapReduce, HDFS, Hive, SQL, Oozie, Sqoop, UNIX Shell Scripting, Yarn, Talend

Managing Director

Confidential

Responsibilities:

  • Successfully launched Web Design, Internet Marketing and PC/Network troubleshooting business in South west England region
  • Designed online website to help local businesses access our services and providing them awareness about IT
  • Managed the south west- Somerset region for IT contractor jobs
  • Managed all aspects of business operations, including advertising/marketing, web design, and social media, accounting, and customer service.
  • Implemented SEO programming and automation process for local businesses.

We'd love your feedback!