We provide IT Staff Augmentation Services!

Cloudera Hadoop Platform Administration Resume

3.00/5 (Submit Your Rating)

SUMMARY

  • Minimum 8 - 10 years of IT experience with 4+ years of relevant Cloudera Hadoop Administration including Cluster management, Installation, Upgrades, patching maintenance and configuration of Hadoop Environment and tools with CDH certified.

TECHNICAL SKILLS

AWS EMR: (Elastic Map Reduce), Presto

Hadoop Stack: HDFS, Yarn, MapReduce, Spark Hive and Impala SQL tool

Security: Kerberos, Authentication and Authorization

Scripting Knowledge: Shell, Python Strong Linux Administration Experience

PROFESSIONAL EXPERIENCE

Cloudera Hadoop Platform Administration

Confidential

Responsibilities:

  • Experience in administering large Cloudera Hadoop clusters in Amazon Cloud including activities like deploy Hadoop cluster, add and remove nodes, keep track of jobs, monitor critical parts of teh cluster, configure name-node high availability, schedule and configure it and take backups.
  • General operational expertise such as good troubleshooting skills, understanding of system's capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks.
  • Responsible for implementation and ongoing administration of Hadoop infrastructure.
  • Aligning with teh AWS systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
  • Working with data analytics delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Impala and MapReduce access for teh new users.
  • Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise,
  • Monitoring, Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
  • Screen Hadoop cluster job performances and capacity planning
  • Monitor Hadoop cluster connectivity and security
  • Manage and review Hadoop log files.
  • File system management and monitoring.
  • HDFS support and maintenance, disk space management
  • Diligently teaming with teh infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
  • Single Point of Contact for Vendor escalation
  • Automate manual tasks.
  • Software installation and configuration, patches and upgrades.

We'd love your feedback!