Hadoop Administrator/engineer Resume
4.00/5 (Submit Your Rating)
Northbrook, IL
SUMMARY:
- An IT Professional with about 7 years of successful track record as an Administrator/Engineer in multiple Big Data and IT engagements for multiple enterprise clients.
- Worked recently on various levels of Hadoop engagements ranging from Groundup design, planning, install and administration of medium to large Production and Test Hadoop Clusters in both Physical and Virtual environments.
- Experience in helping deploy multiple Hadoop framework tools for Data Analysts, Scientists and Developers and help provide necessary support for Big Data Analytic use cases.
TECHNICAL SKILLS:
- Cloudera CDH 4.x, 5.x
- Hortonworks HDP 2.x and Open source Hadoop 1.0 and 2.0
- Cloudera Manager 4.x, 5.x
- Ambari.
- HDFS
- YARN
- MapReducev1/v2
- Apache Spark
- Pig
- Hive
- Impala
- HBase
- Flume
- Sqoop
- Kafka
- Oozie
- Hue
- Tez
- Zookeeper
- R
- MySQL
- C++
- .NET
- Cassandara.
- RHEL/CENTOS Linux
- Windows
- VMware ESX 5.1, 6.0.
- VMAX40k
- VMAX
- DMX
- VNX
- Netapp
- SYMCLI
- CISCO Switches
- DCNM
- Open Replicator.
PROFESSIONAL EXPERIENCE:
Hadoop Administrator/Engineer
Confidential, Northbrook, IL
Responsibilities:- Member of Core Hadoop Architecture and Implementation team participating in design/implementation of a 200 node CDH 4.5.0 cluster for the client.
- Worked closely with SA and Build team to make sure all hardware requirements and software was properly setup for optimum usage of resources.
- Responsible for Installation of Cloudera Manager and deploying the CDH Agents to install CDH across the cluster.
- Responsible for Hadoop configuration, maintenance, monitoring, tuning and troubleshooting of the installed CDH cluster.
- Installed and configured software modules like, YARN with MR2, Pig, Hive, HBase, Sqoop, Flume, Kafka, Spark.
- Experience in commissioning and Decommissioning of nodes.
- Experience with setting up Cluster HA Namenodes, Zookeeper and Journal Nodes.
- Experience with setting up static and dynamic resource pools using YARN in Cloudera Manager.
- Rack aware configuration
- Applying Patches and Perform Version Upgrades to the various modules in the HDFS cluster.
- Day to Day support for cluster issues, job failures and resource management/tuning.
- Work with Development team to fine tune jobs created in Pig, Hive, Impala, Spark and MapReduce
- Work with Cloudera Support Team to fine tune cluster.
Hadoop Administrator
Confidential, SFO, CA
Responsibilities:- Participated in design and development of scalable and custom Hadoop solutions as per dynamic data needs.
- Coordinated with technical team for production deployment of software applications for maintenance.
- Provided operational support services relating to Hadoop infrastructure and application installation.
- Responsible for architecting, installation and administration of HDP 2.2.4, Vanilla Hadoop distributions.
- Setup Hadoop (YARN/MapReduce) cluster on various platforms like RHEL 5.6, CentOS 6.4
- Managed 120+ nodes HDP 2.2.4 cluster with 3 PB of data using Ambari 2.0.
- Configured High - Availability (HA) architecture for Namenode/Resource Manager.
- Experience in decommissioning and commissioning node on running cluster.
- Monitoring Hadoop cluster using tools like Ambari, Ganglia and Cloudera Manager.
- Installed, configured and integrated Hadoop Ecosystem components (Hive, Sqoop, Flume, PIG, and Tez).
- Planned and managed HDFS storage capacity. Advised teams on best practices and optimal processes using Pig, Hive and Sqoop tools.
- Implemented backup and recovery procedures from a Namenode failure using NFS.
- Worked on importing and exporting data from RDBMS into HDFS and Hive using sqoop.
- Collected log data from Webservers and integrated into HDFS using Flume.
- Worked on standards and proof of concept in support of Hortonworks implementation using AWS cloud infrastructure.
- Collaborate with cross-functional teams to ensure that applications are properly tested, configured, and deployed.
- Cluster maintenance including Capacity planning, adding and removing cluster nodes; cluster Monitoring and Troubleshooting
- Rack Aware Configuration
- Cluster HA Setup
- Applying Patches and Perform Version Upgrades.
- Day to day support for the cluster issues and job failures.
- Working with Dev Team to tune Job Knowledge of Writing Hive Jobs.
- Working with Hortonworks Support Team to Fine tune Cluster
- Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop.
- Built a strong working relationship with the client by successfully managing both at Onsite and Offshore.
- Monitoring and maintenance of system health.
Storage Administrator
Confidential, Deerfield, IL
Responsibilities:- Performed Storage Provisioning on VMAX40k, VMAX, DMX-4 and VNX ( 500TB to 1PB Environments)
- Storage Provisioning for New server builds such as Vmware ESX, Windows, Linux, Solaris series as per the requirement.
- Responsible for creating VMAX Thin Pools, FAST policies, and Auto Provisioning Groups.
- Performed zoning on CISCO switches using CISCO CLI and DCNM.
- Involved in troubleshooting routine critical issues including server throughput, ports availability, zoning requirements, one-path down, enabling the path on source or destination arrays, data not migrated, host not seeing storage and storage management problems.
- Migrated data from old storage arrays like VMAX, DMX-4 to new VMAX40k using Open Replicator.
- Extensive use of SYMCLI on VMAX.
- Worked on Lun Expansions.
- Created Aggregates, Volumes, Qtrees, and Quotas on NetApp arrays using NetApp system manager, Filer View and CLI.