Hadoop Administrator Resume
5.00/5 (Submit Your Rating)
Louisville, KY
SUMMARY
- Hands on experience in installing, configuring, and using Hadoop ecosystem components like HadoopMapReduce, HDFS, Hbase, Oozie, Hive, Sqoop, Pig, Zookeeper and Flume.
- Good Understanding of the Hadoop Distributed File System and Eco System (MapReduce, Pig, Hive, Sqoop and Hbase)
- Technical expertise in Big data/Hadoop HDFS, Map Reduce, Spark, HIVE, PIG, Sqoop, Flume, Oozie, NoSQL Data bases HBase, SQL, Unix Scripting.
- Experience with leveraging Hadoop ecosystem components including Pig and Hive for data analysis, Sqoop for data migration, Oozie for scheduling and HBase as a NoSQL data store.
- Experience in importing and exporting data using Sqoopfrom HDFS to Relational Database Systems and vice - versa.
- Experience in installing Hadoop cluster using different distributions of Apache Hadoop, Cloudera, and Hortonworks.
- Around 4years of working experience in setting, configuring and monitoring of Hadoop cluster of Cloudera, Hortonworks distribution.
- Monitor Hadoop cluster using tools like Nagios, Ganglia, Ambari and Cloudera Manager.
- Experience in Hadoop Shell commands, writing MapReduce Programs, verifying managing and reviewing Hadoop Log files.
- Installation, Configuration, and Administration of Hadoop cluster of major Hadoop distributions such as Cloudera Enterprise (CDH3 and CDH4) and Hortonworks Data Platform (HDP1 and HDP2)
- Good Knowledge on Hadoop Cluster architecture and monitoring the cluster.
- Exceptional organizational, multi-tasking, problem solving and leadership skills with result-oriented attitude.
- Effectively work independently and in team with users, project managers, business analysts and developers.
- Ability to interact with developers and product analysts regarding issues raised and following up with them closely.
- Excellent interpersonal and communication skills, creative, research-minded, technically competent and result- oriented with problem solving and leadership skills.
TECHNICAL SKILLS
Hadoop Ecosystems: HDFS, Hive, Yarn, HBase, Sqoop, Flume, Oozie, Pig, Kafka
Operating Tools: Linux (RHEL/CENT OS), Windows (XP/7/8)
Monitoring Tools: Nagios Databases Oracle, MySQL Network TCP/IP, HTTP/HTTPS, SSH, FTP Security Kerberos security.
Programming and Scripting: C & shell.
PROFESSIONAL EXPERIENCE
Hadoop Administrator
Confidential - Louisville, KY
Responsibilities:
- Installed and configured CDH5.0.0 cluster, using Cloud era manager.
- Implemented automatic failover zookeeper and zookeeper failover controller.
- Developed scripts for benchmarking with Terasort / Teragen. worked on commission and decommission of data node.
- Monitored workload, job performance and capacity planning.
- Managing and reviewing Hadoop log files and debugging failed jobs.
- Tuned the cluster by Commissioning and decommissioning the Data Nodes.
- Supported cluster maintenance, Backup and recovery for production cluster.
- Backed up data on regular basis to a remote cluster using distcp
- Knowledge on supporting data analysis projects using Elastic Map Reduce on the Confidential Web Services (Confidential) cloud.
- Fine tuning of Hive jobs for better performance. worked on data processing on AWS EC2 Cluster.
- Automated all the jobs for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
- Collected and aggregated large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
Hadoop Administrator
Confidential - Louisville, KY
Responsibilities:
- Worked on implementation of SSL /TLS implementation.
- Configuration of SSL and trouble shooting in Hue.
- Responsible for building scalable distributed data solutions using Hadoop Cloudera works
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanism
- Enabled Kerberos for authorization and authentication.
- Enabled HA for NameNode, Resource Manager, Yarn Configuration and Hive Megastore.
- Configured Journal nodes and Zookeeper Services for the cluster using Cloudera.
- Monitored Hadoop cluster job performance and capacity planning.
- Monitored and reviewed Hadoop log files.
- Performed Cloudera Manager and CDH upgrades
- Taking backup of Critical data, Hive data and creating snapshots.
- Responsible for building scalable distributed data solutions using Hadoop.
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster
- Monitoring and troubleshooting, and review Hadoop log files.
- Handled importing of data from various data sources, performed transformations using Hive, Ma and loaded data into HDFS.
- Extraction data using Flume. Import/Export to HDFS/RDMS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
- Good Knowledge of NoSQL database like HBase.
- Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
- Installed Oozie workflow engine to run multiple Hive.
- Performance tuning of Impala jobs and resource management in cluster.
Environment: MapReduce, HDFS, Hive, SQL, Oozie, Sqoop, UNIX Shell Scripting, Yarn, Talend
Managing Director
Confidential
Responsibilities:
- Successfully launched Web Design, Internet Marketing and PC/Network troubleshooting business in South west England region
- Designed online website to help local businesses access our services and providing them awareness about IT
- Managed the south west- Somerset region for IT contractor jobs
- Managed all aspects of business operations, including advertising/marketing, web design, and social media, accounting, and customer service.
- Implemented SEO programming and automation process for local businesses.