We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

2.00/5 (Submit Your Rating)

New York, NY

SUMMARY:

  • 7+ years of experience in IT industry includes Hadoop Big data consultant in Banking, Telecom and financial clients.
  • Having 3+ years of comprehensive experience as a Hadoop (HDFS, MAPREDUCE, HIVE, PIG, SQOOP, FLUME, SPARK, KAFKA, KAFKA, ZOOKEEPER, AVRO, OOZIE, HBASE) Hadoop Consultant & Administrator.  
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce concepts. 
  • Well versed in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster. 
  • Hands on experience on major components in Hadoop Ecosystem like Hadoop Map Reduce, HDFS, HIVE, PIG, HBase, Sqoop. 
  • Experience in running workflow jobs with actions that run Hadoop Map/Reduce and Pig jobs. 
  • Experience in managing and reviewing Hadoop Log files. 
  • Experience in Hadoop administration activities such as installation and configuration of clusters using Apache and Cloudera. 
  • Good Knowledge in Amazon AWS concepts like EMR and EC2 web services which provides fast and efficient processing of Big Data. 
  • Experience in handling Hadoop Cluster and monitoring the cluster using Cloudera Manager, Ambari, Nagios and Ganglia
  • Experience in Hadoop Shell commands, verifying managing and reviewing Hadoop Log files.
  • Experience in performing major and minor upgrades of Hadoop clusters in Apache, and Cloudera distributions.
  • Experience in deployment of Hadoop cluster using Puppet tool.

WORK EXPERIENCE:

Confidential, New York, NY

Hadoop Administrator

Responsibilities:

  • Installed and configured CDH cluster, using Cloudera manager for easy management of existing Hadoop cluster.
  • Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop. 
  • Extensively using Cloudera manager for managing multiple clusters with petabytes of data.
  • Having knowledge on documenting processes, server diagrams, preparing server requisition documents 
  • Setting up the machines with Network Control, Static IP, Disabled Firewalls, Swap memory.
  • Managing the configuration of the cluster to the meet the needs of analysis whether I/O bound or CPU bound 
  • Worked on setting up high availability for major production cluster.
  • Performed Hadoop version updates using automation tools.
  • Working on setting up 100 node production cluster and a 40 node backup cluster  Confidential  two different data centers 
  • Experienced in dealing with MapR support team, bug reporting, resolving configuration and other MapR related issues 
  • Importing and exporting structured data from different relational databases into HDFS and Hive using Sqoop.
  • Configured Flume for efficiently collecting, aggregating and moving large amounts of log data from many different sources to HDFS.
  • Involved in setting up hive, hiveserver2, hive authorization and testing the environment 
  • Wrote shell scripts for rolling day - to-day processes and it is automated. 
  • Managed load balancers, firewalls in a production environment.
  • Involved in collecting and aggregating large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Good experience in troubleshoot production level issues in the cluster and its functionality.
  • Worked with application teams to install operating system and Hadoop updates, patches, version upgrades as required. 
  • Managing and scheduling Jobs on a Hadoop cluster.

Environment: Hadoop, HDFS, Hive, Sqoop, Flume, Zookeeper,Cassandra,Mongo DB and HBase,Ruby,Hue,Big Data Cloudera CDH Apache Hadoop, Shell Scripting,Net Backup, Time finder Backups, SQL, NOSQL  

Confidential, Irvine, CA

Hadoop Administrator

Responsibilities:
  • Installed, Configured and Maintained the Hadoop cluster for application development and Hadoop ecosystem components like Hive, Pig, HBase, Zookeeper and Sqoop. 
  • Worked on Hadoop Architecture and various components such as HDFS, Name Node, Data Node, Resource Manager, Node Manager and YARN / Map Reduce programming paradigm. 
  • Monitoring Hadoop Cluster through Cloudera Manager and Implementing alerts based on Error messages. Providing reports to management on Cluster Usage Metrics and Charge Back customers on their Usage. 
  • Extensively worked on commissioning and decommissioning of cluster nodes, replacing failed disks, file system integrity checks and maintaining cluster data replication. 
  • Assigning number of mappers and reducers to Map reduce cluster. 
  • Setting up HDFS Quotas to enforce the fair share of computing resources. 
  • Configuring and maintaining YARN Schedulers (Fair and Capacity). 
  • Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions. 
  • Setting up HBase cluster which includes master and region server configuration, High availability configuration, performance tuning and administration. 
  • Created user accounts and given users the access to the Hadoop cluster. 
  • Involved in loading data from UNIX file system to HDFS. 
  • Worked on ETL process and handled importing data from various data sources, performed transformations. 
  • Coordinate with QA team during testing phase. 
  • Provide application support to production support team. 
  • Implemented Hadoop stack and different Big Data analytic tools, migration from different databases to Hadoop. 
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios. 
  • Monitored workload, job performance and capacity planning.  Environment: Cloudera, HDFS, Hive, Sqoop, Zookeeper and HBase, Unix Linux, Java, HDFS, Map Reduce, Pig, Hive, HBase, Flume, Sqoop, Shell Scripting

Environment: Hadoop, HDFS, MapReduce, Hive, Oozie, Java (jdk1.6), Cloudera, MySQL, Windows Server 2008 R2/2012/2012 R2, MS SQL Server 2012/2008 R2, SSIS, SSRS, SSAS, Erwin, Visual Basic 6.0, SQL Azure, Crystal Reports and Ganglia.

Confidential, Newark, NJ

Hadoop Administrator

Responsibilities:
  • Install, Configure and maintain Single-node and Multi-node cluster Hadoop cluster. 
  • Interacted with Windows server management team to setup multiple virtual Windows Application server on a single Physical box. 
  • Setup cluster environment for Highly Available systems. 
  • Test failovers to secondary and failback to primary on VM clusters for Application and SQL VM's 
  • Test Disaster recovery system by routing traffic to alternate Disaster recovery server with help of LAN team.
  • Installed Apache Hadoop 2.5.2 and Apache Hadoop 2.3.0 on Linux Dev servers 
  • Upgrade Apache Hadoop from version 2.3.0 to 2.5.2 on Linux server 
  • Implementing High Availability system for Hadoop Name node 
  • Installing PIG, HIVE on multi-node cluster 
  • Configuring SQOOP to import data from external database - SQL Server and MYSQL. 
  • Configured users on Hadoop for HDFS and Map Reduce 
  • Setup Hive and NoSQL on remote metastore 
  • Integrating PIG, Hive, Sqoop on Hadoop
  • Monthly Linux server maintenance, shutting down essential Hadoop namenode and data node, job tracker and task tracker. And restarting Hadoop services including Yarn.
  • Plan and Maintain architecture of cluster
  • Assisted with performance tuning and monitoring.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios. 
  • Supported code/design analysis, strategy development and project planning. 
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
  • Work with user to resolve issues related to access and jobs running on cluster.
  • Security design and Implementation.

Confidential, New York, NY

Big Data Admin

Responsibilities:
  • Working as admin in Horton works (HDP 2.2.4.2) distribution for 3 clusters ranges from Dev, PreProd and PROD contains 100+ nodes.
  • Responsible for Cluster maintenance, Cluster Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
  • Day to day responsibilities includes solving developer issues and providing instant solution to reduce the impact and documenting the same and preventing future issues.
  • Installed and configured Apache Ranger and Apache Knox for securing HDFS, HIVE and HBASE.
  • Experience on new component installations and upgrading the cluster with proper strategies.
  • Experience on new Discovery Tools installation and integration with Hadoop Components.
  • Monitoring systems and services, architecture design and implementation of deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Hands on experience on cluster up gradation and patch upgrade without any data loss and with proper backup plans.
  • Installed, Configured, Managed Hadoop Cluster using Cloudera Manager
  • Supported data analyst in running Pig and Hive queries.
  • Performed Data scrubbing and processing with Oozie
  • Job management using Fair scheduler.
  • Efficiently performed Commissioning and Decommissioning of nodes on the Hadoop Cluster. 
  • Successfully Troubles hooted the cluster in case if any node has gone down. And was successfully able to fix it and balance the data distribution on the data nodes using balancer. 
  • Have Monitored the Hadoop Cluster using Ganglia and Nagios and configured them efficiently.

Confidential, San Francisco, CA

Jr. SQL Server DBA

Responsibilities:
  • Managing Production SQL Servers on VIRTUAL and PHYSICAL environments
  • Installing and Configuring SQL Server 2005/2008.
  • Supporting the business 24/7, maintaining 99.999 uptimes depending on the application and business SLA requirements.
  • Involved in upgrading SQL Server 2000 instances to SQL Server 2008.
  • Supported the configuration, deployment and administration of vendor and in-house applications interfacing with Oracle and SQL Server databases.
  • Supported for installation Oracle Applications R12/11i in single node & Multimode
  • Involved in database design, database standards, and Confidential -SQL code reviews.
  • Configured Active/Active and Active/passive SQL Server Clusters.
  • Implemented Mirroring and Log Shipping for Disaster recovery
  • Used Send Mail, Bulk Insert, Execute SQL, Data Flow, Import Export control extensively in SSIS
  • Performed Multi File Imports, Package configuration, Debugging Tasks and Scripts in SSIS
  • Scheduled package execution in SSIS
  • Installing packages on multiple servers in SSIS
  • Extensively worked on SSIS for data manipulation and data extraction, ETL Load
  • Created extensive reports using SSRS (Tabular, Matrix).
  • Configured transactional and snapshot replication and managed publications, articles.
  • Proactively involved in SQL Server Performance tuning Confidential Database level, Confidential -SQL level and operating system level. Maintained database response times and proactively generated performance reports.
  • Responsible for SQL Server Edition upgrades and SQL Server patch management.
  • Created a mirrored database for reporting using Database Mirroring with high performance mode.
  • Created database snapshots and stored procedures to load data from the snapshot database to the report database.
  • Involved in data modeling for the application and created ER diagrams using Erwin and VISIO.

Environment: MS SQL server 2000/2005/2008 , Windows 2003/2008, Enterprise Manager, Query Analyzer, SQL Server Profiler, SSIS, SSRS, Erwin, VISIO.

We'd love your feedback!