We provide IT Staff Augmentation Services!

Resume

5.00/5 (Submit Your Rating)

Hadoop Administrator Atlanta, GA

SUMMARY:

  • Around 7 years of IT experience which include 5 years of experience in Hadoop Administration and 3 years of experience into Linux/UNIX Systems administration.
  • Experienced in installation, configuration, supporting and monitoring Hadoop 2.X clusters using Confidential, Confidential and Confidential distributions.
  • Experience with complete Software Design Lifecycle including design, development, testing and implementation of moderate to advanced complex systems.
  • Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Design Big Data solutions for traditional enterprise businesses.
  • Backup configuration and Recovery from a NameNode failure.
  • Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off - line and on-line Backups.
  • Experience in minor and major upgrades of Hadoop and Hadoop eco system
  • Exceptional skills with NoSQL databases such as HBase and Cassandra.
  • Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
  • Debugging, troubleshooting Oracle PL/SQL packages, stored procedures and functions.
  • Ability to work closely with Devops teams, in order to ensure high quality and timely delivery of builds and releases.
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.

TECHNICAL SKILLS:

  • HDFS, Map Reduce, Pig, Hive, Hbase, Sqoop, Zookeeper, Oozie, Hue, HCatalog, Storm, Kafka, Key Value Store Indexer, and Flume.
  • MySQL, Oracle 8i/9i/10g, SQLServer, PL/SQL.
  • Hbase, Cassandra, Confidential Impala, Mongo DB.
  • HDP Ambari, Confidential Manager, Hue, SolrCloud.
  • Shell scripting, HTMLscripting, Puppet, Ansible.
  • Apache Tomcat, JBOSS and Apache Http web server.
  • Net Beans, Eclipse, Visual Studio, Microsoft SQL Server, MS Office.
  • Kerberos, NagiOS & Ganglia
  • Java, HTML, MVC, Struts, Hibernate, Servlet, spring, Web services.
  • Windows XP, 7, 8, UNIX, MAC, MS DOS.

PROFESSIONAL EXPERIENCE:

HADOOP Administrator

Confidential - Atlanta, GA

Responsibilities:

  • Working on Hadoop distributions like Confidential (CDH 5.7, CDH 5.8) to perform all operations related to Hadoop services.
  • Handling the installation and configuration of a Hadoop cluster.
  • Monitored workload, job performance and capacity planning using the Confidential Manager Interface.
  • Manage Hadoop operations with multi-node HDFS cluster using Confidential Manager.
  • Involved in setup, configuration and management of security for Hadoop clusters using Kerberos and integration with LDAP/AD at an Enterprise level.
  • Spinning clusters in Azure using Confidential director, Implemented this for POC for the cloud migration project.
  • Utilized Kafka and Flume to gain real-time data stream and save it in HDFS and HBase from the different data sources .
  • Performed installation and configuration of Hadoop Cluster of 90 Nodes with Confidential distribution with CDH5.
  • Setting up Hadoop clusters ( Confidential / Confidential ) and performing upgrades, configuration changes of hadoop clusters. Working with Linux commands to maintain Linux RedHat servers and adding them to the Hadoop environment to perform various data operations.
  • Interacting with different application teams in providing hardware architectural guidance, planning and estimating cluster capacity/storage and creating roadmaps for Hadoop cluster deployment.

Environment: Confidential Manager, Ambari, HDFS, HBASE, Hive, Oracle, Sqoop, Flume, Spark, Hue, Kafka, AWS, RDBMS, DataRobot, Unravel.

HADOOP Administrator

Confidential, St. Louis, MO

Responsibilities:

  • Responsible for adding/installation of new components and removal of them through HDP.
  • Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
  • Imported logs from web servers with Flume to ingest the data into HDFS.
  • Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive
  • Involved in installing and configuring Confluent Kafka in R&D line, also Validate the installation with HDFS connector and Hive connectors.
  • Creating event processing data pipelines and handling messaging services using Apache Kafka.
  • Integrated Kafka with Flume in sand box Environment using Kafka source and Kafka sink.
  • Experienced in Design, deploy and support highly available and scalable distributed Cassandra Datastax database solutions for high transactions mission critical applications.
  • Worked on importing and exporting large data from Oracle and DB2 into HDFS and HIVE using Sqoop and automated the sqoop jobs by scheduling in Oozie.
  • Apply Spark Streaming to receive data from Kafka to do the continually data cleaning and aggregating, then store the data in HBase .
  • Expert knowledge of Kafka Topic, Partition, Producer and consumer.
  • Configured internode communication between Cassandra nodes and client using SSL encryption.
  • Installed, configured and deployed a 50 node Confidential Hadoop Cluster for Development and Production
  • Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring, Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.

Environment: HDFS, Map Reduce, Hive, AWS, Pig, GIT, Flume, Cassandra, kafka, Ambari, Oozie, Sqoop, HDP, Informatica, SOLR, python, Nosql, chef, Kerberos, MySQL and Oracle.

Hadoop Administrator

Confidential, Sunnyvale, CA

Responsibilities:

  • Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
  • Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
  • Design the HBase schemes based on the requirements and HBase data migration and validation .
  • Installed and configured Hadoop, MapReduce, HDFS (Hadoop Distributed File System), developed multiple MapReduce jobs for data cleaning.
  • Involved in clustering of Hadoop in the network of 70 nodes.
  • Experienced in loading data from UNIX local file system to HDFS.
  • Developed data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
  • Involved in collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis.

Environment: Environment: Hadoop, MapReduce, Hive, HDFS, PIG, Sqoop, Oozie, Confidential, Flume, HBase, ZooKeeper, CDH3, CDH4, MongoDB, Cassandra, Oracle, NoSQL and Unix/Linux.

Hadoop Administrator

Confidential

Responsibilities:

  • Performed Benchmarking and performance tuning on the Hadoop infrastructure.
  • Automated data loading between production and disaster recovery cluster. Migrated hive schema from production cluster to DR cluster.
  • Provided on-site Linux System Administration support for the Hadoop Cluster and related software stack.
  • Worked on Migrating application by doing POC's from relation database systems. Helping users and teams with incidents related to administration and development.
  • Installed, Configured and Maintained Debian/RedHat Servers at multiple Data Centers. Configured RedHat Kickstart server for installing multiple production servers.
  • Performed high-level, day-to-day operational maintenance, support, and upgrades for the Hadoop operating system, workstations and servers.
  • Coordinate, direct and perform complex software installations and upgrades to operating systems and layered software packages.
  • Worked on monitoring of VMware virtual environments with ESXi 4 servers and Virtual Center. Automated tasks using shell scripting for doing diagnostics on failed disk drives.
  • Continually monitor and tune multiple systems to achieve optimum performance levels.

Environment: HDFS, Map Reduce, Linux Scripting, Shell Scripting, Zoo keeper, cluster health, monitoring security, RedHat Linux.

System Administrator

Confidential

Responsibilities:

  • Patching of RHEL5 and Solaris 8, 9, 10 servers for EMC Powerpath Upgrade for VMAX migration.
  • Configuration of LVM (Logical Volume Manager) to manage volume group, logical and physical partitions and importing new physical volumes.
  • Documented the standard procedure for installation and deployment of VMAX Migration and logical volume manager.
  • Installation, configuration, support and security implementation on following services: DHCP, SSH, SCP.
  • Maintained and monitored all of company servers' operating system and application patch level, disk space and memory usage, user activities on day-to-day basis.
  • User administration on Sun Solaris and RHEL systems, HP-UX machines, management & archiving.
  • Installations of HP Open view, monitoring tool, in more than 300 servers.
  • Attended calls related to customer queries and complaints, offered solutions to them.
  • Creation of VMs, cloning and migrations of the VMs on VMware vSphere 4.0.
  • Worked with DBA team for database performance issues, network related issue on Linux / Unix Servers and with vendors for hardware related issues.

Environment: Red Hat Enterprise Linux 4.x, 5.x, Sun Solaris 8, 9, 10, VERITAS Volume Manager, Oracle 11G, Samba, Oracle RAC/ASM,EMC Power path, DELL PowerEdge 6650, HP Proliant DL 385, 585, 580, Sun Fire v440, SUN BLADE X6250, X6270.

We'd love your feedback!