Resume
Hadoop Administrator Atlanta, GA
SUMMARY:
- Around 7 years of IT experience which include 5 years of experience in Hadoop Administration and 3 years of experience into Linux/UNIX Systems administration.
- Experienced in installation, configuration, supporting and monitoring Hadoop 2.X clusters using Confidential, Confidential and Confidential distributions.
- Experience with complete Software Design Lifecycle including design, development, testing and implementation of moderate to advanced complex systems.
- Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Design Big Data solutions for traditional enterprise businesses.
- Backup configuration and Recovery from a NameNode failure.
- Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off - line and on-line Backups.
- Experience in minor and major upgrades of Hadoop and Hadoop eco system
- Exceptional skills with NoSQL databases such as HBase and Cassandra.
- Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
- Debugging, troubleshooting Oracle PL/SQL packages, stored procedures and functions.
- Ability to work closely with Devops teams, in order to ensure high quality and timely delivery of builds and releases.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
TECHNICAL SKILLS:
- HDFS, Map Reduce, Pig, Hive, Hbase, Sqoop, Zookeeper, Oozie, Hue, HCatalog, Storm, Kafka, Key Value Store Indexer, and Flume.
- MySQL, Oracle 8i/9i/10g, SQLServer, PL/SQL.
- Hbase, Cassandra, Confidential Impala, Mongo DB.
- HDP Ambari, Confidential Manager, Hue, SolrCloud.
- Shell scripting, HTMLscripting, Puppet, Ansible.
- Apache Tomcat, JBOSS and Apache Http web server.
- Net Beans, Eclipse, Visual Studio, Microsoft SQL Server, MS Office.
- Kerberos, NagiOS & Ganglia
- Java, HTML, MVC, Struts, Hibernate, Servlet, spring, Web services.
- Windows XP, 7, 8, UNIX, MAC, MS DOS.
PROFESSIONAL EXPERIENCE:
HADOOP Administrator
Confidential - Atlanta, GA
Responsibilities:
- Working on Hadoop distributions like Confidential (CDH 5.7, CDH 5.8) to perform all operations related to Hadoop services.
- Handling the installation and configuration of a Hadoop cluster.
- Monitored workload, job performance and capacity planning using the Confidential Manager Interface.
- Manage Hadoop operations with multi-node HDFS cluster using Confidential Manager.
- Involved in setup, configuration and management of security for Hadoop clusters using Kerberos and integration with LDAP/AD at an Enterprise level.
- Spinning clusters in Azure using Confidential director, Implemented this for POC for the cloud migration project.
- Utilized Kafka and Flume to gain real-time data stream and save it in HDFS and HBase from the different data sources .
- Performed installation and configuration of Hadoop Cluster of 90 Nodes with Confidential distribution with CDH5.
- Setting up Hadoop clusters ( Confidential / Confidential ) and performing upgrades, configuration changes of hadoop clusters. Working with Linux commands to maintain Linux RedHat servers and adding them to the Hadoop environment to perform various data operations.
- Interacting with different application teams in providing hardware architectural guidance, planning and estimating cluster capacity/storage and creating roadmaps for Hadoop cluster deployment.
Environment: Confidential Manager, Ambari, HDFS, HBASE, Hive, Oracle, Sqoop, Flume, Spark, Hue, Kafka, AWS, RDBMS, DataRobot, Unravel.
HADOOP Administrator
Confidential, St. Louis, MO
Responsibilities:
- Responsible for adding/installation of new components and removal of them through HDP.
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
- Imported logs from web servers with Flume to ingest the data into HDFS.
- Working with data delivery teams to setup new Hadoop users. This job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive
- Involved in installing and configuring Confluent Kafka in R&D line, also Validate the installation with HDFS connector and Hive connectors.
- Creating event processing data pipelines and handling messaging services using Apache Kafka.
- Integrated Kafka with Flume in sand box Environment using Kafka source and Kafka sink.
- Experienced in Design, deploy and support highly available and scalable distributed Cassandra Datastax database solutions for high transactions mission critical applications.
- Worked on importing and exporting large data from Oracle and DB2 into HDFS and HIVE using Sqoop and automated the sqoop jobs by scheduling in Oozie.
- Apply Spark Streaming to receive data from Kafka to do the continually data cleaning and aggregating, then store the data in HBase .
- Expert knowledge of Kafka Topic, Partition, Producer and consumer.
- Configured internode communication between Cassandra nodes and client using SSL encryption.
- Installed, configured and deployed a 50 node Confidential Hadoop Cluster for Development and Production
- Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring, Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
Environment: HDFS, Map Reduce, Hive, AWS, Pig, GIT, Flume, Cassandra, kafka, Ambari, Oozie, Sqoop, HDP, Informatica, SOLR, python, Nosql, chef, Kerberos, MySQL and Oracle.
Hadoop Administrator
Confidential, Sunnyvale, CA
Responsibilities:
- Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Design the HBase schemes based on the requirements and HBase data migration and validation .
- Installed and configured Hadoop, MapReduce, HDFS (Hadoop Distributed File System), developed multiple MapReduce jobs for data cleaning.
- Involved in clustering of Hadoop in the network of 70 nodes.
- Experienced in loading data from UNIX local file system to HDFS.
- Developed data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Involved in collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis.
Environment: Environment: Hadoop, MapReduce, Hive, HDFS, PIG, Sqoop, Oozie, Confidential, Flume, HBase, ZooKeeper, CDH3, CDH4, MongoDB, Cassandra, Oracle, NoSQL and Unix/Linux.
Hadoop Administrator
Confidential
Responsibilities:
- Performed Benchmarking and performance tuning on the Hadoop infrastructure.
- Automated data loading between production and disaster recovery cluster. Migrated hive schema from production cluster to DR cluster.
- Provided on-site Linux System Administration support for the Hadoop Cluster and related software stack.
- Worked on Migrating application by doing POC's from relation database systems. Helping users and teams with incidents related to administration and development.
- Installed, Configured and Maintained Debian/RedHat Servers at multiple Data Centers. Configured RedHat Kickstart server for installing multiple production servers.
- Performed high-level, day-to-day operational maintenance, support, and upgrades for the Hadoop operating system, workstations and servers.
- Coordinate, direct and perform complex software installations and upgrades to operating systems and layered software packages.
- Worked on monitoring of VMware virtual environments with ESXi 4 servers and Virtual Center. Automated tasks using shell scripting for doing diagnostics on failed disk drives.
- Continually monitor and tune multiple systems to achieve optimum performance levels.
Environment: HDFS, Map Reduce, Linux Scripting, Shell Scripting, Zoo keeper, cluster health, monitoring security, RedHat Linux.
System Administrator
Confidential
Responsibilities:
- Patching of RHEL5 and Solaris 8, 9, 10 servers for EMC Powerpath Upgrade for VMAX migration.
- Configuration of LVM (Logical Volume Manager) to manage volume group, logical and physical partitions and importing new physical volumes.
- Documented the standard procedure for installation and deployment of VMAX Migration and logical volume manager.
- Installation, configuration, support and security implementation on following services: DHCP, SSH, SCP.
- Maintained and monitored all of company servers' operating system and application patch level, disk space and memory usage, user activities on day-to-day basis.
- User administration on Sun Solaris and RHEL systems, HP-UX machines, management & archiving.
- Installations of HP Open view, monitoring tool, in more than 300 servers.
- Attended calls related to customer queries and complaints, offered solutions to them.
- Creation of VMs, cloning and migrations of the VMs on VMware vSphere 4.0.
- Worked with DBA team for database performance issues, network related issue on Linux / Unix Servers and with vendors for hardware related issues.
Environment: Red Hat Enterprise Linux 4.x, 5.x, Sun Solaris 8, 9, 10, VERITAS Volume Manager, Oracle 11G, Samba, Oracle RAC/ASM,EMC Power path, DELL PowerEdge 6650, HP Proliant DL 385, 585, 580, Sun Fire v440, SUN BLADE X6250, X6270.