Sr. Hadoop Administrator Resume
Herndon, VA
SUMMARY
- Around 7+ years of professional experience in full life cycle system development and administration in which
- 2.5 Years of experience in Hadoop Administration & Big Data Technologies and 6 years of experience into Cognos/UNIX administration/development and ETL Informatica.
- Experience with complete Software Design Lifecycle including design, development, testing and implementation of moderate to advanced complex systems.
- Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Horton works, Cloudera and MapR.
- Having very good understanding and working knowledge of Hadoop Eco systems such as pig, hive, HBase, Spark, Sqoop and zookeeper.
- Good understanding on cluster capacity planning and configuring the cluster components based on requirements.
- Responsible managing and administering Tableau and Green Plumdatabases instances in various environments.
- Extensive knowledge on performance tuning, cluster Monitoring, Troubleshooting.
- Monitoring performance, disk space, memory usage on the Teradataand ETL server.
- Design Big Data solutions for traditional enterprise businesses.
- Used Network Monitoring Daemons like Nagios, Ganglia.
- Loaded data from SQL database, MS SQL server into HDFS, HIVE.
- Persistence is based on PostgresSQL.
- I designed the schemas for all the services as per application requirements and deployed a clustered setup on Hortonworks as well as planned for proper sharing with growth.
- Adding/removing new nodes to an existing Hadoop cluster with proper plan without data loss.
- Backup configuration and Recovery from a NameNode failure.
- Decommissioning and commissioning the Node on running hadoop cluster.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Installation and configuration of all Hadoop eco system components.
- Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off - line and on-line Backups.
- Guiding new teams for their developments and setting up the environment.
- Making Hadoop cluster ready for development team working on POCs.
- Experience in minor and major upgrades of hadoop and hadoop eco system.
- Experience in monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
- As an admin involved in Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
- Experience in HDFS data storage and support for running map-reduce jobs.
- Installing and configuring hadoop eco system like sqoop, pig, hive.
- Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
- Optimizing performance of Hbase/Hive/Pig jobs.
- Rack aware configuration for quick availability and processing of data.
- Handsome experience in Linux admin activities on RHEL &Cent OS.
- Extensive knowledge on Data warehousing concepts, reporting, relational data bases.
TECHNICAL SKILLS
Hadoop Framework: HDFS, Map Reduce, Yarn, Pig, Hive, HBase, Sqoop, zookeeper, Oozie, Flume, Spark, Storm, Knox, Nagios, Ambari, Ranger, Knox, Hue, Stack iq.
Congo s Tools: Cognos 10.2/10.1.1/8.4 , 7.x/8.x Suite, Report Net.
Microsoft: MS Office, MS Project, MS Visio, MS Visual Studio 2003/ 2005/ 2008.
Databases: Oracle 8i/9i/10g, SQL Server, PL/SQL Developer, PostgresSQL.
Operating Systems: Linux Cent OS,RHEL,Windows 2000/2003/2008/ XP/Vista
Scripting: Shell Scripting, HTML Scripting.
Programming: C, C++, Core Java, PL/SQL.
PROFESSIONAL EXPERIENCE
Confidential, Herndon,VA
Sr. Hadoop Administrator
Responsibilities:
- Currently working as Hadoop Admin and responsible for taking care of everything related to the clusters total of 100 nodes ranges from POC to PROD clusters.
- Experienced on setting up hortonworks cluster and installing all the ecosystem components through Ambari and manually from command line.
- Hand on experience on cluster upgrade and patching without any data loss and with proper backup plans.
- Good experience on enabling NAMENODE HA and YARN capacity schedulers to have HA in cluster and maintaining the cluster resources for balancing the applications in cluster.
- Responsible for cluster Maintenance, Monitoring, Troubleshooting, Tuning, commissioning and Decommissioning of nodes.
- Responsible for adding new eco system components, like spark, storm, flume, Knox with required custom configurations based on the requirements.
- Maintaining the cluster Meta Data backups and make sure we have regular backups to provide high availability of cluster and Manage and review data backups.
- Monitoring systems and services, architecture design and implementation of hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
- Experienced on guiding the upcoming applications in hadoop and guiding them through the entire process like providing the access and creating standard directory setups, service accounts setup and prod deployments and helping them with the tools they are using in cluster to provide better performance.
- Performance tuning of the cluster to increase the jobs performance based on the server hardware availability by using benchmarking activity.
- Changing the configurations based on the requirements of the users for the better performance of the jobs.
- Provided security to the cluster with AMBARI ranger and created policies for each team to maintain security of project data between different teams in hadoop cluster.
- Worked on integration of tools like Informatica, Attunity, Greenplum, Tableau, SSRS reporting and IDQ, SAS with Hadoop.
- Installed and configured flume to get the messages from MDM servers to Hadoop for business analysis.
- Experienced in managing and reviewing Hadoop log files.
- Good knowledge on maintain Mysql databases creation and setting up the users and maintain the backup of databases.
- Good knowledge on importing data import/export from and to Hadoop from data oracle and other data sources.
- Helping the users in production deployments throughout the process.
- Experienced in production support which involves solving the user incidents varies from sev1 to sev5.
- Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes. Communicate and escalate issues appropriately.
- As an admin followed standard Back up policies to make sure the high availability of cluster.
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions, Documented the systems processes and procedures for future s.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Monitoring Hadoop clusters with monitoring tools like Ambari alerts, metrics, Nagios and ganglia to check the regular health of cluster and to take immediate action if anything goes down.
Environment: Hortonworks2.2.4.2,Ambari,Hive,Pig,Sqoop,Zookeeper,Hbase,Ranger,Knox,Spark,Flume,Yarn,Mapreduce2
Confidential, San Jose,CA
Hadoop Administrator
Responsibilities:
- Currently working as Hadoop administrator in Hortonworks distribution for 3 clusters ranges from POC clusters to PROD clusters.
- Responsible for Cluster maintenance, Cluster monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
- Day to day responsibilities includes solving developer issues and providing instant solution to reduce the impact and documenting the same and preventing future issues.
- Installed/Configured/Maintained Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
- Extensively involved in Installation and configuration of Cloud era distribution Hadoop CDH 3.x and CDH 4.x.
- Experience on new component installations and upgrading the cluster with proper strategies.
- Experience on new Discovery Tools installation and integration with Hadoop Components.
- Monitoring systems and services, architecture design and implementation of deployment, configuration management, backup, and disaster recovery systems and procedures.
- Hand on experience on cluster up gradation and patch upgrade without any data loss and with proper backup plans.
- Changing the configurations based on the requirements of the users for the better performance of the jobs.
- Worked on configuration management tools like stack IQ to maintain central and pushing the configurations to the overall cluster for all configuration files like mapred-site.xml, pools.xml, and hdfs-site.xml.
- Experienced in Setting up the project and volume setups for the new projects.
- Involved in snapshots and mirroring to maintain the backup of cluster data and even remotely.
- Implementing the SFTP for the projects to transfer data from External servers to servers.
- Installation of various Eco system sand Daemons.
- Experienced in managing and reviewing log files.
- Working experience on maintaining MySQL, SQL databases creation and setting up the users and maintain the backup of cluster metadata databases.
- Setting up MySQL master and slave replications and helping business applications to maintain their data in MySQL Servers.
- Helping the users in production deployments throughout the process.
- Experienced in production support which involves solving the user incidents varies from sev1 to sev5.
- Managed and reviewed Log files as a part of administration for troubleshooting purposes. Communicate and escalate issues appropriately.
- As an admin followed standard Back up policies to make sure the high availability of cluster.
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions. Documented the systems processes and procedures for future s.
- Worked with systems engineering team to plan and deploy new environments and expand existing clusters.
- Monitored multiple clusters environments using AMBRI Alerts, Metrics and Nagios.
Environment: HDP 2.2.4.2, HUE, Hive, Pig, Sqoop, Flume, Zookeeper and HBase, MySQL Shell Scripting, Red hat Linux.
Confidential, SFO
LINUX Administrator
Responsibilities:
- Managing and maintaining Linux running servers in same environment.
- Responsible for troubleshooting the issues on the servers and provide a solution in a timely fashion.
- Install / Configure VIO servers as well as configuring virtual Ethernet shared Ethernet (SEA), virtual SCSI and NPIV on Confidential Power 6 and 7 systems.
- Taking care of changes made on servers.
- Administration of Red Hat Linux DNS, Web Server.
- Building software packages on Red Hat Linux (RPM).
- Installed, upgraded and configured SUN Solaris 9/10 on Sun Servers.
- Responsible for Listening to the Confidential and provide a solution within SLA.
- Troubleshoot and Fix various OS issues faced by our users/clients. And work with multiple teams to fix the issue.
- Developed KORNand Bash shell-scripts to automate routine activities.
- Troubleshooting issues related to NFS, SSH, NIS, DNS, FTP, VMWARE, NETBACKUP, VERITAS (VCS), ZONES, LVM, RAID, File System, Permissions, Performance Monitoring, IP Bonding, Multipathing, NAS, SAN, Storage, V-Center, Opsware, Power Broker, etc.
- Deployed Puppet master to the server and cached the puppet agent into it.
- Rectifying hardware failure and coordinating with vendors like Symantec, Oracle, Dell, and HP to get them repaired.
- Involve multiple teams sometime for providing a quick resolution to a high priority ticket.
- Creating profiles manifests for new servers, and pushing them to servers using puppet.
- Used Puppet tool for managing system configurations.
- Write Shell/PERL/Python scripts to automate repetitive tasks.
Environment: Red Hat Linux 3,4.x,5,6 VMware ESX 3.5, VERITAS Volume Manager, Sun Solaris 8,9,10, Solaris Volume Manager, VERITAS Cluster Server, Web logic 8.1, Samba, Oracle 9i/10g.
Confidential, NY
Linux Administrator
Responsibilities:
- Administration of RHEL 5/6 which includes installation, testing, tuning, upgrading and loading patches, troubleshooting server issues.
- Configure and automate the deployment ofLinuxand VMware infrastructure through our existing Kick start infrastructure.
- ConfigureLinuxguests in a VMware ESX environment.
- Understand server virtualization technology such as VMware.
- Worked on Cisco USC, virtual infra on VMware, Storage migration and installations.
- Installing, configuring, custom building Oracle10g and preparing servers for database installation which includes adding kernel parameters, software installation, permissions etc.
- Implemented multi-tier application provisioning in Open Stack cloud, integrating it with Puppet.
- Involved in integrated Vsphere hypervisor with Open Stack.
- Configure and maintained FTP, DNS, NFS and DHCP servers.
- Configuring, maintaining and troubleshooting of local development servers.
- Performed configuration of standardLinuxand network protocols, such as SMTP, DHCP, DNS, LDAP, NFS, SMTP, HTTP, SNMP and others.
- Writing shell scripting for automation.
- Worked on virtual and physicalLinuxhost for decommission.
- ServerAdministratorTomcat, Tomcat serving dynamic servlet and JSP requests.
- Managing corn jobs, batch processing and job scheduling.
- Worked on planning for the recovery of critical IT systems and services in a fallback situation following a disaster that overwhelms the resilience arrangements.
- Monitoring System Activities like CPU, Memory, Disk and Swap space usage to avoid any performance issues.
- Tuning the Kernel parameters for the better performance of applications like Oracle.
- Provided 24X7 on-calls production and customer support including trouble shooting problems.
Environment: LINUX, FTP, Shell, UNIX, VMware, NFS, TCP/IP, Oracle Red Hat Linux.