Hadoop Admin Resume
Peachtree City, GA
PROFESSIONAL SUMMARY:
- Senior Software Engineer with 8 years of experience in Hadoop Administration, Linux administration. Worked for various domains including financial, communications, healthcare and understanding the business needs of the applications. Support and Implementation of various stand - alone, client-server enterprise applications.
- Strong architectural understanding and working knowledge / hands-on experience in Hadoop components mainly HDFS, Zookeeper, YARN, HBASE, Hive, Impala, Solr, Spark, Kudu, Kafka.
- Expertise in HDFS - High availability configuration, SSL configuration, encryption zone creation and snapshot management
- Expertise in Hive, Hbase & impala -Load balancer configuration, HA configuration, performance tuning and troubleshooting.
- Responsible for implementation and ongoing administration of Hadoop infrastructure
- Provide technical leadership and collaboration with engineering organization, develop key deliverables for Data Platform Strategy - Scalability, optimization, operations, availability, roadmap.
- Security setup - Kerberos, TLS configuration on CM/agents and services, Load Balancer configuration for CM/hive/impala, KTS and KMS configuration, HSM configuration
- Own the platform architecture and drive it to the next level of effectiveness to support current and future requirements
- Cluster maintenance as well as creation and removal of nodes using tools like Cloudera Manager Enterprise, etc.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines
- Screen Hadoop cluster job performances and capacity planning
- Help optimize and integrate new infrastructure via continuous integration methodologies (DevOps - CHEF)
- Hands on experience on new cluster build, node addition, CDH upgrades, certificate renewal, BDR setup
- Manage and review Hadoop log files with the help of Log management technologies (ELK)
- Provide top-level technical help desk support for the application developers
- Install and upgrade Hortonworks HDP 3.1/Cloudera CDP versions ability to manage security and config changes, address audit/security/performance issues in the Hadoop clusters to address requirements
- Lead articulator in COE (Center of excellence) meetings for Hadoop issues and well aware of deploy/support for Spark, Hive, Solr, Presto and Nifi
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality, availability and security
- Collaborating with application teams to perform Hadoop updates, patches, version upgrades when required
- Clear understanding of Networks and Firewalls concepts.
- Mentor Hadoop engineers and administrators
- Work with Vendor support teams on support tasks
TECHNICAL SKILLS:
Cloud: Microsoft Azure, Amazon Web Services, PCF, OpenStack
Continuous Integration Tools: Jenkins, Hudson, AnthillPro, BuildForge, uBuild,TeamCity
Scripting Languages: Python, Java, HiveQL, Advanced PL/SQL, SQL, IDE's Splunk, Eclipse/Net Beans, SQL Developer, IntelliJ
Languages: C, C++, Core Java, Shell Scripting
Database: Oracle 11g/10g/9i, DB2, MS-SQL Server, MySQL, MS-Access, PostgreSQL, NoSQL (Mongo DB), SQLite3
Software Development Lifecycle : Agile Methodology, Scrum framework
Operating System: Red Hat, Ubuntu, Debian, Fedora, CentOS, Windows, Solaris, Mac OS
Repository Management: Nexus, JFrog, Artifactory
Build Tools: Maven, Ant, Gradle
Web and Application Servers: JBoss, Apache Tomcat, Web Sphere, Web Logic, Nginx
Container Orchestration Tools: Docker Swarm, Kubernetes, EC2 Container Services, Azure Container Services
PROFESSIONAL EXPERIENCE:
Confidential, Peachtree City, GA
Hadoop Admin
Responsibilities:
- Work with Linux administration team to provision set of Linux nodes required for Hadoop clusters
- Installing and configuring Hadoop software stack on set of nodes using automation deployment tools like Apache Ambari etc. based on various Hadoop distributions.
- Adding or decommissioning nodes from Hadoop cluster based on data load to be processed on the system.
- Configuring different Hadoop services like hdfs, mapreduce, hive, pig, sqoop, hbase, spark, storm, kafka, zookeeper etc
- Configured High Availability for Namenode, Resource Manager, Hbase Master etc. in Hadoop cluster to avoid single point of failure issue in the system.
- Identifying and configuring scheduler(fair/capacity) suitable for workload being scheduled on the cluster
- Monitoring health of Cluster components using monitoring tools like Apache Ambari, Nagios, Gangila, Cloudera manager etc.
- Performing periodic health checks on cluster to make sure system is available all the time for fulfilling business requirements.
- Monitoring the HDFS filesytem and logs effectively for any errors.
- Monitoring resources like CPU/Memory and storage on cluster nodes using metrics in automation tools like apache ambari.
- Changing the core Hadoop configuration files like hdfs-site.xml, core-site.xml, mapred-site.xml, yarn-site.xml and specific files related Hadoop services like hive, pig, spark etc. for any performance, scaling, application related issues etc.
- Taking the backup regularly and restore the data whenever required.
- Closely working with the infrastructure, networking, development teams to ensure the timely deliverables.
- Responsible for Configuring Alerts for different types of services which is running in Hadoop Ecosystem
Environment: Hortonworks HDP, MySQL, Oracle 11.2, Linux, Shell Scripting, Sqoop, Hive, Apache spark.
Confidential, Atlanta, GA
Hadoop Admin
Responsibilities:
- Creating and managing users in Hadoop cluster as part of user management.
- Work closely with the database team, network team, BI team and application teams to make sure that all the big data applications are highly available and performing as expected.
- Working on capacity planning activities like scale up/scale down nodes from cluster based on data load to be processed on the system.
- Prepare and maintain documents on configuration changes and best practices.
- Troubleshooting various issues like space related, memory related, application related, different types of failures, performance related, spark jobs execution issues, high availability related, edge node related issues etc.
- Tuning the Cluster Resources, Optimizing Map Reduce Jobs and Benchmarking.
- Rebalancing the data across the data nodes in the cluster using balancer tool.
- Applying patches related to security, stability, fixes, features periodically to Hadoop services whenever they are available.
- Upgrading Hadoop stack on Hadoop cluster to latest version to fulfil the business requirements
- Installing patches and upgrading software as and when needed.
- Data modelling like designing and implementation of data based on recognized standards for Hadoop Data Storage.
- Troubleshoot and debug Hadoop eco system runtime issues.
- Recovering from node failures and troubleshooting common Hadoop cluster issues.
- Document all production scenarios, issues and resolutions.
- Managing and monitoring Hadoop services like Namenode, Datanode & Yarn.
- Performance tuning and solving Hadoop issues using CLI or by WebUI.
- Backup and recovery task by Creating Snapshots Policies, Backup Schedules and recovery from node failure.
- Monitoring HDFS health and taking necessary actions for maintaining it
Environment: Cloudera, AWS, GIT, Jenkins, Linux, Python/Shell script, Tomcat, SVN, Gradle, Nagios, Splunk
Confidential, Northern, NJ
Hadoop Admin
Responsibilities:
- Evaluation of Hadoop infrastructure requirements and design/deploy solutions (high availability, big data clusters, etc.)
- Hadoop integration with large scale distributed data platforms like Teradata, Teradata Aster, Vertica, Greenplum, Netezza, DB2, Oracle, etc.
- Collaborating with application teams to install operating system and Hadoop updates, patches and version upgrades
- Responsible for Designing, Implementation and Administration of Hadoop infrastructure
- Responsible for supporting, managing & monitoring the Cloudera Clusters
- Install, upgrade and configure Hadoop ecosystem tools
- Develop tools and utilities to administer Hadoop jobs, manage queues and monitor performance
- Proactively monitor cluster health and perform performance tuning activities
- Develop process for managing configurations between environments
- Perform capacity planning and expansion activities working across infrastructure and other enterprise services teams
- Working with data delivery teams to setup new Hadoop users
- Involved in Cluster maintenance as well as creation and removal of nodes
- HDFS support and maintenance
- Configuring Kerberos security for Hadoop cluster
- Performance Tuning of Hadoop configurations for optimal utilization of cluster resources
- Troubleshooting Hadoop cluster runtime errors and ensuring that they do not occur again.
- Accountable for storage and volume management of Hadoop clusters.
- Ensuring that the Hadoop cluster is up and running all the time (High availability, big data cluster etc.)
- Evaluation of Hadoop infrastructure requirements and design/deploy solutions
Environment: Hadoop, Ambari, MYSQL, Shell Scripts, Git, Linux, Unix and Windows.
Confidential, DES MOINES, IA
Linux Administrator
Responsibilities:
- Administered RedHat Enterprise Linux 5.x/4.x, OEL 5.x, & Solaris 9 Servers by testing, tuning, upgrading, patching and troubleshooting both Physical & Virtual server problems.
- Used Veritas File system and Veritas Volume Manager 5.0 to configure the RAID 1 & RAID 5 Storage System for more redundancy.
- Installed and maintained regular upgrades of Red Hat Linux Servers using kickstart based network installation.
- Created Disk volumes, Volume groups and Logical volumes (LVM) for Linux operating systems.
- Installed and Configured Apache Tomcat Web Server.
- Configured Proxy Server (Squid), DNS, FTP and DHCP servers on Red Hat Enterprise Linux and maintained system securities using IPTABLES.
- Developed Perl & Shell scripts for automation of the build and release process. Developed automation scripting in Python to deploy some applications.
- Created the LDAP Scripts which monitor the LDAP connectivity and alerts the Admin Group if the connection is closed. Also administered Active Directory Servers.
- Involved in monitoring and troubleshooting Network like TCP/IP, NFS, DNS and SMTP in Linux servers and System Activities like CPU, Memory, Disk and Swap space usage to avoid any performance issues.
- Server, MySQL databases in both development and production.
- Experience in implementing and configuring network services such as HTTP, DHCP, and TFTP.
- Install and configure DHCP, DNS (BIND, MS), web (Apache, IIS), mail (SMTP, IMAP, POP3), and file servers on Linux servers.
- Built ESXi hosts using multiple technologies including HPSA, VUM, Host Profiles, and PowerCLI scripts Performed routine maintenance on VMware environment such as vCenter upgrades, firmware upgrades, Patching.
- Troubleshooting Backup and Restore Problems Creation of LVMs on SAN using Linux utilities
- Troubleshooting Linux network, security-related issues, capturing packets using tools such as IP tables, firewall, and TCP wrapper and NMAP.
- By integrating WLST scripts to Shell scripts, artifacts like war, ear are deployed into WebLogic app server.
- Upgraded Red hat Linux OS on Web Sphere, JBoss and Oracle database servers from V3, V4 to V5.
- Monitored servers, switches, ports etc. with Nagios monitoring tool.
- Responsible for setting up Cronjobs scripts on production servers and Implementation of passwordless (SSH) authentication between servers.
Environment: RedHat Enterprise Linux 5.x/4.x, OEL 5.x, Solaris 9, LVM, RAID, Cronjobs, Oracle, MySQL, TCP/IP
Confidential, CHARLOTTE, NC
Unix Administrator
Responsibilities:
- Design and maintain all system tools for all scripts and automation processes and monitor all capacity planning.
- Integrate all required software and resolve all issues across various technologies and design require enterprise servers and provide back up support.
- Evaluate all documents according to system requirements and evaluate all design and perform tests on all development activities and administer all complex methodologies.
- Develop an infrastructure to provide support to all business requirements and perform regular troubleshoot on system to resolve all issues.
- Monitor everyday systems and evaluate availability of all server resources and perform all activities for Linux servers.
- Assist to configure and deploy all virtual machines and install and provide backup to all configuration procedures.
- Implemented and setup Virtualization environments for AIX LPARs, HP Integrity VMs; and Solaris Zones and Logical Domains
- Updated and created provisioning scripts to setup new operating systems and software for supported platforms
- Consolidated servers at numerous smaller remote data centers to three central data centers
- Stabilized system by disk replacement, firmware upgrade in SAN storage, Solaris Volume Management, clustering environment on scheduled maintenance hours.
- Enhanced business continuity procedure by adding critical middleware server and identified through power-down test activity.
- Resolved issues, planned requests as point-of-contact to vendors, oversaw developers, business users, following change control procedure, reported result Monitor everyday systems and evaluate availability of all server resources and perform all activities for Linux servers.
- Maintain and monitor all patch releases and design various patch installation strategies and maintain all systems according to NIST standardization.
- Administer all performance for various resources and ensure optimization for same and provide support to all applications and ensure optimal level of customer services.
- Maintain and monitor all system frameworks and provide after call support to all systems and maintain optimal Linux knowledge.
- Perform troubleshoot on all tools and maintain multiple servers and provide back up for all files and script management servers.
- Wrote and maintained shell scripts using Perl and Bash.
- Monitored, troubleshot, and resolved issues involving operating systems.
- Applied ITIL approach to incident and problem management.
- Developed and maintained troubleshooting journal for incident management team.
- Participated in on-call rotation to provide 24×7 technical support
- Tested numerous software and hardware configurations during the development stages to recreate the operating environments utilized by customers in an effort to avoid the distribution of releases with bugs and/or erroneous documentation
- Wrote utility scripts using BASH and KORN shell
- Configured UNIX systems to use Active directory, KERBEROS, NTPD, XDMCP, LDAP, SSH, FTP, TFTP and DNS
- Performed problem diagnosis, corrected discrepancies, developed user and maintenance documentation, provided user assistance and evaluated system performance
- Installed, configured third party applications, hardened new and existing servers and desktops
Environment: KERBEROS, monitoring tool HP OpenView ITO (OVO), Redhat Linux, Windows, FTP, Solaris, HP UX with Oracle, Sybase