Senior Hadoop Administrator Resume
Detroit, MI
SUMMARY:
- 8 Years of extensive IT experience with 4 years of experience as a Hadoop Administrator and 4 years of experience in UNIX/Linux Administrator along with SQL developer in designing and implementing Relational Database model as per business needs in different domains.
- Experience in installation, management and monitoring of Hadoop cluster using Cloudera Manager.
- Optimized the configurations of Map Reduce, pig and hive jobs for better performance.
- Backup configuration and Recovery from a Namenode failure.
- Expert level skills in Managing and Scheduling Jobs on a Hadoop cluster.
- Ability to think creatively to help design innovative solutions to complex analytical questions.
- Extensive experience in installation, configuration, management and deployment of Big Data components and the underlying infrastructure of Hadoop Cluster.
- Good working knowledge on importing and exporting data from different databases namely MySQL, PostgreSQL, Oracle into HDFS and Hive using Sqoop.
- Extensive experience in NoSQL and real time analytics.
- Strong knowledge on yarn terminology and the High - Availability Hadoop Clusters.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience in Chef, Puppet or related tools for configuration management.
- Involved in balancing the loads on server and tuning of server for optimal performance of the cluster.
- Expertise in Installing, Configuration and Managing Red hat Linux 5, 6.
- Good experience on scheduling cron jobs in Linux.
- Worked with PLSQL stored procedures in creating reports, which required modified data input from the source.
- Experience in encryption security layer in the Hadoop environment.
TECHNICAL SKILLS:
Hadoop Framework: HDFS, Map Reduce, Hive, Pig, Zookeeper, Sqoop, Hbase, Flume
OS: RedHat Linux, UNIX, Windows 2000/NT/XP
Languages: C, C++, SAS, PL/SQL
Scripting Languages: Unix Shell scripting
Database: Oracle 10g/11g, SQL server, Teradata
Database Tools: Oracle SQL Developer, SQL Plus
Version Control: CVS, SVN
PROFESSIONAL EXPERIENCE:
Senior Hadoop Administrator
Confidential, Detroit, MI
Responsibilities:
- Involved in start to end process of Hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
- Responsible for Cluster maintenance & Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage & review data backups, Manage & review Hadoop log files.
- Monitoring systems & services, architecture design & implementation of Hadoop deployment, configuration management, backup and disaster recovery systems and procedures.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Responsible for Installation and configuration of Hive, Pig, HBase and sqoop on the Hadoop cluster.
- Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml based upon the job requirement.
- Involved in loading data from UNIX file system to HDFS.
- Provisioning, installing, configuring, monitoring, and maintaining HDFS, Yarn, HBase, Flume, Sqoop, Oozie, Pig, Hive.
- Implemented Hadoop stack and different bigdata analytic tools, migration from different databases to Hadoop.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning.
- Upgraded Hadoop to 2.7.2 version.
- Expertise in recommending hardware configuration for Hadoop cluster.
- Installing, Upgrading and Managing Hadoop Cluster on Cloudera distribution.
- Trouble shooting many cloud related issues such as Data Node down, Network failure and data block missing.
- Monitored and configured a test cluster on amazon web services for further testing process and gradual migration.
- Managing and reviewing Hadoop and HBase log files.
- Experience with Unix or Linux, including shell scripting.
- Loading the data from the different Data sources like (Teradata and DB2) into HDFS using sqoop and load into Hive tables, which are partitioned.
- Developed Hive UDF’s to bring all the customers information into a structured format.
- Developed bash scripts to bring the Tlog files from ftp server and then processing it to load into hive tables.
- Built automated set up for cluster monitoring and issue escalation process.
- Administration, installing, upgrading and managing distributions of Hadoop (CDH3, CDH4, Cloudera manager), Hive, HBase.
Environment: Hadoop 2.7.2, HDFS, Map Reduce, Shell Scripting, spark, solr, Pig, Hive, HBase, Sqoop, Flume, Oozie, Zoo keeper, cluster health, monitoring security, Redhat Linux 7.0, impala, Cloudera Manager5.5.4, Hortonworks.
Senior Hadoop Administrator
Confidential, Woburn, MA
Responsibilities:
- Capturing data from existing databases that provide SQL interfaces using Sqoop.
- Implemented Hadoop stack and different bigdata analytic tools, migration from different databases to Hadoop.
- Processed information from Hadoop HDFS. This information will comprise of various useful insights that can be used in the decision making process. All these insights will be presented to the users in the form of Charts.
- Working on different Big Data technologies, good knowledge of Hadoop, Map-Reduce, Hive.
- Developed various POCs over Hadoop, Big data.
- Worked on deployments and automation task.
- Installed and configured Hadoop cluster in pseudo and fully distributed mode environments.
- Involved in developing the data loading and extraction processes for big data analysis.
- Worked on professional services engagements to help customers design, build clusters, applications, troubleshoot network, disk and operating system related issues.
- Administered linux servers, other unix variants, and managed hadoop clusters.
- Installed and configured local Hadoop Cluster with 3 nodes and set up 4 nodes cluster on EC2 cloud.
- Written MapReduce code to process and parsing the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
- Work with HBase and Hive scripts to extract, transform and load the data into HBase and Hive.
- Continuous monitoring and managing of the Hadoop cluster.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
- Installed Oozie workflow engine to run multiple Hive and Pig jobs.
- Developing scripts and batch job to schedule a bundle (group of coordinators) which consists of various Hadoop programs using Oozie.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports.
Environment: Hadoop2.5.2, HDFS, Map Reduce, Hive, Flume, Sqoop, Cloudera CDH4, HBase, Oozie, Pig, AWS EC2 cloud.
Hadoop Administrator
Confidential, Mountain View, CA
Responsibilities:
- Responsible for implementation and ongoing administration of Hadoop infrastructure.
- Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
- Installed and implemented the monitoring tools like ganglia and Nagios on both the clusters.
- Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise.
- Build a new sandbox cluster for the testing purpose and move data from secure cluster to insecure sandbox cluster by using distcp.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
- Screen Hadoop cluster job performances and capacity planning.
- Monitor Hadoop cluster connectivity and security.
- Manage and review Hadoop log files.
- HDFS support and maintenance.
- Written MapReduce code to process and parsing the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
- Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Environment: Cloudera Manager, CDH 5.4.3, HDFS, Yarn, Hue, Sentry, Oozie, Zoo keeper, Impala, Solr, Kerberos, cluster health, Puppet, Ganglia, Nagios, Flume, Sqoop, storm, Kafka, KMS.
Hadoop/System Administrator
Confidential, San Jose, CA
Responsibilities:
- Responsible for implementation and ongoing administration of Hadoop infrastructure.
- Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when require.
- Installed and implemented the monitoring tools like ganglia and Nagios on both the clusters.
- Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise.
- Build a new sandbox cluster for the testing purpose and move data from secure cluster to insecure sandbox cluster by using distcp.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
- Screen Hadoop cluster job performances and capacity planning.
- Monitor Hadoop cluster connectivity and security.
- Manage and review Hadoop log files.
- HDFS support and maintenance.
- Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Environment: Cloudera Manager, CDH 5.1.6, HDFS, Hue, Oozie, Zoo keeper, Kerberos, cluster health, Puppet, Ganglia, Nagios, Flume, Sqoop.
Linux System/ Oracle DBA Admin
Confidential
Responsibilities:
- Worked on Installation, configuration and upgrading of Oracle server software and related products.
- Responsible for installation, administration and maintenance of Linux servers.
- Established and maintain sound backup and recovery policies and procedures.
- Worked with Telnet, FTP, TCP/IP, rlogin, used to inter-operate hosts.
- Take care of the Database design and implementation.
- Implement and maintain database security (create and maintain users and roles, assign privileges).
- Performed database tuning and performance monitoring.
- Worked as part of a team and provide 7x24 support when required.
- Performed general technical trouble shooting on trouble tickets to bring to resolution.
- Interfaced with Oracle Corporation for technical support.
- Patch Management and Version Control.
Unix/ Linux Admin
Confidential
Responsibilities:
- Build Linux servers. Upgrade and patch existing servers. Compile, built and upgrade Linux kernel.
- Setup Solaris Custom Jumpstart server and clients and implement Jumpstart installation.
- Worked with Telnet, FTP, TCP/IP, rlogin, used to inter-operate hosts.
- Contact various systems administration works under CentOS, Redhat Linux environments.
- Performed regular day-to-day system administrative tasks including User Management, Backup, Network Management, and Software Management including Documentation etc.
- Recommend system configurations for clients based on estimated requirements.
- Performed reorganization of disk partitions, file systems, hard disk addition, and memory upgrade.
- Monitored system activities, log maintenance, and disk space management.
- Encapsulated root file systems, and mirrored the file systems were mirrored to ensure systems had redundant boot disks.
- Administer Apache Servers. Published client’s web site in our Apache server.
- Fix all the system problems, based on system email information and users’ complaints.
- Upgrade software, add patches, and add new hardware in UNIX machines.