We provide IT Staff Augmentation Services!

Sr. Hadoop Admin/developer Resume

3.00/5 (Submit Your Rating)

Tomball, TX

PROFESSIONAL SUMMARY:

  • 7+years of professional IT experience in Hadoop Development and Administration using Apache, Cloudera (CDH), Hortonworks (HDP), and extensive years of experience in Linux Administration and system Administration.
  • Excellent understanding and knowledge of Hadoop.
  • Experienced in installation, configuration, supporting and monitoring 100+ node Hadoop cluster using Cloudera manager and Hortonworks distributions.
  • Have experience in installing, configuring, performance tuning and administrating Hadoop cluster for major Hadoop distributions like CDH 4, CDH 5.
  • Experienced in Importing and exporting data using SQOOP from RDBMS to Hadoop and troubleshooting issues related to SQOOP jobs.
  • Experience in analyzing data with Hive and Pig using on read data schema.
  • Setting up Kerberos authentication for Hadoop.
  • Defining job flows in Hadoop environment using tools like Oozie for data scrubbing and processing.
  • Data migration from existing data stores to Hadoop.
  • Upgrading Hadoop cluster using rolling upgrades .
  • Hadoop cluster capacity planning, Optimization of Cluster to meet the SLA.
  • Implementation of data Schemas for Hadoop both in the Persisted model and Read - only model.
  • Working experience on designing and implementing complete end to end Hadoop Infrastructure.
  • Cluster coordination services through Zoo Keeper.
  • Job and user management using Capacity Scheduler , Fair Scheduler, and Hod Scheduler.
  • Experienced in Commissioning and Decommissioning the Data nodes in Hadoop Cluster.
  • Manage & review of Hadoop log files.
  • Extensive experience on Linux and System administration, 24X7 on call support.
  • Experience in performing data validation using HIVE dynamic partitioning and bucketing.
  • Experience on YARN environment with Storm and Spark .
  • Experience in supporting data analysis projects through Elastic Map Reduce on the Amazon Web Services ( AWS ) and RackSpace cloud. Performed Export and import of data into S3.
  • Provisioning, building and support of Linux servers both Physical and Virtual using for Production, QA and Developers environment.
  • Strengths include good team player, excellent communication interpersonal and analytical skills and ability to work effectively in a fast-paced, high volume, deadline-driven environment.
  • Upgrading from tableau from version 8.3 to 9.0 version.
  • Committed to excellence, self-motivator, team-player, and a far-sighted developer with strong problem-solving skills and with zeal to learn new technologies.

TECHNICAL SKILLS:

Technologies/Tools: Hadoop, HDFS, YARN, Cloudera, Cloudera Manager, HBase, Hive, Pig, Oozie, Sqoop, Flume, Storm, Zoo Keeper, AWS, RackSpace, HortonWorks, CDH 4, CDH 5, Shell Scripting, Tableau.

Databases: HiveQL, SQL Server, Oracle 10g, SQL Profiler

Servers: Tomcat

Operating Systems: Linux RedHat, CentOs, Windows Server 2003/2008, Win 7/8

WORK EXPERIENCE:

Confidential, TOMBALL, TX

SR. HADOOP ADMIN/DEVELOPER

RESPONSIBILITIES :

  • Installing, configuring, and administrating Hadoop cluster using CDH 5.2.3
  • Responsible for implementation and ongoing administration of Hadoop infrastructure. Deploy new hardware and software environments required for Hadoop and to expand memory and disks on nodes in existing environments.
  • Handle the data exchange between HDFS and different Web Applications and databases using Flume and Sqoop .
  • Monitor the data streaming between web sources and HDFS.
  • Generated reports using Tableau report designer.
  • Configuring YARN and optimizing Memory related settings.
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
  • Upgrading Hadoop cluster using rolling upgrades.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
  • Screen Hadoop cluster job performances and capacity planning .
  • Configuration change management where changing the configuration properties of the cluster based on volume of the data being processed.
  • Monitor Hadoop cluster connectivity and security.
  • Setting up Kerberos authentication for Hadoop.
  • Importing and exporting data using SQOOP from RDBMS to Hadoop .
  • Working with data delivery teams to setup new Hadoop users which also includes setting up.
  • Data migration from existing data stores to Hadoop.
  • Linux users and testing HDFS, Hive, Pig and Map Reduce access for the new users.
  • Performing Linux systems administration on production and development servers ( RedHat Linux, CentOS and other UNIX utilities ).
  • Commissioning and Decommissioning the Data nodes in Hadoop Cluster.
  • Configured a 20-30 node (Amazon EC2 spot Instance) Hadoop cluster to transfer the data from Amazon S3 to HDFS and HDFS to Amazon S3 and also to direct input and output to the framework.
  • Framework by using Amazon S3 for Input and Output.
  • Job and user management using Capacity Scheduler.
  • Installing Patches and packages on Unix/Linux Servers.
  • Install and Configure vSphere client, Virtual Server creation and resource allocation.
  • Performance Tuning, Client/Server Connectivity and Database Consistency Checks using different Utilities.
  • Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.

ENVIRONMENT: : Hadoop, Cloudera, Redhat Linux/Centos 4, 5, 6, Logical Volume Manager, HDFS, Hive, Pig, Sqoop, Flume, ESX 5.1/5.5, Apache and Tomcat Web Server, Tableau, Oracle 11,12, Oracle Rac 12c, HPSM, HPSA, Kerboros security.

Confidential, DURHAM, NC

HADOOP ADMIN/DEVELOPER

RESPONSIBILITIES:

  • Working on multiple projects spanning from Architecting Hadoop Clusters to pure Java development.
  • Installation, Configuration and Management of Hadoop Cluster using Puppet.
  • Performed an upgrade in development environment from CDH 4.2 to CDH 4.6.
  • Involved in Cluster coordination services through Zookeeper .
  • Remote Management through Nagios and Ganglia.
  • Managed the configuration of the cluster to meet the needs of analysis - I/O bound or CPU bound .
  • Experienced in managing and reviewing Hadoop Log files.
  • Performing bench mark test on Hadoop clusters and tweak the solution based on test results.
  • Commissioning and Decommissioning the Data nodes in Hadoop Cluster.
  • Configuring YARN and optimizing Memory related settings.
  • Importing and exporting data using SQOOP from RDBMS to Hadoop.
  • Performing data validation using HIVE dynamic partitioning.
  • Developed Map Reduce programs to perform data analysis.
  • Supported data analyst in running Pig and Hive queries.
  • Job and user management using Fair Scheduler.
  • Performed Data scrubbing and processing with Oozie .
  • Screen Hadoop cluster job performances and capacity planning.
  • Preparing documentation on the cluster configuration for future reference.
  • Used Flume for Loading log data into HDFS from multiple sources.
  • Performance Tuning, Client/Server Connectivity and Database Consistency Checks using different Utilities.
  • Experience in supporting data analysis projects through Elastic Map Reduce on the Amazon Web Services (AWS) and Rack space cloud. Performed Export and import of data into S3.
  • Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes.

ENVIRONMENT: Cloudera Hadoop, Linux, HDFS, Hive, Pig, Sqoop, Flume, Zookeeper, HBase, YARN, RDBMS, Oozie, AWS.

Confidential -PHILADELPHIA, PA

HADOOP ADMIN/DEVLOPER

RESPONSIBILITIES:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts
  • Involved in installing Hadoop Ecosystem components
  • Responsible to manage data coming from different sources and Involved in HDFS maintenance and loading of structured and unstructured data
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop.
  • Involved in loading data from LINUX file system to HDFS.
  • Experience in managing and reviewing Hadoop log files.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Implemented test scripts to support test driven development and continuous integration.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Worked on tuning the performance Pig queries.
  • Mentored analyst and test team for writing Hive Queries.
  • Installed Oozie workflow engine to run multiple Map Reduce jobs.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.

ENVIRONMENT: : Hadoop, HDFS, Map Reduce, Hive, Pig, Sqoop, Linux, Java, Oozie, HBase

Confidential

SYSTEM ADMINISTRATOR (LINUX)

RESPONSIBILITIES:

  • Installation, configuration and Administration of Red hat Linux Servers, HP-UX and Solaris Servers.
  • Worked on Server migration/Data Center move projects in various stages for all Linux Servers.
  • Provisioning, building and support of Linux servers both Physical and Virtual using for Production, QA and Developers environment
  • Acquired comprehensive requirements from Project Managers, Team Leads pertaining to the servers to be migrated.
  • Setup and configure failover load balancing using IP bonding for network cards.
  • Configured volume groups and logical volumes, extended logical volumes for file system growth needs using Logical Volume Manager (LVM) commands.
  • Updating YUM Repository and Red hat Package Manager (RPM)
  • Applied appropriate support packages/patches to maintain system integrity.
  • Setting up cron schedules for backups and monitoring processes.
  • Server on various business critical database & application production servers.
  • Monitored Linux server for CPU Utilization, Memory Utilization, and Disk Utilization for performance monitoring.
  • Data migration from existing data stores to Hadoop.
  • Setup NFS file systems and shared them to clients.
  • Troubleshooting and configuring NAS and NFS mount points.
  • Performed capacity analysis, monitored and controlled disk space usage on systems.
  • Monitored system activities and fine-tuned system parameters and configurations to optimize performance and ensure security of systems.
  • Implemented the monitoring tools like EHealth.
  • Administration of RHEL 3, 4, 5 which includes installation, configuration, testing on both physical and virtual machines.
  • Installation and configuration on System imager server on RHEL 5

ENVIRONMENT: Red Hat Enterprise Server (RHEL 3, 4 and RHEL 5), HPUX-11i, 11.x, 10.x, Solaris 8, 9 and 10, EMC Power path, Oracle RAC, Perl scripting, ESX Server, Emulex, DELL Power Edge 2850, 2950, 6850, 6950, HP DL 585, 380, 385, SE 1210, HP Blade servers, EMC and Clarion Storage

Confidential

SYSTEM ENGINEER

RESPONSIBILITIES:

  • Monitor and analyze servers and resolve any problems, maintain systems reporting, tuning.
  • Created users, manage user permissions, maintain User & File system quota on Linux servers.
  • Configured volume groups and logical volumes, extended logical volumes for file system growth needs using Logical Volume Manager (LVM) commands.
  • Maintaining integrity of security with the use of group policies across domains.
  • Supporting users through email, on call and troubleshooting.
  • Maintaining inventory of all components including systems and other hardware.
  • Performed User Account management, data backups, and users' logon support.
  • Maintaining user's data backup by creating particular user folder in File Server and applying security permission on folders.
  • Monitored trouble ticket queue to attend user and system calls.
  • Attended team meetings, change control meetings to update installation progress and for upcoming changes in environment.
  • Configured networking for Windows - Like assigning IP address, DNS discovery, enabling AD for authentication.
  • Day to day server related activities, regular maintenance like reading the log files.
  • Event viewer messages and also trouble shooting and documentation.
  • Internetworking of these PC's to each other and to the corporate network.
  • Interfacing of printers and other peripheral devices to various systems.
  • Installation and testing of hardware, basic router configuration, setup office network and troubleshooting LAN.

We'd love your feedback!