We provide IT Staff Augmentation Services!

Sr. Hadoop Administrator Resume

4.00/5 (Submit Your Rating)

Pleasanton, CA

SUMMARY

  • Overall 10+Years of IT experience in Administering, Installing, Configuring and Maintaining Linux and Hadoop clusters
  • Around 3yrs of experience working on Hadoop eco - system
  • Experience in configuring, installing and managing Apache, Cloudera, HortonWorks & MapR Hadoop Distributions
  • Extensive Experience in understanding teh client’s Big Data business requirements and transform it into Hadoop centric technologies.
  • Experience in configuring Hadoop on Amazon Web Services (AWS).
  • Analyzing teh clients existing Hadoop infrastructure and understanding teh performance bottlenecks and provide teh performance tuning accordingly
  • Experience with Installing Hadoop in Confidential servers and rebuild existing servers
  • Performed technical proof of concepts
  • Experience in using Automation tools like Chef for installing, configuring and maintaining Hadoop clusters
  • Strong knowledge with Hadoop cluster connectivity and security.
  • Experience in using Cloudera Manager for installation and management of Hadoop Cluster
  • Experience in using Ambari for installation and management of Horton Works Hadoop Cluster
  • Expertise in writing Shell scripts and Perl scripts and debugging existing scripts
  • Experience in setting up automated 24x7 on monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia
  • Experience in Performance Management of Hadoop Cluster
  • Experience in using Flume to load log files into HDFS
  • Expertise in using Oozie for configuring job flows
  • Managing teh configuration of teh cluster to meet teh needs of data analysis whether I/O bound or CPU bound
  • Experience in using full suite of infrastructure services like DHCP, PXE, DNS, KICKSTART and NFS Mount
  • Developed Hive Queries and automated those queries for analyzing on Hourly, Daily and Weekly basis
  • Strong troubleshooting and performance tuning skills
  • Coordinating Cluster services through Zookeeper
  • Hands on development and Administration experience with reporting tools like OBIEE, TABLEAU
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Experiencein importing andexporting teh preprocessed data into teh commercial analytic database, e.g. RDBMS
  • Proven Expertise of Complete Life Cycle or End-to-End (gathering Business Users Requirements, System Analysis, Design & Data Modeling, Development, Testing) implementation experience of various BI tools (OBIEE,Tableau&MSBI Stack).
  • Sound knowledge of database architecture for OLTP and OLAP applications, Data Analysis, ETL processes in developing data marts and Enterprise Data Warehouse.

TECHNICAL SKILLS

Hadoop Framework: HDFS, Hive, Pig, Flume, Oozie, Zookeeper, HBase, Cassandra and Sqoop

NoSQL Databases: Hbase, Cassandra

Programming/Scripting: C, SQL, PIG LATIN, UNIX shell scripting

Microsoft: MS Office, MS Project, MS Visio, MS Visual Studio 2003/ 2005/ 2008

Databases: MySQL, Oracle, Teradata, DB2

Operating Systems: Linux, Cent OS,RHEL,Windows

WEB Servers: Apache Tomcat, JBOSS and Apache Http web server,Weblogic

Cluster Management Tools: HDP Ambari, Cloudera Manager, Hue, SolrCloud.

IDE: Net Beans, Eclipse, Visual Studio, Microsoft SQL Server, MS Office

Reporting Tools: OBIEE10.x,11.x,12c, Tableau9.x, SSRS

ETL Tools: Informatica, Datastage, SSIS

Monitoring Tools: Nagios,Ganglia

Automation Tools: Chef, Puppet and HP Automation

PROFESSIONAL EXPERIENCE

Confidential, Pleasanton, CA

Sr. Hadoop Administrator

Responsibilities:

  • Installed Hadoop CDH 5.2.1on clustered Environments on Lab/Prod Environments
  • Upgraded teh Hadoop Cluster from CDH 5.2 to CDH 5.4
  • Responsible for Cluster maintenance, Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
  • Day to day responsibilities includes solving developer issues, deployments moving code from one environment to other environment, providing access to Confidential users and providing instant solutions to reduce teh impact and documenting teh same and preventing future issues.
  • Adding/installation of Confidential components and removal of them through Cloudera Manager.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades.
  • Installed and Configured Hbase Master and Region services on cluster
  • Configured High Availability for Control services like Namenode and Job tracker.
  • Configured automatic failover for Namenode to avoid failures of flume jobs if Namenode goes down.
  • Enables Kerberos Security for teh cluster.
  • Contribute in typical system administration and programming skills such as storage capacity management, performance tuning, Setup, configuration and management of security for hadoop clusters.
  • Data loading to Hadoop and Hive using Sqoop from oracle and db2
  • Implemented Bucketing and Partitioning using Hive to assist teh users with data analysis.
  • Day to day support for OS Integration and application installations
  • Did performance tuning on Hadoop cluster running Teragen and Terasort
  • Configured Hue Service and Impala.
  • Supporting users on their issues using Hadoop cluster
  • Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID configurations.
  • Support pre-production and production support teams in teh analysis of critical services and assists with maintenance operations.
  • Automate administration tasks through use of scripting and Job Scheduling using CRON.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.

Environment: Cloudera Manager, HDFS, YARN/MRV2, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Spark, Hue, Teradata and MySQL

Confidential, CA

Hadoop Consultant

Responsibilities:

  • Administering Hadoop clusters using MAPR distribution including activities like deploy Hadoop cluster, add/remove services, add/remove nodes, keep track of jobs, monitor critical parts of teh cluster, and configure name-node high availability and schedule.
  • Maintenance, Production support, incident/change management, service request management, Troubleshooting and Transform data between RDBMS and HDFS.
  • Configured rack awareness using topology scripts in MAPR distribution.
  • Configured FAIR scheduler in MAPR distribution.
  • Performed data completeness, correctness, data transformation and data quality testing using SQL Managing, reviewing Hadoop log files. Performance support to Developers
  • Migrating Application teams Video Streaming data (VOD), pay per clicks and Analytics teams to Hadoop from Mainframes, FTP server and EDW.
  • Data loading to Hadoop and hive using sqoop from MySQL, oracle and db2.
  • Developed scripts to monitor automation jobs and processes required for Hadoop and setup mail service in case of failure.
  • Developed scripts and automated data management from end to end and updating b/w all teh clusters.
  • Developing Oozie Workflows, Coordinators for automation of jobs.

Environment: Java (JDK 1.7), Linux, Shell Scripting, Teradata, SQL server, MAPR Hadoop, Flume, Sqoop, Pig, Hive, Zookeeper and Hbase, Business Objects and Tableau.

Confidential, Foster City, CA

Hadoop Consultant

Responsibilities:

  • Installed Hadoop on clustered Environments on Dev/UAT/Prod Environments
  • Installed Cloudera Manager on CDH3 clusters
  • Configure teh cluster properties to gain teh high cluster performance by taking cluster hardware configuration as key criteria
  • Designed teh rack topology script for teh production Hadoop cluster
  • Develop benchmarking routines by using Teragen and Terasort
  • Implemented teh Hadoop Name-node HA services to make teh Hadoop services highly available
  • Installed Ganglia to monitor Hadoop daemons and Implemented teh changes in configuration parameters and in parallel monitored teh changes in Ganglia
  • By using flume collected web logs from different sources and dumped them into HDFS
  • Implemented Oozie work-flow for ETL Process
  • Developed Hive Scripts and Temporary Functions for Complex Business Analytics
  • Exporting data from RDBMS to HIVE,HDFS and HIVE,HDFS to RDBMS by using SQOOP
  • Implemented shell scripts for log-Rolling day to day processes and made it automated
  • Coordinating FLUME,HBASE nodes and master using zookeeper
  • Commissioned/decommission nodes as needed.
  • Streamlined cluster scaling and configuration
  • Developed teh cron job for storing teh Name-node metadata onto teh NFS mount directory
  • Worked on file system management and monitoring and Capacity planning
  • Execute system and disaster recovery processes
  • Work with teh project and application development teams to implement Confidential business initiatives as they relate to Hadoop.
  • Installed and configured operating systems packages

Environment: Cloudera Manager, HDFS, YARN/MRV2, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Hue, Teradata and MySQL and Oracle

Confidential, Herndon, VA

Hadoop Systems Engineer

Responsibilities:

  • Implemented two physical Clusters in teh company and installed Hadoop
  • Automated Installing Hadoop cluster using puppet
  • Implemented Name Node Metadata backup using NFS Mount
  • Manage teh day to day operations of teh cluster for backup and support
  • Used full suite of infrastructure services like DCHP, PXE, DNS, KICKSTART and NFS
  • Implemented automated 24x7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia
  • Installed and Configured Cassandra on 2-nodes for NOSQL
  • Highly involved in operations and troubleshooting Hadoop clusters
  • Monitored cluster job performance and capacity planning
  • Day to day support for OS Integration and application installations
  • Designed and Implemented Security by gathering teh user requirements.
  • Worked on system and disaster recovery process

Confidential, Charlotte, NC

OBIEE Administrator/Developer

Responsibilities:

  • Gatheird teh requirement, completed teh proof of concept, Designed, Developed and Tested Physical Layer, Business Layer and Presentation Layer of OBIEE
  • Created Dimensional Hierarchy for Dimensions, Level based Measures and Aggregate navigation in BMM layer.
  • Developed and debugged many Dashboards using different Analytics Views (Pivot Table, Chart, and Column Selector), Dynamic / Interactive Dashboards with drill-down capabilities, charts, tabular using global and local Filters.
  • Configured Repository and session variables to get dynamic reports and dashboards
  • Added Confidential Dimensions and Columns to teh Subject Areas to full fill teh Additional Requirements of teh Business Analysts.
  • OBIEE reports performance improved by tuning long running sql. Created indexes, created materialized views, created Aggregate Tables, Configuring Nqsconfig.ini parameters etc
  • Experience in purging teh cache using Event Polling Tables.
  • Developed different brand level Reports, Filters, and Prompts using Oracle BI Answers.
  • Assisting in creation of pivot table requests allowing manager’s to analyze teh business/data from different perspectives.
  • Experience in promoting teh OBI repository and webcat to various environments.
  • Developed user guides for teh end users. User guide contained some screenshots of high level reports and also provides instructions in accessing them.
  • Conducted user training to familiarize them to teh Confidential reporting environment.
  • Involved in Upgrading OBI 10g rpd to OBI 11g rpd and Catalog
  • Implemented Level Based, Parent Child, Skipped and Ragged Hierarchies using OBIEE 11G Administration tool.
  • Having sound knowledge in Time Series Measures (Ago, To date and Periodic Rolling) functions.
  • Implemented teh Security in Console and adding application roles in OBI 11G as per teh Confidential requirements.
  • Modified Pre-built OBI Applications dashboards, Answers, Repository for Sales Finance and Supply Chain.
  • Responsible in providing business users teh out of teh box Dashboards and associating reports for getting their views about teh report columns.
  • Implemented Time Period based security for particular user groups in OBI 11G Admin tool.
  • Hands on Working Experience with Various Views (Scale Marker, Display Slider in Graphs)
  • Proficient in working with Action Links (BI Navigation, Action Link to Analysis)
  • Having good knowledge in creating Pixel Perfect reports using BI-Publisher
  • Experience in implementing KPI’s (Key Performance Indicators) and KPI Watch list to analyze teh business.
  • Consolidated individual SSIS packages into a master package due to teh deployment limit on teh Yardi server.
  • Implemented Logging in SSIS packages.
  • Developed SSIS packages using database as a source and executed teh output to teh flat files.

Environment: OBIEE 11.1.1.X/10.1.3.4.X, Windows2007 Server, Windows Services, Essbase, Informatica, Sqlserver 2005, Windows XP, UNIX/LINUX, Shell Scripting, Oracle 11g/10g, SQL, PL/SQL, and TOAD 6.3/7.0.

Confidential, San Jose, CA

Production Support Consultant

Responsibilities:

  • Create service requests for teh outstanding issues at teh client site and co-ordinate with Oracle Technical support team.
  • Monitoring and verifying teh Workflows for extraction, transformation and loading data
  • Monitored runtime execution of data warehouse applications using DAC and address to teh Email alerts in case of any failures.
  • Monitor users, DAC repository, and application maintenance jobs.
  • Co-ordinate with off shore team to discuss teh status of teh open tickets, ETL jobs and OBIEE Production issues.
  • Primary contact for teh L-2 support at teh client site.
  • Addressed teh Jira/clearquest tickets for teh open ETL, OBIEE and data quality issues.
  • Pulled teh Teradata loader files from teh ETL UNIX servers and worked with teh Oracle Technical support to troubleshoot teh slowly running DAC jobs in production.
  • Assisted teh users with teh OBI report issues and dashboard privileges.

Environment: Oracle Business Intelligence Enterprise Edition (OBIEE) 10.1.3.4, Informatica Power Center 8.1, DAC 7.9.X, Windows XP, UNIX/LINUX, Shell Scripting, Teradata, Oracle 10g, SQL, PL/SQL, Toad

Confidential, San Diego, CA

Linux Administrator/OBIEE Administrator

Responsibilities:

  • Worked closely with Data Modelers in designing teh data model based on teh source system tables.
  • Configured Physical & BMM layer - import tables, Identifying facts, dimensions, created physical tables, join teh tables, creating teh logical tables and logical columns.
  • Worked on Dimensional Hierarchies (Drill down & Drill up), Level Based & Share Measures and Time series Functions based on business requirements.
  • Worked on Repository Variables and Session Variables.
  • Implemented Security by creating users, roles, groups, data level security, object level security and time period level security.
  • Involved in managing appropriate security privileges on subject areas and dashboards according to business requirements.
  • Created Alias Tables for easy reporting and better performance.
  • Implemented Aggregate tables for better performance.
  • Worked on Answers, Dashboards, and Delivers.
  • Developed report with filters, presentation variables, prompts, column selector, view selector, charts, narrative, gauge views, pivot table and other views.
  • Created Dashboards prompts, pages, links, images, embedded content, text, folders and guided navigation links.
  • Patched RHEL5 and Solaris 8, 9, 10 servers for EMC Powerpath Upgrade for VMAX migration.
  • Configured LVM (Logical Volume Manager) to manage volume group, logical and physical partitions and importing Confidential physical volumes.
  • Maintained and monitored all servers' operating system and application patch level, disk space and memory usage, user activities on daily basis, administration on Sun Solaris and RHEL systems, management archiving.
  • Installed, configured, troubleshoot and maintain Linux Servers and Apache Web server, configuration and maintenance of security and scheduling backups, submitting various types of cron jobs.

Environment: Linux 5.x/4.x,Oracle 10g, OBIEE 10.1.3.3, Informatica Power Center 8.1, DAC 7.9.X, Windows Services, Windows 2003 Server UNIX/LINUX, Oracle 9i/10g, SQL, PL/SQL, Toad

Confidential

OBIEE Developer/Performance Engineer

Responsibilities:

  • Worked closely with Data Modelers in designing teh data model based on teh source system tables.
  • Configured Physical & BMM layer - import tables, Identifying facts, dimensions, created physical tables, join teh tables, creating teh logical tables and logical columns.
  • Worked on Dimensional Hierarchies (Drill down & Drill up), Level Based & Share Measures and Time series Functions based on business requirements.
  • Worked on Repository Variables and Session Variables.
  • Implemented Security by creating users, roles, groups, data level security, object level security and time period level security.
  • Involved in managing appropriate security privileges on subject areas and dashboards according to business requirements.
  • Created Alias Tables for easy reporting and better performance.
  • Implemented Aggregate tables for better performance.
  • Worked on Answers, Dashboards,and Delivers.
  • Developed report with filters, presentation variables, prompts, column selector, view selector, charts, narrative,gauge views, pivot table and other views.
  • Created Dashboards prompts, pages, links, images, embedded content, text, folders and guided navigation links.
  • Created pivot tables in many dashboards for providing teh ability to rotate rows, columns and section headings to obtain different perspective of teh same data.
  • TEMPHas integrated different reports in to dashboards and TEMPhas set access permissions accordingly based on user needs.
  • TEMPHas configured teh job manager for working with OBI Delivers, created iBots and scheduled them to display alerts as well as to send teh reports to teh inbox of users.
  • Created different reports like Standard, Dynamic and Ad-hoc reports.
  • Performed unit testing and validating teh repository.
  • Prepared Test scripts, Load Test, Test Data, Test Plan, Test Cases, Execute test, validate results, Manage defects and report results
  • Used to identify teh queries which taking too long and optimize those queries to improve performance
  • Performance Tested SOA based application using Parasoft and also LoadRunner Tool
  • Performance Tested application in SaaS model
  • Handled Complex HCF Check in and Checkout Scripts with Custom DLLs provided by EMC
  • Handled View State and Event Validations for .net Scripts
  • Independently develop LoadRunner test scripts according to test specifications/requirements.
  • Provide support to teh development team in identifying real world use cases and appropriate workflows
  • Performs in-depth analysis to isolate points of failure in teh application
  • Assist in production of testing and capacity certification reports.
  • Investigate and troubleshoot performance problems in a lab environment. This will also include analysis of performance problems in a production environment.

Environment: OBIEE 10.1.3.3, Informatica Power Center 8.1, DAC 7.9.X, Windows Services, Windows 2003 Server UNIX/LINUX, Oracle 9i/10g, SQL, PL/SQL, Toad, LoadRunner, QTP, web -HTML/HTTP, QTP.

We'd love your feedback!