We provide IT Staff Augmentation Services!

Sr. Hadoop Administrator Resume

5.00/5 (Submit Your Rating)

Pleasanton, CA

SUMMARY

  • Overall 10+Years of IT experience in Administering, Installing, Configuring and Maintaining Linux and Hadoop clusters
  • Around 3yrs of experience working on Hadoop eco - system
  • Experience in configuring, installing and managing Apache, Cloudera, HortonWorks & MapR Hadoop Distributions
  • Extensive Experience in understanding the client’s Big Data business requirements and transform it into Hadoop centric technologies.
  • Experience in configuring Hadoop on Amazon Web Services (AWS).
  • Analyzing the clients existing Hadoop infrastructure and understanding the performance bottlenecks and provide the performance tuning accordingly
  • Experience wif Installing Hadoop in Confidential servers and rebuild existing servers
  • Performed technical proof of concepts
  • Experience in using Automation tools like Chef for installing, configuring and maintaining Hadoop clusters
  • Strong knowledge wif Hadoop cluster connectivity and security.
  • Experience in using Cloudera Manager for installation and management of Hadoop Cluster
  • Experience in using Ambari for installation and management of Horton Works Hadoop Cluster
  • Expertise in writing Shell scripts and Perl scripts and debugging existing scripts
  • Experience in setting up automated 24x7 on monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia
  • Experience in Performance Management of Hadoop Cluster
  • Experience in using Flume to load log files into HDFS
  • Expertise in using Oozie for configuring job flows
  • Managing the configuration of the cluster to meet the needs of data analysis whether me/O bound or CPU bound
  • Experience in using full suite of infrastructure services like DHCP, PXE, DNS, KICKSTART and NFS Mount
  • Developed Hive Queries and automated those queries for analyzing on Hourly, Daily and Weekly basis
  • Strong troubleshooting and performance tuning skills
  • Coordinating Cluster services through Zookeeper
  • Hands on development and Administration experience wif reporting tools like OBIEE, TABLEAU
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Experiencein importing andexporting the preprocessed data into the commercial analytic database, e.g. RDBMS
  • Proven Expertise of Complete Life Cycle or End-to-End (gathering Business Users Requirements, System Analysis, Design & Data Modeling, Development, Testing) implementation experience of various BI tools (OBIEE,Tableau&MSBI Stack).
  • Sound knowledge of database architecture for OLTP and OLAP applications, Data Analysis, ETL processes in developing data marts and Enterprise Data Warehouse.

TECHNICAL SKILLS

Hadoop Framework: HDFS, Hive, Pig, Flume, Oozie, Zookeeper, HBase, Cassandra and Sqoop

NoSQL Databases: Hbase, Cassandra

Programming/Scripting: C, SQL, PIG LATIN, UNIX shell scripting

Microsoft: MS Office, MS Project, MS Visio, MS Visual Studio 2003/ 2005/ 2008

Databases: MySQL, Oracle, Teradata, DB2

Operating Systems: Linux, Cent OS,RHEL,Windows

WEB Servers: Apache Tomcat, JBOSS and Apache Http web server,Weblogic

Cluster Management Tools: HDP Ambari, Cloudera Manager, Hue, SolrCloud.

IDE: Net Beans, Eclipse, Visual Studio, Microsoft SQL Server, MS Office

Reporting Tools: OBIEE10.x,11.x,12c, Tableau9.x, SSRS

ETL Tools: Informatica, Datastage, SSIS

Monitoring Tools: Nagios,Ganglia

Automation Tools: Chef, Puppet and HP Automation

PROFESSIONAL EXPERIENCE

Confidential, Pleasanton, CA

Sr. Hadoop Administrator

Responsibilities:

  • Installed Hadoop CDH 5.2.1on clustered Environments on Lab/Prod Environments
  • Upgraded the Hadoop Cluster from CDH 5.2 to CDH 5.4
  • Responsible for Cluster maintenance, Monitoring, commissioning and decommissioning Data nodes, Troubleshooting, Manage and review data backups, Manage & review log files.
  • Day to day responsibilities includes solving developer issues, deployments moving code from one environment to other environment, providing access to Confidential users and providing instant solutions to reduce the impact and documenting the same and preventing future issues.
  • Adding/installation of Confidential components and removal of them through Cloudera Manager.
  • Collaborating wif application teams to install operating system and Hadoop updates, patches, version upgrades.
  • Installed and Configured Hbase Master and Region services on cluster
  • Configured High Availability for Control services like Namenode and Job tracker.
  • Configured automatic failover for Namenode to avoid failures of flume jobs if Namenode goes down.
  • Enables Kerberos Security for the cluster.
  • Contribute in typical system administration and programming skills such as storage capacity management, performance tuning, Setup, configuration and management of security for hadoop clusters.
  • Data loading to Hadoop and Hive using Sqoop from oracle and db2
  • Implemented Bucketing and Partitioning using Hive to assist the users wif data analysis.
  • Day to day support for OS Integration and application installations
  • Did performance tuning on Hadoop cluster running Teragen and Terasort
  • Configured Hue Service and Impala.
  • Supporting users on their issues using Hadoop cluster
  • Managing Disk File Systems, Server Performance, Users Creation and Granting file access Permissions and RAID configurations.
  • Support pre-production and production support teams in the analysis of critical services and assists wif maintenance operations.
  • Automate administration tasks through use of scripting and Job Scheduling using CRON.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.

Environment: Cloudera Manager, HDFS, YARN/MRV2, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Spark, Hue, Teradata and MySQL

Confidential, CA

Hadoop Consultant

Responsibilities:

  • Administering Hadoop clusters using MAPR distribution including activities like deploy Hadoop cluster, add/remove services, add/remove nodes, keep track of jobs, monitor critical parts of the cluster, and configure name-node high availability and schedule.
  • Maintenance, Production support, incident/change management, service request management, Troubleshooting and Transform data between RDBMS and HDFS.
  • Configured rack awareness using topology scripts in MAPR distribution.
  • Configured FAIR scheduler in MAPR distribution.
  • Performed data completeness, correctness, data transformation and data quality testing using SQL Managing, reviewing Hadoop log files. Performance support to Developers
  • Migrating Application teams Video Streaming data (VOD), pay per clicks and Analytics teams to Hadoop from Mainframes, FTP server and EDW.
  • Data loading to Hadoop and hive using sqoop from MySQL, oracle and db2.
  • Developed scripts to monitor automation jobs and processes required for Hadoop and setup mail service in case of failure.
  • Developed scripts and automated data management from end to end and updating b/w all the clusters.
  • Developing Oozie Workflows, Coordinators for automation of jobs.

Environment: Java (JDK 1.7), Linux, Shell Scripting, Teradata, SQL server, MAPR Hadoop, Flume, Sqoop, Pig, Hive, Zookeeper and Hbase, Business Objects and Tableau.

Confidential, Foster City, CA

Hadoop Consultant

Responsibilities:

  • Installed Hadoop on clustered Environments on Dev/UAT/Prod Environments
  • Installed Cloudera Manager on CDH3 clusters
  • Configure the cluster properties to gain the high cluster performance by taking cluster hardware configuration as key criteria
  • Designed the rack topology script for the production Hadoop cluster
  • Develop benchmarking routines by using Teragen and Terasort
  • Implemented the Hadoop Name-node HA services to make the Hadoop services highly available
  • Installed Ganglia to monitor Hadoop daemons and Implemented the changes in configuration parameters and in parallel monitored the changes in Ganglia
  • By using flume collected web logs from different sources and dumped them into HDFS
  • Implemented Oozie work-flow for ETL Process
  • Developed Hive Scripts and Temporary Functions for Complex Business Analytics
  • Exporting data from RDBMS to HIVE,HDFS and HIVE,HDFS to RDBMS by using SQOOP
  • Implemented shell scripts for log-Rolling day to day processes and made it automated
  • Coordinating FLUME,HBASE nodes and master using zookeeper
  • Commissioned/decommission nodes as needed.
  • Streamlined cluster scaling and configuration
  • Developed the cron job for storing the Name-node metadata onto the NFS mount directory
  • Worked on file system management and monitoring and Capacity planning
  • Execute system and disaster recovery processes
  • Work wif the project and application development teams to implement Confidential business initiatives as they relate to Hadoop.
  • Installed and configured operating systems packages

Environment: Cloudera Manager, HDFS, YARN/MRV2, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Hue, Teradata and MySQL and Oracle

Confidential, Herndon, VA

Hadoop Systems Engineer

Responsibilities:

  • Implemented two physical Clusters in the company and installed Hadoop
  • Automated Installing Hadoop cluster using puppet
  • Implemented Name Node Metadata backup using NFS Mount
  • Manage the day to day operations of the cluster for backup and support
  • Used full suite of infrastructure services like DCHP, PXE, DNS, KICKSTART and NFS
  • Implemented automated 24x7 monitoring and escalation infrastructure for Hadoop cluster using Nagios and Ganglia
  • Installed and Configured Cassandra on 2-nodes for NOSQL
  • Highly involved in operations and troubleshooting Hadoop clusters
  • Monitored cluster job performance and capacity planning
  • Day to day support for OS Integration and application installations
  • Designed and Implemented Security by gathering the user requirements.
  • Worked on system and disaster recovery process

Confidential, Charlotte, NC

OBIEE Administrator/Developer

Responsibilities:

  • Gathered the requirement, completed the proof of concept, Designed, Developed and Tested Physical Layer, Business Layer and Presentation Layer of OBIEE
  • Created Dimensional Hierarchy for Dimensions, Level based Measures and Aggregate navigation in BMM layer.
  • Developed and debugged many Dashboards using different Analytics Views (Pivot Table, Chart, and Column Selector), Dynamic / Interactive Dashboards wif drill-down capabilities, charts, tabular using global and local Filters.
  • Configured Repository and session variables to get dynamic reports and dashboards
  • Added Confidential Dimensions and Columns to the Subject Areas to full fill the Additional Requirements of the Business Analysts.
  • OBIEE reports performance improved by tuning long running sql. Created indexes, created materialized views, created Aggregate Tables, Configuring Nqsconfig.ini parameters etc
  • Experience in purging the cache using Event Polling Tables.
  • Developed different brand level Reports, Filters, and Prompts using Oracle BI Answers.
  • Assisting in creation of pivot table requests allowing manager’s to analyze the business/data from different perspectives.
  • Experience in promoting the OBI repository and webcat to various environments.
  • Developed user guides for the end users. User guide contained some screenshots of high level reports and also provides instructions in accessing them.
  • Conducted user training to familiarize them to the Confidential reporting environment.
  • Involved in Upgrading OBI 10g rpd to OBI 11g rpd and Catalog
  • Implemented Level Based, Parent Child, Skipped and Ragged Hierarchies using OBIEE 11G Administration tool.
  • Having sound knowledge in Time Series Measures (Ago, To date and Periodic Rolling) functions.
  • Implemented the Security in Console and adding application roles in OBI 11G as per the Confidential requirements.
  • Modified Pre-built OBI Applications dashboards, Answers, Repository for Sales Finance and Supply Chain.
  • Responsible in providing business users the out of the box Dashboards and associating reports for getting their views about the report columns.
  • Implemented Time Period based security for particular user groups in OBI 11G Admin tool.
  • Hands on Working Experience wif Various Views (Scale Marker, Display Slider in Graphs)
  • Proficient in working wif Action Links (BI Navigation, Action Link to Analysis)
  • Having good knowledge in creating Pixel Perfect reports using BI-Publisher
  • Experience in implementing KPI’s (Key Performance Indicators) and KPI Watch list to analyze the business.
  • Consolidated individual SSIS packages into a master package due to the deployment limit on the Yardi server.
  • Implemented Logging in SSIS packages.
  • Developed SSIS packages using database as a source and executed the output to the flat files.

Environment: OBIEE 11.1.1.X/10.1.3.4.X, Windows2007 Server, Windows Services, Essbase, Informatica, Sqlserver 2005, Windows XP, UNIX/LINUX, Shell Scripting, Oracle 11g/10g, SQL, PL/SQL, and TOAD 6.3/7.0.

Confidential, San Jose, CA

Production Support Consultant

Responsibilities:

  • Create service requests for the outstanding issues at the client site and co-ordinate wif Oracle Technical support team.
  • Monitoring and verifying the Workflows for extraction, transformation and loading data
  • Monitored runtime execution of data warehouse applications using DAC and address to the Email alerts in case of any failures.
  • Monitor users, DAC repository, and application maintenance jobs.
  • Co-ordinate wif off shore team to discuss the status of the open tickets, ETL jobs and OBIEE Production issues.
  • Primary contact for the L-2 support at the client site.
  • Addressed the Jira/clearquest tickets for the open ETL, OBIEE and data quality issues.
  • Pulled the Teradata loader files from the ETL UNIX servers and worked wif the Oracle Technical support to troubleshoot the slowly running DAC jobs in production.
  • Assisted the users wif the OBI report issues and dashboard privileges.

Environment: Oracle Business Intelligence Enterprise Edition (OBIEE) 10.1.3.4, Informatica Power Center 8.1, DAC 7.9.X, Windows XP, UNIX/LINUX, Shell Scripting, Teradata, Oracle 10g, SQL, PL/SQL, Toad

Confidential, San Diego, CA

Linux Administrator/OBIEE Administrator

Responsibilities:

  • Worked closely wif Data Modelers in designing the data model based on the source system tables.
  • Configured Physical & BMM layer - import tables, Identifying facts, dimensions, created physical tables, join the tables, creating the logical tables and logical columns.
  • Worked on Dimensional Hierarchies (Drill down & Drill up), Level Based & Share Measures and Time series Functions based on business requirements.
  • Worked on Repository Variables and Session Variables.
  • Implemented Security by creating users, roles, groups, data level security, object level security and time period level security.
  • Involved in managing appropriate security privileges on subject areas and dashboards according to business requirements.
  • Created Alias Tables for easy reporting and better performance.
  • Implemented Aggregate tables for better performance.
  • Worked on Answers, Dashboards, and Delivers.
  • Developed report wif filters, presentation variables, prompts, column selector, view selector, charts, narrative, gauge views, pivot table and other views.
  • Created Dashboards prompts, pages, links, images, embedded content, text, folders and guided navigation links.
  • Patched RHEL5 and Solaris 8, 9, 10 servers for EMC Powerpath Upgrade for VMAX migration.
  • Configured LVM (Logical Volume Manager) to manage volume group, logical and physical partitions and importing Confidential physical volumes.
  • Maintained and monitored all servers' operating system and application patch level, disk space and memory usage, user activities on daily basis, administration on Sun Solaris and RHEL systems, management archiving.
  • Installed, configured, troubleshoot and maintain Linux Servers and Apache Web server, configuration and maintenance of security and scheduling backups, submitting various types of cron jobs.

Environment: Linux 5.x/4.x,Oracle 10g, OBIEE 10.1.3.3, Informatica Power Center 8.1, DAC 7.9.X, Windows Services, Windows 2003 Server UNIX/LINUX, Oracle 9i/10g, SQL, PL/SQL, Toad

Confidential

OBIEE Developer/Performance Engineer

Responsibilities:

  • Worked closely wif Data Modelers in designing the data model based on the source system tables.
  • Configured Physical & BMM layer - import tables, Identifying facts, dimensions, created physical tables, join the tables, creating the logical tables and logical columns.
  • Worked on Dimensional Hierarchies (Drill down & Drill up), Level Based & Share Measures and Time series Functions based on business requirements.
  • Worked on Repository Variables and Session Variables.
  • Implemented Security by creating users, roles, groups, data level security, object level security and time period level security.
  • Involved in managing appropriate security privileges on subject areas and dashboards according to business requirements.
  • Created Alias Tables for easy reporting and better performance.
  • Implemented Aggregate tables for better performance.
  • Worked on Answers, Dashboards,and Delivers.
  • Developed report wif filters, presentation variables, prompts, column selector, view selector, charts, narrative,gauge views, pivot table and other views.
  • Created Dashboards prompts, pages, links, images, embedded content, text, folders and guided navigation links.
  • Created pivot tables in many dashboards for providing the ability to rotate rows, columns and section headings to obtain different perspective of the same data.
  • Has integrated different reports in to dashboards and has set access permissions accordingly based on user needs.
  • Has configured the job manager for working wif OBI Delivers, created iBots and scheduled them to display alerts as well as to send the reports to the inbox of users.
  • Created different reports like Standard, Dynamic and Ad-hoc reports.
  • Performed unit testing and validating the repository.
  • Prepared Test scripts, Load Test, Test Data, Test Plan, Test Cases, Execute test, validate results, Manage defects and report results
  • Used to identify the queries which taking too long and optimize those queries to improve performance
  • Performance Tested SOA based application using Parasoft and also LoadRunner Tool
  • Performance Tested application in SaaS model
  • Handled Complex HCF Check in and Checkout Scripts wif Custom DLLs provided by EMC
  • Handled View State and Event Validations for .net Scripts
  • Independently develop LoadRunner test scripts according to test specifications/requirements.
  • Provide support to the development team in identifying real world use cases and appropriate workflows
  • Performs in-depth analysis to isolate points of failure in the application
  • Assist in production of testing and capacity certification reports.
  • Investigate and troubleshoot performance problems in a lab environment. This will also include analysis of performance problems in a production environment.

Environment: OBIEE 10.1.3.3, Informatica Power Center 8.1, DAC 7.9.X, Windows Services, Windows 2003 Server UNIX/LINUX, Oracle 9i/10g, SQL, PL/SQL, Toad, LoadRunner, QTP, web -HTML/HTTP, QTP.

We'd love your feedback!