We provide IT Staff Augmentation Services!

Hadoop Consultant Resume

5.00/5 (Submit Your Rating)

SUMMARY:

  • Around 7 years of comprehensive experience in Hadoop ecosystem tools and Big Data technologies.
  • Currently working on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix.
  • Worked on Multi Clustered environment and setting up Cloudera Hadoop eco - System.
  • Implemented in setting up standards and processes for Hadoop based application design and implementation.
  • Implemented Kerberos and Sentry for managing security in Hadoop Ecosystem.
  • Implemented High Availability of different applications in Hadoop cluster.
  • Practical knowledge on Hadoop ecosystem components like Pig, Hive, Sqoop, Flume, Oozie, Zookeeper, Cloudera
  • Manager & Ambari.
  • Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
  • Support and utilize multiple Oracle based applications including: SQL and PL/SQL, TOAD, Oracle views, stored procedures, triggers, and the Microsoft Office suite of tools.
  • Design, create/modify, and implement documented solutions as agreed to by all business partners responsible to design and integrate a computing system, from start to finish.
  • Experience in installation and configuration of HBASE, using HBasemaster and HBase regional server.
  • Worked with system engineering teams to plan and deploy hardware and software environments optimization for
  • Hadoop implementations.
  • Extensive experience in Installation, Configuration, and Migration, Trouble-Shooting and Maintenance of Splunk,
  • Apache Web Server on different UNIX flavors like Linux..
  • Working experience in designing and implementing complete solutions using Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie
  • Involved in maintaining hadoop cluster in development and test environment

PROFESSIONAL EXPERIENCE:

Hadoop Consultant

Confidential

Responsibilities:

  • Strong experience in Hadoop Administration & Big Data Technologies along with Linux & Database administration.
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera.
  • Expertise in Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Design Big Data solutions for traditional enterprise businesses. Hands on experience in HBase, Hbase master, Zookeeper etc.
  • Backup configuration and Recovery from a Name Node failure.
  • Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off-line and on-line Backups.
  • Involved in bench marking Hadoop/Hbase cluster file systems various batch jobs and workloads
  • Expertise in Hadoop cluster ready for development team working on POCs.
  • Experience in minor and major upgrades of Hadoop and Hadoop eco system.
  • Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • Involved in Administration of Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
  • Created system security supporting multi-tier software delivery system by utilizing Active Directory and Kerberos using Ranger.
  • Installed and configured Hadoop eco system like hive, sqoop, pig, and hive.
  • Familiar with writing Oozie workflows and Job Controllers for job automation - shell, hive, scoop automation.

Hadoop Administrator

Confidential

Responsibilities:

  • Managed and reviewed Hadoop log files.
  • Tested raw data and executed performance scripts.
  • Shared responsibility for administration of Hadoop, Hive and Pig. Created pig scripts and map reduce programs to filter the log files and aggregate the data.
  • Worked on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix..
  • Loaded data from Linux file system to HDFS using Pentaho Data Integration
  • Created data flows in Pentaho Data Integration for aggregating the data and loading the data to Hive tables
  • Responsible for building a cluster on HDP 2.0. With Hadoop 2.2.0 using Ambari.
  • Expertise in analyzing data with Hive, Pig and HBase.
  • Expertise in Cluster Capacity Planning.
  • Implemented Kerberos Security Authentication protocol for Hadoop clusters with Knox and Ranger.
  • Expertise with Running Pig and Hive Queries.
  • Implemented Fair scheduler on the job tracker to allocate the fair amount of resources to small jobs.
  • Good experience in troubleshooting production level issues in the cluster and its functionality.
  • Expertise in Manage and review data backups.
  • Backed up data on regular basis to a remote cluster using distcp.
  • Imported logs from web servers with Flume to ingest the data into HDFS.
  • Responsible for creating new users to the hadoop cluster and providing access to the datasets
  • Expertise in disaster recovery processes as required.
  • Implemented Fair scheduler and capacity scheduler to allocate fair amount of resources to small jobs.
  • Dumped the data from MySQL database to HDFS and vice-versa using Sqoop.

Database Administrator

Confidential

Responsibilities:

  • Administration, package installation, configuration of Oracle Enterprise Linux 5.x.
  • Administration of RHEL, which includes installation, testing, tuning, upgrading and loading patches, troubleshooting both physical and virtual server issues.
  • Creating, cloning Linux Virtual Machines.
  • Installing RedHat Linux using kick start and applying security polices for hardening the server based on the company policies.
  • RPM and YUM package installations, patch and other server management.
  • Managing systems routine backup, scheduling jobs like disabling and enabling cron jobs, enabling system logging, network logging of servers for maintenance, performance tuning, testing.
  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster. Assisted with performance tuning and monitoring.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Supported code/design analysis, strategy development and project planning.
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive.
  • Assisted with data capacity planning and node forecasting.
  • Set up user and group login ID, printing parameters, network configuration, password, resolving permissions
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
  • Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.

We'd love your feedback!