Hadoop Consultant Resume
5.00/5 (Submit Your Rating)
SUMMARY:
- Around 7 years of comprehensive experience in Hadoop ecosystem tools and Big Data technologies.
- Currently working on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix.
- Worked on Multi Clustered environment and setting up Cloudera Hadoop eco - System.
- Implemented in setting up standards and processes for Hadoop based application design and implementation.
- Implemented Kerberos and Sentry for managing security in Hadoop Ecosystem.
- Implemented High Availability of different applications in Hadoop cluster.
- Practical knowledge on Hadoop ecosystem components like Pig, Hive, Sqoop, Flume, Oozie, Zookeeper, Cloudera
- Manager & Ambari.
- Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
- Support and utilize multiple Oracle based applications including: SQL and PL/SQL, TOAD, Oracle views, stored procedures, triggers, and the Microsoft Office suite of tools.
- Design, create/modify, and implement documented solutions as agreed to by all business partners responsible to design and integrate a computing system, from start to finish.
- Experience in installation and configuration of HBASE, using HBasemaster and HBase regional server.
- Worked with system engineering teams to plan and deploy hardware and software environments optimization for
- Hadoop implementations.
- Extensive experience in Installation, Configuration, and Migration, Trouble-Shooting and Maintenance of Splunk,
- Apache Web Server on different UNIX flavors like Linux..
- Working experience in designing and implementing complete solutions using Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie
- Involved in maintaining hadoop cluster in development and test environment
PROFESSIONAL EXPERIENCE:
Hadoop Consultant
Confidential
Responsibilities:
- Strong experience in Hadoop Administration & Big Data Technologies along with Linux & Database administration.
- Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera.
- Expertise in Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Design Big Data solutions for traditional enterprise businesses. Hands on experience in HBase, Hbase master, Zookeeper etc.
- Backup configuration and Recovery from a Name Node failure.
- Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off-line and on-line Backups.
- Involved in bench marking Hadoop/Hbase cluster file systems various batch jobs and workloads
- Expertise in Hadoop cluster ready for development team working on POCs.
- Experience in minor and major upgrades of Hadoop and Hadoop eco system.
- Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
- Involved in Administration of Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
- Created system security supporting multi-tier software delivery system by utilizing Active Directory and Kerberos using Ranger.
- Installed and configured Hadoop eco system like hive, sqoop, pig, and hive.
- Familiar with writing Oozie workflows and Job Controllers for job automation - shell, hive, scoop automation.
Hadoop Administrator
Confidential
Responsibilities:
- Managed and reviewed Hadoop log files.
- Tested raw data and executed performance scripts.
- Shared responsibility for administration of Hadoop, Hive and Pig. Created pig scripts and map reduce programs to filter the log files and aggregate the data.
- Worked on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix..
- Loaded data from Linux file system to HDFS using Pentaho Data Integration
- Created data flows in Pentaho Data Integration for aggregating the data and loading the data to Hive tables
- Responsible for building a cluster on HDP 2.0. With Hadoop 2.2.0 using Ambari.
- Expertise in analyzing data with Hive, Pig and HBase.
- Expertise in Cluster Capacity Planning.
- Implemented Kerberos Security Authentication protocol for Hadoop clusters with Knox and Ranger.
- Expertise with Running Pig and Hive Queries.
- Implemented Fair scheduler on the job tracker to allocate the fair amount of resources to small jobs.
- Good experience in troubleshooting production level issues in the cluster and its functionality.
- Expertise in Manage and review data backups.
- Backed up data on regular basis to a remote cluster using distcp.
- Imported logs from web servers with Flume to ingest the data into HDFS.
- Responsible for creating new users to the hadoop cluster and providing access to the datasets
- Expertise in disaster recovery processes as required.
- Implemented Fair scheduler and capacity scheduler to allocate fair amount of resources to small jobs.
- Dumped the data from MySQL database to HDFS and vice-versa using Sqoop.
Database Administrator
Confidential
Responsibilities:
- Administration, package installation, configuration of Oracle Enterprise Linux 5.x.
- Administration of RHEL, which includes installation, testing, tuning, upgrading and loading patches, troubleshooting both physical and virtual server issues.
- Creating, cloning Linux Virtual Machines.
- Installing RedHat Linux using kick start and applying security polices for hardening the server based on the company policies.
- RPM and YUM package installations, patch and other server management.
- Managing systems routine backup, scheduling jobs like disabling and enabling cron jobs, enabling system logging, network logging of servers for maintenance, performance tuning, testing.
- Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster. Assisted with performance tuning and monitoring.
- Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
- Supported code/design analysis, strategy development and project planning.
- Created reports for the BI team using Sqoop to export data into HDFS and Hive.
- Assisted with data capacity planning and node forecasting.
- Set up user and group login ID, printing parameters, network configuration, password, resolving permissions
- Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
- Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.