We provide IT Staff Augmentation Services!

Big Data Hadoop Developer/administrator Resume

4.00/5 (Submit Your Rating)

SUMMARY

  • A detail - oriented professional having over 5 years of IT Experience.
  • Experience in providing Solution Architecture for Big Data projects using Hadoop Eco System. Experienced in setting up of Hadoop cluster (Amazon EC2, CDH3, CDH4, HDP), Performance Tuning, Developing Logical & Physical Data Models using HIVE for Analytics, File processing using PIG, and Data load management using SQOOP.
  • Experience creating reports using Tableau.
  • Experience with Cloudera Manager.
  • Excellent interpersonal, communication skills and a very good team player willing to take on new and varied projects and an ability to handle changing priorities and deadlines.

TECHNICAL SKILLS

Big Data: Hadoop, Hive, Sqoop, Pig, Puppet, Ambari, HBase

Web Servers: Amazon EC2, Cloud era, Ambari Cloud, Microsoft Azure

Query Languages: HiveQL, SQL, PL/SQL, Pig

Databases: Oracle, SQL Server, HBase and MS Access

Programming Languages: Java Script, HTML and ASP

GUI Tools: Visual Basics 5.0 and .Net

BI Reporting Tools: Tableau, Crystal Reporting 4.5 and Power Pivot

Operating System: Red-Hat Linux, Ubuntu 13.02 and Windows

PROFESSIONAL EXPERIENCE

Confidential

Big Data Hadoop Developer/Administrator

Responsibilities:

  • Creating Python Scripts and verifying them.
  • Creating Hive Tables and Views.
  • Creating Pig Scripts.
  • Administering the Hadoop Cluster in EC2.
  • End to End testing of all the scripts and document the time taken for completing one round of initial load and delta loads.
  • Documenting the Solution Architecture.
  • End-User Training.

Confidential

Big Data Hadoop Developer/Administrator

Responsibilities:

  • Apache Hadoop installation & configuration of multiple nodes on AWS EC2 system
  • Setup and optimize Standalone-System/Pseudo-Distributed/Distributed Clusters
  • Build/Tune/Maintain Hive QL and Pig Scripts for user reporting
  • Developed MapReduce Programs
  • Experienced in defining job flows
  • Experienced in managing and reviewing Hadoop log files
  • Supported MapReduce Programs running on the cluster
  • Involved in loading data from UNIX file system to HDFS
  • Installed and configured Hive.
  • Involved in creating Hive tables, loading data, and writing Hive queries
  • Develop Shell scripts to automate routine DBA tasks (i.e. database refresh, backups, monitoring)
  • Tuned/Modified SQL for batch and online processes

Environment:Hadoop (HDFS) multi-node installation, MapReduce, AWS, Hive, Java (JDK 1.6), Flat Files

Confidential

Hadoop Developer.

Responsibilities:

  • Analyzed the Functional Specifications.
  • Installed and configured Hbase, HDFS, PIG, HIVE, Hadoop MapReduce
  • Writing Pig scripts to process the data.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Written Hive queries for data analysis to meet the Business requirements.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Got good experience with NOSQL database.
  • Involved in loading data from UNIX file system to HDFS.
  • Supported MapReduce Programs those are running on the cluster.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduceway.
  • Analyzed functional specifications.
  • Responsible to manage data coming from different sources.
  • Worked on Contacts Relations Management, Human resources Management, Sales and Profit Logistics.
  • Communications Module and Inventory Management System.

Environment: Java, ASP and SQL

We'd love your feedback!