Big Data Hadoop Developer/administrator Resume
SUMMARY
- A detail - oriented professional having over 5 years of IT Experience.
- Experience in providing Solution Architecture for Big Data projects using Hadoop Eco System. Experienced in setting up of Hadoop cluster (Amazon EC2, CDH3, CDH4, HDP), Performance Tuning, Developing Logical & Physical Data Models using HIVE for Analytics, File processing using PIG, and Data load management using SQOOP.
- Experience creating reports using Tableau.
- Experience with Cloudera Manager.
- Excellent interpersonal, communication skills and a very good team player willing to take on new and varied projects and an ability to handle changing priorities and deadlines.
TECHNICAL SKILLS
Big Data: Hadoop, Hive, Sqoop, Pig, Puppet, Ambari, HBase
Web Servers: Amazon EC2, Cloud era, Ambari Cloud, Microsoft Azure
Query Languages: HiveQL, SQL, PL/SQL, Pig
Databases: Oracle, SQL Server, HBase and MS Access
Programming Languages: Java Script, HTML and ASP
GUI Tools: Visual Basics 5.0 and .Net
BI Reporting Tools: Tableau, Crystal Reporting 4.5 and Power Pivot
Operating System: Red-Hat Linux, Ubuntu 13.02 and Windows
PROFESSIONAL EXPERIENCE
Confidential
Big Data Hadoop Developer/Administrator
Responsibilities:
- Creating Python Scripts and verifying them.
- Creating Hive Tables and Views.
- Creating Pig Scripts.
- Administering the Hadoop Cluster in EC2.
- End to End testing of all the scripts and document the time taken for completing one round of initial load and delta loads.
- Documenting the Solution Architecture.
- End-User Training.
Confidential
Big Data Hadoop Developer/Administrator
Responsibilities:
- Apache Hadoop installation & configuration of multiple nodes on AWS EC2 system
- Setup and optimize Standalone-System/Pseudo-Distributed/Distributed Clusters
- Build/Tune/Maintain Hive QL and Pig Scripts for user reporting
- Developed MapReduce Programs
- Experienced in defining job flows
- Experienced in managing and reviewing Hadoop log files
- Supported MapReduce Programs running on the cluster
- Involved in loading data from UNIX file system to HDFS
- Installed and configured Hive.
- Involved in creating Hive tables, loading data, and writing Hive queries
- Develop Shell scripts to automate routine DBA tasks (i.e. database refresh, backups, monitoring)
- Tuned/Modified SQL for batch and online processes
Environment:Hadoop (HDFS) multi-node installation, MapReduce, AWS, Hive, Java (JDK 1.6), Flat Files
Confidential
Hadoop Developer.
Responsibilities:
- Analyzed the Functional Specifications.
- Installed and configured Hbase, HDFS, PIG, HIVE, Hadoop MapReduce
- Writing Pig scripts to process the data.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Written Hive queries for data analysis to meet the Business requirements.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Got good experience with NOSQL database.
- Involved in loading data from UNIX file system to HDFS.
- Supported MapReduce Programs those are running on the cluster.
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduceway.
- Analyzed functional specifications.
- Responsible to manage data coming from different sources.
- Worked on Contacts Relations Management, Human resources Management, Sales and Profit Logistics.
- Communications Module and Inventory Management System.
Environment: Java, ASP and SQL
