Bigdata Intern Resume
PROFESSIONAL SUMMARY:
Software Engineer with good understanding and implementation skills of big data development and operations implementation end to end . Always eager to try, learn and implement new things. I believe in constantly upgrading my skills from the work performed by and around me. I am a people person and work best in a team environment and understand happy and organized team produce a successful product.
TECHNICAL EXPERIENCE:
Data Analytics using EMR: Shipping the server logs to aws S3 and using mapreduce program to parse it and hive to analyze the data with EMR.
Securing Hadoop Cluster: To avoid spoofing Integrating hadoop with MIT Kerberos and simulate production environment.
TECHNICAL SKILLS:
- Expertise in HDFS,Yarn, Mapreduce, Hive, Spark, Sqoop and competent Knowledge in Oozie, Zookeeper, Kafka, HBASE, Kerberos, Ganglia, Nagios, Mysql, Oracle.
- Experience working on Hortonworks and Cloudera distributions.
- Hands on experience on Amazon Ec2, EMR, Open stack Cloud environments:
- Scripting Languages: Python, Sql, Hql, Core Java, shell scripting .
- Experienced and Competent in Linux and windows operating system.
PROFESSIONAL EXPERIENCE:
Bigdata Intern
Confidential
- Implemented Poc for benchmarking Hadoop infrastructure in Bare - metal Vs AWS EC2.
- Installed 5 node Hadoop cluster with Ambari 2.2 and HDP 2.4 in Bare-metal and AWS.
- Tuned Hadoop cluster with Best practices at Linux and Hadoop level.
- Benchmarking of cluster for Hdfs, Map-Reduce, Hive, Spark and kafka.
- Troubleshooting failed benchmarking jobs.
- Collected Metrics from both clusters for benchmark comparison.