We provide IT Staff Augmentation Services!

Big Data - Hadoop Developer Resume

0/5 (Submit Your Rating)

Dallas, TX

SUMMARY

  • 7+ years of IT experience in software development, big data management, data modeling, data integration, implementation and big data frameworks.
  • 3+ years of experience in design, development and deployment of Big Data Analytics using Hadoop Ecosystem including HDFS, Hive, Pig, Hbase, Sqoop, Flume, MapReduce, Spark and Oozie.
  • Strong working experience with ingestion, storage, processing and analysis of big data.
  • Proficient knowledge and hands on experience with writing shell scripts in Linux.
  • Implemented several optimization mechanisms like Combiners, Distributed Cache, Data Compression, and Custom Partitioner to speed up the jobs.
  • Good Experience in writing Map Reduce programs using Java.
  • Hands on performance improvement techniques for data processing in Hive, Pig, Impala & map - reduce using methods dynamic partitioning, bucketing, file compression.
  • Worked with Sqoop in importing and exporting data from different databases like MySql, Oracle into HDFS and Hive.
  • Experience with data formats like JSON, PARQUET, AVRO, RC and ORC formats and compressions like snappy & bzip2.
  • Developed Oozie workflows by integrating all tasks relating to a project and schedule the jobs as per requirements.
  • Worked on analyzing data in NOSQL databases like Hbase
  • Experience in database design, entity relationships, database analysis.
  • Worked on various performance tuning options with Spark SQL including compression codec, compressing the data and storing it into memory, maintaining batch size and converting SQL queries into java bytecode for larger queries.
  • Experience in processing transformations upon Hive Tables using DataFrames using Scala.
  • Developed Spark Streaming by consuming static and streaming data from sources like MySQL, OLTP data stores.
  • Ability to spin up different AWS instances including EC2-classic and EC2-VPC using cloud formation templates.
  • Hands on experience with Amazon Redshift integrating with Spark.
  • Used Azure HDInsight as a sandbox environment to benchmark production cluster for comparing hive queries and spark transformations.

TECHNICAL SKILLS

Big Data Technologies: HDFS, MapReduce, Hive, Pig, Sqoop, Oozie, Flume, Impala, Zookeeper, Kafka, Cloudera CDH5.5

Spark Components: Apache Spark, Data Frames, Spark-SQL

Programming Languages: C, Java, SQL, PLSQL, Scala, Pig Latin, Hive-QL and Unix shell scripting

Databases: Oracle, MySQL, SQL Server, Familiar with NoSQL (HBase)

Web Technologies: JSP, JavaBeans, JDBC, JSF, HTML, CSS, XML, Java Script, Ajax

Operating Systems: Windows, UNIX, Linux distributions (Centos, Ubuntu)

Other Tools: Eclipse, Tableau 9.0, WinSCP 5.7.7, AutoSys, Log4j, Maven

PROFESSIONAL EXPERIENCE

Confidential - Dallas, TX

Big Data - Hadoop Developer

Responsibilities:

  • Digital collects data from vehicle sensors with consent from 1.6 mil customers which is then hosted in a central data lake for analyzing driver behavior, driving activity & other customer demographics.
  • This is intended to reduce insurance premium for selective drivers and provide automated recommendations for underwriting departments.

Environment: CDH, Hadoop, HDFS, Map Reduce, Hive, Pig, Sqoop, Unix, Oozie, Autosys.

Confidential - Delaware, NJ

Big Data/Hadoop Consultant

Responsibilities:

  • The main intension of the project is to create a central, integrated data platform, which can handle big data.
  • Data from RDBMS is pushed to HDFS, Transformations and Actions are applied on the data available in the HDFS using Hive and Map reduce.

Environment: Hadoop, CDH 5.5, Map Reduce, Hive, Pig, Sqoop, Flume, HBase, Java, Spark, Oozie, Linux, UNIX

Confidential

JAVA Developer

Responsibilities:

  • This is a comprehensive and integrated Hospital Management System for Confidential .
  • This package is used to computerize the activities of a servicing system. This package contains four modules.
  • They are Administration, Patient Care System, Health Services and Auxiliary Services.
  • Each module is having sub modules with in this sub modules will maintains storing, registering, details of joining patient details, and over all day activities maintains one particular database.

Environment: J2EE,JSP, JDBC, Servlets, HTML, CSS, ORACLE, Eclipse IDE

Confidential

JAVA/J2EE Developer

Responsibilities:

  • This project was developed for one of the top client in the financial sector.
  • It was developed with a flexible design which provides the platform to offer on-line integrated financial services to the bank customers.
  • Using this web based application customers can conduct activities like on-line retail banking, Secure messaging, Credit Card Payments, Shopping, Corporate banking and even

Confidential

SQL Developer Internship

Responsibilities:

  • Confidential has an integrated health care system that offers patients a continuum of coordinated high-quality care.
  • The system includes primary care and specialty physicians, community hospitals, specialty facilities, community health centers, and other health-related entities.
  • The project focused on developing a health care specific data-mart that enables the health care system to track the activities of existing patients to better serve them and offer them additional products and services

Environment: SQL Server 2000/2005, SQL, PLSQL and Windows XP/2005.

We'd love your feedback!