We provide IT Staff Augmentation Services!

Software Developer Resume

PROFESSIONAL SUMMARY:

  • Around 7 years of experience as a Software Developer in Design, Development, Deployment and Support of Large Scale Distributed Systems
  • About 6 years of experience in using various Hadoop Ecosystem components like Flume, HDFS , Hive, HBase , Kafka , MapReduce, Oozie, Pig, Sqoop , YARN , Zookeeper .
  • 2 years of experience using Spark and Spark Streaming for analysis of streaming data.
  • Extensive experience in creating Spark RDDs , Pair RDDs , Data Frames from various file formats including CSV, JSON, Parquet and textfiles.
  • Hands on experience in developing Spark code using Python as well as Spark Streaming and SparkSQL code in Python for dealing with streaming data.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Experience in working with Hive and HiveQL for creating managed as well as external tables from files on HDFS .
  • Strong experience creating UDFs and UDAFs in Pig and Hive meeting the project requirements.
  • Involved in converting Hive/SQL queries into Spark Transformations using Spark RDDs and Python .
  • Good knowledge about Spark Transforms, Dstream operations, Dataframe functions and working with key-value pairs in a dataset.
  • Experience using Kafka cluster for Data Integration and doing Data summarization, Querying and Analysis of large Datasets stored on HDFS and Amazon S3 filesystem using Hive Query Language (HiveQL).
  • Experienced working with various file formats such as txt, Avro, Parquet, ORC and JSON.
  • Extensive experience in Data Ingestion, Transformation, Analytics using Apache Spark framework, and Hadoop ecosystem components.
  • Strong exposure to NoSQL Databases like HBase , Redis , Cassandra .
  • Good Experience in writing complex SQL queries with databases like Oracle 11g, MySQL, SQL Server and MS SQL Server 2005/2008.
  • Hands on experience using Cloudera and Hortonworks’ distribution for Hadoop , Amazon Web Services(AWS) .
  • Experience in designing both time driven and data driven automated workflows using Oozie.
  • Some understanding about using Google Cloud Platform components like Big Query , Dataflow , Dataproc and PubSub .
  • Experienced in writing Down-Stream and Up-Stream Pipelines using Python OOP.
  • Knowledge about using Tableau for Data Visualization.
  • Experience in Test-driven development, Software Development Life Cycle (SDLC)methodologies like Agile and Scrum .
  • Experience in using source code management tools like Git , SVN , Jira .

TECHNICAL SKILLS:

Big Data Technologies: Hadoop, HDFS, MapReduce, Hive, Pig, Spark, Spark Streaming, Sqoop, Oozie, Kafka, Flume, Zookeeper, Hue

Programming Languages: Python, C, C++, C#

Databases: Cassandra, HBase, Redis, MySQL, MS SQL Server, Oracle

IDEs: IntelliJ, Visual Studio, Pycharm

Version Control: Git, SVN

Operating Systems: Mac OS, Linux (Various Flavors), Windows XP/7/8/10

PROFESSIONAL EXPERIENCE:

Confidential

Software Developer

Responsibilities:

  • Involved in converting Hive/SQL queries into Spark Transformations using Spark RDDs and Python .
  • Good knowledge about Spark Transforms, Dstream operations, Dataframe functions and working with key-value pairs in a dataset.
  • Experience using Kafka cluster for Data Integration and doing Data summarization, Querying and Analysis of large Datasets stored on HDFS and Amazon S3 filesystem using Hive Query Language (HiveQL).
  • Experienced working with various file formats such as txt, Avro, Parquet, ORC and JSON.
  • Extensive experience in Data Ingestion, Transformation, Analytics using Apache Spark framework, and Hadoop ecosystem components.
  • Strong exposure to NoSQL Databases like HBase , Redis , Cassandra .
  • Good Experience in writing complex SQL queries with databases like Oracle 11g, MySQL, SQL Server and MS SQL Server 2005/2008.
  • Hands on experience using Cloudera and Hortonworks’ distribution for Hadoop , Amazon Web Services(AWS) .
  • Experience in designing both time driven and data driven automated workflows using Oozie.
  • Some understanding about using Google Cloud Platform components like Big Query , Dataflow , Dataproc and PubSub .
  • Experienced in writing Down-Stream and Up-Stream Pipelines using Python OOP.
  • Knowledge about using Tableau for Data Visualization.
  • Experience in Test-driven development, Software Development Life Cycle (SDLC)methodologies like Agile and Scrum .
  • Experience in using source code management tools like Git , SVN , Jira .

Hire Now