We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

Charlotte, NC

PROFESSIONAL SUMMARY:

  • 6+ years of extensive IT experience including 4+ years of recent experience in Big Data/Hadoop Ecosystem and 2 year on Development and Implementation of database applications using Oracle 12cg/11g, SQL and PL/SQL.
  • Having hands on experience in using Hadoop Technologies such as HDFS, HIVE, SQOOP, Impala, Flume, Solr.
  • Excellent knowledge of Hadoop Architecture and its related components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Extensive experience in building/writing Pig and Hive scripts for processing and analyzing large volumes of data.
  • Experience in writing Ad - hoc Queries for moving data from HDFS to HIVE and analyzing the data using HIVE QL.
  • Expertise in Spark framework for batch and real-time data processing.
  • Experience with Apache Spark ecosystem using Spark-SQL, Data Frames and RDD’s. Very good at loading data into spark schema RDD's and querying them using Spark-SQL. Good at writing custom RDD's in Scala.
  • Used Spark-SQL to perform transformations and actions on data residing in Hive.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems/ Non-Relational Database Systems and vice-versa.
  • Good understanding of NoSQL databases and hands on experience in writing applications on NoSQL databases like Hbase.
  • Using Hadoop ecosystem components for storage and processing data, exported data into Tableau using live connection. Having good experience on using Tableau Reporting Tool.
  • Having experience on using OOZIE to define and schedule the jobs
  • Experience on Storage and Processing in Hue covering all Hadoop ecosystem components.
  • Experience and good in-depth knowledge of YARN 2.0.
  • Experience in working with different data sources like Flat files, XML files and Databases.
  • Involved in importing Streaming data using FLUME to HDFS and analyzing using PIG and HIVE.
  • Experience in all Phases of Software Development Life Cycle (Requirement gathering, Analysis, Design, Development, Testing, implementation and Maintenance of applications) using Waterfall and Agile methodologies.
  • Extensively used PL/SQL to build Oracle Packages, Stored Procedures, Functions, Triggers, Views and Cursors for processing data.
  • Worked on advanced PL/SQL constructs like Oracle supplied packages, Nested Tables, VARRAYS, Records and Types.
  • Strong analytical, problem solving, multitasking and strategic planning skills.

TECHNICAL SKILLS:

Hadoop Framework: HDFS, MapReduce, Java, Hive, Pig, Hbase, Sqoop, Flume, Solr

Databases: HiveQL, Impala, Oracle 12c/11g

Languages: SQL, PL/SQL, Core Java

Tools: TOAD, Oracle SQL Developer

Operating Systems: Windows, Linux

IDE Tools: Eclipse

Web Technologies: HTML, JavaScript, CSS

Servers: Apache, Tomcat

WORK EXPERIENCE:

Confidential, Charlotte, NC

HADOOP Developer

Responsibilities:

  • Developing Spark programs using Scala API’s to compare the performance of Spark with Hive and SQL.
  • Used Spark API over Hortonworks Hadoop YARN to perform analytics on data in hive.
  • Implemented Spark using Scala and SparkSQL for faster testing and processing of data.
  • Designed and created Hive external tables using shared meta-store instead of derby with partitioning, dynamic partitioning and buckets.
  • Used Impala for querying HDFS data to achieve better performance.
  • Imported data from AWS S3 and into Spark RDD and performed transformations and actions on RDD’s.
  • Used the JSON and XML SerDe’s for serialization and de-serialization to load JSON and XML data into HIVE tables.
  • Used Spark-SQL to Load JSON data and create schema RDD and loaded it into HIVE Tables and handled Structured data using SparkSQL.
  • Develop Spark/MapReduce jobs to parse the JSON or XML data.
  • Involving Spark to improvise the performance and optimization of the existing algorithms in Hadoop using Spark context, spark-SQL, Data Frame, pair RDD's, Spark YARN.
  • Using Scala libraries to process XML data that was stored in HDFS and processed data was stored in HDFS.
  • Load the data into Spark RDD and do in memory data Computation to generate the Output response.
  • Used Spark for interactive queries, processing of streaming data and integration with popular NoSQL database for huge volume of data.
  • Wrote different pig scripts to clean up the ingested data and created partitions for the daily data.
  • Involved in converting Hive/SQL into Spark transformations using Spark RDD’s, Scala and Python.
  • Used Avro, Parquet and ORC data formats to store in to HDFS.
  • Used Oozie workflow to co-ordinate pig and Hive scripts.
  • Working with Kafka to get real-time weblogs data onto big data cluster

Environment: Hadoop, HDFS, Sqoop, Hive, Hbase, Sentry, Spark, Spark-SQL, Kafka, Flume, Oozie, JSON, Avro, Parquet, Talend, AWS, Zookeeper, Cloudera, Hortonworks .

Confidential, Rock Island, IL

Hadoop Developer

Responsibilities:

  • Loading customer data, spending data and credit from legacy warehouses to HDFS
  • Exported analyzed data to RDBMS using Sqoop for data visualization.
  • Used Hive queries to analyze the large data sets.
  • Build reusable Hive UDF's libraries for business requirements.
  • Implemented Dynamic Partitioning and bucketing in Hive.
  • Implement script to transmit sys print information from Oracle to HBase using Sqoop
  • Deployed the Big Data Hadoop application using Talend on cloud AWS (Amazon Web Service).
  • Implemented Map Reduce jobs on XML, JSON, CSV data formats.
  • Developed Map reduce programs which were used to extract and transform the data sets and the resultant dataset is loaded to HBase.
  • Imported the customers log data into HDFS using Flume.
  • Implemented Spark job to improve query performance.
  • Used Impala to handle different file formats
  • Proactively involved in ongoing maintenance, support, and improvements in Hadoop cluster.
  • Used Tableau as a business intelligence tool to visualize the customer information as per the generated records.

Environment: Hadoop, Map Reduce, HDFS, Hive, Sqoop, ZooKeeper, Oozie, Spark, Spark-SQL, Scala, Kafka, Java, Oracle, AWS S3.

Confidential

Software Engineer

Responsibilities:

  • Development of user interface windows for sales order processing, invoicing, customer billing & inventory.
  • Validation rules were implemented by writing procedures during data input.
  • Created stored procedures, Packages, Functions and triggers using SQL & PL/SQL.
  • Involved in analyzing user/business requirements and mapping into technical specifications.
  • Implemented a whole new security for the reports in the oracle database to prevent confidential data from being displayed to all users. The security is based on the standard concept of roles and privileges.
  • All the security roles and privileges are stored in the database tables and the logic is implemented using functions and views. Promoted the security feature to the existing reports with a very minor code change in the actual reports.
  • Managing table spaces, data files, redo logs, tables and its segments
  • Installation of Oracle Database and Oracle products.
  • Creating users and maintaining database security.
  • Managing Users, Privileges and Roles
  • Help users in creating database objects, and enrolling users
  • Prepare and maintained documents of database backup and Oracle database operations.
  • Responsible for setting and managing user accounts, Granting required privileges to user.
  • Writing SQL Loader control files for data loading.
  • Start up and shut down databases, to check for locks, monitor and kill user sessions, monitor database performance and other day-to-day tasks.

Environment: Oracle 12c/11g OEM, PL/SQL, SQL * PLUS, SQL*LOADER Linux, HP-UX, TOAD, STATSPACK, TKPROF, UTL SMTP.

Confidential

Oracle Developer

Responsibilities:

  • Development of user interface windows for sales order processing, invoicing, customer billing & inventory.
  • Validation rules were implemented by writing procedures during data input.
  • Created Custom reports using Reports 2.5.
  • Created stored procedures, Packages, Functions and triggers using SQL & PL/SQL.
  • Developed database triggers where ever necessary.
  • Involved in analyzing user/business requirements and mapping into technical specifications.
  • Performed in depth of data analysis and prepared weekly, biweekly, monthly reports by using reports 2.5.
  • Implemented a whole new security for the reports in the oracle database to prevent confidential data from being displayed to all users. The security is based on the standard concept of roles and privileges.
  • All the security roles and privileges are stored in the database tables and the logic is implemented using functions and views. Promoted the security feature to the existing reports with a very minor code change in the actual reports.

Environment: Oracle 10g, PL/SQL, SQL * PLUS

We'd love your feedback!