We provide IT Staff Augmentation Services!

Bigdata Developer Resume

Phoenix, AZ


  • 9+ Years of work experience in Software Development, Design and Maintenance with technologies Hadoop/Hive/Pig/Sqoop/Hbase/Impala/Kafka, Linux, Oracle Database 9i/10g/11g, SQL and PL/SQL, Core Java/J2EE, REST APIs.
  • 3+ yearsof experience asHadoop Developerand hands on experience in Hadoop Ecosystem.
  • Experience in performing Big Data Analytics usingMapReducefor Apache Hadoop.
  • Expert in Oracle SQL, PL/SQL for Application development and ETL/Datawarehouse.
  • REST API Development in JAVA.
  • In depth knowledge ofHadoop Architecture, HDFS file system, YARN, Serde, file types - AVRO, RC, ORC etc.
  • Exporting data to Relational Databases usingSQOOPand vice versa.
  • Usingflumeto load weblog data into HDFS.
  • Analyzing data usingpig scriptsandhive queries.
  • Good at writing custom Hive UDF, UDAF, UDTF & Pig UDF’s for analysis in Java.
  • Strong in UNIX Shell scripting.
  • Chaining jobs and implement workflows usingOOZIE.
  • Knowledge of Installing and Administering the Hadoop Cluster. UsingCloudera Manager.
  • Working knowledge of Hadoop federation and high-availability features in new release.
  • Knowledge of NoSql and hands on experience inHBaseandMongoDB.
  • Exposure to Spark Streaming to receive real time data from the Kafka and store the stream data to HDFS system.
  • Experience as a Java Developer in Web/intranet, client/server technologiesusing Java, J2EE, Servlets, JSP, Servlets, JSF, EJB, JDBC, HTML, CSS and SQL.
  • Knowledge of Springs, Struts and Hibernate Frameworks.
  • Expert in frontend development using Powerbuilder.
  • Knowledge of SQL Server, DB2 and MySQL, IBM Netezza.
  • Report Development using Tableau, OBIEE.
  • Knowledge and exposure to Strom, Solr, Elastic search.
  • Data modeling using Erwin. Strong work experience on Conceptual, Logican and Physical dimensional modelling.
  • Strong knowledge of ETL Star Schema, Snowflake schema, Implementing Slowly Changing Dimensions, Fact tables, measures etc.


Big Data: Hive, Pig, Hue, Oozie, Kafka, Hbase, Sqoop, Impala, Spark.

Programming Languages: Core Java, Unix shell, Core Java

Software Tools: Power Builder, Tableau, Oracle Forms/Reports 10g, OBIEE.

Databases: Oracle 10g, 11g, SQL Server, DB2

Version Control: Clear Case, Subversion, Harvest

Operating Systems: Windows & UNIX

ETL: IBM Netezza, Informatica.


Confidential, Phoenix AZ

Bigdata Developer


  • MapReduce, hive, Pig development.
  • Worked on fetching and Storing of the data from Hive, Flat files to MAPR-File System.
  • Developed UDFs, UADFs to prepare the data that will be feed to JAVA Mapreduce programs.
  • Implemented Spark SQL and gained knowledge on Spark Streaming.
  • Kafka as source to Spark.
  • Involved in REST API Development using JAVA.
  • Developed Java code that APIs will use and executes HIVE, PIG scripts as a part of Java Code.
  • Worked on the Continues Integration of BigData build process.
  • Expression Evaluations using Java Script.
  • Extensively used Hbase/Mapr-DB Java APIs - PUT, GET, SCAN and Filter APIs.
  • Performance Tuning of Hive and Pig jobs.
  • Extensively worked on Hive Partitions, Buckets, External tables, Serde.
  • Pushing data from Hive to Hbase.

Environment: MAP-R, JAVA/J2EE, Tomcat, Hive, Pig, Sqoop, Oozie, Hbase/MAPR-DB, UNIX.


Sr. Bigdata Developer


  • Architected Ingestion and export to and from HDFS.
  • Loading data to HDFS with Sqoop from sources - Oracle, DB2.
  • XMLs, CSVs and JSON file loading to HDFS and Querying with Hive.
  • Designed Oozie work flow to perform Sqoop, Pig, Hive, Shell script, Mapreduce.
  • Developed custom Mapreduce programs in Java for data processing.
  • Real time Data ingestion using Kafka and Flume.
  • Moving data back to Oracle warehouse from HDFS.
  • Developed Hive UDFs.
  • Loaded data into Hbase Tables.
  • Gained Knowledge on Apache Spark, Hbase.

Environment: CDH, Oracle 11g, Hive, Pig, Sqoop. Oozie, Hbase UNIX, Flume, Kafka, Solr.

Confidential, Greenwood Village, CO

Bigdata Developer


  • Data loading with Sqoop from Warehouse and Flat files to HDFS.
  • Worked on Hive Joins, external table, partitioning, bucketing of tables and Complex queries.
  • Developed Mapreduce programs for data access and manipulation.
  • Developed custom Hive & Pig UDFs for transforming the data.
  • Monitoring Daily Hive Jobs.
  • Storing Data in Avro, Parquet formats.
  • Created Oozie work flows to automate Sqoop, Hive jobs.
  • Developed work flows using Kafka and Flume to collect messaged and storage into HDFS.
  • Strong KnowledgeHadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, Data Node and MapReduce programming paradigm.
  • Exposure to Hbase, Impala.
  • Worked with Hive JsonSerde, AvroSerde.

Environment: Cloudera Hadoop distribution (CDH) 5.x, Oozie, Flume, Hive, Pig, Hbase, Impala, Hue, Beeswax.

Confidential, Jacksonville, FL

IT Developer


  • Worked for Software development in RUP for Diamond Healthcare Claim processing system based on PowerBuilder and Oracle.
  • Have done POC on Hadoop in out Datawarehouse environment.
  • Weblog analysis using Hadoop Hive.
  • Logical and Physical Design of Database.
  • SQL, Oracle PL/SQL programming for the Provider administration, Pricing Benefit data management and Claims processing.
  • Developed over 50 stored procedures, 10+ huge packages, triggers, Views, Synonyms, tables and Sequences to achieve business results.
  • Data loading using SQL*Loader.
  • Managed and monitored Production Control-M job Scheduling.
  • Wored with PL/SQL Data Structures Records, Subtypes, Object Types.
  • Tuned SQL queries and PL/SQL code using explain plan, analyze.
  • Erwin for data modeling.
  • Worked on Oracle Advances Queuing feature.
  • Worked with Ref. Cursors, VARRAYS, Nested tables, DynamicSQL,BulkBinds, and Global Temporary Tables.
  • Programmed Dynamic SQL.
  • Fine-tuning of SQL using explain Plan and TKPROF.
  • Worked with DBMS Oracle supplied package like UTL HTTP(SOAP), UTL FILE, DBMS JOBS, DBMS SCHEDULER, DBMS ALERT, etc.
  • Exclusively developed packages with extensive use of SQL Query, Joining tables, SQL functions in Queries, PL/SQL with the use of Cursors, REF Cursors, and loops.
  • Supporting Oracle 10g to 11g upgrade and Windows 7 migration of applications.
  • Developed PowerBuilder Windows, User Objects, Data Objects.
  • Production Support and finding root causes and providing sizing for R-Tickets.
  • Developed the stored procedures using PL/SQL Tables and records to generate the output in XML format and pull data from Web services and from XML format and parse it and load it into tables.
  • Developed Error handling & Exception system for Claims processing area.
  • Data Obfuscation in Unit environments.
  • Provided ETL scripts for the data loading.
  • Worked on Claims processing in Member Benefits area and modified related Oracle Packages, Procedures.
  • Exposure to applications using Java/J2EE, JSP, Servlets.
  • Worked with DBMS LOBPackage to access and manipulation specific parts of a LOB or complete LOBs
  • Used TOAD and SQL Developer for PL/SQL Development.
  • Developed few reports using tableau.
  • UsedPL/SQLfeatures likePL/SQLtables, Pragmas for error handling, implicit, explicit and Ref cursors for data processing.
  • Used Exception Handling extensively for the ease of debugging and displaying the error messages in the application.

Environment: Oracle 10g/11g, CDH, Hadoop, Hive, pig, sqoop, kafka, Clearcase/Subversion, Erwin, Powerbuilder 9/10.5, Optum/McKesson tools, TOAD, Sharepoint, Windows XP/7,UNIX shell, Ctrl-M, PL/SQL, SQL*Loader, Control M, XML, Tableau, Java/J2EE, JSP, Servlets, Web Services, ETL Scripts, Claims Processing, Oracle Packages.

Confidential, Dallas, TX

Systems Analyst


  • Used direct path loading via SQL*Loader/ SQL INSERT statement, create table as select statements, table spaces, parallelized SQL statement execution where appropriate tables.
  • Developed a package with stored procedures, functions and database triggers for transferring the date from Source to warehouse.
  • Analyzed data to check for Data Integrity and Referential Integrity when loaded to source-staging tables.
  • Created tables to capture errors and data that failed Validation Checks.
  • Developed tables with partitions and created indexes, constraints, triggers, synonyms, database links, table spaces, roles etc in Staging and Destination schemas.
  • Developed dynamic SQL used in procedures and function for the data extraction and loading process.
  • Used ERWIN to develop logical and physical database models.
  • Web application development with Java/J2EE, Servlets, Struts.
  • Created dynamic procedures to drop and create Index for the Staging environment to facilitate faster loading of data.
  • Developed RDBMS including PL/SQL database backend programming with stored procedures, Functions and packages.
  • Used Toad for creating PL/SQL (trigger, sequence, stored procedure).
  • Used UTL FILE, DBMS OUTPUT, DBMS LOB packages effectively.
  • Worked with DBA's in setting up and refreshing the UNIT, Test, Stage environments.
  • Collected all impacted modules from all developers and prepare the deployment script for production release.
  • Developed Shell scripts to automate execution of SQL scripts to check incoming data with master tables, insert the valid and invalid data into error tables.
  • Involved in the database administration and creating objects such as database, users, and schema and table spaces.

Environment: Oracle 10g, TOAD, ERWIN, Clearcase, Java, Servlets, JSP, Struts.

Confidential, Stillwater, OK

Graduate Assistant


  • Work as Graduate assistant at OSU for 10 hrs a week job using skills - CSS, HTML, Javascript and basic SQL and PL/SQL.
  • Web Development using MS Front Page and Dreamweaver.
  • Developed SQL and PL/SQL queries using joins, sub queries, and correlated sub queries.

Environment: MS Front Page, Dreamweaver, Filemaker Pro, Oracle, PL/SQL.

Hire Now