Bigdata Developer Resume
Phoenix, AZ
SUMMARY
- 9+ Years of work experience in Software Development, Design and Maintenance with technologies Hadoop/Hive/Pig/Sqoop/Hbase/Impala/Kafka, Linux, Oracle Database 9i/10g/11g, SQL and PL/SQL, Core Java/J2EE, REST APIs.
- 3+ yearsof experience asHadoop Developerand hands on experience in Hadoop Ecosystem.
- Experience in performing Big Data Analytics usingMapReducefor Apache Hadoop.
- Expert in Oracle SQL, PL/SQL for Application development and ETL/Datawarehouse.
- REST API Development in JAVA.
- In depth knowledge ofHadoop Architecture, HDFS file system, YARN, Serde, file types - AVRO, RC, ORC etc.
- Exporting data to Relational Databases usingSQOOPand vice versa.
- Usingflumeto load weblog data into HDFS.
- Analyzing data usingpig scriptsandhive queries.
- Good at writing custom Hive UDF, UDAF, UDTF & Pig UDF’s for analysis in Java.
- Strong in UNIX Shell scripting.
- Chaining jobs and implement workflows usingOOZIE.
- Knowledge of Installing and Administering the Hadoop Cluster. UsingCloudera Manager.
- Working knowledge of Hadoop federation and high-availability features in new release.
- Knowledge of NoSql and hands on experience inHBaseandMongoDB.
- Exposure to Spark Streaming to receive real time data from the Kafka and store the stream data to HDFS system.
- Experience as a Java Developer in Web/intranet, client/server technologiesusing Java, J2EE, Servlets, JSP, Servlets, JSF, EJB, JDBC, HTML, CSS and SQL.
- Knowledge of Springs, Struts and Hibernate Frameworks.
- Expert in frontend development using Powerbuilder.
- Knowledge of SQL Server, DB2 and MySQL, IBM Netezza.
- Report Development using Tableau, OBIEE.
- Knowledge and exposure to Strom, Solr, Elastic search.
- Data modeling using Erwin. Strong work experience on Conceptual, Logican and Physical dimensional modelling.
- Strong knowledge of ETL Star Schema, Snowflake schema, Implementing Slowly Changing Dimensions, Fact tables, measures etc.
TECHNICAL SKILLS
Big Data: Hive, Pig, Hue, Oozie, Kafka, Hbase, Sqoop, Impala, Spark.
Programming Languages: Core Java, Unix shell, Core Java
Software Tools: Power Builder, Tableau, Oracle Forms/Reports 10g, OBIEE.
Databases: Oracle 10g, 11g, SQL Server, DB2
Version Control: Clear Case, Subversion, Harvest
Operating Systems: Windows & UNIX
ETL: IBM Netezza, Informatica.
PROFESSIONAL EXPERIENCE
Confidential, Phoenix AZ
Bigdata Developer
Responsibilities:
- MapReduce, hive, Pig development.
- Worked on fetching and Storing of the data from Hive, Flat files to MAPR-File System.
- Developed UDFs, UADFs to prepare the data that will be feed to JAVA Mapreduce programs.
- Implemented Spark SQL and gained knowledge on Spark Streaming.
- Kafka as source to Spark.
- Involved in REST API Development using JAVA.
- Developed Java code that APIs will use and executes HIVE, PIG scripts as a part of Java Code.
- Worked on the Continues Integration of BigData build process.
- Expression Evaluations using Java Script.
- Extensively used Hbase/Mapr-DB Java APIs - PUT, GET, SCAN and Filter APIs.
- Performance Tuning of Hive and Pig jobs.
- Extensively worked on Hive Partitions, Buckets, External tables, Serde.
- Pushing data from Hive to Hbase.
Environment: MAP-R, JAVA/J2EE, Tomcat, Hive, Pig, Sqoop, Oozie, Hbase/MAPR-DB, UNIX.
Confidential
Sr. Bigdata Developer
Responsibilities:
- Architected Ingestion and export to and from HDFS.
- Loading data to HDFS with Sqoop from sources - Oracle, DB2.
- XMLs, CSVs and JSON file loading to HDFS and Querying with Hive.
- Designed Oozie work flow to perform Sqoop, Pig, Hive, Shell script, Mapreduce.
- Developed custom Mapreduce programs in Java for data processing.
- Real time Data ingestion using Kafka and Flume.
- Moving data back to Oracle warehouse from HDFS.
- Developed Hive UDFs.
- Loaded data into Hbase Tables.
- Gained Knowledge on Apache Spark, Hbase.
Environment: CDH, Oracle 11g, Hive, Pig, Sqoop. Oozie, Hbase UNIX, Flume, Kafka, Solr.
Confidential, Greenwood Village, CO
Bigdata Developer
Responsibilities:
- Data loading with Sqoop from Warehouse and Flat files to HDFS.
- Worked on Hive Joins, external table, partitioning, bucketing of tables and Complex queries.
- Developed Mapreduce programs for data access and manipulation.
- Developed custom Hive & Pig UDFs for transforming the data.
- Monitoring Daily Hive Jobs.
- Storing Data in Avro, Parquet formats.
- Created Oozie work flows to automate Sqoop, Hive jobs.
- Developed work flows using Kafka and Flume to collect messaged and storage into HDFS.
- Strong KnowledgeHadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, Data Node and MapReduce programming paradigm.
- Exposure to Hbase, Impala.
- Worked with Hive JsonSerde, AvroSerde.
Environment: Cloudera Hadoop distribution (CDH) 5.x, Oozie, Flume, Hive, Pig, Hbase, Impala, Hue, Beeswax.
Confidential, Jacksonville, FL
IT Developer
Responsibilities:
- Worked for Software development in RUP for Diamond Healthcare Claim processing system based on PowerBuilder and Oracle.
- Have done POC on Hadoop in out Datawarehouse environment.
- Weblog analysis using Hadoop Hive.
- Logical and Physical Design of Database.
- SQL, Oracle PL/SQL programming for the Provider administration, Pricing Benefit data management and Claims processing.
- Developed over 50 stored procedures, 10+ huge packages, triggers, Views, Synonyms, tables and Sequences to achieve business results.
- Data loading using SQL*Loader.
- Managed and monitored Production Control-M job Scheduling.
- Wored with PL/SQL Data Structures Records, Subtypes, Object Types.
- Tuned SQL queries and PL/SQL code using explain plan, analyze.
- Erwin for data modeling.
- Worked on Oracle Advances Queuing feature.
- Worked with Ref. Cursors, VARRAYS, Nested tables, DynamicSQL,BulkBinds, and Global Temporary Tables.
- Programmed Dynamic SQL.
- Fine-tuning of SQL using explain Plan and TKPROF.
- Worked with DBMS Oracle supplied package like UTL HTTP(SOAP), UTL FILE, DBMS JOBS, DBMS SCHEDULER, DBMS ALERT, etc.
- Exclusively developed packages with extensive use of SQL Query, Joining tables, SQL functions in Queries, PL/SQL with the use of Cursors, REF Cursors, and loops.
- Supporting Oracle 10g to 11g upgrade and Windows 7 migration of applications.
- Developed PowerBuilder Windows, User Objects, Data Objects.
- Production Support and finding root causes and providing sizing for R-Tickets.
- Developed the stored procedures using PL/SQL Tables and records to generate the output in XML format and pull data from Web services and from XML format and parse it and load it into tables.
- Developed Error handling & Exception system for Claims processing area.
- Data Obfuscation in Unit environments.
- Provided ETL scripts for the data loading.
- Worked on Claims processing in Member Benefits area and modified related Oracle Packages, Procedures.
- Exposure to applications using Java/J2EE, JSP, Servlets.
- Worked with DBMS LOBPackage to access and manipulation specific parts of a LOB or complete LOBs
- Used TOAD and SQL Developer for PL/SQL Development.
- Developed few reports using tableau.
- UsedPL/SQLfeatures likePL/SQLtables, Pragmas for error handling, implicit, explicit and Ref cursors for data processing.
- Used Exception Handling extensively for the ease of debugging and displaying the error messages in the application.
Environment: Oracle 10g/11g, CDH, Hadoop, Hive, pig, sqoop, kafka, Clearcase/Subversion, Erwin, Powerbuilder 9/10.5, Optum/McKesson tools, TOAD, Sharepoint, Windows XP/7,UNIX shell, Ctrl-M, PL/SQL, SQL*Loader, Control M, XML, Tableau, Java/J2EE, JSP, Servlets, Web Services, ETL Scripts, Claims Processing, Oracle Packages.
Confidential, Dallas, TX
Systems Analyst
Responsibilities:
- Used direct path loading via SQL*Loader/ SQL INSERT statement, create table as select statements, table spaces, parallelized SQL statement execution where appropriate tables.
- Developed a package with stored procedures, functions and database triggers for transferring the date from Source to warehouse.
- Analyzed data to check for Data Integrity and Referential Integrity when loaded to source-staging tables.
- Created tables to capture errors and data that failed Validation Checks.
- Developed tables with partitions and created indexes, constraints, triggers, synonyms, database links, table spaces, roles etc in Staging and Destination schemas.
- Developed dynamic SQL used in procedures and function for the data extraction and loading process.
- Used ERWIN to develop logical and physical database models.
- Web application development with Java/J2EE, Servlets, Struts.
- Created dynamic procedures to drop and create Index for the Staging environment to facilitate faster loading of data.
- Developed RDBMS including PL/SQL database backend programming with stored procedures, Functions and packages.
- Used Toad for creating PL/SQL (trigger, sequence, stored procedure).
- Used UTL FILE, DBMS OUTPUT, DBMS LOB packages effectively.
- Worked with DBA's in setting up and refreshing the UNIT, Test, Stage environments.
- Collected all impacted modules from all developers and prepare the deployment script for production release.
- Developed Shell scripts to automate execution of SQL scripts to check incoming data with master tables, insert the valid and invalid data into error tables.
- Involved in the database administration and creating objects such as database, users, and schema and table spaces.
Environment: Oracle 10g, TOAD, ERWIN, Clearcase, Java, Servlets, JSP, Struts.
Confidential, Stillwater, OK
Graduate Assistant
Responsibilities:
- Work as Graduate assistant at OSU for 10 hrs a week job using skills - CSS, HTML, Javascript and basic SQL and PL/SQL.
- Web Development using MS Front Page and Dreamweaver.
- Developed SQL and PL/SQL queries using joins, sub queries, and correlated sub queries.
Environment: MS Front Page, Dreamweaver, Filemaker Pro, Oracle, PL/SQL.