Senior Java And Bigdata Developer Resume
Gillette, WY
SUMMARY
- 9+ Years of work experience in Software Development, Design and Maintenance with technologies Hadoop/Hive/Pig/Sqoop/Hbase/Impala/Kafka, Linux, Oracle Database SQL and PL/SQL, Core Java/J2EE, REST APIs.
- 3+ years of experience as Hadoop Developer and hands on experience in Hadoop Ecosystem.
- Experience in performing Big Data Analytics using MapReduce for Apache Hadoop. Expert in Oracle SQL, PL/SQL for Application development and ETL/Datawarehouse.
- Architected and developed workflows on Hadoop HDFS, NOSQL databases, Sqoop, Pig, hue, Hive, Pig, spark, Kafka, Impala, Hbase, flume, Oozie and Linux.
- Strong experience in using Kafka and Flume to capture messages and store in HDFS. Strong experience on SQL using Oracle, SQL Server, Sybase and MS Access.
- Designed and developed collections, shards and replicas in Solr and successfully implemented Cloudera Search functionality. REST API Development in JAVA.
- In depth knowledge of Hadoop Architecture, HDFS file system, YARN, Serde, file types - AVRO, RC, ORC etc.
- Exporting data to Relational Databases using SQOOP and vice versa. Using flume to load weblog data into HDFS. Analyzing data using pig scripts and hive queries.
- Good Confidential writing custom Hive UDF, UDAF, UDTF & Pig UDF's for analysis in Java. Strong in UNIX Shell scripting.
- Chaining jobs and implement workflows using OOZIE. Knowledge of Installing and Administering the Hadoop Cluster. Using Cloudera Manager.
- Working knowledge of Hadoop federation and high-availability features in new release. Knowledge of NoSql and hands on experience in HBase and MongoDB.
- Exposure to Spark Streaming to receive real time data from the Kafka and store the stream data to HDFS system.
- Experience as a Java Developer in Web/intranet, client/server technologies using Java, J2EE, Servlets, JSP, Servlets, JSF, EJB, JDBC, HTML, CSS and SQL.
- Knowledge of Springs, Struts and Hibernate Frameworks. Expert in frontend development using Powerbuilder. Knowledge of SQL Server, DB2 and MySQL, IBM Netezza.
- Report Development using Tableau, OBIEE.
- Knowledge and exposure to Strom, Solr, Elastic search. Data modeling using Erwin. Strong work experience on Conceptual, Logican and Physical dimensional modelling.
- Strong knowledge of ETL Star Schema, Snowflake schema, Implementing Slowly Changing Dimensions, Fact tables, measures etc.
TECHNICAL SKILLS
Big Data: Hive, Pig, Hue, Oozie, Kafka, Hbase, Sqoop, Impala, Spark.
Programming Languages: Core Java, Unix shell, Core Java
Software Tools: Power Builder, Tableau, Oracle Forms/Reports 10g, OBIEE.
Databases: Oracle 10g, 11g, SQL Server, DB2
Version Control: Clear Case, Subversion, Harvest
Operating Systems: Windows & UNIX
ETL: IBM Netezza, Informatica.
PROFESSIONAL EXPERIENCE
Confidential, Gillette, WY
Senior Java and Bigdata Developer
Responsibilities:
- MapReduce, hive, Pig development. Worked on fetching and Storing of the data from Hive, Flat files to MAPR-File System.
- Developed UDFs, UADFs to prepare the data that will be feed to JAVA Mapreduce programs.
- Implemented Spark SQL and gained knowledge on Spark Streaming. Architected various ingestions using Spark, Kafka and flume with HDFS, Hive and elastic search as destinations.
- Kafka as source to Spark. Involved in REST API Development using JAVA.
- Developed Java code that APIs will use and executes HIVE, PIG scripts as a part of Java Code.
- Experience in developing work flows to ingest data from SQL server, oracle and my sql to hadoop file system using Apache Sqoop. Worked on the Continues Integration of BigData build process.
- Expression Evaluations using Java Script. Extensively used Hbase/Mapr-DB Java APIs - PUT, GET, SCAN and Filter APIs.
- Performance Tuning of Hive and Pig jobs. Extensively worked on Hive Partitions, Buckets, External tables, Serde.
- Pushing data from Hive to Hbase. Presented in-depth product wins and loss trend analysis as a part of quarterly product business reviews to execute team.
- Worked closely with strategic OEM partners like IBM & SAP to launch IP storage solution for BigData focused initiatives within educational and large enterprise customer segment.
Environment: MAP-R, JAVA/J2EE, Tomcat, Hive, Pig, Sqoop, Oozie, Hbase/MAPR-DB, UNIX.
Confidential, Chesterfield, MO
Senior Java and Bigdata Developer
Responsibilities:
- Architected Ingestion and export to and from HDFS.
- Loading data to HDFS with Sqoop from sources - Oracle, DB2.
- XMLs, CSVs and JSON file loading to HDFS and Querying with Hive.
- Architected and developed collections/shards in Solr and successfully implemented Cloudera Search functionality.
- Designed Oozie work flow to perform Sqoop, Pig, Hive, Shell script, Mapreduce.
- Developed custom Mapreduce programs in Java for data processing.
- Real time Data ingestion using Kafka and Flume.
- Moving data back to Oracle warehouse from HDFS.
- Developed Hive UDFs.
- Extensively used Solr Admin portal and Cloudera Manager.
- Loaded data into Hbase Tables.
- Gained Knowledge on Apache Spark, Hbase.
Environment: CDH, Oracle 11g, Hive, Pig, Sqoop. Oozie, Hbase UNIX, Flume, Kafka, Solr.
Confidential, Covington,WA
Bigdata Developer
Responsibilities:
- Data loading with Sqoop from Warehouse and Flat files to HDFS.
- Worked on Hive Joins, external table, partitioning, bucketing of tables and Complex queries.
- Developed Mapreduce programs for data access and manipulation.
- Developed custom Hive & Pig UDFs for transforming the data.
- Monitoring Daily Hive Jobs.
- Storing Data in Avro, Parquet formats.
- Created Oozie work flows to automate Sqoop, Hive jobs.
- Developed work flows using Kafka and Flume to collect messaged and storage into HDFS.
- Strong KnowledgeHadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, NameNode, Data Node and MapReduce programming paradigm.
- Exposure to Hbase, Impala.
- Worked with Hive JsonSerde, AvroSerde.
Environment: Cloudera Hadoop distribution (CDH) 5.x, Oozie, Flume, Hive, Pig, Hbase, Impala, Hue, Beeswax.
Confidential, Flint, MI
IT Developer
Responsibilities:
- Worked for Software development in RUP for Diamond Healthcare Claim processing system based on PowerBuilder and Oracle.
- Have done POC on Hadoop in out Datawarehouse environment.
- Weblog analysis using Hadoop Hive.
- Logical and Physical Design of Database.
- SQL, Oracle PL/SQL programming for the Provider administration, Pricing Benefit data management and Claims processing.
- Developed over 50 stored procedures, 10+ huge packages, triggers, Views, Synonyms, tables and Sequences to achieve business results.
- Data loading using SQL*Loader.
- Managed and monitored Production Control-M job Scheduling.
- Wored with PL/SQL Data Structures Records, Subtypes, Object Types.
- Tuned SQL queries and PL/SQL code using explain plan, analyze.
- Erwin for data modeling.
- Worked on Oracle Advances Queuing feature.
- Worked with Ref. Cursors, VARRAYS, Nested tables, Dynamic SQL, Bulk Binds, and Global Temporary Tables.
- Programmed Dynamic SQL.
- Fine-tuning of SQL using explain Plan and TKPROF.
- Worked with DBMS Oracle supplied package like UTL HTTP(SOAP), UTL FILE, DBMS JOBS, DBMS SCHEDULER, DBMS ALERT, etc.
- Exclusively developed packages with extensive use of SQL Query, Joining tables, SQL functions in Queries, PL/SQL with the use of Cursors, REF Cursors, and loops.
- Supporting Oracle 10g to 11g upgrade and Windows 7 migration of applications.
- Developed PowerBuilder Windows, User Objects, Data Objects.
- Production Support and finding root causes and providing sizing for R-Tickets.
- Developed the stored procedures using PL/SQL Tables and records to generate the output in XML format and pull data from Web services and from XML format and parse it and load it into tables.
- Developed Error handling & Exception system for Claims processing area.
- Data Obfuscation in Unit environments.
- Provided ETL scripts for the data loading.
- Worked on Claims processing in Member Benefits area and modified related Oracle Packages, Procedures.
- Exposure to applications using Java/J2EE, JSP, Servlets.
- Worked with DBMS LOB Package to access and manipulation specific parts of a LOB or complete LOBs
- Used TOAD and SQL Developer for PL/SQL Development.
- Developed few reports using tableau.
- Used PL/SQL features like PL/SQL tables, Pragmas for error handling, implicit, explicit and Ref cursors for data processing.
- Used Exception Handling extensively for the ease of debugging and displaying the error messages in the application.
Environment: Oracle 10g/11g, CDH, Hadoop, Hive, pig, sqoop, kafka, Clearcase/Subversion, Erwin, Powerbuilder 9/10.5, Optum/McKesson tools, TOAD, Sharepoint, Windows XP/7, UNIX shell, Ctrl-M, PL/SQL, SQL*Loader, Control M, XML, Tableau, Java/J2EE, JSP, Servlets, Web Services, ETL Scripts, Claims Processing, Oracle Packages.
Confidential, Dearborn, MI
Systems Analyst
Responsibilities:
- Used direct path loading via SQL*Loader/ SQL INSERT statement, create table as select statements, table spaces, parallelized SQL statement execution where appropriate tables.
- Developed a package with stored procedures, functions and database triggers for transferring the date from Source to warehouse.
- Analyzed data to check for Data Integrity and Referential Integrity when loaded to source-staging tables.
- Created tables to capture errors and data that failed Validation Checks.
- Developed tables with partitions and created indexes, constraints, triggers, synonyms, database links, table spaces, roles etc in Staging and Destination schemas.
- Developed dynamic SQL used in procedures and function for the data extraction and loading process.
- Used ERWIN to develop logical and physical database models.
- Web application development with Java/J2EE, Servlets, Struts.
- Created dynamic procedures to drop and create Index for the Staging environment to facilitate faster loading of data.
- Developed RDBMS including PL/SQL database backend programming with stored procedures, Functions and packages.
- Used Toad for creating PL/SQL (trigger, sequence, stored procedure).
- Used UTL FILE, DBMS OUTPUT, DBMS LOB packages effectively.
- Worked with DBA's in setting up and refreshing the UNIT, Test, Stage environments.
- Collected all impacted modules from all developers and prepare the deployment script for production release.
- Developed Shell scripts to automate execution of SQL scripts to check incoming data with master tables, insert the valid and invalid data into error tables.
- Involved in the database administration and creating objects such as database, users, and schema and table spaces.
Environment: Oracle 10g, TOAD, ERWIN, Clearcase, Java, Servlets, JSP, Struts.
Confidential, Albertville, MN
Systems Analyst
Responsibilities:
- Work as Graduate assistant Confidential OSU for 10 hrs a week job using skills - CSS, HTML, Javascript and basic SQL and PL/SQL.
- Web Development using MS Front Page and Dreamweaver.
- Developed SQL and PL/SQL queries using joins, sub queries, and correlated sub queries.
Environment: MS Front Page, Dreamweaver, Filemaker Pro, Oracle, PL/SQL.