Hadoop/big Data Developer Resume
NJ
PROFESSIONAL SUMMARY
- 15 + years of experience in IT including 3+ years of expertise in design and development ofscalable distributed systems using HadoopEco System tools, Big Data Technologies.
- Comprehensive experience in Big Data processing using Hadoop Ecosystem including Pig, Hive, HDFS, Map Reduce (MRV1 and YARN), Sqoop, Flume, Kafka, Oozie, Zookeeper, Spark, Impala.
- Experience with all flavors of Hadoop distributions, including Cloudera, Hortonworks, Amazon Web Services distribution of Hadoop.
- Expertise in writing Hadoop Jobs for analyzing data using Hive Query Language (HQL), Pig Latin (Data flow language), and custom MapReduce programs in Java.
- Experience in writing Pig and Hive scripts to process structured and unstructured data and extending Hive and Pig core functionality by writing custom UDFs.
- Experience in data loading from Oracle and MySQL databases to HDFS system using SQOOP.
- Good understanding of NoSQL databases like MongoDB, Cassandra, and HBase.
- Expertise in using MongoDB for storing large data objects, real - time analytics, Logging and Full Text search.
- Hands on experience writing applications on HBase and expertise with SQL, PL/SQL database concepts.
- Good Knowledge on general data analytics on distributed computing cluster like Hadoop using Apache Spark and Scala.
- Having experience in developing a data pipeline using Kafka to store data into HDFS.
- Familiar in creating tables in Parquet format in Impala.
- Familiar in creating custom Solr Query components.
- Hands on experience in configuring and working with Flume to load the data from multiple sources directly into HDFS.
- Expertise in Hadoop workflows scheduling and monitoring using Oozie, Zookeeper.
- Detailed understanding of Software Development Life Cycle (SDLC) and sound knowledge of project implementation methodologies including Waterfall and Agile.
- An excellent team player and self-starter with good communication and inter-personal skills and proven abilities to finish tasks before target deadlines.
- 11+ years of experience in development using PL/SQL language.
- Credible experience in Oracle Forms 6i/10g/11g and Oracle Reports 6i/10g/11g.
- Good experience in developing Crystal Reports 8/11.
- Good experience with SAP Business Objects R2/R3.
- Expertise in Unix Shell Scripting, Perl Scripting, Batch Processing,SQL Tuning.
- Extensively used Oracle Designer 6i, Erwin for data modeling.
- 6+ years of experience in Healthcare and property casualty subrogation applications.
- 4+ years of experience in developing financial applications.
- 4 Years of Experience in Project and process management applications.
- Expertise in data loading using Oracle SQL* Loader, UTL Files, SSIS
TECHNOLOGY SKILLS:-
Hadoop/Big Data: Apache Hadoop, HDFS and Map Reduce, Pig, Hive, Sqoop, Flume, Hue, YARN, Oozie, Zookeeper,MapR Converged Data Platform,CDH,HDP,EMR, Apache Spark, Apache Kafka, Apache STORM, Apache Crunch, Avro, Parquet.
Web Technologies: AJAX, HTML5,JavaScript,CSS3,XML,SOAP, WSDL.
IDE Development Tools: Eclipse, Net Beans, My Eclipse, SOAP UI, Ant, Oracle Forms 11g/10g/9i/6i,Developer 2000, Oracle Designer 6i, VisualStudio 6/2005/2008/2010
Frameworks: MVC, Struts, Hibernate, Spring.
Web Servers: Web Logic, Web Sphere, Apache Tomcat, Oracle 11g/10g / 9i AS
NoSQL Databases: HBase, MongoDB, Cassandra.
Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP.
ETL &Reporting Tools: Informatica, Pentaho, SSIS, Cognos BI, Tableau, Hyperion, SSRS, Oracle Reports 10g/9i/6i, Oracle Discoverer, Crystal reports 8/11
Business Intelligence: Business Objects R2/R3
Databases: Oracle 11g/10g/9i/8i., MS SQL Server 7.0/6.5, 2000, 2005, 2008, MySql, DB2
System Programming Languages: PL/SQL, VB 6.0, C, C++, C#, Java,Unix Shell Scripting, Perl Scripting.
U-I Development: Oracle Forms 11g/10g/9i/6i,Developer 2000, Oracle Designer 6i, VisualStudio 6/2005/2008/2010
Operating Systems: Windows 2003/2008 Servers, Unix Server, Windows 95/98/2000 Professional/XP.
Version Control: PVCS,VSS,QVCS Version Manager
Office Suites: Microsoft Office 2007/2003/2000.
Technology Tools: Citrix 9.2, Winscp- FTP,Leach - FTP, PL/SQL Developer, Toad, SQL Loader. SQL Navigator, ERWIN
SDLC / Software Quality Assurance: Business Requirement Design (BRD), Unit Test Case Flow (UTC), System Test Case Flow (STC), System Integration Design (SID), Functional Design Model(FD), Technical Design Flow(TD), SEI-CMM Data Audits, ISO 9001-2000
PROFESSIONAL EXPERIENCE
Confidential, NJ
Hadoop/Big Data Developer
Responsibilities:
- Used Kafka for log aggregation like collecting physical log files off servers and puts them in a central place like HDFS for processing.
- Configured Spark streaming to receive real time data from the Kafka and store the stream data to HDFS.
- Also Used Spark SQL to process structured data in Hive.
- Involved in creating Hive tables, loading data, writing hive queries,generating partitions and buckets foroptimization.
- Developed Simple to complex Map/reduce Jobs using Hive and Pig for data cleaning and preprocessing.
- Analyzed large data sets by runningHive queries and Pig scripts.
- Written Hive UDF’s to sort Structure fields and return complex data type.
- Used different data formats (Text format and ORC format) while loading the data into HDFS.
- Involved in developing Shell scripts to easy execution of all other scripts (Pig, Hive, and MapReduce) andmove the data files within and outside of HDFS.
- Creating indexes and tuned the SQL queries in Hive using HUE.
- Created custom Solr Query components to enable optimum search matching.
- Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data.
- Acted for importing information under HBase utilizing HBase shell Also HBase customer API.
- Building, publishing customized interactive reports and dashboards, report scheduling using Tableau server.
- Created action filters, parameters and calculated sets for preparing dashboards and worksheets in Tableau
- Developed Tableau visualizations and dashboards using Tableau Desktop.
- Used Kafka to rebuild a user activity tracking pipeline as a set of real-time publish-subscribe feeds.
Environment:Hadoop,Cloudera,CDH4, CDH5,HDFS, MapReduce, YARN, Hive, Pig, Sqoop, Hbase, Apache Spark, Oozie Scheduler, Java, UNIX Shell Scripts,Impala, HUE,HCATALOG, Kafka,Solr, Git, Maven, BitBucket.
Confidential . Baltimore, MD
Hadoop Developer
Responsibilities:
- Coordinated with business customers to gather business requirements and worked under agile environment.
- Responsible for importing log files from various sources into HDFS using Flume.
- Processed Big Data using a Hadoop cluster consisting of 45 nodes.
- Performed complex HiveQL queries on Hive tablesto create, alter and drop tables.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Created final tables in Parquet format.
- Developed PIG scripts for source data validation and transformation.
- Developed Shell, Perl and Python scripts to automate and provide Control flow to Pig scripts.
- Developed NoSQL database by using CRUD, Indexing, Replication and Sharding in MongoDB. Sorted thedata by using indexing.
- Extracted and updated the data into MongoDB using Mongo import and export command line utility interface.
- Involved in unit testing using MR unit for Map Reduce jobs.
- Used Hive and Pig to generate BI reports.
- Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types ofHadoop jobs such as Java map-reduce Hive, Pig, and Sqoop.
Environment:Hadoop,HDFS, Pig, Hive, MapReduce, Java, Flume, Oozie, Linux/UNIX Shell Scripting, Avro,MongoDB, Python, Perl,Java (jdk1.7),Git, Maven, Jenkins.
Confidential, Dallas, TX
Hadoop Developer
Responsibilities:
- Involved in start to end process of Hadoop cluster, Hadoop ecosystem,Cloudera Manager Installation, configuration and monitoring using CDH3 Distribution.
- Worked extensively in creating MapReduce jobs for search and analytics in the identification of varioustrends.
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Integrating bulk data into Cassandra file system using MapReduce programs.
- Involved in creating Hive tables, and loading and analyzing data sets using hive queries and Pig scripts which will run internally in map reduce way.
- Extracted files from MySQL tablesin to HDFS using Sqoop.
- Involved in loading data from Linux/UNIX file system to HDFS.
- Written custom Hive and Pig UDF’s based on the requirements.
- Designed and Developed Dashboards using Tableau 6.
Environment: CDH3, Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, HBase, Cloudera Manager, Cassandra, MySQL, LINUX (CentOS), Tableau 6, Java, SQL.
Confidential, Louisville,KY
Sr. Programmer Anlyst
Responsibilities:
- Developed and maintain screens using Oracle Forms 11g/Reports 11g.
- Developed pl/sql procedure to run as Batch process for generating weekly reports.
- Created some Batch files which would run daily basis in order to expunge records.
- Design and development of PL/SQL objects such as Packages, functions, procedures, triggers etc to process complex business requirements.
- Used tools like leech for FTP of files.
- Developed reusable PLL Libraries used by Oracle Forms bringing code efficiency leading to Better performance, standardization and code optimization.
- Participated in the weekly meetings to understand functional requirements.
- Provide production support to the application.
- Planned and implemented data access control and application security at screen level.
- Developed stored procedure using MS SQL Server 2005.
- Developed html reports and oracle Reports using reports 10g
- Migrated reports from Crystal reports 8 to Crystal reports XI.
- Developed reports for various clients using Crystal Report XI and configured with Bussiness Objects R2.
- Extensive analysis to design the Universes and assign them to respective domains.
- Webi Reports were created and deployed through Info View.
- Installed and Fixed maintenance issues for .Net application Self Service Investigation
- Created batch files, Perl Scripts and Shell Scripts for processing end of month schedules.
- Involved in Performance Tuning for Business critical processes by re-coding functions, procedures and packages and creating/adding indexes.
- Client User Interactions for escalations regarding production issues, data corrections, performance improvements and enhancements.
- Developed PL/SQL Procedure to provide standardized XML output data file for external applications across organization.
- Configured APOS definitions and burst PDF report to Business Objects using APOS tool.
Environment: Unix, Oracle 11g/10g, Forms 11g/10g, Reports 11g/10g, .NET, ASP, SQL Server 6.5/7. 2000/2005/2008, Crystal reports 8/11, Business Objects R2, SAP Business Objects 3.1, APOS bursting manager, Microsoft Visual studio 2005/2008.
Confidential, Louisville, KY
Sr. Programmer Analyst
Responsibilities:
- Developed and maintain screens using Oracle Forms 10g.
- Developed reports using Reports 10g.
- Responsible for ensuring proper data analysis requirements are captured and documented to achieve the business objectives of the project.
- Developing and maintaining necessary database objects using PL/SQL which includes creating tables, functions, procedures, packages, table indexes and performance improvement methods.
- Developing Necessary forms Libraries like PLL’s, OLB’s and Form Templates.
- Created some Batch files which would run daily basis in order to expunge records.
- Involved in developing an application for enhancing the security for accessing the forms in the application.
- Used UNIX to make compile and implement the procedures, packages, forms that have been modified and developed into different databases.
- Created partitioned tables for manageability and scalability of the application.
- Performance Tuning, Business critical processes by re-coding and creating/adding indexes. Used Oracle hints for performance optimization.
- Involved in Production Support and data quality investigations.
- Monitoring DBMS Jobs for their success and failures and taking necessary action.
- Created Shell script for wrapping Oracle Procedural calls. Also created shell scripts for data validation, data transfer.
- Testing and taking measures for maximum performance.
Environment: Windows, Oracle 10g, Forms 10g, Reports 10g, Oracle webutil, Unix, MS SQL Server
Confidential
Sr. Software Developer
Responsibilities:
- Development and maintenance of PL/SQL Stored Program units - Packages, triggers.
- Conversion of existing screens from Forms 6i to Forms 10g.
- Developed and maintain screens using Oracle Forms 10g.
- Provide production support to the entire application.
- Planned and implemented data access control and application security at screen level.
- Developed reports using Reports 10g.
- Responsible for ensuring proper data analysis requirements are captured and documented to achieve the business objectives of the project.
- Developing and maintaining necessary database objects using PL/SQL which includes creating tables, functions, procedures, packages, table indexes and performance improvement methods.
- Developing and Deploying Form Interfaces and Reports required by the system using Forms 6i/10g,Reports 6i/10g and Application Server 10g using all important form triggers built in packages MDI application with multiple window and canvas interfaces, Dynamic LOV, Programmatically Record Group handling.
- Developing Necessary forms Libraries like PLL’s, OLB’s and Form Templates.
- Created partitioned tables for manageability and scalability of the application.
- Testing and taking measures for maximum performance.
Environment: Windows, Oracle 10g, Forms 10g, Reports 10g, Oracle webutil, Crystal reports XI, Unix, MS SQL Server
Confidential
Team Lead/Sr. Developer
Responsibilities:
- User requirements analysis and design of screens.
- Involved in Client meetings to gather requirements.
- Developed the database design using Designer 6i and integrated to the other modules.
- Developed and modified Oracle packages, procedures, functions and triggers.
- Calling the sql and report files from shell scripts.
- Extraction of data from different flat files into Oracle Database using SQL* Loader.
- Developed Screens and reports using Oracle Forms 6i and Oracle Reports 6i
- Tuned SQL by using Explain Plan, Indexing and Hints.
- Used Shell scripts to automate Oracle script execution.
- Developed Deviation Scripts between testing and Development databases using TOAD.
- Maintenance and debugging of existing financial modules.
- Involved in integration of finance application with other modules Human resources and Payroll.
- Developed pl/sql procedure to run as Batch process for generating weekly reports.
- Implemented PL/SQL scripts in accordance with the necessary Business rules and procedures using Pl/Sql Developer.
- Extensively involved in coding of the procedures, triggers, and making use of built in packages and libraries while developing Forms.
- Modified existing stored procedures, functions and triggers using PL/Sql Developer.
- Generated SQL and PL/SQL scripts to create and drop database objects including: Tables, Views, Primary keys, Indexes, Constraints, Packages, Sequences and Synonyms.
- Performance tuning of the queries using optimizer hints and creating new Indexes.
- Used tools like leech for FTP of files.
- Optimizing, and fine-tuning the Oracle Reports for efficiency.
- Developed reusable PLL Libraries used by Oracle Forms bringing code efficiency leading to better performance, standardization and code optimization
Environment: Win2000,Oracle 8i/9i, SQL*Loader, Oracle 9i Application Server, Forms 6i /Reports 6i, PL/SQL, Oracle Designer 6i, Unix, TOAD