We provide IT Staff Augmentation Services!

Cloudera Hadoop Developer/admin Resume

5.00/5 (Submit Your Rating)

Pleasanton, CA

PROFESSIONAL SUMMARY:

  • Business driven and result oriented Technologist, with 13+ years of IT experience spread over Architecture, design, development, testing, implementation and support of enterprise Legacy and ETL Data Warehouse(EDW) Business Intelligence applications comprising of in depth business knowledge in Payment Systems/Fraud/Risk, Retail Banking & Securities, Healthcare Membership/Enrollment and Pharmacy/Prescription based business applications, Strategic Project Management.
  • Specializes in data warehousing, Relational Database modelling, ETL methodology and Business Intelligence infrastructure building, data migration, and relational database systems development with Waterfall and Agile SDLC model.
  • Experienced in Requirement Analysis, Developing Business Systems Source to Target Mapping, Test Design, Test Case preparation, Test Execution and Management.
  • Strong understanding of Data Modelling in data warehouse environment such as star schema and snowflake schema.
  • Expertise in Data Integration and Data Conversion ETL development projects, building Business processes/graphs/mappings using ETL tools Ab Initio 3.x, Informatica Powercenter 9.x.
  • Solid hands on experience in developing PL/SQL, Stored Procedure and Unix Shell scripting.
  • Experienced with Batch systems scheduling and processing using Control M, Tivoli, Autosys, Crontab etc.
  • Experienced in Business Intelligence Reporting tools such as Micro strategy, Cognos, and Tableau.
  • Cloudera Certified Hadoop Developer, Cloudera Certified Hadoop Administrator.
  • Sound programming capability using Python, core JAVA along with Hadoop framework utilizing Cloudera Hadoop Ecosystem projects (HDFS, Spark, Sqoop, Hive, HBase, Oozie, Impala, Zookeeper etc.).
  • A successful and experienced Project Management Professional (PMP) - Subject matter expert in change, configuration and release management processes with strong process discipline complemented with extensive experience in software development lifecycle and project management methodologies.
  • Extensive experience in managing Global delivery teams - onsite & offshore comprising of ~25 team members.
  • Has strong development experience working in Legacy IBM Mainframe z O/S/MVS, COBOL, DB2, CICS, JCL, and Enterprise level job schedulers like Control M, TWS, Autosys and Tivoli.
  • Currently leading/managing several Development projects across OPPR program/Pharmacy Data Warehouse for Confidential Care Delivery (CD BIO) Applications.

TECHNICAL SKILLS:

  • Payment Systems: Point of Sale, Fraud, Risk, Authorization business modules for Confidential .: 8 years
  • Retail Banking: Securities & Loans/Deposits business modules for Citi Bank: 1 year
  • Healthcare: Membership, Enrollment, Claim, Adjudication, Prescription Sale, Pharmacy/Clinical, New Member onboarding/communication business modules for Confidential: 5 + years
  • Solution Consultant: Expert in Data Warehouse Architecture, Design and Data Modelling.
  • IT Project Lifecycle: Requirement Analysis, Design, Testing/QA, Implementation, Support, Costing & Budgeting, Project Scheduling.
  • Exposure to both traditional Waterfall model as well as Agile Development Methodology.
  • Core experience in ETL framework: Ab Initito 3.x, Informatica 9.x, Unix Shell Scripting.
  • Sound programming capability using Python, Core Java, Java Enterprise Edition(JEE), C, C++
  • IBM & Cloudera Certified Big Data Consultant - Extensively working in Cloudera Hadoop Framework utilizing Hadoop, Scoop, Flume, Hive, HBase, Impala, Yarn MR2, Zookeeper, Oozie, NoSQL databases.
  • Data Modelling, Data Architecture, Data Mining, Data Analytics.
  • Database exposure to RDBMS/DB2/Oracle: SQL freak.
  • Experienced in BI Reporting suite: Micro strategy, Cognos, and Tableau.
  • Batch Scheduling using: Control M, TWS, Tivoli, Autosys, Crontab.
  • SDLC Release/Change management/tracking using HP Quality Center (QC), BMC Remedy, VIPER.
  • Mainframe, MVS Z O/S, DB2, JCL, VSAM, Infoman, Expeditor, Endeavor, COBOL, FILEAID, DFSORT, CICS, SDSF/IOF,TSO/ISPF, Easytrieve.
  • Project Management Professional (PMP) Custom Software Development, Database Design (RDBMS), System Migrations, and Enterprise wide Integrations.
  • Valueadded leadership Cross: Functional supervision, Team building & mentoring, Client relations & presentations, Business & IT Planning, Vendor Management.

PROFESSIONAL EXPERIENCE:

Confidential, Pleasanton, CA

Cloudera Hadoop Developer/Admin

Responsibilities:

  • Actively involved in designing, developing and implementing a Data Repository to retain last 10 years clinical data, comprising of pharmacy point of sale transactions, covering various scenarios like Sale, Void/Unsell, Return, etc. CDR is much valuable for Kaiser to identify various kind of patterns like Fraud (prescription drug being stolen in pharmacy), trends in returns, max wait time in a particular pharmacy etc., which altogether helps business to improve the patient care/service at Pharmacy and prevent proactively events of fraud etc.
  • Executing adhoc queries for trend analysis and finding various transaction patterns.
  • Building Hadoop clusters/loading to retain 10 years back pharmacy transactions making it readily available for any kind of data analysis, replacing slowly dependency on limited period data retention on traditional EDW.
  • Identifying other related source of data to build a comprehensive and robust data repository with the intention to make it a single marketplace for future generation fast paced Pharmacy Analytics platform.
  • Imported data using Sqoop to load data from relational database/ RDMS/Oracle tables to HDFS on a regular basis.
  • Created Hive tables, loaded data and wrote Hive QL queries to further analyze the data.
  • Created Hive queries that helped Pharmacy prescription sales analysis spot emerging trend/patterns by comparing fresh data with EDW reference tables and historical metrics.
  • Involved in loading data from UNIX/Linux file system to HDFS.
  • Designed and developed automated, complex and efficient ETL processes to match multiple large-scale dataset together.
  • Developed Map Reduce Programs for data analysis and data cleaning.
  • Migration of ETL processes from MySQL to Hive to test the easy data manipulation.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, Pig jobs and Hive jobs.
  • Worked on Hive for further analysis and for generating/transforming files from various analytical formats to text files.
  • Managed and reviewed Hadoop log files.

Used platforms includes: Cloudera Apache Hadoop Ecosystem - Cloudera Hadoop Framework, Hive, HDFS, Impala, Sqoop, Spark, UNIX AIX, Oracle 11g, Flat Files, and IBM AIX, Python.

Confidential, Pleasanton, CA

Data Warehouse Project Lead & Lead Data Analyst

Responsibilities:

  • Designed the application solution framework, developed and implementation of extraction, transformation and load process (ETL), OLTP reporting for the Pharmacy Data Warehouse from heterogeneous source systems using Informatica 9.5.x.
  • Major projects worked are: EED Cutover, Bridgefeed Cutover, EPS General Ledger (GL), POS Transaction Data Store (TDS), Gateway, Performance tuning of EPS, GL Decoupling.
  • Developed slowly changing dimension mappings of type1, type2 using various transformations like Unconnected and Connected lookups, Router, Aggregator, Sorter, ranker, Joiner, Stored procedure and Update Strategy.
  • Used PL/SQL to embed complex business logic in stored procedures.
  • Built Exception Handling Mappings for Data Quality, Data Cleansing and Data Validation.
  • Implemented partition based staging load utilizing pre-SQL and post-SQL in session configurations.
  • Created UNIX shell scripts for file integration and Model scripts.
  • Extensively worked on Code Migration from Development, QA, Systest, UAT, and Production.
  • Actively involved in tuning performance for existing and new source queries - in Informatica and database level.
  • Worked closely with Business owner/regions, SME to capture Business requirements and translate into Technical Specifications/Functional Requirement Document (FRD)/Logical Solution Design Document (LSDD).
  • Performed Source System Analysis (SSA) to identify required source data for new business requirements to move data elements to target tables.
  • Involved in effort estimation, risk identification and project planning/scheduling.
  • Involved in Data Modeling, Design and Development of Informatica objects.
  • Prepared Test Plan, Test Case, and Test Scripts for Unit testing, System Integration testing (SIT), and User Acceptance testing (UAT).
  • Performed Impact Analysis based on the Change request.
  • Effort Estimations for Change Requests/new enhancements of ETL mappings.
  • Change Management, working on Production Code Deployment process involving CRQ, Work Order, tasks in BMC Remedy tool.
  • PDW applications generate various downstream feeds for regions which help them to perform various pharmacy/clinical data research & General Ledger is extremely important for Kaiser Finance team for auditing the Prescription Sales Revenue from Pharmacies across all regions.

Used platforms includes: Informatica Power Center 9.5.x, Oracle 11g/10g, IBM Mainframe, SQL, UNIX Korn Shell, HP Quality Center 11.52, Rally (Agile Methodology), IBM TWS Scheduler, Flat Files, and IBM AIX, Remedy.

Confidential, Pasadena, CA

Data Warehouse Project Lead & Data Analyst

Responsibilities:

  • Developed and implemented ETL solutions for LIS Acumen Business Validation Automation process and further in LACARE business rules using ETL tool Informatica Power Center 9.5.x/ Ab Initio.
  • Created several Informatica Mappings for the CDC change data capture to populate data into Dimensions and Fact tables.
  • Involved in Performance optimization in Informatica and database level.
  • Used SQL, PL/SQL to embed complex business logic in stored procedures and functions.
  • Created UNIX shell and Perl scripts for file integration.
  • Effort Estimations for Change Requests/new enhancements of ETL mappings.
  • Carrying out the Code Reviews and Production deployments process using Confidential Standards.
  • Defined scope and business rules of project, gathering Business Requirement and documenting those using models and further translated into Functional Requirement Document (FRD)/Logical Solution Design Document (LSDD).
  • Created data flow diagrams, data mapping from Source to stage to target mapping documents indicating the source tables, columns, business rules/filters/transformations to be applied.
  • Wrote Complex SQL scripts to perform Data validation and Data integrity testing.
  • Prepared Test Plan, Test Case, and Test Scripts for Unit testing, System Integration testing (SIT), and User Acceptance testing (UAT).
  • Involved in effort estimation, risk identification and project planning/scheduling.

Used platforms includes: Informatica Power Center 9.5.x, Ab Initio, Windows-7, Oracle 10g/9i, IBM Mainframe, SQL, IBM DB2, Korn Shell, IBM TWS Scheduler, MS-Office Tools, Flat Files, and IBM AIX, Remedy.

Confidential, Pasadena, CA

ETL Development Lead

Responsibilities:

  • Having excellent customer and business interface, involved in Customer business requirement gathering meetings and automated business validation audits to enhance the customer feed experience.
  • Developed and implemented ETL solutions for building Part D Enrolment/ Disenrollment Business reporting process and further in the Exclusion of LACARE Dual Demo Members, using ETL tool Ab Initio.
  • Graph Analysis and Code Changes for changing business needs using Ab Initio, Oracle and UNIX.
  • Extensively used PL/SQL to embed complex business logic in stored procedures.
  • Involved in performance tuning from Oracle and job automation.
  • Implemented parallel process on the existing system and reduced significant percentage of Weekly and Monthly run time.
  • Performance tuning of existing ETL process.

Technical platforms used includes: Ab Initio GDE, DB2, SQL Server, and UNIX AIX, Shell scripting, Mainframe, IBM TWS Scheduler, and BMC Remedy.

Confidential, Pasadena, CA

ETL Development Lead

Responsibilities:

  • Requirements gathering and discussions with Clients and Business Analyst to understand the Business Logic.
  • Involved in effort estimation and project planning.
  • Involved in Data Modeling, design and Development of complex graphs.
  • Developed and implemented ETL solutions for building Medicare Letter Optimization (MLO) Process using ETL tool Ab Initio.
  • Having excellent customer and business interface, involved in Customer business requirement gathering meetings and automated business validation audits to enhance the customer feed experience.
  • Prepared Unit Test Case document and perform Unit testing.
  • Effort Estimations for Change Requests/ new enhancements of ETL processes.
  • Performance tuning of existing ETL process.
  • Carried out the Code Reviews and Production deployments process using KP Standards.

Technical platforms used includes: Ab Initio GDE, DB2, SQL Server, and UNIX AIX, Shell scripting, Mainframe, IBM TWS Scheduler, and Remedy.

Confidential, Jersey City, NJ

ETL Designer & Lead Business Analyst

Responsibilities:

  • Lead, designed and implemented key business rules for GIW-GLA (Global Liability Analytics) application using ETL/Ab Initio.
  • Worked as a Technical Architect to design/build SFS Data mart from scratch, which comprises of various security lending and other security assets managed by Citi.
  • Being the Offshore lead, handled 15 member team and delivered the project amidst of tough deadlines and changing requirements.
  • Prepared case study on Performance Tuning Concepts.
  • Built the data warehousing components and provided support for User Acceptance Testing and deployment of application.
  • Performance Tuning in ETL and in Database level.
  • In this project, I was involved in requirement gathering, data modelling, creating business rules mapping, interacting with Microstrategy reporting teams to implement/build the best possible Datamart consisting a number of Fact, Dimension and Summary tables based on customer business needs.
  • Worked closely with Business owner/regions, SME to capture Business requirements and translate into Technical Specifications/Functional Requirement Document (FRD)/Logical Solution Design Document (LSDD).
  • Performed Source System Analysis (SSA) to identify required source data for new business requirements to move data elements to target tables.
  • Involved in effort estimation, risk identification and project planning/scheduling.
  • Involved in Data Modeling, Design and Development of ETL processes.

Technical platforms used includes: Ab Initio GDE, Oracle/SQL Server, and UNIX AIX, Shell scripting, Mainframe, IBM TWS Scheduler, and Microstrategy.

Confidential, Foster City, CA

Lead ETL & Microstrategy Reporting Developer

Responsibilities:

  • Enhanced key ETL/business modules for Visanet Incentive Network data mart and later worked in building key business intelligence reports for Visanet Incentive Network (VIN) loyalty campaigns for merchants and member banks.
  • Created and managed lifecycle of APRIMO Sweepstakes and Campaigns.
  • Took initiative in fine-tuning performance for long running Microstrategy Reporting Queries.
  • Single handedly completed the project amidst of tough deadlines and unforeseen issues.
  • Performed Impact Analysis based on the Change request.

Confidential, Foster City, CA

Lead Application Developer

Responsibilities:

  • TAQC(Transaction Authorization based Qualification and Loyalty point Calculation) derives cardholder’s loyalty points earned based on their authorized purchases in various merchant segments, subject to currently running promotions/campaigns in commercial marketplace.
  • During this time, I have been the lead Application developer for TAQC application.
  • Developed TAQC Daemon process which takes files sent by members at any point of time and load them to Oracle tables which further used by BI applications to get data for various business scenarios related to Loyalty transaction authorization and calculating the loyalty points which cardholders can further redeem towards various purpose in market.
  • Performed Impact Analysis based on the Change request.
  • Involved in creating UNIX shell scripts.
  • Implemented Ab Initio configuration and set up well-tuned environments for Production, Development and Testing.
  • Scheduled the graphs using Job Scheduler and loaded the data into target tables from staging area by using SQL Loader.
  • Involved in creating the Generic graph for pulling the Data from the mainframe data files using the FTP component.
  • Responsible for creating the (MFS) Multi-file, which gives the user the ability to centrally control the distributed data files and they provide the scalability and the kinds of access patterns that parallel applications require.
  • Utilized the EME for version control, and tracking of changes made to new and old graphs.
  • Solved various Issuers promo related queries within the given time frame.
  • Communicating with member banks/customers in order to understand various value added reporting requirements & translating them further into Ab Initio graphs/UNIX scripts/SQL queries to fetch/extract/prepare the report data out of the application data repository.
  • Allocating & tracking tasks/day-to-day deliverables across the team & assist members to execute them timely.

Major technologies used includes: Ab Initio, mainframe, DB2, Oracle, Unix Shell scripting etc.

Confidential, Foster City, CA

Development/Support Lead

Responsibilities:

  • Development lead for VISA point of sale (POS) reporting application, which captures all point of sale transactions processed through VisaNet, in order to provide value added reporting service to member banks and merchants enabling them to get insight of their credit card based transactions for various categories, which further gives them capability to make business decisions to grow their market segment.
  • Created various new member extracts and reports to solve member inquiries and data requests, utilizing the BI framework based on ETL tool Ab Initio, Legacy mainframe feeds, DB2, PL/SQL/Oracle, Unix Shell scripting etc.
  • Provided 24x7 production support for production issues/incidents/change/service requests.
  • Interacted with the client to review new enhancements, providing estimates for executing those enhancements and finally communicating with the offshore team to carry them out.
  • Developed a number of programs for creating reporting metrics & automated alerts for application residing in Mainframe data storage.
  • Analysis of the adhoc requests, code Walk through, enhancement requests review and analysis of the RTN, coding and identification of the impacted modules. Modification of the existing JCLs and COBOL codes based on design document, implementation of version control using ENDEVOR, preparation of test data.

Confidential, Foster City, CA

Development/Support Lead

Responsibilities:

  • Lead Application developer for Authorization Profile & Fraud Reporting application, which provides VISA member banks a summarized statistics of authorization transactions processed by VISANET & member processors depending on various business scenarios and highlights potential fraud patterns and risk scores generated by various business scenario related to fraud/risk.
  • Carried out complex transaction analysis based on complex new patterns of authorization and enhanced business rules further to accommodate the same, adding value to Member reports.
  • Provided 24x7 production support for production issues/incidents/change/service requests.
  • Analysis of the adhoc requests, code Walk through, enhancement requests review and analysis of the RTN, coding and identification of the impacted modules. Modification of the existing JCLs and COBOL codes based on design document, implementation of version control using ENDEVOR, preparation of test data.
  • Interacted with the client to review new enhancements, providing estimates for executing those enhancements and finally communicating with the team to carry them out.

Major technologies used includes: IBM Mainframe, JCL, COBOL, DB2, FILEAID, VSAM, and Easytrieve, Control M scheduler, Endeavor version control tool and MVS operating system.

Confidential, Foster City, CA

Mainframe Developer/Operational Support

Responsibilities:

  • Successfully carried out the day to day user account maintenance, resolving user’s report access issues, and report delivery issues and involved in adding new types of delivery medium, upgrading delivery procedure in the current delivery media for the View Direct Reporting Interface for VISA user community.
  • View Direct is a system through which the Reports produced by various VISA application groups are loaded on Mainframe and VISA internal users interface with Mainframe via the Server through TCP/IP connection which gives them a Window-based image of the reports on their PC.
  • Provided 24x7 production support for production issues/incidents/change/service requests.
  • Since a huge number of members subscribe to View Direct, there are always queries regarding subscriptions, non-receipt of reports etc. There is a separate Mailbox for View Direct to which the queries are directed. Resolving these queries requires a thorough knowledge of the application, sound logical reasoning, and effective communication skills.
  • Developed a number of programs for creating reporting metrics & automated alerts for application residing in Mainframe data storage.

Major technologies used includes: IBM Mainframe, JCL, COBOL, DB2, VSAM, Easytrieve, Control M scheduler, Endevor version control tool, and MVS operating system.

Confidential, Foster City, CA

Mainframe Developer/Operational Support

Responsibilities:

  • Developed several new processes comprising of Mainframe COBOL/DB2/JCL to reflect various new categories of alternations made in Cardholder Data Base (CDB) which gets published regularly in Card Recovery Bulletin (CRB). These reports are valuable to VISA member banks & merchants to update their databases for authenticating financial transactions across the Globe.
  • Interacted with Visa account managers to collect new requirements, estimated efforts for RTN, communicating with team to timely deliver customer needed data feeds and developed new reports utilizing Mainframe legacy suite.
  • Provided 24x7 production support for production issues/incidents/change/service requests.
  • Performed Analysis of the adhoc requests, code Walk through, enhancement requests review and analysis of the RTN, coding and identification of the impacted modules. Modification of the existing JCLs and COBOL codes based on design document, implementation of version control using ENDEVOR, preparation of test data.

Major technologies used includes: IBM Mainframe, JCL, COBOL, DB2, VSAM, Easytrieve, Control M scheduler, Endevor version control tool, and MVS operating system.

We'd love your feedback!