Big Data Pipeline Architect/oracle Data Integrator Lead Architect/bi Apps Lead Developer Resume
San Francisco, CA
SUMMARY:
- Highly motivated Hadoop Developer with over 13 years of experience on file distribution systems and data warehousing solutions.
- Experience in architecting highly scalable, distributed systems using open source tools.
- Experience in designing and optimizing large, multi - terabyte data warehouses.
- Good understanding on complex data processing needs of big data and experience in developing codes & modules to setup a data platform.
- Having experience in integrating state-of-the-art Big Data technologies into the overall data platform architecture.
- In-depth understanding of MapReduce and Hadoop Infrastructure
- Experience with Java, Python, Pig, Hive, Sqoop, Oozie and Hadoop Streaming
- Good understanding of the HDFS (Hadoop file distributing system)
- Configured MySQL database to store Hive metadata
- Loaded unstructured data (JSON/XML payload) into HDFS
- Experience in implementing HDFS on Amazon Web Services (AWS) Cloud platform
- Experience in loading data from HDFS and AWS S3 into Amazon Redshift
- Used spark and scala to do adhoc analysis
- Experience in various ETL tools like IBM infosphere and Oracle Data Integrator
- Expert in Oracle & PL/SQL Programming and Oracle SQL Performance tuning
- Collaborates well offshore-onsite teams and various technology groups
- Good problem solving skills
- Excellent written and oral communication skills
TECHNICAL SKILLS:
Cloud and Big Data: Amazon EMR, S3, Redshift, Hadoop, Spark, Flume, Hive, PIG and Scoop
Languages: Java, Python, SQL, and PL/SQL
ETL Tools: Oracle Data Integrator (ODI), Ascential IBM Datastage and Talend
Reporting Tools: BIRST, Tableau, Business Objects, OBIEE 10.1.3 and Cognos Reporting
ERP: PeopleSoft, Oracle E Business Suite, Oracle Fusion
Databases: MySQL, Oracle 10g/11i and MS SQL Server 2000/7.0/6.5
Tools: Toad, Sql Developer, SQL*Plus, MS Visual Test, PeopleTools & PeopleCode (PeopleSoft), Data Modeling and Quality Erwin
Build Tool: Maven
Source Control: GIT, Perforce
Operating Systems: Unix, Linux and Windows
PROFESSIONAL EXPERIENCE:
Confidential, San Francisco, CA
Big Data Pipeline Architect/Oracle Data Integrator Lead Architect/BI Apps Lead Developer
Responsibilities:
- Setup up Big data platform using Amazon web services (AWS), Hadoop with Elastic Map reduce, Hive and Scoop
- Extensively used Python/Java to parse streaming data with the XML and JSON formats
- Used Python for uploading data into Amazon S3 Buckets, Parse JSON data and to put in ReadShift.
- Involved in migrating 25 TB oracle database data into Amazon S3 (which would act as HDFS for Amazon EMR)
- Extensively used Amazon NoSQL DB’s like DynomoDB and RedShift to stored parsed data to serve Multidimensional Reporting and Cubes. Have a good understanding of HBASE as well.
- Experience in processing streaming data using RabbitMQ/Amazon Kinesis
- Used internally developed Custom ETL’s where complex data processing is required.
- Developed PIG & HIVE scripts to source data from HDFS
- Develop SQOOP scripts to import data from Oracle
- Design the Data Pipeline which can capture data from streaming web data as well as RDBMS source data
- Works closely with Project Managers, Business users, Data producer owners and Business Analysts for defining project requirements for BI Multidimensional reporting/dashboard development
- Analysis, design and development activities on multiple projects, mainly concentrating on Confidential Financials and Ecommerce & Marketing data
- Developed models to capture huge volume of streaming data, process and deliver then for end reporting
- Involved in setting up Big data platform using Amazon web services, Hadoop with Elastic Map reduce, Hive and Scoop
- Worked extensively with Flume to capture the steaming semi structured data into HDFS
- Designs the ETL data flow, develop and maintains the ETL solutions using following tools and environment, Oracle Data Integrator, PL/SQL, Talend, Unix, PostgresSQL
- Participated in data model design and technical reviews for given Confidential Usage data models.
- Developed the ETL flow for maintaining data warehouse with frequent collection of data from various source systems
- Design and develop ETL jobs with Oracle Data Integrator, to build commerce datamart (CDM) and Ecom & Marketing Data
- Enhancing the ETL jobs performance using Oracle Data Integrator with the Confidential specific requirement.
- Create dimension and fact ETL jobs to support the usage and commerce data marts.
- Provided production support for ODI ETL’s daily nightly schedules
- Creating technical design document based on function design document.
- Provided training for on EPM and its dependent components to the newly joined developers
- Designing and developing PL/SQL’s which is being used in ODI
- Involved in business user training and ongoing production support
- Works closely with technical operations team for ETL jobs. Worked closely with GoTo products engineering teams for data sources.
Environment: Oracle Data Integrator (ODI), SAP Financials, BIRST, OBIEE (reporting), Oracle, EBusiness Suite R12, Oracle EPM with FSCM integrations, Talend, My SQL, Amazon Web Services (AWS), S3, Elastic MapReduce(EMR), Hadoop, Spark, HIVE, SQOOP, RedShift, PostgresSQL and Unix
Confidential, CA
Data Management Team Manager/BI Design Architect /Data Modeler/ Principal ETL Developer
Responsibilities:
- Interact with the end users and business leads for requirement gathering to define requirement document
- Architect, design and develop ETL’s, store procedures, configuration files, tables, and views
- Installation, configuration and development for finance and supply chain (FSCM) BI Apps
- Worked extensively in bringing data from OLAP DB to between Physical Layer, Business Layer and Presentation Layer
- Establish relationships between Keys of Physical & Logical tables.
- Implement best practices to maintain good performance
- Added users and configured security in OBIEE 10.1.3.
- Train developers and other groups involved in BI development on ETL development best practices.
- Conduct reviews, done code walkthroughs & review peer code and documentations on the ETL’s.
- Build efficient processing for fact and dimension tables with complex transforms and type 1 and type 2 changes.
- Created and maintain system for capturing data for reporting, and correcting error data and generating exception reports.
- Worked with DBAs and systems personnel in migrating successful code to production.
- Provided on-call support to production system to resolve any issues.
- Requirements gathering and design of data warehouse and data marts.
Environment: Informatica, Ascential DataStage (ETL), OBIEE & Cognos (reporting), Oracle, PeopleSoft EPM 9.1 with FSCM integrations and PeopleTools 8.51
Confidential
Senior ETL Developer/Production Support lead/Deployment Engineer
Responsibilities:
- Extensively involved in the developing Datastage ETL jobs
- Designing and creating the Staging, Dimension and fact jobs with Functional and technical design documents
- Peer reviewing the newly created server jobs
- Production support of daily ETL run using Datastage
- Enhancing the ETL jobs using Datastage 7.5.2 with the Kaiser specific requirements
- Developed Ascential DataStage ETL Jobs and Data Loader definitions in the Enterprise Warehouse based on the client’s requirements.
- Setup and configured metadata including but not limited to: Record Metadata, Tablemap, Datamap, Filter, Constraint, Data Element, Assessment, KPI Definition, Calculation Definition, Calculation Rule, KPI Objects, Target Rules, Strategy, Strategy KPI, Trees, Scorecards, and Security.
- Migrating Trees, Queries, Process Definitions, and Application Designer objects between different database environments.
Environment: Ascential Datastage, Oracle, PeopleSoft EPM 9.1 (Budgeting, Workforce Planning, Workforce Scorecard, HR Warehouse), PeopleTools 8.51, Windows 2000, UNIX, Oracle 11i, nVision, Erwin.
Confidential
OBIA Lead Developer/BI Support Development Lead/Data Modeler
Responsibilities:
- Involved in New Development and Support Development on PeopleSoft EPM Budgeting/Foundation applications 9.1/9.0/8.9/8.8 SP2
- Development for Line Item, position and assert Budgeting
- Development using Component Interface, application engine and PeopleCode
- Performance tuning of slow-running AE programs
- Involved in fit-gap analysis for the clients
- Delivery of resolutions for customer reported complaints in line with the schedule.
- Preparing test cases, reviewing the code and generating other technical and installation documents.
- Performing impact analysis and come up with a solution plan that ensures less or no impacts.
- Ownership for BAM (Business analysis modeler) and ACE (Analytic Calculation Engine)-Single point contact for the customer reported complaints involving BAM & ACE.
- Reviewing the fixes provided by another developer and make sure it is the correct fix & that it works fine and to check that packaging standards/processes are followed
- Packaging the various solutions as a bundle for customer posting.
- Responsible for training existing and new resources in Budgeting.
- Responsible for Handling of Team & Client Communication.
- Follow quality process and standards defined for the project.
- Worked in Data Stage Client components Designer, Manager, Administrator and Director
- Understanding the Data Modeling, source system and the business logic
- Developed jobs based on the ETL specifications
- Prepared test documents which involved preparation of test data, test case and unit testing of the jobs
- Testing the Error Table Logic.
- Developed the Hash jobs which were used as Lookups and driving file in the transformations jobs
- Performance Tuning of the jobs to optimize job run
- Developed main jobs were in all the business logics are being implemented
- Used various stages available in the tool like sequential, hash, transformer, pivot, aggregator for developing jobs
- Developed sequencers for automation of job runs
- Taking Daily Backups of project in DSX format.
Environment: PeopleSoft EPM (Planning & Budgeting, Workforce Planning, Workforce Scorecard, HR Warehouse), Financials & SCM 8.8 (Inventory, Order Management, Purchasing, AP, AR), PeopleSoft EPM, PeopleTools 8.45, SQR 6.2, PeopleCode, App Engine, Workflow, SQL Server, Windows Prof, Ascential.
Confidential
Analytics Lead Developer/Sr Database Developer/Data Modeler
Responsibilities:
- Involved mainly in the customization of the tasks related to the GL, AR, AP, Inventory and Projects.
- Identified possible issues with the BAM Logic Server installation.
- Lead a team of developers to apply code changes to the development and test environments.
- Applied all necessary patches, fixes, and updates to the environments.
- Assisted in troubleshooting any functionality gaps with the Business Planning module, working closely with the Functional Lead.
- Created several deliverables including Technical Project Plan, Customization Effort Level Document, ETL Processes Definition, Enterprise Warehouse Outlined, End-to-End Documentation, and several others.
- Extensively worked with Application Designer to modify and create the record definitions, pages, component, menus and giving access to them by using security administrator. As a team member of project, involved in the development and customization of on-line pages and menus.
- Involved in setting up the core tables for Financials modules.
- Troubleshooting the problem areas in setting up GL tables that serve as the foundation of Financials system and in producing Setup reports.
- Played an important role in application development, organizing unit and system tests. Coordinated with users in designing specifications and functions.
Environment: PeopleSoft EPM (Budgeting, Workforce Planning, Workforce Scorecard, HR Warehouse), Financials 8.4 (GL, AP, AR, Inventory, Projects, Billing), PeopleTools 8.12, Oracle, PL/SQL, SQR, Crystal Reports, Windows NT/98.
Confidential
ERP Developer/Implementation Engineer/ Senior BI Developer
Responsibilities:
- Sql Server Installation & Configuration-such as Sort Order, Character Sets, Networks, etc.
- Database tuning for MS SQL Server and Ramco application tuning.
- Database backup scheduling online and offline.
- Created Triggers & Stored procedures.
- Managing users and resources, Database security
- Data migration from legacy systems (Text, Excel…)
- Scheduled the Database Consistency checks for checking the indexes and allocating Pages.
- Customization, Bug fixing, Implementation and support of Ramco Marshal ERP/ Ramco e.Applications 3.1c covering Sales/ logistics/ Process Production/ budgeting/ Finance/ HRM modules.
- Bug fixing and providing solution technical/functional level for various the modules. Applying of paths using ramco utilities.
- Reports and User defined Screen developments.
- Uploading Data to the live server using Visual test tool.
- Study of client requirement and project specification preparation
- Studied the existing system in the factories and fine-tuned Ramco Marshal EMS / Ramco e.Applications 3.1 for the requirements of the company.
- In the Sales module, anchored SAC/ SOM / Shipping implementations & User Training
- In the Process Production module, anchored Modeling/ planning /Scheduling/ Operations reporting and User Training
- In the finance module, anchored Accounts Payable / Accounts Receivable / Fixed Assets / General Ledger/ Cash flow /Budgeting and User Training
- Generation & customization of reports using Report Writer 3.0 and Crystal Reports 7.0 / 6.0
- Generated different reports for the modules using either of the above tools and provided necessary MIS data via the above reports
- Configuration of clients to ERP Applications
Environment: Ramco ERP, Oracle 8, Sql Server 7/2000, Applications GL, AM, AP, AR, PO Windows 2000