Sr. Data Analyst/ Bi Engineer Resume
San, RamoN
SUMMARY
- 10+ years of experience in manage, analysis, design, develop datawarehouse, datamarts, maintenance/supporting web and client server based applications/products using with Java J2EE technologies.
- 12 years of experience in Business Intelligence and Analytics design dashboards, visualizations, reports using Business Objects, Microstrategy, SAS, Tableau, SSRS and .net technologies.
- Managed design, development and implementation of Big Data initiative for Confidential to detect insurance fraud.
- Lead initatives to customized use - cases developed using expertise of claims management, SIU, Claims adjusters, lawyers, national insurance database and analysis of existing data to prevent claimant and provider fraud.
- Fluent with SDLC methodologies including Waterfall, Agile, SCRUM and hybrid Waterfall.
- Provide technical management and thought leadership, Detail oriented, strategic planner,leader and able communicator.
- Data migration from legacy application to new application and datamarts.
- Ability to translate business requirements to reporting and dashboard requirement documents.
- Provides leadership and mentoring team members and work across departments to complete projects on time.
- Technical architect in migrating from legacy SAL application to TeamConnect application, SME between technical and legal business team.
- Developed interface to in-sync claims and legal information between Claim application(ECF/CARE) and TeamConnect
- Exellent skills in requirement gathering and analysis, design, development, testing of data warehousing application
- Migrated 100+ reports from Business objects and Oracle Discoverer to Microstrategy 10.x
- Developed Business Objects universe and reports on TeamConnect Corporate instance as per business requirements
- Designed and developed oracle databases, tables, Partitioned tables, views, materialized views and created integrity constraints
- Designed Pl/SQL Procedures, packages, triggers and SQL queries for data manipulation and loading data in tables using SQL loader and other ETL tools.
- Performing data manipulation, perfomance tuning, code reviews, explain plans and oracle optimizer hints for SQL procedures and packages
- 2 years of experience with Big Data management and Hadoop related components HDFS, Map-Reduce, Pig, Hive, YARN, Sqoop, Spark, Scala, Kafka based on Big Data platforms.
- Experience in Scala and Spark SQL, Spark Streaming, Spark Core, RDD transformations, Data Frames and actions for faster analyzing and processing of data.
- Experience in build scripts using Maven and SBT and Unix Shell Scripts, Python, PowerShell Scripts.
- ETL hand on experience loading (Sqoop, Flume, Kafka, Talend, DataStage) data from different source systems and transformed data with Hive, Pig, Spark, MapReduce F/Ws then loads into different storage systems like HBase, HDFS, RDBS, MongoDB, XML, JSON, TSV, CSV, PARQUET, AVRO, SOLR.
TECHNICAL SKILLS
Tools: SAP Business Objects 4.0,XI 3.1/3.0, XI R1/R2,6.5, Web Intelligence XI 3.1/3.0/XI R2, Universe Designer XI 3.1/3.0,XI R2, Desktop Intelligence, InfoView,Microstrategy 10.x, Visual Insight,SAS,Power BI, Tableau, IBM Optim and IBM Discovery Studio
Automation Tools: Blue Prizam, Unipath and Automation Anywhere
Applications: Team Connect Legal, SIU and Corporate 3.x to 4.x, Guidewire, ECF, CARE Oracle 9/10g/11 | UNIX (HP-UX, Sun Solaris, AIX) | Microsoft Windows | Hadoop | Amazon Web Service
BigData: Hadoop Stack, HDFS, Hive, Pig, Scoop, Spark, Kubernetes, Kafka
NOSQL: Hbase, Cassandra, MongoDB
Amazon AWS Stack: Athena, Glue, Aurora, DynamoDB, Kinesis, Redshift, EMR, S3, EC2, EMR, RDS
Cloud: AWS Stack, Google Cloud, Microsoft Azure, Oracle Cloud, Internal Cloud, Salesforce
Languages: Python, Java, SQL, PL/SQL, Scripting languages, Unix Shell Scripts.
PROFESSIONAL EXPERIENCE
Confidential, San Ramon
Sr. Data Analyst/ BI Engineer
Responsibilities:
- Document manual collateral management process with business users and SME for automation
- Identified 22 data sources bottleneck and consolidate to 16 sources for the collateral process
- Worked with different file format flat files,CSV,Json and XML and ingested into SQL server using SSIS
- Day to day processing external files from Clients, using ETL some using Legacy DTS Packages
- Work on creating star schema design with fact and dimension tables for staging tables and reporting datamarts.
- Create Infromatica ETL jobs to load data from files from Lightling systems, ISView systems
- Data Quality Health Checks for ETL jobs and developed reconciliation process and generated automated reports in SSRS
- Created complex sqls from enterprise Datawherehouse (EDW) to load into SQL server via SSIS
- Converted 100+ reports from existing MS Access application to SQL Server db and creates reports on Oracle Bi and Power BI
- Create presentations decks for stake holders on current state and proposed future state
- Created BRDs, functional requirements for business and SMEs to stake holders for deliverable and scope of the project
- Worked with EDW, MDM team to integrate latest account related information
- Discover and Identify data relationships across applications and mask data and validate with automation scripts for source and target databases
- Responsible for analyzing applications to identify and protect sensitive and restricted data for non prod environments.
- Developed mapping documents with source and target information with required transframations
- Create EPIC, stories and timelines to complete BRDs with developersboth onshore and offshore
- Worked with UFT and BluePrizm RPA tools to automate data extraction from in-house applications and external websites.
- Created mocked dashboards and KPIs in Excel for executives and stakeholders
- Created dashboards and reports in Tableau and Power BI based on input from stake holders and users
- Schedule key reports for business users for treasury, front office and world markets
- Worked with Appian team to create POC for replacing current MS Access application and processes
Confidential San Francisco
Sr. Data Analyst / BI Engineer / Techincal PM
Responsibilities:
- Worked on design data migration of KYC application database from Appian Oracle to DireKT on Sql Server application.
- Data Analysis, Data Requirement and Data Mapping for ETL in SSIS
- Created mapping documents for ETL team to load data from Appian KYC Application to KYC DireKT database.
- Assisted in desiging hybrid database using flat tables and key value tables for dynamic addition of KYC questions
- Created SSIS data piplines to migrate data from CSV, flat files and Json
- Worked with Executives, Stakeholders and Analytics team for data requirements on Power BI reports and dashboards.
- Created Datapipelines for Data Hubs teams to load KYC historical data for Eagle Eye Vison for fraud analytics.
- Creating Base view, Business View, Derived view connecting multiple data sources
- Created BRDs, worked with business and SMEs to stake holders for deliverable and scope of the project.
- Created functional requirements understanding business needs for BRDs and complete BRDs with IT and stake holders
- Design analytical solutions for interfaces between various applications
- POC comparison AWS and Azure results for measure an portability of application.
- Extracted data efficiently from SQL Server to flat files for BigData migration and initial load.
- Developed S3 data lakes using Hive based partitions and stored in different formats (Parquet, Avro and ORC).
- Worked with Data Science team to integrate new datasets into existing datapipelines, build APIs data consumption.
- Monitor and test datapipelines for optimize perfomanace and apply feature transformation of ML models.
- Designed datapiplines feed from HR, AD, Cunsumer Risk Rating and NACIS information.
- Create private and public consumption layers to provide desired data for downstream applications.
- Creating and Scheduling theDenodo jobs with Web scheduler.
- Involved in designing and creating reports using Tableau, Microstrategy and Denodo.
- Created data cleanup scripts and implemented in SQL SSIS to extract and clean data and load to SQL server.
- Integration with MDM to get golden record for KYC.
- Designed document migration from Appian to OnBase and integrate KYC - DireKT application.
- Migrate documents by converting Appian documents to JASON and load data in MangoDB.
- Design documents mapping for Document Import process and work with OnBase team (Hyland Doc Mgt).
- Design workflow, Dashboard and Keyword search for documents retrival via Webtop and APIs.
- Worked on E-Discovery data requirements to collect data from different applications for litigation for attorneys.
Confidential, San Mateo
Sr. Data Analyst / Data Engineer
Responsibilities:
- Extracted large volume of data from Mainframes, oracle, MS SQL databases and logs data into Hadoop HDFS
- Developed data pipleines to migrate data using Scoop for databases and Spark and Flume for streaming data
- Architected AWS data pipeline to transfer files upon arrival and initiated Hadoop/pig process under EC2/EMR stack to process the results from S3 to downstream data lake for processing.
- Created HIVE tables as per requirements (internal/external tables) with static and dynamic partitions
- Vast Scripting UDFs in Pig and Hive scripts to load data in HBase tables
- Developed Python programs for data cleansing, parcing and formatting unstractured and Spark jobs.
- Worked with Oozie to automate jobs for Scoop, Hive, Spark and JAVA jobs
- Worked with BI team using Tableau and MicroStrategy to create reports and visualization using HIVE and Impala.
- Design ETL process to extract raw data from multiple sources and o ptimize Hive queries to ensure faster results.
- Migrate data from legacy systems to new applications and datamarts.
- Migrated databases into AWS Redshift increase ROI and reduce processing time/ licenses cost
- Create mapping documents for Data modelers and ETL teams for source and target mapping.
- Created Data warehouse with multiple Accumulating Fact tables, Snapshot tables, Dimensions and lookups
- Worked on Data Governence and standardization of Datasets
- Created and Certify Datasets using CHASE Banks transaction (Authorization, Chargeback and Incentives)
- Created L1 and L2 tables for details and summary for transactions.
- Work with offshore team in India, Singapore on data integration and cleanse data from vast range of sources.
- Assess, benchmark and initiated performance tuning for all ETL tasks and DB process
- Perform code reviews, guidelines and data validation ensuring data integriation
Confidential, San Francisco
Sr. Data Analyst / Data Architect
Responsibilities:
- Migration of First Republic Private Wealth Management documents for offices across United States. Worked with PWM project management team to gather information and created new Datamart.
- Design tables and created mapping documents for ETL team to load extracted documents metadata information.
- Work with Data Hubs teams and to load structured and non structured data with different file formats like word, PDF,Excel, CSV and other file formats are in shared drives and SharePoint.
- Created data pipeline in Python to extract and clean data and load to SQL server.
- Design documents mapping for Document Import process and work with OnBase team (Hyland Document Management System) to design workflow and Keyword search for documents.
- Moderate with Business analysts to create mock screens for UI for wealth managers.Worked with Business Analysts and project managers in identifying and defining requirement.
- Designed solution to assessed Blue Prism capabilities to identify and classify 100+ document types.
- Provide expertise in staging and deploying the Blue Prism automation solution into test and live environments.
- Facilitate meetings to demonstrate the capabilities of RPA (Blue Prism) for on-going projects and potential clients.
- Created Dashbards and MIS reports in MS SQL SSIS reporting tools.
Confidential, San Francisco, CA
Sr.Technical Lead /Data Architect /Business Analyst
Responsibilities:
- Expertise in Business Insights and Analytics, specializing in Business Intelligence, Data integration, Data Migration and Data warehouse architecture
- Earned recognition for demonstrating exceptional work performance
- Collaborate with Business users, other architects, business analysts and engineers to define requirements and design solutions
- Create mock-up dashboards and Reports using Excel and other tools.
- Co-ordinate with infrastructure team, facilitate meetings to develop infrastructure road map for project implementation.
- Take charge of drafting technical specification and mapping documents, performing source data analysis and formulating the initial load strategy.
- Create mapping documents for Data modelers and ETL teams for source and target mapping.
- Facilitate design discussions and provided architecture and design guidelines for the client team.
- Co-ordinate closely with BI teams to design and build complex multi-tab reports
- Oversee the development-quality assurance-production migration process and deployment working closely with middle-ware/infrastructure teams.
- Migrate data from legacy systems to new applications and datamarts.
- Developed comprehensive implementation design documentation
- Assess, benchmark and initiated performance tuning for all ETL tasks and DB process
- Perform code reviews, guidelines and data validation ensuring data integrity.
- Assumed responsibility in designing data integration points, ETL framework, error reprocessing, ETL.
- Standards and reusable common components.
- Actively contribute to the creation of ETL components on SAP BODS, IBM DataStage, writing DB stored procedures and complex SQL scripts and calculation views.
- Leveraged skills in creating end-to-end data integration layer for Data Marts, for initial loads, data extraction from multiple sources into DB/BW targets, parallel testing initiatives (comparing data from the old system with the system)
- Applied problem-solving skills in investigating and resolving data issues, which entailed formulating and implementing a workaround for application code or configuration deficiencies
- Execute a range of initiatives that build the foundation of the Legal team’s processes and technology capabilities targeted at creating efficiencies across the Legal Departments.
- Participate in recommending ways to improve business processes through system development and controls.
- Knowledge in eBilling, Matter Management, Contracts Management. Excellent analytical skills and comfort with multiple data sets, with the ability to draw conclusions and make informed recommendations.
- Worked Anti-fraud workbench application using Big Data Hadoop HDFS, Map-Reduce, Sqoop, Hive, Oozie, Cassandra, Zoo Keeper, MongoDB, Spark, SparkSQL, Spark Streaming, Kafka, Flume, HBase, Weka, DynamoDB.
- Experience in Scala and Spark SQL, Spark Streaming, Spark Core, RDD transformations, Data Frames and actions for faster analyzing and processing of data.
Confidential, San Francisco, CA
Technical Lead / Architect
Responsibilities:
- Successfully migrated SAL application to Team Connect application.
- Migrated millions of claims and legal related documents from Hummingbird PCDocs to EMC Documentum.
- Created TeamConnect Data warehouse using Pentaho. Integrated Business Object reporting.
- Migrated TeamConnect SIU instance to InSight application and integrated YellowFin as reporting tool.
- Developed management reports and fraud detection support.
- Support e-discovery projects providing surgical data of claim, payments and bill to Legal chief counsels for legal cases.
- Provided data feeds to Kaiser Permanente, Anthem and Bluecross. Integated Pharmacy billing .
- Developed EDI between CWCI and WCRIB organzations
- Demonstrated skills in designing common components, establishing application development processes and guidelines and utilizing IBM DataStage, IBM Information Server Console, Information Server Manager and Administrator Clients
- Expertly handled the development and design of ETL audit and control model, common components and plug-in modules
- Lead the development and deployment of large scale mission critical reporting applications
- Defined and designed systems with industry best practices meeting Data warehousing standards
- Provide feedback through design and code review of peers
- Closely collaborated with various units during various test phases and user acceptance testing
- Exemplified strong familiarity with software development, system abstraction, project process monitoring and effort distribution
- Directed medium-sized teams during ETL build and unit testing, as well as in creating solution designs, modules, mapping and technical specifications
Confidential
Sr SAP Consultant
Responsibilities:
- Provided consulting services to Confidential Bank on fine tuning ETL process using Informatica and optimized Business Object reports to run 10X faster by creating materialized views.
- Served as primary resource for all ETL process support and enhancement
- Provides workaround and recommendations on technical issues in production system
- Benchmark application performance periodically and fix performance issues
- Took accountability for ETL job change management through DataStage version control
- Expertly reviewed problem reports on ETL code, stored procedures, database administration and SQL loader/bulk load scripts
- Closely collaborated with offshore developers in resolving application code or configuration issues
Confidential, San Francisco, CA
Sr BI Consultant
Responsibilities:
- Closely collaborated with offshore developers in resolving application code or configuration issues
- Coordinate with project management on assigned projects adhering to our Project Lifecycle Methods (including Governance, Project Management, Release Management /SDLC methods) - delivering on-time, within budget.
- Contribute to the provision of a robust, reliable and effective BI and enterprise DW platform.
- Hands-on working with architects and researching new technology and develops business cases to support enterprise-wide business intelligence solutions.
- Partners closely with business partners, systems analysts, development, and architecture teams to facilitate issue resolution, develop solution designs and deliver valuable business solutions .
- Promoting best practice adoption of SAP BI functionality and skills development to maximize its use.
- Articulate and validate the technical and functional requirements to meet architectural guidelines and business requirements.
- Develop and document data and reporting requirements for the development of new or enhanced approved systems.
- Work with enterprise-wide business and IT senior management to understand customer issues and their environment in order to manage enterprise data warehouse solution.
- Solves complex technical problems and guides team on how to implement corresponding changes. Create appropriate and acceptable testing methodology as per business needs
Confidential
Data Analyst/ Programmer Analyst
Responsibilities:
- Work with Data Warehouse architects collaborate with a Business Analyst to define the data requirements and recommend a data structure.
- Responsible in leading and managing projects pertaining to the design, development, and implementation of enterprise data warehouses tool for Business Intelligence (BI) applications that will support the reporting requirements.
- Designing, developing and optimizing data warehouses, with flexible and scalable data and ETL architecture, to support business users and Business Intelligence (BI) applications.
- Analyzing business intelligence data and makes recommendations for data warehouse growth and integration an application/business function.
- Exellent knowledge of ETL tools such as Informatica, DataStage, Oracle Warehouse Builder (OWB), SAP Data Services, Data Migrator (IBI), Oracle Data Integrator (ODI) etc.
- Extensive experience with SQL and SQL scripting for relational databases such as Oracle, Microsoft SQL Server, IBM DB2, MySQL.
- Ability to communicate design solutions to both technical and non-technical audiences.
- Strong analytical and documentation skills.
- Ability to mentor others and provide technical direction on data architecture, reporting, warehousing and OLAP design.