BackBridge Decommission Post Implementation Overview - PowerPoint PPT Presentation

1 / 17
About This Presentation
Title:

BackBridge Decommission Post Implementation Overview

Description:

When COMET was implemented its underlying database, CDB, was intended ... Forte Migration Project. Year End Processing. Ongoing Production Support. Performance ... – PowerPoint PPT presentation

Number of Views:92
Avg rating:3.0/5.0
Slides: 18
Provided by: Akaw6
Category:

less

Transcript and Presenter's Notes

Title: BackBridge Decommission Post Implementation Overview


1
BackBridge Decommission Post Implementation
Overview
  • May 4th, 2006

2
Agenda
  • Project Background
  • Technical Overview
  • Scope and Objectives
  • Technical Challenges
  • Overcoming Technical Challenges
  • Solution Architecture
  • Implementation Approach
  • Performance Summary

3
BackBridge Decommission (BBD) Background
  • When COMET was implemented its underlying
    database, CDB, was intended to be the system of
    record for member and employer data
  • Legacy mainframe applications needing these data
    could not access CDB
  • The enterprise devised a solution that would
    apply CDB changes to MBR, EPR, and MBR Address
    databases through a nightly batch process called
    BackBridge
  • MBR, EPR, and MBR Address are VSAM files
    (key-sequenced indexed sequential)
  • The Backbridge Decommission (BBD) enables legacy
    programs to retrieve data directly from CDB
  • This effort entailed a proof-of-concept (POC)
    that proved the approach. The POC was followed
    by the implementation
  • Programs that called MBR, EPR, and MBR ADDR now
    call CDB

4
Technical Overview
5
BBD Scope and Objectives
  • Establish a sound, tested infrastructure that
    supports applications that rely on member and
    employer data
  • Enable legacy applications to access CDB in
    real-time
  • Decommission the BackBridge infrastructure
  • Decommission MBR, EPR, and MBR Address databases
  • Meet the November-December 2005 deployment
    timeframe
  • Deploy in a manner that is transparent to the
    business areas
  • Provide a safety-net strategy that will allow
    for seamless rollback
  • Implement a maintenance plan for ongoing support

6
Technical Challenges
  • More that 200 batch and online COBOL and Natural
    programs that access VSAM databases
  • Performance
  • Network vs. resident data source
  • Relational database vs. native flat file
  • Sequential and skip-sequential reads
  • Online requests lt 1 second response time
  • Existing batch programs require multiple hours to
    run
  • Ensuring 24/7 access to the database and
    maintainability

7
Solution Architecture
8
Physical Architecture
  • z/OS 1.5
  • COBOL and Natural
  • Common Modules (Assembler)
  • Common Modules (COBOL)
  • SoftwareAG EntireX Broker and RPC Client
  • Windows 2000 SP4 VMWare, 2.2Ghz CPU X 4, 1Gb
    RAM
  • SoftwareAG EntireX RPCServer implemented in Java
  • 4 Online Servers
  • 1 Batch Server
  • JDBC thin driver
  • HPUX 11i 800 MB CPU X 8, 12 Gb RAM
  • ORACLE 9.2.0.5
  • PL/SQL Packaged Procedures
  • 1 Gigabit Network TCP/IP

9
Overcoming Technical Challenges
  • Leverage common modules
  • Re-factor existing PL/SQL
  • Connection pooling (configurable)
  • Pre-fetch and multithreading
  • Read ahead
  • Data caching
  • Splitting large programs
  • Monitoring
  • Ping program

10
High-Level Schedule
11
Status and Accomplishments
  • Proof-of-Concept (two - three months)
  • Built and validated middleware solution for
    Legacy online and batch programs
  • Implemented error management strategy
  • Demonstrated load balancing
  • Accommodated problematic programs
    (skip/sequential)
  • Proved the solution can handle production volumes
  • Less than 30 increase in completion times for
    production batch processing
  • Negligible difference in production online
    response times
  • Implementation (six months)
  • Acquired team commitments and management support
  • Established development and test environments
  • Conducted comprehensive testing
  • Refined infrastructure
  • Minimized modifications to existing programs
  • Deployed ahead of schedule
  • Delivered under budget

12
Accomplishments (Cont.)
  • On November 11th, the Legacy Enrollment
    Database and its associated "backbridge" process
    were decommissioned. For those of you who have
    lived the pain of reconciling data discrepancies
    between COMET and the legacy systems, the
    significance of this achievement is huge! The
    solution went into production seven weeks ahead
    of schedule and well under budget (the budget was
    originally estimated at 4-5 million, but
    delivered for only 1.2 million) and the ongoing
    annual savings to CalPERS is estimated at 500,00
    - 700,000. However, the greatest benefit to
    this project is that we've eliminated a major
    source of our data integrity and redundancy
    problems.
  • - Fred Buenrostro, CEO CalPERS

13
Other Challenges and Risks
  • Ongoing Enterprise Obligations
  • COMET October Release
  • R Street Project
  • Forte Migration Project
  • Year End Processing
  • Ongoing Production Support
  • Performance
  • Sufficient Testing
  • Annual Programs

14
Performance Test Results
  • Summary
  • Batch performance was 1.2 to 3 times longer in
    the new infrastructure
  • Although performance times for selected batch
    programs were as high as 3 times, the overall
    impact did not translate into a similar increase
    in total batch processing time.
  • Programs that access MBR/EPR data represent only
    a percentage, roughly 30, of total batch
    processing time.
  • Performance times for selected batch production
    job processes, consisting of multiple job steps
    and programs, were only 1 to 1.4 times
    longer--well within the acceptable threshold of 2
    times.
  • Change in online performance was unnoticeable.
  • All legacy programs including problematic
    programs that perform sequential reads of can be
    accommodated in the new infrastructure.
  • No noticeable degradation in SmartDesk
    performance after change to access CDB for
    MBR/EPR data.
  • Conclusions
  • Is a technically feasible solution suitable for
    production application
  • Can accommodate legacy programs that perform both
    direct and skip/sequential database accesses with
    minimal or no changes
  • Is scalable to handle large volumes of database
    calls under actual production conditions
  • Can be configured and tuned to achieve similar
    performance results in production as during
    testing

15
Implementation Results
  • Deployed on time and on budget
  • Stable environment
  • No production downtime
  • Supports online and batch without issue
  • Decommissioned old infrastructure
  • Enhancements included 11 annual programs

16
Questions?
17
Appendix POC Test Results
Performance Factor is a comparison of
performance between the new infrastructure (CDB
source) vs. the current backbridge infrastructure
(mainframe MBR/EPR database). For example, a
factor of 1.2 means that the new infrastructure
performs 1.2 times longer than currently.
Write a Comment
User Comments (0)
About PowerShow.com