Title: Multicache-Based%20Content%20Management%20for%20Web%20Caching
1Multicache-Based Content Management for Web
Caching
- Kai Cheng and Yahiko Kambayashi
- Graduate School of Informatics, Kyoto University
- Kyoto JAPAN
2Outline of the Presentation
- Introduction
- Why Content Management
- Contributions of Our Work
- Multicache-Based Content Management
- Content Management Scheme for LRU-SP
- Experimental Evaluation
- Concluding Remarks
31.1. Why Content Management
?
?
?
?
User
Network
Servers
Maximize Hit Rates (r ?/?) (or Weighted HR)
4Can Web Do Without Caching?
- Bandwidth Scarcity Weakest Part
- Unrealistic to Update All Resources
- Hot-Spot Servers
- Unpredictable of Server Overload
- Inherent Latency Light Speed ? Distance
- Even Sufficient Bandwidth and Server Capacity
- Transoceanic Data Transfer 200ms?300ms
Caching Is Necessary To Adaptively Reduce Remote
Data Requests
51.2. Why Content Management
Traditional Caching Web Caching Implications
Process Oriented Human-User Oriented User Preferences
System-Level Application-Level Semantic Information
Data Block Based Document-Based Varying Sizes, Types
Memory-Based Disk-Based Persistent Storage, Large Size,
Replacement policies based on empirical formula
are difficult to deal with these!
6Deploying Content Management
- To Support
- Larger Cache Space
- Sophisticated Control Logic
- To Support
- Sophisticated Replacement Policies With
- User-Oriented Performance Metrics
- Document Treated as Semantic Unit
71.3. Contributions of This Work
- A Multicache Architecture for Implementing
Sophisticated Content Management, Including a New
Cache Definition - A Study of Content Management for LRU-SP
- Simulations to Compare LRU-SP Against Others
8Previous Work
- Classifications in Approximate Implementations of
Complicated Caching Schemes - LRV, LNC-W3-U, etc.
- Segmentation in Traditional Caching As Tradeoffs
Between Performance and Complexity - Segmented FIFO, FBR, 2Q etc.
- Disadvantages
- Both Are Built-in Ad hoc Implementation, Rather
than An Independent Mechanism - Can Not Support Sophisticated Category nor
Semantic-Based Classification
9Managing LFU Contents in Multiple Priority Queues
10Cache Components
- Space
- Limit Storage Space
- Contents
- Objects Selected for Caching
- Policies
- Replacement Policies
- Constraints
- Special Conditions
Space
Space
Constraints
Contents
Policies
11Constraints for Cache
- Admission Constraints
- Define Conditions for Objects Eligible For
Caching - e.g. (size lt 2MB) !(Source local)
- Freshness Constraints
- Define Conditions for Objects Fresh Enough For
Re-Use - e.g. (Type news) (Last-Modified lt 1week)
- Miscellaneous Constraints
- e.g. (Time end-of-day)? (Total-Sizelt
95Cache-Size)
12Multicache Architecture
Web Cache With Multiple Subcaches
IN-CACHE
CONSTRAINTS
CENTRAL ROUTER
Client
Web Servers
Request/Response
CKB
SUBCACHE
SUBCACHE
SUBCACHE
JUDGE
13Components of the Architecture
- Central Router
- Control and Mediate the Cache
- Cache Knowledge Base (CKB)
- A Set of Rule Based To Allocate Objects
- R1. Allocate(X, 1)-url(X, U), match(U,
.jp),content(X, baseball) - Subcaches
- Cache for Keeping Objects With Special Properties
- Cache Judge
- Make Final Decisions From A Set of Eviction
Candidates
14The Procedural Description
- Central Router services each request. Suppose
current request is for document p - Locating p by In-cache Index
- If p is not in cache, download p
- Validate Constraints, if false, loop
- Fire rules in CKB, let subcache ID K
- While no enough space in subcache K for p
- Subcache K selects an eviction
- If space sharing, other subcaches do same
- Judge assesses the eviction candidates
- Purge the victim
- Cache p in subcache K
- If p is in subcache , do i) - iv) re-cache p.
15Content Management for LRU-SP
- LRU (Least Recently Used)
- Primarily Designed for Equal Sized Objects, and
Only Recency of Reference In Use - Extended LRUs
- Size-Adjusted LRU (SzLRU)
- Segmented LRU (SgLRU)
- LRU-SP(Size-Adjusted and Popularity-Aware LRU)
- Make SzLRU Aware of Popularity Degree
16Probability of Re-ReferenceAs a Function of
Current Reference Times
17Cost To-Size Ratio Model
- An Object A In Cache Saves Cost
- nref (1/atime)
- nref is the frequency of reference
- atime is the time since last access, (1/atime) is
the dynamic frequency of A - When Put In Cache, It Takes Up Space size
- Cost-to-size ratio nref /(sizeatime)
- The Object With Least Ratio Is Least Beneficial
One
18Content Management of LRU-SP
- CKB Rule
- Allocate(X, log(size/nref))-Size(X, size),
Freq(X, nref) - Subcaches
- Least Recently Used (LRU)
- Judge
- Find the One With Largest (sizeatime)/nref
- The Larger and Older and Colder, the Fast An
Object Will Be Purged
19Predicted Results
- A higher Hit Rate is expectable for LRU-SP,
because it utilizes three indicators to document
popularity. - However, higher Hit Rates are usually at the cost
of lower Byte Hit Rates, because smaller
documents contribute less to bytes of hit data.
20Experiment Results
21Explanations
- LRU-SP really obtained a much higher Hit Rate
than either SzLRU, SgLRU or LRV. - LRU-SP also obtained a higher Byte Hit Rate, when
cache space exceeds 3 of total required space. - LRU-SP only incurs O(1) time complexity in
content management. - LRU-SP a significantly improved algorithm
22Concluding Remarks
- Multicahe-Based Architecture Has Proved Ideal To
Realize Good Balance Between High Performance and
Low Overhead - It Is Capable of Incorporating Semantic
Information as Well as User Preference In Caching - It Can Work With Data Management Systems to
Support Web Information Integration