Browse Prior Art Database

Cache Management using SLA Policies and QoS Criteria

IP.com Disclosure Number: IPCOM000015991D
Original Publication Date: 2002-Sep-15
Included in the Prior Art Database: 2003-Jun-21
Document File: 1 page(s) / 40K

Publishing Venue

IBM

Abstract

Cache management techniques for web caches and host memory attempt to increase the benefit of cached content by maximizing the hit rate of object requests. Traditional methods for achieving better hit rates have been to use a Least Recently Used or Least Frequently Used algorithm for determining which pages to evict from the cache. These techniques consider cache entries from all services independently when choosing entries to evict from the cache.

This text was extracted from a PDF file.
This is the abbreviated version, containing approximately 52% of the total text.

Page 1 of 1

Cache Management using SLA Policies and QoS Criteria

Cache management techniques for web caches and host memory attempt to increase the benefit of cached content by maximizing the hit rate of object requests. Traditional methods for achieving better hit rates have been to use a Least Recently Used or Least Frequently Used algorithm for determining which pages to evict from the cache. These techniques consider cache entries from all services independently when choosing entries to evict from the cache.

We propose a cache management enhancement that includes the notion of multiple services hosted on a single server. Using a web server as an example, the cache management system can be improved in two distinct areas to provide an increased ability to meet the demands of multiple customers.

The first improvement is to separate the partitioning of the cache into services, such that each service is guaranteed enough cache memory to meet the minimum requirements of the Service Level Agreement in the cache. Using customer response time as a metric for service level agreements, it is possible to show that the amount of memory allocated to a particular service affects the cache hit/miss rates for the request streams of a given service. The cache miss rate directly affects the number of requests that flow from the cache to the storage system (and increase the time necessary to generate a response). By adjusting the cache memory allocated to a particular service it is possible to control the average response time of client requests (note that this can be performed dynamically based on changing arrival rates and also depends on the popularity distribution characteristics of a given web request stream). After the memory allocation is performed to meet the minimum requirements of the SLA's for all hosted services, additional memory allocation decisions can be made based on the perceived benefit in terms of response time for each service (again, this decision can be performed bas...