The LRU-K page replacement algorithm for database disk buffering
SIGMOD '93 Proceedings of the 1993 ACM SIGMOD international conference on Management of data
Answering queries using views (extended abstract)
PODS '95 Proceedings of the fourteenth ACM SIGACT-SIGMOD-SIGART symposium on Principles of database systems
Optimizing queries using materialized views: a practical, scalable solution
SIGMOD '01 Proceedings of the 2001 ACM SIGMOD international conference on Management of data
Mid-tier caching: the TimesTen approach
Proceedings of the 2002 ACM SIGMOD international conference on Management of data
Web caching for database applications with Oracle Web Cache
Proceedings of the 2002 ACM SIGMOD international conference on Management of data
VLDB '98 Proceedings of the 24rd International Conference on Very Large Data Bases
A predicate-based caching scheme for client-server database architectures
The VLDB Journal — The International Journal on Very Large Data Bases
MTCache: Transparent Mid-Tier Database Caching in SQL Server
ICDE '04 Proceedings of the 20th International Conference on Data Engineering
Cache tables: paving the way for an adaptive database cache
VLDB '03 Proceedings of the 29th international conference on Very large data bases - Volume 29
Value complete, column complete, predicate complete
The VLDB Journal — The International Journal on Very Large Data Bases
A middleware-based approach to database caching
ADBIS'06 Proceedings of the 10th East European conference on Advances in Databases and Information Systems
Concurrency and replica control for constraint-based database caching
ADBIS'10 Proceedings of the 14th east European conference on Advances in databases and information systems
Hi-index | 0.00 |
Caching data reduces user-perceived latency and often enhances availability in case of server crashes or network failures. DB caching aims at local processing of declarative queries in a DBMS-managed cache close to the application. Query evaluation must produce the same results as if done at the remote database backend, which implies that all data records needed to process such a query must be present and controlled by the cache, i. e., to achieve "predicate-specific" loading and unloading of such record sets. Hence, cache maintenance must be based on cache constraints such that "predicate completeness" of the caching units currently present can be guaranteed at any point in time. We explore how cache groups can be maintained to provide the data currently needed. Moreover, we design and optimize loading and unloading algorithms for sets of records keeping the caching units complete, before we empirically identify the costs involved in cache maintenance.