Evolution of Buffer Management in Database Systems: From Classical Algorithms to Machine Learning and Disaggregated Memory
By: Prudhvi Gadupudi, Suman Saha
Potential Business Impact:
Makes computers remember information faster.
Buffer management remains a critical component of database and operating system performance, serving as the primary mechanism for bridging the persistent latency gap between CPU processing speeds and storage access times. This paper provides a comprehensive survey of buffer management evolution spanning four decades of research. We systematically analyze the progression from foundational algorithms like LRU-K, 2Q, LIRS, and ARC to contemporary machine learning-augmented policies and disaggregated memory architectures. Our survey examines the historical OS-DBMS architectural divergence, production system implementations in PostgreSQL, Oracle, and Linux, and emerging trends including eBPF-based kernel extensibility, NVM-aware tiering strategies, and RDMA-enabled memory disaggregation. Through analysis of over 50 seminal papers from leading conferences (SIGMOD, VLDB, OSDI, FAST), we identify key architectural patterns, performance trade-offs, and open research challenges. We conclude by outlining a research direction that integrates machine learning with kernel extensibility mechanisms to enable adaptive, cross-layer buffer management for heterogeneous memory hierarchies in modern database systems.
Similar Papers
Survey of Disaggregated Memory: Cross-layer Technique Insights for Next-Generation Datacenters
Distributed, Parallel, and Cluster Computing
Lets computers share memory to work faster.
The Future of Memory: Limits and Opportunities
Hardware Architecture
Makes computers faster by putting memory closer.
Dynamic Dual Buffer with Divide-and-Conquer Strategy for Online Continual Learning
Machine Learning (CS)
Keeps computer learning new things without forgetting.