Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. HPCache: Memory-Efficient OLAP Through Proportional Caching
 
conference paper

HPCache: Memory-Efficient OLAP Through Proportional Caching

Nicholson, Hamish
•
Chrysogelos, Periklis
•
Ailamaki, Anastasia
2022
DaMoN'22: Data Management on New Hardware
SIGMOD/PODS '22: International Conference on Management of Dataent on New Hardware

Analytical engines rely on in-memory caching to avoid disk accesses and provide timely responses by keeping the most frequently accessed data in memory. Purely frequency- & time-based caching decisions, however, are a proxy of the expected query execution speedup only when disk accesses are significantly slower than in-memory query processing. On the other hand, fast storage offers loading times that approach or even outperform fully in-memory query execution response times, rendering purely frequency-based statistics incapable of capturing impact of a caching decision on query execution. For example, caching the input of a frequent query that spends most of its time processing joins is less beneficial than caching a page for a slightly less frequent but scan-heavy query. As a result, existing caching policies waste valuable memory space to cache input data that offer little-to-no acceleration for analytics. This paper proposes HPCache, a buffer management policy that enables fast analytics on high-bandwidth storage by efficiently using the available in-memory space. HPCache caches data based on their speedup potential instead of relying on frequency-based statistics. We show that, with fast storage, the benefit of in-memory caching varies significantly across queries; therefore, we quantify the efficiency of caching decisions and formulate an optimization problem. We implement HPCache in Proteus and show that i) estimating speedup potential improves memory space utilization, and ii) simple runtime statistics suffice to infer speedup expectations. We show that HPCache achieves up to 12% faster query execution over state-of-the-art caching policies, or 75% less in-memory cache footprint without deteriorating query performance. Overall, HPCache enables efficient use of the in-memory space for input caching in the presence of fast storage, without any requirement for workload predictions.

  • Details
  • Metrics
Type
conference paper
DOI
10.1145/3533737.3535100
Author(s)
Nicholson, Hamish
Chrysogelos, Periklis
Ailamaki, Anastasia
Date Issued

2022

Publisher

ACM

Publisher place

New York

Published in
DaMoN'22: Data Management on New Hardware
ISBN of the book

978-1450393-78-2

Start page

1

End page

9

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
DIAS  
Event nameEvent placeEvent date
SIGMOD/PODS '22: International Conference on Management of Dataent on New Hardware

Philadelphia PA, USA

June 13, 2022

Available on Infoscience
September 9, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/190599
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés