4 minute read · October 9, 2025

The Value of Dremio’s End-to-End to Caching

Will Martin

Will Martin · Technical Evangelist

Caching dramatically reduces latency and computational costs by storing frequently accessed data closer to where it's needed. Instead of repeated expensive operations - such as fetching from object storage, planning complex queries, or executing SQL - the data you need is provided in fast, local memory.

To deliver on this, Dremio implements different layers of caching throughout the analytics workflow. There’s a query plan cache for speeding up query compilation, a results cache for reusing answers to identical queries, and the columnar cloud cache, which stores frequently accessed data right on the compute nodes. Read on to learn more about each of these mechanisms and how they improve performance speeds and reduce costs.

Query Plan Cache

When running a SELECT query, the query must be planned before it is executed. This planning is split into two stages:

  • Logical: the order of operations for the query
  • Physical: the steps or actions needed to execute the query

The query plan cache works by storing the physical component of any executed plan for future reuse. When the same query is run again, rather than creating a new plan this cached plan is used instead. This improves performance of repeated queries by reducing the query process down to just execution. This is particularly useful for BI workloads, where queries are run multiple times in order to populate dashboards for different users.

Results Cache

The most straightforward type of caching - reusing results. You save time and compute by neither generating nor running a query plan. The results cache will use answers to previous queries, as long as the corresponding query plan (and user) are the same. However, if the schema or datasets change in any way, the query plans will no longer match and the cached result will be invalid. 

The results cache will automatically cache and reuse results with no configuration needed. This cache is also client-agnostic, meaning however you execute your query - in Dremio or through clients such as JDBC or Arrow Flight - the results cache will always be used to accelerate your queries.

Try Dremio’s Interactive Demo

Explore this interactive demo and see how Dremio's Intelligent Lakehouse enables Agentic AI

Columnar Cloud Cache

The columnar cloud cache (C3) is a distributed cache designed to bring your data closer to your compute. C3 works by storing frequently accessed data in-memory on the Dremio cluster nodes’ built-in NVMe/SSD storage.

Figure 1.0: Dremio's Columnar Cloud Cache.

With C3, data retrieval is much quicker than fetching from remote object storage, thanks to both being in-memory and not requiring network data transfers. By selectively caching data, either subsets or entire tables, C3 reduces S3/ADLS/GCS I/O costs by over 90%.

Conclusion

Cache is money. Leveraging highly-performant caching mechanisms throughout the analytics workflow dramatically reduces latency and computational costs. Whether repeatedly analysing the same datasets or running the same queries, Dremio's range of caching features will optimise your workflows.

Ready to Get Started?

Enable the business to accelerate AI and analytics with AI-ready data products – driven by unified data and autonomous performance.