Featured Articles
Popular Articles
-
Product Insights from the Dremio Blog
Apache Iceberg Table Performance Management with Dremio’s OPTIMIZE
-
Product Insights from the Dremio Blog
Minimizing Iceberg Table Management with Smart Writing
-
Product Insights from the Dremio Blog
Apache Iceberg Table Storage Management with Dremio’s VACUUM TABLE
-
Dremio Blog: Various Insights
Handling Complex Data Types in Dremio
Browse All Blog Articles
-
Product Insights from the Dremio Blog
Apache Iceberg Table Performance Management with Dremio’s OPTIMIZE
Performance management for Apache Iceberg tables isn’t just about cleaning up small files, it’s about ensuring your data layout evolves in step with your ingestion patterns and query workloads. Dremio’s OPTIMIZE command provides the precision engineers need: merging, splitting, and reclustering data into efficient layouts while keeping metadata lean. With its flexible parameters, you can tailor compaction jobs to strike the right balance between optimization depth, runtime, and cost. At the same time, Dremio’s auto-optimization features mean you don’t always have to run these jobs manually. By letting Dremio continuously monitor and optimize Iceberg tables in the background, your most critical datasets stay query-ready without the overhead of constant maintenance. -
Product Insights from the Dremio Blog
Minimizing Iceberg Table Management with Smart Writing
The real secret to minimizing Iceberg table maintenance isn’t running more optimization jobs, it’s writing smarter data from the very beginning. By combining batch and streaming ingestion best practices, designing thoughtful partitioning and clustering strategies, tuning table properties, and monitoring file health, you can dramatically reduce the frequency and cost of downstream operations like OPTIMIZE. -
Product Insights from the Dremio Blog
Apache Iceberg Table Storage Management with Dremio’s VACUUM TABLE
Apache Iceberg’s snapshot model is a game-changer for time travel, auditing, and recovery, but it comes with a responsibility: old data must be managed carefully. Without proactive cleanup, tables can accumulate unnecessary files, driving up storage costs, slowing queries, and even creating compliance risks. Dremio’s VACUUM TABLE command provides the control data engineers and architects need to: Expire outdated snapshots, keeping only the versions that align with retention policies. Permanently remove deleted data to meet GDPR and CCPA requirements. Clean up orphan files to ensure storage remains lean and predictable. -
Dremio Blog: Various Insights
Handling Complex Data Types in Dremio
Overview Dremio provides out-of-the-box methods of handling complex data types in, for example JSON and parquet datasets. Common characteristics are embedded “columns within columns” and “rows within columns”. In this blog, we will demonstrate how Dremio can discover and handle these types of data. The examples have been tested on the following Dremio versions: Preparation […] -
Dremio Blog: Various Insights
Why Agentic AI Needs a Data Lakehouse
Agentic AI is an artificial intelligence system that is designed to operate autonomously. With minimal human supervision it can be expected to make decisions and perform tasks with specifically trained agents. This is thanks in large part to Large Language Models (LLMs) which provide agentic AI with enhanced reasoning and the ability to understand context. […] -
Product Insights from the Dremio Blog
Using Dremio’s MCP Server with Agentic AI Frameworks
This is exactly where MCP and A2A come together. MCP ensures that agents can securely interact with enterprise tools like Dremio, accessing trusted data through well-defined interfaces. A2A, in turn, provides the framework for those agents to collaborate, delegating tasks, exchanging results, and orchestrating end-to-end workflows. -
Product Insights from the Dremio Blog
Data Regulations in Food & Agriculture Supply Chains and Dremio’s Lakehouse Solution
Regulatory change in the food and agriculture supply chain is no longer about periodic paperwork, it’s about maintaining continuous, trustworthy, and accessible data. Whether it’s proving product lineage within 24 hours for the FDA, reporting Scope 3 emissions under the EU’s CSRD, or ensuring supplier compliance through due diligence laws, the common denominator is data complexity. -
Dremio Blog: Various Insights
Why Education Companies Need Secure Data Platforms: Navigating Privacy Regulations and How Dremio Helps
As education becomes increasingly data-driven, the stakes for protecting sensitive information have never been higher. Regulations like FERPA, COPPA, GDPR, and state-level privacy laws demand rigorous compliance, while rising cyber threats highlight the urgent need for robust security and governance. At the same time, educators and edtech companies cannot afford to sacrifice innovation, students expect personalized learning, administrators need real-time insights, and institutions are exploring AI-driven opportunities to improve outcomes. -
Product Insights from the Dremio Blog
Why Dremio is the Ideal Secure Data Platform for Transportation & Automotive Companies
The transportation and automotive industry sits at a unique inflection point: vehicles are now data platforms as much as they are machines. With each car, truck, or bus generating gigabytes of telemetry, location, and sensor data every hour, the potential for innovation is enormous. Yet the regulatory landscape, from U.S. privacy laws to EU mandates on data portability and global cybersecurity standards, makes it clear that this data must be handled with care. Non-compliance isn’t just a legal risk; it undermines consumer trust and can derail new business models before they gain traction. -
Product Insights from the Dremio Blog
Why Dremio is an Ideal Data Platform for Telecom Companies: Navigating Data Regulations and Security
Telecom companies cannot afford to rely on platforms that “might” meet regulatory standards , they need proven, auditable compliance. Dremio is designed with security and certification at its core, giving providers confidence that their data practices align with industry and legal obligations. -
Product Insights from the Dremio Blog
From Grid to Insight: Building a Compliant, Secure Lakehouse for Energy & Utilities with Dremio
The Energy and Utilities industry faces some of the most demanding regulatory and operational pressures in the world. From NERC CIP requirements in the power grid, to TSA directives in pipelines, to the EU’s NIS2 obligations, the stakes are high: protect sensitive data, demonstrate compliance, and maintain uninterrupted service to millions of customers. -
Product Insights from the Dremio Blog
Navigating Finance and Insurance Data Regulations with Dremio’s Intelligent Lakehouse
For financial and insurance companies, compliance is not just a legal requirement, it’s a strategic imperative. The overlapping patchwork of regulations, from NAIC Model Laws and HIPAA to Basel III, MiFID II, GDPR, and PCI DSS, makes it clear that data governance and security must be woven into the very fabric of an organization’s data strategy. At the same time, customers expect real-time experiences, transparent policies, and the assurance that their most sensitive information is being handled with care. -
Product Insights from the Dremio Blog
Building a Secure Healthcare Data Platform: Why Dremio is the Right Choice
Healthcare organizations face a unique challenge: they must manage some of the world’s most sensitive data while also unlocking its potential to improve patient outcomes, accelerate research, and streamline operations. Regulations like HIPAA, HITECH, GDPR, and the 21st Century Cures Act ensure accountability, but they also raise the bar for data security, governance, and interoperability. -
Product Insights from the Dremio Blog
Governance Without Friction: How Dremio’s Semantic Layer Keeps AI Agents Accurate and Secure
AI without governance is a liability. Governance without speed is a bottleneck. Dremio’s semantic layer eliminates this trade-off by providing governance without friction. With a single, governed interface exposed to AI agents via the MCP server, organizations can unlock AI-powered analytics that are fast, secure, and business-friendly. -
Product Insights from the Dremio Blog
Autonomous Reflections and Agentic AI: Why Sub-Second Responses Matter in the Lakehouse
In the AI era, speed is the differentiator. Dremio’s autonomous reflections make it possible to achieve sub-second responses without endless tuning or fragile performance layers.
- 1
- 2
- 3
- …
- 33
- Next Page »