December 10, 2025
Scaling Peta-byte Scale Distributed Writes to Apache Iceberg
Apache Iceberg has revolutionized data lake architectures with its promise of ACID transactions and transactional consistency. However, beneath its elegant facade lies a fundamental scaling challenge: the atomic commit mechanism that ensures data integrity becomes a bottleneck under high-concurrency workloads. This session explores the real-world difficulties of distributed Iceberg writes, demonstrates how seemingly simple operations can lead to cascading failures at scale, and presents proven architectural solutions to unlock massive write throughput while preserving consistency guarantees.
We’ll begin by examining Iceberg’s commitTable API and its strict atomic requirements, showing how only one metadata update can succeed at a time. Through concrete examples of concurrent write scenarios, we’ll demonstrate how this design leads to commit conflicts, writer starvation, and ultimately dropped transactions when multiple distributed processes attempt simultaneous updates. Using real-world scenarios from high-frequency data ingestion pipelines, we’ll illustrate how these challenges manifest in production environments and why traditional retry mechanisms often fail to provide adequate relief.
Topics Covered
Sign up to watch all Subsurface 2025 sessions