Microsoft Fabric Updates Blog

Announcement: Eventhouse Standard Storage Billing

As part of the consumption model for Eventhouse, we are enabling Standard Storage consumption. This change is scheduled to start rolling out the week of September 16th. This means that you will start seeing billable consumption of the OneLake Storage Data Stored meter from the Eventhouse and KQL Database artifacts. For details on all things KQL Database and Eventhouse consumption check out this article.

OneLake Standard Storage used to persist and store all queryable data. When you set the retention policy, you affect this storage tier. For instance, if you need to maintain 365 days of queryable data you can set the retention to 365 days. This storage tier is comparable to the Azure ADLS (Azure Data Lake Storage) hot tier.

The amount of standard storage is controlled via your retention policy. To understand how to configure you retention policy check out this article

You can monitor how much OneLake Standard storage you are utilizing via the Microsoft Fabric Capacity Metrics App under the storage tab. Usage from KQL Database and Eventhouse will be reported via the KustoStandardStorage Operation Name. The “Billing Type” will change from “Non-billable” to “Billable”.

Fabric Capacity Metric App

Useful Links

Eventhouse/KQL Consumption

Fabric Capacities

Fabric Pricing

Fabric Capacity App

Caching Policy

Eventhouse Minimum Consumption

Postagens relacionadas em blogs

Announcement: Eventhouse Standard Storage Billing

janeiro 20, 2026 de Xu Jiang

The exchange of real-time data across different data platforms is becoming increasingly popular. The Cribl source (preview) is now available in Real-Time Intelligence, allowing real-time data to flow into Fabric RTI Eventstream through our collaboration with Cribl, enabling you to take full advantage of Fabric Real-Time Intelligence’s robust analytics tools for their real-time needs. Collaborating to broaden … Continue reading “Expanding Real-Time Intelligence data sources with Cribl source (Preview)”

janeiro 20, 2026 de Ye Xu

Copy job is the go-to solution in Microsoft Fabric Data Factory for simplified data movement, whether you’re moving data across clouds, from on-premises systems, or between services. With native support for multiple delivery styles, including bulk copy, incremental copy, and change data capture (CDC) replication, Copy job offers the flexibility to handle a wide range … Continue reading “Simplifying data movement across multiple clouds with Copy job – Enhancements on incremental copy and change data capture”