Microsoft Fabric Updates Blog

Faster data processing with Native Execution Engines in Microsoft Fabric Runtime 1.3: efficient memory management explained

Accelerating data processing at scale demands far more than just raw compute power. As the complexity and volume of data workloads grow, every layer of the execution pipeline—especially memory management—plays a critical role in determining overall performance. We’ve introduced unique changes how Spark handles memory with the Native Execution Engine for Fabric Spark, reducing performance bottlenecks and making large-scale data analytics dramatically more efficient. 

Spark relies heavily on the Java Virtual Machine (JVM) to manage memory on-heap. Any memory that is allocated and used from this on-heap memory is managed by the JVM’s garbage collector (GC). Now, imagine a more direct and agile memory management model—one that bypasses the JVM’s overhead while still utilizing existing Spark memory settings and tracking. That’s exactly what Native Execution Engine for Fabric Spark delivers. We’ve bundled a dynamic memory manager that borrows memory from the on-heap quota allocated to Spark and tracks the total memory usage by Spark and native alike, to intelligently identify when a native memory allocation can be safely granted without leading to any out of memory (OOM) errors. By borrowing from on-heap memory, we avoid the need to update any of Spark’s internal memory settings (and therefore avoid any potential side effects like unintentionally granting Spark operators access to off-heap memory) but still have an overview of the memory utilization by Spark and the native engine.  

Key ways how Native Execution Engine for Fabric Spark boosts performance: 

  1. Unified Memory Allocation: A single manager tracks memory utilization in Spark and native and grants allocation requests made by the native engine.
  1. Reduced GC Overhead: Although natively executed queries borrow memory from the on-heap quota, this memory is not controlled by the GC. 
  1. Adaptive Scaling: Memory usage adjusts dynamically based on the workload at hand, ensuring that allocation requests are granted based on the current memory availability. 

A major goal in integrating Native Execution Engine for Fabric Spark is reducing the operational burden for users. We know customers want to focus on faster data pipelines and analytics insights, not memory tuning. That’s why, with Native Execution Engine for Fabric Runtime 1.3 (based on Apache Spark 3.5), we’ve simplified memory management to the point where customers can trust the system to do the heavy lifting. This approach not only streamlines operations but also lowers the risk of unexpected memory-related performance issues. Real-world customer feedback guided these enhancements. Early adopters who experienced OOM challenges helped us refine off-heap allocation strategies. Their insights led to dynamic improvements in how memory is distributed and reclaimed, ensuring that as data volumes scale and complexity grows, Spark remains agile, responsive, and efficient. 

With Native Execution Engine now fully integrated into Fabric Runtime 1.3, the path is clear: Faster data processing, more stable performance, and a user experience free of memory management headaches. 

Entradas de blog relacionadas

Faster data processing with Native Execution Engines in Microsoft Fabric Runtime 1.3: efficient memory management explained

febrero 23, 2026 por Ankita Victor-Levi

Introduction In today’s data landscape, as organizations scale their analytical workloads, the demand for faster, more cost-efficient computation continues to rise. Apache Spark has long been the backbone of largescale data processing with its in‑memory processing and powerful APIs, but today’s workloads demand even better performance. Microsoft Fabric addresses this challenge with the Native Execution … Continue reading “Under the hood: an introduction to the Native Execution Engine for Microsoft Fabric”

febrero 3, 2026 por Bogdan Crivat

As executives plan the next phase of their data and AI transformation, the bar for analytics infrastructure continues to rise. Enterprises are expected to support traditional business intelligence, increasingly complex analytics, and a new generation of AI-driven workloads—often on the same data, at the same time, and with far greater expectations for speed and cost … Continue reading “A turning point for enterprise data warehousing “