Microsoft Fabric Updates Blog

Spark Connector for Fabric Data Warehouse (DW) – Preview

We are pleased to announce the availability of the Fabric Spark connector for Fabric Data Warehouse (DW) in the Fabric Spark runtime. This connector enables Spark developers and data scientists to access and work with data from Fabric DW and the SQL analytics endpoint of the lakehouse, either within the same workspace or across different workspaces, using a simplified Spark API. The connector will be included as a default library within the Fabric Runtime, eliminating the need for separate installation.

Read Support

The connector supports reading data from tables or views from both the Data Warehouse and the SQL analytics endpoint. It is designed with security in mind, requiring minimal permission to work with Fabric SQL engines and adhering to security models such as Object Level Security (OLS), Row Level Security (RLS), and Column Level Security (CLS) defined at the SQL engine level.

Write Support

The connector now supports writing data of a dataframe to a Fabric DW table. It employs a two-phase write process: initially staging the Spark dataframe data into intermediate storage, followed by using the COPY INTO command to ingest the data into the Fabric DW table. This approach ensures scalability with increasing data volumes and supports multiple modes for writing data to a DW table.

PySpark Support

We are also excited to announce PySpark support for this connector, in addition to Scala. This means that you no longer need to use a workaround to utilize this connector in PySpark, as it is now available as a native capability.

To learn more about Spark Connector for Fabric Data Warehouse (DW), please refer to the documentation at: Spark connector for Fabric Data Warehouse

Postagens relacionadas em blogs

Spark Connector for Fabric Data Warehouse (DW) – Preview

junho 12, 2025 de RK Iyer

Introduction Whether you’re building analytics pipelines or conversational AI systems, the risk of exposing sensitive data is real. AI models trained on unfiltered datasets can inadvertently memorize and regurgitate PII, leading to compliance violations and reputational damage. This blog explores how to build scalable, secure, and compliant data workflows using PySpark, Microsoft Presidio, and Faker—covering … Continue reading “Privacy by Design: PII Detection and Anonymization with PySpark on Microsoft Fabric”

junho 11, 2025 de Eren Orbey

Earlier this year, we released AI functions in public preview, allowing Fabric customers to apply LLM-powered transformations to OneLake data simply and seamlessly, in a single line of code. Since then, we’ve continued iterating on AI functions in response to your feedback. Let’s explore the latest updates, which make AI functions more powerful, more cost-effective, … Continue reading “Introducing upgrades to AI functions for better performance—and lower costs”