Microsoft Fabric Updates Blog

Run Spark Job Definitions in Pipelines with Service Principal or Workspace Identity

The Spark job definition activity in Microsoft Fabric Data Factory pipelines now supports connection property, unlocking a more secure and production-ready way to run your SJDs.

What’s New?

With this update, you can configure Notebook activities to run as Service Principal (SPN) or Workspace Identity (WI). These authentication methods are our recommended approach for production environments, ensuring:

  • Operational reliability: Avoid issues tied to user credentials, such as password changes or account deactivation.
  • Enterprise-grade security: Service-based authentication reduces risk and simplifies compliance.
  • Consistent automation: Keep your pipelines running smoothly without manual intervention.
SJD supports connection property now

Why it Matters

Previously, many customers relied on user authentication for pipeline runs, which can lead to broken workflows when users leave the organization or tokens expire. By enabling SPN and WI through the connection property, you gain:

  • Scalable orchestration for notebooks in complex workflows.
  • Better governance with centralized identity management.
  • Future-proof automation for production workloads.

How to Get Started

  1. In your pipeline, add a Spark job definition activity.
  2. Under Connection, you can configure a new connection or use an existing one.
  3. Provide the required credentials or identity configuration (you can use SPN and WI now).
  4. Run your pipeline and enjoy a secure, automated experience.

To learn more, check out our documentation on the Spark Job Definition activity.

Related blog posts

Run Spark Job Definitions in Pipelines with Service Principal or Workspace Identity

April 20, 2026 by Penny Zhou

Coordinating dbt runs with upstream ingestion and downstream consumption often requires complex solutions and different tools. You can now add a dbt job activity (Preview) directly to your Fabric pipelines. This lets you orchestrate dbt transformations alongside other pipeline activities, so you can build end-to-end data workflows without switching tools. Why this matters Run dbt … Continue reading “Orchestrate dbt jobs activity in your Fabric pipelines (Preview)”

April 16, 2026 by Nick Salch

As the Data Integration Customer Advisory Team (CAT) lead, I spent a lot of time talking to customers at the recent FabCon/SQLCon about Fabric Data Factory, and I came away with a clear picture of what’s on customers’ minds when it comes to the future of data integration. Many of the same questions came up … Continue reading “Answers to common questions about Fabric Data Factory”