Microsoft Fabric Updates Blog

Run Spark Job Definitions in Pipelines with Service Principal or Workspace Identity

The Spark job definition activity in Microsoft Fabric Data Factory pipelines now supports connection property, unlocking a more secure and production-ready way to run your SJDs.

What’s New?

With this update, you can configure Notebook activities to run as Service Principal (SPN) or Workspace Identity (WI). These authentication methods are our recommended approach for production environments, ensuring:

  • Operational reliability: Avoid issues tied to user credentials, such as password changes or account deactivation.
  • Enterprise-grade security: Service-based authentication reduces risk and simplifies compliance.
  • Consistent automation: Keep your pipelines running smoothly without manual intervention.
SJD supports connection property now

Why it Matters

Previously, many customers relied on user authentication for pipeline runs, which can lead to broken workflows when users leave the organization or tokens expire. By enabling SPN and WI through the connection property, you gain:

  • Scalable orchestration for notebooks in complex workflows.
  • Better governance with centralized identity management.
  • Future-proof automation for production workloads.

How to Get Started

  1. In your pipeline, add a Spark job definition activity.
  2. Under Connection, you can configure a new connection or use an existing one.
  3. Provide the required credentials or identity configuration (you can use SPN and WI now).
  4. Run your pipeline and enjoy a secure, automated experience.

To learn more, check out our documentation on the Spark Job Definition activity.

Related blog posts

Run Spark Job Definitions in Pipelines with Service Principal or Workspace Identity

April 16, 2026 by Nick Salch

As the Data Integration Customer Advisory Team (CAT) lead, I spent a lot of time talking to customers at the recent FabCon/SQLCon about Fabric Data Factory, and I came away with a clear picture of what’s on customers’ minds when it comes to the future of data integration. Many of the same questions came up … Continue reading “Answers to common questions about Fabric Data Factory”

March 31, 2026 by Mark Kromer

Introduction The world of data integration is rapidly evolving, and staying up to date with the latest technologies is crucial for organizations seeking to make the most of their data assets. Available now are the newest innovations in Fabric Data Factory pipelines and Apache Airflow job orchestration, designed to empower data engineers, architects, and analytics … Continue reading “Announcing the latest innovations in Fabric Data Factory: Apache Airflow jobs and pipelines”