Microsoft Fabric Updates Blog

Service principal support to connect to data in Dataflow, Datamart, Dataset and Dataflow Gen 2

Today I am very excited to announce that Azure service principal has been added as an authentication type for a set of data sources that can be used in Dataset, Dataflow, Dataflow Gen2 and Datamart. 

Azure service principal is a security identity that is application based and can be assigned permissions to access your data sources. Service principals are used to safely connect to data, without a user identity. Learn more about service principals.

SPN authentication type for SQL data source won’t work with DQ dataset

Supported data sources

  • Azure Synapse analytics
  • Azure SQL database
  • Azure data Lake gen 2
  • Azure data lake
  • Azure blob storage
  • Web
  • Dataverse
  • SharePoint online

Note: Service principal authentication is not supported for SQL data source with Direct Query in datasets

How to use service principals to connect to your data in dataflows gen 2 

In this example, we will show how you can use service principal to connect to an Azure Data Lake gen 2 through dataflows gen 2. Later in the article we show how you can use service principals in datasets.  

Prerequisite

  1. Create a service principal using the Azure portal
  1. Grant permission for the application to have read access on the data source. In the example of data lake, make sure the application has storage blob data reader access.  

Connect to your data using Service Principal within dataflows gen 

  1. Navigate to https://app.fabric.microsoft.com/ 
  1. Create a dataflow gen 2 
  1. Select Azure Data Lake Storage Gen2 as source 
  1. Fill in the URL and select create new connection 
  1. Change Authentication kind to Service Principal 
Select service principal as authentication type.
  1. Fill in Tenant ID. You can find the tenant ID in the Azure Portal. 
Find tenant ID in Azure portal.
  1. Fill in the Service principal ID 
Find client ID in Azure portal

8. Fill in the Service principal key 

  1. Click Next 

How to use service principals to connect to your data in datasets 

Prerequisite

  1. Create a service principal using the Azure portal.[ Create an Azure AD app and service principal in the portal – Microsoft Entra | Microsoft Learn
  1. Grant permission for the application to have read access on the data source. 
  1. Have a dataset published to te service. 

Connect to data using Service Principal within datasets 

  1. Navigate to https://app.fabric.microsoft.com/ 
  1. Navigate to the dataset settings page 
Open dataset settings page
  1. Navigate to the data source credentials and click edit credentials 
  1. Fill in Tenant ID. You can find the tenant ID in the Azure Portal. 
Find tenant ID in Azure portal
  1. Fill in the Service principal ID 
Find client ID in Azure portal
  1. Fill in the Service principal key 
  1. Click Sign in 

Other resources

  • Join the Fabric community to post your questions, share your feedback, and learn from others.
  • Visit Microsoft Fabric Ideas to submit feedback and suggestions for improvements and vote on your peers’ ideas!
  • Check our Known Issues page for up to date on product fixes!

Have any questions or feedback? Leave a comment below!

Related blog posts

Service principal support to connect to data in Dataflow, Datamart, Dataset and Dataflow Gen 2

December 11, 2025 by Faisal Mohamood

A Leader in Data Integration We’re thrilled to share that Microsoft has been named a Leader in the 2025 Gartner Magic Quadrant for Data Integration Tools—our fifth consecutive year in this position. In our view, this recognition reflects our commitment to help organizations unify their data estate and unlock the full potential of AI. To … Continue reading “Microsoft Recognized as a Leader in the 2025 Gartner® Magic Quadrant™ for Data Integration Tools”

December 10, 2025 by Sindhu Bharadwaj

As organizations modernize their data integration workflows, moving from Azure Data Factory (ADF) to Fabric Data Factory is an important milestone. Before starting the migration, it’s crucial to determine which pipelines and activities are ready to move, which require modifications, and which are not yet supported in Fabric. The built-in migration assessment tool in ADF … Continue reading “Assessing Your Azure Data Factory for Migration to Fabric Data Factory”