Microsoft Fabric Updates Blog

Copy Data from Lakehouse in Another Workspace Using Data pipeline

In this post, we will introduce the practice of copying data between Lakehouse that cross different workspaces via Data pipeline.

In this example, we will copy data from a Lakehouse in another workspace to a Lakehouse in the current workspace by leveraging parameters to specify the workspace and Lakehouse.

Follow the steps below to achieve that:

  1. In a Data pipeline, add a copy activity into the canvas. Under the Source tab, specify a Lakehouse in another workspace as the connection.
  • In the Connection section, select Use dynamic content at the bottom of the drop-down list.
  • In the pop-up Add dynamic content pane, under the Parameters tab, select +
Screenshot showing the Add dynamic content page.
  • Specify the name for your parameter and specify a default value. A white background with black lines

Description automatically generated

Note that the parameter value should be the Lakehouse object ID. To get your Lakehouse object ID, open your Lakehouse, and the ID is after /lakehouses/ in the URL.

Screenshot showing the Lakehouse object ID.
  • Select Save to go back to the Add dynamic content pane. Then select your parameter so it appears in the expression box. Then select OK. A screenshot of a computer

Description automatically generated
  1. You’ll go back to the pipeline page and can see the parameter expression specified in the Connection section. Then specify the ID of the workspace where your Lakehouse is located.

To get your Workspace ID, open your workspace, and the ID is after /groups/ in your URL.

  1. Select Refresh after Table name and select OK in the pop-up pane since you already specified the default value for the parameter. Then select the table to copy from the drop-down list.
  1. In the Destination tab, select the Lakehouse that you want to copy data to in the drop-down list of Connection. Then specify the destination table.
  2. Select Run. In the pop-up Pipeline run pane, select OK. After the Data pipeline runs, you can see the table is copied to the destination Lakehouse.

We have a plan to introduce a native experience for the cross-workspace experience. This will allow you to select Lakehouses and other objects in another workspace directly. Please stay tuned!

Have any questions or feedback? Leave a comment below!

Related blog posts

Copy Data from Lakehouse in Another Workspace Using Data pipeline

July 24, 2024 by Jason Himmelstein

Welcome to the July 2024 update. Here are a few, select highlights of the many we have for Fabric. Creating and managing Git branches & connected workspaces with Git integration just got with the latest enhancements to Fabric Git integration. You now have the capability to perform restore-in-place of a warehouse in Microsoft Fabric through … Continue reading “Microsoft Fabric July 2024 Update”

July 19, 2024 by Penny Zhou

We are thrilled to share that you can easily browse and connect to your Azure resources automatically with the modern get data experience of Data Pipelines. With the Browse Azure experience, you don’t need to manually fill in the endpoint, URL or server and database name when connecting to Azure resources. Instead, the connection will … Continue reading “Quickly Connect to your Azure Resources in Fabric with the Data Pipeline Modern Get Data Experience”