Microsoft Fabric Updates Blog

Copy Data from Lakehouse in Another Workspace Using Data pipeline

In this post, we will introduce the practice of copying data between Lakehouse that cross different workspaces via Data pipeline.

In this example, we will copy data from a Lakehouse in another workspace to a Lakehouse in the current workspace by leveraging parameters to specify the workspace and Lakehouse.

Follow the steps below to achieve that:

  1. In a Data pipeline, add a copy activity into the canvas. Under the Source tab, specify a Lakehouse in another workspace as the connection.
  • In the Connection section, select Use dynamic content at the bottom of the drop-down list.
  • In the pop-up Add dynamic content pane, under the Parameters tab, select +
Screenshot showing the Add dynamic content page.
  • Specify the name for your parameter and specify a default value. A white background with black lines

Description automatically generated

Note that the parameter value should be the Lakehouse object ID. To get your Lakehouse object ID, open your Lakehouse, and the ID is after /lakehouses/ in the URL.

Screenshot showing the Lakehouse object ID.
  • Select Save to go back to the Add dynamic content pane. Then select your parameter so it appears in the expression box. Then select OK. A screenshot of a computer

Description automatically generated
  1. You’ll go back to the pipeline page and can see the parameter expression specified in the Connection section. Then specify the ID of the workspace where your Lakehouse is located.

To get your Workspace ID, open your workspace, and the ID is after /groups/ in your URL.

  1. Select Refresh after Table name and select OK in the pop-up pane since you already specified the default value for the parameter. Then select the table to copy from the drop-down list.
  1. In the Destination tab, select the Lakehouse that you want to copy data to in the drop-down list of Connection. Then specify the destination table.
  2. Select Run. In the pop-up Pipeline run pane, select OK. After the Data pipeline runs, you can see the table is copied to the destination Lakehouse.

We have a plan to introduce a native experience for the cross-workspace experience. This will allow you to select Lakehouses and other objects in another workspace directly. Please stay tuned!

Have any questions or feedback? Leave a comment below!

Susiję tinklaraščio įrašai

Copy Data from Lakehouse in Another Workspace Using Data pipeline

spalio 30, 2024 – Patrick LeBlanc

Welcome to the October 2024 Update! Here are a few, select highlights of the many we have for Fabric this month. API for GraphQL support for Service Principal Names (SPNs). Introducing a powerful new feature in Lakehouses: Sorting, Filtering, and Searching capabilities. An addition to KQL Queryset that will revolutionize the way you interact with … Continue reading “Fabric October 2024 Monthly Update”

spalio 29, 2024 – Leo Li

We’re excited to announce several powerful updates to the Virtual Network (VNET) Data Gateway, designed to further enhance performance and improve the overall user experience. These new features allow users to better manage increasing workloads, perform complex data transformations, and simplify log management. Expanded Cluster Size from 5 to 7 One of the key improvements … Continue reading “New Features and Enhancements for Virtual Network Data Gateway”