Microsoft Fabric Community Conference 2025, March 31 - April 2, Las Vegas, Nevada. Use code MSCUST for a $150 discount.
Register nowGet certified as a Fabric Data Engineer: Check your eligibility for a 50% exam voucher offer and join us for free live learning sessions to get prepared for Exam DP-700. Get started
I am facing an issue with the Notebook activity in Fabric Data Factory, where the workspace ID is not being resolved dynamically. I extracted the JSON definition of my pipeline from Azure DevOps and noticed that the workspaceId is not explicitly defined for the Notebook activity.
Any insights or best practices would be greatly appreciated!
Pipeline JSON sample:
{
"properties": {
"activities": [
{
"type": "TridentNotebook",
"typeProperties": {
"notebookId": "c83f4cb0-d163-48cb-9a11-b4f029e0820f"
},
"policy": {
"timeout": "0.12:00:00",
"retry": 0,
"retryIntervalInSeconds": 30,
"secureInput": false,
"secureOutput": false
},
"name": "Check new folders and files",
"description": "Check if new folders and Ingest folder entry in config table and files present in the folder in file config table.",
"dependsOn": [
{
"activity": "fetch env info from ctrl tbl",
"dependencyConditions": [
"Succeeded"
]
}
]
},
Solved! Go to Solution.
Hi @sudarshan5 ,
Here is my personal opinion on your question:
To ensure that the Notebook activity correctly resolves the workspace, you should do these steps below:
1. When creating a pipeline, add a Notebook activity and configure it to point to the correct notebook within your workspace. Make sure to set the workspace and notebook parameters correctly in the activity settings.
2. You can dynamically retrieve the list of notebooks within a workspace using REST APIs. This can help in scenarios where you need to dynamically select notebooks based on certain conditions.
3. Enable high concurrency mode in your workspace settings to optimize the execution of multiple notebooks. This can help in minimizing the startup time by reusing existing Spark sessions.
Creating a config table to store notebook IDs for different environments (Dev, Test, Prod) is a good practice. This approach helps in managing and dynamically selecting the correct notebook ID based on the environment.
Here is also a ducument that you can read:
Microsoft Fabric: Dynamically Provide Notebook ID in Data Factory Pipeline
Best Regards
Yilong Zhou
If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.
Hi @sudarshan5 ,
Here is my personal opinion on your question:
To ensure that the Notebook activity correctly resolves the workspace, you should do these steps below:
1. When creating a pipeline, add a Notebook activity and configure it to point to the correct notebook within your workspace. Make sure to set the workspace and notebook parameters correctly in the activity settings.
2. You can dynamically retrieve the list of notebooks within a workspace using REST APIs. This can help in scenarios where you need to dynamically select notebooks based on certain conditions.
3. Enable high concurrency mode in your workspace settings to optimize the execution of multiple notebooks. This can help in minimizing the startup time by reusing existing Spark sessions.
Creating a config table to store notebook IDs for different environments (Dev, Test, Prod) is a good practice. This approach helps in managing and dynamically selecting the correct notebook ID based on the environment.
Here is also a ducument that you can read:
Microsoft Fabric: Dynamically Provide Notebook ID in Data Factory Pipeline
Best Regards
Yilong Zhou
If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.
March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount!
If you love stickers, then you will definitely want to check out our Community Sticker Challenge!
User | Count |
---|---|
8 | |
2 | |
2 | |
1 | |
1 |