Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
My company is starting to use Fabric Notebooks for various activities, and to secure our credentials, we implemented the usage of a Key Vault. After setting this up and getting it to work, people noticed massive performance hits in starter pools and Python experience start-ups. For instance, the Python experience went from 5 to 10 seconds, starting to over 2 minutes on average and never below 1 minute. During research, I was not able to find an explanation until I stumbled upon a note about spark sessions (see image below). It seems that you have to decide between security and performance. Does anyone have a suggestion on how to use Key Vault without private links? Most of these scripts are going to be running in the background and Fabric is not a trusted resource so this was the only way I could get it to work.
Sincerly,
Michal
Solved! Go to Solution.
Hi @mwsmoley ,
Thank you for your insights, @lbendlin . You are right this involves not only a trade-off between security and performance but also cost. Keeping Spark pools warm can become expensive.
To balance these factors, consider the following optimizations.
1. Reduce Key Vault Calls - Cache secrets in memory or use environment variables to reduce authentication overhead.
2. Use Managed Identity - Switching to Managed Identity can enhance security and performance while simplifying access management.
3. Optimize Spark Pool Usage - Schedule workloads efficiently to minimize compute costs and improve resource allocation.
Hope this helps.
If there is any post helps, then please consider Accept it as the solution to help the other members find it more quickly.
Hi @mwsmoley ,
Could you let us know if your issue has been resolved or if you need any additional information? If the issue is resolved, please mark it as the accepted solution. This helps other members who might be facing similar issues.
Thank You.
Not really solved but I will mark it as accepted.
Hi @mwsmoley ,
Thank you for your insights, @lbendlin . You are right this involves not only a trade-off between security and performance but also cost. Keeping Spark pools warm can become expensive.
To balance these factors, consider the following optimizations.
1. Reduce Key Vault Calls - Cache secrets in memory or use environment variables to reduce authentication overhead.
2. Use Managed Identity - Switching to Managed Identity can enhance security and performance while simplifying access management.
3. Optimize Spark Pool Usage - Schedule workloads efficiently to minimize compute costs and improve resource allocation.
Hope this helps.
If there is any post helps, then please consider Accept it as the solution to help the other members find it more quickly.
It seems that you have to decide between security and performance.
It's a bit more complex than that. You have to decide between security, performance, and cost. Keeping Spark pools "warm" is getting expensive really quickly.
User | Count |
---|---|
13 | |
4 | |
3 | |
3 | |
3 |
User | Count |
---|---|
8 | |
8 | |
7 | |
6 | |
5 |