Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Get Fabric Certified for FREE during Fabric Data Days. Don't miss your chance! Request now

Reply
choran
Regular Visitor

Notebooks taking a considerable time to run when deployed in data pipelines

Hi all!

 

I am currently experiencing substantial runtimes for my data pipelines due to long runtimes for my notebooks. 

A breakdown of the time that it takes for a notebook to run (when no other pipelines, or notebooks are running, that may consume Spark capacity):

- Notebook activity overall runtime: 4-7mins

- Actual time to run activities within notebook: Less than 60s

 

I have shown an example screenshot when I view the run details of the notebook activity from within the pipeline: 

choran_0-1725015030740.png

 

- The notebook is being run in a custom environment as I need to import the AzureOpenAI library. 

- I am currently utilising F8 SKU.

 

I think it is taking a long-time to run as it takes time to connect to a Spark cluster. Is there a way of speeding this up with configurations, or is it a case of increasing the SKU?

 

1 ACCEPTED SOLUTION
frithjof_v
Super User
Super User

I think using custom environments increase the start-up time.

 

Also, if you're not using the starter pools, it will increase the start-up time.

 

I'm not sure about the queuing you're experiencing. I'm not sure if that is due to using a custom environment or if it's about something else. 

 

Is this the only notebook in your pipeline? Or do you run another notebook just before this one?

 

Could it be that you or another user is running a notebook or somehow using spark at the same time?

 

Here is a couple of articles about queuing:

 

https://learn.microsoft.com/en-us/fabric/data-engineering/job-queueing-for-fabric-spark

 

https://learn.microsoft.com/en-us/fabric/data-engineering/spark-job-concurrency-and-queueing

View solution in original post

1 REPLY 1
frithjof_v
Super User
Super User

I think using custom environments increase the start-up time.

 

Also, if you're not using the starter pools, it will increase the start-up time.

 

I'm not sure about the queuing you're experiencing. I'm not sure if that is due to using a custom environment or if it's about something else. 

 

Is this the only notebook in your pipeline? Or do you run another notebook just before this one?

 

Could it be that you or another user is running a notebook or somehow using spark at the same time?

 

Here is a couple of articles about queuing:

 

https://learn.microsoft.com/en-us/fabric/data-engineering/job-queueing-for-fabric-spark

 

https://learn.microsoft.com/en-us/fabric/data-engineering/spark-job-concurrency-and-queueing

Helpful resources

Announcements
November Fabric Update Carousel

Fabric Monthly Update - November 2025

Check out the November 2025 Fabric update to learn about new features.

Fabric Data Days Carousel

Fabric Data Days

Advance your Data & AI career with 50 days of live learning, contests, hands-on challenges, study groups & certifications and more!

FabCon Atlanta 2026 carousel

FabCon Atlanta 2026

Join us at FabCon Atlanta, March 16-20, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.