The ultimate Microsoft Fabric, Power BI, Azure AI, and SQL learning event: Join us in Stockholm, September 24-27, 2024.
Save €200 with code MSCUST on top of early bird pricing!
Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and more. Get started
Im running 4 background jobs
An event stream
2 pipelines running a notebook using pyspark streaming.
A pipeline running SQL to copy for Lakehouse to ware house.
I have been careful to add timeouts in the pipelines and notebooks
Amount of data is like 20 messages an hour.
What im seeing is during the day the amount of CU these background jobs use increase.
However when i stop and start Fabric instance it goes back down again.
How do i go about solving this ?
staggering did not help
2 of the notebooks are like this.
During day restart same. I cant believe that this is coincidence of more jobs running at the same peak load. There is only 4 jobs and it was doing the same when there were 2 jobs. ( Both pipelines running every 11 minutes running a notebook with a 2 minute trigger).
I know i can prob work around it by running the notebook once per day and leave it running but i want to know why this is happening and how to get detail ?
Hi @bklooste ,
Based on the description, try staggering job start times to distribute the load more evenly throughout the day.
Besides, try to reduce the polling frequency of the event stream and pipelines.
Best Regards,
Wisdom Wu
If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.
Join the community in Stockholm for expert Microsoft Fabric learning including a very exciting keynote from Arun Ulag, Corporate Vice President, Azure Data.
Check out the August 2024 Power BI update to learn about new features.
Learn from experts, get hands-on experience, and win awesome prizes.
User | Count |
---|---|
52 | |
22 | |
11 | |
11 | |
9 |
User | Count |
---|---|
112 | |
32 | |
30 | |
20 | |
19 |