Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
Hi,
I have created a modelling notebook. It contains four models and usually takes 12 hours to run. When I put it in a Fabric Pipeline, it generates this error after 3 hours:
I an unsure of how to solve this.
Edit: Added configurations
Solved! Go to Solution.
thanks for sharing the details
Non-distributed Python training (SARIMAX/GAM) creates single-threaded workloads that don’t leverage Spark’s parallelism
Each notebook call in a pipeline creates a new Spark session (even for simple I/O), causing Livy session throttling.
Nested loops create 1600 variables × 4 models = 6,400 sequential tasks, overwhelming small compute nodes
Some possible solutions
Enable High Concurrency Mode
Reduces Spark session overhead by 70% through session sharing
Use Spark Job Definitions
Queueable batch jobs avoid interactive capacity limits
Restructure the training workflow using Spark’s native distributed processing instead of Python loops. Combine with batch job definitions and proper resource allocation to stay within Fabric’s capacity limits while maintaining throughput
if this is helpful please accept the answer
Hi @FatimaArshad,
Thank you for reaching out to Microsoft Fabric Community.
Thank you @nilendraFabric for addressing the issue.
As we haven’t heard back from you, we wanted to kindly follow up to check if the solution provided by the super user resolved your issue? or let us know if you need any further assistance.
If our super user response resolved your issue, please mark it as "Accept as solution" and click "Yes" if you found it helpful.
Thanks and regards,
Anjan Kumar Chippa
Hi @FatimaArshad,
We wanted to kindly follow up to check if the solution provided by the super user resolved your issue.
If our super user response resolved your issue, please mark it as "Accept as solution" and click "Yes" if you found it helpful.
Thanks and regards,
Anjan Kumar Chippa
Hi @FatimaArshad,
As we haven’t heard back from you, we wanted to kindly follow up to check if the solution provided by the super user resolved your issue.
If our super user response resolved your issue, please mark it as "Accept as solution" and click "Yes" if you found it helpful.
Thanks and regards,
Anjan Kumar Chippa
thanks for sharing the details
Non-distributed Python training (SARIMAX/GAM) creates single-threaded workloads that don’t leverage Spark’s parallelism
Each notebook call in a pipeline creates a new Spark session (even for simple I/O), causing Livy session throttling.
Nested loops create 1600 variables × 4 models = 6,400 sequential tasks, overwhelming small compute nodes
Some possible solutions
Enable High Concurrency Mode
Reduces Spark session overhead by 70% through session sharing
Use Spark Job Definitions
Queueable batch jobs avoid interactive capacity limits
Restructure the training workflow using Spark’s native distributed processing instead of Python loops. Combine with batch job definitions and proper resource allocation to stay within Fabric’s capacity limits while maintaining throughput
if this is helpful please accept the answer
My intial thinking is going for resource and compute constraints. Which SKU you are on, what are the sizes of these models etc. please share further details
Linear Regress, SARIMAX, GAM, XGboost. Training is done on 1600 dependent variables. One loop goes through set of features, second goes through models, and third applies models to all these variables
I am using Spark 3.4, compute small 1-4 nodes. But I only use spark to read and write to delta tables. I use Python mainly
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
74 | |
48 | |
16 | |
12 | |
7 |
User | Count |
---|---|
81 | |
81 | |
27 | |
8 | |
7 |