Supplies are limited. Contact info@espc.tech right away to save your spot before the conference sells out.
Get your discountScore big with last-minute savings on the final tickets to FabCon Vienna. Secure your discount
Hi all,
I'm using F4 and run a pipeline that calls some child Fabric notebooks to load data. I'm facing with the execution is too long 7-8 mins. While In my notebook, all cells take total execution time only 15s for 6000 rows (read dataframe, write to next storage).
And when I checked the monitor screen, It took over 7 mins with Running stage from 5:57 to 5:59 - 2 mins only.
I already configureed high concurrency and try to test with some spark configurations such as spark.ms.autotune.enabled.
How can I resolve this?
Hi @D4zk ,
Are you using parallelisation in DAG. I guess if you do parallel execution of all child notebook , this can further save execution time.
Regards,
Srisakthi
Hi @D4zk ,
It looks like your pipeline is taking much longer to execute than a single notebook unit.
Are there any other concurrent interactions during the time you are running the pipeline? Make sure these dependencies are not causing delays. Network latency or slow response from external systems can affect the overall execution time.
Alternatively, refer to the following document for a description of the detailed metrics of pipeline operation.
How to monitor pipeline runs - Microsoft Fabric | Microsoft Learn
Best Regards,
Adamk Kong
If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.
Hi, I already find an approach to help reduce the time to start, queue and idle of node, that is using DAG to call all notebooks from 1 main notebook. It can help me reduce from 2 hours to 37 minutes for 37 tables.
Thanks for you response.
User | Count |
---|---|
2 | |
2 | |
1 | |
1 | |
1 |