Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
I currently have a pipeline in which I copy data from two different sources and create lakehouse tables. Then I added a waiting period of 160seconds to be sure to leave enough time for the lakehouse to refresh, and then run a DataFlow Gen2 in which I append both tables, do some transformations and add the final table to the DataWarehouse (using Overwrite).
Recently, I have noticed that sometimes in the final table of the data warehouse all data from one of the two sources is missing, or new rows of data are missing. Everytime I have noticed this happenning, the DataFlow ran successfully and I also verify that the data is available in the lakehouse tables.
Is there something else I could monitor to make sure that my data warehouse table gets updated successfully?
Has anyone else noticed this? Could it be a bug?
Solved! Go to Solution.
Hi ilseeb, I have noticed that sometimes 160 seconds will not be enough to update the SQL-endpoint of a lakehouse. I found a blog posts that contains a Python script to refresh the SQL-endpoint programmatically and wait for the lakehouse to be refreshed before moving on. You can find the blog post here: https://www.obvience.com/blog/fix-sql-analytics-endpoint-sync-issues-in-microsoft-fabric-data-not-sh.... You can add this in a notebook and run that after copying the data and before appending and transforming.
Hi ilseeb, I have noticed that sometimes 160 seconds will not be enough to update the SQL-endpoint of a lakehouse. I found a blog posts that contains a Python script to refresh the SQL-endpoint programmatically and wait for the lakehouse to be refreshed before moving on. You can find the blog post here: https://www.obvience.com/blog/fix-sql-analytics-endpoint-sync-issues-in-microsoft-fabric-data-not-sh.... You can add this in a notebook and run that after copying the data and before appending and transforming.
@ilseeb @FabianSchut @Anonymous
Unfortunately, this is a gamble. Being an undocumented and unsupported solution by Microsoft, as soon as one morning some Microsoft tech bro parachutes into his cubicle all fresh and perky, full of unwarranted optimism for the day ahead fueled by too much caffeine, and on a whim decides to change some code in the API, this 'so-called' solution will come crashing down while you run it in production.
Who wants to gamble with production workloads? Not me.
But since Microsoft has been working on this and they say they will release something in Q2, like a REST API endpoint to query in order to refresh the lakehouse metadata (I guess they are just officializing this undocumented solution), why not send them a message, loud and clear, by voting for the idea found at the link below and that will give us a new pipeline activity that's easy to set up from within a pipeline, set and forget, and will do the job without having to mess around with a Python Notebook? (I have in mind here those users that expect user-friendliness, and not those who are coding veterans.)
Here is the idea link: LAKEHOUSE I/O WRITE DELAY MITIGATION - Microsoft Fabric Community
Thanks a lot for this link! I'll try it out
Hi @ilseeb
Thank you very much FabianSchut for your prompt reply.
Can you tell me if your problem is solved? If yes, please accept it as solution.
Regards,
Nono Chen
If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |
User | Count |
---|---|
4 | |
3 | |
2 | |
1 | |
1 |