The ultimate Microsoft Fabric, Power BI, Azure AI, and SQL learning event: Join us in Stockholm, September 24-27, 2024.
Save €200 with code MSCUST on top of early bird pricing!
Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and more. Get started
I've been trying to get a medium-sized (100,000 records for now) dataset to load from a SQL database (on prem and Azure) to a Fabric Warehouse from a Gen2 dataflow. I can get it to load 1000 or 10000 records, but once I step up to higher numbers, I get the error below. And yes--port 1433 is open. Also--I can load all the data no problem with a native Power BI dataset or a Gen1 dataflow.
It's almost like the connection to the source database times out after a certain period of time.
Mashup Exception Data Source Error Couldn't refresh the entity because of an issue with the mashup document MashupException.Error: DataSource.Error: Microsoft SQL: A transport-level error has occurred when receiving results from the server. (provider: TCP Provider, error: 0 - An existing connection was forcibly closed by the remote host.) Details: DataSourceKind = Lakehouse;DataSourcePath = Lakehouse;Message = A transport-level error has occurred when receiving results from the server. (provider: TCP Provider, error: 0 - An existing connection was forcibly closed by the remote host.);ErrorCode = -2146232060;Number = 10054;Class = 20
So have been doing some more testing (things magically started working again last week) and am starting to think there might be a max # of times per [time period] that I can load data to the warehouse. I tried with a dataset with 85000 rows and got a failure--host closed the connection--but then tried an hour later and it worked. When I hit the 85000 mark, I'd been gradually ramping up from 1000, so 4-5 runs in the space of under an hour. Still testing to see if this is the real issue, but would be great if someone from Microsoft could let me know whether there is indeed some kind of throttling going on that prevents repreated loads into the same table...
Things seem to have stabilized for me now and I'm able to successfully load full datasets from onprem SQL directly to a Warehouse or Lakehouse. The one thing I'm still seeing--and just noticed this yesterday--is that if I have a brand new warehouse/lakehouse and try do a load to a new table with a Gen2 dataflow, it will fail if I attempt to load the full dataset the first time. If I load 100 rows, it succeeds and then I'm able to pull the filter and load everything.
Hi motoray,
This is unexpected behavior with the SQL endpoint and we are still actively debugging the issue. The issue (connection forcibly closed by remote host) is sporadic and there isn't a specific limit or throttle that you're running into.
Thanks
Thanks for working with me on this. It was really good to hear from you and have the issue confirmed yesterday.
Hi motoray,
Thanks for confirming. We are continuing the investigation.
Hi motoray & ShunOshidari,
A question to help us with the investigation: Have you modified/deleted the automatically generated staging artifacts that were generated in the workspace? These are the automatically generated Lakehouse, Warehouse, etc. with "Staging" in the name.
Thanks
I have not deleted the staging artifacts. Learned early on that everything breaks when you delete those, so I deleted those workspaces and started fresh and left the staging lakehouses/warehouses alone. Also--I don't put data in the staging areas or use dataflows to pull data from them either. At most, I open and run a query to check and see if the data is loaded to staging.
Hi motoray & ShunOshidari,
Thank you for all the information you've provided. The request ids have helped us better identify the symptoms and we are now investigating the root cause. This is not a widespread issue, but some users loading to Lakehouses and Warehouses appear to be getting errors from the SQL endpoint. We are investigating.
Thanks
Hi ShunOshiadari,
Could you please share the error message you're receving in Refresh History? If you're able to share the Request ID, that would also be helpful to better understand the failure.
Thanks
My original post (2/2)
---
My original post (1/2)
---
I am conducting a post test because my responses keep disappearing within seconds no matter how many times I post them. If this post does not disappear, I will then proceed with my original post.
thank you for sharing all of this info! We're investigating internally, but please do reach out to our support team so that they can create an incident case and get the support directly from our team.
Use this link to reach the support team: https://support.fabric.microsoft.com/support
I'll update this thread once we have more information to share.
Thanks miguel! I will do it.
Already there--case # 2307240010003563
Hi!
It appears that I'm encountering a similar problem, which relates to the number of rows in a file. I can successfully load files with over 100K rows from Azure ALDS Gen2 Storage to a Lakehouse. However, when attempting to transfer files between two Lakehouses, only the files with more than 100K rows encounter failure, while files with fewer rows transfer without any issues.
Now none of my Gen2 dataflows will complete so I'm not sure what's going on anymore. Time to call in Microsoft.
Join the community in Stockholm for expert Microsoft Fabric learning including a very exciting keynote from Arun Ulag, Corporate Vice President, Azure Data.
Check out the August 2024 Fabric update to learn about new features.
Learn from experts, get hands-on experience, and win awesome prizes.
User | Count |
---|---|
3 | |
2 | |
2 | |
2 | |
1 |
User | Count |
---|---|
3 | |
2 | |
2 | |
2 | |
2 |