Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
I had the same problem with some staging tables that have a truncate and load pattern.
What I'm doing now is to drop the table and recreate it. That appears to have gotten rid of all the delta parquet history files. This does mean also recreating any indexes, and any table-level permissions, so it's not the normal pattern on a T-SQL database, but the Fabric Warehouse is not a T-SQL database.
I used this to find the real size of the Fabric Warehouse, since the Azure Storage Explorer doesn't count it all.
"Whats the size of my Fabric Workspace / Lakehouse / Warehouse" - Mark Pryce-Maher
Another thought I had was to put the staging table in a Fabric Lakehouse, because that should be reachable from T-SQL in the Fabric Warehouse using a 3-part name. Then I would be able to adjust the retention and/or manually run the vacuum command. But the drop and recreate was fully T-SQL and had less setup. I do have an existing Fabric Lakehouse in the same workspace, but I didn't fully look into if adjusting the retention was table specific, or applies to the whole Lakehouse, meaning I might need a Lakehouse solely for staging tables.
Hello, everyone,
Any updates on this? My One Lake Storage is growing enormously every day, and we need a solution as soon as possible. I look forward to hearing from you.
Thanks!
Thank @govindarajan_d for the valuable insights and document link.
Here is my understanding:
This is expected behavior from the perspective of delta tables and parquet files. Because the parquet file cannot be modified, all operations on the data, including inserts, deletes, and updates, will generate new parquet files instead of updating the existing parquet file. Meanwhile, due to the time travel function, old parquet files cannot be deleted, making it possible to query prior versions of data.
Currently, the data retention for time travel queries in Fabric Warehouses is thirty calendar days. However Fabric doesn't allow us to modify the retention duration for Warehouses.
Best Regards,
Jing
Community Support Team
Hi Jing,
We have data older than 30 days, which is causing the enormous growth in OneLake. Please find the below screenshot for your references.
@datakulture How are you saying this particular folder is from Aug? The folders that you see most likely would be partitions. And a file generated in Aug doesn't mean that it is up for deletion. Let's say you deleted some records or did some OPTIMIZE on top of a delta table that reduced the number of files, if you run VACUUM and if the file is older than retention period, then it will be deleted.
But if you still believe your delta table is not working properly, I would recommend creating a MS support ticket.
Firstly, since it's a Fabric Warehouse, VACUUM is completely ruled out.
I am performing a truncate-and-load operation for the particular table, and the folder belongs to that specific table. Auto-deletion should occur for files older than 30 days, but that is not happening currently. Let me reach out to Microsoft for a resolution. Thanks.
@datakulture Yeah I understand VACUUM is not possible manually, but what I meant was the automatic run process. If it is a truncate and load operation, then definitely the previous version files should be there if it is older than 30 days.
Thanks for your valuable insights.
The data is still persisting in Delta tables for more than 30 days. Please find the screenshot below for your reference. The August 2024 data is still exists in the one lake.
@govindarajan_d Question - How are you saying this particular folder is from Aug? Please refer above screenshot.
Hi @datakulture,
Delta tables have version mechanism and that is how it supports this feature - Time travel in Warehouse within Microsoft Fabric - Microsoft Fabric | Microsoft Learn.
Using time travel feature, you can query and clone data as it was in the past with a limit of upto 30 days. So the retention time is 30 days. While in Lakehouse you can manually run VACUUM to clear up the old files with whichever retention period you need but in Warehouse currently it is not possible.
Hi @lbendlin
This is a warehousing problem only. I'm having the identical storage problem with my ongoing project.
Are you sure this is a Warehouse issue?
Microsoft Fabric Lakehouse OPTIMIZE and VACUUM for Table Maintenance
Yes it's a MS Fabric warehouse issue.
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |
User | Count |
---|---|
5 | |
3 | |
3 | |
2 | |
2 |