Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered

Reply
paulv
Frequent Visitor

Error on missing checkpoints in Lakehouse - Delta table in Azure Datalake Gen2 unsing shortcut

Hello all, 

 

we have the following problem: 

 

We are pushing data on a regular basis from an on prem system to a delta table in an Azure datalake using deltatable python library.  

 

Files are written every 10 minutes. Then we shortcut this table to a Fabric Lakehouse. 

 

Now I get the following error: 

 

Delta table 'Head_Sk2x' has atleast '100' transaction logs, since last checkpoint. For performance reasons, it is recommended to regularly checkpoint the delta table more frequently than every '100' transactions. As a workaround, please use SQL or Spark to retrieve table schema.

 

I am running a daily clean up job on the table, which takes all the (relativly small) files using OPTIMIZE and VACUUM. But I still get the error. 

 

I am also not able to create checkpoints via this clean up notebook. 

 

Am I missing somethink? 

 

A second question, can we directly push the data to the lakehouse table without using the workaround via ADL? 

 

Any help is welcome  

 

 

1 ACCEPTED SOLUTION
Anonymous
Not applicable

Hi @paulv ,

Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.

 

 

%%spark

import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()

 

 


Note: The above code is in Scala.

Maybe it will work for you.  Please let me know incase of further queries.

View solution in original post

5 REPLIES 5
paulv
Frequent Visitor

Hi @Anonymous, 

 

We are collecting data from production side via Siemens Edge devices. 

 

The data can be preprocessed there and we run a python based container where we use the deltalake library to write the data to ADL. 

 

paulv_0-1710140719863.png

Here is the screenshot of the issue. 

 

 

 

Anonymous
Not applicable

Hi @paulv ,

Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.

 

 

%%spark

import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()

 

 


Note: The above code is in Scala.

Maybe it will work for you.  Please let me know incase of further queries.

Hi @Anonymous , 

 

Thanks a lot, this solves the error. 

 

I includes this in the clean up notebook and it works fine. 

Anonymous
Not applicable

Hi @paulv ,

Glad to know your issue got resolved. Please continue using Fabric Community for your further queries.

Anonymous
Not applicable

Hi @paulv ,

Thanks for using Fabric Community.
I would like to understand where are you running the delta table python library? 
How are you moving the data from On prem to Delta Table - is it using Fabric Notebooks? 

If possible can you please share the screenshot of the issue?

Helpful resources

Announcements
Join our Fabric User Panel

Join our Fabric User Panel

This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.

May FBC25 Carousel

Fabric Monthly Update - May 2025

Check out the May 2025 Fabric update to learn about new features.

June 2025 community update carousel

Fabric Community Update - June 2025

Find out what's new and trending in the Fabric community.