Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered

Reply
paulv
Frequent Visitor

Error on missing checkpoints in Lakehouse - Delta table in Azure Datalake Gen2 unsing shortcut

Hello all, 

 

we have the following problem: 

 

We are pushing data on a regular basis from an on prem system to a delta table in an Azure datalake using deltatable python library.  

 

Files are written every 10 minutes. Then we shortcut this table to a Fabric Lakehouse. 

 

Now I get the following error: 

 

Delta table 'Head_Sk2x' has atleast '100' transaction logs, since last checkpoint. For performance reasons, it is recommended to regularly checkpoint the delta table more frequently than every '100' transactions. As a workaround, please use SQL or Spark to retrieve table schema.

 

I am running a daily clean up job on the table, which takes all the (relativly small) files using OPTIMIZE and VACUUM. But I still get the error. 

 

I am also not able to create checkpoints via this clean up notebook. 

 

Am I missing somethink? 

 

A second question, can we directly push the data to the lakehouse table without using the workaround via ADL? 

 

Any help is welcome  

 

 

1 ACCEPTED SOLUTION
Anonymous
Not applicable

Hi @paulv ,

Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.

 

 

%%spark

import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()

 

 


Note: The above code is in Scala.

Maybe it will work for you.  Please let me know incase of further queries.

View solution in original post

5 REPLIES 5
paulv
Frequent Visitor

Hi @Anonymous, 

 

We are collecting data from production side via Siemens Edge devices. 

 

The data can be preprocessed there and we run a python based container where we use the deltalake library to write the data to ADL. 

 

paulv_0-1710140719863.png

Here is the screenshot of the issue. 

 

 

 

Anonymous
Not applicable

Hi @paulv ,

Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.

 

 

%%spark

import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()

 

 


Note: The above code is in Scala.

Maybe it will work for you.  Please let me know incase of further queries.

Hi @Anonymous , 

 

Thanks a lot, this solves the error. 

 

I includes this in the clean up notebook and it works fine. 

Anonymous
Not applicable

Hi @paulv ,

Glad to know your issue got resolved. Please continue using Fabric Community for your further queries.

Anonymous
Not applicable

Hi @paulv ,

Thanks for using Fabric Community.
I would like to understand where are you running the delta table python library? 
How are you moving the data from On prem to Delta Table - is it using Fabric Notebooks? 

If possible can you please share the screenshot of the issue?

Helpful resources

Announcements
May FBC25 Carousel

Fabric Monthly Update - May 2025

Check out the May 2025 Fabric update to learn about new features.

June 2025 community update carousel

Fabric Community Update - June 2025

Find out what's new and trending in the Fabric community.