Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
Hello all,
we have the following problem:
We are pushing data on a regular basis from an on prem system to a delta table in an Azure datalake using deltatable python library.
Files are written every 10 minutes. Then we shortcut this table to a Fabric Lakehouse.
Now I get the following error:
Delta table 'Head_Sk2x' has atleast '100' transaction logs, since last checkpoint. For performance reasons, it is recommended to regularly checkpoint the delta table more frequently than every '100' transactions. As a workaround, please use SQL or Spark to retrieve table schema.
I am running a daily clean up job on the table, which takes all the (relativly small) files using OPTIMIZE and VACUUM. But I still get the error.
I am also not able to create checkpoints via this clean up notebook.
Am I missing somethink?
A second question, can we directly push the data to the lakehouse table without using the workaround via ADL?
Any help is welcome
Solved! Go to Solution.
Hi @paulv ,
Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.
%%spark
import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()
Note: The above code is in Scala.
Maybe it will work for you. Please let me know incase of further queries.
Hi @Anonymous,
We are collecting data from production side via Siemens Edge devices.
The data can be preprocessed there and we run a python based container where we use the deltalake library to write the data to ADL.
Here is the screenshot of the issue.
Hi @paulv ,
Can you please run this code in a notebook, then refresh the lakehouse table that had the erro.
%%spark
import org.apache.spark.sql.delta.DeltaLog
DeltaLog.forTable(spark,"Tables/yourtablenamehere").checkpoint()
Note: The above code is in Scala.
Maybe it will work for you. Please let me know incase of further queries.
Hi @Anonymous ,
Thanks a lot, this solves the error.
I includes this in the clean up notebook and it works fine.
Hi @paulv ,
Glad to know your issue got resolved. Please continue using Fabric Community for your further queries.
Hi @paulv ,
Thanks for using Fabric Community.
I would like to understand where are you running the delta table python library?
How are you moving the data from On prem to Delta Table - is it using Fabric Notebooks?
If possible can you please share the screenshot of the issue?
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
79 | |
46 | |
17 | |
11 | |
7 |
User | Count |
---|---|
83 | |
82 | |
27 | |
8 | |
7 |