Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered

Reply
tammekas
Regular Visitor

Data storing from json into lakehouse tables fails in runtime 1.3 but not in runtime 1.2

Hi!

I am having issues when creating and populating new table with data in notebook in runtime 1.3.

 

We are having code what reads in list of jsons. Every item in json has attribute content what is dynamic and can be quite big. also it can be go down so many levels. when reading we are giving along the schema for data. In schema there is declared the first level of content atribute child items and said they are strings. Thats cause we want to avoid fixt structure because items there can be random depth, ranom size etc. The content data is there for later reading only and not used for searching or filtering.

 

Saving happens as this:

 

 

 

 

 

 

 

        df.write.format('delta').mode('append').save(target_path)

 

 

 

 

 

 

 

 

What I have found.

Current code works fine in runtime 1.2 and not droping any errors.

When runing with runtime 1.3  and dataframe contains smaller data it creates and stores data fine. but when data is bigger it creates table with error sign and there is no data in.

 

tammekas_0-1733489056767.png

tammekas_0-1733492726110.png

 

When digging into the logs I ended up with following error:

2024-12-04 14:46:26,037 ERROR Utils [Executor task launch worker for task 0.0 in stage 21.0 (TID 33)]: Aborting task
org.apache.spark.sql.delta.DeltaRuntimeException: [DELTA_STATS_COLLECTION_COLUMN_NOT_FOUND] nullCount stats not found for column in Parquet metadata: [content, body, content].

I googled it and ended up https://learn.microsoft.com/en-us/azure/databricks/delta/data-skipping#specify-delta-statistics-colu...

 

But I do not understand if it is related to my error or not.

and what I need to do to get working the file saving in runtime 1.3 ?

 

1 ACCEPTED SOLUTION
Anonymous
Not applicable

Hi @tammekas ,

 

The error message indicates that Delta Lake failed to find a transaction log entry for table art.

 

I have a couple of suggestions:

 

  • Ensure that the _delta_log folder exists in the path where the table is located.

vhuijieymsft_0-1733728481269.png

vhuijieymsft_1-1733728481270.png

 

  • If the folder does not exist, try recreating the transaction log folder.
  • Clear Spark's cache to ensure that no old cached data is interfering with the new write operation.
spark.catalog.clearCache()

 

If you have any other questions please feel free to contact me.

 

Best Regards,
Yang
Community Support Team

 

If there is any post helps, then please consider Accept it as the solution  to help the other members find it more quickly.
If I misunderstand your needs or you still have problems on it, please feel free to let us know. Thanks a lot!

View solution in original post

2 REPLIES 2
Anonymous
Not applicable

Hi @tammekas ,

 

Is my follow-up just to ask if the problem has been solved?

 

If so, can you accept the correct answer as a solution or share your solution to help other members find it faster?

 

Thank you very much for your cooperation!

 

Best Regards,
Yang
Community Support Team

 

If there is any post helps, then please consider Accept it as the solution  to help the other members find it more quickly.
If I misunderstand your needs or you still have problems on it, please feel free to let us know. Thanks a lot!

Anonymous
Not applicable

Hi @tammekas ,

 

The error message indicates that Delta Lake failed to find a transaction log entry for table art.

 

I have a couple of suggestions:

 

  • Ensure that the _delta_log folder exists in the path where the table is located.

vhuijieymsft_0-1733728481269.png

vhuijieymsft_1-1733728481270.png

 

  • If the folder does not exist, try recreating the transaction log folder.
  • Clear Spark's cache to ensure that no old cached data is interfering with the new write operation.
spark.catalog.clearCache()

 

If you have any other questions please feel free to contact me.

 

Best Regards,
Yang
Community Support Team

 

If there is any post helps, then please consider Accept it as the solution  to help the other members find it more quickly.
If I misunderstand your needs or you still have problems on it, please feel free to let us know. Thanks a lot!

Helpful resources

Announcements
Join our Fabric User Panel

Join our Fabric User Panel

This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.

June FBC25 Carousel

Fabric Monthly Update - June 2025

Check out the June 2025 Fabric update to learn about new features.

June 2025 community update carousel

Fabric Community Update - June 2025

Find out what's new and trending in the Fabric community.