Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Enhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends August 31st. Request your voucher.

Reply
joshuaking
Regular Visitor

write_deltalake with Python Notebook is creating an "Unidentified" folder in my lakehouse.

Hi there, all!

 

Due to the smaller size of our data at my organization, we decided to use Polars and PyArrow for our data transformation in lieu of PySpark, Spark SQL, and other distributed data processing applications.

 

After making a transformation and attempting to write a DataFrame, it drops off into this "Unidentified" folder instead of registering as a delta table. I am able to query it from the SQL endpoint and read it from the "/lakehouse/default/Tables/" directory using pl.read_delta, but it doesn't seem to identify the delta table like it should (see image below)

joshuaking_1-1739222466741.png

 

I have tried to write this code in numerous ways, such as:
- Referencing both the ABFS path and the local path (/lakehouse/default/Tables)

- Kept dataframe as Polars DataFrame, and have tried switching it to PyArrow DF and Pandas DF.

- Used polars' native write_delta, and the write_deltatable function in the deltalake package.

- I've even tried Microsoft Fabric's "Write data to delta table" code snippet for the Python notebooks and that also places items under "Unidentified".

 

What am I doing wrong? I'm beginning to run out of ideas.

Thanks a bunch for any help you all can offer.

2 ACCEPTED SOLUTIONS
nilendraFabric
Super User
Super User

Hello @joshuaking 

 

Here is what I am thinking could happened. 

when writing a Delta table using Python functions like `write_deltalake` or Polars’ `write_delta`, the underlying data files are correctly created and queryable, yet the lakehouse metadata may not recognize or register the table properly. This can lead to the table being shown under an “Unidentified” folder

 

Writing data to a specific path does not always automatically update the lakehouse catalog. 

try using ddl command like in your notebook 
`CREATE TABLE student USING DELTA LOCATION '/lakehouse/your_schema/your_path/student';`
to inform the system about the logical table name

 

if this helps please accept the solution and give kudos

View solution in original post

Yes it is automatic in Fabric

 

When files in Delta format (Parquet + `_delta_log` folder) are placed in the managed area of a Lakehouse, Fabric automatically registers them as tables in the catalog.

 

In most thr cases properly structured delta file from Unidentified folder eventually moves to Table

View solution in original post

7 REPLIES 7
Thibauld_c
Regular Visitor

For those wanting to use Polars you can write to your Lakehouse with write_delta by using the full URL. For example:

 

import polars as pl

data = {"a": [1, 2], "b": [3, 4]}
df = pl.DataFrame(data)

df.write_delta("abfss://workspace-name@onelake.dfs.fabric.microsoft.com/lakehouse-name.Lakehouse/Tables/table-name")
v-tsaipranay
Community Support
Community Support

Hi @joshuaking,

Thanks for reaching out to the Microsoft fabric community forum.

 

As @nilendraFabric  mentioned in his post, could you confirm whether it resolved your issue? Please feel free to reach out if you have any further questions. If the response addressed your query, please accept it as a solution and give a 'Kudos' so other members can easily find it.

 

Thankyou.

Hi @joshuaking ,

 

May I ask if you have resolved this issue? If so, please mark the helpful reply and accept it as the solution. This will be helpful for other community members who have similar problems to solve it faster.

 

Thank you.

Hello @joshuaking ,

 

Could you please confirm if the issue has been resolved? It would be greatly appreciated if you could share your insights. Feel free to reach out if you have any further questions. please accept it as a solution and give a 'Kudos' so other members can easily find it.

 

Thank you.

nilendraFabric
Super User
Super User

Hello @joshuaking 

 

Here is what I am thinking could happened. 

when writing a Delta table using Python functions like `write_deltalake` or Polars’ `write_delta`, the underlying data files are correctly created and queryable, yet the lakehouse metadata may not recognize or register the table properly. This can lead to the table being shown under an “Unidentified” folder

 

Writing data to a specific path does not always automatically update the lakehouse catalog. 

try using ddl command like in your notebook 
`CREATE TABLE student USING DELTA LOCATION '/lakehouse/your_schema/your_path/student';`
to inform the system about the logical table name

 

if this helps please accept the solution and give kudos

I'll give this a try and see what happens. While I'm working on it, do you know if the lakehouse catalog updates on its own periodically? If not, what do you assume would be the best way to register new tables automatically?

Yes it is automatic in Fabric

 

When files in Delta format (Parquet + `_delta_log` folder) are placed in the managed area of a Lakehouse, Fabric automatically registers them as tables in the catalog.

 

In most thr cases properly structured delta file from Unidentified folder eventually moves to Table

Helpful resources

Announcements
Fabric July 2025 Monthly Update Carousel

Fabric Monthly Update - July 2025

Check out the July 2025 Fabric update to learn about new features.

August 2025 community update carousel

Fabric Community Update - August 2025

Find out what's new and trending in the Fabric community.