Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Get certified in Microsoft Fabric—for free! For a limited time, the Microsoft Fabric Community team will be offering free DP-600 exam vouchers. Prepare now

Reply
WomanToBlame
Advocate I
Advocate I

Notebook in Pipeline: Please attach a lakehouse to proceed

Hello,

 

When trying to run a notebook as part of a pipeline I am frequently getting the following error message:

 

Spark SQL queries are only possible in the context of a lakehouse. Please attach a lakehouse to proceed.

 

I do have a lakehouse attached to the notebooks. I can run the individual notebooks successfully and occassionally they will run in the pipeline, but 9 out of 10 times it fails with this error message. Any ideas would be greatly appreciated!

1 ACCEPTED SOLUTION
Anonymous
Not applicable

Hi @WomanToBlame ,

 

Save a copy will not work. You should open that notebook. Edit it to trigger auto save. You can also re-attach lakehouse again in notebook ux

 

Can you please try and let me know if it works?

View solution in original post

19 REPLIES 19
ajarora
Microsoft Employee
Microsoft Employee

This looks like a product bug. In case you dont receive the needed response here, the best way to proceed further is to log a support ticket.

Anonymous
Not applicable

Hi @WomanToBlame ,

Thanks for using Fabric Community and reporting this.

 

Apologies for the issue you have been facing. I would like to check are you still facing this issue? 

It's difficult to tell what could be the reason for this performance. I would request you to wait for sometime and try again.

If the issue still persists, can you please share few more details about the nature of the failure like screenshots? error? sessionID? repro steps?

I would be able to guide you better once you provide these details.

Hello @Anonymous,

 

Thanks for your response. The issue still occurs. It happens frequently enough that it is very easy to reproduce (it fails more often than succeeds). 

 

I have 4 notebooks that I have added to a pipeline. I have them set to run on the success of the previous one:

WomanToBlame_3-1699046869616.png

 

Here are some screen shots (I have the retry set to 3):

WomanToBlame_0-1699046690483.pngWomanToBlame_1-1699046717197.png

 

WomanToBlame_2-1699046786036.png

 

sessionId":"aedb126f-3b59-4490-8808-cea0f218e31d"

Anonymous
Not applicable

Hi @WomanToBlame ,

Thanks for using Fabric Community and reporting this.

 

Apologies for the issue you have been facing.  We are reaching out to the internal team to get more information related to your query and will get back to you as soon as we have an update.

Thank you, @Anonymous !

Anonymous
Not applicable

Hi @WomanToBlame ,

Thanks for using Fabric Community and reporting this,

 

You haven't done anything wrong. We've identified this as a bug in the front-end that was affecting the creation of notebooks from Lakehouse or Data Factory.

A simple workaround for this issue is to save all the notebooks once more and try again.
In case if it doesn't resolve your issue, can you please share the session id?

 

Hope this is helpful. Please let me know in case of further queries.

Thanks, @Anonymous , unfortunately that did not work. I saved a copy of the notebook and included it in the pipeline, but I still get the same error.

 

rrunId: 5XXX0f6-XXXX-4cc2-XXXX-448cXXXX5833

 

sessionId: 372XXXX9-XXXX-463b-XXXX-60b4XXXXc5a0

 

message: Spark SQL queries are only possible in the context of a lakehouse. Please attach a lakehouse to proceed.

 

Anonymous
Not applicable

Hi @WomanToBlame ,

Apologies for the issue you have been facing.  We are reaching out to the internal team to get more information related to your query and will get back to you as soon as we have an update.

Anonymous
Not applicable

Hi @WomanToBlame ,

 

Save a copy will not work. You should open that notebook. Edit it to trigger auto save. You can also re-attach lakehouse again in notebook ux

 

Can you please try and let me know if it works?

Where do you get the option on attaching lakehouse to a notebook ?

Haha, it worked! Thank you! My goodness, in my next life, I want to be a Buddhist, a philosophy teacher, or something less weird... Just kidding! But seriously, thanks a lot!

Thank you, @Anonymous, it works! Leaving the notebook in the pipeline, I detached and reattached the lakehouse. Just editing did not seem to work. Also, my original mistake was thinking the notebook needed to be removed from the pipeline, edited, and re-added.

This error still exists. Just editing did not seem to work.

Workaround is to detach and reattached the lakehouse.

Where do you get the option of attaching lakehouse to a notebook ? can you please show it?

In the notebook that is throwing the error, in the panel on the left, left click on your lakehouse name. A context menu will appear with the option to "Remove all Lakehouses". Select that option, then do "Add Lakehouse" to re-add it. Once you select "Add Lakehouse" you will have the option to add an existing Lakehouse.

 

WomanToBlame_0-1708690241190.png

 

Anonymous
Not applicable

Hi @WomanToBlame ,

Glad to know that your query got resolved. Please continue using Fabric Community for any help regarding your queries

I'm getting the same error "Spark SQL queries are only possible in the context of a lakehouse. Please attach a lakehouse to proceed." on a notebook that I've been using once a month without problems. I have tried the remove all Lakehouses and adding it back option like 5 or 6 times and is still not working, I'm also noticing from the screenshots shared that when I re-add the lakehouse I don't see it listed and with the default pinned icon. 

cmunozCP_0-1716316170931.png

 

This bug seems to have resurfaced, I'm trying to run OPTIMIZE command from notebook against a lakehouse table, the lakehouse explorer is open at the side and I can see the tables and files 

 

When I run OPTIMIZE it returns the "Spark SQL queries are only possible in the context of a lakehouse. Please attach a lakehouse to proceed." error.

 

If I click on "Lakehouse" above the "Tables" folder on the explorer, it says no results, when I open menu to add a lakehouse the lakehouse which is already attached is greyed out

I tried again today and it was working as expected with the datalake attached. Seems that it was a random bug yesterday.

Helpful resources

Announcements
September Hackathon Carousel

Microsoft Fabric & AI Learning Hackathon

Learn from experts, get hands-on experience, and win awesome prizes.

October NL Carousel

Fabric Community Update - October 2024

Find out what's new and trending in the Fabric Community.