Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Get certified in Microsoft Fabric—for free! For a limited time, get a free DP-600 exam voucher to use by the end of 2024. Register now

Reply

Read data from lakehouse table from azure function

Hi,

I must read data from a lakehouse table in Fabric with an Azure Function (Python). However, I cannot find any code examples. Can anyone provide an example, or how to achieve this? 

Regards from Norway!

1 ACCEPTED SOLUTION

Hi, I found that the easiest way to solve this was to write the data I need in the Azure Function as a csv file from a dataframe in a Python notebook and and read the data in Python code in the Azure with a Pandas dataframe:

This is the Azure Python Function code:

from azure.storage.filedatalake import DataLakeServiceClient
from azure.identity import DefaultAzureCredential
import pandas as pd
from io import StringIO

# Set your account and workspace details
ACCOUNT_NAME = "onelake"
WORKSPACE_NAME = "GUID"  # Workspace GUID
DIRECTORY_PATH = "GUID/Files/FILENAME"  # Directory path containing the CSV file

def main():
    # Create a service client using the default Azure credential
    account_url = f"https://{ACCOUNT_NAME}.dfs.fabric.microsoft.com"
    token_credential = DefaultAzureCredential()
    service_client = DataLakeServiceClient(account_url, credential=token_credential)

    # Create a file system client for the workspace
    file_system_client = service_client.get_file_system_client(WORKSPACE_NAME)

    # List all files in the specified directory
    paths = file_system_client.get_paths(path=DIRECTORY_PATH)

    # Find the CSV file in the directory
    csv_file_path = None
    for path in paths:
        if path.name.endswith('.csv', csv_file_path = path.name
            break  # Stop after finding the first CSV file

    if csv_file_path:
        print(f"Found CSV file: {csv_file_path}")
       
        # Create a file client for the specific CSV file
        file_client = file_system_client.get_file_client(csv_file_path)

        # Download the file content
        download = file_client.download_file()
        downloaded_bytes = download.readall()

        # Convert the downloaded bytes to a StringIO object for use with pandas
        csv_data = StringIO(downloaded_bytes.decode('utf-8'))

        # Load the CSV data into a pandas DataFrame
        df = pd.read_csv(csv_data)

        # Print or process the DataFrame
        print(df)
    else:
        print("No CSV file found in the directory.")
 

View solution in original post

2 REPLIES 2

Hi, I found that the easiest way to solve this was to write the data I need in the Azure Function as a csv file from a dataframe in a Python notebook and and read the data in Python code in the Azure with a Pandas dataframe:

This is the Azure Python Function code:

from azure.storage.filedatalake import DataLakeServiceClient
from azure.identity import DefaultAzureCredential
import pandas as pd
from io import StringIO

# Set your account and workspace details
ACCOUNT_NAME = "onelake"
WORKSPACE_NAME = "GUID"  # Workspace GUID
DIRECTORY_PATH = "GUID/Files/FILENAME"  # Directory path containing the CSV file

def main():
    # Create a service client using the default Azure credential
    account_url = f"https://{ACCOUNT_NAME}.dfs.fabric.microsoft.com"
    token_credential = DefaultAzureCredential()
    service_client = DataLakeServiceClient(account_url, credential=token_credential)

    # Create a file system client for the workspace
    file_system_client = service_client.get_file_system_client(WORKSPACE_NAME)

    # List all files in the specified directory
    paths = file_system_client.get_paths(path=DIRECTORY_PATH)

    # Find the CSV file in the directory
    csv_file_path = None
    for path in paths:
        if path.name.endswith('.csv', csv_file_path = path.name
            break  # Stop after finding the first CSV file

    if csv_file_path:
        print(f"Found CSV file: {csv_file_path}")
       
        # Create a file client for the specific CSV file
        file_client = file_system_client.get_file_client(csv_file_path)

        # Download the file content
        download = file_client.download_file()
        downloaded_bytes = download.readall()

        # Convert the downloaded bytes to a StringIO object for use with pandas
        csv_data = StringIO(downloaded_bytes.decode('utf-8'))

        # Load the CSV data into a pandas DataFrame
        df = pd.read_csv(csv_data)

        # Print or process the DataFrame
        print(df)
    else:
        print("No CSV file found in the directory.")
 
collinq
Super User
Super User

Hi @AslakJonhaugen ,

 

To do this you must first connect to your Lake - for Python I believe that DirectLake mode is the best but I am not an expert in this area.

 

Then, install your python scripts from here -   File > Options and settings > Options > Python scripting

 

Then, use your script to connect and do your operations.




Did I answer your question? Mark my post as a solution!

Proud to be a Datanaut!
Private message me for consulting or training needs.




Helpful resources

Announcements
November Carousel

Fabric Community Update - November 2024

Find out what's new and trending in the Fabric Community.

Live Sessions with Fabric DB

Be one of the first to start using Fabric Databases

Starting December 3, join live sessions with database experts and the Fabric product team to learn just how easy it is to get started.

Las Vegas 2025

Join us at the Microsoft Fabric Community Conference

March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount! Early Bird pricing ends December 9th.

Nov PBI Update Carousel

Power BI Monthly Update - November 2024

Check out the November 2024 Power BI update to learn about new features.