Check your eligibility for this 50% exam voucher offer and join us for free live learning sessions to get prepared for Exam DP-700.
Get StartedDon't miss out! 2025 Microsoft Fabric Community Conference, March 31 - April 2, Las Vegas, Nevada. Use code MSCUST for a $150 discount. Prices go up February 11th. Register now.
Hi There,
I am building a pipeline that copies some tables from one Database and plonks it into an Azurew SQL Database, in which I can then save the tables into a Lakehouse as Parquet files. My pipeline is currently failing, with the following error code at the Copy Each Table point:
After investigating, it turns out that a new column has been added to a table and the column contains all Null values. I have set the pipeline up to be dynamic (in the sense that regardless of table changes, it should still work).How would I go about excluding columns that have all null values?
Any help would be really appreciated!
Just to confirm,
Hi @AndyDDC , thanks for responding!
In answer to your bullet points:
Thanks.
So this is just my opinion on this, but you could write from the source database directly to the Lakehouse Files section (your raw/bronze layer). That would mean that any new columns/altered columns in the source will just get automatically written to Parquet files in the lakehouse files section.
The issue as you've discovered is that there is no auto-update on the azure sql database table schema, you'd have to either:
You could go one step further and just write the source data to the lakehouse tables and use Overwrite (although it does "overwrite" the data in the destination table, although keeps history using delta versioning).
Hi @heastham ,
We haven’t heard from you on the last response and was just checking back to see if your query got resolved. Please let us know if you have any further queries.
Hi @v-nikhilan-msft - I've been frozen out of the fatory for the past few days due to 'capacity' issues. We don't have nearly enough data for this to even be a problem yet so I'm logging a ticket currently with Microsoft to see what the issue is. Once I'm successfully back in then I can attempt the solutions given above. Thanks!
Just a thought, but if you use the Overwrite option when writing to a Lakehouse table (not Files as the schema could evolve anyway if you were just writing parquet files) then that will evolve the schema of the lakehouse table automatically when source columns get added to the source.
Not sure if that could help in your situation? Overwrite would replace the data downstream with the latest incoming so not sure if that's what you want
March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount!
Arun Ulag shares exciting details about the Microsoft Fabric Conference 2025, which will be held in Las Vegas, NV.
User | Count |
---|---|
6 | |
2 | |
2 | |
1 | |
1 |
User | Count |
---|---|
13 | |
3 | |
3 | |
2 | |
2 |