Power BI is turning 10! Tune in for a special live episode on July 24 with behind-the-scenes stories, product evolution highlights, and a sneak peek at what’s in store for the future.
Save the dateEnhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends August 31st. Request your voucher.
Hi,
I've implemented a dataflow gen2 to perform some transformations respect to source data and then copying them in the destination.
Moreover, I've created a data pipeline to use this dataflow but I've noticed and tried that when running the pipeline the destination is overwritten.
Now, does a dataflow allow to write data in the destination in append mode and not in overwrite mode, please?
I don't understand because a such option doesn't exist. Thanks
Solved! Go to Solution.
If you turn off the automatic settings,
you can use append function rather than replace in dataflow gnen 2
If you turn off the automatic settings,
you can use append function rather than replace in dataflow gnen 2
@NandanHegde In case of Append, if we are using a source like Sharepoint Folder or otherwise and want to bring the incremental changes only, the data is getting duplicated in Lakehouse. Is there a way to get around this.
If you want to append only new records, you can achieve that by following the pattern in these docs https://learn.microsoft.com/en-us/fabric/data-factory/tutorial-setup-incremental-refresh-with-datafl...
If you want to do Upsert, then I think you need to use Notebook. It is shown code examples for Upsert in this tutorial: https://microsoftlearning.github.io/mslearn-fabric/Instructions/Labs/03b-medallion-lakehouse.html
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
Check out the June 2025 Fabric update to learn about new features.