Power BI is turning 10, and we’re marking the occasion with a special community challenge. Use your creativity to tell a story, uncover trends, or highlight something unexpected.
Get startedJoin us at FabCon Vienna from September 15-18, 2025, for the ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM. Get registered
Hello,
I'm working on a pipeline with a Notebook (Python) and a Dataflow Gen2, but I'm struggeling with a problem.
Here is the basic setup:
1. I've setup a Python-Notebook that runs every night an fetches data for the previous day from a web api. This data is in a flat format and consists of dim- and fact-data in one row.
2. The daily data is written to a Lakehouse table "RawData", the mode is "overwrite" to ensure, that only the data from the previous day is being stored in raw format and not all data from past days.
3. I've also setup a Dataflow Gen2 that reads the data from the table "RawData" and separates it in dim data and fact data.
4. Dim data is written to a table "DimData" in the Lakehouse, fact data is written to a table "FactData".
5. The update method in the dataflow is set to "Append".
Here are the results:
1. The daily data is written to table "RawData". RawData contains only the data from the previous day. This is the expected behaviour and works fine.
2. The Dataflow separates the RawData in dim data and fact data and writes this data to separate Lakehouse tables "DimData" and "FactData". This works in principle, but not as expected (see 3).
3. The daily data should be appended to "DimData" and "FactData", so that DimData and FactData contain the date from the previous day as well as existing data from past days. This doesn't work. Instead existing data in DimData and FactData is ovewritten, so that both tables contain only data from the preivious day.
I hope this was understandable. Does anyone have any ideas whats going wrong?
Thanks,
Udo
Solved! Go to Solution.
Hi Ibendlin and Fabian,
Thanks for offering your help.
It appears that the dataflow was corrupt. I have completely rebuilt the process and now the appending of data via append seems to work.
Hi Ibendlin and Fabian,
Thanks for offering your help.
It appears that the dataflow was corrupt. I have completely rebuilt the process and now the appending of data via append seems to work.
You should only append items to the DimData table that are not already there. Otherwise you end up with primary key violations.
Can you share some screenshots of the dataflow? The steps seem correct. We may see something in the settings.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.