Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and more. Get started
Hello, I'm struggling to understand the utility of Power BI dataflows over using a "golden" dataset (promoted/shared for use in multiple reports). I haven't found a post that pitted these two against each other in this way yet.
Assume that I have an average-quality data-warehouse function in my org. Assume also that I don't care about sharing prepared data with other parts of the Power Platform; I'm only interested in optimal Power BI architecture and implementation.
Here are my thoughts. Please feel free to correct me where I'm wrong.
All guidance is greatly appreciated.
Solved! Go to Solution.
Here's my biased opinion: Dataflows are glorified CSV / Parquet files. They can be useful when you have a slow, badly performing data source and you want to shield your developers from that. The dataflow data can be spooled efficiently into a dataset.
A golden dataset is not only a bunch of tables, it also represents the data model (although that can be circumvented). If well maintained (with some incremental refresh and selective partition refresh sprinkled in) it will be a great starting point for a unified company wide data model. Compared to the professional data stewardship tools it is more of a toy, though.
Golden datasets come with their own baggage - specifically all the access issues around dataset chaining, but also design limitations in composite data models.
Dataflows do nothing for Power BI report consumers. To create a pleasant experience for report users you should favor (import mode) datasets over dataflows, and accept the added cost for developers.
Recently someone asked if they should use an incremental dataset on top of an incremental dataflow. That's on the same level as Direct Query against dataflows.
Here's my biased opinion: Dataflows are glorified CSV / Parquet files. They can be useful when you have a slow, badly performing data source and you want to shield your developers from that. The dataflow data can be spooled efficiently into a dataset.
A golden dataset is not only a bunch of tables, it also represents the data model (although that can be circumvented). If well maintained (with some incremental refresh and selective partition refresh sprinkled in) it will be a great starting point for a unified company wide data model. Compared to the professional data stewardship tools it is more of a toy, though.
Golden datasets come with their own baggage - specifically all the access issues around dataset chaining, but also design limitations in composite data models.
Dataflows do nothing for Power BI report consumers. To create a pleasant experience for report users you should favor (import mode) datasets over dataflows, and accept the added cost for developers.
Recently someone asked if they should use an incremental dataset on top of an incremental dataflow. That's on the same level as Direct Query against dataflows.
Thanks for sharing your perspective!
Check out the September 2024 Power BI update to learn about new features.
Learn from experts, get hands-on experience, and win awesome prizes.
User | Count |
---|---|
90 | |
46 | |
25 | |
21 | |
19 |