Starting December 3, join live sessions with database experts and the Microsoft product team to learn just how easy it is to get started
Learn moreGet certified in Microsoft Fabric—for free! For a limited time, get a free DP-600 exam voucher to use by the end of 2024. Register now
Hello Power BI Community!
I don't know that I see many topics here looking for a discussion, but that's what I'm hoping for here. I work for a university that has been trying to adopt Power BI as an enterprise solution. We're going for a mostly self-service model, but we do have some data that we expect to model within IT and deliver to users. If context helps, maybe think about a centralized HR dataset within an IT workspace and we publish a report with RLS to another workspace that others have access to. I have a few questions I'd like to hear others' opinions on:
The approach you described makes sense. Creating golden datasets is a good way to achieve a self-service model (see article below). Dataflows are an excellent tool for harvesting and transforming data, and offer numerous advantages including reusability. Keep in mind that not all connectors in PBI Desktop are available in PBI Service, and in some cases the PBI Desktop connectors are more robust than those in PBI Service (e.g., in the PBI Service Snowflake connector, there's no gear icon to edit custom SQL). The workaround is to create the M code in PBI Desktop, copy it, and paste it into a blank query in PBI Service.
It's recommended to do report development in PBI Desktop. You can develop reports in PBI Service (based on a published dataset), but if the dataset gets deleted, reports created in PBI Service become permanently broken (see article below). PBI Service is a great option for ad hoc reports, but enterprise reports should be developed in PBI Desktop.
Row-level security can be implemented in various ways. The article below describes how to implement it dynamically, with a solution to the many-to-many relationship that may exist.
These are great discussion points, and I look forward to hearing from the community. 🙂
https://exceleratorbi.com.au/new-power-bi-reports-golden-dataset/
https://nisorness.com/data/2020/11/19/how-to-prevent-the-deletion-of-datasets-in-power-bi
https://radacad.com/dynamic-row-level-security-with-profiles-and-users-in-power-bi
Proud to be a Super User!
Hello! Thanks for your thoughtful and detailed reply.
The linked article on golden datasets does seem to be the route I would like to adopt. Long-term I think having datasets that users can re-use where we already have the tables modeled correctly and some well-formulated measures premade is a great way to ensure data is reported correctly across departments.
One thing I am struggling with in adopting this model is permissions. Given the size of our university, we could have hundreds to thousands of employees trying to access reports/dashboards/apps based on these underlying datasets. What is the best practice is setting up these permissions? I had thought it would be to:
Is the only way to ensure access to reports is to allow access to all? Is there an easy way to do this without opening the workspace up to everyone?
Having a golden workspace and multiple workspaces for functional teams is the approach I use. You'll need to grant Build permission to users so they can access datasets in the golden workspace.
https://docs.microsoft.com/en-US/power-bi/connect-data/service-datasets-build-permissions
Proud to be a Super User!
Starting December 3, join live sessions with database experts and the Fabric product team to learn just how easy it is to get started.
March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount! Early Bird pricing ends December 9th.
User | Count |
---|---|
93 | |
87 | |
85 | |
70 | |
49 |
User | Count |
---|---|
141 | |
118 | |
112 | |
59 | |
59 |