March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount! Early bird discount ends December 31.
Register NowBe one of the first to start using Fabric Databases. View on-demand sessions with database experts and the Microsoft product team to learn just how easy it is to get started. Watch now
Hi,
I have a database with many columns some of which have spurious values 'na', 'n/a', 'tbd'.
Unfortunately the entries can't be validated and caught at the time of entry.
To find and resolve these entries my current plan is to concatate all columns into a single string with a delimiter e.g.
|Orange|123.23|Banana||12|Apple|na|Grape|n/a|Peach|tbd|
Then, a conditional column would be used to search for the relevant string |n/a|, |na|, |tbd|
Although this would work it seems innefficient, especially since it would require ongoing maintenance as further columns were added.
Can anyone advise on a better approach to find the rows where the incorrect entries have occurred?
Thanks,
Richard
Solved! Go to Solution.
@RichardJ Oh, thank you for clarifying, that does make sense now. For that, I would then suggest you unpivot your data in Power Query, so you have all your field names in an Attribute column and all your data in a Value column. Then you can do your counts on Value and show by column in a report.
Respectfully,
Zoe Douglas (DataZoe)
Follow me on LinkedIn at https://www.linkedin.com/in/zoedouglas-data
See my reports and blog at https://www.datazoepowerbi.com/
Hi @DataZoe,
Thanks for the reply. I should have been clearer in my original question.
I don't wish to clean up the data in Power BI but would prefer if the users of the original database cleaned up their entries with useful, meaningful data rather than the n/a, na, tbd generic statements.
I was hoping to find a way to scan all columns in the table at once for those values and identify which rows/records should be visible to the users/Data Quality Manager in a Power BI Dashboard so that the records could be fixed by the people who didnt enter the correct info originally.
Hope that makes sense.
Thanks,
Richard
@RichardJ Oh, thank you for clarifying, that does make sense now. For that, I would then suggest you unpivot your data in Power Query, so you have all your field names in an Attribute column and all your data in a Value column. Then you can do your counts on Value and show by column in a report.
Respectfully,
Zoe Douglas (DataZoe)
Follow me on LinkedIn at https://www.linkedin.com/in/zoedouglas-data
See my reports and blog at https://www.datazoepowerbi.com/
Very nice solution @DataZoe . This is far more efficient than what I was thinking. Fabulous.
@RichardJ In the transform data of the power bi file you can select multiple columns (or all of them) in a table and then replace values, changing them to null or another value of your choosing.
Respectfully,
Zoe Douglas (DataZoe)
Follow me on LinkedIn at https://www.linkedin.com/in/zoedouglas-data
See my reports and blog at https://www.datazoepowerbi.com/
March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount!
Your insights matter. That’s why we created a quick survey to learn about your experience finding answers to technical questions.
Arun Ulag shares exciting details about the Microsoft Fabric Conference 2025, which will be held in Las Vegas, NV.
User | Count |
---|---|
124 | |
89 | |
84 | |
70 | |
51 |
User | Count |
---|---|
206 | |
143 | |
97 | |
79 | |
68 |