Don't miss your chance to take the Fabric Data Engineer (DP-700) exam on us!
Learn moreWe've captured the moments from FabCon & SQLCon that everyone is talking about, and we are bringing them to the community, live and on-demand. Starts on April 14th. Register now
Hi, I'm working on an idea and it will involve combining around 1,000 excel files each with about 5,000 rows. I will then perform fuzzy match in just one column to look up some keywords which will potentially yield more tna 1 potential match (based on .40 threshold). What would be the best way to approach this considering the volume of the data? thanks!
I'm not exactly sure on how you will automate getting 1000 excel files. I guess that PBI will be very slow in getting so many different files. I would consolidate first in a single excel and load that.
Anyway 5 million rows is large but not extremely large for PBI. You can apply your formulas on a fuzzy match with a calculated column.
the files are saved in a SP library, and PBI will be used to combine them and then perform the fuzzy match. How can I apply fuzzy match using calculated columns?
If you have recently started exploring Fabric, we'd love to hear how it's going. Your feedback can help with product improvements.
A new Power BI DataViz World Championship is coming this June! Don't miss out on submitting your entry.
Share feedback directly with Fabric product managers, participate in targeted research studies and influence the Fabric roadmap.
| User | Count |
|---|---|
| 55 | |
| 35 | |
| 31 | |
| 19 | |
| 17 |
| User | Count |
|---|---|
| 75 | |
| 72 | |
| 38 | |
| 35 | |
| 25 |