The ultimate Fabric, Power BI, SQL, and AI community-led learning event. Save €200 with code FABCOMM.
Get registeredCompete to become Power BI Data Viz World Champion! First round ends August 18th. Get started.
Hi!
The files that I have to work with have 2 column rows. Therefore, in "Transform Sample File" I transpose the 2 column rows to make each file have 1 column row. Before I do that though, I duplicate my "Transform Sample File" to be able to append the rest of the sample table later (to that 1 column row that I create). The result though looks like this:
File A has 1 column row and attached rows of File A
File B has 1 column row and attached rows of File A
File C has 1 column row and attached rows of File A
File D has 1 column row and attached rows of File A
Desired result:
File A has 1 column row and attached rows of File A
File B has 1 column row and attached rows of File B
File C has 1 column row and attached rows of File C
File D has 1 column row and attached rows of File D
How do I make it work? Thanks in advance!
Solved! Go to Solution.
Hi @Anonymous ,
I'm not sure why you're duplicating your Transform Sample File query. The whole point of the folder combine process is that Power BI performs exactly the same transformations on every file in the folder (based on your Transform Sample File steps) then appends them for you.
If you don't want to do exactly the same transformations on each file, then you should just be importing them as individual sources/queries into Power Query and transforming them separately.
Pete
Proud to be a Datanaut!
Hi @Anonymous ,
I'm not sure why you're duplicating your Transform Sample File query. The whole point of the folder combine process is that Power BI performs exactly the same transformations on every file in the folder (based on your Transform Sample File steps) then appends them for you.
If you don't want to do exactly the same transformations on each file, then you should just be importing them as individual sources/queries into Power Query and transforming them separately.
Pete
Proud to be a Datanaut!