Check your eligibility for this 50% exam voucher offer and join us for free live learning sessions to get prepared for Exam DP-700.
Get StartedDon't miss out! 2025 Microsoft Fabric Community Conference, March 31 - April 2, Las Vegas, Nevada. Use code MSCUST for a $150 discount. Prices go up February 11th. Register now.
Hello,
I place multiple Excel files in a Sharepoint folder, process them with the Power BI service's Data Flow, and create a Data Store for reporting. The main purpose of using the Power BI service is for automatic updates, which Power BI Desktop cannot replace it. The workspace license is Pro.
In such a situation, when reading and processing multiple Excel files in Power BI Service, the number of data has increased too much and a timeout occurs. The number of Excel files is about 800 and the total number of rows is about 1.5 million.
It is taking a long time, especially in the area of merging multiple tables. In the current is as follows.
```
previous_table = ...
output_table = Table.Combine(previous_table[table_column])
...
```
What action would allow me to finish this process within 10 minutes of the query evaluation?
Solved! Go to Solution.
The use of incremental refresh is good in many situations (not just dataflows), but merge steps can be costly for performance, especially if you end up processing all your Excel files multiple times. Do you need to do the merges? Could you have separate tables in your model instead?
While building, the tip to "Keep Top Rows" is a good one, so you only work with data from 2-3 files while building.
Also, see this article - Chris Webb's BI Blog: Excel.Workbook() And The delayTypes Option In Power Query/Power BI Chris Webb'...
Pat
Please see this video that shows how to set up incremental refresh with files stored on SharePoint.
https://www.youtube.com/watch?v=IVMdg16yBKE
Pat
Thanks @ppm1,
I think your idea would be very useful when I have a trouble while updating Data Flow in Power BI Service. However, in my case, the problem is that while developing ETL using Power Query. So the target is a bit different.
I would appreciate it if you could share another idea with me.
The use of incremental refresh is good in many situations (not just dataflows), but merge steps can be costly for performance, especially if you end up processing all your Excel files multiple times. Do you need to do the merges? Could you have separate tables in your model instead?
While building, the tip to "Keep Top Rows" is a good one, so you only work with data from 2-3 files while building.
Also, see this article - Chris Webb's BI Blog: Excel.Workbook() And The delayTypes Option In Power Query/Power BI Chris Webb'...
Pat
Thanks @MAwwad
> Use the "Partition" feature in the Power BI service to break up the data into smaller chunks. You can use the "Partition Size" option in the "Advanced" tab of the Data Store settings to specify the number of rows that should be processed at a time.
Would you mind if I ask you to share a screenshot of this option? I cannot detect this in my screen...
f you are experiencing a timeout when reading and processing multiple Excel files in the Power BI service, there are a few things you can try to improve performance:
Use the Power Query Editor in Power BI Desktop to optimize the data transformation process. You can use the "Remove Rows" and "Keep Rows" options to remove unnecessary rows, and the "Remove Columns" and "Keep Columns" options to remove unnecessary columns. You can also use the "Filter Rows" and "Sort Rows" options to reduce the amount of data that needs to be processed.
Use the "Partition" feature in the Power BI service to break up the data into smaller chunks. You can use the "Partition Size" option in the "Advanced" tab of the Data Store settings to specify the number of rows that should be processed at a time.
Use the "Index" column in Power Query to optimize the merge process. When you merge tables using the "Table.Combine" function, Power Query will automatically create an "Index" column to identify each row. You can use this column to optimize the merge process by sorting the tables by the "Index" column before merging them.
Consider using a different data source that is better suited for large data sets, such as a SQL Server database or an Azure Data Lake.
March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount! Prices go up Feb. 11th.
Check out the January 2025 Power BI update to learn about new features in Reporting, Modeling, and Data Connectivity.
User | Count |
---|---|
14 | |
13 | |
13 | |
12 | |
12 |