Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Score big with last-minute savings on the final tickets to FabCon Vienna. Secure your discount

Reply
Fleeups
New Member

How to optimize big data load

So, here's the thing. I had a database that's over 289k rows from Google Big Query. 

 

so, I've tried to split this query in 2:

 

Sentimentos_old and Sentimentos.

 

Sentimentos_old are pulling data from before 2022-11-01, and I've locked it from refreshing, it returned 287k rows.

 

Table Sentimentos pulls data after this date, and combine the old ones.

 

I was expecting to instantly had the 287 rows from Sentimentos_old (that should be static from now, since it was not refreshing), and only refresh the 2k rows from Sentimentos, but that didn't happened, it still take so long to refresh, just like if it was on a unique query.

 

Is there a way to do something like this? 

1 ACCEPTED SOLUTION
ppm1
Solution Sage
Solution Sage

Incremental refresh is the way to do this. Please see this documentation.

Incremental refresh for datasets and real-time data in Power BI - Power BI | Microsoft Learn

 

Pat

Microsoft Employee

View solution in original post

4 REPLIES 4
djurecicK2
Super User
Super User

@Fleeups , Yes, it is. 

djurecicK2
Super User
Super User

Agree with incremental refresh with large data, but 289k rows does not seem that big- maybe something else is going on? Can you reduce the number of columns you are importing?

I've read somewhere that some merges can really slow down the refresh, and I've got a merge that should be slowing it down. 

 

Every row on that table refers to comments on instagram posts. I have another table that had the columns campaign ID and key words to map comments that has been adherent to the campaign.

 

So, I use a merge (to bring only the right words to each campaign), that result on a column with tables. Turn that tables into a list, like that:

 

Fleeups_0-1671477196245.png

 

then I create a new column to check if there any of that words on the list match with the comment column, if there's a match returns 1, else 0, so I can just sum this column as relevant comments, and calculate adherence rate. here's the formula:

 

= Table.AddColumn(#"Added Custom", "Comentários Aderentes", each if
List.AnyTrue(List.Transform([Custom],
(x) => Text.Contains([texto_sem_acento], x, Comparer.OrdinalIgnoreCase))) = true
then 1 else 0)

 

maybe all this process are slowing down the refresh? 

ppm1
Solution Sage
Solution Sage

Incremental refresh is the way to do this. Please see this documentation.

Incremental refresh for datasets and real-time data in Power BI - Power BI | Microsoft Learn

 

Pat

Microsoft Employee

Helpful resources

Announcements
August Power BI Update Carousel

Power BI Monthly Update - August 2025

Check out the August 2025 Power BI update to learn about new features.

August 2025 community update carousel

Fabric Community Update - August 2025

Find out what's new and trending in the Fabric community.

Top Solution Authors