Starting December 3, join live sessions with database experts and the Microsoft product team to learn just how easy it is to get started
Learn moreGet certified in Microsoft Fabric—for free! For a limited time, get a free DP-600 exam voucher to use by the end of 2024. Register now
Missed the previous parts of this series? See Become your organization’s strategic advisor by using Machine Learning and Power BI
In part one, we considered how correlation plots can easily be integrated in Power BI reports to show how various factors relate to one another. That has been a useful starting point for me when looking at new data for the first time. It can also be useful for challenging old wisdoms or to perform an initial ranking of potential investments. A correlation plot however only compares two columns at a time, e.g. how correlated is revenue and inventory levels.
Unfortunately, the world is often more complicated than that. For instance, revenue and inventory may be positively correlated during the high season but negatively correlated during the other months of the year (if a lack of stock hurts sales). A simple correlation plot would not be able to show that. In this post, we will therefore look at a simple approach to detect simultaneous interactions between multiple variables.
Used correctly, this type of analysis can be a great tool for understanding a business and its underlying sensitivities. This can be incredibly valuable when deciding which investments to prioritize. There are multiple possible approaches to this type of multivariate analysis. A simple and effective one is a binary tree. It tries to explain one thing by simultaneously using all other available information. Say for instance that we are trying to explain (or predict) variations in store revenue using the month of the year, marketing investments and product promotions. The tree algorithm would attempt to group the rows into multiple subsets with the goal of minimizing the range of revenue in each subset. In other words, it tries to find subsets that have as similar values for the target column as possible.
For this example, the rows with the highest revenue may all have coincided with significant marketing activities, product promotions and be in the month of December. The tree would depict this with three or more splits on these variables. The sequence of the splits is determined by how much of the variations in revenue they explain. If December has 300% higher revenue than any other month, while marketing on average gives a 50% boost to sales, vs. 30% for product promotions then Month = December would be the first split. The second split would not however necessarily be marketing activities since for that the algorithm now only considers the data rows for December. This is the important difference to a correlation plot which always compares all rows and the tree which can discern strong correlations for a subset. In this example management, may for instance realize that product promotions only have a negligible return in December but are a good investment all other months of the year.
Below is another simple example of how trees can complement correlation plots built using the below table.
Here two factors (A and B) are used to try and predict the third column (Outcome). Individually both factors seem to have some positive impact on the outcome. However, when both factors have the value “2” the outcome is much higher. Of course, this is not enough data for statistical significance but it illustrates the point.
Using a correlation plot we can see the average 12% correlation between the outcome and Factor A, vs. 32% for Factor B.
If we instead use a tree we see a much richer explanation of the interactions. To get to the highest outcome (bottom right with an average value of 12) we read the tree from the top:
In other words, what this tree tells us is that the highest value for the column “Outcome”, our target value, is achieved when Factor B and A are both: not less than 1.5 and not greater than 2.5. Since we know from our data that both factors are discrete numbers we can simplify this further to say we observe the highest values for “Outcome” when the two factors both have value 2.
In addition to being very valuable tools for analysis; trees are a great introduction to more complex Machine Learning scenarios which we will cover in subsequent posts.
Here we see the same data as in the two previous posts. The bar chart to the left shows the all-up revenue by Product ID and is mainly used to select one product at a time (here product 6 is selected). The tree plot in the top-right shows the same data as the correlation plot below it. Its first input column is revenue and that is the number that it is trying to forecast by splitting the dataset at different thresholds for the other columns. For product 6 the best split is on seasonality (see post 2), followed by promotion. The highest revenue bucket for product 6, with an average of $35k per month, is expected when seasonality is greater than or equal to 0.95 and there is an active promotion for that product:
The second highest revenue bucket with an average of $33k per month is expected when seasonality is greater than or equal to 1.2, regardless of whether there is an active promotion:
Based on that, one conclusion could be that the return on promotions is less in the top season.
The importance of seasonality to explain revenue could in this case have been guessed by looking at the correlation plot but an insight about diminishing returns on promotions, as the annual cycle picked up, would have been missed:
The tree and correlation plots were both made using R-visuals with the same columns as input.
Using the report
For this report to work in Power BI Desktop you need the packages rpart, rpart.plot, RColorBrewer, RGtk2 and rattle installed. If you have not previously used these packages on your computer, install them by going to your R console in R Studio or other R GUI and copy/paste install.packages(c("rattle", "rpart", "rpart.plot", "RColorBrewer", "RGtk2")).
By clicking on the different products, we can examine which combination of factors are most important in determining a given product’s revenue. Using Power BI’s standard filters we can also drill-in to specific subsets of the data. If we for instance use product 6 again we see that the first split that the model does is on the column “Seasonality”.
To validate the accuracy of this we can add Seasonality to the Page level filters:
This leaves all but the leftmost branch and reduces the all-up correlation between Product 6’s revenue and Seasonality.
In a second step, we can then add Active promotion to the Page level filters pane and set that to “is greater than 0.5”. That will show NA values in the correlation plot for Active promotion since that column now only has one value, namely “1”. It might therefore be helpful to convert that visualization to a scatter plot, e.g. SandDance to explore the remaining rows of data. For small datasets, a table might be the best option. Here I converted the correlation plot to a standard table and added Revenue again, this time aggregating it as Standard deviation (by using the quick calculation option, under “Values” in the right-hand side menu - visible after clicking the down-arrow button). I also changed the aggregation form for some of the other columns, below to help me understand this subset of data.
As you can see the average of revenue is $35k, identical to the leaf in the tree that we filtered down to. From the values in that column and from the standard deviation of $16.8k we can however see that just filtering on Seasonality and Active promotion still leaves us with a lot of variance in revenue and with 14 data points we need to be cautious reading in too much in these results.
Recreating the report
R visuals will automatically work in Power BI when the report is published to the service, where the most common packages have already been installed, but to get them to work in Power BI Desktop you need a local R installation (please see my previous post for installation instructions).
To create the binary tree R visual:
Once expanded you should see the below, with space for the R script.
Copy/paste the script below into the script editor (any line that starts with # or ## is a comment only and is not executed):
## Load the required libraries ## If you have not installed these packages you will get an error message saying that the libraries were not found. In that case install them by copy/pasting: install.packages(c("rattle", "rpart", "rpart.plot", "RColorBrewer", "RGtk2")) directly in your R Console. The function c() combines values into lists in R. library(rpart) library(rattle) ## rpart, like our previous examples is sensitive to "NAs" and missing values. na.omit() deletes those rows. ## Alternatively, you can use the Power BI filter functions to filter out rows with missing values treenona <- na.omit(dataset) ## In this example I used month id to ensure I aggregated by month. However, I don't want to include it in the calculation so I delete the last column. treenona <- treenona[,1:(ncol(dataset)-1)] ## I'm trying to predict Revenue but here I'll rename it to Sales which I will later reference names(treenona)[1]<-"Sales" ## Calculate the splits to explain/predict column "Sales" ## with a minimum of 7 (21/3) observations per bucket tree.1 <- rpart(Sales ~ .,data=treenona,control=rpart.control(minsplit=21,cp=0.015)) ## Note on the above, the value for "cp" determines if a split is worthwhile, lower value leads to more splits/grannularity ## Plot the tree with text size 110% (1.1) fancyRpartPlot(tree.1, cex = 1.1, sub = NULL)
To recreate this visual for your own data the steps will be almost identical. It is important however to remember to get the granularity right.
Links and downloads
Decision trees in the R showcase on Power BI
* Third-party programs. This software enables you to obtain software applications from other sources. Those applications are offered and distributed by third parties under their own license terms. Microsoft is not developing, distributing or licensing those applications to you, but instead, as a convenience, enables you to use this software to obtain those applications directly from the application providers.
By using the software, you acknowledge and agree that you are obtaining the applications directly from the third-party providers and under separate license terms, and that it is your responsibility to locate, understand and comply with those license terms. Microsoft grants you no license rights for third-party software or applications that is obtained using this software.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.