Perform a Relative Weights Analysis in R

**Relative Weights Analysis (RWA)** is a method of calculating relative importance of predictor variables in contributing to an outcome variable. The method implemented by this function is based on Toniandel and LeBreton (2015), but the origin of this specific approach can be traced back to Johnson (2000), *A Heuristic Method for Estimating the Relative Weight of Predictor Variables in Multiple Regression*. Broadly speaking, RWA belongs to a family of techiques under the broad umbrella ‘Relative Importance Analysis’, where other members include the ‘Shapley method’ and ‘dominance analysis’. This is often referred to as ‘Key Drivers Analysis’ within market research.

This package is built around the main function `rwa()`

, which takes in a data frame as an argument and allows you to specify the names of the input variables and the outcome variable as arguments.

The `rwa()`

function in this package is compatible with dplyr / tidyverse style of piping operations to enable cleaner and more readable code.

You can install the stable CRAN version of **rwa** with:

`install.packages("rwa")`

Alternatively, you can install the latest development version from GitHub with:

```
install.packages("devtools")
devtools::install_github("martinctc/rwa")
```

RWA decomposes the total variance predicted in a regression model (R2) into weights that accurately reﬂect the proportional contribution of the various predictor variables.

RWA is a useful technique to calculate the relative importance of predictors (independent variables) when independent variables are correlated to each other. It is an alternative to multiple regression technique and it addresses the *multicollinearity* problem, and also helps to calculate the importance rank of variables. It helps to answer “which variable is the most important and rank variables based on their contribution to R-Square”.

See https://link.springer.com/content/pdf/10.1007%2Fs10869-014-9351-z.pdf.

When independent variables are correlated, it is difficult to determine the correct prediction power of each variable. Hence, it is difficult to rank them as we are unable to estimate coefficients correctly. Statistically, multicollinearity can increase the standard error of the coefficient estimates and make the estimates very sensitive to minor changes in the model. It means the coefficients are biased and difficult to interpret.

Key Drivers Analysis methods do not conventionally include a score sign, which can make it difficult to interpret whether a variable is *positively* or *negatively* driving the outcome. The `applysigns`

argument in `rwa::rwa()`

, when set to `TRUE`

, allows the application of positive or negative signs to the driver scores to match the signs of the corresponding linear regression coefficients from the model. This feature mimics the solution used in the Q research software. The resulting column is labelled `Sign.Rescaled.RelWeight`

to distinguish from the unsigned column.

As Tonidandel et al. (2009) noted, there is no default procedure for determining the statistical significance of individual relative weights:

The difficulty in determining the statistical significance of relative weights stems from the fact that the exact (or small sample) sampling distribution of relative weights is unknown.

The paper itself suggests a Monte Carlo method for estimating the statistical significance, but this is currently not available or provided in the package, but the plan is to implement this in the near future.

You can pass the raw data directly into `rwa()`

, without having to first compute a correlation matrix. The below example is with `mtcars`

.

Code:

```
library(rwa)
library(tidyverse)
mtcars %>%
rwa(outcome = "mpg",
predictors = c("cyl", "disp", "hp", "gear"),
applysigns = TRUE)
```

Results:

```
$predictors
[1] "cyl" "disp" "hp" "gear"
$rsquare
[1] 0.7791896
$result
Variables Raw.RelWeight Rescaled.RelWeight Sign Sign.Rescaled.RelWeight
1 cyl 0.2284797 29.32274 - -29.32274
2 disp 0.2221469 28.50999 - -28.50999
3 hp 0.2321744 29.79691 - -29.79691
4 gear 0.0963886 12.37037 + 12.37037
$n
[1] 32
```

The main `rwa()`

function is ready-to-use, but the intent is to develop additional functions for this package which supplement the use of this function, such as tidying outputs and visualisations.

Please feel free to submit suggestions and report bugs: https://github.com/martinctc/rwa/issues

Also check out my website for my other work and packages.

Azen, R., & Budescu, D. V. (2003). The dominance analysis approach for comparing predictors in multiple regression. *Psychological methods*, 8(2), 129.

Budescu, D. V. (1993). Dominance analysis: a new approach to the problem of relative importance of predictors in multiple regression. *Psychological bulletin*, 114(3), 542.

Grömping, U. (2006). Relative importance for linear regression in R: the package relaimpo. *Journal of statistical software*, 17(1), 1-27.

Grömping, U. (2009). Variable importance assessment in regression: linear regression versus random forest. *The American Statistician*, 63(4), 308-319.

Johnson, J. W., & LeBreton, J. M. (2004). History and use of relative importance indices in organizational research. *Organizational research methods*, 7(3), 238-257.

Lindeman RH, Merenda PF, Gold RZ (1980). *Introduction to Bivariate and Multivariate Analysis*. Scott, Foresman, Glenview, IL.

Tonidandel, S., & LeBreton, J. M. (2011). Relative importance analysis: A useful supplement to regression analysis. *Journal of Business and Psychology*, 26(1), 1-9.

Tonidandel, S., LeBreton, J. M., & Johnson, J. W. (2009). Determining the statistical significance of relative weights. *Psychological methods*, *14*(4), 387.

Wang, X., Duverger, P., Bansal, H. S. (2013). Bayesian Inference of Predictors Relative Importance in Linear Regression Model using Dominance Hierarchies. *International Journal of Pure and Applied Mathematics*, Vol. 88, No. 3, 321-339.

Also see Kovalyshyn for a similar implementation but in Javascript.