RWA Web: A Free, Comprehensive, Web-Based, and User-Friendly Tool for Relative Weight Analyses
- 5.4k Downloads
- 39 Citations
Abstract
Over the last 15 years, a number of methodological developments have enabled researchers to draw more accurate inferences concerning the relative contribution (i.e., relative importance) among multiple (often correlated) predictor variables in a regression analysis. One such development has been relative weight analysis (RWA). Researchers can use a RWA to decompose the total variance predicted in a regression model (R^{2}) into weights that accurately reflect the proportional contribution of the various predictor variables. Prior to RWA, researchers were forced to rely on traditional statistics (e.g., correlations; standardized regression weights), which are known to yield faulty or misleading information concerning variable importance (especially when predictor variables are correlated with one another, which is often the case in organizational research). Although there has been a surge of interest in RWA over the last 10 years, integration of this statistical tool into organizational research has been hampered by the lack of a user-friendly statistical package for implementing RWA. Indeed, most popular statistical packages (e.g., SPSS, SAS) have yet to include RWA protocols into their regression modules. The purpose of this paper is to present a new, free, comprehensive, web-based, user-friendly resource, RWA-Web, which may be used by anyone having simple access to the internet. Our paper is structured as a tutorial on using RWA-Web to examine relative importance in the classic multiple regression model, the multivariate multiple regression model, and the logistic regression model. We also illustrate how RWA-Web may be used to conduct null hypothesis significance tests using advanced bootstrapping procedures.
Keywords
Relative weight analysis Relative importance analysis Multiple regression Predictor importanceMultiple regression is perhaps one of the most widely used analytical procedures in the organizational sciences. Although many researchers use multiple regression to identify the set of variables that will maximize the prediction of some outcome variable or criterion, another important use of multiple regression is for theory testing and explanatory purposes. In these instances, researchers are most interested in understanding how each variable contributes toward the variance explained in the outcome or criterion. Alas, it is widely recognized (Darlington 1968) that multicollinearity makes the partitioning of variance among multiple correlated predictors difficult. Despite these difficulties, researchers continue to rely on commonly available indices (e.g., simple correlation coefficients, partial correlation coefficients, standardized regression coefficients) when partitioning the variance in the criterion that is predicted by multiple (typically correlated) predictors.
Fortunately, two alternative techniques, dominance analysis (Budescu 1993) and relative weight analysis (Johnson 2000), have been developed that permit more accurate partitioning of variance among correlated predictors. Simulation research clearly shows that these measures of importance perform quite well across a variety of conditions and are recommended for this purpose (LeBreton et al. 2004b). Despite calls advocating for the wider use of these indices (Tonidandel and LeBreton 2011), researchers seem reluctant to do so. One limiting factor for many researchers might lie in the practicalities associated with obtaining these metrics for their own datasets. The output typically generated by virtually all commercial statistical software fails to routinely include either of these indices in their standard output. Although some researchers have created macros and other similar variants that will perform these calculations, these resources rely on disparate statistical packages and fail to offer a simple, easy to use, integrated solution that would be attractive to most users. To that end, the purpose of this paper is to introduce a free, comprehensive, web-based, user-friendly resource to conduct relative weight analyses. In what follows, we will provide a brief overview of relative importance analyses, provide some examples from the published literature that makes use of these techniques, and introduce our comprehensive web-based solution for performing these calculations.
Relative Importance Analysis
Within the context of linear regression, the term relative importance refers to the proportionate contribution each predictor variable makes to the total predicted criterion variance taking into account a variable’s contribution by itself and in combination with other predictor variables (Johnson and LeBreton 2004). This definition of variable importance is not to be confused with other commonly used connotations of importance such as statistical or practical significance. Information regarding a variable’s contribution to predictable variance is helpful when considering the practical utility of a variable, but aspects of the particular situation must also be considered to fully gauge practical importance (Cortina and Landis 2009). In some situations, a variable may explain only a small proportion of predictable variance and yet be very meaningful (Martell et al. 1996), whereas in other situations, a variable may account for a larger percentage of the variance but may provide little practical utility (Cortina and Landis 2009).
Relative Weight Analysis
Relative weight analysis (RWA; Johnson 2000) addresses the problem caused by correlated predictors by using a variable transformation approach to create a set of new predictors (Z_{Xk}) that are the maximally related to the original predictors (X_{j}) but are orthogonal to one another. Because these new transformed predictors are uncorrelated to one another, the criterion variable (Y) can be regressed onto this new set of predictors producing a series of standardized regression coefficients (β_{k}). Because these coefficients are generated using the orthogonal transformations of the original predictors, they no longer suffer from problems associated with collinearity. These regression coefficients are then rescaled back to the original variables by combining them with the standardized regression coefficients (λ_{jk}) obtained by regressing the original predictors (X_{j}) on their orthogonal counterparts (Z_{Xk}) producing an estimate of relative importance for each predictor variable.
The interested reader is directed to Johnson (2000), Johnson and LeBreton (2004), LeBreton and Tonidandel (2008), and Tonidandel and LeBreton (2011) for a more detailed discussion of relative weight analysis. The relative weight statistic has been shown to provide extremely good estimates of the relative importance of predictor variables when those predictor variables are correlated. This has been found in both simulation studies (LeBreton et al. 2004b) and primary studies (LeBreton et al. 2004a; LeBreton et al. 2007). Recently, new developments in relative weight analysis have expanded its applications from traditional multiple regression to more complicated regression models such as multivariate multiple regression (LeBreton and Tonidandel 2008), regression models containing higher order terms such as cross-product terms, quadratic terms, or other polynomial terms (LeBreton et al. 2013), logistic regression (Tonidandel and LeBreton 2010), and multivariate analysis of variance (MANOVA; Tonidandel and LeBreton 2013). In addition to obtaining the relative weights themselves, more useful information can be gleaned by supplementing the weights with confidence intervals, testing the statistical significance of the weights, and evaluating the statistical significance of the difference between weights either within or between samples (Johnson 2004; Tonidandel et al. 2009).
Applications of Relative Weight Analysis
Relative weight analysis has been used across a variety of domains in the organizational sciences literature. For example, Lopina et al. (2012) used relative weight analysis to examine the relative importance of individual difference variables for predicting turnover. In a related vein, Major et al. (2013) used relative weight analysis to understand how men and women differ in terms of the most important predictors of organizational commitment. Turning to leadership, Braddy et al. (2013) examined the relative importance of different leader behaviors for predicting derailment, whereas Snell et al. (2013) used a similar approach to understand how political skill predicts managerial effectiveness. Concerning job performance, Dalal et al. (2012) examined the relative importance of employee engagement, personality, and job attitudes as predictors of performance, and Gentry et al. (2013) used relative weight analysis to understand how different character virtues might be more or less important for managers at different levels of an organization. Scherer et al. (2013) used relative weight analysis to examine the contributions of sub-clinical psychopathy and the global traits of the Five Factor Model in the prediction of counter-productive workplace behaviors. As a final example, Lievens and Peeters (2008) used relative weight analysis to understand how different impression management tactics are related to evaluations of interviewees. As is evident from these examples, relative weight analysis is applicable to many different areas of research.
RWA-Web
To assist those wishing to perform relative weight analysis in their own work, we endeavored to create a more general solution that would integrate the disparate types of relative importance calculations within a single easy-to-use framework that would be available to all researchers regardless of their preferred platform or fluency with a particular platform. To that end, we created an interactive website that performs all of these calculations using the free, open-source R statistical package as its backbone. The website is designed such that users who are familiar and comfortable with R can use the website to generate the R code necessary to conduct the various RWAs. These users can simply download the code and execute it locally on their workstation.
However, many individuals may not be familiar with R (and may not want to learn R). Consequently, we have also designed the website so that users may simply point-and-click through the website to generate the R syntax and then submit the syntax and a data file to our R web server. This server will then run the desired analyses and send the results to the user via e-mail. Thus, users can successfully conduct these relative importance analyses with no prior knowledge of R or experience using R.
Visiting the Website and Selecting a RWA Model
This website also contains a link to the data file used for the illustrative examples in the current paper. Readers wanting to confirm their understanding of relative weight analysis and the RWA-Web tool are encouraged to replicate our findings using these data. RWA-Web was designed with three key features in mind: (1) ease of use, (2) inclusiveness, and (3) flexibility. Below, we briefly comment on these features before presenting a tutorial on using RWA-Web including the interpretation of results obtained using several illustrative datasets.
Ease of Use
The interface of each page is designed as an editable web form where users can select certain options by choosing various check boxes and provide additional detailed information in text boxes. For each type of analysis (multiple regression, multivariate regression, and logistic regression), the user only needs to provide three basic pieces of information in order to obtain results: the name and location of the data file, the name of the criterion variable(s), and the name of the predictor variables (note that R is case sensitive). The default file format for the data file is a comma-separated values (.csv) file, which is a standard exporting option in Excel and most other statistical software packages. With these minimal pieces of information, the website will generate raw and rescaled relative weights for each predictor (scaled as the percentage of explainable variance), the bootstrapped confidence intervals around these weights, and the corresponding tests for significance of these weights.
While disparate resources exist for calculating relative weights, this is the first platform to integrate the relative weight results with confidence intervals and tests of statistical significance. Moreover, the only other program currently available for calculating confidence intervals does not rely on the bias corrected accelerated method of bootstrapping, which was recommended by Tonidandel et al. (2009) after extensive simulation work.
The ability to easily obtain confidence intervals and tests of significance will hopefully improve the accessibility of relative importance statistics within the organizational sciences. Without having this information readily available, researchers have tended to rely on the flawed practice of merely looking at the rank ordering of predictor importance without taking into account the sampling distributions of these statistics (Tonidandel and LeBreton 2011). By focusing solely on rank order, one might incorrectly conclude a predictor to be more important than another predictor despite a considerable degree of overlap in their respective sampling distributions. Thus, by having the additional information of confidence intervals and tests of significance readily available, one can make more informed judgments regarding the true contribution of predictors to explainable variance.
Inclusiveness: Built-in Comprehensive Functionality
In addition to integrating the relative weights, confidence intervals, and tests of statistical significance, within a single platform, the website also possesses crucial additional features to make it a truly comprehensive resource for these analyses. Frequently, in addition to wanting to determine the relative importance of various predictors, researchers often have hypotheses that might predict that two predictors differ significantly from one another in terms of their relative importance or that a particular predictor’s relative importance differs in two different samples. To address either of these questions, one needs to be able to evaluate the statistical significance of the difference between relative weights either within or between samples. This new resource can perform both of these types of tests using bias corrected accelerated methods. To test for the statistical significance of the difference between relative weights within a sample, a user only needs to supply the name of a predictor variable, and the relative weight for that predictor will be tested against all of the other predictors in the model. To evaluate the statistical significance between samples, one simply provides the name of the grouping variable that identifies the different samples within the dataset and the two levels of that grouping variable one wishes to compare. The program will then automatically generate tests for significance for all of the predictors in the model between the two samples.
Flexibility
We have designed the website in an effort to provide a maximum amount of flexibility to accommodate the needs of various users. To that end, the point-and-click interface allows users to select from a variety of options including: conducting relative importance analyses on both raw data or on a correlation matrix, choosing the alpha value for the confidence intervals and tests of statistical significance as well as the number of bootstrapped replications they wish to use for those analyses, and electing different techniques for handling missing data should they have any. Finally, individuals comfortable with the R statistical program can elect to download the R code and can then modify it to meet their own personal needs.
Illustrative Example—Standard Multiple Linear Regression
To illustrate the features of RWA-Web, we will use a simulated dataset containing a sample of 604 employees. The employees completed a climate survey measuring: Customer Orientation (custom: higher scores = more customer orientation), Team Orientation (team: higher scores = greater emphasis on teamwork), Diversity Orientation (diverse: higher scores = greater emphasis on promoting ethnic and gender diversity in the workplace), Safety Orientation (safety: higher scores = greater emphasis on maintaining a safe work environment), and Communication (comm: higher scores = more effective communication from the organization’s leaders to the employees). Three months later, the same employees completed a job attitude survey measuring: Organizational Commitment (commit: higher scores = stronger commitment to the organization), Job Satisfaction (jobsat: higher scores = higher overall levels of job satisfaction), and Leader Effectiveness (ldr_eff: higher scores = higher overall perceptions of effectiveness). Finally, one year after the initial survey, data were collected concerning voluntary turnover from the organization (turnover: 0 = employee remained, 1 = employee attrited).
Using RWA-Web to Estimate Relative Weights
Using these data, we sought to examine the relative importance of the climate variables in predicting Job Satisfaction, Organizational Commitment, and Leader Effectiveness. Specifically, we wanted to estimate relative weights (Johnson 2000) along with bootstrapped standard errors for use with significance testing (Tonidandel et al. 2009).
- (1)
Rather than downloading the R code and running it locally, we opted to “Send to R Server.”
- (2)
Consequently, we needed to specify a dataset that would be sent to the R Server along with the R Code that we would generate in subsequent steps. Our data were originally saved as an SPSS “.sav” data file. However, we converted this file into a “.csv” file by opening the data file in SPSS, selecting “File” and “Save as” from the drop down menus. In the new dialog box, we selected “Comma delimited (*.csv)” under the “Save as type” options.
- (3)
Because we were sending the R code and data to the R server, we also entered an e-mail address to which the results could be sent. We noted that the data were “Raw data” and that we wanted to delete data using “Pairwise deletion,” even though the dataset had complete data.
- (4)
Next, we specified the first relative weight analysis. We entered “jobsat” as our criterion variable and entered the names of our climate variables (i.e., custom, team, diverse, safety, comm) as our predictor variables (hitting “enter” or “return” after each variable was entered in the list).
- (5)
We retained the default bootstrapping options (k = 10,000 replications; alpha = 0.05). We also requested a comparison of whether the relative importance of the “safety” variable differed significantly from the remaining predictor variables. That is, we sought to test a hypothesis that the “safety” variable was statistically larger than all of the remaining predictor variables.
- (6)
We also requested an analysis comparing whether the relative weights differed across the “gender” variable.
- (7)
After making these selections, we clicked “submit” and waited for the e-mail from the R server. We received the e-mail about 10 min later.
- (8)
Finally, we repeated steps 1–6 for both the “commit” and “ldr_eff” outcome variables.
Interpreting the Output from RWA-Web
RWA-Web: results for a traditional relative weight analysis (criterion = job satisfaction)
> #R-squared For the Model | ||
> RSQ.Results | ||
[1] 0.49230515 | ||
> | ||
> | ||
> #The Raw and Rescaled Weights | ||
> RW.Results | ||
Variables | Raw.RelWeight | Rescaled.RelWeight |
1 custom | 0.12347889 | 25.081778 |
2 team | 0.10222762 | 20.765092 |
3 diverse | 0.04667241 | 9.480383 |
4 safety | 0.04106346 | 8.341059 |
5 comm | 0.17886277 | 36.331688 |
> #BCa Confidence Intervals around the raw weights | ||
CI.Results | ||
Variables | CI.Lower.Bound | CI.Upper.Bound |
1 custom | 0.09466013 | 0.15414487 |
2 team | 0.070329043 | 0.13733754 |
3 diverse | 0.02738673 | 0.06997227 |
4 safety | 0.02404133 | 0.06291130 |
5 comm | 0.13516420 | 0.22519974 |
> #BCa Confidence Interval Tests of significance | ||
> #If Zero is not included, Weight is Significant | ||
CI.Significance | ||
Variables | CI.Lower.Bound | CI.Upper.Bound |
1 custom | 0.09419578 | 0.15490512 |
2 team | 0.07041818 | 0.13829928 |
3 diverse | 0.02603065 | 0.07147323 |
4 safety | 0.02284249 | 0.06334282 |
5 comm | 0.13531691 | 0.22362039 |
> #BCa Confidence Interval Tests of significance | ||
> #Comparing one predictor with all others | ||
> #If Zero is not included, Weights are Significantly different from one another | ||
CI.Predictor.Comparison | ||
Labels2 | CI.Lower.Bound | CI.Upper.Bound |
1 custom | 0.04790609 | 0.11940084 |
2 team | 0.02284219 | 0.10136002 |
3 diverse | -0.02033992 | 0.03242466 |
4 safety | 0.08748920 | 0.19370958 |
> #The Raw and Rescaled Weights for gender Level=0 | ||
Group1.RW.Results | ||
Variables | Raw.RelWeight | Rescaled.RelWeight |
1 custom | 0.13992299 | 24.932520 |
2 team | 0.10337767 | 18.420603 |
3 diverse | 0.07549385 | 13.452056 |
4 safety | 0.04100290 | 7.306202 |
5 comm | 0.20140936 | 35.888620 |
> #The Raw and Rescaled Weights for gender Level=1 | ||
Group2.RW.Results | ||
Variables | Raw.RelWeight | Rescaled.RelWeight |
1 custom | 0.11730668 | 24.671628 |
2 team | 0.10215503 | 21.484974 |
3 diverse | 0.04027808 | 8.471179 |
4 safety | 0.04129866 | 8.685823 |
5 comm | 0.17443353 | 36.686395 |
> #BCa Confidence Interval Tests of significance | ||
> #Comparing predictors across 2 groups | ||
> #If Zero is not included, Weights are Significantly different between the groups | ||
CI.Group.Comparison | ||
Labels | CI.Lower.Bound | CI.Upper.Bound |
1 custom | −0.04080760 | 0.08587230 |
2 team | −0.06412972 | 0.07337245 |
3 diverse | −0.01333891 | 0.08770565 |
4 safety | −0.04040427 | 0.04171060 |
5 comm | −0.06670757 | 0.12162894 |
#The Raw and Rescaled Weights
This section contains a summary of the relative importance weights for each of the climate variables. The column labeled “Raw.RelWeight” provides estimates of variable importance using the metric of relative effect sizes (LeBreton et al. 2007). Specifically, these weights represent an additive decomposition of the total model R^{2} and can be interpreted as the proportion of variance in job satisfaction that is appropriately attributed to each climate variable. For example, communication explained ~18 % of the variance in job satisfaction, whereas diversity orientation and safety orientation each explained ~4 % of the variance in job satisfaction. By summing the raw weights we obtain the total model R^{2} = 0.49 (i.e., 0.1234 + 0.1022 + 0.0467 + 0.0410 + 0.1789 = 0.4922), which matches the R^{2} provided in the first line of the output.
The values listed under the “Rescaled.RelWeight” column were obtained by dividing each raw relative weight by the model R^{2}. These rescaled weights provide estimates of relative importance using the metric of percentage of predicted variance attributed to each variable. For example, communication explains ~36 % of the predicted variance in job satisfaction. That is, communication explains ~36 % of the total R^{2} (i.e., 0.1789/0.4922 = 0.3635 × 100 = 36.35 %). The difference between the raw weights and the rescaled weights is simply in the interpretation one draws. Thus, it is important for researchers to be clear as to whether they are reporting raw weights based on the metric of relative effect sizes or rescaled weights based on the metric of percentage (or proportion) of predicted or explained variance in the criterion that should be attributed to each predictor.
#BCa Confidence Intervals around the Raw Weights
The next section of output provides confidence intervals (CIs) around the raw weights. These CIs are useful for explaining the precision of the raw relative weights. Larger CIs indicate less precision; smaller CIs indicate greater precision. Because relative weights can never be zero,^{1} the CIs will tend to be positively skewed, especially for relative weights that are close to zero. As noted above, we retained the alpha value of 0.05, thus these represent 95 % CIs around the point estimates for the raw relative weights.
#BCa Confidence Interval Tests of Significance
Whereas the prior section of output provides confidence intervals that can be used to evaluate the precision of the sample raw relative weight, the next two sections provide information about the statistical significance of the raw relative weights obtained by calculating bias corrected and accelerated CIs as described by Tonidandel et al. (2009). Briefly, because raw weights are scaled in terms of relative effect sizes (i.e., variance explained) they cannot be zero (or negative; Johnson 2004; Tonidandel et al. 2009). Thus, Tonidandel et al. describe a procedure that compares the observed relative weights to the relative weight obtained by including a randomly generated variable in the regression analysis. Because the variable is completely random, it has a population relative weight of zero. Consequently, it is possible to test whether the observed relative weights significantly differ from the relative weight for the random variable. The logic of their procedure is roughly analogous with the logic of using random data to determine the number of factors in a dataset via parallel analysis. An examination of the CIs immediately following the section labeled “#If Zero is not included, Weight is Significant” reveals that all of the raw weights presented in the first part of the output are statistically significant.
The next section of output also involves significance testing, but this section is focused on “#Comparing one predictor with all others.” Specifically, we requested a comparison of whether the raw relative weight for safety orientation was significantly different from the relative weights for the remaining climate variables. As noted in the output, if zero is included in the BCa CI, then the two weights are not significantly different from one another. In our example, we see that safety orientation is significantly different from customer orientation, team orientation, and communication. However, it is not different from diversity orientation. An intuitive examination of the raw weights confirms these results. We see that both diversity orientation and safety orientation have raw weights of ~0.04. And, we see that the weights for the other climate variables are much larger than the weight for safety.
#The Raw and Rescaled Weights for Gender
The next section of output presents the raw and rescaled relative weights separately for each level of our gender variable. Recall, we requested a comparison of whether the relative weights for our climate variables differed across levels of employee gender. The final section of output, “#Comparing predictors across 2 groups” presents the BCa CIs comparing the differences in raw weights as a function of employee gender. As noted on the output, “#If Zero is not included, weights are significantly different between the groups.” An examination of this final section of output indicates that there were no statistically significant differences in the raw relative weights when comparing men to women. This analysis allows one to examine whether the relative importance of particular predictors differs as a function of gender. This approach is akin to testing the difference in regression weights using moderated multiple regression, but the focus here is on relative importance instead of prediction.
Summary
Summary of a traditional relative weight analysis
Predictor | b | β | RW | CI-L | CI-U | RS-RW (%) |
---|---|---|---|---|---|---|
Criterion = Job Satisfaction (R^{2} = .4923; F[5,598] = 115.96, p < .001) | ||||||
Intercept | 0.20 | |||||
Custom^{a} | 0.23* | 0.21 | 0.1235* | 0.0942 | 0.1549 | 25.08 |
Team^{a} | 0.23* | 0.19 | 0.1022* | 0.0704 | 0.1383 | 20.77 |
Diverse | 0.04 | 0.05 | 0.0467* | 0.0260 | 0.0715 | 09.48 |
Safety | 0.06 | 0.06 | 0.0411* | 0.0228 | 0.0633 | 08.34 |
Comm^{a} | 0.34* | 0.36 | 0.1789* | 0.1353 | 0.2236 | 36.33 |
Criterion = Organizational Commitment (R^{2} = .4788; F[5,598] = 109.89, p < .001) | ||||||
Intercept | 0.23 | |||||
Custom^{a} | 0.21* | 0.18 | 0.1037* | 0.0675 | 0.1352 | 21.67 |
Team^{a} | 0.29* | 0.23 | 0.1164* | 0.0826 | 0.1500 | 24.31 |
Diverse | 0.08* | 0.10 | 0.0634* | 0.0343 | 0.0946 | 13.26 |
Safety | 0.16* | 0.16 | 0.0762* | 0.0424 | 0.1142 | 15.91 |
Comm^{a, b} | 0.24* | 0.24 | 0.1190* | 0.0782 | 0.1628 | 24.86 |
Criterion = leader effectiveness (R^{2} = .4658; F[5,598] = 104.30, p < .001) | ||||||
Intercept | 0.47* | |||||
Custom^{a} | 0.15* | 0.14 | 0.0947* | 0.0661 | 0.1225 | 20.33 |
Team^{a} | 0.34* | 0.29 | 0.1342* | 0.0947 | 0.1751 | 28.82 |
Diverse | 0.07* | 0.09 | 0.0562* | 0.0300 | 0.0875 | 12.07 |
Safety | 0.07* | 0.08 | 0.0499* | 0.0273 | 0.0767 | 10.70 |
Comm^{a} | 0.25* | 0.27 | 0.1308* | 0.0891 | 0.1758 | 28.08 |
“A relative weight analysis (Johnson 2000) was conducted using RWA-Web (Tonidandel and LeBreton 2014); results from this analysis are summarized in Table 2. Confidence intervals for the individual relative weights (Johnson 2004) and all corresponding significance tests were based on bootstrapping with 10,000 replications, an approach recommended by Tonidandel et al. (2009). Bias corrected and accelerated confidence intervals were used because of their superior coverage accuracy as recommended by Tonidandel et al. (2009). In all cases, 95 % CIs were used (corresponding to a significance testing alpha level of 0.05). Briefly, these results indicate that a weighted linear combination of our five climate variables explained roughly half of the variance in the job satisfaction criterion (R^{2} = 0.49). An examination of the relative weights revealed that all five variables explained a statistically significant amount of variance in job satisfaction as none of the 95 % CIs for the tests of significance contained zero, with the most important variables being Customer Orientation (RW = 0.12), Team Orientation (RW = 0.10), and Communication (RW = 0.18).
The relative weight results differ slightly from what was obtained from the traditional multiple regression analysis. Specifically, in the traditional analysis neither Diversity Orientation nor Safety Orientation provided a statistically significant incremental effect in the prediction of Job Satisfaction, holding constant all of the remaining climate variables. The lack of concordance in the significance of the regression coefficients and the relative weights is not uncommon (Tonidandel et al. 2009) and simply reflects that these two statistics are addressing different research questions. Regression weights are focused on incremental prediction and when predictors are correlated, variables that yield a significant bivariate relationship may not yield a significant incremental relationship; relative weights are focused on explaining which predictors are explaining non-trivial variance in our outcomes (even in the presence of additional, correlated predictors). Thus, these results work in a supplementary fashion (Tonidandel and LeBreton 2011) and suggest that both of these variables are explaining non-trivial variance in Job Satisfaction, but the correlations they share with one another (and the other climate variables) results in them explaining little unique, incremental variance.
We were also interested in testing whether the relative contribution of safety orientation to the overall R^{2} was significantly different from the remaining climate variables. Results indicated that relative weight of safety orientation (RW = 0.04) was significantly lower than all of the remaining climate variables (i.e., CIs for the comparisons did not include zero), except diversity orientation (RW = 0.05). Finally, we also tested whether there were differences in the magnitude of the relative weights as a function of employee gender. Results indicated there were no statistically significant differences as a function of gender; the CI for the male–female comparison did include zero. In sum, it appears that the majority of the explained/predicted variance in job satisfaction is attributed to customer orientation (25 % of model R^{2}), team orientation (21 % of model R^{2}), and communication (36 % of model R^{2}).”
Table 2 also includes the results for the other two outcome variables (i.e., organizational commitment and leader effectiveness). We leave obtaining these results via RWA-Web as an exercise for the interested reader.
Using RWA-Web for Other Procedures
Using the dataset described under Example #1, one could very easily execute relative weight analyses using both multivariate multiple regression and logistic regression. For multivariate relative weight analysis, one would simply select “Multivariate Regression” from the main landing, and the same general options could be selected from the web form with the exception of including multiple criterion variables (i.e., jobsat, commit, ldr_eff) in the analysis.
Whereas the traditional relative weight analysis yields raw weights that sum to R^{2}, the multivariate analysis yields weights that sum to a multivariate analog of R^{2}. LeBreton and Tonidandel (2008) reviewed several potential multivariate analogs of R^{2} and RWA-Web estimates multivariate weights using one of the recommended analogs: \( P^{2}_{\text{yx}} \). The content and structure of output mirrors that of Table 1.
For logistic relative weight analysis (Tonidandel and LeBreton 2010), one could examine the relative importance of our five climate variables in predicting the “turnover” variable in our example dataset by simply selecting “Logistic Regression” from the main landing page.
Similar to the multivariate analyses, the estimation of logistic relative weights necessitates the identification of a logistic analog to R^{2}. RWA-Web estimates logistic relative weights using one of the recommended analogs: \( R^{2}_{\text{O}} \). The content and structure of the results again mirrors that of Tables 1.
Conclusion
In conclusion, we hope that this article has documented how easy it is for scholars to integrate relative weight analysis into their research activities. We believe that relative importance analyses provide important supplemental information that should be regularly included along with the results typically provided when using regression analyses.
The typical indices produced by regression are useful but do not accurately partition variance among correlated predictors, whereas relative weight analysis is properly suited for this function. Moreover, relative importance analyses provide valuable information to researchers who can address substantive hypotheses that are not easily addressed by traditional regression results, and because they are scaled in the metric of variance explained, importance weights may be interpreted as estimates of relative effect size. RWA-Web fills an important void by providing scholars with the tools to conduct relative importance analyses in a user-friendly format irrespective of the statistical platform to which one is most accustomed. Thus, with minimal effort, a researcher can conduct relative importance analyses as a supplement to their planned regression analyses.
Footnotes
- 1.
Although zero is a theoretically possible value if the correlation between a variable and all other variables in the model are zero, one would never observe a relative weight of zero in practice because sampling error would ensure that the observed correlations deviate from zero. The implication of this is that the sample relative weights will always be positive even if the true population relative weight is zero. For a more detailed discussion of this and related issues, please see Tonidandel et al. (2009).
References
- Azen, R., & Budescu, D. V. (2003). The dominance analysis approach for comparing predictors in multiple regression. Psychological Methods, 8, 129–148. doi:10.1037/1082-989X.8.2.129.CrossRefPubMedGoogle Scholar
- Braddy, P. W., Gooty, J., Fleenor, J. W., & Yammarino, F. J. (2013). Leader behaviors and career derailment potential: A multi-analytic method examination of rating source and self–other agreement. The Leadership Quarterly,. doi:10.1016/j.leaqua.2013.10.001.Google Scholar
- Budescu, D. V. (1993). Dominance analysis: A new approach to the problem of relative importance of predictors in multiple regression. Psychological Bulletin, 114, 542–551. doi:10.1037/0033-2909.114.3.542.CrossRefGoogle Scholar
- Cortina, J. M., & Landis, R. S. (2009). When small effect sizes tell a big story, and when large effect sizes don’t. In C. E. Lance & R. J. Vandenberg (Eds.), Statistical and methodological myths and urban legends: Doctrine, verity and fable in the organizational and social sciences (pp. 287–308). New York: Routledge.Google Scholar
- Dalal, R. S., Baysinger, M., Brummel, B. J., & LeBreton, J. M. (2012). The relative importance of employee engagement, other job attitudes, and trait affect as predictors of job performance. Journal of Applied Social Psychology, 42, 295–325. doi:10.1111/j.1559-1816.2012.01017.x.CrossRefGoogle Scholar
- Darlington, R. B. (1968). Multiple regression in psychological research and practice. Psychological Bulletin, 69, 161–182. doi:10.1037/h0025471.CrossRefPubMedGoogle Scholar
- Gentry, W. A., Cullen, K. L., Sosik, J. J., Chun, J., Leupold, C. R., & Tonidandel, S. (2013). Integrity’s place among the character strengths of middle-level managers and top-level executives. The Leadership Quarterly, 24, 395–404. doi:10.1016/j.leaqua.2012.11.009.CrossRefGoogle Scholar
- Johnson, J. W. (2000). A heuristic method for estimating the relative weight of predictor variables in multiple regression. Multivariate Behavioral Research, 35, 1–19. doi:10.1207/S15327906MBR3501_1.CrossRefGoogle Scholar
- Johnson, J. W. (2004). Factors affecting relative weights: The influence of sampling and measurement error. Organizational Research Methods, 7, 283–299. doi:10.1177/1094428104266018.CrossRefGoogle Scholar
- Johnson, J. W., & LeBreton, J. M. (2004). History and use of relative importance indices in organizational research. Organizational Research Methods, 7, 238–257. doi:10.1177/1094428104266510.CrossRefGoogle Scholar
- LeBreton, J. M., & Tonidandel, S. (2008). Multivariate relative importance: Extending relative weight analysis to multivariate criterion spaces. Journal of Applied Psychology, 93, 329–345. doi:10.1037/0021-9010.93.2.329.CrossRefPubMedGoogle Scholar
- LeBreton, J. M., Binning, J. F., Adorno, A. J., & Melcher, K. M. (2004a). Importance of personality and job-specific affect for predicting job attitudes and withdrawal behavior. Organizational Research Methods, 7, 300–325. doi:10.1177/1094428104266015.CrossRefGoogle Scholar
- LeBreton, J. M., Ployhart, R. E., & Ladd, R. T. (2004b). A Monte Carlo comparison of relative importance methodologies. Organizational Research Methods, 7, 258–282. doi:10.1177/1094428104266017.CrossRefGoogle Scholar
- LeBreton, J. M., Hargis, M. B., Griepentrog, B., Oswald, F. L., & Ployhart, R. E. (2007). A multidimensional approach for evaluating variables in organizational research and practice. Personnel Psychology, 60, 475–498. doi:10.1111/j.1744-6570.2007.00080.x.CrossRefGoogle Scholar
- LeBreton, J. M., Tonidandel, S., & Krasikova, D. (2013). Residualized relative importance analysis: A technique for the comprehensive decomposition of variance in higher-order effects regression models. Organizational Research Methods, 16, 449–473. doi:10.1177/1094428113481065.CrossRefGoogle Scholar
- Lievens, F., & Peeters, H. (2008). Interviewers’ sensitivity to impression management tactics in structured interviews. European Journal of Psychological Assessment, 24, 174–180. doi:10.1027/1015-5759.24.3.174.CrossRefGoogle Scholar
- Lopina, E. C., Rogelberg, S. G., & Howell, B. (2012). Turnover in dirty work occupations: A focus on pre-entry individual characteristics. Journal of Occupational and Organizational Psychology, 85, 396–406. doi:10.1111/j.2044-8325.2011.02037.x.CrossRefGoogle Scholar
- Major, D. A., Morganson, V. J., & Bolen, H. M. (2013). Predictors of occupational and organizational commitment in information technology: Exploring gender differences and similarities. Journal of Business and Psychology, 28, 301–314. doi:10.1007/s10869-012-9282-5.CrossRefGoogle Scholar
- Martell, R. F., Lane, D. M., & Emrich, C. (1996). Male–female differences: A computer simulation. American Psychologist, 51, 157–158. doi:10.1037/0003-066X.50.3.145.CrossRefGoogle Scholar
- Scherer, K. T., Baysinger, M., Zolynsky, D., & LeBreton, J. M. (2013). Predicting counterproductive work behaviors with sub-clinical psychopathy: Beyond the Five Factor Model of personality. Personality and Individual Differences, 55, 300–305. doi:10.1016/j.paid.2013.03.007.CrossRefGoogle Scholar
- Snell, S. J., Tonidandel, S., Braddy, P., & Fleenor, J. (2013). The relative importance of political skill dimensions for predicting managerial effectiveness. European Journal of Work and Organizational Psychology,. doi:10.1080/1359432X.2013.817557.Google Scholar
- Tonidandel, S., & LeBreton, J. M. (2010). Determining the relative importance of predictors in logistic regression: An extension of relative weights analysis. Organizational Research Methods, 13, 767–781. doi:10.1177/1094428109341993.CrossRefGoogle Scholar
- Tonidandel, S., & LeBreton, J. M. (2011). Relative importance analyses: A useful supplement to multiple regression analyses. Journal of Business and Psychology, 26, 1–9. doi:10.1007/s10869-010-9204-3.CrossRefGoogle Scholar
- Tonidandel, S., & LeBreton, J. M. (2013). Beyond step–down analysis: A new test for decomposing the importance of dependent variables in MANOVA. Journal of Applied Psychology, 98, 469–477. doi:10.1037/a0032001.CrossRefPubMedGoogle Scholar
- Tonidandel, S., & LeBreton, J. M. (2014). RWA web: A free, comprehensive, web-based, and user-friendly tool for relative weight analyses. Journal of Business and Psychology. doi:10.1007/s10869-014-9351-z.
- Tonidandel, S., LeBreton, J. M., & Johnson, J. W. (2009). Determining the statistical significance of relative weights. Psychological Methods, 14, 387–399. doi:10.1037/a0017735.CrossRefPubMedGoogle Scholar