Abstract
Over the last 15 years, a number of methodological developments have enabled researchers to draw more accurate inferences concerning the relative contribution (i.e., relative importance) among multiple (often correlated) predictor variables in a regression analysis. One such development has been relative weight analysis (RWA). Researchers can use a RWA to decompose the total variance predicted in a regression model (R ^{2}) into weights that accurately reflect the proportional contribution of the various predictor variables. Prior to RWA, researchers were forced to rely on traditional statistics (e.g., correlations; standardized regression weights), which are known to yield faulty or misleading information concerning variable importance (especially when predictor variables are correlated with one another, which is often the case in organizational research). Although there has been a surge of interest in RWA over the last 10 years, integration of this statistical tool into organizational research has been hampered by the lack of a userfriendly statistical package for implementing RWA. Indeed, most popular statistical packages (e.g., SPSS, SAS) have yet to include RWA protocols into their regression modules. The purpose of this paper is to present a new, free, comprehensive, webbased, userfriendly resource, RWAWeb, which may be used by anyone having simple access to the internet. Our paper is structured as a tutorial on using RWAWeb to examine relative importance in the classic multiple regression model, the multivariate multiple regression model, and the logistic regression model. We also illustrate how RWAWeb may be used to conduct null hypothesis significance tests using advanced bootstrapping procedures.
Multiple regression is perhaps one of the most widely used analytical procedures in the organizational sciences. Although many researchers use multiple regression to identify the set of variables that will maximize the prediction of some outcome variable or criterion, another important use of multiple regression is for theory testing and explanatory purposes. In these instances, researchers are most interested in understanding how each variable contributes toward the variance explained in the outcome or criterion. Alas, it is widely recognized (Darlington 1968) that multicollinearity makes the partitioning of variance among multiple correlated predictors difficult. Despite these difficulties, researchers continue to rely on commonly available indices (e.g., simple correlation coefficients, partial correlation coefficients, standardized regression coefficients) when partitioning the variance in the criterion that is predicted by multiple (typically correlated) predictors.
Fortunately, two alternative techniques, dominance analysis (Budescu 1993) and relative weight analysis (Johnson 2000), have been developed that permit more accurate partitioning of variance among correlated predictors. Simulation research clearly shows that these measures of importance perform quite well across a variety of conditions and are recommended for this purpose (LeBreton et al. 2004b). Despite calls advocating for the wider use of these indices (Tonidandel and LeBreton 2011), researchers seem reluctant to do so. One limiting factor for many researchers might lie in the practicalities associated with obtaining these metrics for their own datasets. The output typically generated by virtually all commercial statistical software fails to routinely include either of these indices in their standard output. Although some researchers have created macros and other similar variants that will perform these calculations, these resources rely on disparate statistical packages and fail to offer a simple, easy to use, integrated solution that would be attractive to most users. To that end, the purpose of this paper is to introduce a free, comprehensive, webbased, userfriendly resource to conduct relative weight analyses. In what follows, we will provide a brief overview of relative importance analyses, provide some examples from the published literature that makes use of these techniques, and introduce our comprehensive webbased solution for performing these calculations.
Relative Importance Analysis
Within the context of linear regression, the term relative importance refers to the proportionate contribution each predictor variable makes to the total predicted criterion variance taking into account a variable’s contribution by itself and in combination with other predictor variables (Johnson and LeBreton 2004). This definition of variable importance is not to be confused with other commonly used connotations of importance such as statistical or practical significance. Information regarding a variable’s contribution to predictable variance is helpful when considering the practical utility of a variable, but aspects of the particular situation must also be considered to fully gauge practical importance (Cortina and Landis 2009). In some situations, a variable may explain only a small proportion of predictable variance and yet be very meaningful (Martell et al. 1996), whereas in other situations, a variable may account for a larger percentage of the variance but may provide little practical utility (Cortina and Landis 2009).
Relative Weight Analysis
Relative weight analysis (RWA; Johnson 2000) addresses the problem caused by correlated predictors by using a variable transformation approach to create a set of new predictors (Z _{Xk}) that are the maximally related to the original predictors (X _{j}) but are orthogonal to one another. Because these new transformed predictors are uncorrelated to one another, the criterion variable (Y) can be regressed onto this new set of predictors producing a series of standardized regression coefficients (β _{k}). Because these coefficients are generated using the orthogonal transformations of the original predictors, they no longer suffer from problems associated with collinearity. These regression coefficients are then rescaled back to the original variables by combining them with the standardized regression coefficients (λ _{jk}) obtained by regressing the original predictors (X _{j}) on their orthogonal counterparts (Z _{Xk}) producing an estimate of relative importance for each predictor variable.
The interested reader is directed to Johnson (2000), Johnson and LeBreton (2004), LeBreton and Tonidandel (2008), and Tonidandel and LeBreton (2011) for a more detailed discussion of relative weight analysis. The relative weight statistic has been shown to provide extremely good estimates of the relative importance of predictor variables when those predictor variables are correlated. This has been found in both simulation studies (LeBreton et al. 2004b) and primary studies (LeBreton et al. 2004a; LeBreton et al. 2007). Recently, new developments in relative weight analysis have expanded its applications from traditional multiple regression to more complicated regression models such as multivariate multiple regression (LeBreton and Tonidandel 2008), regression models containing higher order terms such as crossproduct terms, quadratic terms, or other polynomial terms (LeBreton et al. 2013), logistic regression (Tonidandel and LeBreton 2010), and multivariate analysis of variance (MANOVA; Tonidandel and LeBreton 2013). In addition to obtaining the relative weights themselves, more useful information can be gleaned by supplementing the weights with confidence intervals, testing the statistical significance of the weights, and evaluating the statistical significance of the difference between weights either within or between samples (Johnson 2004; Tonidandel et al. 2009).
Applications of Relative Weight Analysis
Relative weight analysis has been used across a variety of domains in the organizational sciences literature. For example, Lopina et al. (2012) used relative weight analysis to examine the relative importance of individual difference variables for predicting turnover. In a related vein, Major et al. (2013) used relative weight analysis to understand how men and women differ in terms of the most important predictors of organizational commitment. Turning to leadership, Braddy et al. (2013) examined the relative importance of different leader behaviors for predicting derailment, whereas Snell et al. (2013) used a similar approach to understand how political skill predicts managerial effectiveness. Concerning job performance, Dalal et al. (2012) examined the relative importance of employee engagement, personality, and job attitudes as predictors of performance, and Gentry et al. (2013) used relative weight analysis to understand how different character virtues might be more or less important for managers at different levels of an organization. Scherer et al. (2013) used relative weight analysis to examine the contributions of subclinical psychopathy and the global traits of the Five Factor Model in the prediction of counterproductive workplace behaviors. As a final example, Lievens and Peeters (2008) used relative weight analysis to understand how different impression management tactics are related to evaluations of interviewees. As is evident from these examples, relative weight analysis is applicable to many different areas of research.
RWAWeb
To assist those wishing to perform relative weight analysis in their own work, we endeavored to create a more general solution that would integrate the disparate types of relative importance calculations within a single easytouse framework that would be available to all researchers regardless of their preferred platform or fluency with a particular platform. To that end, we created an interactive website that performs all of these calculations using the free, opensource R statistical package as its backbone. The website is designed such that users who are familiar and comfortable with R can use the website to generate the R code necessary to conduct the various RWAs. These users can simply download the code and execute it locally on their workstation.
However, many individuals may not be familiar with R (and may not want to learn R). Consequently, we have also designed the website so that users may simply pointandclick through the website to generate the R syntax and then submit the syntax and a data file to our R web server. This server will then run the desired analyses and send the results to the user via email. Thus, users can successfully conduct these relative importance analyses with no prior knowledge of R or experience using R.
Visiting the Website and Selecting a RWA Model
RWAWeb is available at: http://relativeimportance.davidson.edu/. This website will continue to be updated as new procedures relevant to RWA are developed. In its current form, RWAWeb permits researchers to examine the relative importance of multiple (correlated or uncorrelated) predictors using (1) a standard multiple regression model, (2) a multivariate multiple regression model containing multiple (correlated or uncorrelated) dependent variables, and (3) a standard logistic regression model. A screen shot from the landing page is presented in Fig. 1. From this landing page, users simply select which type of RWA analysis they would like to perform. Each procedure has a separate page that contains instructions and an easytouse interface for conducting the relative importance analyses (Fig. 2). In addition, there is a list of references for the user seeking additional information about relative weights (in general) or the specific options available through RWAWeb.
This website also contains a link to the data file used for the illustrative examples in the current paper. Readers wanting to confirm their understanding of relative weight analysis and the RWAWeb tool are encouraged to replicate our findings using these data. RWAWeb was designed with three key features in mind: (1) ease of use, (2) inclusiveness, and (3) flexibility. Below, we briefly comment on these features before presenting a tutorial on using RWAWeb including the interpretation of results obtained using several illustrative datasets.
Ease of Use
The interface of each page is designed as an editable web form where users can select certain options by choosing various check boxes and provide additional detailed information in text boxes. For each type of analysis (multiple regression, multivariate regression, and logistic regression), the user only needs to provide three basic pieces of information in order to obtain results: the name and location of the data file, the name of the criterion variable(s), and the name of the predictor variables (note that R is case sensitive). The default file format for the data file is a commaseparated values (.csv) file, which is a standard exporting option in Excel and most other statistical software packages. With these minimal pieces of information, the website will generate raw and rescaled relative weights for each predictor (scaled as the percentage of explainable variance), the bootstrapped confidence intervals around these weights, and the corresponding tests for significance of these weights.
While disparate resources exist for calculating relative weights, this is the first platform to integrate the relative weight results with confidence intervals and tests of statistical significance. Moreover, the only other program currently available for calculating confidence intervals does not rely on the bias corrected accelerated method of bootstrapping, which was recommended by Tonidandel et al. (2009) after extensive simulation work.
The ability to easily obtain confidence intervals and tests of significance will hopefully improve the accessibility of relative importance statistics within the organizational sciences. Without having this information readily available, researchers have tended to rely on the flawed practice of merely looking at the rank ordering of predictor importance without taking into account the sampling distributions of these statistics (Tonidandel and LeBreton 2011). By focusing solely on rank order, one might incorrectly conclude a predictor to be more important than another predictor despite a considerable degree of overlap in their respective sampling distributions. Thus, by having the additional information of confidence intervals and tests of significance readily available, one can make more informed judgments regarding the true contribution of predictors to explainable variance.
Inclusiveness: Builtin Comprehensive Functionality
In addition to integrating the relative weights, confidence intervals, and tests of statistical significance, within a single platform, the website also possesses crucial additional features to make it a truly comprehensive resource for these analyses. Frequently, in addition to wanting to determine the relative importance of various predictors, researchers often have hypotheses that might predict that two predictors differ significantly from one another in terms of their relative importance or that a particular predictor’s relative importance differs in two different samples. To address either of these questions, one needs to be able to evaluate the statistical significance of the difference between relative weights either within or between samples. This new resource can perform both of these types of tests using bias corrected accelerated methods. To test for the statistical significance of the difference between relative weights within a sample, a user only needs to supply the name of a predictor variable, and the relative weight for that predictor will be tested against all of the other predictors in the model. To evaluate the statistical significance between samples, one simply provides the name of the grouping variable that identifies the different samples within the dataset and the two levels of that grouping variable one wishes to compare. The program will then automatically generate tests for significance for all of the predictors in the model between the two samples.
Flexibility
We have designed the website in an effort to provide a maximum amount of flexibility to accommodate the needs of various users. To that end, the pointandclick interface allows users to select from a variety of options including: conducting relative importance analyses on both raw data or on a correlation matrix, choosing the alpha value for the confidence intervals and tests of statistical significance as well as the number of bootstrapped replications they wish to use for those analyses, and electing different techniques for handling missing data should they have any. Finally, individuals comfortable with the R statistical program can elect to download the R code and can then modify it to meet their own personal needs.
Illustrative Example—Standard Multiple Linear Regression
To illustrate the features of RWAWeb, we will use a simulated dataset containing a sample of 604 employees. The employees completed a climate survey measuring: Customer Orientation (custom: higher scores = more customer orientation), Team Orientation (team: higher scores = greater emphasis on teamwork), Diversity Orientation (diverse: higher scores = greater emphasis on promoting ethnic and gender diversity in the workplace), Safety Orientation (safety: higher scores = greater emphasis on maintaining a safe work environment), and Communication (comm: higher scores = more effective communication from the organization’s leaders to the employees). Three months later, the same employees completed a job attitude survey measuring: Organizational Commitment (commit: higher scores = stronger commitment to the organization), Job Satisfaction (jobsat: higher scores = higher overall levels of job satisfaction), and Leader Effectiveness (ldr_eff: higher scores = higher overall perceptions of effectiveness). Finally, one year after the initial survey, data were collected concerning voluntary turnover from the organization (turnover: 0 = employee remained, 1 = employee attrited).
Using RWAWeb to Estimate Relative Weights
Using these data, we sought to examine the relative importance of the climate variables in predicting Job Satisfaction, Organizational Commitment, and Leader Effectiveness. Specifically, we wanted to estimate relative weights (Johnson 2000) along with bootstrapped standard errors for use with significance testing (Tonidandel et al. 2009).
To perform these analyses we selected the “Multiple Regression” link from the home page (see Fig. 1). This took us to the RWAWeb page for traditional multiple regression. From here, we completed the web form as follows:

(1)
Rather than downloading the R code and running it locally, we opted to “Send to R Server.”

(2)
Consequently, we needed to specify a dataset that would be sent to the R Server along with the R Code that we would generate in subsequent steps. Our data were originally saved as an SPSS “.sav” data file. However, we converted this file into a “.csv” file by opening the data file in SPSS, selecting “File” and “Save as” from the drop down menus. In the new dialog box, we selected “Comma delimited (*.csv)” under the “Save as type” options.

(3)
Because we were sending the R code and data to the R server, we also entered an email address to which the results could be sent. We noted that the data were “Raw data” and that we wanted to delete data using “Pairwise deletion,” even though the dataset had complete data.

(4)
Next, we specified the first relative weight analysis. We entered “jobsat” as our criterion variable and entered the names of our climate variables (i.e., custom, team, diverse, safety, comm) as our predictor variables (hitting “enter” or “return” after each variable was entered in the list).

(5)
We retained the default bootstrapping options (k = 10,000 replications; alpha = 0.05). We also requested a comparison of whether the relative importance of the “safety” variable differed significantly from the remaining predictor variables. That is, we sought to test a hypothesis that the “safety” variable was statistically larger than all of the remaining predictor variables.

(6)
We also requested an analysis comparing whether the relative weights differed across the “gender” variable.

(7)
After making these selections, we clicked “submit” and waited for the email from the R server. We received the email about 10 min later.

(8)
Finally, we repeated steps 1–6 for both the “commit” and “ldr_eff” outcome variables.
Interpreting the Output from RWAWeb
Table 1 contains the abbreviated output for the job satisfaction criterion. Specifically, we copied and pasted the results portion of the email into Table 1. We did not bother copying and pasting the earlier portion of the email that includes the R code for these procedures. We now provide a brief interpretation of the results from each section of the abbreviated output. It is also important to acknowledge that researchers will not be able to perfectly reproduce the results discussed in our paper because those results are based on a bootstrapping procedure. However, substantive conclusions should be generally consistent, and the numeric values in our tables should be fairly similar to those that would be obtained as others replicate our analyses.
#The Raw and Rescaled Weights
This section contains a summary of the relative importance weights for each of the climate variables. The column labeled “Raw.RelWeight” provides estimates of variable importance using the metric of relative effect sizes (LeBreton et al. 2007). Specifically, these weights represent an additive decomposition of the total model R ^{2} and can be interpreted as the proportion of variance in job satisfaction that is appropriately attributed to each climate variable. For example, communication explained ~18 % of the variance in job satisfaction, whereas diversity orientation and safety orientation each explained ~4 % of the variance in job satisfaction. By summing the raw weights we obtain the total model R ^{2} = 0.49 (i.e., 0.1234 + 0.1022 + 0.0467 + 0.0410 + 0.1789 = 0.4922), which matches the R ^{2} provided in the first line of the output.
The values listed under the “Rescaled.RelWeight” column were obtained by dividing each raw relative weight by the model R ^{2}. These rescaled weights provide estimates of relative importance using the metric of percentage of predicted variance attributed to each variable. For example, communication explains ~36 % of the predicted variance in job satisfaction. That is, communication explains ~36 % of the total R ^{2} (i.e., 0.1789/0.4922 = 0.3635 × 100 = 36.35 %). The difference between the raw weights and the rescaled weights is simply in the interpretation one draws. Thus, it is important for researchers to be clear as to whether they are reporting raw weights based on the metric of relative effect sizes or rescaled weights based on the metric of percentage (or proportion) of predicted or explained variance in the criterion that should be attributed to each predictor.
#BCa Confidence Intervals around the Raw Weights
The next section of output provides confidence intervals (CIs) around the raw weights. These CIs are useful for explaining the precision of the raw relative weights. Larger CIs indicate less precision; smaller CIs indicate greater precision. Because relative weights can never be zero,^{Footnote 1} the CIs will tend to be positively skewed, especially for relative weights that are close to zero. As noted above, we retained the alpha value of 0.05, thus these represent 95 % CIs around the point estimates for the raw relative weights.
#BCa Confidence Interval Tests of Significance
Whereas the prior section of output provides confidence intervals that can be used to evaluate the precision of the sample raw relative weight, the next two sections provide information about the statistical significance of the raw relative weights obtained by calculating bias corrected and accelerated CIs as described by Tonidandel et al. (2009). Briefly, because raw weights are scaled in terms of relative effect sizes (i.e., variance explained) they cannot be zero (or negative; Johnson 2004; Tonidandel et al. 2009). Thus, Tonidandel et al. describe a procedure that compares the observed relative weights to the relative weight obtained by including a randomly generated variable in the regression analysis. Because the variable is completely random, it has a population relative weight of zero. Consequently, it is possible to test whether the observed relative weights significantly differ from the relative weight for the random variable. The logic of their procedure is roughly analogous with the logic of using random data to determine the number of factors in a dataset via parallel analysis. An examination of the CIs immediately following the section labeled “#If Zero is not included, Weight is Significant” reveals that all of the raw weights presented in the first part of the output are statistically significant.
The next section of output also involves significance testing, but this section is focused on “#Comparing one predictor with all others.” Specifically, we requested a comparison of whether the raw relative weight for safety orientation was significantly different from the relative weights for the remaining climate variables. As noted in the output, if zero is included in the BCa CI, then the two weights are not significantly different from one another. In our example, we see that safety orientation is significantly different from customer orientation, team orientation, and communication. However, it is not different from diversity orientation. An intuitive examination of the raw weights confirms these results. We see that both diversity orientation and safety orientation have raw weights of ~0.04. And, we see that the weights for the other climate variables are much larger than the weight for safety.
#The Raw and Rescaled Weights for Gender
The next section of output presents the raw and rescaled relative weights separately for each level of our gender variable. Recall, we requested a comparison of whether the relative weights for our climate variables differed across levels of employee gender. The final section of output, “#Comparing predictors across 2 groups” presents the BCa CIs comparing the differences in raw weights as a function of employee gender. As noted on the output, “#If Zero is not included, weights are significantly different between the groups.” An examination of this final section of output indicates that there were no statistically significant differences in the raw relative weights when comparing men to women. This analysis allows one to examine whether the relative importance of particular predictors differs as a function of gender. This approach is akin to testing the difference in regression weights using moderated multiple regression, but the focus here is on relative importance instead of prediction.
Summary
Table 1 provides substantial information, but how should a researcher go about integrating that information and presenting it in his or her article or technical report? Although there is not a single best way to present results, we have extracted the key elements from the output and assembled them in Table 2. And, based on these results, a researcher might say something along the lines of:
“A relative weight analysis (Johnson 2000) was conducted using RWAWeb (Tonidandel and LeBreton 2014); results from this analysis are summarized in Table 2. Confidence intervals for the individual relative weights (Johnson 2004) and all corresponding significance tests were based on bootstrapping with 10,000 replications, an approach recommended by Tonidandel et al. (2009). Bias corrected and accelerated confidence intervals were used because of their superior coverage accuracy as recommended by Tonidandel et al. (2009). In all cases, 95 % CIs were used (corresponding to a significance testing alpha level of 0.05). Briefly, these results indicate that a weighted linear combination of our five climate variables explained roughly half of the variance in the job satisfaction criterion (R ^{2} = 0.49). An examination of the relative weights revealed that all five variables explained a statistically significant amount of variance in job satisfaction as none of the 95 % CIs for the tests of significance contained zero, with the most important variables being Customer Orientation (RW = 0.12), Team Orientation (RW = 0.10), and Communication (RW = 0.18).
The relative weight results differ slightly from what was obtained from the traditional multiple regression analysis. Specifically, in the traditional analysis neither Diversity Orientation nor Safety Orientation provided a statistically significant incremental effect in the prediction of Job Satisfaction, holding constant all of the remaining climate variables. The lack of concordance in the significance of the regression coefficients and the relative weights is not uncommon (Tonidandel et al. 2009) and simply reflects that these two statistics are addressing different research questions. Regression weights are focused on incremental prediction and when predictors are correlated, variables that yield a significant bivariate relationship may not yield a significant incremental relationship; relative weights are focused on explaining which predictors are explaining nontrivial variance in our outcomes (even in the presence of additional, correlated predictors). Thus, these results work in a supplementary fashion (Tonidandel and LeBreton 2011) and suggest that both of these variables are explaining nontrivial variance in Job Satisfaction, but the correlations they share with one another (and the other climate variables) results in them explaining little unique, incremental variance.
We were also interested in testing whether the relative contribution of safety orientation to the overall R ^{2} was significantly different from the remaining climate variables. Results indicated that relative weight of safety orientation (RW = 0.04) was significantly lower than all of the remaining climate variables (i.e., CIs for the comparisons did not include zero), except diversity orientation (RW = 0.05). Finally, we also tested whether there were differences in the magnitude of the relative weights as a function of employee gender. Results indicated there were no statistically significant differences as a function of gender; the CI for the male–female comparison did include zero. In sum, it appears that the majority of the explained/predicted variance in job satisfaction is attributed to customer orientation (25 % of model R ^{2}), team orientation (21 % of model R ^{2}), and communication (36 % of model R ^{2}).”
Table 2 also includes the results for the other two outcome variables (i.e., organizational commitment and leader effectiveness). We leave obtaining these results via RWAWeb as an exercise for the interested reader.
Using RWAWeb for Other Procedures
Using the dataset described under Example #1, one could very easily execute relative weight analyses using both multivariate multiple regression and logistic regression. For multivariate relative weight analysis, one would simply select “Multivariate Regression” from the main landing, and the same general options could be selected from the web form with the exception of including multiple criterion variables (i.e., jobsat, commit, ldr_eff) in the analysis.
Whereas the traditional relative weight analysis yields raw weights that sum to R ^{2}, the multivariate analysis yields weights that sum to a multivariate analog of R ^{2}. LeBreton and Tonidandel (2008) reviewed several potential multivariate analogs of R ^{2} and RWAWeb estimates multivariate weights using one of the recommended analogs: \( P^{2}_{\text{yx}} \). The content and structure of output mirrors that of Table 1.
For logistic relative weight analysis (Tonidandel and LeBreton 2010), one could examine the relative importance of our five climate variables in predicting the “turnover” variable in our example dataset by simply selecting “Logistic Regression” from the main landing page.
Similar to the multivariate analyses, the estimation of logistic relative weights necessitates the identification of a logistic analog to R ^{2}. RWAWeb estimates logistic relative weights using one of the recommended analogs: \( R^{2}_{\text{O}} \). The content and structure of the results again mirrors that of Tables 1.
Conclusion
In conclusion, we hope that this article has documented how easy it is for scholars to integrate relative weight analysis into their research activities. We believe that relative importance analyses provide important supplemental information that should be regularly included along with the results typically provided when using regression analyses.
The typical indices produced by regression are useful but do not accurately partition variance among correlated predictors, whereas relative weight analysis is properly suited for this function. Moreover, relative importance analyses provide valuable information to researchers who can address substantive hypotheses that are not easily addressed by traditional regression results, and because they are scaled in the metric of variance explained, importance weights may be interpreted as estimates of relative effect size. RWAWeb fills an important void by providing scholars with the tools to conduct relative importance analyses in a userfriendly format irrespective of the statistical platform to which one is most accustomed. Thus, with minimal effort, a researcher can conduct relative importance analyses as a supplement to their planned regression analyses.
Notes
Although zero is a theoretically possible value if the correlation between a variable and all other variables in the model are zero, one would never observe a relative weight of zero in practice because sampling error would ensure that the observed correlations deviate from zero. The implication of this is that the sample relative weights will always be positive even if the true population relative weight is zero. For a more detailed discussion of this and related issues, please see Tonidandel et al. (2009).
References
Azen, R., & Budescu, D. V. (2003). The dominance analysis approach for comparing predictors in multiple regression. Psychological Methods, 8, 129–148. doi:10.1037/1082989X.8.2.129.
Braddy, P. W., Gooty, J., Fleenor, J. W., & Yammarino, F. J. (2013). Leader behaviors and career derailment potential: A multianalytic method examination of rating source and self–other agreement. The Leadership Quarterly,. doi:10.1016/j.leaqua.2013.10.001.
Budescu, D. V. (1993). Dominance analysis: A new approach to the problem of relative importance of predictors in multiple regression. Psychological Bulletin, 114, 542–551. doi:10.1037/00332909.114.3.542.
Cortina, J. M., & Landis, R. S. (2009). When small effect sizes tell a big story, and when large effect sizes don’t. In C. E. Lance & R. J. Vandenberg (Eds.), Statistical and methodological myths and urban legends: Doctrine, verity and fable in the organizational and social sciences (pp. 287–308). New York: Routledge.
Dalal, R. S., Baysinger, M., Brummel, B. J., & LeBreton, J. M. (2012). The relative importance of employee engagement, other job attitudes, and trait affect as predictors of job performance. Journal of Applied Social Psychology, 42, 295–325. doi:10.1111/j.15591816.2012.01017.x.
Darlington, R. B. (1968). Multiple regression in psychological research and practice. Psychological Bulletin, 69, 161–182. doi:10.1037/h0025471.
Gentry, W. A., Cullen, K. L., Sosik, J. J., Chun, J., Leupold, C. R., & Tonidandel, S. (2013). Integrity’s place among the character strengths of middlelevel managers and toplevel executives. The Leadership Quarterly, 24, 395–404. doi:10.1016/j.leaqua.2012.11.009.
Johnson, J. W. (2000). A heuristic method for estimating the relative weight of predictor variables in multiple regression. Multivariate Behavioral Research, 35, 1–19. doi:10.1207/S15327906MBR3501_1.
Johnson, J. W. (2004). Factors affecting relative weights: The influence of sampling and measurement error. Organizational Research Methods, 7, 283–299. doi:10.1177/1094428104266018.
Johnson, J. W., & LeBreton, J. M. (2004). History and use of relative importance indices in organizational research. Organizational Research Methods, 7, 238–257. doi:10.1177/1094428104266510.
LeBreton, J. M., & Tonidandel, S. (2008). Multivariate relative importance: Extending relative weight analysis to multivariate criterion spaces. Journal of Applied Psychology, 93, 329–345. doi:10.1037/00219010.93.2.329.
LeBreton, J. M., Binning, J. F., Adorno, A. J., & Melcher, K. M. (2004a). Importance of personality and jobspecific affect for predicting job attitudes and withdrawal behavior. Organizational Research Methods, 7, 300–325. doi:10.1177/1094428104266015.
LeBreton, J. M., Ployhart, R. E., & Ladd, R. T. (2004b). A Monte Carlo comparison of relative importance methodologies. Organizational Research Methods, 7, 258–282. doi:10.1177/1094428104266017.
LeBreton, J. M., Hargis, M. B., Griepentrog, B., Oswald, F. L., & Ployhart, R. E. (2007). A multidimensional approach for evaluating variables in organizational research and practice. Personnel Psychology, 60, 475–498. doi:10.1111/j.17446570.2007.00080.x.
LeBreton, J. M., Tonidandel, S., & Krasikova, D. (2013). Residualized relative importance analysis: A technique for the comprehensive decomposition of variance in higherorder effects regression models. Organizational Research Methods, 16, 449–473. doi:10.1177/1094428113481065.
Lievens, F., & Peeters, H. (2008). Interviewers’ sensitivity to impression management tactics in structured interviews. European Journal of Psychological Assessment, 24, 174–180. doi:10.1027/10155759.24.3.174.
Lopina, E. C., Rogelberg, S. G., & Howell, B. (2012). Turnover in dirty work occupations: A focus on preentry individual characteristics. Journal of Occupational and Organizational Psychology, 85, 396–406. doi:10.1111/j.20448325.2011.02037.x.
Major, D. A., Morganson, V. J., & Bolen, H. M. (2013). Predictors of occupational and organizational commitment in information technology: Exploring gender differences and similarities. Journal of Business and Psychology, 28, 301–314. doi:10.1007/s1086901292825.
Martell, R. F., Lane, D. M., & Emrich, C. (1996). Male–female differences: A computer simulation. American Psychologist, 51, 157–158. doi:10.1037/0003066X.50.3.145.
Scherer, K. T., Baysinger, M., Zolynsky, D., & LeBreton, J. M. (2013). Predicting counterproductive work behaviors with subclinical psychopathy: Beyond the Five Factor Model of personality. Personality and Individual Differences, 55, 300–305. doi:10.1016/j.paid.2013.03.007.
Snell, S. J., Tonidandel, S., Braddy, P., & Fleenor, J. (2013). The relative importance of political skill dimensions for predicting managerial effectiveness. European Journal of Work and Organizational Psychology,. doi:10.1080/1359432X.2013.817557.
Tonidandel, S., & LeBreton, J. M. (2010). Determining the relative importance of predictors in logistic regression: An extension of relative weights analysis. Organizational Research Methods, 13, 767–781. doi:10.1177/1094428109341993.
Tonidandel, S., & LeBreton, J. M. (2011). Relative importance analyses: A useful supplement to multiple regression analyses. Journal of Business and Psychology, 26, 1–9. doi:10.1007/s1086901092043.
Tonidandel, S., & LeBreton, J. M. (2013). Beyond step–down analysis: A new test for decomposing the importance of dependent variables in MANOVA. Journal of Applied Psychology, 98, 469–477. doi:10.1037/a0032001.
Tonidandel, S., & LeBreton, J. M. (2014). RWA web: A free, comprehensive, webbased, and userfriendly tool for relative weight analyses. Journal of Business and Psychology. doi:10.1007/s108690149351z.
Tonidandel, S., LeBreton, J. M., & Johnson, J. W. (2009). Determining the statistical significance of relative weights. Psychological Methods, 14, 387–399. doi:10.1037/a0017735.
Author information
Authors and Affiliations
Corresponding author
Additional information
Editor’s note from Steven Rogelberg: In 2011, I invited Scott and James to write a Methods Corner piece on relative importance analyses. Their piece was important and impactful. In this editorial, I invited them to follow up on their earlier work by introducing a free tool they designed to enable scholars to easily implement relative weights analyses. I just want to thank Scott and James for this terrific service they provided our discipline.
Rights and permissions
About this article
Cite this article
Tonidandel, S., LeBreton, J.M. RWA Web: A Free, Comprehensive, WebBased, and UserFriendly Tool for Relative Weight Analyses. J Bus Psychol 30, 207–216 (2015). https://doi.org/10.1007/s108690149351z
Published:
Issue Date:
DOI: https://doi.org/10.1007/s108690149351z