Survey data
The survey involved 37 researchers from 17 Polish and foreign scientific institutions engaged in research in the field of technical and engineering sciences. Most of them are tenured faculty members at Universities in Poland, USA, Canada, and Australia although some of them declared employment in research institutes. The vast majority of respondents declared the position of a full professor or equivalent.Footnote 6 A few persons held the prestigious title of distinguished professor.
Every participant of the survey had to answer six questions and, thus, determine six ratios: \(\frac{w_{1}}{w_{2}},\frac{w_{1}}{w_{3}},\frac{w_{1}}{w_{4}},\frac{w_{2}}{w_{3}},\frac{w_{2}}{w_{4}},\frac{w_{3}}{w_{4}}\). The answers allowed for formation of the partial PC matrix M
r
in the form:
$$ M_{r}= \left[ \begin{array}{llll} 1 &\frac{w_{1}}{w_{2}} & \frac{w_{1}}{w_{3}} & \frac{w_{1}}{w_{4}}\\ \frac{w_{2}}{w_{1}} & 1 & \frac{w_{2}}{w_{3}} & \frac{w_{2}}{w_{4}}\\ \frac{w_{3}}{w_{1}} & \frac{w_{3}}{w_{2}} & 1 & \frac{w_{3}}{w_{4}}\\ \frac{w_{4}}{w_{1}} & \frac{w_{4}}{w_{2}} & \frac{w_{4}}{w_{3}} & 1 \end{array}\right] $$
(16)
To synthesize the final results the authors used almost all the gathered matrices M
r
. The only exceptions were five result sets with the very high inconsistency index \({\fancyscript{K}(M_{r})}\) (over 0.836), and the inconsistency index Ic(M
r
) higher than 0.1. Although all the rejected cases differ in detail, most of the rejected authors indicated very significant importance of the first criterion (scientific and/or creative achievements) over other arbitrarily chosen criteria. Unfortunately, due to the large inconsistency (in the literature Ic(M
r
) higher than 0.1 is considered as unacceptable Saaty 2005) their opinions have not been taken into account
Footnote 7 in the synthesized matrix \(\widehat{M}\).
All the 32 admissible partial results M form the following final output matrix \(\widehat{M}\), which looks as follows:
$$ \widehat{M}= \left[ \begin{array}{llll} 1 &1.813 & 1.503 & 1.784\\ 0.552 & 1 & 0.952 & 1.296\\ 0.666 & 1.05 & 1 & 1.302\\ 0.561 & 0.772 & 0.768 & 1 \end{array} \right] $$
(17)
The normalized weight vector ω derived from \(\widehat{M}\) using the geometric mean method is as follows:
$$ \omega= \left[ \begin{array}{llll} 0.36 &0.22 & 0.236 & 0.184 \end{array} \right]^{T} $$
(18)
which means that the invited experts found that rank(c
1)—the relative importance of scientific and/or creative achievements criterion is 0.36, rank(c
2)—scientific potentiality criterion is 0.22, rank(c
3)—tangible benefits of the scientific activity criterion is 0.236, and finally rank(c
4)—intangible benefits of the scientific activity criterion is 0.184.
The inconsistency indices for \(\widehat{M}\) are low. The more sensitive for local perturbations Koczkodaj’s index \({\fancyscript{K}(\widehat{M})=0.241}\) whilst \(Ic(\widehat{M})=0.002\). The standard geometric deviation for the appropriate \(\widehat{m}_{ij}\) defined as:
$$ \sigma_{g}(\widehat{m}_{ij})=\exp\left(\sqrt{\frac{\sum_{k=1}^{32}\left(\ln m_{ij}^{(k)}-\ln\widehat{m}_{ij}\right)^{2}}{32}}\right) $$
(19)
forms the matrix \(\widehat{M}_{\sigma}=[\sigma_{g}(\widehat{m}_{ij})]\) as follows:
$$ \widehat{M}_{\sigma}= \left[ \begin{array}{llll} 1 &2.093 & 2.03 & 1.787\\ 2.093 & 1 & 1.913 & 1.831\\ 2.03 & 1.913 & 1 & 1.61\\ 1.787 & 1.831 & 1.61 & 1 \end{array} \right] $$
(20)
It is easy to see (Eq. 20) that the most controversial (with the highest standard geometric deviation) comparison is between the scientific and/or creative achievements
c
1, and the scientific
potentiality
c
2. On the other hand experts were most unanimous comparing c
3 and c
4 (the standard geometric deviation of \(\sigma_{g}(\widehat{m}_{34})=\sigma_{g}(\widehat{m}_{43})=1.61\) is the closest to 1).
Results: different perspectives
Experts were chosen at random among those who know the specificity of Polish technical scientific units. Most of the experts are affiliated at the Polish universities or research institutes. Three experts are affiliated at foreign universities, although they worked at Polish universities in the past. Since the aim of the survey was to propose the weights \(w_{1},\ldots,w_{4}\) for technical scientific units (including such units as the departments of mathematics, physics or computer science), hence most experts is working or has worked in such institutions. On the other hand, it was important for the authors of the survey that the experts came from different research centers. The best represented university is AGH UST—the place of work of the second and the third author. The representations of other 16 scientific units count from one to three experts. Out of the all respondents the authors chose the VIP group of six the most influential people consisting of distinguished professors and former or current members of official governmental and scientific bodies, including CERU. The overall results taking into account two special groups: VIP group and experts employed at the best represented AGH UST are shown below (Fig. 1).
Among the experts whose opinion have been taken into account there may be distinguished a group of full professors (or professor ordinarius)—18 persons, associate professors (or doctors with habilitation)—6 persons, assistant professors (or doctors)—7 persons, assistants—1 person. The ranking result with respect of these groups (assistant professors and assistants are treated as a single group) are shown below (Fig. 2).
Montecarlo discrepancy validation
As a validation method for the survey data the authors adopt ten times repeated twofold cross-validation procedure Kohavi (1995). In every repetition the survey sample is randomly split into two disjoint sets \(S_{1}=\{M_{1},\ldots,M_{16}\}\) and \(S_{2}=\{M_{17},\ldots,M_{32}\}. \) Both groups are used to synthesize matrices \(\widehat{M}_{1}\) and \(\widehat{M}_{2}\), next two ranking vectors \(a=\left[a_{1},\ldots,a_{4}\right]^{T}\) and \(b=\left[b_{1},\ldots,b_{4}\right]^{T}\) are computed. The vector a is called the reference rank vector, whilst b is called the validation rank vector. For each pair of vectors a and b the discrepancy vector \(d=\left[\left|a_{1}-b_{1}\right|,\ldots,\left|a_{4}-b_{4}\right|\right]^{T}\) is computed.
The values \(d_{1},\ldots,d_{4}\) are adopted as a measures of fit. They provide information on how much the synthesized ranking values for the criteria \(c_{1},\ldots,c_{4}\) provided by the first group of experts differ from the ranking values provided by the second group. Intuitively speaking the adopted procedure simulates the situation where two disjoint group of experts provide two competitive rankings. Then the first ranking is validated by the second one. The validation procedure has been repeated ten times, so there are ten vectors \(a^{(1)},\ldots,a^{(10)}\) and ten vectors \(b^{(1)},\ldots,b^{(10)}\). The final reference rank a
avg and the discrepancy (fit indicator) vector d
avg are computed as arithmetic means:
$$ a^{{\rm avg}}=\left[\frac{1}{10}\sum_{i=1}^{10}a_{1}^{(i)}, \ldots,\frac{1}{10}\sum_{i=1}^{10}a_{4}^{(i)}\right] $$
(21)
and
$$ d^{{\rm avg}}=\left[\frac{1}{10}\sum_{i=1}^{10}d_{1}^{(i)}, \ldots,\frac{1}{10}\sum_{i=1}^{10}d_{4}^{(i)}\right] $$
(22)
As a result of the conducted experiment, the following numerical values are obtained:
$$ a^{{\rm avg}}= \left[ \begin{array}{llll} 0.364& 0.219 & 0.234 & 0.183 \end{array} \right]^{T} $$
(23)
$$ d^{{\rm avg}}= \left[ \begin{array}{llll} 0.053 & 0.037 & 0.043 & 0.023 \end{array} \right]^{T} $$
(24)
It is easy to see, that the obtained rank result is (on average) similar to the overall result of the survey (Eq. 18). In particular both vectors a
avg
(Eq. 23) and ω (Eq. 18) propose the same order of criteria importance. Their individual numerical values are also close to each other. The absolute average absolute difference between individual values in vectors a
(i) and b
(i) seem to be reasonably small since they are almost an order of magnitude less than the values in a
avg. They suggest that regardless of the selection of the group criterion c
1 should be the most important one a
avg1
− d
avg1
> a
avg
i
+ d
avg
i
. Unfortunately there is no similar guarantee in the case of any other criterion. The values \(a_{1}^{{\rm avg}}\pm d_{1}^{{\rm avg}},\ldots,a_{4}^{{\rm avg}}\pm d_{4}^{{\rm avg}}\) indicate the discrepancy intervals in which the weights of criteria \(c_{1},\ldots,c_{4}\) established by the competitive team of experts are expected to be found.
The results a
avg and d
avg (Eqs. 23, 24) were calculated on the assumption that S
1 and S
2 are equal in size. Thus, in our case both of them count 16 elements. Of course when the size of the set of experts is changing the values a
avg and d
avg may get changed. For example, according to the intuition (confirmed in tests), the smaller set S
1 (and the larger S
2) the higher discrepancies \(d_{1}^{{\rm avg}},\ldots,d_{4}^{{\rm avg}}\). For example for \(\left|S_{1}\right|=6\) and \(\left|S_{2}\right|=26\) the sample reference rank and the discrepancy vectors are:
$$ a_{6}^{{\rm avg}}= \left[ \begin{array}{llll} 0.336 & 0.221 & 0.252 & 0.189 \end{array} \right]^{T} $$
(25)
$$ d_{6}^{{\rm avg}}= \left[ \begin{array}{llll} 0.083& 0.048 & 0.051 & 0.027 \end{array} \right]^{T} $$
(26)
The adopted Montecarlo discrepancy validation procedure tries to model a realistic situation in which one group of experts provides one rank, whilst the other group (disjoint with the first one) creates another rank. Both groups call into question the results of its opponent. As demonstrated by the tests carried out when both groups are composed of experts with a similar scientific background the discrepancies might not be to high.
Also the further research on the inconsistency of synthesized PC matrix \(\widehat{M}\) seem to be interesting. In particular the relationship between the values of inconsistency indices \(Ic(\widehat{M})\) and \({\fancyscript{K}(\widehat{M})}\) and the deviations of the individual expert judgements in matrices \(M_{1},\ldots,M_{r}\) need better explanation.