1 Introduction

The first written evidence for the pairwise comparisons (PC) method dates back to the thirteenth century, when Ramon Llull from Majorca wrote a seminal piece “Artifitium electionis personarum” (The method for the elections of persons) about voting and elections [5, 7], followed by two subsequent works forming a practical study on election processes.Footnote 1 Nowadays, PC as a voting method is a way of deciding on the relative utility of alternatives used in decision theory [25] and other fields like economics [22], psychometrics and psychophysics [28] and so on. PC theory is developed by many research teams in different fields and approaches. Characteristic approaches can be noted, such as AHP–analytic hierarchy process proposed and constantly developed by Saaty [25, 27], the approximate PC reasoning [12, 20], the issue of incomplete PC relations [1], nonreciprocal PC relation properties [8], and the problem of non-numerical rankings [11]. There are also many case studies and applications [13, 19].

The heuristic rating estimation (HRE) method which enables the user to explicitly define a reference set of concepts, for which the ranking values are a priori known, is currently being developed [16, 17]. The base heuristics used in HRE proposes to determine the relative values of a single non-reference concept as a weighted arithmetic mean of all the other concepts. This proposition leads to the linear equation system defined by the matrix \(\mathbf {A}\) and the strictly positive vector of constant terms \(b\).

In this work, the authors show that using a geometric mean, rather than an arithmetic one, to determine the relative priorities of concepts may in some cases be more convenient. The main benefit of the proposed solution stems from the guarantee of the existence of a solution.Footnote 2 Hence, unlike the original proposal, a ranking list can be created for all even a non-reciprocal matrices. This guarantee is paid for by an increase in computational complexity. The presented solution is accompanied by two numerical examples.

The HRE approach can potentially be useful in many situation when the reference data are available. In particular it might be interesting for the AHP users as a viable alternative for the eigenvalue based [25] or the geometric mean method [6].

The presented work is a continuation of research initiated in [16, 17]. It redefines the main heuristics of HRE and the method of calculating the solution. The HRE approach proposed in the previous articles is briefly outlined in (Sect. 2). There is also a short summary of a number of important properties of M-matrices (Sect. 2.3), which are essential to the properties of the presented method. The next section (Sect. 3) describes the proposed solution and discusses two important properties: solution existence (Sect. 3.2) and optimality (Sect. 3.3). These theoretical considerations are accompanied by two meaningful examples showing how the presented method can be used in practice (Sect. 4). A brief summary is provided in (Sect. 5).

2 Preliminaries

2.1 Basic concepts of pairwise comparisons method

The input for the PC method is the PC matrix \(M=(m_{ij})\), where \(m_{ij}\in \mathbb {R}_{+}\) and \(i,j\in \{1,\ldots ,n\}\). The values \(m_{ij}\) and \(m_{ji}\) represent subjective expert judgments as to the relative importance, utility or quality indicators of concepts \(c_{i}\) and \(c_{j}\). Thus, according to the best knowledge of experts, it should hold \(c_{i}=m_{ij}c_{j}\) .

Definition 1

A matrix \(M\) is said to be reciprocal if for all \(i,j\in \{1,\ldots ,n\}\) it holds that \(m_{ij}=\frac{1}{m_{ji}}\), and \(M\) is said to be consistent if for all \(i,j,k\in \{1,\ldots ,n\}\) is \(m_{ij}\cdot m_{jk}\cdot m_{ki}=1\).

Since the data in the PC matrix represents subjective opinions of experts, they might be inconsistent. Hence, there may exist a triad \(m_{ij},m_{jk},m_{ki}\) of entries in \(M\) for which \(m_{ik}\cdot m_{kj}\ne m_{ij}\). This leads to a situation in which the relative importance of \(c_{i}\) with respect to \(c_{j}\) is either \(m_{ik}\cdot m_{kj}\) or \(m_{ij}\). This observation underlies two related concepts: a priority deriving method that transforms even an inconsistent matrix \(M\) into a consistent priority vector, and an inconsistency index that describes to what extent the matrix \(M\) is inconsistent. The problem of inconsistency has been addressed by Saaty in his seminal work [25], later intensively studied also by other researchers [4, 14]. There are a number of priority deriving methods [9] and inconsistency indexes [2, 3]. Following Saaty a good inconsistency indicator is the principal eigenvalue of \(M\).

Definition 2

Saaty’s inconsistency index \(\fancyscript{S}\) of \(n\times n\) and (\(n>2)\) reciprocal matrix \(M\) is equal to

$$\begin{aligned} \fancyscript{S}(M)\overset{{df}}{=}\frac{\lambda _{\max }-n}{n-1} \end{aligned}$$
(1)

where \(\lambda _{\max }\) is the principal eigenvalue of \(M\).

According to [25] the matrix is sufficiently consistent when \(\fancyscript{S}(M)\le 0.1\) (the usual limit of acceptability for the matrix \(4\times 4\) is \(0.08\) [see [26]]).

The result of the pairwise comparisons method is a ranking–a function that assigns values to the concepts. Formally, it can be defined as follows.

Definition 3

The ranking function for \(C\) (the ranking of \(C\)) is a function \(\mu :C\rightarrow \mathbb {R}_{+}\) that assigns to every concept from \(C\) a positive value from \(\mathbb {R}_{+}\).

Thus, \(\mu (c)\) represents the ranking value for \(c\in C\). The \(\mu \) function is usually defined as a vector of weights \(\mathbf {\mu }\overset{{df}}{=}\left[ \mu (c_{1}),\ldots , \mu (c_{n})\right] ^{T}\). According to the most popular eigenvalue based approach proposed by Saaty [25] the final ranking \(\mu _{{ev}}\) is determined as the principal eigenvector of the \(PC\) matrix \(M\), rescaled so that the sum of all its entries is \(1\), i.e.

$$\begin{aligned} \mu _{{ev}}=\left[ \frac{\mu _{{max}}(c_{1})}{s_{{ev}}},\ldots ,\frac{\mu _{{{max}}}(c_{n})}{s_{{ev}}}\right] ^{T} \quad \text{ and } \quad s_{{ev}}=\underset{i=1}{\overset{n}{\sum }}\mu _{{max}}(c_{i}) \end{aligned}$$
(2)

where \(\mu _{{ev}}\)–the ranking function, \(\mu _{{max}}\overset{{df}}{=}\left[ \mu _{{max}}(c_{1}),\ldots , \mu _{{max}}(c_{n})\right] ^{T}\)—the principal eigenvector of \(M\). Another popular approach proposes the rescaled geometric mean (GM) of rows of \(M\) as the ranking result, i.e.

$$\begin{aligned} \mu _{{gm}}=\left[ \frac{p_{1}}{s_{{gm}}},\ldots ,\frac{p_{n}}{s_{{gm}}}\right] ^{T} \end{aligned}$$
(3)

where

$$\begin{aligned} p_{i}=\left( \prod _{j=1}^{n}m_{ij}\right) ^{\frac{1}{n}} \quad \text {and} \quad s_{{gm}}=\underset{i=1}{\overset{n}{\sum }}\left( \prod _{j=1}^{n}m_{ij}\right) ^{\frac{1}{n}} \end{aligned}$$
(4)

It can be shown that for the fully consistent matrix \(M\) the ranking vectors \(\mu _{{ev}}\) and \(\mu _{{gm}}\) are identical.

The geometric mean minimizes the sum of logarithmic errors \(e\) [6, 10] given as:

$$\begin{aligned} e(\mu (c_{1}),\ldots ,\mu (c_{n}))=\sum _{i=1}^{n}\sum _{j=1}^{n}\left( \ln (m_{ij})-\ln \left( \frac{\mu (c_{i})}{\mu (c_{j})}\right) \right) ^{2} \end{aligned}$$
(5)

Hence, a direct solving of a logarithmic least squares optimization problem can also be used as the priority deriving procedure. As it will be shown later in (Sect. 3.3) under certain conditions also the presented geometric approach minimizes (5). Thus, despite the fact that the presented approach is a ranking procedure, it can serve as a method of solving constrained optimization problem.

A more complete overview including other priority deriving methods can be found in [2, 9].

2.2 Pairwise comparisons method with the reference set

Usually, when using the pairwise comparisons method, the ranking values \(\mu (c_{1}),\ldots ,\mu (c_{n})\) are initially unknown. Hence, they need to be determined by the priority deriving procedure. In some cases, however, there are concepts for which the priorities are known from elsewhere. Hence, the decision makers may have additional knowledge about the group of elements \(C_{K}\subseteq C\) that allows them to determine \(\mu (c)\) for \(C_{K}\) in advance.

For example, let \(c_{1},c_{2}\) and \(c_{3}\) represent oil paintings that an auction house plans to put up for auction. The sequence of paintings during the auction should correspond to their approximate valuation. In order to determine the indicative price of paintings the auction house asked experts to evaluate them in pairs, taking into account that two other paintings from the same period of time were previously auctioned for \(\mu (c_{4})\) and \(\mu (c_{5})\).

The situation as described above prompted the first author [16, 17] to propose a heuristic rating estimation (HRE) model. According to HRE the set of concepts \(C\) is composed of unknown concepts \(C_{U}=\{c_{1},\ldots ,c_{k}\}\) and known (reference) concepts \(C_{K}=\{c_{k+1},\ldots ,c_{n}\}\), where \(C_{U},C_{K}\ne \emptyset \) and \(C_{U}\cap C_{K}=\emptyset \). The values \(\mu (c_{i})\) for \(c_{i}\in C_{K}\) are known, whilst the values \(\mu (c_{j})\) for elements \(c_{j}\in C_{U}\) need to be calculated. Following the heuristics of averaging with respect to the reference values [17], the solution proposed by HRE is to adopt as \(\mu (c_{j})\), for every \(c_{j}\in C_{U}\), the arithmetic mean of all the other values \(\mu (c_{i})\) multiplied by factor \(m_{ji}\):

$$\begin{aligned} \mu (c_{j})=\frac{1}{n-1}\sum _{i=1,i\ne j}^{n}m_{ji}\mu (c_{i}) \end{aligned}$$
(6)

If the expert judgments gathered in the matrix \(M\) were fully consistent (Definition 1), then every component of the sum (6) in the form \(m_{ji}\mu (c_{i})\) would equal \(\mu (c_{j})\). Because they are generally not, then every component is only an approximation of \(\mu (c_{j})\). Thus, the arithmetic mean of the individual approximations has been adopted as the most probable value of \(\mu (c_{j})\). To determine unknown values \(\mu (c_{j})\) for \(c_{j}\in C_{U}\) the problem, formalised as (6), can be written down as the linear equation system \(A\mu =b\), where:

$$\begin{aligned} A=\left[ \begin{array}{ccc} 1 &{} \cdots &{} -\frac{1}{n-1}m_{1,k}\\ -\frac{1}{n-1}m_{2,1} &{} \cdots &{} -\frac{1}{n-1}m_{2,k}\\ \vdots &{} \ddots &{} \vdots \\ -\frac{1}{n-1}m_{k,1} &{} \cdots &{} 1 \end{array}\right] \end{aligned}$$
(7)

and

$$\begin{aligned} b=\left[ \begin{array}{c} \frac{1}{n-1}\sum _{i=k+1}^{n}m_{1,i}\mu (c_{i})\\ \frac{1}{n-1}\sum _{i=k+1}^{n}m_{2,i}\mu (c_{i})\\ \vdots \\ \frac{1}{n-1}\sum _{i=k+1}^{n}m_{k,i}\mu (c_{i}) \end{array}\right] \end{aligned}$$
(8)

The solution \(\mu =\left[ \mu (c_{1}),\ldots ,\mu (c_{k})\right] ^{T}\) determines the values of \(\mu \) for elements from \(C_{U}\). Together with the known \(\mu (c_{k+1}),\ldots , \mu (c_{n})\), the vector \(\mu \) forms the complete result list, which after sorting can be used to build a ranking. Although the values \(\mu (c)\) for \(c\in C\) are called priorities, they usually have a specific meaning. In the case of the previously mentioned example they represent the expected price of paintings.

According (Definition 3) the ranking results must be strictly positive, hence only strictly positive vectors \(\mu \) are considered to be feasible. It can be shown that the equation \(A\mu =b\) has a feasible solution if \(A\) is strictly diagonally dominant by rows [17]. It has recently been shown [18] that the equation has a feasible solution when the Koczkodaj’s inconsistency index [14] is not to high.

2.3 M-matrices

Very often the real life problem can be reduced to the linear equation system \(A\mu =b\), where the matrix \(A\) has some special structure. Frequently the matrix \(A\) has positive diagonal and nonpositive off-diagonal entries. Due to their importance in practice this type of matrix has been especially thoroughly studied by researchers [23, 24]. To define it formally a few more notions and definitions are needed.

Let \(\fancyscript{M}_{\mathbb {R}}(n)\) be a set of \(n\times n\) matrices over \(\mathbb {R}\), and \(\fancyscript{M}_{\mathbb {Z}}(n)\) the set of all \(A=[a_{ij}]\in \fancyscript{M}_{\mathbb {R}}(n)\) with \(a_{ij}\le 0\) if \(i\ne j\) and \(i,j\in \{1,\ldots ,n\}\). Furthermore, assume that for every matrix \(A\in \fancyscript{M}_{\mathbb {R}}(n)\) and vector \(b\in \mathbb {R}^{n}\) the notation \(A\ge 0\) and \(b\ge 0\) will mean that every \(m_{ij}\) and \(b_{k}\) are non-negative and neither \(A\) nor \(b\) equals \(0\). The spectral radius of \(A\) is defined as \(\rho (A)\overset{{df}}{=}\max \{|\lambda |:\det (\lambda I-A)=0\}\).

Definition 4

An \(n\times n\) matrix that can be expressed in the form \(A=sI-B\) where \(B=[b_{ij}]\) with \(b_{ij}\ge 0\) for \(i,j\in \{1,\ldots ,n\}\), and \(s\ge \rho (B)\), the maximum of the moduli of the eigenvalues of B, is called an M-matrix.

Following [23] some M-matrix properties are recalled below in the form of the Theorem 1.

Theorem 1

For every \(A\in \fancyscript{M}_{\mathbb {Z}}(n)\) each of the following conditions is equivalent to the statement: \(A\) is a nonsingular M-matrix.

  1. 1.

    \(A\) is inverse positive. That is, \(A^{-1}\) exists and \(A^{-1}\ge 0\)

  2. 2.

    There exists a positive diagonal matrix \(D\) such that \(AD\) has all positive row sums.

It is worth noting that for every matrix equation in the form \(A\mu =b\), where \(A\) is a nonsingular M-matrix, \(\mu =A^{-1}b\) holds. Since \(A^{-1}\ge 0\), \(b>0\) also implies that \(\mu >0\).

3 HRE: geometric approach

3.1 Heuristics of the geometric averaging with respect to the reference values

The pairwise comparisons method is most often used to transform a PC matrix into a ranking list of mutually compared concepts. During the transformation a priority is assigned to each concept. Therefore, this transformation is often called a priority deriving method. There are many priority deriving methods. Besides the eigenvalue based method (2), where the ranking values \(\mu (c_{i})\) are approximated as the arithmetic means of \(m_{ij}\cdot \mu (c_{j})\), the geometric mean of rows is also used (3). This may suggest that for a ranking problem with the reference set [17], the arithmetic mean (6) might also be replaced by the geometric mean. This observation prompted the author to formulate and investigate geometric averaging with respect to reference values heuristics. According to this proposition the following non-linear equation is used to determine the unknown values \(\mu (c_{j})\) for \(c_{j}\in C_{U}\):

$$\begin{aligned} \mu (c_{j})=\left( \prod _{i=1,i\ne j}^{n}m_{ji}\mu (c_{i})\right) ^{\frac{1}{n-1}} \end{aligned}$$
(9)

After raising both sides to the power of \(n-1\) the geometric averaging heuristics equation (9) leads to a non-linear equation system in the form:

(10)

Of course, since the ranking values for \(c_{k+1},\ldots ,c_{n}\in C_{K}\) make a reference set where the values \(\mu (c_{j})\) are known and fixed, some products in the form \(m_{ji}\mu (c_{i})\) are initially known constants. Let us denote:

$$\begin{aligned} g_{j}=\prod _{i=k+1}^{n}m_{ji}\mu (c_{i}) \end{aligned}$$
(11)

for \(j=1,\ldots ,k\) as the constant part of each Eq. (10). Thus, the non-linear equation system can be written as:

Hence \(\mu (c_{j}),\, m_{ij},\, g_{j}\in \mathbb {R}_{+}\), let us denote \(\log _{\xi }\mu (c_{j})\overset{{df}}{=}\widehat{\mu }(c_{j})\), \(\widehat{m}_{ij}\overset{{df}}{=}\log _{\xi }m_{ij}\) and \(\widehat{g}_{j}\overset{{df}}{=}\log _{\xi }g_{j}\) for some \(\xi \in \mathbb {R}_{+}\). It is easy to see that the above non-linear equation system is equivalent to the following one:

(12)

By grouping all the constant terms on the right side of each above equation we obtain the linear equation system

(13)

where \(b_{i}\overset{{df}}{=}\sum _{j=1,j\ne i}^{k}\widehat{m}_{1,j}+\widehat{g}_{i}\) for \(i=1,\ldots ,k\), which can be easily written down in the matrix form

$$\begin{aligned} \widehat{A}\widehat{\mu }=b \end{aligned}$$
(14)

where

$$\begin{aligned} \widehat{A}&= \left[ \begin{array}{cccc} (n-1) &{} -1 &{} \cdots &{} -1\\ \vdots &{} \ddots &{} &{} \vdots \\ \vdots &{} &{} \ddots &{} \vdots \\ -1 &{} -1 &{} \cdots &{} (n-1) \end{array}\right] ,\end{aligned}$$
(15)
$$\begin{aligned} \widehat{\mu }&= \left[ \begin{array}{c} \widehat{\mu }(c_{1})\\ \widehat{\mu }(c_{2})\\ \vdots \\ \widehat{\mu }(c_{k}) \end{array}\right] ,\quad \text {and}\quad b=\left[ \begin{array}{c} b_{1}\\ b_{2}\\ \vdots \\ b_{k} \end{array}\right] \end{aligned}$$
(16)

Therefore, the solution \(\widehat{\mu }\) of the linear equation system (14) automatically provides the solution to the original non-linear problem as formulated in (10). Indeed the ranking vector \(\mu \) can be computed following the formula:

$$\begin{aligned} \mu =\left[ \xi ^{\widehat{\mu }(c_{1})},\ldots ,\xi ^{\widehat{\mu }(c_{k})}\right] ^{T} \end{aligned}$$
(17)

Importantly, as is shown below, a feasible solution of (14) always exists. Hence, the heuristics of averaging with respect to the geometric mean always provides the user with an appropriate ranking function.

3.2 Existence of solution

The form of \(\widehat{A}\) is specific. The positive diagonal and the negative off-diagonal real entries mean that \(\widehat{A}\in \fancyscript{M}_{\mathbb {Z}}(k)\) (see Sect. 2.3). Let us put:

$$\begin{aligned} D=\left[ \begin{array}{ccc} 1 &{} \cdots &{} 0\\ \vdots &{} \ddots &{} \vdots \\ 0 &{} \cdots &{} 1 \end{array}\right] \end{aligned}$$

and \(D\in \fancyscript{M}_{\mathbb {R}}(k)\). Of course \(D\) is a positively dominant matrix. Thus, the product \(\widehat{A}\cdot D=\widehat{A}\). The sum of each row in \(\widehat{A}\) equals

$$\begin{aligned} (n-1)+\sum _{i=1}^{k-1}(-1)=n-k \end{aligned}$$

Since \(C_{K}\) is nonempty, its cardinality \(\left| C_{K}\right| =n-k\) is greater than \(0\). This means that the sum of each row of \(\widehat{A}\cdot D\) is positive. Hence, due to the Theorem 1, \(\widehat{A}\) is a nonsingular M-matrix (Definition 4), the rank of \(\widehat{A}\) is \(k\) and also the rank of the extended matrix \([\widehat{A}|b]\) is \(k\). Thus, \(\widehat{A}^{-1}\) exists (i.e. \(\widehat{\mu }=\widehat{A}^{-1}b\)) and the equation (14) always has exactly one solution in \(\mathbb {R}^{k}\). Due to the form of the solution to the main problem (17) and the fact that for any fixed \(x>0\) and \(y>0\) the exponential function \(f(x)=x^{y}\) is a bijection in \(\mathbb {R}_{+}\), the existence and uniqueness of solution of (14) implies the existence and uniqueness of solution of (10). Note that \(\mu \) is a vector in \(\mathbb {R}_{+}^{k}\), i.e. each of its entries is strictly positive.

It is worth noting that, similarly as the geometric mean method (3), the presented approach provides a feasible solution also for non-reciprocal matrices. In other words, unlike the original proposition [17], there is a guarantee that the heuristics of the geometric averaging with respect to the reference values always provides a feasible ranking result to the user. Footnote 3

3.3 Optimality condition

One of the reasons for introducing the geometric mean method (3) is minimizing the multiplicative error \(e_{ij}\) [9] defined as:

$$\begin{aligned} m_{ij}=\frac{p_{i}}{p_{j}}e_{ij} \end{aligned}$$
(18)

In the case of the geometric averaging heuristics the multiplicative error equation takes the form:

$$\begin{aligned} m_{ij}=\frac{\mu (c_{i})}{\mu (c_{j})}e_{ij} \end{aligned}$$
(19)

The multiplicative error is commonly accepted to be log normal distributed (in the same way the additive error would be assumed to be normally distributed). Let \(e:\mathbb {R}_{+}^{n}\rightarrow \mathbb {R}\) be the sum of multiplicative errors (5). As is shown in the Theorem below the heuristics (9) is often optimal with respect to the value of the multiplicative error function \(e\).

Theorem 2

The geometric averaging with respect to the reference values heuristics minimizes the sum of multiplicative errors \(e(\mu (c_{1}),\ldots ,\mu (c_{n}))\) if

$$\begin{aligned} \mu (c_{i})<(n-1)\sum _{j=1,j\ne i}^{n}\mu (c_{j}) \end{aligned}$$
(20)

for \(i=1,\ldots ,n\).

Proof

To determine the minimum of (5) let us forget for a moment that \(\mu (c_{k+1}),\ldots ,\mu (c_{n})\) are constants (the reference values), and let us treat them as any other arguments of \(e\). In order to determine the minimum of (5) the first derivative needs to be calculated. Thus,

$$\begin{aligned} \frac{\partial e}{\partial \mu (c_{i})}&= \frac{1}{\mu (c_{i})}\left( \sum _{r=1,r\ne i}^{n}4(n-1)\ln \mu (c_{i})-4\sum _{j=1,j\ne i}^{n}\ln \mu (c_{j})\right. \nonumber \\&\left. +2\sum _{r=1,r\ne i}^{n}\ln (m_{ri})-2\sum _{j=1,j\ne i}^{n}\ln (m_{ij})\right) \end{aligned}$$
(21)

for \(i=1,\ldots ,n\). Due to the reciprocity of \(M\), i.e. \( m_{ij}={1}/{m_{ji}}\), the Eq. (21) can be written as:

$$\begin{aligned} \frac{\partial e}{\partial \mu (c_{i})}&=-4\left( \frac{\sum _{j=1,j\ne i}^{n}(\ln \mu (c_{j})+\ln (m_{ij}))-(n-1)\ln \mu (c_{i})}{\mu (c_{i})}\right) \end{aligned}$$
(22)

The function \(e\) reaches the minimum if \({\partial e}/{\partial \mu (c_{i})}=0\). This leads to the postulate that

$$\begin{aligned} \sum _{j=1,j\ne i}^{n}(\ln \mu (c_{j})+\ln (m_{ij}))-(n-1)\ln \mu (c_{i})=0 \end{aligned}$$
(23)

for \(i=1,\ldots ,n\). Thus,

$$\begin{aligned} \ln \mu (c_{i})=\frac{1}{n-1}\left( \sum _{j=1,j\ne i}^{n}\ln m_{ij}\mu (c_{j})\right) \end{aligned}$$
(24)

which is directly equivalent to (9). In other words any solution to the equation system (10) is a good candidate to be a minimum of (5). It remains to settle the matrix \(H\) of second derivative of \(e\). When \(H\) is positive definite then the solution of (10) actually minimizes the function \(e\). As a result of further differentiation is determined that the diagonal elements of \(H\) are

$$\begin{aligned} \frac{\partial ^{2}f}{\partial \mu (c_{i})\partial \mu (c_{i})}=\frac{4(n-1)}{\mu ^{2}(c_{i})}-\frac{1}{\mu (c_{i})}\frac{\partial f}{\partial \mu (c_{i})} \end{aligned}$$
(25)

where \(i=1,\ldots ,n\), and the other elements for which \(i\ne j\) and \(i,j=1,\ldots ,n\) take the form:

$$\begin{aligned} \frac{\partial ^{2}f}{\partial \mu (c_{i})\partial \mu (c_{j})}=-\frac{4}{\mu (c_{i})\mu (c_{j})} \end{aligned}$$
(26)

Since the matrix \(H\) is considered for \(e\) in the point \(\left( \mu (c_{1}),\ldots , \mu (c_{n})\right) \) such that (9) holds, the first derivative of \(e\) is \(0\). Therefore, the Hessian matrix \(H\) takes the form:

$$\begin{aligned} H=\left[ \begin{array}{cccc} \frac{4(n-1)}{\mu ^{2}(c_{1})} &{} -\frac{4}{\mu (c_{1})\mu (c_{2})} &{} \cdots &{} -\frac{4}{\mu (c_{1})\mu (c_{n})}\\ \vdots &{} \frac{4(n-1)}{\mu ^{2}(c_{2})} &{} \vdots &{} \vdots \\ \vdots &{} \vdots &{} \ddots &{} \vdots \\ -\frac{4}{\mu (c_{n})\mu (c_{1})} &{} -\frac{4}{\mu (c_{n})\mu (c_{2})} &{} \cdots &{} \frac{4(n-1)}{\mu ^{2}(c_{n})} \end{array}\right] \end{aligned}$$
(27)

According to [24, p. 29], if \(H\) is strictly diagonally dominant by rows, symmetrical, and with positive diagonal entries, then it is also positive definite. To meet the first strict diagonal dominance criterion (others are satisfied) it is required that:

$$\begin{aligned} \left| \frac{n-1}{\mu ^{2}(c_{i})}\right| >\sum _{j=1,j\ne i}^{n}\left| -\frac{1}{\mu (c_{i})\mu (c_{j})}\right| \end{aligned}$$
(28)

for \(i=1,\ldots ,n\). Thus,

$$\begin{aligned} \mu ^{2}(c_{i})<(n-1)\mu (c_{i})\sum _{j=1,j\ne i}^{n}\mu (c_{j}) \end{aligned}$$
(29)

Since every \(\mu (c_{i})>0\), it is easy to verify that the above equation is equivalent to the desired condition (20). \(\square \)

4 Numerical examples

The HRE method can be useful in many situations in which, based on expert subjective opinions and actual data, new concepts, objects or entities need to be assessed. In order to show how the method might work in practice the following two numerical examples are presented. The first one, more abstract, discusses using the method to solve the non-linear equation system. The second, more complex, tries to put the method into an actual business context, where it can be successfully used.

In both examples the set of concepts consists of \(C_{K}\)–the reference (known) and \(C_{U}\)–the initially unknown elements. The numerical results Footnote 4 were calculated using the PairwiseComparisons Mathematica \(^{\mathrm{TM}}\) package.Footnote 5

4.1 Example I (Scientific entities assessment)

Let \(c_{1}, \ldots , c_{5}\) represent research units,Footnote 6 where two of them \(\mathbf {c_{2},c_{3}}\in C_{K}\) are the reference units. Their values have been arbitrarily set by experts to \(\upmu (\mathbf{c}_{\mathbf{2}})=\mathbf{5}\) and \(\upmu (\mathbf{c}_{\mathbf{3}})=\mathbf{7}\), whilst the other values \(\mu (c_{1}),\mu (c_{4})\) and \(\mu (c_{5})\) are considered unknowns. To determine the missing values the experts compare \(c_{1},\ldots ,c_{5}\) in pairs and propose the values \(m_{ij}\) defining the relative priority of \(c_{i}\) with respect to \(c_{j}\). An exception is the pair \((c_{2},\, c_{3})\) for which the ratios \(m_{23}=5/7\) and \(m_{32}=7/5\) follow directly from the initially known values \(\mu (c_{2})\) and \(\mu (c_{3})\).

The expert analysis of the scientific achievements of the units \(c_{1},c_{4}\) and \(c_{5}\) leads to the following PC matrix Footnote 7 \(M=[m_{ij}]\):

$$\begin{aligned} M=\left[ \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} \frac{3}{5} &{} \frac{4}{7} &{} 2 &{} \frac{16}{7}\\ \frac{5}{3} &{} 1 &{} {\frac{\mathbf {5}}{7}} &{} \frac{5}{2} &{} \frac{10}{3}\\ \frac{7}{4} &{} {\frac{\mathbf {7}}{5}} &{} 1 &{} \frac{7}{2} &{} 4\\ \frac{1}{2} &{} \frac{2}{5} &{} \frac{2}{7} &{} 1 &{} \frac{4}{3}\\ \frac{7}{16} &{} \frac{3}{10} &{} \frac{1}{4} &{} \frac{3}{4} &{} 1 \end{array}\right] \end{aligned}$$
(30)

Hence, for example, according to the expert judgment the relative priority of \(c_{1}\) with respect to \(c_{2}\) is \(m_{12}=3/5\). Since the matrix is reciprocal it holds that \(m_{ij}=1/m_{ji}\).

In accordance with the proposed hypothesis (9), to calculate the rank using HRE with the geometric averaging heuristics, the following system of non-linear equations (compare with 10) needs to be solved:

(31)

thus, after raising both sides of the equations to the power,

(32)

Substituting the logarithm of both sides of the equations, we get the following system:

(33)

which leads to the intermediate, linear logarithmic equation system:

$$\begin{aligned} \begin{array}{l} 4\lg \mu (c_{1})-\lg \mu (c_{4})-\lg \mu (c_{5}) = b_{1}\\ -\lg \mu (c_{1})+4\lg \mu (c_{4})-\lg \mu (c_{5}) = b_{4}\\ -\lg \mu (c_{1})-\lg \mu (c_{4})+4\lg \mu (c_{5}) = b_{5} \end{array} \end{aligned}$$
(34)

where

$$\begin{aligned} \begin{array}{ccc} b_{1} &{} \overset{{df}}{=} &{} \lg \left( m_{1,2}\mu (c_{2})m_{1,3}\mu (c_{3})m_{1,4}m_{1,5}\right) \\ b_{4} &{} \overset{{df}}{=} &{} \lg \left( m_{4,1}m_{4,2}\mu (c_{2})m_{4,3}\mu (c_{3})m_{4,5}\right) \\ b_{5} &{} \overset{{df}}{=} &{} \lg \left( m_{5,1}m_{5,2}\mu (c_{2})m_{5,3}\mu (c_{3})m_{5,4}\right) \end{array} \end{aligned}$$
(35)

Then, according to the procedure proposed in (Sect. 3.1), the linear equation system (14) where the unknown values \(\hat{{\upmu }}(c_{i})\overset{{df}}{=}\lg \left( {\upmu }(c_{i})\right) \) for \(i=1,4,5\) takes the form:

$$\begin{aligned} \left[ \begin{array}{ccc} n-1 &{} -1 &{} -1\\ -1 &{} n-1 &{} -1\\ -1 &{} -1 &{} n-1 \end{array}\right] \left[ \begin{array}{c} \widehat{\mu }(c_{1})\\ \widehat{\mu }(c_{4})\\ \widehat{\mu }(c_{5}) \end{array}\right] =\left[ \begin{array}{c} b_{1}\\ b_{4}\\ b_{5} \end{array}\right] \end{aligned}$$
(36)

hence, numerically:

$$\begin{aligned} \left[ \begin{array}{ccc} 4 &{} -1 &{} -1\\ -1 &{} 4 &{} -1\\ -1 &{} -1 &{} 4 \end{array}\right] \left[ \begin{array}{c} \widehat{\mu }(c_{1})\\ \widehat{\mu }(c_{4})\\ \widehat{\mu }(c_{5}) \end{array}\right] =\left[ \begin{array}{c} 1.739\\ 0.425\\ -0.064 \end{array}\right] \end{aligned}$$
(37)

Solving the linear equation system provides us with \(\widehat{\mu }(c_{1})=0.557\), \(\widehat{\mu }(c_{4})=0.295\) and \(\widehat{\mu }(c_{5})=0.197\) which leads to the desired result \(10^{\widehat{\mu }(c_{1})}=3.613\), \(10^{\widehat{\mu }(c_{4})}=1.973\) and \(10^{\widehat{\mu }(c_{5})}=1.574\). The non-scaled weight vector \(\mu \) supplemented by the known values \(\mu (c_{2})=5\) and \(\mu (c_{3})=7\) takes the form:

$$\begin{aligned} \mu =\left[ 3.613,\mathbf {5,7},1.973,1.574\right] ^{T} \end{aligned}$$
(38)

and after rescaling:

$$\begin{aligned} \mu _{n}=\left[ 0.189,\mathbf {0.261,0.365},0.103,0.0822\right] ^{T} \end{aligned}$$
(39)

Note that \(|C_{U}|=3\) implies that the dimensions of matrix \(\hat{A}\) are \(3\times 3\), moreover \(\det (\hat{A})\ne 0\) and \(\upmu (c_{i})>0\) for \(i=1,4,5\) (see Sect. 3.2). The inconsistency index is \(\fancyscript{S}(M)=0.003<0.1\). Thus, according to Saaty criterion, the matrix is consistent enough to be a reliable input to the ranking procedure.

Forgetting for a moment that the values \(\mu (c_{2})\) and \(\mu (c_{3})\) are known, the ranking for \(c_{1},\ldots ,c_{5}\) can be calculated on the basis of M only using the eigenvalue based and geometric mean methods (2, 4). The appropriate (scaled) weight vectors are: \(\mu _{{ev}}\approx \mu _{{gm}}=[0.189\), \(0.274\), \(0.35\), \(0.103\), \(0.082]^{T}\). Solving the logarithmic least square optimization problem (5) using the Nelder-Mead method [21] with constraints \(\mu (c_{2})/\mu (c_{3})=5/7\) and \(\mu (c_{1})>0,\ldots ,\mu (c_{5})>0\), leads to the priorities \(\mu _{{nm}}=[0.189\), \(0.261\), \(0.365\), \(0.103\), \(0.0822]^{T}\). The weight vector obtained using the HRE additive method [17] is \(\mu _{{HREa}}=[0.19\), \(0.26\), \(0.364\), \(0.103\), \(0.0824]^{T}\).

It is worth to note that all the rankings preserve the same order of preferences, i.e. \(c_{3}\) gets the highest rank, next there are \(c_{2},c_{1},c_{4}\) and \(c_{5}\). They differ in the numeric details. As can be expected (Sect. 3.3) the vectors \(\mu _{n}\) and \(\mu _{{nm}}\) are almost identical.

4.2 Example II (Choosing the best TV show)

A certain TV broadcaster wants to produce a new entertainment TV show in a European country. It considers purchasing the license for one of five different entertainment shows produced in the United States. So far in Europe three similar programs have been broadcasted. Through market research the approximate size of their European audience is known. They are 5,500,000, 4,500,000 and 4,950,000 persons for programs \(c_{6},c_{7}\) and \(c_{8}\) respectively. The production costs of these programs are similar. In order to select possibly the most profitable TV show the station hires a few seasoned media experts. During the expert panel they prepare the following PC matrix \(M\), representing the relative attractiveness of all the considered programs.

$$\begin{aligned} M=\left[ \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 0.8 &{} 0.968 &{} 0.7 &{} 0.5 &{} 0.682 &{} 0.908 &{} 0.667\\ 1.25 &{} 1 &{} 1.211 &{} 0.994 &{} 0.757 &{} 0.75 &{} 1.27 &{} 0.833\\ 1.033 &{} 0.826 &{} 1 &{} 0.821 &{} 0.551 &{} 0.704 &{} 0.826 &{} 0.625\\ 1.429 &{} 1.006 &{} 1.218 &{} 1 &{} 0.714 &{} 0.857 &{} 1.050 &{} 0.952\\ 2 &{} 1.321 &{} 1.817 &{} 1.401 &{} 1 &{} 1.2 &{} 1.467 &{} 1.333\\ 1.467 &{} 1.333 &{} 1.42 &{} 1.167 &{} 0.833 &{} 1 &{} \mathbf {1.222} &{} \mathbf {1.111}\\ 1.101 &{} 0.8 &{} 1.211 &{} 0.952 &{} 0.682 &{} \mathbf {0.818} &{} 1 &{} \mathbf {0.909}\\ 1.499 &{} 1.2 &{} 1.6 &{} 1.05 &{} 0.75 &{} \mathbf {0.9} &{} \mathbf {1.1} &{} 1 \end{array}\right] \end{aligned}$$
(40)

In the matrix \(M\) every entry \(m_{ij}\) corresponds to a ratio describing attractiveness of the TV show \(c_{i}\) with respect to the attractiveness of TV show \(c_{j}\). Since the values of attractiveness for \(c_{6},c_{7}\) and \(c_{8}\) are known (they are approximated by the number of people watching the given TV show), thus the appropriate ratios \(m_{ij}\) for \(i,j=6,7,8\) are not the subject of expert judgment. Instead, they are calculated based on data from market research. For example:

$$\begin{aligned} \mathbf {m}_{\mathbf {6,7}}=\frac{\mu (c_{6})}{\mu (c_{7})}=\frac{5{,}100{,}000}{4{,}500{,}000}=\mathbf {1.222} \end{aligned}$$
(41)

or

$$\begin{aligned} \mathbf {\mathbf {m}_{6,8}}=\frac{\mu (c_{6})}{\mu (c_{8})}=\frac{5{,}100{,}000}{4{,}950{,}000}=\mathbf {1.111} \end{aligned}$$
(42)

The other entries of \(M\) represent the subjective judgements of experts.

Like before, to find a solution with the help of HRE supported by the geometric averaging heuristics, the system of equations (10) must be solved. The desired values \({\upmu }(c_{i})\) for \(i=1\ldots ,5\) will be derived from the formula \(\widehat{\mu }(c_{i})=\log {\upmu }(c_{i})\). Because \(|C_{U}|=5\), the dimensions of matrix \(\widehat{A}\) are \(5\times 5\). The linear equation system need to be solved as follows:

$$\begin{aligned} \left[ \begin{array}{ccccc} n-1 &{} -1 &{} -1 &{} -1 &{} -1\\ -1 &{} n-1 &{} -1 &{} -1 &{} -1\\ -1 &{} -1 &{} n-1 &{} -1 &{} -1\\ -1 &{} -1 &{} -1 &{} n-1 &{} -1\\ -1 &{} -1 &{} -1 &{} -1 &{} n-1 \end{array}\right] \left[ \begin{array}{c} \begin{array}{c} \widehat{\mu }(c_{1})\\ \widehat{\mu }(c_{2})\\ \widehat{\mu }(c_{3}) \end{array}\\ \begin{array}{c} \widehat{\mu }(c_{4})\\ \widehat{\mu }(c_{5}) \end{array} \end{array}\right] =\left[ \begin{array}{c} b_{1}\\ b_{2}\\ b_{3}\\ b_{4}\\ b_{5} \end{array}\right] \end{aligned}$$
(43)

where

$$\begin{aligned} \begin{array}{ccc} b_{1} &{} \overset{{df}}{=} &{} \lg \left( m_{1,2}m_{1,3}m_{1,4}m_{1,5}m_{1,6}\mu (c_{6})m_{1,7}\mu (c_{7})m_{1,8}\mu (c_{8})\right) \\ b_{2} &{} \overset{{df}}{=} &{} \lg \left( m_{2,1}m_{2,3}m_{2,4}m_{2,5}m_{2,6}\mu (c_{6})m_{2,7}\mu (c_{7})m_{2,8}\mu (c_{8})\right) \\ b_{3} &{} \overset{{df}}{=} &{} \lg \left( m_{3,1}m_{3,2}m_{3,4}m_{3,5}m_{3,6}\mu (c_{6})m_{3,7}\mu (c_{7})m_{3,8}\mu (c_{8})\right) \\ b_{4} &{} \overset{{df}}{=} &{} \lg \left( m_{4,1}m_{4,2}m_{4,3}m_{4,5}m_{4,6}\mu (c_{6})m_{4,7}\mu (c_{7})m_{4,8}\mu (c_{8})\right) \\ b_{5} &{} \overset{{df}}{=} &{} \lg \left( m_{5,1}m_{5,2}m_{5,3}m_{5,4}m_{5,6}\mu (c_{6})m_{5,7}\mu (c_{7})m_{5,8}\mu (c_{8})\right) \end{array} \end{aligned}$$
(44)

hence, (43) numerically:

$$\begin{aligned} \left[ \begin{array}{ccccc} 7 &{} -1 &{} -1 &{} -1 &{} -1\\ -1 &{} 7 &{} -1 &{} -1 &{} -1\\ -1 &{} -1 &{} 7 &{} -1 &{} -1\\ -1 &{} -1 &{} -1 &{} 7 &{} -1\\ -1 &{} -1 &{} -1 &{} -1 &{} 7 \end{array}\right] \left[ \begin{array}{c} \begin{array}{c} \widehat{\mu }(c_{1})\\ \widehat{\mu }(c_{2})\\ \widehat{\mu }(c_{3}) \end{array}\\ \begin{array}{c} \widehat{\mu }(c_{4})\\ \widehat{\mu }(c_{5}) \end{array} \end{array}\right] =\left[ \begin{array}{c} \begin{array}{c} 19.137\\ 20.037 \end{array}\\ 19.234\\ \begin{array}{c} 20.117\\ 21.285 \end{array} \end{array}\right] \end{aligned}$$
(45)

The intermediate result vector is:

$$\begin{aligned} \widehat{\mu }=\left[ 6.551,\,6.664,\,6.563,\,6.674,\,6.82\right] ^{T} \end{aligned}$$
(46)

Hence, following the rule \(\mu (c_{i})=\xi ^{\widehat{\mu }(c_{i})}\), where \(\xi =10\) is the logarithm base, the final result vector is calculated.

$$\begin{aligned} \mu =\left[ \begin{array}{c} 3,556,858\\ 4,608,487\\ 3,657,533\\ 4,716,180\\ 6,601,181 \end{array}\right] \end{aligned}$$
(47)

The rescaled full priority vector is \(\mu _{n}=[0.0934,\) \(0.121,\) \(0.096,\) \(0.124,\) \(0.173,\) \(\mathbf {0.144},\) \(\mathbf {0.118},\) \(\mathbf {0.13}]^{T}\). Thus, according to the expert judgments and the market research TV show number \(5\) (denoted as \(c_{5}\)) has the chance to gather nearly \(6.5\) million people in front of their TVs, whilst the second one in line can gather “only” \(4.6\) million people. Based on this estimate, the board of directors representing the broadcaster decides to recommend purchasing the license for the fifth presented TV show. Similarly as before \(\fancyscript{S}(M)=0.002\) is much smaller than \(0.1\). Thus, the matrix \(M\) seems to be consistent enough to recognize the result obtained as relevant.

The eigenvalue based and geometric mean ranks calculated for \(M\) (solely) are \(\mu _{{ev}}\approx \mu _{{gm}}=[0.0933,\,0.121,\,0.0959,\,0.124,\,0.173,\,0.144,\,0.113,\,0.136]^{T}\). The HRE additive method for the pair \(M\) and \(C_{K}\) returns \(\mu _{{HREa}}=[0.093,\) \(0.122,\) \(0.096,\) \(0.124,\) \(0.173,\) \(\mathbf {0.144},\) \(\mathbf {0.117},\) \(\mathbf {0.129}]^{T}\). Solving the logarithmic least square optimization problem (5) using the Nelder-Mead method [21] with the additional constraints \(\mu (c_{6})/\mu (c_{7})=1.222\), \(\mu (c_{6})/\mu (c_{7})=1.111\), \(\mu (c_{7})/\mu (c_{8})=0.909\), and \(\mu (c_{1})>0,\ldots ,\mu (c_{8})>0\), leads to the following priority vector: \(\mu _{{nm}}=[0.0934,\,\) \(0.121,\) \(0.096,\) \(0.123,\) \(0.173,\) \(0.144,\) \(0.118,\) \(0.13]^{T}\). Similarly as before, all the vectors preserve the same rank order, which is \(c_{5}\), \(c_{6}\), \(c_{8}\), \(c_{4}\), \(c_{2}\), \(c_{7}\) and \(c_{1}\). Similarly as before the vectors \(\mu _{n}\) and \(\mu _{{nm}}\) are almost identical.

5 Summary

The presented geometric HRE approach is another solution to the problem of rankings with the reference set. It proposes to use a geometric mean instead of the arithmetic one used in [16, 17]. The advantage of this approach is the robustness of the procedure. As it was shown in (Sect. 3.2) the proposed solution works for an arbitrary set of input data producing an admissible vector of weights. The resulted ranking very often turns out to be optimal in the sense of the magnitude of multiplicative errors. According to the formulated and proven condition (Sect. 3.3), this happens when the differences between the resulting priorities are not too large. Therefore, the geometric HRE approach is also a method of solving the logarithmic least square constrained optimization problem.

The HRE approach may be useful in many different situations, including ranking creation, valuation of goods and services, risk assessment and others. Due to the lack of restrictions on the input PC matrix (there is a guarantee that a method with the geometric mean always produces an admissible result), the scope of the applicability of the HRE method increases. Thus, the presented method covers cases (e.g. rankings with non-reciprocal PC matrices) which can not always be dealt with using the arithmetic mean heuristics. The HRE method may also be useful for the AHP users. In this context, for the non-empty reference set \(C_{K}\), it may serve as an alternative for the standard priority deriving methods such as eigenvalue based method [25] or geometric mean method [6].

Despite these encouraging results, much remains to be done. In particular, the role of the inconsistency in the input matrix \(M\) should be more deeply investigated. Also the existence of solution for the “additive” HRE approach needs to be better researched. Of course, the more examples are studied the better. Thus, further development of the method will be particularly focused on the study and analysis of examples of its use.