1 Introduction

It is commonly known that covariance matrices for observations play an important role in constructing the stochastic models for geodetic networks. Possessing reliable values of the elements of covariance matrices in satellite and ground positioning systems is of considerable significance in estimating the actual accuracy of final coordinates (Wang et al. 2012). Also, the design of such systems based on the knowledge of potential effects of observation correlations may contribute to achieving high reliability of the systems.

The covariance matrices, but actually the correlation matrices contained in them, are the weak (in terms of accuracy) elements of stochastic models built for networks. For correlation matrices that are obtained by estimation in data post-processing the accuracy of non-diagonal elements can undergo steady improvement due to the research being carried out on the methods of estimating the covariance matrices (e.g., Ananga et al. 1994; Leandro et al. 2005). The most difficult situation and not solved as yet takes place at the stage of network design where the knowledge on non-diagonal terms in correlation matrices is usually small, or sometimes even none. That calls for the need to work out the methods of constructing something like substitute matrices on basis of the possessed knowledge on correlation terms.

Seeking the tools for investigating the effect of observation correlations on network reliability, we have then to cover both the case of a specified correlation matrix and the substitute one. So, the tools should be applicable to any arbitrary correlation matrix.

The problem of the effect of observation correlations on network internal reliability has been undertaken in Prószyński and Kwaśniak (2016). The investigations were confined to the elementary case termed there uniform correlation, i.e., where all the non-diagonal elements of the correlation matrix are of equal values (positive or negative). Such a matrix that played there a role of a substitute matrix is known in financial mathematics as constant correlation matrix, e.g., Tiit and Helemae (1997), Dufresne (2005). In those publications, one can find a rigorously derived, necessary, and sufficient condition for positive definiteness of such a matrix. The set of constant correlation matrices \((n\times n)\) can be immediately ordered according to the increasing values of the constant term. The problem is with ordering the set of possible configurations of an arbitrary \((n \times n)\) correlation matrix. An appropriate measure would be needed for this purpose.

The study of the elementary case of constant correlation matrices indicated the way to a more generalized approach and is assumed as reference material in the present investigations.

The objectives of the paper, covering the problems outlined above, can be formulated in the following way:

  • to extend the former studies by seeking a global measure of observation correlations covering all possible configurations of correlation matrices. The measure should allow ranking of matrix configurations;

  • to create theoretical basis for operating with any correlation matrices in the analyses of correlation effect;

  • to construct analytical tools for investigating the structure of correlation effect on internal and external reliability of networks for any arbitrary correlation matrices;

  • to propose levels of global correlation in networks.

As in the previous paper of authors, the research is confined to a single outlier case.

Taking into account all the problems to be solved, reflected in the above objectives, the primary motivation to undertake the research was the fact that at the stage of network design the knowledge on non-diagonal terms in a correlation matrix for observations is usually small or none.

2 A proposed global measure of observation correlations in a network

Determinant of a correlation matrix is a measure of collinearity of explanatory variables in linear regression. High correlation between these variables has a negative influence upon the effectiveness of estimates of the model parameters. The determinant takes the values within the interval \(\left\langle {0,1} \right\rangle \). Since 0 corresponds to full correlation (i.e., 1) and 1 corresponds to the lack of correlation (i.e., 0), to get a measure ordered according to the increase of correlation, we introduce for a positive definite matrix \(\mathbf{C}_{\mathrm{s}}\) the following quantity

$$\begin{aligned} {\uprho }_{\mathrm{G}} =\sqrt{1-\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \quad {\uprho }_{\mathrm{G}} \in <0,1) \end{aligned}$$
(1)

where \({\uprho }_{\mathrm{G}} \) is termed a global measure of stochastic relationships between the observations in a network, or for short—a measure of global correlation. It only takes positive values like a coefficient of determination (e.g., Anderson-Sprecher 1994). For the quality of any measure, it is recommended that it should be a one-to-one mapping of all possible values of the characterized quantity. Looking from the algebraic point of view, we can observe that the function \(f :\mathbf{C}_{\mathrm{s}} \rightarrow \hbox {det}\mathbf{C}_{\mathrm{s}} \) (or \(f :\mathbf{C}_{\mathrm{s}} \rightarrow {\uprho }_{\mathrm{G}} )\) is not injective (i.e., is not a one-to-one function). This means that although a specified \(\mathbf{C}_{\mathrm{s}} \) configuration corresponds to only one value of \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \), a specified value of \(\hbox {det}{} \mathbf{C}_{\mathrm{s}}\) (or \({\uprho }_{\mathrm{G}} )\) may correspond to a certain number of \(\mathbf{C}_{\mathrm{s}} \) configurations. Such numbers expressed in terms of frequencies are shown in Fig. 1 on exemplary diagrams obtained by numerical simulation of the \(\mathbf{C}_{\mathrm{s}} \) configurations for \(n~=~3, 4, 5\) (3 sets of 50,000 simulations were used for each graph). The simulations were based on the algorithm termed “accept-reject.” It consisted in drawing each non-diagonal element of the matrix \(\mathbf{C}_{\mathrm{s}}\) from the uniform distribution and rejecting those of the resulting matrix configurations that were not positive definite matrices. As stated in Numpacharoen and Atsawarungruangkit (2012) and Budden et al. (2008), such an algorithm is effective only for small values of n. In the present research, it was found empirically that this algorithm, being free of any algebraic operations, does not affect the frequencies of matrix configurations as is the case with some faster algorithms.

Fig. 1
figure 1

Frequencies of \(\mathbf{C}_{\mathrm{s}}\) configurations corresponding to \({\uprho }_{\mathrm{G}}\) intervals of 0.01 width (averages of 3 sets of 50,000 simulations) for \(n~=~3, 4, 5\)

In Fig. 1 we can observe an advantageous trend that with the increase of n, the percent of \(\mathbf{C}_{\mathrm{s}} \) configurations increases significantly only for ever greater values of \({\uprho }_{\mathrm{G}}\) where \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \) starts asymptotic ascend toward 1. Hence, the quality of the measure in the analyzed aspect of the function injectivity improves with the increase of n.

To approximate the actual numbers of \(\mathbf{C}_{\mathrm{s}}\) configurations corresponding to a specified value of \({\uprho }_{\mathrm{G}} \) more accurately than on basis of the 0.01-width intervals, the numbers sought for were determined for the 0.0001-width intervals located at \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \) values 0.1, 0.2,..., 0.9 and 0.9999 (see Table 1). The results of the above-mentioned simulations were used for this purpose. The zeroes in Table 1, being more numerous for greater n, are merely the result of the problems with generating the correlation matrices and very low probabilities of the investigated events.

Initially, we tried to get the above curves (hopefully, also for greater n) by means of the probabilistic approach involving determination of density of \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \), treated as a random variable (e.g., Hanea and Nane 2016). That would require a rigorous probabilistic definition of the elements of \(\mathbf{C}_{\mathrm{s}}\) and would yield the results dependent on that definition. Finally, we chose the above-mentioned algebraic approach as satisfactory to show the degree of departure from injectivity for the function f :\(\mathbf{C}_{\mathrm{s}} \rightarrow \hbox {det}\mathbf{C}_{\mathrm{s}} \).

Table 1 Numbers of the \(\mathbf{C}_{s}\) configurations obtained within 0.0001-width intervals (mean values of 3x50,000 simulations)

In spite of the function \(\mathbf{C}_{\mathrm{s}} \rightarrow \hbox {det}{} \mathbf{C}_{\mathrm{s}} \) (or \(\mathbf{C}_{\mathrm{s}} \rightarrow {\uprho }_{\mathrm{G}}\)) being not injective, the use of \({\uprho }_{\mathrm{G}} \) makes possible, although with some ambiguity, the ranking of configurations of a correlation matrix. As follows from Fig. 1, the ambiguities for n = 5 are relatively small for up to \({\uprho }_{\mathrm{G}} < 0.85\) (i.e., frequency \({<}\)2%). However, they will increase abruptly for higher values of \({\uprho }_{\mathrm{G}} \).

With constant correlation matrices, we get a specific measure of observation correlations denoted in Prószyński and Kwaśniak (2016) by the symbol a. The measure provides ambiguity-free ranking of constant correlation matrices. It is related to \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \) by the formula (Dufresne 2005)

$$\begin{aligned} \hbox {det }{} \mathbf{C}_{\mathrm{s}}^n (a)=(1-a)^{n-1}\left[ {1+(n-1)a} \right] \qquad -\frac{1}{n-1}<a<1 \end{aligned}$$
(2)

where \(\mathbf{C}_{\mathrm{s}}^n \) denotes \(\mathbf{C}_{\mathrm{s}} (n\times n)\).

Due to (1) we get on basis of (2)

$$\begin{aligned}&{\uprho }_{\mathrm{G}}^n (a)=\sqrt{1-(1-a)^{n-1}\left[ {\hbox {1}+(n-1)a} \right] } \nonumber \\&~~~~~~~~~~~\quad -\frac{1}{n-1}<a<1. \end{aligned}$$
(3)

The relationship between a and \({\uprho }_{\mathrm{G}} \) for \(n~=~2, 3, {\ldots }, 50\) is shown in Fig. 2. All the curves approach \({\uprho }_{\mathrm{G}} = 1\) asymptotically. We can see that with the increase of n, the same values of \({\uprho }_{\mathrm{G}} \) correspond to diminishing values of a.

Fig. 2
figure 2

The relationship between a and \({\uprho }_{\mathrm{G}} \)

We can easily connect the global measure \({\uprho }_{\mathrm{G}} \) with the index of multiple correlation \({\uprho }_{(i)} \) for the ith observation \((i~=~1, {\ldots }, n)\), i.e.,

$$\begin{aligned} {\uprho }_{(i)}^2 =1-\frac{\hbox {det}{} \mathbf{C}_{\mathrm{s}} }{\hbox {det}{} \mathbf{C}_{\mathrm{s}(i)} }=1-\frac{1-{\uprho }_{\mathrm{G}}^2 }{\hbox {det}{} \mathbf{C}_{\mathrm{s}(i)} } \end{aligned}$$
(4)

where \(\hbox {det}{} \mathbf{C}_{\mathrm{s}(i)} \) is obtained after deleting the ith row and the ith column in \(\mathbf{C}_{\mathrm{s}} \).

3 Auxiliary indices connecting a global measure \({\uprho }_{\mathrm{G}}\) with the elements of correlation matrix

The definition of the global measure (1) does not provide information about the magnitudes of non-diagonal elements of \(\mathbf{C}_{\mathrm{s}} \). Hence, the following auxiliary indices that may provide some more detailed description of a global measure \({\uprho }_{\mathrm{G}} \) are proposed and investigated

$$\begin{aligned}&\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{max}} ; \quad \left( {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right) _{\mathrm{qm}} =\sqrt{\frac{\sum {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij}^2 } }{n(n-1)/2}}; \nonumber \\&\quad i, j~=~1, {\ldots }, n; j~>~i \end{aligned}$$
(5)

where \(\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} \) is a non-diagonal element of \(\mathbf{C}_{\mathrm{s}} \); the subscript “qm” denotes quadratic mean.

The indices are related to each individual \(\mathbf{C}_{\mathrm{s}} \) configuration that can be generated for a specified value of \({\uprho }_{\mathrm{G}} \).

Here are the properties of the above indices, proven immediately for a specific case \(n~=~2\) and due to complexity of the task for \(n\ge 3\) confirmed only on basis of numerical testing (for \(n~=~3, 5, 9\)).

Fig. 3
figure 3

Illustration of Property 1 and Property 2 (for \(n \ge 3\)); MAV—interval concerning maximum absolute value, QM—interval concerning quadratic mean

Fig. 4
figure 4

Variability of \({\uprho }_{G}, a^{+}\) and \(\left| {a^{-}} \right| \) with respect to \(\hbox {det}{} \mathbf{C}_\mathrm{s}\) for some values of n

Property 1

For a positive definite correlation matrix \(\mathbf{C}_{\mathrm{s}} (n\times n),\hbox { }n\ge 2\) and an arbitrarily chosen value of a global correlation measure \({\uprho }_{\mathrm{G}} \), denoted as , the following relationships occur

(6)
(7)

where is that smaller of the two real roots of the function satisfying (3).

The specific case (6) can be proved immediately. We can write , where x denotes \(\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{12} \), and hence, we get the equality (6).

The numerically based confirmation for a general case (7) is presented in Appendix B.

Property 2

For a positive definite correlation matrix \(\mathbf{C}_{\mathrm{s}} (n\times n),\hbox { }n\ge 2\) and an arbitrarily chosen value of a global correlation measure \({\uprho }_{\mathrm{G}} \), denoted as , the following relationships occur

(8)
(9)

where (as above) and are real roots of the function satisfying (3).

The specific case (8) can be proved immediately. We can write , where x denotes \(\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{12} \). Since \(\left( {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{12} } \right) _{\mathrm{qm}} =\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{12} } \right| \), we get the equality (8).

The numerically based confirmation for a general case (9) is presented in Appendix B.

Both the properties are illustrated in Fig. 3. The intervals given by (7) and (9) are denoted there as MAV and QM, respectively.

The problem of finding \(a^{-}\) and \(a^{+}\) for a specified value of \({\uprho }_{\mathrm{G}} \) and \(n\ge 3\) is a complex task of finding the real roots of the polynomial of the nth order, each root satisfying the condition \(-\frac{1}{n-1}<a<1\). This can be done in an iterative way. We used for this purpose a Solver being a Microsoft Excel add-in program.

Fig. 5
figure 5

Variability of the scale factor q with respect to \({\uprho }_{\mathrm{G}} \) and with respect to a

For uniform correlation matrices, i.e., \(\mathbf{C}_{\mathrm{s}} (a)\) as in Sect. 2, we have \(\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{max}} =a\) and \( \left( {\left\{ {\mathbf{C}_{\mathrm{s}} (a)} \right\} _{ij} } \right) _{\mathrm{qm}} =a.\)

For proper interpretation of Property 1 and Property 2, we show in Fig. 4 a mutual arrangement of the \({\uprho }_{G}\), \(a^{+}\) and \(\left| {a^{-}} \right| \) curves with the increase of n (except for \(n~=~2\)).

For \(n~=~2\) (not shown in Fig. 4) the \(\hbox {a}^{+}\)and \(\left| {a^{-}} \right| \) curves coincide with the \({\uprho }_{G}\) curve. With the increase of n, the \(a^{+}\)and \(\left| {a^{-}} \right| \) curves move away downward from the \({\uprho }_{G}\) curve while diminishing their mutual separation. For \(n~=~100\), they only slightly depart from the \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \) axis on a long section, but close to \(\hbox {det}{} \mathbf{C}_{\mathrm{s}}= 0\) the \(a^{+}\) curve starts to ascend asymptotically to 1.

4 A representation of correlation matrix based on the proposed global measure

Using a formula for the inverse of the correlation matrix \(\mathbf{C}_{\mathrm{s}}\) (positive definite), i.e.,

$$\begin{aligned} \mathbf{C}_{\mathrm{s}}^{-1} =\frac{1}{\hbox {det}{} \mathbf{C}_{\mathrm{s}} }\mathbf{C}_{\mathrm{s}}^{*} \end{aligned}$$
(10)

where \(\mathbf{C}_{\mathrm{s}}^{*} \) is the adjugate (positive definite), we may represent \(\mathbf{C}_{\mathrm{s}} (n\times n),\hbox { }n\ge 2\) in the form

$$\begin{aligned} \mathbf{C}_{\mathrm{s}} =\hbox {det}{} \mathbf{C}_{\mathrm{s}} \cdot \left( {\mathbf{C}_{\mathrm{s}}^{*} } \right) ^{-1} \end{aligned}$$
(11)

where \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \) is a scale factor common for all the n observations in a network.

Since \(\hbox {det}{} \mathbf{C}_{\mathrm{s}}^{*} \ne 1\), \(\hbox {det}\mathbf{C}_{\mathrm{s}} \) as in (11) is not a complete scale factor, and the remaining part of it must be contained in \(\left( {\mathbf{C}_{\mathrm{s}}^{*} } \right) ^{-1}\). A representation of \(\mathbf{C}_{\mathrm{s}} \) showing a complete scale factor is specified by the theorem below. The theorem was formulated and proved within the present research.

Theorem 1

A positive definite correlation matrix \(\mathbf{C}_{\mathrm{s}} (n\times n),n\ge 2\) can be represented as

$$\begin{aligned} \mathbf{C}_{\mathrm{s}} =q\cdot \mathbf{R} \end{aligned}$$
(12)

where

$$\begin{aligned} q= & {} \left( {\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \right) ^{1/n}\text {- a complete scale factor},\\ \mathbf{R}= & {} \left( {\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \right) ^{(n-1)/n}\cdot \left( {\mathbf{C}_{\mathrm{s}}^{*} } \right) ^{-1}\text {- a matrix with determinant}\\&\text { equal to 1} \end{aligned}$$

The proof is immediate.

Assuming \(\mathbf{C}_{\mathrm{s}} =q\cdot \mathbf{R}\), where detR = 1, we get \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} =q^{n}\cdot \hbox {det}\mathbf{R}=q^{n}\) and hence, \(q=\left( {\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \right) ^{1/n}\). On basis of (11) we write \(\hbox {det}{} \mathbf{C}_{\mathrm{s}} \cdot \left( {\mathbf{C}_{\mathrm{s}}^{*} } \right) ^{-1}=\left( {\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \right) ^{\hbox {1/}n}\cdot \mathbf{R}\), and finally we get \(\mathbf{R}=\left( {\hbox {det}{} \mathbf{C}_{\mathrm{s}} } \right) ^{(n-\hbox {1)/}n}\left( {\mathbf{C}_{\mathrm{s}}^{*} } \right) ^{-1}\), what ends the proof.

We can readily prove that the representation (12) applies to non-singular matrices \((n\times n),\hbox { }n\ge 2\) with a determinant greater than zero, and also to non-singular matrices \((n\times n),\hbox { }n\ge 3\) where n is an odd number. The asymmetry of a matrix does not affect this type of representation.

In analogy to the well-known relationship between the covariance matrix and the weight matrix, i.e., \(\mathbf{C}=\sigma _{\mathrm{o}}^2 \cdot \mathbf{P}^{-1}\) with \(\sigma _{\mathrm{o}}^2 \) being a variance factor, we express (11) in a notation that introduces the proposed global correlation measure \({{\uprho } }_{\mathrm{G}} \) and considers R as a rescaled correlation matrix

$$\begin{aligned} \mathbf{C}_{\mathrm{s}} =q({\uprho }_{\mathrm{G}} )\cdot \mathbf{P}_{\mathrm{s}}^{-1} \end{aligned}$$
(13)

where

$$\begin{aligned}&q({\uprho }_{\mathrm{G}} )\text { the complete scale factor expressed in terms of }{\uprho }_{\mathrm{G}} ;\\&\qquad q({\uprho }_{\mathrm{G}} )=\left( {1-{\uprho }_{\mathrm{G}}^2 } \right) ^{1/n}\\&\mathbf{P}_{\mathrm{s}}\text {-internal weight matrix; det}{} \mathbf{P}_{\mathrm{s}} ~=~1. \end{aligned}$$

We can readily check that for \({\uprho }_{\mathrm{G}} \rightarrow 0\) or \({\uprho }_{\mathrm{G}} \rightarrow 1\) the complete scale factor \(q({\uprho }_{\mathrm{G}} )\) tends asymptotically to 1 or 0, respectively. The relationship \(q({\uprho }_{\mathrm{G}} )=\left( {1-{\uprho }_{\mathrm{G}}^2 } \right) ^{1/n}\) is shown in Fig. 5a, and as a function of a constant correlation term a in Fig. 5b.

With the increase in global correlation, the scale factor decreases. The greater the values of n, the longer are the intervals where the scale factor does not depart from 1 significantly (see Fig. 5a).

5 Transforming basic reliability expressions to learn about the character of correlation effect

We show that representation of the correlation matrix as in the formula (13) can be useful in analyses of the structure of network reliability measures with respect to observation correlations.

Below, there is a list of basic expressions and quantities used in theory of reliability for correlated observations. They all refer to the modified (i.e., standardized) form of the Gauss–Markov model (GMM), that exposes the correlation matrix \(\mathbf{C}_{\mathrm{s}} \) (Prószyński 2010), i.e.,

$$\begin{aligned} \mathbf{A}_{\mathrm{s}} \mathbf{x}+\mathbf{e}_{\mathrm{s}} =\mathbf{y}_{\mathrm{s}} \quad \mathbf{e}_{\mathrm{s}} \sim (0,\mathbf{C}_{\mathrm{s}} ) \end{aligned}$$
(14)

where: \(\mathbf{x}(u\times 1)\), \(\mathbf{y}_{\mathrm{s}} (n\times 1)\), \(\mathbf{A}_{\mathrm{s}} (u\times n)\), rank \(\mathbf{A}_{\mathrm{s}} \le u\), \(\mathbf{C}_{\mathrm{s}} (n\times n)\) (positive definite).

Let us remind here that the standardized model such as (14) is obtained by multiplying both sides of the original GMM by \({{\varvec{\upsigma }} }^{-1}\), where \({{\varvec{\upsigma }}}=(\hbox {diag }\mathbf{C})^{1/2}\), and transforming the covariance matrix C (positive definite) accordingly.

Following the essential classification of network reliability (Baarda 1968), the expressions and quantities are listed in two groups, the first pertaining to the internal reliability and the second pertaining to the external reliability

a) internal reliability

$$\begin{aligned} \mathbf{H}= & {} \mathbf{I}-\mathbf{A}_{\mathrm{s}} (\mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{A}_{\mathrm{s}} )^{+}{} \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{C}_{\mathrm{s}}^{-1} ; \, \mathbf{W}=\mathbf{H}-\mathbf{H}^{\mathrm{T}}\mathbf{H}; \, h_{ii} \nonumber \\= & {} \left\{ \mathbf{H} \right\} _{ii} ; \, w_{ii} =h_{ii} -\{\mathbf{H}^{\mathrm{T}}{} \mathbf{H}\}_{ii} \end{aligned}$$
(15)

\(h_{ii} ,w_{ii}\,\)—the response-based measures of internal reliability for the ith observation

$$\begin{aligned} \mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}=\mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}; \quad r_i =\{\mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}\}_{ii} ; \quad \hbox {MDB}_{\mathrm{s},i} =\frac{\sqrt{\lambda }}{\sqrt{r_i }} \end{aligned}$$
(16)

\(r_i\,\)—reliability number for the ith observation; \(\hbox {MDB}_{\mathrm{s},i}\)—minimal detectable bias for the ith observation (standardized), where \(\lambda \) is non-centrality parameter in a global model test (Wang and Chen 1994; Teunissen 1990, 2006)

$$\begin{aligned} {\uprho }_{ij} =\frac{\left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}} \right\} _{ij} }{\sqrt{\left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}} \right\} _{ii} }\sqrt{\left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}} \right\} _{jj} }} \end{aligned}$$
(17)

\({\uprho }_{ij}\)—coefficient of correlation between the outlier w-test statistics for the ith and the jth observation, as a quantity related to internal reliability

b) external reliability

$$\begin{aligned}&{} \mathbf{K}=(\mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{A}_{\mathrm{s}} )^{+}{} \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{C}_{\mathrm{s}}^{-1} ; \, {\hat{\mathbf{x}}}=\mathbf{Ky}_{\mathrm{s}} ; \nonumber \\&\Delta {\hat{\mathbf{x}}}_{(i)} =\mathbf{K}\cdot \Delta \mathbf{y}_{\mathrm{s}(i)} =\mathbf{K}\cdot \left[ {{\begin{array}{l} 0 \\ {\hbox {MDB}_{\mathrm{s},i} } \\ 0 \\ \end{array} }} \right] \nonumber \\&{\updelta }_i =\left( {\Delta {\hat{\mathbf{x}}}_{(i)} } \right) ^{\mathrm{T}}{} \mathbf{C}_{{\hat{\mathbf{x}}}}^+ \cdot \Delta {\hat{\mathbf{x}}}_{(i)} \end{aligned}$$
(18)

\({\hat{\mathbf{x}}}\)—LS estimator for x as in (14); \(\hbox {MDB}_{\mathrm{s,}i} \) as in (16); \(\Delta {\hat{\mathbf{x}}}_{(i)} \)—the vector of increments in \({\hat{\mathbf{x}}}\) due to \(\hbox {MDB}_{\mathrm{s,}i} \), \({\updelta } _i \)—a measure of external reliability (Wang and Chen 1994), given there with the use of regular inverse of \(\mathbf{C}_{{\hat{\mathbf{x}}}}\).

Additionally, we include in this analysis a correlation-dependent quantity based on a matrix H, but not considered as a measure of network reliability

$$\begin{aligned} \mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}=\mathbf{HC}_{\mathrm{s}} ; \, \mathbf{C}_{\mathbf{v} _{\mathrm{s}}}= & {} \mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}; \, {\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 =\left\{ {\mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}} \right\} _{ii}\nonumber \\= & {} 1-{\upsigma }_{\hat{{y}}_{\mathrm{s,}i} }^2 \end{aligned}$$
(19)

\(\mathbf{C}_{\mathbf{v} _{\mathrm{s}}}\)—covariance matrix of the vector of standardized LS residuals; \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2\)—variance of the ith standardized LS residual.

First, we consider the matrices H and K appearing in all the expressions in the groups (a) and (b) as above. Applying (13), we get

$$\begin{aligned} \mathbf{H}= & {} \mathbf{I}-\mathbf{A}_{\mathrm{s}} \left( \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \frac{1}{q({\uprho }_{\mathrm{G}})}{} \mathbf{P}_{\mathrm{s}} \mathbf{A}_{\mathrm{s}} \right) ^{+}{} \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \frac{1}{q({\uprho }_{\mathrm{G}} )}\mathbf{P}_{\mathrm{s}} \nonumber \\= & {} \mathbf{I}-\mathbf{A}_{\mathrm{s}} \left( \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{P}_{\mathrm{s}} \mathbf{A}_{\mathrm{s}}\right) ^{+}{} \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{P}_{\mathrm{s}} ~=~\hbox {f}_\mathbf{H} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} ) \end{aligned}$$
(20)
$$\begin{aligned} \mathbf{K}= & {} \left( \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \frac{1}{q({\uprho }_{\mathrm{G}} )}{} \mathbf{P}_{\mathrm{s}} \mathbf{A}_{\mathrm{s}}\right) ^{+}{} \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \frac{1}{q({\uprho }_{\mathrm{G}} )}{} \mathbf{P}_{\mathrm{s}} \nonumber \\= & {} \left( \mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{P}_{\mathrm{s}} \mathbf{A}_{\mathrm{s}}\right) ^{+}\mathbf{A}_{\mathrm{s}}^{\mathrm{T}} \mathbf{P}_{\mathrm{s}} ~=~\hbox {f}_\mathbf{K} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} ) \end{aligned}$$
(21)

We can see that both the matrices H and K do not depend on \(q{(\uprho }_{\mathrm{G}} )\) and, hence, do not depend on a global correlation measure \({\uprho }_{\mathrm{G}} \). They are functions of the structural matrix \(\mathbf{A}_{\mathrm{s}}\) (standardized) and the internal weight matrix \(\mathbf{P}_{\mathrm{s}} \).

Now, basing on the above properties of H and K, we analyze other quantities listed under a) and b).

Ad a) The matrix W, and hence all its diagonal elements \(w_{ii} (i=1,\ldots ,n)\) depend on the structural matrix \(\mathbf{A}_{\mathrm{s}}\) and the internal weight matrix \(\mathbf{P}_{\mathrm{s}} \). Therefore, the measures \(h_{ii} \) and \(w_{ii} \) for a given network can only undergo changes in mutual diversification for different configurations of \(\mathbf{C}_{\mathrm{s}} \) and, obviously, for different accuracies of observations.

$$\begin{aligned} \mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}= & {} \frac{1}{q({\uprho }_{\mathrm{G}} )}{} \mathbf{H}^{\mathrm{T}}{} \mathbf{P}_{\mathrm{s}} \mathbf{H}\hbox { and hence } \left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{C}_{\mathrm{s}}^{-1} \mathbf{H}} \right\} _{ii} \nonumber \\= & {} \frac{1}{q({\uprho }_{\mathrm{G}} )}\left\{ {\mathrm{H}^{\mathrm{T}}{} \mathbf{P}_{\mathrm{s}} \mathbf{H}} \right\} _{ii} \end{aligned}$$
(22)

where \(\mathbf{H}^{\mathrm{T}}{} \mathbf{P}_s \mathbf{H}\) and \(\left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{P}_{\mathrm{s}} \mathbf{H}} \right\} _{ii} \) are functions of \(\mathbf{A}_{\mathrm{s}} \) and \(\mathbf{P}_{\mathrm{s}}\).

$$\begin{aligned} \hbox {MDB}_{\mathrm{s,}i} =\sqrt{q({\uprho }_{\mathrm{G}} )}\frac{\sqrt{\lambda }}{\sqrt{\left\{ {\mathbf{H}^{\mathrm{T}}\mathbf{P}_{\mathrm{s}} \mathbf{H}} \right\} _{ii} }} \end{aligned}$$
(23)

We can easily find that the correlation coefficients \({\uprho }_{ij}\) as in (17) do not depend on \({\uprho }_{\mathrm{G}} \), but on \(\mathbf{A}_{\mathrm{s}} \) and \(\mathbf{P}_{\mathrm{s}} \) only, i.e., \({\uprho }_{ij} =\hbox {f}_{ij} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\).

Ad b) The vector \({\hat{\mathbf{x}}}\) does not depend on \({\uprho }_{\mathrm{G}} \), but on \(\mathbf{A}_{\mathrm{s}} \) and \(\mathbf{P}_{\mathrm{s}} \) only, i.e., \({\hat{\mathbf{x}}}=\hbox {f}_{\mathrm{x}} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\). This result is quite understandable since it is well known that the LS estimator can be obtained equivalently either by using the covariance matrix or the weight matrix.

As derived in Appendix A, the measure of external reliability \(\delta _i \) can be represented in the following form

$$\begin{aligned} \delta _i =\lambda \left[ {\frac{\left\{ {\mathbf{P}_{\mathrm{s}} } \right\} _{ii} }{\left\{ {\mathbf{H}^{\mathrm{T}}{} \mathbf{P}_{\mathrm{s}} \mathbf{H}} \right\} _{ii} }-1} \right] \end{aligned}$$
(24)

This means that it does not depend on the global correlation measure \({\uprho }_{\mathrm{G}} \), but on \(\mathbf{A}_{\mathrm{s}} \) and \(\mathbf{P}_{\mathrm{s}} \) only. We notice that there is a common factor \(\lambda \), which depends on type I error \(\upalpha \), type II error \(\upbeta \), and redundancy f of a network.

The additional quantity \(\mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}\) will have the representation as below

$$\begin{aligned} \mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}= & {} q({\uprho }_{\mathrm{G}} )\cdot \mathbf{HP}_{\mathrm{s}}^{-1} \mathbf{H}^{\mathrm{T}}\hbox { and hence }\left\{ {\mathbf{HC}_{\mathrm{s}} \mathbf{H}^{\mathrm{T}}} \right\} _{ii}\\= & {} q({\uprho }_{\mathrm{G}} )\cdot \left\{ {\mathbf{HP}_{\mathrm{s}}^{-1} \mathbf{H}^{\mathrm{T}}} \right\} _{ii} \end{aligned}$$

where \(\mathbf{HP}_{\mathrm{s}}^{-1} \mathbf{H}^{\mathrm{T}}\) and \(\left\{ {\mathbf{HP}_{\mathrm{s}}^{-1} \mathbf{H}^{\mathrm{T}}} \right\} _{ii} \) are functions of \(\mathbf{A}_{\mathrm{s}} \) and \(\mathbf{P}_{\mathrm{s}}\).

Before summing up the results of the above analysis, we denote by \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}} \) the quantity related to the ith observation \((i=1,{\ldots }, n)\) dependent on the level of global correlation \({\uprho }_{\mathrm{G}} \) and by \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{ind}} \) the quantity related to the ith observation (\(i~=~1,{\ldots }, n)\) independent of that level.

The results show that the representation of the correlation matrix as in (13) makes it possible to express the reliability measure \(\hbox {z}_{i}\) for the ith observation in the following form:

$$\begin{aligned} - \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}} ={\upmu }({\uprho }_{\mathrm{G}} )\cdot \hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \end{aligned}$$
(25)

where \({\upmu }({\uprho }_{\mathrm{G}} )\) can be either of \(\left( {q({\uprho }_{\mathrm{G}} )} \right) ^{-1}\), \(\sqrt{{\lambda }(f)\cdot }\sqrt{q({\uprho }_{\mathrm{G}} )}\)

$$\begin{aligned} - \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{ind}} =\hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \end{aligned}$$
(26)

For additional quantity \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 \), being \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}} \), we get in (25) \({\upmu }({\uprho }_{\mathrm{G}} )=q({\uprho }_{\mathrm{G}} )\).

Fig. 6
figure 6

Variability of the scale factors \(\upmu \) with respect to \({\uprho }_{\mathrm{G}} \) (\(\lambda \) computed for \(\alpha ~=~0.05\), \(\beta ~=~0.80\), \(f~=~0.5n\))

In the factorization of reliability measure as in (25), \({\upmu }({\uprho }_{\mathrm{G}} )\) is a scale factor common for all the n observations in a network, whereas \(\hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \) is a factor related only to the ith observation. The factors \(\hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \) for \(i~=~1,~2,\ldots ,n\) form together a set of mutually diversified values, as also is the case in relationship (26). Figure 6 shows graphs of the above-mentioned three types of the function \({\upmu }({\uprho }_{\mathrm{G}} )\), i.e., \({\upmu }_1 ({\uprho }_{\mathrm{G}} )=\left( {q({\uprho }_{\mathrm{G}} )} \right) ^{-1}\), \({\upmu }_2 ({\uprho }_{\mathrm{G}} )\) = \(\sqrt{{\lambda }(f)\cdot }\sqrt{q({\uprho }_{\mathrm{G}} )}\), \({\upmu }_3 ({\uprho }_{\mathrm{G}} )=q({\uprho }_{\mathrm{G}} )\) for \(n~=~2, 6\) and 10. Since beside the statistical parameters, \(\lambda \) is a function of redundancy f in a network, to get a common basis for different n, we assume for constructing the graphs that \(f~=~0.5n\), which corresponds to a level of internal reliability \(h~=~0.5\) for \(\mathbf{C}_{\mathrm{s}} =\mathbf{I}\).

Figure 6 shows that the scale factor \({\upmu }_1 ({\uprho }_{\mathrm{G}} )\) is increasing with the increase in \({\uprho }_{\mathrm{G}} \). The compound scale factor \({\upmu }_2 ({\uprho }_{\mathrm{G}} )\) displaying the resulting decrease with the increase in \({\uprho }_{\mathrm{G}} \) contains two mutually opposing effects, i.e., the increasing one (\(\sqrt{{\lambda } (f)})\) due to network redundancy and the decreasing one (\(\sqrt{q({\uprho }_{\mathrm{G}} )})\) due to the global correlation. The scale factor \({\upmu }_3 ({\uprho }_{\mathrm{G}} )\) is decreasing with the increase of \({\uprho }_{\mathrm{G}} \).

Additionally, by transforming (25) to a form

$$\begin{aligned} \hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) =\left\lfloor {\upmu ({\uprho }_{\mathrm{G}} )} \right\rfloor ^{-1}\cdot \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }\mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}}, \end{aligned}$$
(27)

we get an idea what part of the total value of the reliability measure for the ith observation represents a diversifying effect of observation correlations.

By simple modification of (27), we may also decompose the reliability measure for the ith observation into a part independent of \({\uprho }_{\mathrm{G}} \) and a part dependent on \({\uprho }_{\mathrm{G}} \), as shown below

$$\begin{aligned} \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}}= & {} \hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) +\left\{ {\hbox { }1-\left[ {\upmu ({\uprho } _{\mathrm{G}} )} \right] ^{-1}} \right\} \nonumber \\&\cdot \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{dep}} \end{aligned}$$
(28)

For completeness, we quote (26), which corresponds to \({\upmu }({\uprho }_{\mathrm{G}} )~=~1\) in (28)

$$\begin{aligned} \hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) _{\mathrm{ind}} =\hbox {f}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \end{aligned}$$

6 Numerical examples illustrating the use of derived formulas

In numerical examples, we maintain the approach applied in Prószyński and Kwaśniak (2016) of expressing each correlation-dependent quantity as a function of the form \(\hbox {z}(\mathbf{A},{{\varvec{\upsigma }}}=\mathbf{1}, \mathbf{C}_{\mathrm{s}} )\) where the column vector \({\varvec{\upsigma }}\) of ones represents unitary accuracies of observations. For simplicity, the form is denoted as \(\hbox {z}(\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\) where \(\mathbf{A}_{\mathrm{s}} =\mathbf{A}\).

A test network in two options, as shown in Fig. 7, is taken from the above-mentioned publication. It is a horizontal network treated as a free network. The brief characteristics of both the options, placed under each sketch, contain the following features:

  • the range of internal reliability indices for \(\mathbf{C}_{\mathrm{s}} =\mathbf{I}\) (in square brackets);

  • number of observations n, redundancy f.

Fig. 7
figure 7

Options of the test network. a [0.11, 0.48]; n = 15; f = 5. b [0.40, 0.82]; n = 28, f = 18

The arrows for GPS vectors in Fig. 7 indicate the orientation of coordinate differences assumed in the corresponding GMM models. The option (b) is the option (a) strengthened with 13 angular observations, what is reflected in a considerable increase in network internal reliability. The coordinates of the network point are given in Table 2.

Table 2 Point coordinates for horizontal network
Fig. 8
figure 8

Effect of reducing \(\hbox {MDB}_{\mathrm{s,}i} (\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\) to \(\hbox {MDB}_{\mathrm{s,}i} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\) for a network in Fig. 7a \({\upmu }_{2}\, ({\uprho }_{\mathrm{G}} ) = \sqrt{\lambda (f)\cdot }\sqrt{q({\uprho }_{\mathrm{G}} )}\)

For generating the correlation matrices, a slightly modified version of an algorithm as that in Davies and Higham (2000) was used, being much faster for greater n than the “accept-reject” algorithm described in Sect. 2.

The examples refer only to quantities that depend on a global correlation \({\uprho }_{\mathrm{G}} \), i.e., \(\hbox {MDB}_{\mathrm{s,}i} \), \(r_i \) and \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 \). First, we illustrate the effect of reducing \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) \) to \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \), where \(\hbox {z}_i =\hbox {MDB}_{\mathrm{s,}i} \), by using a scale factor \({\upmu }({\uprho }_{\mathrm{G}} )\) as in the formula (27).

In Figures 8 and 9 we can see that due to the reduction, for quantities being the reliability measures (i.e., \(\hbox {MDB}_{\mathrm{s,}i} \) and \(r_i\)) the graphs for different values of \({\uprho }_{\mathrm{G}} \) become considerably compacted, especially for the latter measure. The compacted graphs enable one to learn about the mutual diversification of the \(\hbox {z}_i \) values for different observations, i.e., for \(i~=~1,\ldots ,n\), independent of the global correlation \({\uprho }_{\mathrm{G}} \). In the case of an additional quantity \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 \) (Fig. 10), the values of \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\, (i=1,\ldots , n)\) for different values of \({\uprho }_{\mathrm{G}} \) are more extended along the vertical axis and much less mutually diversified for individual observations than in \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 (\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\).

For a more detailed analysis, we could determine the degree of mutual diversifications of the \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{P}_{\mathrm{s}} } \right) \) (\(i=1,\ldots ,n\)) values and compare them with the corresponding ones for \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) \) (\(i~=~1,\ldots ,n\)), for some chosen values of \({\uprho }_{\mathrm{G}} \). The degree of mutual diversifications might then be defined as a mean square deviation from the average value.

Fig. 9
figure 9

Effect of reducing \(r_{\mathrm{s,}i} (\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\) to \(r_{\mathrm{s,}i} (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\) for a network in Fig. 7b \({\upmu }_{1} ({\uprho }_{\mathrm{G}} )\) = \(\left( {q({\uprho }_{\mathrm{G}} )} \right) ^{-1}\)

Fig. 10
figure 10

Effect of reducing \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 (\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\) to \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 (\mathbf{A}_{\mathrm{s}} ,\mathbf{P}_{\mathrm{s}} )\) for a network in Fig. 7a \({\upmu }_{3} ({\uprho }_{\mathrm{G}} )~=~q({\uprho }_{\mathrm{G}} )\)

Now, the additive representation of \(\hbox {z}_i \left( {\mathbf{A}_{\mathrm{s}} ,\hbox { }{} \mathbf{C}_{\mathrm{s}} } \right) \) as in the formula (28) is illustrated in the examples for a network in Fig. 7a).

Fig. 11
figure 11

Additive representation of \(\hbox {MDB}_{\mathrm{s,}i} (\mathbf{A}_s ,\mathbf{C}_{\mathrm{s}} )\)

Fig. 12
figure 12

Additive representation of \(r_{\mathrm{s,}i} (\mathbf{A}_s ,\mathbf{C}_{\mathrm{s}} )\)

Fig. 13
figure 13

Additive representation of \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 (\mathbf{A}_{\mathrm{s}} ,\mathbf{C}_{\mathrm{s}} )\)

For \(\hbox {MDB}_{\mathrm{s,}i} \) and \(r_i \), due to \({\upmu }({\uprho }_{\mathrm{G}} )>1\) (see diagram in Fig. 6), the \({\uprho }_{\mathrm{G}} \)-dependent components assume positive values as shown in Figs. 11 and 12. In the case of \({\upsigma }_{\mathbf{v}_{\mathrm{s,}i} }^2 \), where \({\upmu }({\uprho }_{\mathrm{G}} )<1\) (see diagram in Fig. 6), the \({\uprho }_{\mathrm{G}} \)-dependent components are of negative values (see Fig. 13).

7 Proposed global correlation levels in networks

For a pair of random variables, there are established correlation levels. The problem arises in the case of global correlation discussed in this paper. We may express the correlation levels either directly in terms of the global correlation measure \({\uprho }_{\mathrm{G}} \) or in terms of the constant correlation coefficient a. Since the latter approach refers to a specific type of correlation matrices only, preference should be given to the former approach covering a full set of matrix configurations.

The following global correlation levels are proposed:

  1. 1.

    \(0<{\uprho }_{\mathrm{G}} \le 0.3\) weak correlation

  2. 2.

    \(0.3<{\uprho }_{\mathrm{G}} \le 0.6\) moderate correlation

  3. 3.

    \(0.6<{\uprho }_{\mathrm{G}} <1\) significant/strong correlation

Figure 14 shows the relationship between the above levels and the corresponding intervals of the constant correlation parameter a. As one might expect on basis of the curves in Fig. 2, with the increase of n, the intervals of a corresponding to weak and moderate global correlation are more and more lowered and narrower.

This testifies that the choice of the correlation levels was to a high extent imposed by the character of variability of \({\uprho }_{\mathrm{G}} \) versus a for different values of n. The case \(n=2\) is satisfyingly consistent with the commonly used levels of correlation between two random variables.

Fig. 14
figure 14

The proposed correlation levels as functions of a and n

The proposed global correlation levels provide a concise description of each given correlation matrix and make possible the comparison of different-size correlation matrices. It seems that they can be helpful in the design of systems with correlated observations. At the phase of design, we usually do not possess a specified correlation matrix, but may have some general knowledge on its elements as, for instance, the expected value of \(\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{max}} \) or \(\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{qm}} \) (see (5)). It is obvious that with this knowledge we may only find the approximate value of the global correlation index and the corresponding correlation level. We can do it by making the simplifying assumption that \(\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{max}} =a_1 \) or \(\left| {\left\{ {\mathbf{C}_{\mathrm{s}} } \right\} _{ij} } \right| _{\mathrm{qm}} =a_2 \), and using the formula (3) for \(a_{1} \) or \(a_{2} \).

We may use the matrices \(\mathbf{C}_{\mathrm{s}} (a_{1} )\) or \(\mathbf{C}_{\mathrm{s}} (a_{2})\) as substitute correlation matrices for determining some approximate reliability characteristics of the designed network.

8 Concluding remarks

Representation of the correlation matrix, based on the proposed measure of network global correlation, made it possible to investigate the structure of correlation-dependent quantities. Each analyzed quantity can be qualified either as dependent on a global correlation measure or as independent of such a measure. Those of the first type can be represented by a scale factor common for all the observations in a network and a factor being, beside a network structure, a function of the internal weight matrix yielding inter-observational diversifications. The quantities of the second type are solely the functions of the network structure and the internal weight matrix. Another way of representing the analyzed quantity of the first type is its decomposition into a part independent of the global correlation measure and a part dependent on that measure.

The above ways of representing the correlation matrices, together with the proposed levels of global correlation, can be applied in the analyses of systems with correlated observations.

For investigating the structure of correlation-dependent quantities at greater values of \({\uprho }_{\mathrm{G}} \) and n, we may use the easily created correlation matrices \(\mathbf{C}_{\mathrm{s}} (a)\) with a constant term a computed for the required values of \({\uprho }_{\mathrm{G}} \) and n. Obviously, the thus-obtained matrices will only be specific substitutes of the correlation matrices sought for.

In the paper the proposed analytical tools were applied to basic reliability measures for a single outlier case. It is easy to deduce that due to generality of the theoretical basis of the tools their potential application field can be much wider. Analyzing the findings of some publications that contributed to development of the theory of network reliability and outlier detection methods, we may state that the tools can also be applied in investigating the effect of observation correlations on the quantities such as for instance:

  • the outlier test statistics and MSBs (minimally separable biases) for a single outlier case as in Knight et al. (2010),

  • the outlier test statistic and MDB (minimal detectable bias) for multiple outliers as in Wang and Knight (2012).