1 Introduction

The usability of Digital Twins is highly dependent on the quality of the sensor data provided [1, 2]. Malfunctions and failures of sensors can not only disturb the monitoring of the physical twin through the Digital Twin, but also lead to serious malfunctions of the physical twin, due to being controlled by the feedback of the Digital Twin. This problem can be mitigated by the integration of redundant sensors, however, the installation of several physical sensors leads to additional financial and constructive expenditure. For this reason, this publication examines the question of how redundancy can be built up by introducing soft sensors, which calculate the quantity of interest instead of measuring them. This is to be used to detect sensor failures with the resulting extended system boundary of the Digital Twin of a rolling bearing test bench. In the long term the results can be applied to different types of Digital Twins, such as test benches, production machines or customer products.

2 State of the art

2.1 The Digital Twin concept

Due to the high degree of novelty of the concept of Digital Twins, a multitude of partly contradictory understandings and definitions exist. In order to create a uniform understanding the following definition of a Digital Twin applies in the context of this publication:

A digital twin is a digital representation of a real product instance (physical twin). The representation uses models that are fed with real-time data, e.g. by sensors installed on the physical twin, to simulate its behavior. The simulation results are then fed back into the physical world via a bidirectional connection and made use of there [3,4,5,6].

In order to distinguish the digital twin from its environment, it is defined by a system boundary [7, 8]. The system boundary of the digital twin is based on the system boundary of the physical twin [9] and is drawn according to the intended use case [10]. The system boundary is primarily used to describe whether the respective digital twin is a single unit or an aggregated system of subsystems [8, 9, 11]. A more detailed consideration from a system theoretical view to identify an optimum of the system boundary between twin and environment does not take place. It is therefore not defined where the system boundary should be drawn and whether the sensors should be included as an information interface between the physical and Digital Twin. However, the strong dependence of Digital Twins on correct sensor data [1, 2] leads to the conclusion that, in general, the behavior of sensors is not described by Digital Twins. For this reason, malfunctions and failures of sensors can disturb the monitoring of the physical twin by the Digital Twin. The calculations and simulations of the Digital Twin are based on sensor data and can be used to regulate or control the physical twin. With incorrect input values this can lead to serious malfunctions of the latter.

2.2 Soft sensors

In contrast to physical sensors, soft sensors describe an algorithm which calculates the quantity of interest in a product instead of measuring it with a physical entity [12]. The algorithms can be based on physical, empirical or data-driven models. The latter include machine learning approaches [13]. These algorithms use the measurement data of various physical sensors integrated in the product as input variables to calculate the quantity of interest. The thus calculated variable of the soft sensor neither needs a local proximity or the same measured variable as the installed physical sensors. However, it must be ensured that a correlation is given and the soft sensor Ssoftsensor can be represented in the form of a function of the physical sensors Si,physical  shown in Eq. 1 [14].

$$S_{\text{softsensor}}=f\left(S_{1{,}\text{physical}\,}{,}\ldots {,}S_{n{,}\text{physical}\,}\right)$$
(1)

In this way, data can be simulated for locations where placing physical sensors is difficult due to design restrictions or operating conditions. In the literature, soft sensors are also called virtual sensors [14,15,16,17], extended Kalman filters [18] or, in the context of control engineering, state observers, estimators or predictors [19]. Some authors describe the soft sensor as a Digital Twin of a physical sensor [13, 14, 19]. Others use Digital Twins as comprehensive soft sensors of complex systems [20, 21]. In order to do justice to the previously mentioned definition of the term Digital Twin, the terms Digital Twin and soft sensor are not used synonymously in a general way. However, assuming that a physical counterpart exists in the product, a soft sensor will be regarded as a Digital Twin of the corresponding physical sensor in the scope of this publication. Soft sensors can, just like physical sensors, be used as input variables for a Digital Twin [22,23,24].

2.3 Redundancy through soft sensors

One possible approach to mitigate the effects of sensor failures is the redundant implementation of physical sensors. Redundancy can be differentiated into hot and cold redundancy. In the case of hot redundancy, two or more systems work in parallel at any given time, but can also fulfil the task on their own. In cold redundancy only one system is operated at a time. The replacement system is kept on standby and activated only in case of failure of the first system [25]. Since a redundant sensor system records the same measured variables and compares the measured values, it is classified as hot redundancy. Discrepancies between the measured values allow sensor failures or misbehavior to be identified and appropriate measures to be initiated. A redundant design of physical sensors is associated with a number of disadvantages. One drawback is increased cost, due to the usage of more sensors. Furthermore, the integration of sensors is accompanied by an increased design effort in order to meet requirements for functional fulfilment or construction space.

One alternative strategy is a data driven approach to validate sensor data and detect sensor failures like the implementation of redundant soft sensors. Approaches in recent literature utilize the idea that soft sensors can not only be used to determine unknown measured variables, but also to perform a calculation of variables already measured from existing physical sensors. There are multiple ways to calculate the required soft sensor values. The Input can be either the last few values of the respective physical sensor [17] or the values all sensors except the respective physical sensor [16]. Another alternative is to identify reliable and unreliable sets of sensors in advance. Soft sensors are then created for the unreliable ones from the data of the reliable ones [17].

Using these methods, a redundant sensor system is built without the need to integrate additional physical sensors. This is referred to in the literature as analytical [2, 17] or virtual redundancy [13, 14, 16, 19]. In the context of redundancy alignment, the measured values of the physical sensors and the predicted values of the soft sensors can be compared and deviations can be determined. The deviations are utilized to verify the measured data from the physical sensors [2]. This way the condition of the physical sensor can be monitored and failures can be detected [16, 19, 26]. Soft sensors can also be used as backups to replace the physical sensors in case of a deviation [13, 17]. In this way, process stability can be improved [14].

Redundant soft sensors cannot be used on their own, but must be processed and interpreted by a suitable data dissemination system. One possible approach is processing by a Digital Twin, for which some initial applications can be found in the literature.

Staudter et al. [2] consider soft sensors in a holistic investigation of data-induced conflicts in Digital Twins. In this context, soft sensors are used to build analytical redundancy and are used within a holistic framework to identify conflicts. However, an isolated consideration of the practical generation and use is not the focus of this publication.

He et al. [26] use a multi-block PLS approach to model the system with the aim of monitoring the process through a digital twin. In addition to the (soft) sensors, the actuators and possible process faults are modelled in order to identify the cause of a process fault when it is detected. However, the identification of the process error itself does not take place through the soft sensors.

Darvishi et al. [17, 27] in contrast, use soft sensors specifically to identify sensor faults. For this purpose, soft sensors are used to calculate sensor values from unreliable physical sensors and the occurring deviations are compared. In a first approach, the soft sensors consist of a predictor that takes into account the temporal development and an estimator that uses the remaining reliable sensors [27]. In a follow-up publication, an algorithm is used which takes both perspectives into account at the same time [17]. In each case, multilayer perceptron (MLP) neural networks are used to generate the soft sensors, using 70% [17] and 85% [27] of the available data to train each. The need for large amounts of data is explicitly mentioned as a limiting factor of neural networks. Darvishi et al. further state that the inclusion of correlations between the sensors can increase the performance compared to the use of all sensors simultaneously [17].

In this contribution, the concept of soft sensors is considered in isolated fashion in order to avoid implications that arise from interactions with related systems. For this purpose, the effects of the system boundary of Digital Twins are first examined and the soft sensors are delimited.

In the practical side, redundant soft sensors are used to identify sensor errors. Neural networks are deliberately avoided, as they are not always suitable due to the lack of transparency, as well as the high demand for data. Instead, various alternative machine learning methods are used. To increase the performance of the soft sensors, the correlations of the individual physical sensors are included. The entire procedure for creating the soft sensors is replicable and is presented in a methodical procedure model.

3 Extension of the system boundary of a Digital Twin

As described in Sect. 2.1, a Digital Twin is a virtual representation of a physical counterpart (physical twin). Real-time data from the physical twin is captured by sensors and transferred to the Digital Twin. The behavior of the sensors is often not represented by Digital Twins. Instead, sensor data is taken as given. This corresponds to a system boundary that is drawn around the physical twin, but excludes the sensors. This can be seen in Fig. 1a.

Fig. 1
figure 1

Schematic presentation of the Digital Twin concept with the sensor outside (a) or inside (b) the system boundary

The ability to simulate and anticipate the behavior of physical sensors through a redundant soft sensor system extends this system boundary to include the physical sensors as shown in Fig. 1b. In this way, the Digital Twin can verify sensor data and detect sensor failures and malfunctions at an early stage. This helps prevent erroneous inputs to the Digital Twin and potentially harmful behavior.

The information flow from the physical twin to the Digital Twin is of particular interest for the consideration and expansion of the system boundary through soft sensors. The soft sensor system is located in this information flow. Figure 2 schematically shows the information flow in detail.

Fig. 2
figure 2

Schematic presentation of the information flow from the physical twin to the Digital Twin including the processing by soft sensors

The measured values collected by the physical sensors A, B and C are fed into the soft sensor system. In this system, the soft sensor C* is initially created from the measured values of the physical sensors A and B and later supplied by them. The values of the physical sensor C and the soft sensor C* are compared before the information is fed into the Digital Twin. For reasons of clarity, only the creation and comparison of sensor C and soft sensor C* is shown in the illustration. In the same way, this must also be done for sensors A and B.

In Fig. 2, the soft sensor system, which is the focus of this publication, is highlighted in red. It is not relevant for the scope of this publication how the information on the physical twin is collected in the form of physical sensors or how the information obtained is utilized in the Digital Twin.

To create the soft sensors, different machine learning models are investigated. Neural networks are deliberately excluded due to their lack of transparency. The algorithms for the creation and usage of the soft sensors were realized in Python. To clarify the procedure a flow chart according to DIN 66001 [28] is created, which contains both the creation and use of the soft sensors. The flow chart is shown in Fig. 3, the two algorithms are described in the following sections.

Fig. 3
figure 3

Flowchart of the algorithm for the creation (a) and usage (b) the soft sensors

3.1 Creation of soft sensors

New soft sensors are created for each individual use case, this is shown in in Fig. 3a. For this, data is collected at the beginning of the runtime, to create a soft sensor Si, softsensor for each physical sensor Si, physical. The soft sensors are based on correlations between the measured values of the individual physical sensors. In order to utilized these correlations in the best possible way, different machine learning models are investigated. In the context of this work, the models linear regression, polynominal regression, random forrest regressor and decision trees were used. For each sensor i, the k different model types are created.

All physical sensors are taken as input for the model, with the exception of the particular physical sensor i for which the respective redundancy is to be generated. The latter data are used as labels to train the model. Equation 2 shows the creation of the models using the example of linear regression with weights αi,j and intercept variables βi.

$$\left[\begin{array}{c} S_{1{,}\text{softsensor}}\\ \vdots \\ \vdots \\ S_{n{,}\text{softsensor}} \end{array}\right]=\begin{bmatrix} 0 & \alpha _{1{,}2} & \ldots & \alpha _{1{,}n}\\ \alpha _{2{,}1} & 0 & \, & \vdots \\ \vdots & \, & \ddots & \alpha _{n-1{,}n}\\ \alpha _{n{,}1} & \ldots & \alpha _{n{,}n-1} & 0 \end{bmatrix}\left[\begin{array}{c} S_{1{,}\text{physical}}\\ \vdots \\ \vdots \\ S_{n{,}\text{physical}} \end{array}\right]+\left[\begin{array}{c} \beta _{1}\\ \vdots \\ \vdots \\ \beta _{n} \end{array}\right]$$
(2)

Then, using data collected up to this point, the model scores of the different models are calculated. For this purpose, the respective score functions of the different model types of the Scikit-learn environment are used [29,30,31,32]. The best possible model for each individual sensor is selected and used as the core of the corresponding soft sensor i. Depending on the correlations between the measured values, the individual soft sensors can only predict the measured values with a certain accuracy. For this reason, an individual failure criterion must be determined for each soft sensor. To do this, each soft sensor predicts a value for the corresponding physical sensor. These values are compared and the normalized deviation over the value range of the data is calculated. This is shown in Eq. 3.

$$\frac{\left| S_{i{,}\text{physical}}-S_{i{,}\text{softsensor}}\right| }{S_{i{,}\text{physical}{,}\max }-S_{i{,}\text{phyiscal}{,}\min }}=\Updelta S_{n}$$
(3)

The failure criterion is then determined using the average deviation over the set of measured values. After a soft sensor is created for each physical sensor, the algorithm ends. The resulting models for each sensor, as well as the failure criteria in the form of the limit, exported for the later usage.

3.2 Usage of soft sensors

The created soft sensors are applied during the rest of the runtime of the physical twin operation. For this purpose the saved soft sensors in the form of models of each physical sensor as well as the individual failure criteria are loaded in for the utilization. This is shown in Fig. 3b.

For each time step, the data of the physical sensors to be examined is imported. The model of each soft sensor is fed with the sensor data of the other sensors. Depending on the respective trained weights, these are used to predict the value of the corresponding soft sensor.

The predicted values of the soft sensor are then compared with the values of the physical sensors and the deviations are calculated. If these deviations exceed the previously defined limits, a warning is issued and appropriate measures are initiated. If the deviation of all measured variables is within their individual limits, the data of the entire time step is released for use and passed on to the actual Digital Twin of the superordinate system or product.

4 Application on a test bench

4.1 Introduction of the test bench

At the Department of Product Development and Machine Elements at TU Darmstadt, rolling bearings can be examined on the rolling bearing test bench “Athene” (see Fig. 4; [33,34,35]). This test bench can apply loads to rolling bearings at up to 8000 rpm in four test chambers. For this purpose, two hydraulic cylinders apply up to 40 kN axially and radially. The reactions of the bearing in the form of radial and axial vibration and impact levels are measured via sensors. Furthermore, sensors for speed, torque, axial and radial force are installed and the voltage and current is recorded. A listing of all sensors can be seen in Fig. 5 or Table 1. Since the test bench is already equipped with extensive sensor systems, the corresponding sensor data can be used to feed a Digital Twin. In contrast to consumer products, no integration of further sensor technology is necessary.

Fig. 4
figure 4

Rolling bearing test bench “Athene” of the TU Darmstadt

Fig. 5
figure 5

Correlation matrix of the sensors installed on the test bench

Table 1 Model score and deviations of selected sensors on the test bench

Since the necessary IT infrastructure for real-time processing of the data is currently still under development, stored data sets are used. The data sets are not loaded as a whole but row wise in discrete time steps of one second, so a real time data input is simulated and the transferability of the results is increased. This applies to both the Digital Twin and the soft sensors implemented later.

In the following, a soft sensor system is created and applied for the sensors of this test bench. As illustrated in Fig. 2, the digital twin of the test bench itself is not the focus of this work and will therefore not be discussed further below.

4.2 Creation of the soft sensors

Data from a long-term test was used to create/train the soft sensors. As described in Sect. 3.2, the first hour of the experiment was used to create the models of the soft sensors. The inconsistent run-in phase was deliberately excluded. Figure 5 shows a correlation matrix of the physical sensor data during the training period. The brighter a single cell of the matrix is, the higher the correlation of the two corresponding physical sensors. It is emphasized that the correlations are differently pronounced.

The data of the training period is first split using a train test split (0.75; 0.25). The training data is used to create four different machine learning models (linear regression (lr), polynomial regression (pr), random forrest regressor (rfr) and decision trees (dt)) for each of the physical sensors and to determine the model scores. The model with the highest model score is then automatically selected and used as the basis for the corresponding soft sensors.

The test data is then used to determine the deviations between data of the physical sensors and the predicted values of the soft sensors. For each sensor 10,000 random data points are used and the average deviation is determined. The deviations are then normalized to the range of values of the data.

Table 1 shows the most suitable ML models for each sensor, their model score and the average deviation of the predictions that can be obtained with them. The normalization to the value range of the data leads to sensors with quasi-static measured values with a small value range showing high average deviations. The temperatures fluctuate with about ± 1 ℃ around a constant value range, so that small deviations of 0.5 ℃ already lead to the high deviations shown here. The model scores and average deviations of the soft sensors for axial force and torque are also striking. This may be due to defective sensors during data collection, but requires closer investigation.

For the purposes of this example, the failure criteria for each sensor are set as ten times the average deviation.

4.3 Usage of the soft sensors

For each physical sensor, the corresponding previously created model is loaded. This is fed with the corresponding physical sensor data and thus the values of the respective soft sensors are calculated based on the weights of the model.

The calculated values of the soft sensor are then compared with the real measured values of the physic sensor and the relative deviation is calculated. This is then compared with the limit values derived from the mean deviation. If this limit value is exceeded for a certain time, an error message is displayed.

Figure 6 shows an exemplary dashboard with three soft sensors in comparison with the physical. For the axial vibration the deviation is determined.

Fig. 6
figure 6

Dashboard of calculated values of the soft sensors (normal operation)

4.4 Detection of sensor failures

To evaluate the function of the soft sensors developed, a physical sensor is artificially damaged. For this purpose, an offset is applied to the physical sensor of the axial vibration from second 60 of the observed time range. This leads to the violation of the failure criterion of the maximum permissible deviation in form of the yellow line. The sensor of the axial vibration is indicated as defective. This is shown in Fig. 7.

Fig. 7
figure 7

Dashboard of calculated values of the soft sensors (sensor failure)

5 Discussion

It was shown that it is possible to build a redundant sensor system using soft sensors. The sensor behavior is modeled, which extends the system boundary of the Digital Twin onto the sensors and makes it possible to detect sensor errors. The theoretical concepts were applied exemplarily at a rolling bearing test bench of the TU Darmstadt.

At the current time, the IT infrastructure required for real-time processing is still under development, which is why the exemplary implementation was only possible with existing data sets. Care was taken to process these data sets row wise in one-second increments, which increases the transferability of the results, but this still needs to be finally evaluated. Furthermore, the models are trained for individual use cases. In the future, the use of measurement data from four-quadrant experiments can be used, to create soft sensors, which can be used with various use cases at different operating points. Furthermore, the choice of failure criteria should be reconsidered. Although the derivation of the selection criteria from the mean deviations during the training leads to satisfying results, the interdependencies of the different soft sensors cannot be covered sufficiently. Even if the evaluation an the rolling bearing test bench shows that the soft sensors fulfill their purpose, the determination of the failure criteria from the average deviations must be critically questioned and extensively verified.