1 Introduction

Coronavirus Disease 2019 (COVID-19) is currently the world’s most severe epidemic, posing a global public health problem and challenge. Diagnostic COVID-19 early and accurately is essential to controlling the spread of the epidemic. The most popular diagnostic method in biochemical is Real-Time Polymerase Chain Reaction (RT-PCR), which is currently the most popular diagnostic method against COVID-19 [1]. Although RT-PCR has a lower cost, it is time-consuming and may produce false positives [2]. In radiology, analyzing Chest CT and X-ray images of COVID-19 is also a key technology. Studies have shown that Chest CT scans are more sensitive than RT-PCR in the early detection of COVID-19 infection [3, 4]. When RT-PCR is negative, Chest CT image analysis is critical for diagnosing COVID-19. However, it is impossible to use CT on a large scale as CT is an expensive and unsafe test. The American College of Radiology (ACR) warns that the CT disinfection necessary following screening COVID-19 individuals potentially cause radiological service availability to be disrupted, and recommends that portable chest radiography be explored to minimize the likelihood of cross-infection [5]. Also, a positive X-ray test may not require CT if COVID-19 is suspected at a clinical high [6]. Therefore, Chest X-ray image testing is a feasible solution for the early diagnosis and control of COVID-19.

Artificial intelligence (AI), which is concerned with imitating human reasoning in computation, has made tremendous advances in a variety of fields [7, 8]. AI-assisted healthcare systems have lately attracted interest, with the aim of producing diagnostic technologies and clinical decision-making more instantaneous, self-sufficient, and efficient [9,10,11]. As we all know, image segmentation is crucial in computer image processing and analysis, and Multilevel threshold Image Segmentation (MIS) is a simple and efficient image segmentation technique [12]. Medical images are a particular class of images, their feature space is usually more complex, and more attention is paid to the accuracy and efficiency of image segmentation[13]. In recent years, many research results have been published on applying the Meta-Heuristic Algorithm (MHA) to optimize MIS on medical images [14,15,16]. The advantages of MHA in medical image segmentation are becoming more and more prominent. Since the outbreak of COVID-19, many scholars have conducted much research on MIS based on the MHA [13, 17,18,19,20,21,22,23]. These results indicate that the advantages of using MHA to optimize the segmentation of COVID-19 Chest X-ray Images (COVID-19-CXIs) are apparent. However, to achieve better segmentation results and improve the diagnostic quality of COVID-19, more research is still needed to propose better MHAs and segmentation schemes. As a result, this paper focuses on studying the MIS based on the MHA and proposes a high-performance scheme to segment COVID-19-CXIs.

Image segmentation is the process of dividing an image into different areas based on specific rules. These areas do not intersect, and each region has a universal consistency. Formally, it can be defined as follows: if \(g\left(x,y\right)\) is the set of all pixels, where \(\left(x,y\right)\) denotes the spatial coordinate, then segmentation is a partitioning of the set \(g\) into a set of connected subsets \(\left({g}_{1},{g}_{2}\dots {g}_{n}\right)\). These subsets satisfy the following conditions [24]:

  • \({g}_{i}\) is a connected region, \(1\le i\le n\).

  • \(\bigcup_{i=1}^{n}{g}_{i}=g\).

  • \({g}_{i}\cap {g}_{j}=\mathrm{\varnothing },1\le x\le n,1\le y\le n,i\ne j\).

  • \({g}_{i}\) satisfies certain rules.

Image segmentation can be achieved using a variety of techniques. Pal [12] provided an overview of image segmentation methods, which include: gray-level thresholding [25,26,27,28], iterative pixel classification [29,30,31], surface-based segmentation [32, 33], segmentation of color images [34], edge detection [35], methods based on fuzzy set theory [36]. In addition, Pham et al. [37] divided these methods into eight categories: thresholding approaches, region growing approaches, classifiers, clustering approaches, Markov Random Field (MRF) models, artificial neural networks, deformable models, and atlas-guided approaches. Regardless of classification, thresholding is an old, simple, and popular image segmentation technique [12]. Thresholding is classified as Bilevel threshold Image Segmentation (BIS) or MIS based on the number of thresholds. BIS is the process of dividing the image into two regions: the object and the background. When multiple thresholds are chosen, the image can be divided into several objects and backgrounds, a process known as MIS. MIS is used in the majority of real-world scenes.

There are many ways of MIS, such as histogram shape-based [38, 39], clustering-based [25, 40], similarity-based [41, 42], entropy-based [28], etc. Among these, the MIS based on entropy is widely used due to its ease of implementation and accurate results. For example, Kapur’s entropy [28], fuzzy entropy [43], Tsallis entropy [44], etc. Kapur proposed Kapur’s entropy image segmentation in 1985. The essence of Kapur’s entropy image segmentation is to determine a set of segmentation thresholds to maximize Kapur’s entropy. Determining a set of segmentation thresholds is the key to Kapur’s entropy image segmentation. The traditional exhaustive method is time-consuming and ineffective [17, 18]. Determining the threshold is equivalent to solving a constrained optimization problem, precisely what the MHA is designed to solve.

The MHA is an iterative algorithm that uses the advantageous different data distribution of particle swarms. This new search method solves complex problems faster than traditional optimization analysis algorithms and is more adaptable. Therefore, many researchers were highly concerned about the MHA soon after it was first proposed. Various excellent MHAs have been continuously proposed, including Multi-Verse Optimizer (MVO) [45], Whale Optimization Algorithm (WOA)[46], Grey Wolf Optimization (GWO)[46], Harris Hawk Optimization (HHO) [47], Bat Algorithm (BA)[46], Particle Swarm Optimization (PSO) [48], Hunger Games Search (HGS) [49], Colony Predation Algorithm (CPA) [50], Slime Mould Algorithm (SMA) [51], Harris Hawks Optimization (HHO) [47], Weighted Mean of Vectors (INFO) [52], Runge Kutta Optimizer (RUN) [53], etc. They are widely used to solve optimization problems in many fields, such as image segmentation [54, 55], optimization of machine learning model [56], economic emission dispatch problem [57], medical diagnosis [58, 59], scheduling problems [60,61,62], plant disease recognition [63], practical engineering problems [64, 65], solar cell parameter Identification [66], feature selection [67, 68], bankruptcy prediction [69, 70], expensive optimization problems [71, 72], combination optimization problems [73], and multi-objective problem [74, 75]. There have also been numerous achievements in image segmentation.

Li et al. [76] improved the barnacle mating optimizer using a logistic model and chaotic mapping, and a multilevel color image segmentation algorithm was applied combined with Masi entropy. Li et al. [77] improved the pollination algorithm used for Tsallis entropy image segmentation. Li et al. [78] proposed a Modified Artificial Bee Colony (MABC) optimizer and solved the MIS problem. In 2021, Li et al. [79] used an improved coyote optimization algorithm to achieve fuzzy MIS. Houssein et al. [80] improved manta ray foraging optimization for MIS. Khairuzzaman [81] applied the GWO to Otsu MIS in 2017. Further, many studies based on Kapur’s entropy have also been published. Akay [82] applied PSO and Artificial Bee Colony (ABC) to Kapur’s entropy MIS, respectively, and compared the performance, Bhandari et al. [83] applied the Cuckoo Search (CS) algorithm and Wind Driven Optimization (WDO) to Kapur’s entropy MIS respectively in 2014, in the second year, they improved the ABC and combined using Kapur, Otsu, and Tsallis functions respectively for satellite image segmentation [84].

In 2016, Mirjalili et al. [45] proposed a new MHA called MVO. MVO, like other MHAs, suffers from slow convergence and is prone to falling into local optimum. However, it has a miracle that the structure is simple and the parameters are few. Since MVO was proposed in 2016, it has drawn the concentration of many scholars. In recent years, research results on MVO have been published continuously. In 2016, Faris et al. [85] proposed to use MVO to train the Multi-Layer Perceptron (MLP) neural network. PSO, Differential Evolution (DE), Firefly (FF), and CS were experimentally compared and then compared with two traditional gradient-based training methods (BP and LM algorithms). The experimental results showed that the MVO algorithm to retrain the MLP neural network is competitive. In 2017, Faris et al. [86] applied MVO to the machine learning algorithm in the paper. They proposed using MVO to optimize the Support Vector Machine (SVM) parameters and select the optimal features. The experimental results showed that MVO is used. The MVO algorithm can efficiently feature the number and improve the accuracy of predictions. Ewees et al. [87] proposed the Chaotic MVO (CMVO) algorithm in the paper. The CMVO algorithm combines the chaotic map based on the MVO algorithm, which effectively improves the algorithm’s performance and was successfully used to solve the problem of feature selection. Mirjalili et al. [88] proposed a Multi-Objective MVO (MOMVO) for solving problems in a multi-objective search space. Fathy et al. [89] applied MVO to determine the optimal parameter selection for Proton Exchange Membrane Fuel Cells (PEMFCs) under specific operating conditions. Yilmaz et al. [90] proposed a hybrid algorithm, IMVOSA, mixing MVO, and the simulated annealing algorithm. In the paper [91], MVO was improved, called the enhanced version of MVO EMVO. EMVO was used as a task scheduler in the cloud computing environment. Experiments showed that it could effectively improve resource utilization and minimize manufacturing time. Pothiraj et al. [92] used MVO to optimize 3D IC floor planning.

It is clear that it can improve MVO and apply it to real-world problems. However, as far as the author knows, there has not been much MVO research in the direction of Kapur’s entropy MIS. Considering all this, this paper improves the shortcomings of MVO and applies it to COVID-19-CXI segmentation.

This paper proposes the RDMVO algorithm to address the shortcomings of MVO. RDMVO is based on MVO and incorporates the Rosenbrock Method (RM) and the Diffusion Mechanism (DM). To test the performance of RDMVO. This paper first selected IEEE CEC2017 [93] as benchmark functions and used the Wilcoxon signed-rank test [94] and the Friedman test [95] to compare some mainstream MHAs. The experimental results show that RDMVO effectively improves the global search and local search capabilities of MVO and the convergence speed and accuracy of most test functions. The comprehensive performance is better than other MHAs involved in the experiment. Then, at different threshold levels, RDMVO was applied to Kapur’s entropy MIS experiment and compared to some MHAs. Using the Peak Signal-To-Noise Ratio (PSNR) [96], Structural Similarity Index (SSIM) [97], and Feature Similarity Index (FSIM) [98], three indicators for image segmentation effect evaluation, the results show that RDMVO’s performance is satisfactory. At the same time, Berkeley images (BKIs) and the COVID-19-CXIs were selected for the test images. The BKIs are widely used in image processing. It is used to test the comprehensive performance of RDMVO in MIS. COVID-19-CXIs are some of the Chest X-rays of patients with confirmed COVID-19. It is also the problem to be solved in this paper. The MIS experimental effects are all satisfactory. In addition, to clarify the generality of the algorithm, we use the Friedman test, and the test results also show that RDMVO is statistically significant. The main contributions of this paper are as follows:

  • An improved MVO algorithm is proposed, called RDMVO, which significantly enhances the convergence speed, accuracy, and ability to jump out of the local optimum of MVO.

  • The performance comparison experiment of RDMVO and some mainstream MHAs were conducted on IEEE CEC2017. The experimental data reveal that RDMVO’s performance is better than the other MHA's.

  • A novel MIS scheme (RDMVO-MIS) is proposed. Kapur’s entropy is used as the objective function, RDMVO determines the threshold, and the image segmentation experiment was successfully carried out on BKIs and COVID-19-CXIs.

  • MIS comparison experiments with some MHAs were carried out. The effect and generality of Kapur’s entropy MIS based on RDMVO are the best.

The remaining work is scheduled as follows in this paper: in Sect. 2, the work is related to image segmentation and improved Kapur’s entropy image segmentation method. In Sect. 3, the original MVO algorithm is presented. In Sect. 4, this paper mainly describes the proposed RDMVO algorithm. In Sect. 5, comparative experiments on benchmark functions and image segmentation experimentations are run to verify RDMVO’s and RDMVO-MIS’s performance. Section 6 gives a discussion. Finally, in Sect. 7, conclusions and the direction of future work are summarized.

2 Multilevel Threshold Image Segmentation

Thresholding is the most commonly used image segmentation method. In essence, thresholding is to select an attribute to divide the gray value of an image into two or more sets. Thresholding is commonly performed based on the histogram generated by the image’s gray level. The image is not disturbed by noise in an ideal situation, the histogram of the segmented image has two or more peaks, the threshold is set at the trough, and the image can be divided into multiple objects and backgrounds as needed. In a real picture, however, the image will be disturbed by various noises, and the gray value of the image is not the correct data. The histogram is disturbed by noise, there is no peak on the histogram, or there may be multiple peaks. Image segmentation results with thresholds on the troughs will be poor or incorrect. In this case, many threshold determination methods have been proposed [99,100,101]. To evaluate the threshold selection results, Pun [102] proposed the concept of entropy in 1980 and determined the optimal threshold by calculating the threshold corresponding to the maximum entropy. Kapur then proposed an improved threshold segmentation method based on maximum entropy based on Pun in 1985, which is simple to calculate and has a good segmentation effect.

The traditional calculation of Kapur’s entropy is based on the image’s one-dimensional gray value, which has poor anti-noise interference ability. Buades et al. [103] proposed non-local means 2D histogram in 2005. This technology can effectively reduce noise interference and has a good segmentation effect when the image is polluted by noise. Kapur’s entropy MIS used in this paper is based on non-local means 2D histogram. The specific process is to obtain the grayscale image corresponding to the original image and then perform non-local mean noise reduction processing on the grayscale image. The obtained image is called the NLM image. Then non-local means 2D histogram is obtained according to the grayscale and NLM images. Perform the maximum Kapur’s entropy calculation based on the 2D histogram, then obtain a set of thresholds corresponding to the maximum Kapur’s entropy, and finally perform image segmentation according to the thresholds.

Figure 1 shows an image segmentation example in this paper. The specific related content is introduced in the following. In addition, three mainstream image segmentation evaluation methods used in this paper will also be introduced one by one.

Fig. 1
figure 1

Image segmentation example in this paper

2.1 Kapur’s Entropy

Kapur’s entropy MIS is based on the gray value of the image. We define the gray value of the image to be stored in 8 bits, and the gray value ranges from 0 to 255. Let \(L\)=256, \({n}_{i}\) represents the number of pixels whose gray value is \(i\), then the Kapur’s entropy \(H\) is defined according to the following formula:

$$\begin{array}{c}N=\sum\limits_{i=0}^{L-1}{n}_{i}\end{array}$$
(1)
$$\begin{array}{c}{p}_{i}=\frac{{n}_{i}}{N}\end{array}$$
(2)
$$\begin{array}{c}H=-\sum\limits_{i=0}^{L-1}{p}_{i}ln{p}_{i}\end{array}$$
(3)

where \({p}_{i}\) is the probability of occurrence of gray value \(i\), \(H\) is Kapur’s entropy. For BIS, images are divided into two subclasses, \({C}_{0}\) and \({C}_{1}\). \({C}_{0}=\{\mathrm{1,2},3\dots {t}_{1}-1\}\), \({C}_{1}=\{{t}_{1},\dots L-1\}\), the Kapur’s entropy \({H}_{C}\) is described according to the following formula:

$$\begin{array}{c}{H}_{C}={H}_{{C}_{0}}+{H}_{{c}_{1}}\end{array}$$
(4)
$$\begin{array}{c}{H}_{{C}_{0}}=-\sum\limits_{j=0}^{{t}_{1}-1}\frac{{p}_{j}}{{\omega }_{0}}ln\frac{{p}_{j}}{{\omega }_{0}}\end{array}$$
(5)
$$\begin{array}{c}{\omega }_{0}=\sum\limits_{n=0}^{{t}_{1}-1}{p}_{n}\end{array}$$
(6)
$$\begin{array}{c}{H}_{{C}_{1}}=-\sum\limits_{j={t}_{1}}^{L-1}\frac{{p}_{j}}{{\omega }_{1}}ln\frac{{p}_{j}}{{\omega }_{1}}\end{array}$$
(7)
$$\begin{array}{c}{\omega }_{1}=\sum_{n={t}_{1}}^{L-1}{p}_{n}\end{array}$$
(8)
$$\begin{array}{c}{t}^{\star }=argMax\left({H}_{c}\right)\end{array}$$
(9)

where \({t}^{\star }\) is the split point set when \({H}_{c}\) takes the maximum value, it is also the determined threshold. Similar to BIS, for MIS, suppose the image gray value is divided into \(m\) subsets, \({C}_{0}=\{\mathrm{1,2},3\dots {t}_{1}-1\}\), \({C}_{1}=\{{t}_{1},\dots {t}_{2}-1\}\), \({C}_{2}=\{{t}_{2},\dots {t}_{3}-1\}\),…, \({C}_{m-1}=\{{t}_{m-1},\dots L-1\}\), the Kapur’s entropy is described according to the following formula:

$$\begin{array}{c}{H}_{C}=\sum\limits_{i=0}^{m-1}{H}_{{C}_{i}}\end{array}$$
(10)
$$\begin{array}{c}{H}_{{C}_{i}}=-\sum\limits_{j={t}_{i}}^{{t}_{i+1}-1}\frac{{p}_{j}}{{\omega }_{i}}ln\frac{{p}_{j}}{{\omega }_{i}}\end{array}$$
(11)
$$\begin{array}{c}{\omega }_{i}=\sum\limits_{n={t}_{i}}^{{t}_{i+1}-1}{p}_{j}\end{array}$$
(12)
$$\begin{array}{c}{{\varvec{t}}}^{\star }=argMax\left({H}_{c}\right)\end{array}$$
(13)

2.2 Non-local Means 2D Histogram

Buades et al. [103] proposed non-local means 2D histogram in 2005; this innovative technology uses redundant information for denoising and maintaining the maximum detailed features of the image. Assuming that the original image is \(O\), the non-local mean denoised image is \(N\), and the gray value of the pixels \(p\) in the image \(O\) is denoted as \(O(p)\), then after non-local mean filtering, the gray value \(N(p)\) of pixels \(p\) can be obtained by the following formula:

$$\begin{array}{c}N\left(p\right)=\frac{\sum_{q\epsilon O}O\left(q\right)\omega \left(p,q\right)}{\sum_{q\epsilon O}\omega \left(p,q\right)}\end{array}$$
(14)
$$\begin{array}{c}\omega \left(q,p\right)={exp}^{-\frac{{\left|\mu \left(p\right)-\mu \left(q\right)\right|}^{2}}{{\partial }^{2}}}\end{array}$$
(15)
$$\begin{array}{c}\mu \left(p\right)=\frac{\sum_{i\epsilon O\left(p\right)}O\left(i\right)}{m \times m}\end{array}$$
(16)
$$\begin{array}{c}\mu \left(q\right)=\frac{\sum_{i\epsilon O\left(q\right)}O\left(i\right)}{m \times m}\end{array}$$
(17)

where \(\omega \left(q,p\right)\) is the weight of pixel \(p\) and\(q\), and \(L(p)\) and \(L(q)\) are local images and centered on pixel \(p\) and \(q\), respectively, and the size is\(m \times m\). \(\mu \left(p\right)\) and \(\mu \left(q\right)\) are the local mean of pixel \(p\) and \(q\), respectively, and the mean of \(L(p)\) and\(L(q)\). \(\partial\) is the standard deviation.

So far, there are corresponding grayscale image \(O\) and image \(N\) filtered by non-local mean. Combining them, the abscissa is the gray value corresponding to the pixel of the original image \(O\), and the ordinate is the gray value corresponding to the image \(N\) filtered by the non-local mean. Then we can get a 2D view of the 2D histogram, shown in Fig. 2. The non-local means 2D histogram according to the following formula:

Fig. 2
figure 2

2D view of the 2D histogram

$$\begin{array}{c}{P}_{ij}=\frac{{h}_{ij}}{m \times n}\end{array}$$
(18)

where \(i\) refers to the value of image \(O(x,y)\) pixels, \(j\) refers to the value of image \(N(x,y)\) pixels, and \(h(i,j)\) denotes the number of times the point \((i,j)\) appears on the gray value vector \((s,t)\), the total number of pixels in this image is \(m \times n\).

2.3 Kapur’s Entropy-based 2D Histogram

A 2D view histogram based on the above 2D histogram is given in Fig. 2. The main diagonal of the 2D histogram contains adequate image information. This paper calculates Kapur’s entropy as the objective function and Kapur’s entropy of the subregions on the main diagonal using Eq. (19). The optimal solution found by in {\(t,{t}_{2}\dots {t}_{n}-1\)}is the optimal threshold.

$$H\left(s,t\right)=-\sum_{i=0}^{{s}_{1}}\sum_{j=0}^{{t}_{1}}\frac{{P}_{ij}}{{P}_{1}}ln\frac{{P}_{ij}}{{P}_{1}}-\sum_{i={s}_{1}+1}^{{s}_{2}}\sum_{j={t}_{1}+1}^{{t}_{2}}\frac{{P}_{ij}}{{P}_{2}}ln\frac{{P}_{ij}}{{P}_{2}}\dots -\sum_{i={s}_{L-2}+1}^{{s}_{L-1}}\sum_{j={t}_{L-2}+1}^{{t}_{L-1}}\frac{{P}_{ij}}{{P}_{L-1}}ln\frac{{P}_{ij}}{{P}_{L-1}}$$
(19)

And \({P}_{1}=\sum_{i=0}^{{s}_{1}}\sum_{j=0}^{{t}_{1}}{P}_{ij}\), \({P}_{2}=\sum_{i={s}_{1}+1}^{{s}_{2}}\sum_{j={t}_{1}+1}^{{t}_{2}}{P}_{ij}\), \({P}_{L-1}=\sum_{i={s}_{L-2}+1}^{{s}_{L-1}}\sum_{j={t}_{L-2}+1}^{{t}_{L-1}}{P}_{ij}\).

2.4 Image Segmentation Evaluation

We all know how important it is to conduct practical evaluations of image segmentation experiments. The evaluation of mathematical computations is a crucial juncture that requires benchmarking, sufficient data, and appropriate metrics for a credible evaluation [104,105,106]. Many methods for evaluating segmentation results have been proposed, each with advantages and disadvantages. The three most commonly used evaluation methods are PSNR [96], SSIM [97], and FSIM [98]. As a result, they will be used to analyze the experimental results in this paper.

  • PSNR. PSNR is a full-reference image quality evaluation index, and it is the most commonly used image quality assessment metric. PSNR relies on MSE to measure the degree of distortion of the image. The calculation of PSNR is expressed as follows [96]:

    $$\begin{array}{c}PSNR\left(f,G\right)=10\times lo{g}_{10}\frac{L}{MSE\left(f,G\right)}\end{array}$$
    (20)
    $$\begin{array}{c}MSE\left(f,G\right)=\frac{{\sum }_{i=0}^{M-1}{\sum }_{j=0}^{N-1}{\left[f\left(i,j\right)-G\left(i,j\right)\right]}^{2}}{M\times N}\end{array}$$
    (21)

    \(M\) and \(N\) are the numbers of rows and columns of the image, respectively, \(f\) represents the original image, and \(G\) represents the segmented image. \(f\left(i,j\right)\) represents the pixel gray-scale value of the original image, and \(G\left(i,j\right)\) is the pixel gray-scale value of the segmented image. \(L\) is the scale range of the image. For an 8-bit image, \(L={2}^{8}-1=255\).

  • SSIM. SSIM is a full-reference image quality evaluation index. Furthermore, it is a widely used image quality evaluation metric based on the assumption that the human eye extracts structured information from an image when viewing it. It calculates the brightness, contrast, and structure comparison functions of the image, respectively, then makes a comprehensive evaluation. Its calculation process is as follows [97]:

    $$\begin{array}{c}SSIM\left(x,y\right)=\frac{\left(2{\mu }_{x}{\mu }_{y}+{C}_{1}\right)\left(2{\sigma }_{xy}+{c}_{2}\right)}{\left({\mu }_{x}^{2}+{\mu }_{y}^{2}+{C}_{1}\right)\left({\sigma }_{x}^{2}+{\sigma }_{y}^{2}+{C}_{2}\right)}\end{array}$$
    (22)

    where \(x\) represents the image block of the original image, \(y\) represents the image block of the segmented image. \({\mu }_{x}\) and \({\mu }_{y}\) are the means of the image block \(x\) and the image block \(y\), respectively, and it reflect the brightness information of the image. \({\sigma }_{x}\) and \({\sigma }_{y}\) are the standard deviations of the image block \(x\) and the image block \(y\), respectively, and it reflects the contrast information of the image. \({\sigma }_{xy}\) is the correlation coefficient between the image block \(x\) and the image block \(y\). And it reflects the structural similarity information of the image.

  • FSIM. FSIM is a relatively new full reference image quality evaluation index. It is based on two major features, Phase Consistency (PC) and Gradient Magnitude (GM). The calculation of FSIM is expressed as follows [98]:

    $$\begin{array}{c}FSIM=\frac{\sum_{I\in\Omega }{S}_{L}\left(X\right)P{C}_{m}\left(X\right)}{\sum_{I\in\Omega }P{C}_{m}\left(X\right)}\end{array}$$
    (23)
    $$\begin{array}{c}{S}_{L}\left(X\right)={S}_{PC}{\left(X\right)}^{\alpha }{S}_{G}{\left(X\right)}^{\beta }\end{array}$$
    (24)
    $$\begin{array}{c}P{C}_{m}\left(X\right)=\frac{E\left(X\right)}{(\epsilon +\sum_{m}{A}_{n}(X))}\end{array}$$
    (25)
    $$\begin{array}{c}{S}_{PC}\left(X\right)=\frac{2P{C}_{1}\left(X\right)P{C}_{2}\left(X\right)+{T}_{1}}{P{C}_{1}^{2}\left(X\right)P{C}_{2}^{2}\left(X\right)+{T}_{1}}\end{array}$$
    (26)
    $$\begin{array}{c}{S}_{G}\left(X\right)=\frac{2{G}_{1}\left(X\right){G}_{2}\left(X\right)+{T}_{2}}{{G}_{1}^{2}\left(X\right){G}_{2}^{2}\left(X\right)+{T}_{2}}\end{array}$$
    (27)
    $$\begin{array}{c}G=\sqrt{{G}_{X}^{2}+{G}_{Y}^{2}}\end{array}$$
    (28)

FSIM is coupled by PC term and GM term, \({S}_{L}\) refers to the similarity score, \(\alpha\) and \(\beta\) take 1 by convention. \(P{C}_{m}\left(X\right)\) is \(Max\left(P{C}_{1}\left(X\right),P{C}_{2}\left(X\right)\right)\), \(\epsilon\) is a very small positive number, preventing a denominator of 0. \({T}_{1}\) and \({T}_{2}\) are a constant, \(E\left(X\right)\) indicates the local energy, \({A}_{n}\left(X\right)\) is the amplitude value.

3 Overview of Original MVO

MVO’s mathematical model and algorithm are based on the multi-verse theory of cosmology’s three concepts of white holes, black holes, and wormholes. According to cosmological theory, white and black holes are two amazing celestial bodies, and their properties are diametrically opposed. White holes only eject matter and energy to the outside and are considered the main part of the universe. Black holes only absorb matter and energy in the universe. Wormholes are tunnels that connect parallel universes, bridging the gap between white and black holes, and allowing objects to move instantly between universes and time-spaces.

In the MVO algorithm, the concepts of white holes and black holes are modeled to represent the exploration process in the search space, and the wormhole model simulates the exploitation process. MVO proposes the concept of the expansion rate and believes that the universe is constantly changing, and objects in the multi-verse are constantly evolving to the most stable universe according to the expansion rate, white holes, black holes, and wormholes. The MVO algorithm specifies the following rules [45]:

  • The greater the expansion rate, the greater the possibility of a white hole, and the lower the possibility of a black hole.

  • A universe with a higher expansion rate sent matter through a white hole.

  • Universes with lower expansion rates receive more matter through black holes.

  • Regardless of the expansion rate, all objects in the universe can move randomly through wormholes toward the optimal universe.

Therefore, MVO starts by generating random universes. In each iteration, objects use white/black holes to move from universes with high expansion rates to other universes with low expansion rates. Furthermore, objects in any universe are randomly teleported to the best universe through a wormhole. Repeating these processes until the final criteria are met.

Based on the above theory, MVO first randomly generates a multi-verse:

$$\begin{array}{c}X=\left[\begin{array}{ccc}{{\varvec{X}}}_{1}^{1}& {{\varvec{X}}}_{1}^{2}& \begin{array}{cc}\cdots & {{\varvec{X}}}_{1}^{d}\end{array}\\ {{\varvec{X}}}_{2}^{1}& {{\varvec{X}}}_{2}^{2}& \begin{array}{cc}\cdots & {{\varvec{X}}}_{2}^{d}\end{array}\\ \begin{array}{c}\vdots \\ {{\varvec{X}}}_{n}^{1}\end{array}& \begin{array}{c}\vdots \\ {{\varvec{X}}}_{n}^{2}\end{array}& \begin{array}{c}\begin{array}{cc}\vdots & \vdots \end{array}\\ \begin{array}{cc}\cdots & {{\varvec{X}}}_{n}^{d}\end{array}\end{array}\end{array}\right]\end{array}$$
(29)

where \(n\) represents the number of universes, each universe represents a candidate solution, \(d\) represents the amount of matter in each universe, and the matter in the universe represents the parameters in the solution. The update of the universe is based on the following formula:

$$\begin{array}{c}{{\varvec{X}}}_{i}^{j}=\left\{\begin{array}{c}\begin{array}{cc}{{\varvec{X}}}_{k}^{j},& {r}_{1}<NI\left({{\varvec{X}}}_{i}\right)\end{array}\\ \begin{array}{cc}{{\varvec{X}}}_{i}^{j},& {r}_{1}\ge NI\left({{\varvec{X}}}_{i}\right)\end{array}\end{array}\right.\end{array}$$
(30)

Among them, \({{\varvec{X}}}_{i}\) is the \(ith\) universe, \(NI\left({{\varvec{X}}}_{i}\right)\) is the normalized expansion rate of the \(ith\) universe, \({{\varvec{X}}}_{i}^{j}\) refers to the \(jth\) matter in the \(ith\) universe, and \({{\varvec{X}}}_{k}^{j}\) is the \(jth\) substance of the \(kth\) universe, \(k\) is generated by the roulette selection mechanism, and \({r}_{1}\) is a random number of \(\left[\mathrm{0,1}\right]\). As can be seen from these equations, the selection and determination of white holes are made by roulette, which is based on the normalized expansion rate. The lower the expansion rate, the more likely it is that an object will pass through the tunnel of a white hole. Beyond that, there could be wormholes in the universe that randomly alters objects in the universe, regardless of their expansion rate. The mechanism is expressed as follows:

$${\varvec{X}}_i^j = \left\{ {\begin{array}{*{20}{l}} {\left\{ {\begin{array}{*{20}{l}} {{\varvec{X}}_{best}^j + TDR \times \left( {\left( {ub - lb} \right) \times {r_4} + lb} \right),}&{{r_3} < 0.5} \\ {{\varvec{X}}_{best}^j - TDR \times \left( {\left( {ub - lb} \right) \times {r_4} + lb} \right),}&{{r_3} \geq 0.5} \end{array}} \right.}&{{r_2} < WEP} \\ {{\varvec{X}}_i^j,}&{{r_2} \geq WEP} \end{array}} \right.$$
(31)

Among them, \({{\varvec{X}}}_{i}^{j}\) refers to the \(jth\) substance in the \(ith\) universe, \({{\varvec{X}}}_{k}^{j}\) is the \(jth\) substance in the \(kth\) universe, \({{\varvec{X}}}_{best}^{j}\) is the \(jth\) substance in the current best universe, \(ub, lb\) is the upper and lower bounds of the variable, \({r}_{2}, {r}_{3}, {r}_{4}\) are random numbers of \(\left[0, 1\right]\), \(TDR\) is the travel distance rate, which defines that an object can be teleported by a wormhole to the currently obtained best universe around the distance rate, which increases in iterations. Moreover, \(WEP\) is the wormhole existence rate, which defines the probability that wormholes exist in the universe and increases linearly in iterations. The adaptive formulas of \(\mathrm{TDR}\) and \(\mathrm{WEP}\) are as follows:

$$\begin{array}{c}WEP=min+l\times \left(\frac{\mathrm{max}-\mathrm{min}}{L}\right)\end{array}$$
(32)
$$\begin{array}{c}TDR=1-\left(\frac{{l}^{1/p}}{{L}^{1/p}}\right)\end{array}$$
(33)

where \(\mathrm{min}\) is the minimum value, \(\mathrm{max}\) is the maximum value, \(l\) represents the current iteration number, and \(L\) represents the maximum iteration number. Where \(p\) defines the development precision in iterations. The higher \(p\) is, the faster and more accurate the local search is. Algorithm 1 shows the pseudo-code of MVO.

figure a

4 Proposed RDMVO

4.1 Rosenbrock Method (RM)

RM is a local search method proposed by Rosenbrock [107] in 1960. It can adapt to the search direction and size and conduct multiple searches until a relatively optimal solution is found in a specific limited area. Regarding the determination of the search size, RM will first determine \(a\) random number \(\varepsilon\) as the search size and then determine whether \(a\) better solution is obtained under the search size through calculation and comparison. If \(a\) better solution is obtained, the next search size is \(\varepsilon \times \alpha\) and\(\alpha >1\), otherwise the next search size is \(\varepsilon \times \beta\) and \(0<\beta <1\). Regarding the determination of the search direction, RM refers to the orthonormal basis of the n-dimensional space rather than experimenting in a single direction.

The basic RM is effective for unimodal functions, and it is not easy to jump out of the local optimum for multimodal functions. The basic RM was modified in 2011 by Kang et al. [108] to solve this problem. Li et al. [109] applied RM modified to the HHO in 2021 and proved the mechanism's performance. The modified RM is used to enhance MVO’s performance in this paper. Algorithm 2 shows the pseudo-code of modified RM.

figure b

Eq refers to \({{\varvec{\delta}}}_{i}=\sqrt[2]{\frac{{\sum }_{k=1}^{n}\left({{\varvec{x}}}_{ki}-avg{{\varvec{x}}}_{i}\right)}{n}}+{\varepsilon }_{1},i=\mathrm{1,2}\dots d\). and \(avg{{\varvec{x}}}_{i}=\frac{{\sum }_{k=1}^{n}{{\varvec{x}}}_{ki}}{n}\). Where \(n\) is the number of prospect particles, \(avg{{\varvec{x}}}_{i}\) refers to the average of the prospect particles in the \(ith\) dimension, \({{\varvec{x}}}_{ki}\) refers to the value of the \(kth\) prospect particles in the \(ith\) dimension. \({\varepsilon }_{1}=1.0{e}^{-150}\) is a very small variable for preventing the initial value from being 0.

4.2 Diffusion Mechanism (DM)

The DM is mentioned in Literature [110, 111], which effectively alleviates the optimum local problem and dramatically improves the possibility of finding the global optimum. The diffusion process refers to the random diffusion of new particles around the original agent at different locations. There are two main methods for generating new particles: Levy flight and Gaussian distribution [111]. They are applied based on Eqs. (4.1) and (4.2), respectively.

$$\begin{array}{c}{{\varvec{x}}}_{i}^{q}={{\varvec{x}}}_{{\varvec{i}}}+{\boldsymbol{\alpha }}_{i}^{q}\otimes Levy\left(\lambda \right)\end{array}$$
(34)
$$\begin{array}{c}{{\varvec{x}}}_{i}^{q}={{\varvec{x}}}_{i}+\beta \times Gaussian\left({{\varvec{P}}}_{i},\left|{\varvec{B}}{\varvec{P}}\right|\right)-\left(\gamma \times {\varvec{B}}{\varvec{P}}-{\gamma }_{1}\times {{\varvec{P}}}_{i}\right)\end{array}$$
(35)

In Eq. (34), \({x}_{i}\) refers to the original agent, and \(q\) refers to the number of new particles obtained after diffusion. \(\alpha\) is a variable used to control the convergence speed. The symbol \(\otimes\) refers to the Hadamard multiplications. In Eq. (35), likewise, \({{\varvec{x}}}_{i}\) refers to the original agent, and \(q\) refers to the number of new particles obtained after diffusion. \(\beta\) is equal to \((log(t))/t\), and \(t\) is the number of iterations. \(\gamma\) and \({\gamma }_{1}\) are random numbers between \([\mathrm{0,1}]\), and \({\varvec{B}}{\varvec{P}}\) is the best position of new particles so far. \(Gaussian\left({{\varvec{P}}}_{i},\left|{\varvec{B}}{\varvec{P}}\right|\right)\) is the Gaussian distribution, and the mean and standard deviation are \({{\varvec{P}}}_{i}\) and \(\left|{\varvec{B}}{\varvec{P}}\right|\), respectively.

In this paper, reference paper [112], We use the mathematical model of the DM as follows:

$$\begin{array}{c}{{\varvec{X}}}_{i}=Gaussian\left({\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}},{\varvec{\varepsilon}},{\varvec{m}}\right)+\sigma \times \left({\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}}-{{\varvec{X}}}_{i}\right)\end{array}$$
(36)

where \(Gaussian\left({\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}},{\varvec{\varepsilon}},m\right)\) means to generate a random matrix that obeys Gaussian distribution. \({\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}}\) and \({\varvec{\varepsilon}}\) represent the mean and standard deviation, respectively, and \({\varvec{m}}\) is a vector that determines the form of the generated matrix. \({\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}}\) denotes the position of the global optimal solution. \(\sigma\) is a random number between \([\mathrm{0,1}]\). The calculation of \({\varvec{\varepsilon}}\) is expressed as follows:

$$\begin{array}{c}\varepsilon =\frac{log\left(t\right)}{t}\times \left|{{\varvec{X}}}_{i}-{\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}{\varvec{p}}{\varvec{o}}{\varvec{s}}{\varvec{i}}{\varvec{t}}{\varvec{i}}{\varvec{o}}{\varvec{n}}\right|\end{array}$$
(37)

The RDMVO algorithm introduces the RM and the DM into the standard MVO algorithm, enhancing the global search ability and local exploration ability of the MVO algorithm and effectively improving the convergence speed and accuracy. Algorithm 3 shows the pseudo-code of RDMVO. The flowchart of RDMVO is shown in Fig. 3.

Fig. 3
figure 3

Flowchart of RDMVO

figure c

4.3 Algorithm Complexity Analysis

Figure 3 shows the main process of RDMVO. The time complexity of RDMVO is mainly determined by the initialization, the RM, the DM, and the iterative updates of MVO. In addition, the parameters also affect the algorithm complexity. These parameters are mainly: the algorithm’s maximum number of iterations (\(T\)), the dimension (\(d\)) and the population size (\(N\)). Therefore, the main time complexity of RDMVO is O (RDMVO) = O (initialization) + O (fitness evaluation) + O (MVO iterative updates) + O (RM) + O (DM) ≈ O (\(n\times d\)) + O (\(T\times n\)) + O (\(n\times d\)) + O (\(n\times d\)) + O (\(n\)).

5 Experimental Results and Analysis

A series of experiments were pushed to verify RDMVO’s performance. To begin, we choose the IEEE CEC2017 as benchmark functions. On IEEE CEC2017, we selected some contemporary mainstream MHAs to conduct an algorithm comparison experiment. The experimental data show that RDMVO has apparent convergence speed and accuracy advantages. Also, we conducted parametric experiments on IEEE CEC2017 to choose a better \(p\). Furthermore, we applied RDMVO in conjunction with Kapur’s entropy to segment BKIs and COVID-19-CXIs, compared the performance with some mainstream MHAs, and selected PSNR, SSIM, and FSIM as evaluation indicators. The experimental results are also positive. The following section will briefly introduce these experimental procedures and results. In addition, this paper uses MATLAB coding. All experiments were performed on the same computer. Computer parameters and coding software versions are given in Table 14.

5.1 Experiment on IEEE CEC2017 Benchmark Functions

The IEEE CEC2017 benchmark functions and related parameter settings are first introduced in this section. The parameter set includes two parts, the first part is the setting of public parameters, and the second part is the parameters of the algorithm participating in the experiment. Common parameters mean that to ensure the experiment’s fairness, as per other computational science works [113], and the differences in the algorithm itself, other parameters must be unified during the experiment. The settings of public parameters are shown in Table 1. Where \(N\) is the number of particles in the population, \(D\) is the dimension of the problem, \(MaxFEs\) is the Maximum number of evaluations, \(F\) is the number of independent runs of the experiment, and we set F to 30 to reduce the influence of randomness on the experimental results. In addition, the relevant parameter settings of the MHA involved in this experiment are given in Table 16.

Table 1 Unified parameter settings

Then, we analyzed the experimental results from multiple perspectives. The first is the mean and standard deviation. To reduce the experimental error, we ran the experiment 30 times independently and recorded the minimum value of the test function obtained by each relevant algorithm. Thus, we would get 30 on each test function result for each algorithm. We computed the mean and standard deviation of these values and recorded Avg and Std, respectively. Furthermore, use Avg and Std to evaluate the algorithm’s performance. We expect them to be as small as possible, and the minimum value is shown in bold in the relevant tables. In addition, to verify whether the experimental results are statistically significant, we applied nonparametric statistical tests with a significance level of 0.05, Wilcoxon signed-rank test, and Friedman test to the experimental results.

The symbol “±/=” in the related table is the result summary of the Wilcoxon signed-rank test, and ARV represents the mean value of the Friedman test results of the overall test function of the algorithm.

5.1.1 Benchmark Functions

IEEE CEC2017 [93] is the relatively new test function set used to test the performance of the MHA, which can better measure the superiority of algorithm improvement strategies. The test function of IEEE CEC2017 is shown in Table 15. F1 to F3 are unimodal functions, F4 to F10 are simple multimodal functions, F11 to F20 are hybrid functions, and F21 to F30 are composition functions.

5.1.2 Influence of Two Mechanisms

In Sect. 4, we propose using DM and RM to enhance the performance of the MVO algorithm. Experiments must be carried out to determine whether these two mechanisms are effective. We added a single mechanism to MVO, named RDMVO1, RDMVO2, in addition to the original algorithm MVO and the new algorithm RDMVO, for a total of four algorithms. Table 2 contains a detailed description of each algorithm. The letter Y means that the mechanism is included, while the letter N means no such mechanism exists. For the convenience of description, RDMVO1, RDMVO2, and MVO are called variants of RDMVO. We used IEEE CEC2017 to experiment with these four algorithms, and the experimental results are shown in Tables 17, 18, and Fig. 4.

Table 2 Results of the combination of mechanisms
Fig. 4
figure 4

Convergence curves of 9 benchmark functions

The ARV indicator after the Friedman test in Table 17 (The best results are bolded) shows that the RDMVO has the best overall ranking across all test functions; RDMVO outperformed the other variants. The “±/=” indicators also show that RDMVO outperforms its variants on most test functions, especially the original MVO, which outperforms its original MVO in 27 out of 30 test functions. In general, the experimental results demonstrate that the combination of RM and DM improves the performance of MVO the most.

Table 18 shows the P values of the experimental results via the Wilcoxon signed-rank test with 5% significance, and P values ⩾ 0.05 are shown in boldface. As can be seen from the P values in Table 18, RDMVO and RDMVO2 are statistically different in each test function, RDMVO and MVO are statistically different in each test function except for F9 and F17.

Figure 4 is the convergence curve of RDMVO and its variants on some test functions. As shown in Fig. 4, the convergence speed and accuracy of the variant RDMVO2 are worse than those of the original algorithm MVO. RDMVO2 only adds DM to the MVO, which means that only adding DM has not affected the improvement of MVO or had a negative effect. When comparing the variants RDMVO1 and MVO, the convergence accuracy of RDMVO1 is significantly improved, as is the convergence speed. That is, RM speeds up the overall global search speed of MVO and effectively enhances the ability of MVO to jump out of the local optimum. Finally, we compared RDMVO and RDMVO1, and the convergence rate of RDMVO1 was further enhanced under the action of DM. From the above analysis, it can be concluded that the combined action of the RM and DM have the most powerful effect on MVO: effectively enhancing MVO’s global search ability and local exploration ability, improving the problem that MVO is easy to fall into local optimum, and improving the search accuracy of MVO.

5.1.3 Parameter Test

In Sect. 3, the MVO algorithm is introduced in detail. The \(TDR\) in Eq. (31) directly affects the algorithm’s development efficiency and accuracy. \(p\) in Eq. (33) is the only parameter that affects \(TDR\). In RDMVO, it is necessary to determine an appropriate \(p\). So, this section selected multiple \(p\) to conduct experiments on IEEE CEC2017. For the convenience of description, the algorithm name in this section is RDMVO- \(p\){\(x\)}, representing that the \(p\) in RDMVO is set to \(x\). Table 3 shows the Friedman test rankings when \(p\) is 4, 5, 6, and 7, respectively. From Table 3, it can be seen that the performance of RDMVO is relatively best when \(p\) is 6 if we set \(p\) to any of 4, 5, 6, and 7. Accordingly, we have set the \(p\) at 6 in the subsequent experiments in this paper.

Table 3 Friedman test ranking on the benchmark functions

5.1.4 Comparison with Other Algorithms

In this part, RDMVO was mainly compared with DE [46], WOA, GWO, HHO [47], BA [46], PSO [46], and MVO [45], seven mainstream MHAs on IEEE CEC 2017. All the algorithms involved were compared under the same conditions, and the experimental results are shown in Tables 4, 19, 20, and Fig. 9.

Table 4 Friedman test ranking of all algorithms on the benchmark functions

Table 19 displays the performance of RDMVO and 7 other MHAs on benchmark functions, and the best results are bolded. The “±/=” indicators in Table 19 show that RDMVO outperforms other competitors in most test functions. RDMVO outperforms the efficient algorithm DE on 20 test functions and the traditional algorithm PSO on up to 28. RDMVO outperforms the original algorithm MVO in 26 test functions, with two being equal to MVO and two being weaker. According to the ARV metric after the Friedman test, RDMVO has the highest overall ranking among all tested functions, followed by the DE algorithm.

For further detailed analysis, the Friedman test ranking of all algorithms on the benchmark functions is shown in Table 4. It is clear to find that RDMVO scored first place in F5, F7, F8, F10, F12, F16, F17, F18, F19, F20, F21, and F22, and second place in F1, F4, F6, F9, F11, F13, F14, F15, F23, F24, F26, F27, F28, F29, and F30. This shows that RDMVO has excellent performance on multimodal functions and performs well on some Hybrid Functions and Composition Functions.

Table 20 shows the P values of the experimental results via the Wilcoxon signed-rank test with 5% significance, and P values ≥ 0.05 are shown in boldface. As can be seen from the P values in Table 20, RDMVO and HHO, BA are statistically different on each test function, and RDMVO and MVO are statistically different on each test function except F9. In most cases, RDMVO was statistically significantly different from its competitors.

To visually better see the difference, the convergence curve on some test functions (F5, F7, F8, F10, F16, F18, F19, F20, F22) is given in Fig. 9. The convergence speed and accuracy of RDMVO and other competitors are compared. RDMVO has a faster convergence trend than other algorithms, achieving good convergence results early in the algorithm iteration.

5.2 Experiment on Image Segmentation

In this section, we applied RDMVO to image segmentation. First, RDMVO is combined with Kapur’s entropy. Specifically, Kapur’s entropy is used as a fitness function, and the position of the universe is used as a set of thresholds at a certain threshold level. Slightly adjust RDMVO logic to find the maximum value of Kapur’s entropy. This is simple to accomplish. The image segmentation and evaluation are performed under the resulting set of thresholds. The specific experimental process is shown in Fig. 5.

Fig. 5
figure 5

Flowchart of Kapur’s entropy image segmentation based RDMVO

In other AI-related publications, there are proposals for fair experimental comparisons of two or more computational approaches on a certain dataset, which need allocating the same computer resources to each methodology [115,116,117,118]. Same as Sect. 5.1, to ensure the fairness of the experiment and compare the differences in the algorithms themselves, we set other experimental parameters uniformly. As shown in Table 5, where N is the number of particles in the population, T is the number of iterations, and we set the number of experiments F to 20, so each algorithm was run independently 20 times to reduce the influence of randomness on the experimental results. Table 16 shows the relevant parameter settings of the MHA used in this experiment.

Table 5 Unified parameter settings

We ran three sets of experiments. The first set was a mechanism comparison experiment on BKIs to verify the effectiveness of RM and DM; the second set was an algorithm comparison experiment on BKIs. We chose some MHAs that have been applied to image segmentation. Moreover, compare the image segmentation effects of RDMVO and these algorithms to determine whether the RDMVO algorithm is more advantageous; the third group conducted algorithm comparison experiments on COVID-19-CXIs, and the difference from the second group is that the test image set has been changed. In addition, considering the complexity of the actual situation, all our experiments set multiple threshold levels to verify the adaptability of the algorithm. Finally, we used the Friedman test to perform a statistical test on the results. Furthermore, use PSNR, SSIM, and FSIM as evaluation indicators. All experiments are set out in detail below.

5.2.1 Influence of the Two Mechanisms

As in Sect. 5.1.2, a mechanism comparison experiment on image segmentation was first performed. We selected nine complex representative images from the BKIs as the research objects, as shown in Fig. 10. Figure 11 depicts the grayscale, histogram, and 2D histograms of these images. The first column in Fig. 11 is the grayscale image, the second column is the histogram, and the third column is the 2D histogram. The experimental results are shown in Tables 6, 21, and Fig. 6.

Table 6 Friedman test results for each algorithm on PSNR, FISM and SSIM
Fig. 6
figure 6

Maximum fitness of RDMVO and variants at different thresholds

Fitness is the maximum of Kapur’s entropy obtained by segmentation, representing the amount of information in the segmented image. A suitable segmentation method will have a more significant amount of information. Simultaneously, with the improvement of the threshold level, the amount of information obtained by a suitable segmentation method will also increase. The fitness values of RDMVO and variant algorithms at different segmentation levels are given in Table 21. The bold shows the maximum fitness value obtained for each experimental image at each threshold level. In addition, Fig. 6 shows the maximum fitness of all images of RDMVO and variants under different thresholds. In Table 21, as the threshold level increases from 2 to 4, the fitness obtained by RDMVO and variant algorithms increases gradually. It can be observed more directly in Fig. 6. In Fig. 6, the abscissa is the different test images, the ordinate is the maximum fitness, and curves of different colors represent different threshold levels. It can be seen that the curve with a larger threshold level is always above the curve smaller than its threshold level. The above shows that RDMVO and variant algorithms are suitable for image segmentation on Kapur’s entropy.

Table 6 shows three evaluation data of PSNR, FSIM, and SSIM of the segmentation results of RDMVO and variant algorithms and a more extended Friedman test. Among them, Avg is the average value of the relevant evaluations across all images. The larger the average value, the better the comprehensive result of image segmentation. Rank is the corresponding ranking. The best Avg and Rank values are bolded. RDMVO achieves top results for all three evaluations on experimental images at all threshold levels except the FSIM metric with a threshold level of 2.

In summary, RDMVO is more competitive in image segmentation experiments. That is, the introduced RM and DM have meaning.

5.2.2 Comparison with Other Algorithms on Berkeley

In this section, we compared some mainstream algorithms with the original algorithm MVO. These mainstream algorithms included ACOR, BA, WOA, SSA, DE, AVOA, CLPSO and HHO. Some MHAs have been successfully applied to image segmentation or some relatively new algorithms [18, 81, 119,120,121,122,123,124,125]. Nine test images are shown in Figs. 10 and 11. In addition, we set the threshold levels to 2, 3, 4, and 5 for experiments.

Table 22, 23, and 24 show three evaluation values of segmentation quality at different threshold levels, including the mean and standard deviation of PSNR, SSIM, and FSIM evaluations, respectively. The data are derived from the results of 20 independent runs of the experiment. Avg refers to the average value obtained by running the experiment independently 20 times. The higher the Avg, the better the corresponding evaluation result. The Std denotes the standard deviation from running the experiment independently 20 times. The smaller the Std, the more stable the algorithm. The most significant mean and standard deviation are bold in the table. From Tables 22, 23, and 24, we can see that RDMVO can achieve an enormous average value and obtain a minor standard deviation in most cases. This shows that RDMVO has better performance and is more stable than other competitors.

Further, Tables 22, 23, and 24 are also subjected to the Friedman test. Tables 7, 8, and 9 show the average test results.

Table 7 Overall Friedman test results for each algorithm on PSNR
Table 8 Overall Friedman test results for each algorithm on SSIM
Table 9 Overall Friedman test results for each algorithm on FISM

PSNR measures the difference between the obtained segmented image and the original image, specifically the degree of distortion of the image. The higher the PSNR value, the closer the segmented image is to the original image. Table 22 shows the evaluation results of the PSNR of all algorithms. Table 22 shows that compared with other competitors, the RDMVO algorithm can obtain higher PSNR in most cases. Moreover, under the same test image, with the increase of the threshold level, the PSNR gradually becomes larger, which indicates that the RDMVO segmentation result is more accurate. Table 7 is the average result of the Friedman test on Table 22, avg is the average result over all test images, and rank is the ranking of the average result. The best Avg and Rank values are bolded. After the Friedman test, RDMVO can achieve first place at all threshold levels. This means that the RDMVO segmentation method based on Kapur’s entropy is statistically significant, indicating that RDMVO is better than other competitors.

SSIM evaluates the visual difference between the obtained segmented and original images. The larger the SSIM value, the closer the segmented image is to the original image visually. Table 23 shows the evaluation results of SSIM of all algorithms. It can be seen in Table 23 that compared with other competitors, the RDMVO algorithm can obtain higher SSIM in most cases. Moreover, under the same test image, as the threshold level increases, the SSIM gradually becomes larger, which indicates that the RDMVO segmentation result is more in line with human visual observation. Table 8 is the average result of the Friedman test in Table 23. The best results are bolded. After the Friedman test, RDMVO can also achieve first place at all threshold levels. Therefore, from the SSIM point of view, the RDMVO segmentation method based on Kapur’s entropy is also more statistically significant.

FSIM is to evaluate the feature similarity between the obtained segmented image and the original image. The evaluation results of FSIM of all algorithms are given in Table 24. In Table 24, it can be seen that compared to other competitors, the same as the above evaluation indicators, the RDMVO algorithm can obtain higher FSIM in most cases. Moreover, under the same test result, as the threshold level increases, FSIM gradually becomes larger. Table 9 is the average result of the Friedman test in Table 24. The best results are bolded. Similarly, RDMVO can achieve first place at all threshold levels after the Friedman test. Therefore, from the FSIM point of view, the RDMVO algorithm is also the best.

In addition, Tables 25, 26, and 27 give the Friedman test results of different images under different thresholds for all algorithms. These data are the data basis of Tables 7, 8, and 9.

Figures 12, 13, 14 and 15 are the convergence information of all algorithms when the threshold levels are 2, 3, 4, and 5, respectively. The best value is the average fitness value obtained by running each algorithm 20 times independently, and the abscissa is the number of iterations. It can be seen from the figure that under all threshold levels, RDMVO can achieve the best convergence speed and the most prominent fitness. The performance of RDMVO is satisfactory. Especially the comparison effect of RDMVO and the original algorithm MVO. It is obvious that the convergence speed and optimal value of RDMVO have significant advantages. The performance of the RDMVO algorithm we propose dramatically improves.

Figure 16 shows the specific thresholds obtained by all related algorithms when the threshold level is 3 for image 35,070. Figure 17 shows the segmented images obtained by different algorithms when the threshold level is 3 for image 35070. Figure 18 is the segmentation image obtained using RDMVO when the thresholds are 2, 3, 4, and 5. The first column has a threshold level of 2, the second column has a threshold level of 3, the third column has a threshold level of 4, and the fourth column has a threshold level of 5. Obviously, as the threshold level increases, the segmentation effect is better.

5.2.3 Comparison with Other Algorithms on COVID-19

Testing for COVID-19, we selected nine medical images. These are the X-ray images of nine patients with COVID-19, clinically recorded between January 21 and March 10, 2020, at a tertiary-level A hospital in eastern China. Our study was approved by the Ethics Committee of the Affiliated Yueqing Hospital of Wenzhou Medical University (Yueqing, China; protocol number 202000002) and complied with the Helsinki declaration. Except for the different test image sets, the experimental parameters and procedures are the same as in Sect. 5.2.3. Below we give the specific experimental results without further explanation. For the specific analysis of the results, please refer to Sect. 5.2.3. Figure 19 is the original test image, and Figure 20 is the grayscale image, grayscale value histogram, and 2D histogram corresponding to the original image.

Figure 7 is a graph of fitness convergence curves in all images corresponding to all algorithms. From a longitudinal perspective, there are 4 columns corresponding to different thresholds. The first column has a threshold level of 2, the second column has a threshold level of 3, the third column has a threshold level of 4, and the fourth column has a threshold level of 5. It can be seen from the figure that with different thresholds, the convergence speed of RDMVO can be the fastest, and the maximum fitness value can be obtained at the same time. From the horizontal view, there are 9 lines corresponding to different test images, from top to bottom, corresponding to the test images 03, 04, 05, 09, 12, 15, 20, 21, and 25. As can be seen from the figure, for all the test images, the performance of RDMVO far exceeds that of other competitors at different threshold levels. We know that selecting the threshold level generally depends on the actual practical needs. That is to say, the adaptability of RDMVO is greater.

Fig. 7
figure 7figure 7

Kapur’s entropy convergence curve of all algorithms at different thresholds

Table 10 shows the results of PSNR, SSIM, and FISM of different algorithms after the Wilcoxon signed-rank test with 5% significance at different threshold levels. The “+” denotes the number by which RDMVO is better than other algorithms. “-” means that RDMVO is worse than the number of other algorithms. “=” represents the number of RDMVO that is not significantly different from other algorithms. When the threshold level is 3, 4, and 5, and any PSNR, SSIM, and FSIM are used as evaluation indicators, RDMVO outperforms other algorithms on all images. Except that there is one image on which RDMVO performance is weaker than ACOR when the threshold level is 3 and the PSNR metric; there is one image on which RDMVO performance is weaker than DE when the threshold level is 5, and under the SSIM metric. When the threshold level is 2 and the PSNR metric is used, the performance of RDMVO is weaker than other algorithms on more images, but RDMVO performs better than the comparison algorithm in more images when compared to the number of “+” and “−”.

Table 10 The result of the Wilcoxon signed-rank test with 5% significance at different thresholds

Table 11 contains the average results of the three indicators PSNR, SSIM, and FSIM of all algorithms under different threshold levels after the Friedman test, and the optimal data are bolded. It can be seen in Table 11 that under the Friedman test, RDMVO is the first at all threshold levels on all images. The thresholds obtained by all algorithms at all threshold levels of the test images are given in Table 28.

Table 11 PSNR, SSIM, FISM results of all algorithms at different thresholds

Furthermore, we analyze the test image 03 at a threshold level of 5, and Fig. 8, Tables 12, and 3 are the experimental results of image 03. Figure 8 compares the grayscale and jet colormap images obtained by all algorithms based on Kapur’s entropy to segment the test image. Table 12 shows the mean values of PSNR, SSIM, and FSIM obtained by running 20 segmentation experiments independently for all algorithms, and the optimal data are bolded. Table 13 shows the standard deviations of PSNR, SSIM, and FSIM obtained by running 20 segmentation experiments independently for all algorithms, and the optimal data are bolded. It can be seen that the mean of RDMVO is the largest, and the standard deviation is the smallest. The all-around performance of the RDMVO algorithm is consistent with the overall performance in picture 03.

Fig. 8
figure 8

The grayscale image and jet colormap image obtained by all algorithms on image 03

Table 12 PSNR, SSIM, FSIM mean of all algorithms on image 03
Table 13 PSNR, SSIM, FSIM standard deviation of all algorithms on image 03

6 Discussions

This paper proposes an innovative algorithm RDMVO based on the MVO. Moreover, designs an image segmentation scheme RDMVO-MIS that is Kapur’s entropy MIS based on RDMVO to segment COVID-19-CXIs quickly and accurately. Then, a series of experiments were carried out to validate three aspects: the general performance of RDMVO, the feasibility of RDMVO-MIS, and the performance of RDMVO-MIS for BKIs and COVID-19-CXIs.

RDMVO introduces two mechanisms based on MVO, RM and DM. The impact of introducing the two mechanisms is demonstrated in the mechanism comparison experiment on IEEE CEC2017. The combined RM and DM can significantly enhance MVO’s ability to jump out of the local optimum and improve MVO’s solution accuracy. According to the convergence curves on some test functions, the RM mainly improves MVO’s ability to jump out of the local optimum. Only adding DM has no effect on improving MVO or has a negative effect, but the action of DM in the MVO improved by RM can further improve the convergence speed. In conclusion, combining RM and DM to improve MVO performance makes sense.

Then, in the performance comparison experiment on IEEE CEC2017 with some mainstream MHAs, RDMVO performed admirably, ranking first among all related experimental algorithms. RDMVO outperforms other algorithms, particularly in multimodal and hybrid functions. Although RDMVO performs well in composition functions compared to other excellent algorithms such as HHO, but there is still much room for improvement.

Further, MIS experiments on the BKIs prove that RDMVO-MIS is feasible. Compared with other MHAs, RDMVO has better performance. RDMVO consistently outperforms other MHAs in terms of the convergence speed on Kapur’s entropy function, the obtained maximum Kapur’s entropy, and the algorithm’s stability. Moreover, RDMVO outperforms other MHAs in the three most commonly used image segmentation evaluation indicators, PSNR, SSIM, and FSIM. Finally, we implemented RDMVO-MIS to segment COVID-19-CXIs and compared it with other MHAs. RDMVO also has the best performance and segmentation effect. In addition, the experiment’s threshold levels were set to 2, 3, 4, and 5, respectively. The threshold level selection in this paper is relatively small compared to these papers [13, 19,20,21,22,23]. This is mainly based on considering studies supplementing segmentation horizons with small thresholds. At the same time, it should be noted that RDMVO’s performance at high threshold levels has not been validated, and this is also something to consider for future works.

In summary, RDMVO has the best performance in all experiments, so it also can be applied to many other fields in the future, such as human activity recognition [126], information retrieval services [127,128,129], power flow optimization [130], location-based services [131, 132], endoscope imaging [106], image dehazing [113], image-to-image translation [133], autism spectrum disorder classification [134], disease identification and diagnosis [135, 136], dynamic module detection [137, 138], and pharmacoinformatic data mining [139, 140]. Different from the paper [17], the author only verified the COVID-19-CXI segmentation performance of the improved algorithm and did not verify the improved algorithm itself, and the performance of the improved algorithm universality is difficult to confirm. In this paper, in addition to COVID-19-CXI segmentation experiments, we also comprehensively tested the performance of RDMVO on the IEEE CEC2017 benchmark functions and BKIs. Moreover, the performance of RDMVO is satisfactory. As a result, it is highly recommended that managers try RDMVO when managers need to segment other types of images rather than COVID-19-CXIs. Of course, the parameters used in this paper are unsuitable for all problems as every problem has its character. The managers should adjust the parameters according to the specific problem to achieve maximum performance. In addition, according to the previous analysis of the time complexity of RDMVO, the dimension of the problem will directly affect the RDMVO’s running time. When using RDMVO, whether the time cost caused by the dimension is acceptable is also a problem that managers should consider for some high-dimensional problems.

7 Conclusions and Future Works

As COVID-19 spreads throughout the world, its diagnosis is vital. Chest X-ray image analysis and diagnosis are considered effective means. Experts and scholars worldwide have invested a lot of energy and time in improving the efficiency of COVID-19-CXI analysis. To better improve the effect of image segmentation, a novel improved MVO variant, namely RDMVO, was proposed and successfully applied to MIS and segmented COVID-19-CXIs. Experimental results show that our proposed RDMVO enhances MVO’s ability to explore globally and jump out of local optimum. Moreover, it outperforms most MHAs on 2D Kapur’s entropy MIS.

There is much work to be done in the future. First, MHA is a class of algorithms based on probability and randomness, although MHAs are widely used in various fields. However, the results of mathematical proof of MHAs are currently relatively lacking. Rigorous mathematical proofs are of great significance for the development of MHAs. More research is needed to provide rigorous mathematical proofs in the future, whether MVO, RDMVO, or other MHAs. Second, our experiments on RDMVO are only on MIS of BKIs and COVID-19-CXIs. Although, the images of different diseases have different characteristics. However, considering they all belong to medical images, we can still try to use RDMVO for image segmentation of other diseases. In addition, MHAs have already produced many research results in feature selection and engineering optimization. So, using RDMVO for feature selection and engineering optimization is also the future research direction. The most important and urgent work is to apply RDMVO to the actual COVID-19 diagnosis as soon as possible to help overcome COVID-19 and restore the regular operation of human social life.