1 Introduction

The use of interfaces for remote medical data visualization is currently in high demand. These interfaces provide access to patient records containing text and image data [1]. This sharing of this data, and in particular images on the Internet, exposes them to unauthorized manipulation [2], despite existing security tools such as access control. Watermarking has therefore been proposed to increase the sharing security by allowing keeping the patients’ data confidential and verifying the medical images' integrity [3]. Researchers have been increasingly interested in this field and several techniques have already been proposed in this field. However, the medical image has its own specificities [4]. The watermarked image must present the same clinical reading as the original image, so it must not suffer any degradation that affects the diagnosis [5]. The image watermarking methods can be classified according to the insertion domain. Spatial methods consist of inserting the mark directly into the image [6]. They have the advantage of being easy to implement and the insertion and signature detection operations are very fast and have low computation time but are not robust to geometrical attacks [7]. Frequency methods remain more recent methods whose principle is to insert the mark not directly in the image but in the transformed domain. To obtain the marked image, inverse transform is performed [8]. These methods are more resistant to geometric attacks [9]. We propose in this work a watermarking approach for the medical image protection. The integrated watermark contains information about the treated patient as well as data about the establishment. This will allow verifying during the extraction that the medical image really corresponds to the treated patient. This approach is applied in the frequency domain. The watermark is integrated into the coefficients obtained by transforms. The wavelet transform of a signal allows representing the signal on a two-dimensional space called the time-scale plane, providing joint information on the signal in time and frequency [10]. The paving of the time–frequency plane induced by this transforms has the particularity of allowing a fine temporal resolution at high frequencies and a fine frequency resolution at low frequencies [11]. This property often allows an interesting analysis of the signal but remains rigid. The interest of this transform is the optimization of the choice of locations and the strength of the signature in the image as well as its multi-scale aspect which offers a more robust distribution to the watermark [12]. Adding the watermark to the low-frequency coefficients can cause visible perceptual distortion, as the insertion of the watermark in the high-frequency coefficients is vulnerable to some types of attack like compression [13]. To obtain a good compromise between robustness and imperceptibility, the integration is performed in the medium frequencies of the image. To experiment this two approaches, we used the COVID-19 radiography database [14], and the proposed schemes were applied to watermark the radiographic images of this base. Distortion and similarity measurements (between the original and watermarked images) were then calculated to determine the imperceptibility of our approaches. To determine the robustness of our approaches, the most commonly used image watermarking attacks (Compression, filtering…etc.) are applied to the watermarked images. The results obtained will thus allow us to determine the most appropriate approach to the medical image protection. The paper is organized as follows, Sect. 2 will review the various related works that we will use for comparison, and Sect. 3 describes the proposed approach. Section 4 will present the experiments conducted and the results obtained. Finally, the conclusion of our work and the perspectives are proposed in Sect. 5.

2 Background

The two main areas of representation used in signal processing are time and frequency. The classical tool used to switch from one domain to the other is the Fourier Transform (TF). The TF allows estimating the frequencies present in a signal, but it does not locate the place where these frequencies appear or disappear, because it acts on the entire signal [15]. To have time–frequency type information, it is necessary to calculate the TF of a small window dragged from one end of the signal to the other. We then obtain a time–frequency representation. The disadvantage of this technique, besides its low joint time/frequency resolution, is that the window size is constant. It would be more relevant to adapt the size of the analysis window to the local characteristics of the signal: small window when the signal varies rapidly (high frequencies) and larger window when its variations are slow (low frequencies).Wavelet analysis aims to provide a solution to this problem by decomposing the signal on the basis of elementary signals [16]. Wavelets are obtained by dilating and shifting a basic wavelet.

2.1 The discrete wavelet transforms

The discrete wavelet transform is a mathematical formalization that allows to easily constructing orthonormal wavelet bases. The principle consists in decomposing the signal to be analyzed into a series of approximation coefficients and details, each series of approximation being decomposed again into approximation and detail. At each level of this decomposition, the detail coefficients correspond to the wavelet coefficients of the signal at a given scale. The Discrete Wavelet Transform (DWT) decomposes an image into four sub-bands: an LL sub-band of approximation coefficients and three sub-bands of horizontal (HL), vertical (LH) and diagonal (HH) detail coefficients [17]. To obtain the coefficients of each sub-band, we can calculate with the Haar filter as follows:

$$LL(x,y)=\frac{\mathrm{p}\left(\mathrm{x},\mathrm{y}\right)+\mathrm{p}\left(\mathrm{x},\mathrm{y}+1\right)+\mathrm{p}\left(\mathrm{x}+1,\mathrm{y}\right)+\mathrm{p}\left(\mathrm{x}+1,\mathrm{y}+1\right)}{2}$$
(1)
$$LH(x,y)=\frac{\mathrm{p}\left(\mathrm{x},\mathrm{y}\right)+\mathrm{p}\left(\mathrm{x},\mathrm{y}+1\right)-\mathrm{p}\left(\mathrm{x}+1,\mathrm{y}\right)-\mathrm{p}(\mathrm{x}+1,\mathrm{y}+1)}{2}$$
(2)
$$HL(x,y)=\frac{\mathrm{p}\left(\mathrm{x},\mathrm{y}\right)-\mathrm{p}\left(\mathrm{x},\mathrm{y}+1\right)+\mathrm{p}\left(\mathrm{x}+1,\mathrm{y}\right)-\mathrm{p}(\mathrm{x}+1,\mathrm{y}+1)}{2}$$
(3)
$$HH(x,y)=\frac{\mathrm{p}\left(\mathrm{x},\mathrm{y}\right)-\mathrm{p}\left(\mathrm{x},\mathrm{y}+1\right)-\mathrm{p}\left(\mathrm{x}+1,\mathrm{y}\right)+\mathrm{p}(\mathrm{x}+1,\mathrm{y}+1)}{2}$$
(4)

The information redundancy contained in this transforms can be generally interesting in the context of signal analysis. Indeed, it is interesting as long as the conditions to be checked by the wavelets are not too constraining. The digital reconstruction of the signal is stable. On the other hand, with the introduction of orthonormal wavelet bases, wavelet theory has become important in the sense that this orthonormal base concept remains a key tool in signal processing; it has speed constraints requiring the minimum amount of possible data. [18].

2.2 Non-subsampled contourlet transform

The non-subsampled contourlet transform (NSCT) is a translation invariant version of the contourlet transform. The contourlet transform uses a Laplacian pyramid for multi-scale decomposition and banks of directional filters for directional decomposition. To ensure translational invariance, the NSCT is implemented using a non-sub-sampled pyramid structure and non-sub-sampled directional filter banks [19]. The principle is to successively combine two translation-invariant decomposition stages. A non-subsampled pyramidal structure (NSP), which ensures the multi-scale property and a structure of non-subsampled directional filter banks. Non-Subsampled Directional Filter Banks (NSDFB), which allows decomposition according to different orientations [20]. The result (Fig. 1) is flexible, multi-scale, multi-directional, and translation-invariant image decomposition.

Fig. 1
figure 1

Non-subsampled contourlet transform: a NSFB structure, b idealized frequency partitioning

2.3 The non-subsampled shearlet transform

A shearlet system can be obtained by anisotropic expansion, translation and division of a finite number of generating functions. These operations are performed by applying a translation operator, an anisotropic expansion operator and a division operator. Unlike wavelet transforms, shearlets only concern 2D or 3D signals but are unsuitable for 1D signal [21]. Non-Subsampled shearlet transform (NSST) has been applied to serve as the shift-invariant and multi-scale expansion of the shearlet transform (Fig. 2), and it has the ability to eliminate the down-samplers and up-samplers. Typically, the implementation of the NSST includes two key steps: the application of Non-Subsampled pyramid (NSP) filter banks and Non-Subsampled shearing (NSS) filter banks [22].

Fig. 2
figure 2

Non-subsampled shearlet transform

The NSST is a shift-invariant version of the shearlet transform. It differs from the shearlet transform by eliminating of the down-samplers and up-samplers. This transform is a fully shift-invariant, multi-scale and multi-directional transform that combines the non-subsampled Laplacian pyramid transform with different combinations of the shearing filters. The analysis of the non-subsampled Laplacian pyramid can be done through an iterative procedure, where f is an image, NSLPj + 1 is the detail coefficients at scale j + 1, \({Ah}_{j}^{1}\) and \({Ah}_{k}^{0}\) are the low-pass and high-pass filters at scales j and k, respectively.

$${NSLP}_{j+1}=\left({Ah}_{j}^{1}\prod_{k=1}^{j-1}{Ah}_{k}^{0}\right)f$$
(5)

2.4 The discreet cosine transform

The discrete cosine transform (DCT) is widely used in digital signal and image processing [23]. It is a mathematical transformation that transforms a set of data from the spatial to frequency domain and vice versa using the IDCT (Inverse Discrete Cosine Transform). The DCT of an N × N block size for i, j = 1, 2… N is calculated as follows:

$${F}_{u,v}=\frac{1}{\sqrt{2N}} {C}_{u}\cdot {C}_{v}\sum_{x=1}^{N}\sum_{y=1}^{N}{f}_{x,y}\cdot \mathrm{cos}[\frac{\left(2x+1\right)u\pi }{2N}] \mathrm{cos}[\frac{\left(2y+1\right)v\pi }{2N}]$$
(6)
$$\mathrm{Cu}=\left\{\begin{array}{ll}\frac{1}{\sqrt{2N}} & \quad {\text{if}}\; u=0\\ 1 & \quad {\text{if}} \;u>0\end{array}\right.$$
(7)

Next, thresholding is performed to increase the efficiency of the compression. If the absolute values of the (non-zero) coefficients of the DCT matrix obtained are below a certain threshold [24], these coefficients will be eliminated (set to zero).The quantization phase then reduces the importance of all DCT coefficients by dividing them by a quantization step and replacing them with their quotient with this step. Quantification is therefore a simple Euclidean division of the DCT coefficients by a certain divisor (the step), which replaces the original coefficients by the quotient of the division. It is therefore at this step that we will lose part of the information, because after quantization the rest of the division is lost [25]. To calculates the quantization matrix, the quantization value Q is chosen (Q represents the number of bits needed to code each element of the DCT matrix). The two MAX and MIN values of the DCT matrix (DCTmax, DCTmin) are determined. Next, the DCTQ matrix is calculated by the following formula:

$$DCTQ(i,j)=\frac{(-1+{2}^{\mathrm{Q}})(\mathrm{DCT}\left(\mathrm{i},\mathrm{j}\right)-\mathrm{DCTmin})}{\mathrm{DCTmax}-\mathrm{DCTmin}}$$
(8)

2.5 Schur decomposition

In linear algebra, Schur decomposition [26] of a complex square matrix M is a decomposition of the form:

$$M=U\times S\times {U}^{^{\prime}},$$
(9)

where U is a unitary matrix (\(U\times {U}^{^{\prime}}=I\)) and S the upper triangular matrix.

The successful application of singular values decomposition in digital watermarking suggests the use of Schur decomposition for the same purpose. However, Schur decomposition represents an alternative to the singular values decomposition. Indeed, singular value decomposition requires about 11N3 flops as compared to a Schur decomposition which requires only 8N3/3. This represents about three times less computing time. However, the calculations should not be expected to be reduced by the same amount. This is mainly due to the fact that the integration and extraction procedures involve operations other than the Schur decomposition.

2.6 Related works

2.6.1 Watermarking approach using NSST

Wang et al. presents a novel robust watermarking algorithm in Non-Subsampled Shearlet Transform (NSST) domain using quaternion polar harmonic transform (PHT) and least squares support vector regression (LS-SVR). In this approach, the low pass subband of Y, U, and V color components are used to embed watermark information in NSST domain [27]. To more significantly improve the resilience of the watermarking algorithm against image processing attacks, the author adopt the redundant watermark embedding strategy, in which the watermark information is embedded repeatedly into the non-overlapping low frequency NSST coefficients blocks. Zheng et al. propose an adaptive embedding strength watermarking algorithm based on shearlets’ capture directional features [28]. In this approach, a DWT is first applied to obtain the subband decomposition. Then, a NSST is applied to the low frequency sub-band. Singular value decomposition (SVD) is applied to the watermark image, and then the principle components are extracted. The principle components, which represent the unique features of the image, are embedded into the host image to eliminate the false positive problem. Favorskaya &al propose a robust watermarking scheme based on the prerequisites of the HVS. In this approach, a perceptual channel decomposition of the host image is done to provide the appropriate regions for embedding. A non-subsampled shearlet transform is then applied to the selected regions [29]. SVD is then applied to the obtained NSST coefficient. Finally, the watermark bits are embedded by replacing the least significant bit of the singular values. To enhance the security, the watermark is scrambled by the Arnold’s transform.

2.6.2 Watermarking approach using NSCT

Chen et al. present a robust blind watermarking scheme based on Non-Subsampled Contourlet Transform and Singular Value Decomposition. In this approach, the original image is decomposed using the NSCT to obtain the coarse sub-band and band-pass sub-bands [30]. The resulting low pass approximation sub-band is decomposed into bloc (8 × 8) without overlapping. Singular value decomposition is then applied to each bloc. Each singular value is used to embed one bit watermark information. Gao et al. propose a non-subsampled Contourlet Transform based watermarking scheme [31]. In this approach, the cover image is transformed by NSCT to obtain low-pass subband. For enhancing the algorithm robustness, the directional subband with maximum energy is selected in mid-frequency subbands. For the watermark embedding, the region with the most abundant texture is selected based on the characteristics of human visual system (HVS). To enhance the watermark robustness, watermark bit information is repeatedly embedded into coefficients of a block selected randomly by logistic chaotic map with a key.

2.6.3 Watermarking approach using DCT

Hu et al. proposes a blind watermarking scheme based on discrete cosine transform (DCT) to integrate binary information into images [32]. This scheme modulates the Partially Signed Average (PSA) of some DCT coefficients derived from an 8 × 8 picture block. The variation margin of each coefficient is regulated by consulting a JPEG quantization table to reinforce the imperceptibility of the watermark. In the approach proposed by Moosazadeh et al. a DCT of the image is calculated [33], then, the parameters and position of the coefficients to be modified is determined automatically by a method of Teaching–Learning-Based Optimization. The proposed method uses TLBO, which has rarely been applied in watermarking and allows to automatically determining the integration parameters and the appropriate position for the watermark insertion. Assini et al. [34] proposes a robust watermarking scheme combining discrete wavelet transform (DWT), discrete cosine transform (DCT) and singular value decomposition (SVD). A DWT is applied to the medical image to obtain different frequency sub-bands. Then a DCT is applied to the HH sub-band. An SVD is then applied to the obtained DCT coefficients. Finally, the watermark is integrated into the singular values obtained from the HH high-frequency sub-band of the medical image.

2.6.4 Watermarking approach using DWT

Verma et al. present a hybrid technique [35] combining DWT and SVD for the watermarking of medical images (to integrate the advantages of both techniques). A DWT is applied to the image to obtain the different frequency sub-bands. Singular value decomposition is then applied to the LL sub-band. The watermark is integrated into the S-value using a scaling factor. The combination of the modified part S and the original parts U and V provides a modified LL sub-band. An IDWT with the modified LL sub-band is finally applied to obtain the watermarked image. Anand et al. proposes a multi-watermark embedding method for medical image [36]. The embedding watermark is composed of a text and an image. A four sub-band decomposition of the host image is first performed by applying a DWT. An SVD is applied to the HH sub-band, and the text is embedded in the singular value S. The image watermark is divided into two sub images, these two parts are then integrated (after applying an SVD to both sub bands) into the singular S values of the HL and LH sub bands of the cover image. To enhance security, the watermarked image is then encrypted using Hyperchaotic encryption algorithm. In the approach presented by Gangadhar et al. [37], a sub-band decomposition of the medical image is first performed by applying a DWT. Then, an SVD is applied to the LL subbands. To determine the most suitable regions for the integration process, the particle swarm optimization algorithm is then applied to the LL subbands. Alshanbari et al. [38] combine two different watermarking insertion methodologies. In this work, multiple watermarks are integrated into the medical image. A secret key is first generated based on SHA-256 algorithm using the image ROI. The ROI is compressed with Lempel–Ziv–Welch (LZW) algorithm and then, combined with signature information. The generated watermark is then inserted into the RONI region of host image. Devi et al. [39] present a blind watermarking scheme that combines DWT and SVD. A DWT is applied to the blue component of the cover image; the resulting LL sub-band is then decomposed into 8 × 8 blocks. For each block, a fast Walsh Hadamard transformation is applied. Singular value decomposition is then applied to the blocks of the LL sub-band formed by the Hadamard transform. The watermark bits are then integrated into the singular values obtained. To increase security, the watermark image is scrambled using the Arnold Transform.

3 Proposed watermarking scheme

A watermarking scheme generally consists of three important processes, the mark generation, the integration process and the extraction process [40].

3.1 Watermark generation

The watermark used in our approach consists of patient information and image acquisition data. This information is concatenated and a hash is computed using the MD5 algorithm (Fig. 3). This hash will guarantee the integrity of the extracted watermark. The extraction of this data will precisely allow identifying the patient and avoiding any confusion.

Fig. 3
figure 3

Watermark generation process

3.2 DWT–Schur watermark integration and extraction process

For the watermark integration, a DWT is first applied to the image to obtain the four subbands LL, HL, LH and HH. This representation makes possible to distinguish very specific regions around the image contours and textured areas. These regions correspond to areas where there is a high density of significant (or dominant) wavelet coefficients. Some works are based on modifying the coefficients of all the sub-bands; others rely on inserting in HH or LL. To improve robustness to some attacks like compression, filtering and scaling while maintaining a good imperceptibility rate. In our approach the LH and HL subbands are used for the watermark integration (Fig. 4). Schur decomposition is then applied to the LH and HL subbands. The upper triangular matrix is then used for the watermark bit integration. Every three successive values of this matrix are modulated to integrate the watermark bits according to the rules in Fig. 3. These rules allow modifying only the parity of a single value, so for the integration of two bits of the watermark a single bit will be modified, the inequality must be modified to match the equations to the watermark bits, this will allow the extraction of the watermark in blind mode (without using the original image).

Fig. 4
figure 4

Integration process for the DWT–Schur algorithm

The receiver will just have to apply equations (Fig. 5) to get the watermark bits without using the original image. Finally, an inverse Schur decomposition is applied to obtain the modified subbands. HL and LH, an inverse DWT is then applied to the four subbands LL, HL, LH and HH to obtain the watermarked image.

Fig. 5
figure 5

Substitution rules

For the extraction process, a DWT is applied to the watermarked image, as shown in the diagram block in Fig. 6. Then, Schur decomposition is applied to the resulting LH and HL subbands. As for the DCT–Schur algorithm, the watermark bits are extracted from the upper triangular matrix.

Fig. 6
figure 6

Extraction process for the DWT–Schur algorithm

3.3 DCT–Schur watermark integration and extraction process

Calculating the DCT over the entire image would take a considerable time. To avoid this, in our approach, we apply DCT on fixed-length blocks (8 × 8 pixels), this choice gives a better compromise between quality and computation time. The image is thus divided into blocks (8 × 8) without overlapping (Fig. 7), and then a 2D DCT is applied to each block. This transformation will allow separating the low frequencies from the high frequencies. All the information of the image is in the low frequencies. On the other hand, the details of the image are located in the high frequencies. If the mark is inserted in the low frequencies, the watermark will be robust but the mark will be visible. If the mark is inserted in the high frequencies, it will be imperceptible but the watermark will be less robust because of the compression operations. To find a compromise between imperceptibility and robustness, medium frequencies are selected for the integration process. Schur decomposition is then applied to the medium frequencies obtained; the upper triangular matrix is then used for the watermark bits integration. Every three successive values of this matrix are modulated to integrate the watermark bits according to the rules in Fig. 5. Finally, an inverse Schur decomposition is applied to obtain the modified mid band frequency coefficients. These blocks are combined with the other blocks, and then an Inverse DCT is applied to obtain the watermarked image.

Fig. 7
figure 7

Integration process for the DCT–Schur algorithm

For the extraction process, a DCT is applied to the watermarked image, as shown in the diagram block in Fig. 8. Then, Schur decomposition is applied to the mid band frequency. The watermark bits are then extracted from the upper triangular matrix.

Fig. 8
figure 8

Extraction process for the DCT–Schur algorithm

3.4 NSCT–Schur watermark integration and extraction process

For watermark integration, a non-subsampled contourlet transform is applied to the medical image. The NSCT allows obtaining the approximate low-pass subband B and directional subbands C. The watermark bits are then integrated into the directional subbands coefficients with maximum energy in the mid-frequency subbands. The calculation of the subband energies is calculated according to the formula:

$$\mathrm{E}=\frac{1}{\mathrm{M}\times \mathrm{N}}\sum_{m=1}^{M}\sum_{n=1}^{N}{\left(C\left(m,n\right)\right)}^{2},$$
(10)

where M and N represent respectively the width and height of the directional subband.

Schur decomposition is then applied to the obtained medium frequencies (Fig. 9), the upper triangular matrix is then used for the watermark bits integration. Every three successive values of this matrix are modulated to integrate the watermark bits according to the rules in Fig. 10.

Fig. 9
figure 9

Integration process for the NSCT–Schur algorithm

Fig. 10
figure 10

Substitution rules

Finally, an inverse Schur decomposition is applied to obtain the modified mid band frequency coefficients. These blocks are combined with the other blocks, and then an Inverse NSCT is applied to obtain the watermarked image. For the extraction process, a NSCT is applied to the watermarked image, as shown in the diagram block in Fig. 11. Then, a Schur decomposition is applied to the mid band frequency. The watermark bits are then extracted from the upper triangular matrix. The binary sequence obtained is divided into three parts, and the first part represents the patient’s personal information, and the second part contains the identity of the doctor in charge of this patient as well as the image acquisition parameters. The third part contains the hash of the two first parts; this binary sequence is very useful to verify the integrity of the extracted data. This ensures the patient or image information has not been altered during transfer or by an unauthorized entity.

Fig. 11
figure 11

Extraction process for the NSCT–Schur algorithm

3.5 NSST–Schur watermark integration and extraction process

For the watermark integration, the image is divided into blocks (8 × 8) without overlapping, and then a NSST is first applied to the image. This representation makes possible to distinguish very specific regions around the image contours and textured areas. These regions correspond to areas where there is a high density of significant (or dominant) coefficients.

Schur decomposition is then applied to the low frequency NSST sub-bands (Fig. 12). The upper triangular matrix is then used for the watermark bit integration. Every three successive values (C1, C2 and C3) of this matrix are modulated to integrate the watermark bits X and Y according to this rule:

Fig. 12
figure 12

Integration process for the NSCT–Schur algorithm

$$\left\{\begin{array}{l}C1++, \, {\text{I}}{\text{f}}((X\ne (C1-C2) \% \;2)\; {\text{and}}\; (Y\ne (C1 - C3) \%\; 2))\\ C2++, \;{\text{If}}\;\left(\left(X= \left(C1- C2\right)\%\; 2\right)\; {\text{and}}\; \left(Y\ne \left(C1- C3\right)\%\; 2\right)\right)\\ C3++, \;{\text{If}}\;((X\ne (C1- C2) \%\; 2)\; {\text{and}}\;(Y= (C1 - C3) \%\; 2))\end{array}\right.$$
(11)

These rules allow modifying only the parity of a single value, so for the integration of two bits of the watermark a single bit will be modified, the inequality must be modified to match the equations to the watermark bits, this will allow the extraction of the watermark in blind mode (without using the original image). Finally, an inverse Schur decomposition is applied to obtain the modified Low frequency subbands. These blocks are combined with the other blocks, and then an Inverse NSST is applied to obtain the watermarked image. For the extraction process, a NSST is applied to the watermarked image, as shown in the diagram block in Fig. 13. Next, Schur decomposition is applied to the Low frequency subbands. The watermark bits are then extracted from the upper triangular matrix.

Fig. 13
figure 13

Extraction process for the NSCT–Schur algorithm

4 Experimental results and discussions

In this section, several experiments are taken to evaluate the performance of our proposed watermarking scheme. The experiments were performed using images from the COVID-19 radiography database (Fig. 14). This database contains chest X-ray images for COVID-19 positive cases along with Normal and Viral Pneumonia images.

Fig. 14
figure 14

Medical images used for the experimentations

In this release, there are 219 COVID-19 positive images, 1341 normal images and 1345 viral pneumonia images. All the images are in Portable Network Graphics (PNG) file format and resolution is 1024-by-1024 pixels. In this experiment, all methods are performed on the same platform of 2.27GHZ CPU, 2.00 GB RAM, Win 7 and MATLAB 8.5.0 (R2015a). The experimental results are divided into two parts: the first is focused on measuring the imperceptibility while the second is dedicated to the robustness analysis against some types of attacks. In medical images, the imperceptibility concept remains the most important property and differs from the case of general public multimedia.

Indeed, watermarking can hide subtle signs or introduce new artifacts that will falsify the image's interpretation. The consequences will be more or less serious, even life-threatening. As we can see on Fig. 15, it is difficult to make the difference between the original and watermarked images; the degradation of the watermarked images is visually imperceptible.

Fig. 15
figure 15

Watermarking results, a original image, b image watermarked using DCT–Schur, c image watermarked using DWT–Schur, d Image watermarked using NSST–Schur, e image watermarked using NSCT–Schur

4.1 Imperceptibility test

To evaluate the imperceptibility of our approaches, we used the PSNR (peak signal to noise ratio) which is a distortion measure widely used in digital imaging and especially in image compression [41]. The aim is to quantify the performance of encoders by measuring the quality of the watermarked image compared to the original image. While the PSNR is useful to measure the proximity of the watermarked image to the original, it does not take into account the visual quality of reconstruction and cannot be considered as an objective measure of the visual quality of an image [42]. For this purpose the structural similarity index measurement (SSIM) is calculated. Also, after extraction of the mark, the normalized correlation coefficient (NCC) is calculated using the original mark and the extracted mark. This coefficient is used to evaluate the existence and accuracy of the extracted mark. The number of values changed (NVC) by the integration process is also calculated between the original and watermarked image. The standard deviation (STD) is also calculated. The average values obtained for the 200 images used will thus be represented in all the following tables.

As we can see in Table 1, the PSNR rate offered by the DWT–Schur approach is relatively higher than that of the DCT–Schur algorithm, as well as the mean square error which is slightly lower for the DWT–Schur approach, the rate of structural similarity index measurement is almost identical, however the number of values modified by the DCT–Schur approach is significantly lower than the rate of values modified by the DWT–Schur algorithm. To evaluate our two proposed algorithms, the imperceptibility results obtained are compared with the approaches presented in Sect. 2 (Table 2). However, since the watermark used is composed of a hash, the alteration of a single bit of this hash will result in the loss of the embedded fingerprint message. This will then make impossible the verification of the extracted watermark's integrity.

Table 1 Imperceptibility results
Table 2 Imperceptibility results comparison with related work

4.2 Robustness test

In order to preserve the patient's or doctor's data, the watermarking algorithm must resist to the various attacks. In order to verify the robustness of the proposed algorithms, we tested both algorithms against various types of attacks, including JPEG compression, Gaussian noise, Sal and Pepper noise, median filter and other attacks. The normalized cross-correlation is then calculated between the original and extracted watermark (Table 3).

Table 3 Robustness comparison with related work

4.3 Discussion

The image decomposition into sub-bands has allowed separating the low-frequency components, which constitute an insertion space for the signature less sensitive to modifications. On the other hand, the spatial content of the image remains preserved after a multi-resolution transformation, this similar content can then be used to locate the mark after a geometric transformation. In this work, we also reviewed recent transformations. These multi-scale decompositions, operating according to a multitude of frequency orientations, provide a satisfactory compromise between the scattered (or compact) representation of visual information and the perceptual quality of the reconstructed image. In particular, the Non-Subsampled Contourlet Transform. The NSCT represent a discrete version, therefore adapted to digital images, which offers better directional selectivity and maintains the property of translational invariance. Unlike the contourlet transform, NSCT is a fully translation-invariant, multi-scale, multi-directional image decomposition that has better frequency selectivity and fast implementation. However, it is important to highlight that the wavelet transform is not optimal for modeling the anisotropic features of an image. This is due to the isotropic nature of the operators used. Since these operators process the two axes of the image equally, 2D high-frequency wavelets are well unsuited for singularities other than vertical, horizontal or diagonal singularities. The shearlet transform, which is a recent decomposition based on anisotropic operators, has been used to overcome this limitation. The imperceptibility results obtained remain acceptable. The use of the DCT in the first approach allows for high coefficient dissociation, as most of the information contained in an image is concentrated in a reduced number of coefficients, which facilitates integration. In addition, the signal energy included in the transformed domain will be distributed over all pixels in the spatial domain, which considerably improves the imperceptibility. Representations at different resolutions in the second approach (DWT–Schur) allow the extraction of the image's features into a small number of coefficients, while precisely locating the discontinuities. Hence, the most significant areas of the image are highlighted. The robustness results obtained are practically similar for the four approaches. Both approaches allow generating a watermark robust to filtering and Gaussian noise attacks; however we can see that the DCT approach is more robust to compression attacks. The discrete cosine transform makes the watermark more robust to JPEG compression, since they exploit the same space used for image coding. It also reduces the spatial correlation between pixels in an image. In addition, each block is independently transformed in the DCT domain which improves the algorithm's robustness. Moreover, Eigen values in the Schur decomposition are remarkably stable, so the use of the upper triangular matrix for the watermark integration allows a better robustness.

5 Conclusion

The digital image watermarking finds an application in the medical imaging domain and particular in the field of telemedicine. Indeed, given the importance and growth of the practice of remote medicine, watermarking can be proposed to contribute to the security of medical images shared on the Internet. The medical image, given its specificities, must be handled with considerable precautions. An image degradation which would seem insignificant is often not accepted in this field because it could lead to an incorrect diagnosis. Currently, a great interest is given to the substitutive watermark which represents the methods for which the mark is not added but substituted and formed by modification of certain image components. In this work we have proposed a substitutive watermarking approach for the medical image protection. These schemes allow the protection of the patient's information and therefore ensure the confidentiality of personal data. Moreover, these schemes ensure the integrity of the hidden data through a hash embedded in the image. Imperceptibility and robustness experimental results show that the proposed methods maintain a high quality of watermarked images and are very robust against several conventional attacks. However, the principal disadvantage of this approach is that if the hash is altered, it makes impossible to verify the integrity of the extracted information. The proposed scheme can be integrated into a medical document authentication application where any user can check whether a document from the database is authentic or not.