An efficient hidden marking approach for forensic and contents verification of digital images

Recently, forensic tools have been presented for detecting the forged image and illegal image manipulations. This paper presents an efficient and simple image contents verifying approach acting as a forensic technique. The presented approach mechanism is built by adding a hidden mark in a secret images, this mark achieves image integrity verification and detecting the tampering or forgery in the secret/authentic images. The secret/authentic/authentic image is divided to two main partitions, each portion is segmented to small blocks. These blocks in one partition are used to mark the blocks of the second partition using data transform technique. Firstly, the sensitive image is marked according to a self-embedding method. Then, a transform domain is utilized in order to embed a block-based signature into another block of the same image. Common discrete transform domains like DWT, DCT, and DFT are examined individually. Different analyses and comparison measurements are employed. The DCT is proved to be the most suitable and efficient transform domain to be used with the proposed scheme. At the receiver side, the reverse process is performed to verify image integrity. As proved from the experiments, this mark-algorithm is not visible or observable and robust against various attacks.


Introduction
Data integrity is one of the essential security requirements, it is the assurance that a sensitive data cannot be modified or accessed, and the only authorized persons do accessing or modifying in the classified data. In a highly sensitive applications such as; nuclear and military applications when the integrity of a related sensitive data is violated, a vital process may be disrupted as a consequence. For the reason of rapid advances in data processing techniques, the data content can be easily attacked, accessed and modified. Therefore, the capability of the integrity verification techniques should be focused on to be more sensitive to detect any unintentionally modifications (different noises added through transmission and unintentionally user misuse) or intentionally modifications (masquerade and forgeries). Digital watermarking methods are not only utilized for approving the owner identity, but also for protecting the multimedia content integrity.
The data forgery and image manipulation process, these issue are happened every day around the world. With increasing the amount of data transferring over the internet under the Internet of Things (IOT) or Cloud computing titles for different applications, this big amount of data need to protected from any manipulation and forgery. The level of the cyber crime will raise with the increasing data transferring. on the other hand, with the time, all people in different countries will depend on the internet. All these factors clears the need of a simple data contents protection approach. The rate of cyber crime around the world increases rapidly. That clears in the flow graphs of the cyber crime in the internet. So, the presented self-marked approach achieves the data contents protecting with a simple algorithm.
To verify image content integrity and detect various types of image forgery, different techniques have been emerged. These techniques should be capable of sensing any small image tampering and should be utilizable in many applications such as; law, journalism, national defense, etc. [7]. Image integrity could be verified using digital watermarking. Generally, watermarking technique embeds a message assigned as a watermark that's may be visible or invisible into an image, and then on the other hand the embedded watermark is extracted to verify the image content integrity. Traditional watermarking schemes require the receiver side to own the watermarks that are used by the sender to be compared with the extracted one as in [9,49], which means that the watermark has to be created in a separate manner and transmitted in a secure route to a particular receiver which may compromise the security of the scheme.
Other common limitations of several watermarking schemes are; the necessity of the receiver to have a former knowledge about the transmitted image, i.e. image size, private keys, lookup table, etc. [14,45]. Algorithms that depend on interrelationships among neighboring pixels during the watermarking process are not able to detect the cropping on the watermarked image [44]. Some algorithms require an extra processing overhead in watermark creation, embedding, extraction, and integrity verification, such as in [47]. In this method, the authors proposed an authentication and image integrity verification scheme using a public key cryptosystem and a content-based watermarks. On the transmission, a characteristic map of the underlying image is obtained as a watermark and then partitioned into blocks. A watermarked image to be created a neighboring feature map and watermark blocks are mixed and encrypted for embedding process. On the reception, the characteristic map is extracted from the obtained image and matched with the retrieved watermark to confirm the authenticity and integrity. This scheme is capable of detecting addition of foreign objects, removal of original objects, geometric transformation, identifying rigging, and cropping. Additionally, this public system does not need protected channel to deliver additional information like image dimensions and/ or secret key. However, with RSA cryptosystem protection, it requires extra processing overheads in watermark extraction, embedding, and verification processes.
Also, the generality of previous efforts related to the same issues have concentrated on investigating methods under only the ideal case of noise-free channel [8,35]. These methods are not practical when utilized to transmit sensitive images over the error-prone (noisy) wireless channel. Subsequently, there are a number of solutions that have been proposed for verifying image authentication in random packet loss existence case [46]. However, they often have a higher computational complexity, so their application in mobile communication may become critical, where the signature scheme should be efficient enough to verify authentication without any extra delays. The authors in [40] tried to overcome the security and data transmission constraints imposed by the wireless medium by proposing a signature-based image authentication technique. Although it tries to balance between the security and robustness, but a higher processing capability is still required.
The literature introduced many techniques [4,5,8,12,20,23,24,27,29,35,40,44,46,47] that have critical problems with watermark verification, extraction and complexity considerations. They have low robustness and imperceptibility, i.e., in the presence of multimedia attacks, they failed in achieving adequate watermark and in extracting watermark subjective and objective qualities. Furthermore, in watermark embedding process, they failed in picking the most suitable regions inside the images. Thus, they increase the computations as well as effecting the quality and imperceptibility of the watermarked images. Considering the limitations of the state-of-the-art image integrity and verification watermarking techniques, this paper presents an efficient and secure image integrity and verification communication system to preserve both robustness and imperceptibility. Moreover, compared to the traditional image integrity and verification watermarking techniques, the proposed work has minimal computational overhead, high quality, high robustness, good imperceptibility, and sufficient resistance to most forms of multimedia attacks [2, 3, 6, 41-43, 48, 52].
It is know that the machine learning based is an approach for image classification [16,26,35]. It is also, employed as a method for verifying the data contents [22]. In [22], authors presented image forgery approach based on the Support Vector Machine (SVM). The social media security attacked the focusing of several researchers [17,25,50]. Really, it is very essential research point. In [50], Image authentication approach on the social media has been proposed based on the deep learning concepts utilizing the neural network. The presented approach in this research paper can be more efficient and acceptance for applicability and lower complex than the other approaches. The deep learning based approach utilizing for the data contents verification works according to two partition, the training and learning (feature extraction), the second partition is the classification process. On the other hand, the dependable approach for data verification is presented in this research paper, it is evaluated with presence the different attacks for measuring its stability and applicability. The presented approach aims to verify the image contents by put invisible mark to the image, this mark is very small of piece from image replaced by other piece with considering this mark is not observable. The decryption process here means verifying the contents of marked image. If there is no manipulation, the algorithm passes the verification process or if there is any littlie change (1 bit) or more, the verification process will detect this change and its place with high accuracy. Also, the reliability of this approach has been studied with presence the different attacks.

Contribution
In this section, our contribution in this research paper has been presented. Simple algorithm with lower complexity is proposed for achieving sufficient integrity of image contents. This algorithm is built based on data transform tools for hiding mark within the secret/authentic images. Utilizing this algorithm in the contents verification process provides capability for detecting any tiny tampering or forgery in the images due to any effects or attacks. Various data transforms techniques has been tested for choosing the suitable tool for performing the algorithm. Also, this research paper presents brief analytical describing of image processing steps and various format for executing the presented algorithm on the secret/authentic images. This paper presents an approach for protecting and verifying integrity of image data content by exploiting one of the most popular transform domain techniques. The following sections will present in detail the proposed scheme and evaluate its performance in different situations of image integrity violations. The arrangement of the remnant of this paper is as follows. Section 2 gives a briefing review on DCT. Section 3 provides a systematic description of the proposed scheme. Then, the results analysis, comparisons, and the advantages and applications of proposed approach are presented and explained in section 4. Finally, Section 5 concludes this paper.

The discrete cosine transform
A Discrete Cosine Transform (DCT) generally can express a dedicated information point series like a summation of cosine functions oscillating at various frequencies. The 2D DCT is the most common category of DCT, which often used in signal and image processing [29]. The DCT when used with image data separates it into high, middle, and low frequency parts (FH, FM, and FL) as presented in Fig. 1. In low frequency component, much of the signal energy is located that includes significant visual portions of an image. In the other side, high frequency components are usually removed through compression. Therefore, a mid-band frequency is chosen for the use in watermarking. In this case, the watermark cannot be removed by compression and the embedding process will not alter the image visibility [19].
The parameters of Eq. (1) is presented as follows:-N and M represent image dimensions, rows and columns, respectively. Symbol f(i, j) represents the pixels intensity for pixel P(i, j), F(u, v) refers to the coefficients of DCT transform of pixel P(k 1 , k 2 ). The DCT transform contains three bands of frequencies, low frequency (LF) in upper left corner in the DCT coefficient matrix, medium frequency (MF) and high frequency (HF) which lies in low right location in DCT coefficient matrix. The DCT separates the image into parts of different frequencies and the top left coefficients correlate to low frequencies of authentic image block which the human eye is more sensitive to their changes, and as we move away from top-left corner in all directions, the DCT coefficients correlate to higher frequencies of authentic image block which the human eye is less sensitive to their changes, so they can be discarded in compression process, and hence this concept is basically exploited in the construction of the proposed integrity verification algorithm which the changes in the last bottom row, and the last right column for each image block does not affect the image visibility significantly.

Verification algorithm description
The model investigated in this work can be considered as a multi-task scheme that can be used for confidential image integrity verification. It is considered self watermarking algorithm without cover need. There are several research papers proposed image security techniques [10,11,21,51]. These presented techniques in the previous research work tried to develop the traditional image security tools which are the encryption or the hiding through the watermarking or the steganography [1,28,37]. The main contribution of the proposed work that it tries to avoid most of popular limitations in traditional watermarking techniques and can be characterized by: & Suitable for image confidence verification, which means that it can efficiently differentiate between received marked and non-marked image. & Applicable to be applied on a wireless channel at different noise levels. & It has an acceptable applicability to verify the integrity of an encrypted version of the image, so a higher security level can be obtained (confidentiality and integrity achieved at the same time).
The proposed model consists of two sub-models. The first sub-model implements the embedding process while the second one implements the verification process. Figure 2 describes briefly the proposed algorithm where an input secret image is marked using embedding process. The marked image is possibly manipulated through distribution by different manipulation reasons and intentions. On the other side when a marked (possibly manipulated) image is received, it will be subjected to a verification process which acts as a decision-making process. If the received image is safely distributed (not manipulated), the output of verification process will be a truly extracted copy of an authentic image and its integrity is verified, otherwise the verification process output will be a distorted image. The visual properties and statistical analysis will tend directly to the amount of modification which the marked image is subjected on the distribution process.
The two models are explained in sections 3.1 and 3.2. In addition, to determine which one of the three commonly used frequency domain transforms (DWT, DCT, and DFT) is more applicable to be used in the proposed model, a comparison among those transforms has been introduced with respect to having acceptable image analysis results.
In the following: simple mathematical represents the presented algorithm, where secret/ authentic image is represented in a matrix within different format.

-Mathematical Expression
In the following, a simple mathematical model of the image segmentation in the markembedding algorithm is presented. Firstly, the algorithm is built based on the self mirror concept, two equals partitions work as a mirror together. In Eq. (2), the original data is represented as a matrix, it dimensions are (MxN), and it's elements are given in Eq. (3). The data matrix F is converted to square matrix as in Eqs. (4) and (5), the data form in Eq. (3) is formatted to square form, it' dimensions (YxY).
The image is segmented to two equal partitions, up and down partitions, as given in Eq. (5) (The) F Square (y, y) is divided to equal partitions as shown in Eq.
The elements of the processed image in the square matrix form are formulated in Eq. (6) as follows: F Square y; y ð Þ ¼ The second step is square image dividing to two partition, Eq. (7) represents the square matrix with considering the segmentation process, as shown in this Eq. (8), there are another variable is appeared (β), this symbol equals the small blocks size as mentioned in the previous section.
The symbol β denotes to the size of small blocks, it proves the flexibility of this algorithm, it can be executed employing the different β values, β = 4 means the blocks size is 4 × 4 matrix, β = 8 means the blocks size is 8 × 8 matrix and β = 16 means the blocks size is 16 × 16 matrixes and etc. The symbol β can be employed in the combined different security levels.
The every block can be encrypted individually utilizing the different secret key, where the secret key length is determined based on the size of the encrypted matrix. The encryption process merging in the marking embedding algorithm is considered additional feature in this presented work. Moreover, there are two approaches for the merging process, the marked image encryption approach, in this approach, the mark embedding process is followed by the encryption to produce the combined multi-levels of security. The second approach is the marking of the encrypted image, this approach starts with the encryption process, this encrypted image represents the input image of the algorithm for producing the marked encrypted image. The suitability of these two approaches is determined according to the level of the image sensitivity and its contents importance. I. In Eq. (9) example for the processed image segmentation to small equal blocks in case of β = 8.
In Fig. 3, the mirror concept is applied on the square matrix as shown. There are several options for the mirror applying as shown in Fig.1a-e. As shown in this figure, there are four styles can be used for Segmenting the data square matrix form to two equal partitions, These styles are denoted by M 1 , M 2 , M 3 and M 4 . In every style, the two partitions are segmented to equal small blocks, as the two partitions have the same size, the blocks in each partition contains the same number of blocks and the same size. In the following: the simple mathematical equations have been presented to formulate these four styles.
The style M 1 and M 3 are known by the ordinary styles, they can be expressed mathematically in simple form. On the other hand, the styles M 2 and M 4 are not applicable for the images processing, it may accepted and applicable for the audio or text information. In general, the mathematical model of these styles is discussed and presented. The presented mathematical model of the data contents verifications integrity must be invisible and not observable by the human eyes, this is the only condition for the applicability in the real applications for the recorded/archived multimedia information or for the image, audio, video, and text transmission over the different wireless noisy channels. Figure 3b is the used style in this presented algorithm.
In our presented research paper, the style M 1 is considered in the presented approach for achieving the image contents verifications against any forgery of illegal manipulation. For enhancing the security and confidentiality of the presented approach the style models "M 2 , M 3 and M 4 " can be utilized for marking the secret/authentic/authentic images. As shown in Fig. 3, styles M 3 and M 4 are complex more than styles M 1 and M 2 .

Embedding process
This section presents how to mark highly sensitive image data that may be archived for a dedicated use or transmitted over untrusted communication medium. This embedding process takes a confidential image, as an input, and produces a signed block-based image (Marked Image), as an output. Utilizing a standard gray scale image as the authentic image, the embedding process is described briefly in the block diagram shown in Fig. 4 is depicted in stepwise as follows: The operation mechanism of the marking process can be briefed as follows: firstly, the processed image is divided to upper and lower partition, these two partition equal in the dimensions. Also, the number of blocks in partition f1 equals the blocks in the partition f2. The partitions f1and f2 form the image, these partitions are processed together. The image always is one piece, But, the marked image will generated from the original one by replacing the very

Verification process
This model implements the reverse process of the embedding scenario. It takes the marked image as an input and reconstructs the true authentic image in case of no modification occurs through distribution. Figure 5 and the following steps describe the flow of extraction process with a DCT as an efficient TD used with embedding process which will be shown clearly in the comparison results in the following sections: 1-Input the marked image.  4 Results analysis and comparative study MATLAB R2018a with windows7 environment were utilized to establish simulation experiments. The performance of the proposed model was evaluated in case of popular transform domains. In DFT case, the pervious steps are almost implemented. However, in DWT case, a 2-level Haar DWT is utilized to every block that results in the creation of the four bands, i.e. HH1, LH1, HL1, and LL, with size (4 × 4). The LL band is chosen for inserting a mark of the same order block in the second part of the image. All tests were applied on various benchmark gray scale images with size of (256 × 256) as an authentic image, which are presented in Table 1. These images are used for testing the applicability and validity of the presented algorithm. The testing images contain standard Matlab images and X-ray image (medical image), it is not standards image, real image (reactor image) is not standards image for representing the variety in the testing process.

Evaluation of embedding process
Quality metrics are utilized to evaluate the performance analysis of the proposed scheme. These quality metrics are calculated and discussed for every transform domain individually to determine which transform domain is more suitable and efficient for using with the proposed algorithm.

The image histogram analysis
Histogram is widely used as a metric for image encryption/decryption technique [32]. In general, Histogram is considered a visual graph presentation for summarizing the data/image and it represents the frequency graph of numerical data by rectangles graph. The origin of "Histogram" term is given from "Histos", which is Greek word means pole/g (chart-graph). The quality of image encryption is indicated by the Histogram, flat graph means ideal encrypted image and efficient encryption algorithm. Therefore, the Histogram of encrypted image can be considered one of the most methods to clear the quality of image encryption technique.
Histogram graph provides the comparative number of occurrences of each pixel through an image and considered as one of the most important methods for the evaluation of the quality of an image. Authentic image and marked image (output of embedding process) with their histograms for every transform domain with two different images (Cameraman and Reactor design) are shown in Figs. 6, 7, 8, 9, 10 and 11, respectively. In this research paper, the Histogram is utilized for ensuring the invisibility/ not observable mark of the presented algorithm.  The results indicate that when the proposed algorithm is implemented using DCT, the two images, i.e. authentic and marked, appear to be the same and their histograms are approximately the same. Thus the results verifies that utilizing DCT produces the most acceptable results related to imperceptibility and image histogram analysis.

Correlation (C r )
It is a measure that is utilized to evaluate the degree of closeness between the two functions, it can be used to determine the extent to which two images are remained close to each other. The correlation ratio gives a direct measure of the efficiency of the proposed algorithm. The closer the correlation ratio to unity, the efficient the algorithm [48].

Mean square error (MSE)
MSE is the median measure of the squares of the difference between the intensities of two examined images. It can be mathematically represented by Eq. (10).
where f (i, j) and f'(i, j) are assigned for the authentic and marked images respectively. The higher the value of MSE, the lower the quality of the image and vice versa [34].

Peak signal-to-noise ratio (PSNR)
The PSNR is an image quality metric used for differentiating between two examined images. It can be formulated mathematically in Eq. (11). A higher value of PSNR is an optimum case [34].
The PSNR, MSE, and correlation are computed for the three utilized transform domains with two different images (Cameraman and Reactor) and are utilized for testing the presented technique for contents verifications through the computer experiments.  The results of these initial experiments are listed in Table 2 and the images samples with the Histogram of the original and marked images have been shown in Figs. 6, 7, 8, 9, 10 and 11. Higher PSNR, lower MSE, and higher correlation ratio are acquired in case of realizing the proposed system using DCT, which the replaced rows and columns through the algorithm procedure are selected to be the eighth row, and eighth column (k = 8) for each block of size (8 × 8).
As shown in the images samples of the results in Figs. 6 and 7 for cameraman and reactor, respectively. The marking of contents is not visible, the original and marked image appear as the same. Also, there is no difference in the Histogram of the authentic and marked images. Figures 6 and 7 show the results of image verification based on DCT transform. Figures 8, 9, 10 and 11 give the results of the algorithm based on DFT and DWT transform techniques.
The Original authentic image and marked cameraman image are given in Fig. 6a and b, respectively. As cleared form these images, the mark don't affects on the image and it is not observable by human eye. Figure 5c and d show the Histogram of original and marked cameraman image, there is little difference between them.
DCT-based algorithm achieves better than other scenarios. The Original authentic image and marked Reactor image are given in Fig. 7a and b, respectively. As cleared form these images, the mark don't affects on the image and it is not observable by human eye. Figure 7c and d show the Histogram of original and marked cameraman image, there is little difference between them.
The image content verification algorithm based on DFT and DWT testing experiment have been given in Figs. 8, 9, 10 and 11. As shown in Fig. 8a  The previous results of the different computer experiments reveal the suitability and applicability of the DCT-based image content verification algorithm more than the other data transform techniques due to invisibility of the mark. Figures 10 and 11 show results of DWT-based algorithm, it is cleared the difference in Histogram of original authentic and marked images. The results of the previous experiments have been tabulated in Table 2.
In the previous results samples, the block size is 8 × 8, it is chosen for testing the performance of different data transform techniques. In the next section, the performance of the various block size has been studied. As shown in the computer simulation experiment, the DCT transform performs better than the other techniques with the presented scenario of the simulation parameters as shown in Table 2.

Evaluation of embedding process
Assuming that no attacks tampered the marked image and the transmission medium is of free errors. The extraction process is implemented using DCT domain only as an efficient transform to be used with the proposed algorithm which achieved best performance results through embedding process. Figures 12 and 13 show the authentic image, extracted image, and their histograms for two different images. In case of no modification, the authentic image is extracted successfully, and image integrity is verified.
The image quality metrics are listed in Table 3. These metrics assure the acceptable efficiency of the proposed verification algorithm. Every block is exposed to DCT and its reverse with their standardized quantization, and at every time some error is introduced during quantization process. Therefore, a tiny variation in the correlation ratio between the authentic and obtained images is presented. For more evaluation of the proposed image integrity verification algorithm, its performance is evaluated at different block sizes. The algorithm behavior at block sizes of 4 × 4 and 16 × 16 is compared with its previously discussed behavior at 8 × 8 block size. The performance metrics (MSE, PSNR, and Cr) are measured and compared during embedding process between the authentic and marked images for the three assigned block sizes and introduced in Table 4. Also, these metrics are measured during the verification process between authentic and verified extracted image in case of no image manipulations and listed in Table 5. The results reveal that the best performance of the proposed system is achieved with a block of size 16 × 16. Hence, in the rest evaluating simulation experiments the size block is 16 × 16.
To examine the practicality of the proposed image integrity verification algorithm, it applied on different images with different visual properties. Table 6 shows the extracted verified image and related quality metrics between the authentic and extracted verified images which are measured individually for every image sample. The results reveal the sufficient applicability of the proposed algorithm to verify image integrity.
As shown in Figs. 14 and 15, the marked image appears as the original image. These figures show the various versions of the processed images. Firstly, Fig. 14 shows the results of standard Matlab (Cameraman image) for testing the visibility and delectability of the algorithm. Figure 14b gives the verified/extracted standard Cameraman image. The experiment repeated using the reactor image as not standard image. Figure 15a and b give the embedding hidden mark and verification processes images samples, respectively. As shown, in Fig. 15b, extracted/verified image appears like/same original image.

Advantages and applications
The following sections describe experimentally the advantages and some different applications of the proposed image content integrity verification approach.

Modification detection sensitivity
The detection sensitivity is considered a major aspect related to the integrity verification algorithm evaluation, which describes the ability of the algorithm to detect any small image manipulation even it was visibly undetectable. To achieve a comprehensive assessment of the proposed algorithm performance, the detection sensitivity analysis is built upon two aspects; a) Visual detection, which depends on the visual properties of the verification algorithm output image. b) Statistical analysis, which depends on some measured parameters between the input and the output of the verification process.
The detection sensitivity analysis is focused on measuring the performance of the proposed algorithm to sense the visually undetectable manipulation attacks to highlight its importance.
Because the most image manipulations which have a forensic nature are intended to be visually undetectable and the image receiver cannot visually detect the forensic act that is usually having a severe impact on the future or tampered image dependable issues. Visual detection is a method used to detect the image integrity violation, depending on the visual properties of integrity verification process for the output image. The quality of the output image is inversely proportional to the amount of modification occurred. The proposed scheme performance is studied in the case of two categories of manipulations; a) Large scale and visually undetectable manipulation attacks, including; image resizing attack, image blurring attack, Gaussian noise, and salt & pepper noise. b) Small scale and visually un-detectable manipulation attacks, including copy-move forgery, and sub-image blurring attack.

I. Large Scale and Visually Undetectable Image Manipulation Attacks
Large scale image manipulation attacks are those manipulations that affect a large relative area of an intended image. These modifications may be mainly for forgery intent, such an image resize, image blurring, image cropping, and another geometrical image processing. On the other hand, there is an unintentionally image modifications such different noises that affect transmitted image like Gaussian noise, and salt & pepper noise. In this context, an image   Figure 16 shows the different types of scale-based image manipulation attacks [39,53]. Figures 17 and 18 show the comparison between the received noisy image and the obtained verified image in the case of Gaussian white noise of mean (0) and variance value (0.00001). In addition, these images are examined in the case of salt-and-pepper noise with rate (0.00005) and are presented in Figs. 19 and 20 respectively. In relation to these two noise types, the results indicate that the proposed model is sensitive to even any small modifications in the image. While the received noisy image and the extracted image appear to be visibly the same. The extracted image (the output of the proposed model) demonstrates a major change compared to the manipulated marked image.
Also, image resizing modification is examined, which a marked image of size (256 × 256) is converted to an image with size (300 × 300) that appears visually unchanged. The received resized image and the output image of the verification process are shown in Figs. 21 and 22, respectively.
The algorithm is proved to have a high sensitivity to detect this type of image manipulation attacks, although the received manipulated image appears to be visually unchanged. The image blurring attack is one of the common sensitive image attacks, which intentionally change the clarity of a whole or sub-image detail. This results in difficult visual extraction of valuable information hidden in blurred areas of a sensitive image that affects the dependable future operations and decisions. It is usually tampering images related to nuclear and medical applications which fine Here, the marked image is blurred using a circular averaging filter with radius 2 to make a tampering act undetectable. Blurred marked image, extracted verified image are compared in case of blurring attack and shown in Figs. 23 and 24, respectively. The results also reveal the high sensitivity of the proposed approach to detect this type of image tampering, although the received tampered image appears to be visually not manipulated.

II. Small Scale and Visually Undetectable Image Manipulation Attacks
Small scale imaging manipulation attacks are those manipulations that affect a small relative area of an intended image. In almost cases, these attacks are mainly intended to a temper, act and usually the tampering cannot be visually detected. This work will focus on studying the capability of the proposed algorithm to identify a copy-move forgery, sub- & Copy-move forgery (Tampered Regions Detection) Copy-move forgery is a familiar type of image forgery, in which a dedicated portion of an image is intentionally reproduced and moved to a new part of the same image to change its originality. As an example; replacing included personals faces with others or changing their positions. The proposed algorithm can be also used to detect the tampered areas, which appears clearly in the following results. The marked image is tempered by replacing two small image regions (number of blocks with another number of blocks) which have the same appearance and visible feature. The tampered marked image is intended to appear visibly not  Fig. 25 for reactor design and cameraman images. By processing the tampered image using the verification algorithm, the tampering is detected, in which a forged copy-moving indication appeared visually in the upper half of the extracted image as shown in Fig. 26.

& Sub-image Blurring Attack
A sub-image blurring attack is a type of image manipulations which a small region of the image (small number of blocks) is blurred intentionally to hide an included valuable detail, so the blurred image receiver cannot detect a tampering act. Here a small region of the marked image is blurred using circular averaging filter with radius 10 which the blurring act cannot be detected as shown in Fig. 27. Also; the verification algorithm can detect the tampering in which a forger blurring indication appeared in the upper half of the extracted image as shown in Fig. 28. This type of image manipulations is discussed to examine the detection sensitivity of the proposed approach when only one-pixel value of marked image is changed intentionally in an imperceptible manner. So, the received manipulated marked image appears to be not tampered as shown in Fig. 29. The output of the verification algorithm in Fig. 30 shows a visual change in the lower half of the extracted image as an indication to one-pixel change manipulation which confirms higher detection sensitivity of the integrity verification algorithm.
In [18], color image steganography is investigated, authors in this paper presented channel dependent payload partition strategy steganography based on amplifying channel modification probabilities. For improving the empirical steganographic security against the channel co- occurrences detection, the embedding capacity among RGB channels the embedding impacts could be clustered. The experiments of the proposed steganography performance evaluation revealed that its robust and against the modern steganalysis of color images [36].
Data hiding research field importance and its role in the emerging technologies have been discussed in [32]. Authors in [32] presented efficient steganography techniques based on Image Texture Features utilizing adaptive payload distribution in multiple image steganography. Performance of the proposed multiple images steganography has compared to the ordinary single image steganography technique against the modern single image steganalysis, as cleared form the experiments in [31,38], the payload distribution based steganography performs better against security threats. An efficient Forensic detection technique based on CNN has been proposed in [30].
In the future research plan, the integrity medical data wireless transmission [15] will be considered for enhancing the security Wireless Body Area Networks (WBANs). Also, the encrypted image verification will be presented for double the security of the sensitive images and any secret/authentic documents. The open environment applications require highly secured link for providing reliable transmission media, the encrypted/marked images transmission over the noisy medium will be studied and analyzed with considering the error control schemes utilizing. Also, in the future plan, the complexity of the presented approach will be analyzed with applying another efficient transform technique for decreasing the complexity.

Conclusion
Hidden marked algorithm has been presented in this paper to achieve image integrity verification. This DCT based algorithm detects any tiny tampering or forgery in the secret/authentic images. DCT proved to be the most applicable transform to be exploited according to imperceptibility and performance analysis. Both visual and statistical analysis results revealed the higher sensitivity of the approach to detect any image manipulation even they were visually undetectable. So, it can be utilized as an efficient image forgery detection tool for a high confidential data such as nuclear and military applications. In the future work, we are motivated to incorporate the application of a self-multi-level secure scheme with the use of different encryption, steganography and watermarking techniques for vigorous and dependable image communication. Also, the future research plan will include the deep learning-based security techniques for image transmission.  Data availability Author Confirms the data availability which is related to this manuscript in case the requirements for research targets.

Declarations
Conflict of interest Authors declare wish to confirm that there are no known conflicts of interest associated with this publication and there has been no significant financial support for this work that could have influenced its outcome.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .