Next Article in Journal
Learning Pathways and Students Performance: A Dynamic Complex System
Previous Article in Journal
Can the Sci-Tech Innovation Increase the China’s Green Brands Value?—Evidence from Threshold Effect and Spatial Dubin Model
Previous Article in Special Issue
Low Computational Coding-Efficient Distributed Video Coding: Adding a Decision Mode to Limit Channel Coding Load
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Anti-Screenshot Watermarking Algorithm for Archival Image Based on Deep Learning Model

1
School of Computer Science and Technology, Anhui University, Hefei 230039, China
2
Department of Computer Science, University of Warwick, Coventry CV47AL, UK
3
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(2), 288; https://doi.org/10.3390/e25020288
Submission received: 8 January 2023 / Revised: 30 January 2023 / Accepted: 31 January 2023 / Published: 3 February 2023

Abstract

:
Over recent years, there are an increasing number of incidents in which archival images have been ripped. Leak tracking is one of the key problems for anti-screenshot digital watermarking of archival images. Most of the existing algorithms suffer from low detection rate of watermark, because the archival images have a single texture. In this paper, we propose an anti-screenshot watermarking algorithm for archival images based on Deep Learning Model (DLM). At present, screenshot image watermarking algorithms based on DLM can resist screenshot attacks. However, if these algorithms are applied on archival images, the bit error rate (BER) of the image watermark will increase dramatically. Archival images are ubiquitous, so in order to improve the robustness of archival image anti-screenshot, we propose a screenshot DLM “ScreenNet”. It aims to enhance the background and enrich the texture with style transfer. Firstly, a preprocessing process based on style transfer is added before the insertion of an archival image into the encoder to reduce the influence of the screenshot process of the cover image. Secondly, the ripped images are usually moiréd, so we generate a database of ripped archival images with moiréd by means of moiréd networks. Finally, the watermark information is encoded/decoded through the improved ScreenNet model using the ripped archive database as the noise layer. The experiments prove that the proposed algorithm is able to resist anti-screenshot attacks and achieves the ability to detect watermark information to leak the trace of ripped images.

1. Introduction

Archival images have become one of the most important documentary references in the history of modern social development. It is the valuable historical image directly formed by state institutions, social organizations and individuals during their social activities. The archival images play an unique and valuable role for both individuals and organizations. It is an information link between the past and the future that perpetuates human memory. By recording people’s experiences, their development history can be preserved intact. Archival images play an important role in the history of human development, while actively promoting the advancement of individuals. The institutions that provide the service of keeping archival images are the local talent market, each district and county talent market and street office, etc. According to laws and regulations, archival images belong to the national statutory, mandatory and vested public information. No individual may retain and destroy these images.
Moreover, archival images record the main experience, political and ideological style of each person and serve as a proof, basis and reference. For the employer, the Human Resources (HR) department can use archival images to evaluate employees and select talents. It provides an important basis for the salary setting and employee promotion of the employer. The individual appraisal, reward and punishment of the work unit are definitely put into the archival images. Therefore, personnel management can work better in a uniform way. Some workplaces are required to provide archival images for recruitment. This can be a serious problem if the copyright of archival images is not well-protected.
Image watermarking algorithms should have strong robustness because the watermarked image may be attacked in the transmission process. Common attacks include cropping, compression, Gaussian noise, etc., all of which may change the pixel value of the watermark image, and then the watermark information cannot be extracted smoothly. Screenshot attack is a very complex image attack that includes perspective deformation and a moiré pattern. In this paper, the noise between the screen and the camera exists commonly. However, the traditional robust watermarking algorithms show low robustness, so it is necessary to introduce a DLM to improve the performance of the algorithm.
Recently, DLMs such as Stegastamp [1] have been widely applied in computer vision, natural language processing and other fields. DLMs have achieved excellent results in experiments. Therefore, researchers also explore image watermarking algorithm based on DLM. Zhu et al. [2] initiated the HiDDeN model, which introduces a variety of noises between the encoder and decoder to improve the robustness of the auto-encoder. However, since the noise layer only considers Gaussian blur, cropping, JPEG compression and other attacks, it cannot effectively resist the problems of a moiré attack, perspective deformation and color transformation. In order to simulate screenshot attacks in real scenes, Wengwowski et al. [3] proposed the light field messaging (LFM) system. LFM specially trains a network to simulate the effect of cover image to screenshot process, so a dataset of 1,000,000 images is established. However, the dataset takes up a lot of memory and it is too difficult to obtain. Moreover, LFM has a good effect on screenshot attacks, and it has limitations for print attacks.
Many researchers apply the DLM directly as the encoder or decoder. Fang et al. [4] generated several centrosymmetric Gaussian noises on the red channel of the color image to determine the embedding position of the watermark information. They trained a decoder with DLM to extract the watermark on the blue channel of the color image. To resist the screenshot attack, Fang et al. [4] designed an enhancing subnet to improve the accuracy of watermark extraction. In a subsequent work, they applied the anti-screenshot DLM to dynamic images [5]. This algorithm embeds watermark information in the red or blue channel of the image in turn. Due to the frequent switching of the image in a short time, the color interchanges make the watermark information undetectable to the human eye. Fang et al. [5] proposed DLM-based decoder with attention mechanism to improve the accuracy of the model. Consequently, their algorithm has significantly improved in imperceptibility and robustness compared with [4].
DLM can be used not only to embed or extract watermark information directly, it can also be used to assist in image watermark extraction. Li et al. [6] used DLM for finding feature points that can resist screenshot attacks. The watermark information is embedded in the region around the feature points, so feature point localization has a very important position in watermark extraction. DLM can find the feature points that contain the image watermark in the case of screenshot attack on the image. Additionally, Li et al. [7] added an anti-screenshot noise layer to the end-to-end DLM. The noise layer is presented between the encoder and decoder and used to simulate possible attacks on watermarked images. This increases the training time of the DLM to some extent, but the robustness of the anti-screenshot watermarking algorithm is significantly improved. This ensures that when a screenshot is taken of a watermarked image, the decoder can still extract the watermark information accurately.
Matthew et al. [1] proposed the Stegastamp model, in which the differentiable image perturbations were applied between the encoder and decoder to simulate the distortion caused by print attacks and screenshot attacks. The excellent watermarking algorithm is robust to both screen attack and print attack, but the Stegastamp model is low robustness against the cropping attack. Therefore, the image cropping attack has limitations, which leads to the poor quality of a dense image.
In this paper, we propose a robust archival image watermarking algorithm based on the Stegastamp model. The proposed algorithm structure includes five parts: the preprocessing network adds color and texture into the archival image. The encoder embeds the watermark information into the cover image successfully. It also minimizes the difference between the watermarked image and the cover image, and finally generates the watermarked image with good visual quality. The noise layer of Stegastamp is designed to distribute the watermarked image generated by the encoder to simulate the effect of screenshot attacks. The purpose of the moiré network is to further add real moiré to the distributed watermarked image. The decoder is to extract watermark information from the moiré-watermarked image. The main contributions of this paper are as follows:
  • A style transfer network is applied to improve the richness of background texture for archival images. So, the robustness of the image watermarking algorithm can be maintained.
  • A moiré network is constructed to produce the moiré archival image dataset. In addition, moiré images taken in realistic environments are added to build a rich database of screenshot images.
  • The anti-screenshot capability of the Stegastamp model is improved by placing these datasets in the noise layer. It solves the problem of a lack of standard moiré dataset in digital watermarking field.

2. Related Works

In this section, we introduce the Stegastamp model and style transfer in details.

2.1. Stegastamp

Recently, plenty of learning-based algorithms leveraging encoder and decoder to embed different kinds of watermark in images have been proposed, such as the Stegastamp. The Stegastamp applies DLM to the screenshot watermarking algorithm by modeling more realistic distortions of watermarked image. Firstly, the cover image and the expanded watermarked sequence are fed into the encoder. Then, the watermarked image output from the encoder goes through the printing and screenshot process. To improve the robustness of the algorithm, during the model training, watermarked image is added with some attacks by the noise layer, such as perspective warp, motion/defocus blur, color manipulation, noise and JPEG compression. The noise layer makes the watermarked image closer to the real application scenario. Therefore, the decoder can robustly decode the watermarked image even during the screenshot process. The decoder locates, corrects and extracts the watermarked image in the captured image to form the corrected image. Finally, the decoder extracts the watermark information from the corrected image.
Stegastamp mainly improves the robustness of the image watermarking algorithm against moiré attacks, which are interference streaks that appear on surface of the photographed object. For example, during screenshot, the image displayed on the shooting screen will appear as a moiré pattern.
However, the Stegastamp model is vulnerable to visible noise on the watermarked image during the encoding process, which very unfavorable for the archival images with a single color background, so that it reduces the robustness and imperceptibility of the watermarking algorithm. Some researchers have refined Stegastamp in the past two years; however, the performance of the model is still not well suited for archival images [8,9]. Therefore, the performance of the StegaStamp model for archival images needs to be improved.

2.2. Style Transfer

Style transfer is to remain the content of one image unchanged, but its style is changed to another image. Style transfer is often applied to simulate the creative style of artists who have passed away. Due to the excellent performance of DLM, it has been used to accomplish research work on style transfer [10,11,12,13]. For example, Chen et al. [10] proposed a DLM for cartoon style transfer (CartoonGAN). This Generative Adversarial Network (GAN) is well able to migrate the style of images to cartoon style. Li et al. [11] proposed CariGAN to generate caricatures. Their model ensures that the face of the cartoon characters does not show unsightly deformations and generates more details in key areas. Style transfer is not limited to image processing; it can also be applied to automatically generate fonts. Zhao et al. [12] proposed CycleGAN for both text and image. They introduced a new loss function to train the DLM, which better guarantees the style of the synthesized images. In order to apply style transfer in real scenarios, Zhang et al. [13] proposed CSST-Net for tile watermarking algorithm, which has a good performance in watermarking capacity and robustness in addition to good style transfer effect.
Archival images include a large number of white areas, so they are very unsuitable for embedding watermark information directly. Therefore, in this paper, we propose to apply the VGG [14] model to perform style transfer on archival images to improve the imperceptibility of the watermarking algorithm.

3. Proposed Algorithm

The flow diagram of the proposed robust watermarking algorithm is shown in Figure 1. Firstly, the cover image is preprocessed. The watermark information with size 100 bits is fully connected to obtain the tensor. Then, the tensor is upsampled to make the size consistent with the image size after preprocessing. After a series of convolution operations, the watermark is embedded into watermarked image. Then, the noise layer simulates the screenshot attack in a real scene with watermarked image. After the watermarked image with noise is attacked by real moiré, the decoder extracts the watermark information through a series of convolution operations and fully connects.

3.1. Archival Image Preprocessing

Style transfer allows us to modify the style of an image to another style without changing the content of the image. This is great for archival images where content needs to be protected. More importantly, style transfer makes the background of the archive image more suitable for embedding watermark information. An archival image has a large range of single background. However, the watermarking in a single background will be very conspicuous, which leads to terrible algorithmic imperceptibility. Therefore, the archival image needs to be preprocessed before being inserted into the model. In this work, we apply the style transfer algorithm to preprocess the archival image. This causes the archival image to add different colors and textures, which is more conducive to embedding an image watermark, improving both the imperceptibility and robustness of the proposed algorithm. The structure of the preprocessed archival image is shown in Figure 2.
The preprocessing model applies the trained VGG network [14] as the encoder to extract features. The structures of the transfer encoder and loss encoder are identical. These encoders cover and output the features of different levels of convolution layer by jumping connection. Then, adaptive instance normalization (AdaIN) is applied to adjust the first-order and second-order statistics of features. The spatial mask automatically adjusts the level of stylization. The structure of the decoder is basically symmetrical to the encoder, and it needs to be trained from the beginning. The discriminator is used to predict the style types of the image and distinguish the authenticity of the transfer image at the same time.

3.2. Encoder Construction

Encoder’s architecture is similar to the U-net [15]. The encoder is mainly applied to embed the watermark information into the transfer image and minimize the difference between the watermarked image and the transfer image. It concatenates the preprocessed cover image and watermark information to form a tensor. To balance the watermark capacity and the image quality of the generated watermark-containing images, a compromise is adopted to encode 100 bits of watermark information uniformly on a 400 × 400 × 3 carrier image. Figure 3 shows the structure of Encoder.
Before inputting the carrier image and watermark information into the encoder, some pre-processing is needed. First, the 100-bit watermark information is fully concatenated to obtain a 7500-dimensional tensor, and the dimension is modified to obtain a 50 × 50 × 3 tensor, and then the tensor is upsampled to make the dimension of the tensor consistent with that of the carrier image to obtain a 400 × 400 × 3 tensor. The tensor is then cascaded with the carrier image to obtain a 400 × 400 × 6 tensor. The tensor is fed into the encoder for a series of convolution operations. During the convolution process, the high level feature map is upsampled. So, the size of the upsampled high level feature map and its corresponding low level feature map are the same. Finally, the high-level feature map and the low-level feature map are cascaded (concatenate) by channel dimension, so that the low level features can be preserved.
The tensor undergoes a series of convolution, pooling, and downsampling operations, and then combines the previously extracted features for upsampling. Conv_1~Conv_10 represent the convolution operations, using 3 × 3 convolution kernels with the field of perception, the step size of Conv_1 and Conv_6~Conv_10 are set to 1, the step size of other convolutions is set to 2, and the step size of Conv_11 is set to 1. Up_6~Up_9 represent the upsampling operations, using 2 × 2 convolution kernels with the step size set to 1. The activation functions are all ReLU functions, the fill type is “same”, and the initializer of the convolution kernel is he_normal. The cascade operation is performed in the channel dimension. Finally, when the loss function reaches the standard, the encoder generates the watermarked image with good visual quality.

3.3. Noise Layer

The noise layer is consistent with the Stegastamp cover noise layer. Different from the design of HiDDeN model noise layer, the noise layer of HiDDeN mainly contains six different types of noise attacks, such as cropping, Gaussian noise, JPEG compression, etc. No suitable differentiable function has been designed to simulate the attack types generated in the screenshot process. Stegastamp’s noise layer is designed to resist the attacks generated in the screenshot process and printing process. In this work, we regard these attacks as the superposition of a series of attacks, including perspective deformation, motion and defocusing blur, color transformation, Gaussian noise and JPEG compression.

3.4. Moiré Network

In this paper, we first use the standard dataset in the field of moiré removal, which was proposed by Sun et al. [16]. This dataset contains 100,000 pairs of cover images and moiré images. We select 6000 pairs of images as the training set. On this basis, in order to improve the robustness of the watermarking algorithm in the archival image, we add some archival images into the dataset. The dataset image is shown in Figure 4. The shooting distance and angle are shown in Figure 5a. The distance of the phone from the monitor is about 20~30 cm, and the image obtained from the shooting needs to be further corrected for processing. Finally, 2000 pairs of moiré images are obtained as needed. Figure 5b shows the correct images. In order to improve the generalization ability of moiré network and to prevent the phenomenon of overfitting. Furthermore, the moiré network can accurately approximate the nonlinear relationship between the original image and the moiré image, and it saves the time cost of producing the dataset. The 2000 pairs of images we collected are augmented. Due to the lack of a standard moiré dataset, referring to paper [4,5], we decided to perform the geometric transformation on the existing images. The images are horizontally flipped, vertically flipped and scaled to expand the number of datasets. In the end, we obtained 6000 pairs of self-built moiré dataset.
In this paper, we propose a novel moiré network to improve the robustness of the algorithm by adding realistic moiré disturbance to the watermarked image. Unlike the cover Stegastamp, the noise layer use differentiable functions to simulate screenshot attacks. The moiré network does not apply the established functions to learn attacks, but it applies the learning characteristics of Convolutional Neural Network (CNN) to learn the moiré attacks in realistic scenes.
Figure 6 shows the network structure of the moiré network based on the U-net. The differently colored arrows in the figure represent different operations. Additionally, the two orange rectangles represent that they will be connected. The network architecture consists of three main parts: the encoding stage on the left, the decoding stage on the right and the hop connection. The encoding stage is used to down sample the cover image and extract high-level semantic information. The convolution in the encoding stage is a 3 × 3 convolution kernel with a step size of 1. The size of the convolution kernel for upsampling is 2 × 2. The decoding stage is used to decode high-level semantic features and generate the final moiré image after the moiré attack. Finally, the decoding stage changes the features into 3 channels by 1 × 1 convolution to obtain the noisy image after the moiré attack. The convolutional activation function is the Sigmoid function. All the other convolutions use 3 × 3 convolutional kernels and ReLU function. All padding methods in the network are of type SAME. The initializer of the convolution kernel is he_normal. Skip connection is applied to fuse the features of the two stages to make the result more accurate. Due to the weak of moiré in this dataset, we use iPhone 8, Hongmi K30 and Huawei hi nova9 mobile phones to collect the moiré images. Five hundred images are randomly selected from Div2k dataset [17] and displayed on the screen for shooting. The display used is HP N246v. The distance and angle of the mobile phone shooting are shown in Figure 7.
The distance between the mobile phone and the monitor is about 20~30 cm. An example of the image pair of the self-built moiré dataset is shown in Figure 7. In order to improve the generalization ability of moiré network, the collected 2000 pairs of images are augmented [18].

3.5. Decoder Construction

The decoder is used to extract the watermark from the moiré image. Some geometric attacks such as translation, scaling and rotation may be encountered in the transmission of moiré images. In order to balance the robustness of these geometric attacks and the watermark extraction accuracy, some preprocessing is needed before watermark extraction from the moiré image. Therefore, a spatial transformation network (STN) is used in the decoder, which can learn in an end-to-end manner without changing the loss function and improve the performance of the model. The final watermark information is obtained by Sigmoid function. The size of the convolution kernels of Conv_1~Conv_7 is 3 × 3. The activation function is ReLU. The filling method is same. The step size of Conv_1, Conv_3, Conv_5, Conv_6 and Conv_7 is 2, and the step size of other convolution operations is 1. The flatten operation is performed after Conv_7 to reduce the tensor dimension to 1 dimension. Then, the decoder performs two full concatenation operations to achieve the goal of successfully extracting the watermark information. The structure of the decoder is shown in Figure 8.

3.6. Model Training

DLM needs to select the appropriate loss function to guide network training. The loss function carries on the back propagation through the error between the predicted result and the ground truth. It also guides the learning of network parameters. In this paper, we choose different loss functions for different sub-networks. The mean square error (MSE) loss function and Learned Perceptual Image Patch Similarity (LPIPS) sensing loss function [19,20] are applied in the encoder, which are recorded as L R 1 and L P . MSE loss function and LPIPS perceived loss function are shown in Equations (1) and (2), respectively.
L R 1 ( I Cover , I Stego ) = 1 C × H × W I Cover I Stego 2 2
L P ( I Cover , I Stego ) = l 1 H l W l h , w w l × ( y ^ l y ^ 0 l ) 2 2
where I Cover is the cover image, and I Stego is the watermarked image. 2 denotes summing the squares of each element and then calculating the square root of the sum. The l feature maps are extracted from the convolutional layer and normalized in channel dimension y ^ l R H l × W l × C l and y ^ 0 l R H l × W l × C l , respectively. We compute distance L 2 by scaling activation in the channel dimension by W l R C l . Finally, the average distance in space and the sum in channel is calculated.
The loss function L R 2 of moiré network also uses MSE as shown in Equation (3)
L R 2 ( I Cover , I Screen ) = 1 C × H × W I Cover I Screen 2 2
where I Screen is the watermarked image after the moiré attack.
The decoder uses the cross entropy loss function L M [21], which is widely applied in classification tasks. The formula for L M is shown in Equation (4).
L M ( M , M ) = 1 N i = 1 N [ M i log ( M i ) + ( 1 M i ) log ( 1 M i ) ]
where M is the original watermark, M is the extracted watermark, and N is the length of image watermark. The total loss function L of the model is shown in Equation (5)
L = λ R 1 L R 1 + λ P L P + λ R 2 L R 2 + λ M L M
where λ R 1 , λ P , λ R 2 and λ M are the parameters of corresponding loss function. Finally, the encoder, the moiré network and the decoder are trained against each other, which make the model iterates continuously until convergence [22].

4. Experimental Results

4.1. Experimental Environment and Settings

The MIRFLICKR [23] dataset is used to evaluate the performance of the proposed algorithm ScreenNet. In the experiments, 20,000 images are selected for training, and the image size is adjusted to 400 × 400 pixels. The graphics card used is NVIDIA RTX 2080ti GPU training. The code is written in tensorflow framework. From the MIRFLICKR dataset, 2000 images which have no intersection with the training set are selected to form the verification set to evaluate the robustness under common noise attacks.
Watermarking information is a random sequence composed of 0 s and 1 s. Through experiments on different capacities of watermarking bits, the final choice is to embed 100 bits of watermarking information in the cover image. The distance similarity measures PSNR and SSIM are used to measure the quality of watermarked images. From Table 1, if the watermarking information bits increase, the value of PSNR and SSIM decrease. Consequently, the image quality will be poor, whereas the watermark capacity will be improved. Based on the comprehensive consideration of watermark capacity and image quality, 100 bits of watermarking information is selected to embed into the cover image. In this paper, the accuracy of extracting watermark information is used to evaluate the robustness of the proposed algorithm against a screenshot attack. The formula of the accuracy is shown in Equation (6).
A c c u r a c y = 1 - B E R

4.2. Performance Comparison before and after Preprocessing

We compare the performance of the proposed ScreenNet algorithm before and after the preprocessing of the archival images. Figure 9 shows some cover archival images and the corresponding preprocessed archival images. We can notice that the text and color of the archival images can be identified from the preprocessed archival image. However, the archival images after preprocessing are similar to the archival images stored in reality.

4.3. Network Training Speed Comparison

We propose a novel ScreenNet algorithm based on a moiré network. Due to the addition of the moiré network in Stegastamp, the overall network architecture is more complex. So, the ScreenNet algorithm has a slower convergence speed compared to Stegastamp. Table 2 shows the training speed comparison between Stegastamp and ScreenNet. The training time of ScreenNet is about 2.3 h longer than Stegastamp. In the follow-up work, we need to standardize the data and select the appropriate learning rate to improve the training speed.

4.4. Network Training Speed Comparison

In this section, five common noise attacks are applied to evaluate the robustness of the proposed ScreenNet model. Figure 10 shows some visual images of watermarked images, noisy images and difference images. From Figure 10, the first row represents the watermarked images obtained by the proposed ScreenNet model. The second row shows the noisy images after the watermarked image is attacked. The third row represents the difference images between the watermarked image and the noisy image. The residual diagram I r e between watermarked image and noisy images can be calculated from Equation (7) as follows:
I r e = | I e n I n o |
where I e n denotes the watermarked image and I n o denotes the noisy image after the attack.
As shown in Table 3, five common noise attacks used in hidden are selected as the basis for evaluating the robustness, including Dropout (0.3), Cropout (0.3), Crop (0.035), Gaussian (2) and JPEG Compression (50). The numbers between brackets represent the noise intensity. From Table 3, we can notice that the accuracy of the proposed ScreenNet model is robust against Cropout, Dropout, Gaussian and JPEG compression. However, for crop attacks, the accuracy is only about 64%. The main reason is that the cover noise layer of Stegastamp does not consider the situation of crop attack in the training process, so the differentiable function is not applied to simulate this process. As a result, the watermark information is almost covered in the whole cover image, and a large amount of watermark information will be lost when the cropping attack is carried out.

4.5. Robustness Comparison of Screenshot at Different Distances

We evaluate the performance of the proposed ScreenNet model compared with HiDDeN and Stegastamp against screenshot attacks using different distances. Table 4 shows the anti-screenshot experiment.
The experimental results show that the HiDDeN model has poor robustness to screenshot attacks, and the accuracy of the watermark information extraction is in the range of 66~80%. This is because the noise layer of HiDDeN model only considers the common attacks such as cropping, JPEG compression, Gaussian blur, etc. However, it does not consider the screenshot attack. The Stegastamp model maintains a good watermark extraction accuracy under four different distances, and the accuracy rate is above 92%. This is mainly due to the application of various differentiable perturbations in the noise layer of Stegastamp to approach the screenshot attack. Compared with Stegastamp, the proposed ScreenNet model is more robust against screenshots. The proposed ScreenNet model also uses archival image dataset of moiré to add real disturbances. Therefore, the proposed ScreenNet has stronger antiscreenshot robustness. The average accuracy of watermark extraction with ScreenNet is higher than that of Stegastamp and HiDDeN models. Figure 11a shows the accuracy of screenshot attacks against different distances and angles. Furthermore, Figure 11a intuitively shows that the proposed ScreenNet model has a stronger robustness against screenshot under different distances.

4.6. Robustness Comparison of Screen Shots at Different Angles

In this experiment, we selected different perspective angles to take screenshots for the watermarked images. The relevant experimental data are shown in Table 5, and the graphical accuracy is shown in Figure 11b.
The experimental results show that the HiDDeN model has poor robustness to screenshot attacks under different perspective angles. The accuracy rate of the watermark information extraction is less than 60% for HiDDeN. The Stegastamp model also maintains good watermark extraction accuracy under different perspective angles, and its extraction accuracy of watermark information reaches about 80% of ScreenNet. Under different perspective angles, the accuracy of watermark extraction for the proposed ScreenNet model is slightly higher than the accuracy of Stegastamp and HiDDeN models. As ScreenNet applies the self-built moiré with more obvious moiré traces, the accuracy of watermark information extraction is improved. It can be noticed that the proposed ScreenNet algorithm is more robust to screenshot attacks.

5. Conclusions

In this paper, a robust screenshot watermarking auto-encoder ScreenNet based on Stegastamp model is proposed. It is mainly composed of the encoder, Stegastamp cover noise layer, moiré network and decoder. Differently from the previous CNN-based anti-screenshot robust watermarking algorithms, we no longer use differentiable functions to simulate the moiré noise in real scenes. Instead, we train the constructed moiré network by removing the standard moiré dataset and the self-built moiré dataset, so as to add real moiré noise to the cover image. Due to the lack of a suitable moiré dataset, a collection of large amount of data is presented. Finally, the encoder, Stegastamp noise layer, moiré network and decoder are trained against each other, and a robust anti-screenshot watermarking algorithm based on U-net is finally realized. The experimental results show that the watermarked image generated by the model has a strong robustness against screenshot attacks at different distances and perspective angles. Moreover, it has a high watermarking information extraction accuracy, which can effectively solve some practical problems. Additionally, archival images have important contents and a white background, which have many similarities with document images. Therefore, the algorithm in this paper is also well suited for document images.
In the future, we are planning to add an attention mechanism to the DLM to improve the performance of the model. Furthermore, we are aiming to improve the training speed of DLM by a reasonable way, such as adjusting unnecessary convolutional layers or well preprocessing the dataset.

Author Contributions

Conceptualization, W.G. and L.L.; software, Y.F.; formal analysis, Y.B. and C.-C.C.; writing—original draft, Y.B. and W.G.; writing—review and editing, L.T. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by Science and Technology Projects of National Archives Administration of China (No. 2020-X-058) and National Natural Science Foundation of China (No. 62172132).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tancik, M.; Mildenhall, B.; Ng, R. Stegastamp: Invisible hyperlinks in physical photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition CVPR, Seattle, WA, USA, 13–19 June 2020; pp. 2117–2126. [Google Scholar]
  2. Zhu, J.; Kaplan, R.; Johnson, J.; Fei-Fei, L. HiDDeN: Hiding data with deep networks. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 682–697. [Google Scholar]
  3. Wengrowski, E.; Dana, K. Light field messaging with deep photographic steganography. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1515–1524. [Google Scholar]
  4. Fang, H.; Chen, D.; Huang, Q.; Zhang, J.; Ma, Z.; Zhang, W.; Yu, N. Deep template-based watermarking. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1436–1451. [Google Scholar] [CrossRef]
  5. Fang, H.; Chen, D.; Wang, F.; Ma, Z.; Liu, H.; Zhou, W.; Zhang, W.; Yu, N. TERA: Screen-to-Camera image code with transparency, efficiency, robustness and adaptability. IEEE Trans. Multimed. 2022, 24, 955–967. [Google Scholar] [CrossRef]
  6. Li, L.; Bai, R.; Zhang, S.; Chang, C.C.; Shi, M. Screen-Shooting resilient watermarking scheme via learned invariant keypoints and QT. Sensors 2021, 21, 6554. [Google Scholar] [CrossRef] [PubMed]
  7. Bai, R.; Li, L.; Zhang, S.; Lu, J.; Chang, C.C. SSDeN: Framework for screen-shooting resilient watermarking via deep networks in the frequency domain. Appl. Sci. 2022, 12, 9780. [Google Scholar] [CrossRef]
  8. Wang, K.; Li, L.; Luo, T.; Chang, C.C. Deep Neural Network Watermarking Based on Texture Analysis. In Artificial Intelligence and Security; Springer: Singapore, 2020; pp. 558–569. [Google Scholar]
  9. Niu, Y.; Zhang, J. An image steganography method based on texture perception. In Proceedings of the IEEE 2nd International Conference on Data Science and Computer Application, Dalian, China, 28–30 October 2022. [Google Scholar]
  10. Chen, Y.; Lai, Y.K.; Liu, Y.J. Cartoongan: Generative adversarial networks for photo cartoonization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9465–9474. [Google Scholar]
  11. Li, W.; Xiong, W.; Liao, H.; Huo, J.; Gao, Y.; Luo, J. CariGAN: Caricature generation through weakly paired adversarial learning. Neural Netw. 2020, 132, 66–74. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, Y.; Fu, G.; Wang, H.; Zhang, S. Image style migration based on improved CycleGAN. Sci. J. Intell. Syst. Res. 2021, 3, 36–42. [Google Scholar]
  13. Zhang, S.; Su, S.; Li, L.; Lu, J.; Zhou, Q.; Chang, C.C. CSST-Net: An arbitrary image style transfer network of coverless steganography. Vis. Comput. 2021, 38, 2125–2137. [Google Scholar] [CrossRef]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  16. Sun, Y.; Yu, Y.; Wang, W. Moiré photo restoration using multiresolution convolutional neural networks. IEEE Trans. Image Process. 2018, 23, 4160–4172. [Google Scholar] [CrossRef]
  17. Agustsson, E.; Timofte, R. NTIRE 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
  18. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. AutoAugment: Learning augmentation policies from data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 113–123. [Google Scholar]
  19. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  20. Wen, Z.; Guan, J.; Zeng, T.; Li, Y. Residual network with detail perception loss for single image super-resolution. Comput. Vis. Image Underst. 2020, 199, 103007–103016. [Google Scholar] [CrossRef]
  21. Li, L.; Doroslovacki, M.; Loew, M.H. Approximating the gradient of cross-entropy loss function. IEEE Access 2020, 8, 111626–111635. [Google Scholar] [CrossRef]
  22. Li, X.W.; Yang, Y.J.; Zeng, W.; Bi, Y.L.; Xu, J.L.; Xu, G. Area-Preserving hierarchical NURBS surfaces computed by the optimal freeform transformation. Comput. Aided Des. 2022, 143, 103134. [Google Scholar] [CrossRef]
  23. Huiskes, M.J.; Lew, M.S. The MIR flickr retrieval evaluation. In Proceedings of the ACM International Conference on Multimedia Information Retrieval, Vancouver, BC, Canada, 30–31 October 2008; pp. 39–46. [Google Scholar]
Figure 1. Flow Diagram of the Proposed Robust Watermarking Algorithm.
Figure 1. Flow Diagram of the Proposed Robust Watermarking Algorithm.
Entropy 25 00288 g001
Figure 2. Archival image preprocessing based on VGG network.
Figure 2. Archival image preprocessing based on VGG network.
Entropy 25 00288 g002
Figure 3. Encoder network structure.
Figure 3. Encoder network structure.
Entropy 25 00288 g003
Figure 4. Example of our dataset image.
Figure 4. Example of our dataset image.
Entropy 25 00288 g004
Figure 5. Collection and correction of moiré image. (a) Screen image; (b) Correct image.
Figure 5. Collection and correction of moiré image. (a) Screen image; (b) Correct image.
Entropy 25 00288 g005
Figure 6. Structure of moiré network.
Figure 6. Structure of moiré network.
Entropy 25 00288 g006
Figure 7. Example of self-built moiré image dataset. (a) A color image selected from the Div2k dataset. (b) The archival image. (c,d) The corrected images after screenshot, respectively.
Figure 7. Example of self-built moiré image dataset. (a) A color image selected from the Div2k dataset. (b) The archival image. (c,d) The corrected images after screenshot, respectively.
Entropy 25 00288 g007
Figure 8. Decoder network structure.
Figure 8. Decoder network structure.
Entropy 25 00288 g008
Figure 9. Archival images before and after preprocessing. (a,d,g,j) are archival images. (b,e,h,k) are four style images. (c,f,i,l) are their corresponding transfer images.
Figure 9. Archival images before and after preprocessing. (a,d,g,j) are archival images. (b,e,h,k) are four style images. (c,f,i,l) are their corresponding transfer images.
Entropy 25 00288 g009aEntropy 25 00288 g009b
Figure 10. Watermarked archives image after noise attack and its residual image.
Figure 10. Watermarked archives image after noise attack and its residual image.
Entropy 25 00288 g010
Figure 11. Screenshot attack accuracy against different distances and angles. (a) Distance; (b) Angle.
Figure 11. Screenshot attack accuracy against different distances and angles. (a) Distance; (b) Angle.
Entropy 25 00288 g011
Table 1. PSNR and SSIM of watermarked image with different watermark capacity.
Table 1. PSNR and SSIM of watermarked image with different watermark capacity.
Evaluation Standards50 bits100 bits150 bits200 bits
PSNR30.4330.0228.8723.48
SSIM0.9230.9190.8450.808
Table 2. Comparison of network training time.
Table 2. Comparison of network training time.
MethodIteration TimePer Iteration Time (s)Total Time (min)
Stegastamp140,0000.185432
ScreenNet140,0000.244569
Table 3. Comparison of accuracy under different noise attacks.
Table 3. Comparison of accuracy under different noise attacks.
AttackAttack RangeStegastampScreenNet
JPEG Compression200.99880.9931
Dropout0.30.95940.9963
Cropout0.30.82440.8673
Crop0.050.60480.6405
Gaussian20.99740.9823
Table 4. Accuracy comparison of screenshot attack at different distances.
Table 4. Accuracy comparison of screenshot attack at different distances.
Distance20 cm30 cm40 cm50 cm
Watermarked
Image
Entropy 25 00288 i001 Entropy 25 00288 i002 Entropy 25 00288 i003 Entropy 25 00288 i004
Corrected
Image
Entropy 25 00288 i005 Entropy 25 00288 i006 Entropy 25 00288 i007 Entropy 25 00288 i008
HiDDeN0.80230.77890.74300.6692
Stegastamp0.96580.95080.92670.9286
ScreenNet0.97390.97630.99690.9790
Table 5. Accuracy comparison of screen shot attack under different angle.
Table 5. Accuracy comparison of screen shot attack under different angle.
Angle30° down45° down30° up45° up
Watermarked
Image
Entropy 25 00288 i009 Entropy 25 00288 i010 Entropy 25 00288 i011 Entropy 25 00288 i012
Corrected Image Entropy 25 00288 i013 Entropy 25 00288 i014 Entropy 25 00288 i015 Entropy 25 00288 i016
HiDDeN0.59640.56340.58300.5591
Stegastamp0.78570.78310.81970.7906
ScreenNet0.97130.98830.97940.9873
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gu, W.; Chang, C.-C.; Bai, Y.; Fan, Y.; Tao, L.; Li, L. Anti-Screenshot Watermarking Algorithm for Archival Image Based on Deep Learning Model. Entropy 2023, 25, 288. https://doi.org/10.3390/e25020288

AMA Style

Gu W, Chang C-C, Bai Y, Fan Y, Tao L, Li L. Anti-Screenshot Watermarking Algorithm for Archival Image Based on Deep Learning Model. Entropy. 2023; 25(2):288. https://doi.org/10.3390/e25020288

Chicago/Turabian Style

Gu, Wei, Ching-Chun Chang, Yu Bai, Yunyuan Fan, Liang Tao, and Li Li. 2023. "Anti-Screenshot Watermarking Algorithm for Archival Image Based on Deep Learning Model" Entropy 25, no. 2: 288. https://doi.org/10.3390/e25020288

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop