Next Article in Journal
Semantic Arithmetic Coding Using Synonymous Mappings
Previous Article in Journal
Quantum Error Mitigation in Optimized Circuits for Particle-Density Correlations in Real-Time Dynamics of the Schwinger Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spread Spectrum Image Watermarking Through Latent Diffusion Model

School of Information Science and Engineering, Huaqiao University, Xiamen 361021, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(4), 428; https://doi.org/10.3390/e27040428
Submission received: 10 March 2025 / Revised: 6 April 2025 / Accepted: 12 April 2025 / Published: 15 April 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
The rapid development of diffusion models in image generation and processing has led to significant security concerns. Diffusion models are capable of producing highly realistic images that are indistinguishable from real ones. Although deploying a watermarking system can be a countermeasure to verify the ownership or the origin of images, the regeneration attacks arising from diffusion models can easily remove the embedded watermark from the images, without compromising their perceptual quality. Previous watermarking methods that hide watermark information in the carrier image are vulnerable to these newly emergent attacks. To address these challenges, we propose a robust and traceable watermark framework based on the latent diffusion model, where the spread-spectrum watermark is coupled with the diffusion noise to ensure its security and imperceptibility. Since the diffusion model is trained to reduce information entropy from disordered data to restore its true distribution, the transparency of the hidden watermark is guaranteed. Benefiting from the spread spectrum strategy, the decoder structure is no longer needed for watermark extraction, greatly alleviating the training overhead. Additionally, the robustness and transparency are easily controlled by a strength factor, whose operating range is studied in this work. Experimental results demonstrate that our method performs not only against common attacks, but also against regeneration attacks and semantic-based image editing.

1. Introduction

Digital image watermarking enables information to be added to an image to protect the copyright of the image without compromising its visual quality. To achieve effective watermarking, the watermark should be recovered at the receiver side even when suffering from various distortions like JPEG compression, blurring, etc.
Traditional methods typically embed watermarks in the spatial domain, like the LSB embedding [1], or modify the coefficients in the transform domain using transformations such as discrete Fourier transform (DFT), discrete Cosine transform (DCT), and discrete Wavelet transform (DWT) [2,3]. As deep learning advances, the paradigm of image watermarking shifts from explicit spatial or transform domain to the implicit latent space of high dimensions. An end-to-end encoder–decoder framework was initially proposed in [4]. To tackle the robustness issue, researchers optimize the design of the noise layer via data augmentation. Tancik et al. [5] embed an invisible hyperlink to the images, utilizing the UNet style structure [6] to encode and decode the secret hyperlink information. The authors add various differential operations, approximating the print-shooting distortions in the noise layer to enhance its robustness. To counter the compression attack, the authors in [7] propose a mini-batch scheme, which changes the noise layer from real JPEG, simulated JPEG, and noise-free ones. Furthermore, Wang et al. [8] introduced an adaptive network which optimizes the embedding factor. A decoupled noise layer against screen-capture attacks is proposed in [9], where style transfer is employed to mimic channel distortions and aid in network training. Fernandez et al. [10] integrated the watermark into the latent space of a neural network pre-trained via self-supervised learning, effectively building feature space to hide the watermark. In recent years, with the development of diffusion models, researchers have discovered their potential in the field of watermarking. Zhao et al. [11] implanted watermarks into the training images to achieve copyright protection when generating images with diffusion models. By fine-tuning the latent decoder in the latent diffusion model, embedding messages in generated images was achieved in [12]. A stable diffusion based image watermark pipeline is proposed in [13], which embeds a tree-ring pattern [14] in the Fourier domain of the latent space. Tan et al. [15] achieved perceptual image watermarking by embedding the watermark in the low-level feature space generated by diffusion models.
In spite of the merits of the new modality of image watermarking via deep learning frameworks, the rapid development of generative models also raises new security concerns. A recent study in [16] showed that processing images with the latent diffusion model can remove watermarks from most existing methods. As shown in Figure 1, this regeneration attack can inject noise into the latent space of images to disrupt the original distribution, and then denoise it to regenerate images. This regeneration process can remove watermarks while preserving the fidelity of images. Additionally, text-to-image diffusion models can easily edit images with prompts, resulting in the unavailability of the embedded watermarks. However, most watermarking approaches did not take the regeneration attack and local editing into account, resulting in the vulnerability of the watermarking systems.
Since the regeneration attack and image editing are launched with the diffusion model, the latent representation of images integrated with the diffusion model can be an ideal place to hold the watermark. The spread spectrum watermark that draws on spread spectrum technology from communications has been proven to be a success in traditional watermarking schemes [17], as it spreads the watermark signal over a wide range of frequency bands, guaranteeing its security and imperceptibility. Therefore, it is employed to enhance the security and robustness of the watermark. Our contributions can be summarized as follows:
  • The proposed scheme embeds the watermark information in a spread spectrum manner. It does not require a pre-training embedding network or a decoder structure, and therefore, the network is shown to be more lightweight.
  • The noisy feature representation of the latent vector is explored to hide the watermark message. Benefiting from the reversible property of diffusion models, the watermarked image can be produced without compromising the perceptual quality. Additionally, the VAE encoder acts as a robust feature extractor, which also provides a sweet spot to accommodate the watermark. Furthermore, the diffusion operated on the lower dimensional vector yielded by the VAE-encoder is computationally efficient.
  • Experimental results show that our framework performs robustly under common attacks like JPEG compression, brightness adjustment, and blurring. More importantly, it is resistant to image regeneration and image editing.

2. Method

Figure 2 shows the framework of the proposed method. The clean image is fed into the pipeline to obtain noisy latent space. Next, we embed modulated spread spectrum watermark into the noisy latent vectors. Finally, the denoising network and the decoder for the latent vector are used to generate watermarked images. Unlike the existing deep learning-based watermarking schemes, which typically require a decoder network to restore the watermark, blind detection is achieved in this work by calculating the correlation between the watermark patterns and the feature representation vector in noisy latent space.

2.1. Diffusion and Inversion

The diffusion process in DDPM [18] adds Gaussian noise to the original data in T -steps through a Markov chain to obtain a noisy image x T as
q ( x t | x t 1 ) = N ( x t ; 1 β t x t 1 , β t I ) ,
where β t ( 0 , 1 ) is the variance at the step of t and t ( 0 , 1 , , T 1 ) . In the reverse process, the generation requires a noise predictor with which the noise is removed to restore the images. However, this diffusion process is stochastic and is therefore detrimental to the reconstruction of watermarked images. In contrast, DDIM inversion [19] implicitly adds predicted noise at each diffusion time step and outputs the final distribution as structured noise. This process can be defined as
x ^ 0 t = x t 1 α ¯ t ϵ θ ( x t ) α ¯ t ,
x t + 1 = α ¯ t + 1 x ^ 0 t + 1 α ¯ t + 1 ϵ θ ( x t ) ,
where α ¯ t = i = 0 t ( 1 β t ) , t ( 0 , 1 , , T 1 ) and ϵ θ ( · ) is a noise predictor based on a time-conditional UNet [6]. For efficiency, DDIM is employed in this work to generate the watermarked images.
To enhance the robustness and reduce computational complexity, we employ the variational autoencoder to transform images into the latent distribution and apply sampling in latent space [20]. As the latent space is much smaller compared to the original image space, the denoising and its reverse process run much faster.

2.2. Watermark Embedding

The goal of our watermarking scheme is to enable the extraction of the embedded watermark under various attacks and to provide high fidelity of the original image. To achieve this goal, we generate random watermark information M = [ m 0 , m 1 , , m k 1 ] , m i { 0 , 1 } and orthonormal code vectors P i = [ p i 0 , p i 1 , , p i ( n 1 ) ] , n = 2 k and i [ 0 , k 1 ] . The modulation process for generating the spread spectrum watermark W = [ w 0 , w 1 , , w n 1 ] can be simplified to:
w j = i = 0 k 1 ( ( 2 m i 1 ) · p i j ) ,
where j = 0 , 1 , , n 1 . To embed the watermark W , the clean image I ori R 3 × W × H is mapped to the latent space using the VAE encoder to obtain the latent vector z 0 R 4 × W / f × H / f , where f is the scale factor set to 8 in this work. Estimated noise is applied to z 0 in T time steps, yielding z T R 4 × W / f × H / f :
z t + 1 = α ¯ t + 1 z ^ 0 t + 1 α ¯ t + 1 ϵ θ ( z t ) ,
where α ¯ t = i = 0 t ( 1 β t ) and ϵ θ ( · ) is a noise estimator. Then, the noisy latent vector z T is split into four slices along the channel dimension. It is empirically found that embedding the watermark in the fourth channel leads to better image quality, in spite of the fact that the embedding capacity would profit from using all the four channels. Considering the invisibility of watermarks, watermark embedding is performed in 8 × 8 DCT blocks of z T . We extract the seventh high-frequency coefficients on the main diagonal from each block and concatenate them to construct the embedding matrix F co = [ F 0 , F 1 , , F n 1 ] . The watermark embedding is depicted as:
F w = F co + S · W ,
where S is the strength factor of the embedding. IDCT is applied to F w to update z T as z T . Finally, the watermarked image I gen is yielded by
I g e n = D e c ( ϵ θ ( z T , T ) ) ,
where Dec ( · ) is the VAE decoder to transform the latent vector back into image space.
To further improve the image quality, the original image is partially added to the generated image to obtain the watermarked image I w . The enhancement is achieved by
I w = λ I o r i + ( 1 λ ) I g e n ,
where λ is a weighting parameter.

2.3. Watermark Extraction

As the inverse operation of watermark embedding, watermark extraction is also conducted in the noisy latent space. Therefore, the VAE encoder, as in the embedding process, is imposed on the watermarked image followed by the inversion process, yielding the latent vector z Tw . Then, we apply DCT on the channel that was selected for watermark embedding. The extracted DCT coefficients are denoted as F w = [ F 0 , F 1 , , F n 1 ] . Note that the indexes of DCT coefficients in F w are the same as that of the embedding. The original watermark is recovered by calculating the correlation between the spread spectrum code and the detection matrix:
m i = 1 , if F w , P i 0 , 0 , otherwise ,
where · , · denotes the inner product operation.

2.4. Loss Function

Although watermarking is performed in the noisy latent space, the loss of image quality is still inevitable. In order to minimize the loss of image quality, the mean squared error (MSE) between I ori and I gen is used to optimize the trainable latent vector:
L I = 1 C H W I o r i I g e n 2 2 ,
where C , H , W are the image channels, height, and width. Additionally, SSIM loss [21] is also included to achieve structural similarity to the original image
L S = 1 SSIM ( I ori , I gen ) .
To get closer to human perception, perceptual loss [22] is also included,
L P = 1 C j H j W j ϕ j ( I o r i ) ϕ j ( I g e n ) 2 2 ,
where ϕ j ( · ) represents the output of the j -th layer of a pre-trained VGG model [23], C j , H j , W j are the number of channels, height, and width of the feature map in the j -th layer. The total loss is formulated as
L = γ 1 L I + γ 2 L S + γ 3 L P ,
where γ 1 , γ 2 , and γ 3 are weight factors.

3. Experiments

3.1. Experimental Settings

To evaluate our scheme, we adopt two public datasets, i.e., MS-COCO [24] and DiffusionDB [25]. We randomly sample 200 images sized 512 × 512 from each dataset. The pre-trained stable diffusion model v2.1 [20] and the DDIM sampler [19] with a sampling time step of 50 are employed as the generator of watermarked images. For the weight factors of the loss function, γ 1 , γ 2 and γ 3 are set to 10, 1, and 0.1, respectively. The Adam optimizer [26] is employed with an initial learning rate of 10 2 for 30 epochs. The learning rate is then decayed at a rate of 0.3 at the 31st and the 81st epoch. The framework is implemented by PyTorch 2.0.1 and trained on a single NVIDIA RTX 4080. The inference time of our model for watermarking an image is 404.87 s, and the time cost for watermark extraction is about 2.5 s per image. As robustness is one of the primary goals of watermarking techniques, the bit error rate (BER) is employed to evaluate the robustness of watermarks under various attacks. Some common attacks are considered, e.g., brightness or contrast adjustment, JPEG compression (Q = 50), Gaussian noise with σ = 0.05, blur with a filter size of 5, crop with a ratio of 0.5, and BM3D denoising [27]. Considering the latest attacks from deep generative models, VAE-based image compression with Q = 3 [28,29] and regeneration attack [16] are launched. For imperceptibility evaluation, we choose three metrics, i.e., PSNR, SSIM, and LPIPS [30]. Comparisons are made with four state-of-the-art (SOTA) deep learning-based methods HiDDeN [4], Stegastamp [5], SSL [10], and Stable Signature [12]. All methods set the watermark bit length of 32, except that stable signature uses the pre-trained model for 48 bits.

3.2. Imperceptibility Test

Table 1 presents the comparison of image quality with various schemes derived from deep learning models, while Figure 3 displays the visual comparisons. As Table 1 and Figure 3 show, HiDDeN performs the best in terms of overall image quality. However, it is very limited in terms of robustness, as shown in Table 2, especially when facing attacks with global modification. For our method, the loss of image quality is mainly due to the reduced dimension of the latent space, while it is still comparable to other deep learning watermarking schemes. In Figure 3, it can be seen that the proposed method embeds watermarks in the edge details of images, thus leading to better imperceptibility. Additionally, the proposed method modifies fewer pixels of the original image, as illustrated by the comparison of residual images in Figure 3. In contrast to Stable Signature, which is also based on the diffusion model, our training strategy in the noisy latent space achieves less reconstruction loss and thus leads to better quality of watermarked images in terms of all the three metrics.

3.3. Robustness Test

It is observed in Table 2 that most of the methods are vulnerable to the regeneration attack, whereas our proposed scheme performs robustly under all the three regeneration attacks. Figure 4a visualizes the noise distribution introduced by the diffusion attack in latent space. It is seen that its distribution closely resembles the noise distribution used in training, and hence, the impact of this attack on our proposed scheme is negligible. Benefiting from this embedding process, the proposed method performs better in terms of generalization against regeneration attacks compared to those based on noise layer training. To verify the effectiveness of the spread spectrum scheme, the radar chart in Figure 4b compares the robustness of using QIM [31] as the embedding scheme, with the quantization step set to 50. The chart highlights the superiority of the spread spectrum watermarking in latent space.
Table 2. Performance comparison of robustness under different attacks. The numbers 1 to 7 represent JPEG, brightness, contrast, Gaussian noise, blur, crop, and bm3d attacks, respectively. And the abbreviations B, C, D, P, and U stand for Bmshj [28], Cheng [29], Diffusion [16], Instruct-Pix2Pix [32], and UltraEdit [33], respectively. The best result of each column is marked in bold, while the second-best is underlined.
Table 2. Performance comparison of robustness under different attacks. The numbers 1 to 7 represent JPEG, brightness, contrast, Gaussian noise, blur, crop, and bm3d attacks, respectively. And the abbreviations B, C, D, P, and U stand for Bmshj [28], Cheng [29], Diffusion [16], Instruct-Pix2Pix [32], and UltraEdit [33], respectively. The best result of each column is marked in bold, while the second-best is underlined.
MethodsCommon AttacksRegeneration AttacksImage Editing
1234567BCDPU
COCO
HiDDeN0.3100.1510.1540.2230.1880.1820.2950.4170.3780.4160.4610.457
Stegastamp0.0010.0010.0020.0030.0010.1860.0010.0020.0020.1370.0580.200
SSL0.0320.0010.0010.0360.0010.2010.1220.3180.2900.2070.1310.409
Stable Signature0.1240.0060.0080.0250.0660.0120.1690.3770.2840.5220.5110.583
Proposed0.0140.0010.0010.0410.0010.2170.0210.0670.0700.0180.0100.079
Proposed (64)0.0130.0010.0010.0320.0010.2210.0190.0590.0700.0130.0140.063
DiffusionDB
HiDDeN0.3190.1550.1550.2260.1840.1870.3050.4200.3850.4220.4810.448
Stegastamp0.0010.0020.0030.0030.0010.1890.0010.0020.0030.1520.0730.258
SSL0.0380.0010.0010.0310.0010.2110.1390.3220.2980.2600.1630.376
Stable Signature0.1130.0050.0060.0130.0620.0080.1460.3750.2750.5210.4820.552
Proposed0.0250.0010.0010.0490.0010.2270.0290.0770.0810.0280.0180.100
Proposed (64)0.0230.0010.0010.0470.0010.2200.0300.0810.0890.0180.0250.094
To verify the effectiveness of the proposed under image editing attack, Instruct-Pix2Pix [32] and UltraEdit [33] are applied. The former is an instruction-driven image editing framework, enabling precise image modifications with user-specified commands, while the latter allows high-quality regional editing through the guidance of a mask. Figure 5 shows the editing styles imposed on the images, including local content replacement, background change and human action modification. These kinds of editing can modify the content of images easily and disrupt the embedded watermark information. The editing pipelines are controlled by the Text CFG (Classifier-free Guidance) and the Image CFG, which are the parameters for adjusting the relevance of the generated image to the prompt. In our experiment, the text and image CFG are set to 7.5 and 1.5, respectively. As demonstrated by the results shown in Table 2, the proposed method performs robustly to all kinds of image editing except image cropping. Although Stegastamp also shows very competitive results in terms of robustness to the attacks launched by generative models, our scheme excels a lot in terms of imperceptibility.

3.4. Capacity

In conventional watermarking systems, the embedding capacity is positively correlated with the size of the cover image. In contrast, our proposed method takes the latent representation of an image to hold the watermark. Since it is a low-dimensional representation of the image, it is much smaller in size than the cover image, limiting the embedding capacity of the watermark. On the flip side, embedding watermarks in the latent space ensures the robustness of watermarking systems, since transmission distortions and regeneration attacks tend to affect the high-frequency components of the image. To further evaluate the impact of the increased embedding capacity, we employ watermarks of 64 and 96 bits for the test. For 64 bits, the sixth coefficient on the main diagonal of each DCT block is additionally chosen for watermark embedding while the eighth coefficient is additionally selected for 96-bit embedding. From the results in Table 3, it is observed that a higher capacity like 96 bits leads to a significant degradation of image quality, in spite of a robustness gain as the reward. However, the watermarked image with an embedding capacity of 64 only suffers from slight image degradation, suggesting the trade-off between embedding capacity and imperceptibility of watermarks with our proposed scheme.

3.5. Discussions

In this section, we discuss some of the key parameters that affect the robustness and imperceptibility of the proposed watermarking scheme. In traditional watermarking schemes, selecting an appropriate strength factor to balance the invisibility and the robustness of the watermark is a vital concern. Generally, a larger strength factor would lead to better robustness, while compromising the image quality. In Figure 6a, the impact of the strength factor on image quality and watermark robustness is plotted, measured in terms of PSNR and BER. As Figure 6a suggests, the proposed scheme is fragile given a mild strength factor of less than 50, even when no attack is imposed. This is because watermark embedding is performed in the noisy latent space and the noise predictor might be inaccurate when the amplitude is small. Drawing on this, Figure 6b presents the range of BER under different attacks, by applying the strength factor from 50 to 100. It is demonstrated that an optimal option exists at S = 70 since S larger than 70 can bring about severe image degradation. Considering the available editing techniques and the potential distortions brought by channel transmission, we prefer to have higher priority on robustness of the watermark given an acceptable quality of the watermarked image.
In addition, the parameter λ in Equation (8) controls the weight of raw data to further enhance image quality. Meanwhile, λ also affects the robustness of the watermark, so we sample at equal intervals in the range of [0,0.5] to determine the choice of λ . From Figure 7, a value of 0.3 is opted for considering the tradeoff between image quality and robustness against attacks.

4. Conclusions

In this paper, we propose an image watermarking framework that achieves spread spectrum watermarking in latent diffusion models, thus guaranteeing watermark imperceptibility and robustness. Essentially, the proposed design delivers comparable visual quality to the state-of-the-art methods, but outperforms in terms of resisting most recent attacks stemming from generative models like image editing and regeneration. In future work, we will further address the challenge from image cropping, rotation, and explore solutions for improving the watermark capacity in diffusion models.

Author Contributions

Conceptualization, H.W. and X.L.; methodology, H.W. and X.L.; software, H.W.; validation, X.L.; formal analysis, G.T.; writing—original draft preparation, H.W. and X.L.; writing—review and editing, G.T; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities grant number ZQN-1112 and the National Natural Science Foundation of China under Grant 62002124.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Van Schyndel, R.G.; Tirkel, A.Z.; Osborne, C.F. A digital watermark. In Proceedings of the 1st International Conference on Image Processing, Washington, DC, USA, 13–16 November 1994; IEEE: New York, NY, USA, 1994; Volume 2, pp. 86–90. [Google Scholar]
  2. Hwang, M.J.; Lee, J.; Lee, M.; Kang, H.G. SVD-based adaptive QIM watermarking on stereo audio signals. IEEE Trans. Multimed. 2017, 20, 45–54. [Google Scholar] [CrossRef]
  3. Guan, H.; Zeng, Z.; Liu, J.; Zhang, S. A novel robust digital image watermarking algorithm based on two-level DCT. In Proceedings of the 2014 International Conference on Information Science, Electronics and Electrical Engineering, Sapporo, Japan, 26–28 April 2014; IEEE: New York, NY, USA, 2014; Volume 3, pp. 1804–1809. [Google Scholar]
  4. Zhu, J.; Kaplan, R.; Johnson, J.; Li, F. Hidden: Hiding data with deep networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 657–672. [Google Scholar]
  5. Tancik, M.; Mildenhall, B.; Ng, R. Stegastamp: Invisible hyperlinks in physical photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 14–19 June 2020; pp. 2117–2126. [Google Scholar]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  7. Jia, Z.; Fang, H.; Zhang, W. Mbrs: Enhancing robustness of dnn-based watermarking by mini-batch of real and simulated jpeg compression. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 41–49. [Google Scholar]
  8. Wang, B.; Wu, Y.; Wang, G. Adaptor: Improving the robustness and imperceptibility of watermarking by the adaptive strength factor. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 6260–6272. [Google Scholar] [CrossRef]
  9. Fang, H.; Chen, K.; Qiu, Y.; Liu, J.; Xu, K.; Fang, C.; Zhang, W.; Chang, E.C. DeNoL: A Few-Shot-Sample-Based Decoupling Noise Layer for Cross-channel Watermarking Robustness. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 7345–7353. [Google Scholar]
  10. Fernandez, P.; Sablayrolles, A.; Furon, T.; Jégou, H.; Douze, M. Watermarking images in self-supervised latent spaces. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 3054–3058. [Google Scholar]
  11. Zhao, Y.; Pang, T.; Du, C.; Yang, X.; Cheung, N.M.; Lin, M. A recipe for watermarking diffusion models. arXiv 2023, arXiv:2303.10137. [Google Scholar]
  12. Fernandez, P.; Couairon, G.; Jégou, H.; Douze, M.; Furon, T. The stable signature: Rooting watermarks in latent diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 22466–22477. [Google Scholar]
  13. Zhang, L.; Liu, X.; Martin, A.V.; Bearfield, C.X.; Brun, Y.; Guan, H. Robust Image Watermarking using Stable Diffusion. arXiv 2024, arXiv:2401.04247. [Google Scholar]
  14. Wen, Y.; Kirchenbauer, J.; Geiping, J.; Goldstein, T. Tree-ring watermarks: Fingerprints for diffusion images that are invisible and robust. arXiv 2023, arXiv:2305.20030. [Google Scholar]
  15. Tan, Y.; Peng, Y.; Fang, H.; Chen, B.; Xia, S.T. WaterDiff: Perceptual Image Watermarks Via Diffusion Model. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 3250–3254. [Google Scholar] [CrossRef]
  16. Zhao, X.; Zhang, K.; Su, Z.; Vasan, S.; Grishchenko, I.; Kruegel, C.; Vigna, G.; Wang, Y.X.; Li, L. Invisible image watermarks are provably removable using generative AI. arXiv 2023, arXiv:2306.01953. [Google Scholar]
  17. Cox, I.J.; Kilian, J.; Leighton, F.T.; Shamoon, T. Secure spread spectrum watermarking for multimedia. IEEE Trans. Image Process. 1997, 6, 1673–1687. [Google Scholar] [CrossRef] [PubMed]
  18. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  19. Song, J.; Meng, C.; Ermon, S. Denoising diffusion implicit models. arXiv 2020, arXiv:2010.02502. [Google Scholar]
  20. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar]
  21. Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss functions for image restoration with neural networks. IEEE Trans. Comput. Imaging 2016, 3, 47–57. [Google Scholar] [CrossRef]
  22. Czolbe, S.; Krause, O.; Cox, I.; Igel, C. A loss function for generative neural networks based on watson’s perceptual model. Adv. Neural Inf. Process. Syst. 2020, 33, 2051–2061. [Google Scholar]
  23. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  24. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  25. Wang, Z.J.; Montoya, E.; Munechika, D.; Yang, H.; Hoover, B.; Chau, D.H. Diffusiondb: A large-scale prompt gallery dataset for text-to-image generative models. arXiv 2022, arXiv:2210.14896. [Google Scholar]
  26. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  27. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  28. Ballé, J.; Minnen, D.; Singh, S.; Hwang, S.J.; Johnston, N. Variational image compression with a scale hyperprior. arXiv 2018, arXiv:1802.01436. [Google Scholar]
  29. Cheng, Z.; Sun, H.; Takeuchi, M.; Katto, J. Learned image compression with discretized gaussian mixture likelihoods and attention modules. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Event, 14–19 June 2020; pp. 7939–7948. [Google Scholar]
  30. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  31. Chen, B.; Wornell, G.W. Quantization index modulation: A class of provably good methods for digital watermarking and information embedding. IEEE Trans. Inf. Theory 2001, 47, 1423–1443. [Google Scholar] [CrossRef]
  32. Brooks, T.; Holynski, A.; Efros, A.A. InstructPix2Pix: Learning to Follow Image Editing Instructions. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 18392–18402. [Google Scholar] [CrossRef]
  33. Zhao, H.; Ma, X.; Chen, L.; Si, S.; Wu, R.; An, K.; Yu, P.; Zhang, M.; Li, Q.; Chang, B. UltraEdit: Instruction-based Fine-Grained Image Editing at Scale. 2024. Available online: http://arxiv.org/abs/2407.05282 (accessed on 6 December 2024).
Figure 1. Image regeneration attack.
Figure 1. Image regeneration attack.
Entropy 27 00428 g001
Figure 2. The proposed framework.
Figure 2. The proposed framework.
Entropy 27 00428 g002
Figure 3. Visual quality comparison with various methods. From left to right are clean images, HiDDeN, Stegastamp, SSL, Stable Signature, and proposed. The second and fourth rows are the residual images.
Figure 3. Visual quality comparison with various methods. From left to right are clean images, HiDDeN, Stegastamp, SSL, Stable Signature, and proposed. The second and fourth rows are the residual images.
Entropy 27 00428 g003
Figure 4. (a) Comparison of noise distribution; and (b) Visualization of the robustness comparison between using spread spectrum and QIM for watermarking.
Figure 4. (a) Comparison of noise distribution; and (b) Visualization of the robustness comparison between using spread spectrum and QIM for watermarking.
Entropy 27 00428 g004
Figure 5. Performance of different prompts for image editing. The first group is the result with Instruct-Pix2Pix, the second is with UltraEdit.
Figure 5. Performance of different prompts for image editing. The first group is the result with Instruct-Pix2Pix, the second is with UltraEdit.
Entropy 27 00428 g005
Figure 6. (a) PSNR and BER under different strength factors; (b) Range of BER under different attacks.
Figure 6. (a) PSNR and BER under different strength factors; (b) Range of BER under different attacks.
Entropy 27 00428 g006
Figure 7. Parameter analysis about λ .
Figure 7. Parameter analysis about λ .
Entropy 27 00428 g007
Table 1. Performance comparison of quantitative image quality among different methods. The best result is marked in bold, while the second-best is underlined.
Table 1. Performance comparison of quantitative image quality among different methods. The best result is marked in bold, while the second-best is underlined.
COCODiffusionDB
Methods PSNR ↑ SSIM ↑ LPIPS ↓ PSNR ↑ SSIM ↑ LPIPS ↓
HiDDeN31.700.930.0231.420.940.02
Stegastamp28.730.890.0728.050.890.07
SSL32.070.870.1132.110.880.10
Stable Signature26.430.750.0625.780.750.06
Proposed30.850.900.0530.490.900.05
Proposed (64)28.280.880.1028.630.890.09
Table 3. Performance comparison of the quantitative image quality and robustness under different embedding capacities in COCO dataset. The average BER of the three categories of attacks in Table 2 is computed to reflect the robustness. The best result is marked in bold.
Table 3. Performance comparison of the quantitative image quality and robustness under different embedding capacities in COCO dataset. The average BER of the three categories of attacks in Table 2 is computed to reflect the robustness. The best result is marked in bold.
TransparencyRobustness
Capacity PSNR ↑ SSIM ↑ LPIPS ↓ Common Regeneration Editing
3230.850.900.050.0420.0520.045
Proposed6428.280.880.100.0410.0710.039
9622.730.780.270.0260.0220.026
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, H.; Lin, X.; Tan, G. Spread Spectrum Image Watermarking Through Latent Diffusion Model. Entropy 2025, 27, 428. https://doi.org/10.3390/e27040428

AMA Style

Wu H, Lin X, Tan G. Spread Spectrum Image Watermarking Through Latent Diffusion Model. Entropy. 2025; 27(4):428. https://doi.org/10.3390/e27040428

Chicago/Turabian Style

Wu, Hongfei, Xiaodan Lin, and Gewei Tan. 2025. "Spread Spectrum Image Watermarking Through Latent Diffusion Model" Entropy 27, no. 4: 428. https://doi.org/10.3390/e27040428

APA Style

Wu, H., Lin, X., & Tan, G. (2025). Spread Spectrum Image Watermarking Through Latent Diffusion Model. Entropy, 27(4), 428. https://doi.org/10.3390/e27040428

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop