Next Article in Journal
Improving the Robustness and Quality of Biomedical CNN Models through Adaptive Hyperparameter Tuning
Previous Article in Journal
Study on Size Optimization of a Protective Coal Pillar under a Double-Key Stratum Structure
Previous Article in Special Issue
Local-Moment-Driven Robust Reversible Data Hiding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identification of Content-Adaptive Image Steganography Using Convolutional Neural Network Guided by High-Pass Kernel

1
Department of Computer Science and Engineering, Amity School of Engineering & Technology, Amity University Uttar Pradesh, Noida 201313, India
2
Department of Software Convergence, Andong National University, Gyeongbuk 36729, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(22), 11869; https://doi.org/10.3390/app122211869
Submission received: 14 October 2022 / Revised: 16 November 2022 / Accepted: 19 November 2022 / Published: 21 November 2022
(This article belongs to the Special Issue Development of IoE Applications for Multimedia Security)

Abstract

:
Digital images are very popular and commonly used for hiding crucial data. In a few instances, image steganography is misused for communicating with improper data. In this paper, a robust deep neural network is proposed for the identification of content-adaptive image steganography schemes. Multiple novel strategies are applied to improve detection performance. Two non-trainable convolutional layers is used to guide the proposed CNN with fixed kernels. Thirty-one kernels are used in both non-trainable layers, of which thirty are high-pass kernels and one is the neutral kernel. The layer-specific learning rate is applied for each layer. ReLU with customized thresholding is applied to achieve better performance. In the proposed method, image down-sampling is not performed; only the global average pooling layer is considered in the last part of the network. The experimental results are verified on BOWS2 and BOSSBase image sets. Content-adaptive steganography schemes, such as HILL, Mi-POD, S-UNIWARD, and WOW, are considered for generating the stego images with different payloads. In experimental analysis, the proposed scheme is compared with some of the latest schemes, where the proposed scheme outperforms other state-of-the-art techniques in the most cases.

1. Introduction

Image steganography schemes are applied to hide secret data/information that cannot be noticed by the naked eye. In most content-adaptive steganography schemes, syndrome-trellis code (STC) is considered to make the schemes less vulnerable. In this paper, four notable content-adaptive steganography schemes, such as HILL [1], Mi-POD [2], S-UNIWARD [3], and WOW [4], are analyzed. In steganography, there is only one unit value (addition or subtraction) in some image pixels, which is in accordance with the steganography scheme and message. STC ensures that there is no visible change in the image. In comparison to other classification problems, such as object recognition and texture classification, steganalysis is more challenging due to the low value of stego noise. Difference arrays of cover and stego images in the different steganography schemes are displayed in Figure 1, where the cover image [5] is shown in the first row. Difference arrays of cover and HILL, Mi-POD, S-UNIWARD, and WOW stego images are shown in the second and third rows for 0.2 bit per pixel (bpp) and 0.4 bpp payloads, respectively. Although it is difficult to interpret the difference between different steganography schemes, the changes are more evident with a higher value of the payload.
The effect of steganography with different schemes cannot be understood effectively using natural images, as displayed in Figure 1. A computer-generated (CG) image with (0–255) is displayed in the first row of Figure 2 to understand a better effect of different steganography schemes. There is a difference of one unit value in adjacent pixels, with the exception of major diagonal pixel values, as well as a difference of 255 unit values in major diagonal pixel values. In different images of cover and stego images, it can be observed that each scheme has different behavior. In the HILL scheme, the changes in pixel values are more diversified than Mi-POD, followed by S-UNIWARD. However, WOW behavior is entirely different. The variations in the behavior of schemes are also reflected in experimental analysis.
Several attempts have been made for the detection of image steganography. Initially, the thriving steganalysis is performed using Markov model-based features. Moreover, some researchers proposed steganalysis schemes based on texture descriptors. The first order and higher order difference arrays and residual 2D arrays are exploited for large-size feature vectors. As the large-size feature vector is more effective in steganalysis due to low stego noise, a feature vector is extracted from the spatial and/or frequency domain. Lyu and Farid [6] exploited the statistics of magnitude and phase information in the frequency domain for the detection of F5 [7] and Outguess [8] steganography schemes. Li et al. [9] used the texture and Markov features, with a dimension of 22,153, for the detection of HUGO [10] steganography scheme. Fridrich and Kodovsky [11] introduced thirty high-pass kernels that are frequently recognized as SRM kernels. These kernels are based on the first and higher order difference arrays and multiple quantization factors. Additionally, SRM kernels are applicable in most deep networks. The feature vector with a dimension of 34,671 is extracted using a Markov model, where the scheme is verified on HUGO and ±1 variants steganography schemes. Penvy et al. [12] extracted Markov features of the second order from the first order difference arrays for the detection of the ±1 steganography scheme. The variants of the scheme [11], abbreviated as maxSRM and maxSRMd2, are proposed by Denemark et al. [13]. The formation process of the co-occurrence matrix is modified to decrease the feature vector size in the maxSRM and maxSRMd2 schemes. The Denemark scheme is verified on WOW, S-UNIWARD, and S-UNIGARD steganography schemes. Furthermore, Tang et al. [14] utilized the Markov model, although the dense and edge region is exploited for better results. The local correlation pattern [15] is proposed for identifying HUGO, S-UNIWARD, and the LSB matching revisited LSBMR [16] steganography schemes. Li et al. [17] proposed the steganalysis scheme based on texture operator features and verified it on HUGO and WOW steganography schemes. In this scheme, images are preprocessed with high-pass kernels. Moreover, PCA is used to decrease the size of feature vector and increase the performance. In [18,19,20], popular texture operators, such as the variant of the local binary pattern (LBP) with some improvements, are proposed for steganalysis. Moreover, Markov features are combined for better performance. In [18,19], schemes are verified on HILL, Mi-POD, and S-UNIWARD steganography schemes. A steganalysis scheme [20] is applied for the detection of HILL, CMD-HILL [21], and Mi-POD steganography schemes. At present, many recent steganalysis schemes depend on deep networks, as the high computing power hardware is not an obstacle. Qian et al. [22,23] introduced the deep network for the detection of HUGO, S-UNIWARD, and WOW steganography schemes. Images are preprocessed by one 5 × 5 high-pass kernel. However, the results are inferior to some manual schemes. In the Xu et al. method [24], the tanh is applied rather than ReLU for the identification of HILL and S-UNIWARD steganography schemes. In [25,26], numerous residual 2D arrays and one 5 × 5 high-pass kernel are used to enhance the performance on many types of stego images. Ye et al. [27] used the residual array after processing with thirty high-pass kernels [11]. Truncated linear unit (TLU) is considered rather than ReLU to outperform previous schemes on WOW, HILL, and S-UNIWARD. A deep network known as steganalysis residual network (SRNet) [28] is introduced, where skip connections and pooling layers with other types of layers are used to create three categories of blocks. Moments are extracted from a trained deep network to identify HILL, S-UNIWARD, and WOW for uncompressed images and J-UNIWARD and UED-JC [29] for compressed images. Yedroudj et al. [30] utilized thirty high-pass kernels [11] in a non-trainable convolutional layer. In [31], a new concept of shared batch normalization is proposed to improve the performance. The images are preprocessed with twenty high-pass kernels [11], which is verified on HILL, HUGO, S-UNIWARD, and WOW. In Zhang et al. [32], thirty high-pass kernels [11] are utilized in a non-trainable layer. In particular, two concepts are incorporated with the suggested CNN, bottleneck approach and spatial pyramid pooling, for the detection of three popular types of stego images. Furthermore, Xiang et al. [33] utilized thirty high-pass kernels in a non-trainable convolutional layer with some improvements in the order of layers. Wang et al. [34] preprocessed the image with thirty high-pass kernels and included the transfer learning approach. The authors exploited both spatial and frequency domains. The network was initialized with trained CNN from low-capacity stego images and verified on S-UNIWARD and WOW stego images. As can be seen from previous research on image steganalysis, the majority of the previous articles can improve the performance by exploiting the high-pass kernels in preprocessing to increase the stego noise for better classification. In this paper, two measures are considered to reduce the detection error without increasing the computational cost. The key measurements of the proposed scheme are outlined as follows:
  • In the proposed scheme, thirty-one kernels are used; thirty are high-pass kernels and one is the neutral kernel.
  • Two non-trainable convolutional layers are considered using thirty-one kernels; one layer is used at the beginning of the network and the second before the middle of the network.
  • To retain a complete statistical information, down-sampling is not performed.
  • The layer-specific learning rate is considered for better results.
  • The clipped ReLU layer is applied with a customized cut-off value for better control on the CNN.
  • Softmax classifier is the popular choice in CNN. However, several classifiers are investigated and the SVM classifier is the most suitable.
  • The outcomes of the proposed scheme are equated with the popular schemes Zhu-Net, SRNet, Yedroudj-Net, and Ye-Net.
  • The comprehensive outcomes are discussed for HILL, Mi-POD, S-UNIWARD, and WOW steganography schemes with 0.2, 0.3, and 0.4 bpp payloads.
  • The proposed scheme is discussed in detail in Section 2. The experimental evaluation is carried out in Section 3 and the conclusions are presented in Section 4.

2. The Proposed Scheme

Numerous image steganography schemes are used to hide some secret data in an image. At present, deep networks are found to be effective in state-of-the-art steganalysis schemes. In this paper, a robust deep neural network for the steganalysis of content-adaptive steganography schemes is proposed. The variation in different steganography schemes can be understood from Figure 2, and only one standard solution can be proposed for steganalysis. The effect of HILL, Mi-POD, S-UNIWARD, and WOW steganography schemes can be understood by covariance plots of image entropy in Figure 3. A total of twenty thousand images from BOWS2 [35] and BOSSBase [36] datasets is considered as a cover image. The same number of corresponding stego images with payload PL = {0.1, 0.2, 0.3, 0.4} is created using HILL, Mi-POD, S-UNIWARD, and WOW steganography schemes. In Figure 3, it is evident that a very small gap exists for cover and HILL PL = 0.1, which indicates that the detection of HILL may be challenging. However, the gap is more evident in Mi-POD followed by S-UNIWARD and WOW. Moreover, the behavior of covariance plot is reflected in the classification of cover and stego images. The highest detection error is found in HILL, followed by Mi-POD, S-UNIWARD, and WOW.
The block diagram of the proposed scheme is shown in Figure 4. The proposed convolutional neural network has two non-trainable convolutional layers. Thirty high-pass kernels and one neutral kernel of size 5 × 5 are used in non-trainable convolutional layers. In Figure 5, one neutral kernel and three high-pass kernels are displayed to understand the effect of kernels. Each kernel provides a different type of information. High-pass kernels increase the stego noise. Conventionally, in [27,30,31,32,33,34], only one non-trainable convolutional layer was used with thirty high-pass kernels. The neutral kernel is applied to retain the unprocessed information from the preceding layer. Moreover, a neutral filter increases the statistical information as unprocessed information is added with high-pass kernel operated information. The effect is verified in experimental analysis. The proposed CNN is guided by two non-trainable convolutional layers. It has already been established that a single non-trainable convolutional layer with thirty high-pass kernels enhances the detection performance for cover and stego images. In experimental analysis, it is discovered that the performance can improve significantly using an additional non-trainable convolutional layer. However, there is an adverse effect when three non-trainable convolutional layers are tried. Therefore, two non-trainable layers are considered. The second non-trainable layer provides the best results when used not very far from the first non-trainable layer. If the distance between non-trainable layers becomes very far, then there is a drop in the performance of the network.
The network considers ten trainable convolutional layers, and their weights are initialized using Glorot and Bengio [37] scheme. Each trainable convolutional layer is followed by batch normalization and a clipped ReLU layer [38]. The layer-specific learning rate is considered for each trainable convolutional layer. In previous research, a fixed learning rate, irrespective of the layer, was considered. Layer-specific learning rate [39] can improve the performance substantially which is found to be effective in many other applications [40,41]. The layer-specific learning is adapted by rigorous experimental analysis.
The ReLU layer is utilized in the majority of the earlier networks. ReLU replaces the negative elements with zero and remains intact in non-negative elements. The ReLU can be defined as:
E x = x , x 0 0 , x < 0
In this paper, the clipped ReLU (CReLU) is used to give an improved control in the proposed CNN. Similar to ReLU, CReLU replaces the negative components with zero while replacing the non-negative items that are higher than a threshold value with threshold value (t). Positive elements smaller than the threshold value remain intact. The CReLU can be defined as:
C x = 0 , x < 0 x , 0 x < t t , x t
In the proposed CNN, down-sampling is not performed in order to gain the utmost statistical information from the preceding layers. After the last clipped ReLU layer, the global average pooling (GAP) layer is applied [30,42]. The GAP layer is effective in decreasing the detection error. Stochastic gradient descent (SGD) optimizer is used to train the CNN. The initial learning rate of 1 × 10−2 and L2 regularization factor 1 × 10−4 are considered. One hundred epochs are considered to fine-tune the proposed CNN. Mini-batch of size 10 is considered. In the experimental analysis, the SVM classifier with the quadratic kernel gave better results than the softmax classifier. Therefore, the SVM classifier is utilized in the proposed CNN to classify cover and stego images. In Figure 6, the effect of four kernels in layers three (second row), thirteen (third row), and twenty-eight (fourth row) is displayed after training the CNN. Each layer has a different type of effect on the image. The depth of the network has a significant effect on the kernels of the layer. As the depth increases, the specific details become more prominent.

3. Experimental Analysis

Digital image steganography is very popular for covert communication. Steganalysis is challenging as there are little changes with a lower impact on the stego image. In this paper, four popular content-adaptive steganography schemes are evaluated by classifying cover and stego images. HILL, Mi-POD, S-UNIWARD, and WOW steganography schemes are tested using the proposed scheme on BOWS2 [35] and BOSSBase [36] image sets. Each set has ten thousand images with a dimension of 512 × 512 pixels of different varieties. Experimental analysis is performed after changing the size of the image to 256 × 256 pixels using interpolation. In most of the previous schemes, 256 × 256 pixel images are considered. The proposed scheme is compared with the latest schemes, SRNet [28], Ye-Net [27], Yedroudj-Net [30], and Zhu-Net [32]. BOSSBase image set is used to create cover and stego images for experimental analysis and experimental results are shown in Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6. Five thousand images of each class, stego and cover, are considered for training the CNN. One thousand and four thousand images are considered for validation and testing, respectively. In Table 7 and Table 8, experimental results are shown using BOWS2 [35] and BOSSBase [36] image sets. The results are defined in terms of detection error. The small value of detection error represents a better performance.
In Table 1, the detection errors are compared using thirty and thirty-one kernels in non-trainable layers. As can be seen in Table 1, a neutral filter can improve the performance and reduce the detection error. In earlier articles, only thirty kernels are used. The additional kernel passes some unchanged information to subsequent layers, which can increase the detection accuracy. The improvement in experimental results can be seen in four steganography schemes with three payloads.
Table 1. Detection errors using thirty and thirty-one kernels.
Table 1. Detection errors using thirty and thirty-one kernels.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
Thirty Non-trainable kernels0.38750.32420.27120.34750.25830.23780.33080.21660.17020.25230.18880.1512
Thirty-one Non-trainable kernels0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
In general, one preprocessing or non-trainable convolutional layer is considered in previous networks. In the proposed scheme, CNN is guided by two non-trainable convolutional layers that contain thirty-one kernels. The first non-trainable layer is considered after the image input layer, and the second non-trainable layer after three trainable convolutional layers. In experiments, three non-trainable layers are also considered. However, the best performance is achieved when two non-trainable layers are used. There is a noticeable reduction in detection error when compared with the single non-trainable layer, as depicted in Table 2.
Table 2. Detection errors using different non-trainable convolutional layers.
Table 2. Detection errors using different non-trainable convolutional layers.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
Single Non-trainable Layer0.39250.32870.27390.34890.26240.24200.33370.22170.17490.25450.19050.1521
Two Non-trainable Layers0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
Three Non-trainable Layers0.38410.32460.26960.34790.25780.23900.32890.21570.17070.25250.18940.1503
The layer-specific learning rate is considered in the proposed scheme. Typically, a common learning rate is considered in each convolutional layer. Although the customized learning rate reduces the detection error moderately as shown in Table 3, customized learning rates are found after an exhaustive experimental analysis. Four steganography methods with three different payloads can be more effective in the proposed steganalysis scheme.
Table 3. Detection errors of fixed and layer-specific learning rate.
Table 3. Detection errors of fixed and layer-specific learning rate.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
Fixed
learing rate
0.38980.32780.27150.35060.26060.23880.32890.21640.17310.25480.18980.1516
Layer-specific learning rate0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
In Table 4, experimental results are compared using ReLU and clipped ReLU. High-pass kernels in non-trainable layers increase the values, although CReLU is considered to control the increase. As discussed in the previous section, some specific values are controlled by the threshold in clipped ReLU. Similar to ReLU, CReLU replaces the negative components with zero while replacing the non-negative items that are higher than a threshold value with a particular threshold value. Positive elements below the threshold value are retained.
Table 4. Detection errors using ReLU and clipped ReLU.
Table 4. Detection errors using ReLU and clipped ReLU.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
ReLU0.38560.32360.27200.34620.25810.23970.33020.21700.17210.25400.19000.1519
Cipped ReLU0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
Support vector machine (SVM) is a very popular classifier. The proposed scheme replaces the softmax classifier with an SVM classifier with a quadratic kernel for a better detection of cover and stego images. There is a fair deduction in detection errors using the SVM classifier as shown in Table 5. A detailed analysis regarding the outperformance of SVM over softmax classifier is performed by Tang [43]. While SVM seeks to maximize the margin between data points belonging to different classes, the softmax classifier minimizes the cross-entropy. The proposed steganography scheme can successfully classify the cover and four steganography techniques with 0.2, 0.3, and 0.4 bpp payloads.
Table 5. Detection errors using softmax and SVM classifiers.
Table 5. Detection errors using softmax and SVM classifiers.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
Softmax Classifier0.38540.32620.27350.34760.26200.24220.32980.22020.17570.25530.19330.1567
SVM Classifier0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
The proposed scheme is compared with the recent schemes SRNet [28], Ye-Net [27], Yedroudj-Net [30], and Zhu-Net [32] as shown in Table 6. The proposed scheme outperforms other schemes with the exception of two cases. As represented in Figure 2, the spread of stego noise is more uniform than other steganography schemes, making the detection of HILL very challenging. HILL has the maximum detection error than the other schemes. WOW has the least detection error as the stego noise is more concentrated in some specific areas. The proposed scheme outperforms other schemes with a payload of 0.2 bpp, with the exception of Mi-POD and S-UNIWARD. A detection error reduction from 2.37% to 13.77% is observed in most cases when the proposed scheme is used, with an average decline of 6.92% in classification error.
Table 6. Detection errors of the proposed and other schemes using BOSSBase image set.
Table 6. Detection errors of the proposed and other schemes using BOSSBase image set.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
SRNet0.45600.37890.33050.42210.34170.28060.35680.26860.22250.28500.22260.1783
Ye-Net0.46720.42020.37360.43110.37330.34700.40580.33010.27060.32280.29220.2214
Yedroudj-Net0.47100.42160.33720.42940.37980.29520.41220.31890.27570.30740.26520.2071
Zhu-Net0.38880.33390.28780.33850.28280.25760.31670.23910.19510.26890.23390.1489
Proposed Scheme0.37960.31980.26610.34140.25480.23430.32340.21320.16820.24910.18570.1484
In Table 7, both BOWS2 and BOSSBase image sets are considered for experimental analysis. Seven thousand images from each dataset are considered as the cover image for training the CNN. Corresponding stego images are created using a particular steganography scheme with specific payloads. The trained CNN is evaluated using six thousand remaining images of BOWS2 and BOSSBase image sets. There is a significant reduction in detection error after considering twenty-eight thousand images of both classes for training and twelve thousand images for testing the proposed CNN. Moreover, the experimental results show that the categorization error increases as the payload increases, which implies that a low payload makes the classification difficult.
Table 7. Detection errors of proposed and other schemes using BOWS2 and BOSSBase image sets.
Table 7. Detection errors of proposed and other schemes using BOWS2 and BOSSBase image sets.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
SRNet0.41800.34680.31700.39670.32120.26380.33920.24950.20870.25050.18320.1401
Ye-Net0.46010.39940.35270.40520.35090.32620.38640.30650.25110.30090.24020.1947
Yedroudj-Net0.44260.38190.32570.40360.35700.27750.38350.29680.25430.28310.21300.1677
Zhu-Net0.37180.31690.24140.32090.26580.24210.28750.21020.16950.23150.17020.1057
Proposed Scheme0.36080.28670.24760.31090.24950.22030.26910.19220.15460.20660.15610.1011
In Table 8, experimental results are displayed after augmenting BOWS2 and BOSSBase image sets. One hundred thousand images are considered for training the CNN, and twenty-five thousand images for testing the trained CNN. There is a tremendous reduction in detection error using augmented images. Then, ten thousand images (Table 6) and twenty-eight thousand images (Table 7) are considered for training the CNN. The proposed scheme outperforms the other schemes and Zhu-Net by decreasing the detection error from 4% to 10%. Thirty high-pass kernels were applied by Ye et al. [27], Yedroudj et al. [30], and Zhang et al. [32]. The proposed scheme outcomes have been improved significantly with the use of thirty-one kernels and two non-trainable convolutional layers.
Table 8. Detection errors of proposed and other schemes using augmented images.
Table 8. Detection errors of proposed and other schemes using augmented images.
Steganography SchemeHILLMi-PODS-UNIWARDWOW
Payload (bpp)0.20.30.40.20.30.40.20.30.40.20.30.4
SRNet0.37390.31350.29200.30950.25050.20580.29300.21360.18070.22260.15920.1185
Ye-Net0.43240.36820.32200.31610.27370.25440.35990.27610.21350.28470.21750.1697
Yedroudj-Net0.40930.34560.31400.31480.27850.21650.34790.26730.22920.26080.17740.1381
Zhu-Net0.34690.29810.22910.27030.20730.18890.23120.19360.14000.19320.14130.0795
Proposed Scheme0.31660.26550.21140.24820.18680.17180.21290.18590.12860.17650.12960.0718

4. Conclusions

In this paper, a robust steganalysis scheme has been proposed for the detection of content-adaptive steganography schemes. Two non-trainable convolutional layers with fixed thirty-one kernels were incorporated to guide the proposed CNN, and multiple novel strategies were considered to increase the performance of CNN. Moreover, customized learning rates were applied to convolutional layers. In the proposed scheme, the clipped ReLU was considered rather than ReLU for improvement and the SVM classifier was used as an alternative to softmax classifier. Furthermore, the proposed scheme was verified on popular content-adaptive steganography schemes, such as HILL, Mi-POD, S-UNIWARD, and WOW. A remarkable reduction in the detection errors from 2% to 10% was observed in the classification of cover and stego images. Furthermore, in the most cases, the performance of the proposed scheme was more outstanding than other state-of-the-art techniques.

Author Contributions

Both authors discussed the details of the manuscript. S.A. designed and wrote the manuscript. S.A. implemented the proposed technique and provided the experimental results. K.-H.J. drafted and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Brain Pool program funded by the Ministry of Science and ICT through the National Research Foundation of Korea (2019H1D3A1A01101687) and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2021R1I1A3049788).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this paper are publicly available, and their links are provided in the reference section.

Acknowledgments

We thank the anonymous reviewers for their valuable suggestions that improved the quality of this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, B.; Wang, M.; Huang, J.; Li, X. A new cost function for spatial image steganography. In Proceedings of the 2014 IEEE International Conference on Image Processing, ICIP 2014, Paris, France, 27–30 October 2014; pp. 4206–4210. [Google Scholar] [CrossRef]
  2. Sedighi, V.; Cogranne, R.; Fridrich, J. Content-Adaptive Steganography by Minimizing Statistical Detectability. IEEE Trans. Inf. Forensics Secur. 2016, 11, 221–234. [Google Scholar] [CrossRef] [Green Version]
  3. Holub, V.; Fridrich, J.; Denemark, T. Universal distortion function for steganography in an arbitrary domain. EURASIP J. Inf. Secur. 2014, 2014, 1. [Google Scholar] [CrossRef] [Green Version]
  4. Holub, V.; Fridrich, J. Designing steganographic distortion using directional filters. In Proceedings of the WIFS 2012—Proceedings of the 2012 IEEE International Workshop on Information Forensics and Security, Costa Adeje, Spain, 2–5 December 2012; pp. 234–239. [Google Scholar] [CrossRef] [Green Version]
  5. Weber, A.G. The USC-SIPI Image Database: Version 5. USC-SIPI Rep. 2006, 315, 1–24. [Google Scholar]
  6. Lyu, S.; Farid, H. Steganalysis Using Higher-Order Image Statistics. IEEE Trans. Inf. Forensics Secur. 2006, 1, 111–119. [Google Scholar] [CrossRef]
  7. Westfeld, A. F5—A Steganographic Algorithm. In International Workshop on Information Hiding; Springer: Berlin/Heidelberg, Germany, 2001; pp. 289–302. [Google Scholar]
  8. Provos, N.; Honeyman, P. Detecting Steganographic Content on the Internet. 2001. Available online: http://niels.xtdnet.nl/papers/detecting.pdf/ (accessed on 21 January 2022).
  9. Li, B.; Huang, J.; Shi, Y.Q. Textural features based universal steganalysis. In Proceedings of the Security, Forensics, Steganography, and Watermarking of Multimedia Contents X, San Jose, CA, USA, 27–31 January 2008; p. 681912. [Google Scholar] [CrossRef]
  10. Pevný, T.; Filler, T.; Bas, P. Using High-Dimensional Image Models to Perform Highly Undetectable Steganography. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2010; Volume 6387, pp. 161–177. [Google Scholar]
  11. Fridrich, J.; Kodovsky, J. Rich Models for Steganalysis of Digital Images. IEEE Trans. Inf. Forensics Secur. 2012, 7, 868–882. [Google Scholar] [CrossRef] [Green Version]
  12. Pevny, T.; Bas, P.; Fridrich, J. Steganalysis by Subtractive Pixel Adjacency Matrix. IEEE Trans. Inf. Forensics Secur. 2010, 5, 215–224. [Google Scholar] [CrossRef] [Green Version]
  13. Denemark, T.; Sedighi, V.; Holub, V.; Cogranne, R.; Fridrich, J. Selection-channel-aware rich model for Steganalysis of digital images. In Proceedings of the 2014 IEEE International Workshop on Information Forensics and Security, WIFS 2014, Atlanta, GA, USA, 3–5 December 2015; pp. 48–53. [Google Scholar] [CrossRef]
  14. Tang, W.; Li, H.; Luo, W.; Huang, J. Adaptive steganalysis against WOW embedding algorithm. In Proceedings of the 2nd ACM workshop on Information hiding and multimedia security—IH&MMSec ’14, Salzburg, Austria, 11–13 June 2014; pp. 91–96. [Google Scholar] [CrossRef]
  15. Xu, X.; Dong, J.; Wang, W.; Tan, T. Local correlation pattern for image steganalysis. In Proceedings of the 2015 IEEE China Summit and International Conference on Signal and Information Processing (ChinaSIP), Chengdu, China, 12–15 July 2015; pp. 468–472. [Google Scholar] [CrossRef]
  16. Mielikainen, J. LSB matching revisited. IEEE Signal Process. Lett. 2006, 13, 285–287. [Google Scholar] [CrossRef]
  17. Li, F.; Zhang, X.; Cheng, H.; Yu, J. Digital image steganalysis based on local textural features and double dimensionality reduction. Secur. Commun. Netw. 2016, 9, 729–736. [Google Scholar] [CrossRef]
  18. Li, B.; Li, Z.; Zhou, S.; Tan, S.; Zhang, X. New Steganalytic Features for Spatial Image Steganography Based on Derivative Filters and Threshold LBP Operator. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1242–1257. [Google Scholar] [CrossRef]
  19. Wang, P.; Liu, F.; Yang, C. Towards feature representation for steganalysis of spatial steganography. Signal Process. 2020, 169, 107422. [Google Scholar] [CrossRef]
  20. Ge, H.; Hu, D.; Xu, H.; Li, M.; Zheng, S. New Steganalytic Features for Spatial Image Steganography Based on Non-negative Matrix Factorization. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2020; Volume 12022, pp. 337–351. [Google Scholar]
  21. Li, B.; Wang, M.; Li, X.; Tan, S.; Huang, J. A Strategy of Clustering Modification Directions in Spatial Image Steganography. IEEE Trans. Inf. Forensics Secur. 2015, 10, 1905–1917. [Google Scholar] [CrossRef]
  22. Qian, Y.; Dong, J.; Wang, W.; Tan, T. Deep learning for steganalysis via convolutional neural networks. Media Watermarking Secur. Forensics 2015, 2015, 94090J. [Google Scholar] [CrossRef]
  23. Qian, Y.; Dong, J.; Wang, W.; Tan, T. Learning and transferring representations for image steganalysis using convolutional neural network. In Proceedings of the International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016. [Google Scholar] [CrossRef]
  24. Xu, G.; Wu, H.-Z.; Shi, Y.-Q. Structural Design of Convolutional Neural Networks for Steganalysis. IEEE Signal Process. Lett. 2016, 23, 708–712. [Google Scholar] [CrossRef]
  25. Wu, S.; Zhong, S.H.; Liu, Y. Steganalysis via deep residual network. In Proceedings of the International Conference on Parallel and Distributed Systems—ICPADS, Wuhan, China, 13–16 December 2016; pp. 1233–1236. [Google Scholar] [CrossRef]
  26. Wu, S.; Zhong, S.; Liu, Y. Deep residual learning for image steganalysis. Multimed. Tools Appl. 2017, 77, 10437–10453. [Google Scholar] [CrossRef]
  27. Ye, J.; Ni, J.; Yi, Y. Deep Learning Hierarchical Representations for Image Steganalysis. IEEE Trans. Inf. Forensics Secur. 2017, 12, 2545–2557. [Google Scholar] [CrossRef]
  28. Boroumand, M.; Chen, M.; Fridrich, J. Deep Residual Network for Steganalysis of Digital Images. IEEE Trans. Inf. Secur. 2019, 14, 1181–1193. [Google Scholar] [CrossRef]
  29. Guo, L.; Ni, J.; Shi, Y.Q. Uniform Embedding for Efficient JPEG Steganography. IEEE Trans. Inf. Forensics Secur. 2014, 9, 814–825. [Google Scholar] [CrossRef]
  30. Yedroudj, M.; Comby, F.; Chaumont, M. Yedroudj-Net: An Efficient CNN for Spatial Steganalysis. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2092–2096. [Google Scholar] [CrossRef] [Green Version]
  31. Wu, S.; Zhong, S.; Liu, Y. A Novel Convolutional Neural Network for Image Steganalysis with Shared Normalization. IEEE Trans. Multimed. 2020, 22, 256–270. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, R.; Zhu, F.; Liu, J.; Liu, G. Depth-Wise Separable Convolutions and Multi-Level Pooling for an Efficient Spatial CNN-Based Steganalysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 1138–1150. [Google Scholar] [CrossRef]
  33. Xiang, Z.; Sang, J.; Zhang, Q.; Cai, B.; Xia, X.; Wu, W. A New Convolutional Neural Network-Based Steganalysis Method for Content-Adaptive Image Steganography in the Spatial Domain. IEEE Access 2020, 8, 47013–47020. [Google Scholar] [CrossRef]
  34. Wang, Z.; Chen, M.; Yang, Y.; Lei, M.; Dong, Z. Joint multi-domain feature learning for image steganalysis based on CNN. EURASIP J. Image Video Process. 2020, 2020, 28. [Google Scholar] [CrossRef]
  35. Bas, P.; Furon, T. Break Our Watermarking System. 2008. Available online: http://bows2.ec-lille.fr/ (accessed on 21 January 2022).
  36. Bas, P.; Filler, T.; Pevný, T. “Break Our Steganographic System”: The Ins and Outs of Organizing BOSS. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2011; Volume 6958, pp. 59–70. [Google Scholar]
  37. Glorot, Y.B.X. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  38. Hannun, A.; Case, C.; Casper, J.; Catanzaro, B.; Diamos, G.; Elsen, E.; Prenger, R.; Satheesh, S.; Sengupta, S.; Coates, A.; et al. Deep Speech: Scaling up end-to-end speech recognition. arXiv 2014, arXiv:1412.5567. [Google Scholar]
  39. Singh, B.; De, S.; Zhang, Y.; Goldstein, T.; Taylor, G. Layer-Specific Adaptive Learning Rates for Deep Networks. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015; pp. 364–368. [Google Scholar] [CrossRef]
  40. Hermessi, H.; Mourali, O.; Zagrouba, E. Deep feature learning for soft tissue sarcoma classification in MR images via transfer learning. Expert Syst. Appl. 2019, 120, 116–127. [Google Scholar] [CrossRef]
  41. Alarifi, J.S.; Goyal, M.; Davison, A.K.; Dancey, D.; Khan, R.; Yap, M.H. Facial Skin Classification Using Convolutional Neural Networks. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2017; Volume 10317, pp. 479–485. [Google Scholar]
  42. Xu, G. Deep convolutional neural network to detect J-UNIWARD. In Proceedings of the 5th ACM Workshop on Information Hiding and Multimedia Security, Philadelphia, PA, USA, 20–22 June 2017. [Google Scholar] [CrossRef]
  43. Tang, Y. Deep Learning Using Linear Support Vector Machines. arXiv 2013, arXiv:1306.0239. [Google Scholar]
Figure 1. Cover image with residual arrays of cover and stego images.
Figure 1. Cover image with residual arrays of cover and stego images.
Applsci 12 11869 g001
Figure 2. CG cover image and residual arrays of cover-stego images.
Figure 2. CG cover image and residual arrays of cover-stego images.
Applsci 12 11869 g002aApplsci 12 11869 g002b
Figure 3. Covariance plots of cover and stego images entropy.
Figure 3. Covariance plots of cover and stego images entropy.
Applsci 12 11869 g003
Figure 4. Block diagram of the proposed scheme.
Figure 4. Block diagram of the proposed scheme.
Applsci 12 11869 g004
Figure 5. Neutral kernel and high-pass kernels.
Figure 5. Neutral kernel and high-pass kernels.
Applsci 12 11869 g005
Figure 6. Effect of different layer kernels. (a) An image, (b) layer 3 kernels effect, (c) layer 13 kernels effect, (d) layer 28 kernels effect.
Figure 6. Effect of different layer kernels. (a) An image, (b) layer 3 kernels effect, (c) layer 13 kernels effect, (d) layer 28 kernels effect.
Applsci 12 11869 g006aApplsci 12 11869 g006b
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Agarwal, S.; Jung, K.-H. Identification of Content-Adaptive Image Steganography Using Convolutional Neural Network Guided by High-Pass Kernel. Appl. Sci. 2022, 12, 11869. https://doi.org/10.3390/app122211869

AMA Style

Agarwal S, Jung K-H. Identification of Content-Adaptive Image Steganography Using Convolutional Neural Network Guided by High-Pass Kernel. Applied Sciences. 2022; 12(22):11869. https://doi.org/10.3390/app122211869

Chicago/Turabian Style

Agarwal, Saurabh, and Ki-Hyun Jung. 2022. "Identification of Content-Adaptive Image Steganography Using Convolutional Neural Network Guided by High-Pass Kernel" Applied Sciences 12, no. 22: 11869. https://doi.org/10.3390/app122211869

APA Style

Agarwal, S., & Jung, K. -H. (2022). Identification of Content-Adaptive Image Steganography Using Convolutional Neural Network Guided by High-Pass Kernel. Applied Sciences, 12(22), 11869. https://doi.org/10.3390/app122211869

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop