Next Article in Journal
Parameter Optimisation in Selective Laser Melting on C300 Steel
Next Article in Special Issue
Stripe-Based Futurism-like Image—A New Type of Art and Its Application to Covert Communication
Previous Article in Journal
Temporal and Spatial Variation Characteristics of Actual Evapotranspiration in the Yiluo River Basin Based on the Priestley–Taylor Jet Propulsion Laboratory Model
Previous Article in Special Issue
Joint Hamming Coding for High Capacity Lossless Image Encryption and Embedding Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SSDeN: Framework for Screen-Shooting Resilient Watermarking via Deep Networks in the Frequency Domain

1
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
2
Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou 310018, China
3
Hangzhou Dianzi University, Shangyu Institute of Science and Engineering, Shaoxing 312000, China
4
Department of Information Engineering and Computer Science, Feng Chia University, Taichung 40724, Taiwan
*
Author to whom correspondence should be addressed.
Submission received: 15 August 2022 / Revised: 14 September 2022 / Accepted: 15 September 2022 / Published: 28 September 2022
(This article belongs to the Special Issue Recent Advances in Steganography and Multimedia Security)

Abstract

:
Mobile devices have been increasingly used to take pictures without leaving a trace. However, the application system can lead to confidential information leaks. A framework for screen-shooting-resilient watermarking via deep networks (SSDeN) in the frequency domain is put forward in this study to solve this problem. The proposed framework can extract the watermark from the leaked photo for copyright protection. SSDeN is an end-to-end process that combines convolutional neural network (CNN) with residual block to embed and extract watermarks in the DCT domain. We simulate some screen-shooting attacks to ensure the networks embed the watermark robustly. Our framework achieves the state-of-the-art performance on existing learning architectures for screen-shooting-resilient watermarking.

1. Introduction

Digital media techniques in the network information age provide excellent opportunities for information retrieval, access and spreading, while the electronic reading environment presents difficult- or impossible-to-obtain information on paper. These features have remarkably changed and continue to change the way we read. Notably, screen reading is a new trend that has become a necessity for modern people. The use of TV, computer, laptop computer and mobile phone screens and electronic dictionaries in screen reading has changed the way people read from paper to obtain information. At the same time, taking photographs has become a simple and efficient way of information transmission with the prevalence of mobile cameras, thereby increasing the threat to information security. The increased usage of mobile devices to take pictures without leaving a trace has led to many leaks of internal confidential information. A robust watermarking algorithm that extracts watermarks from screen-shooting photographs is necessary to protect confidential information given that images displayed on the screen can embed imperceptible watermarks.
Several screen-shooting-resilient watermarking schemes have been presented in recent years. Robust watermarking techniques are divided into spatial [1] and frequency domains. Compared with that based on the spatial domain, the frequency-domain watermarking method can achieve stronger watermarking robustness without causing serious image distortion. Frequency-domain watermarking techniques, such as DCT [2,3], QDFT [4] and tensor decomposition [4], can help improve imperceptibility and robustness. Fang et al. [2] introduced an intensity-based SIFT framework to extract confidential watermark information from DCT coefficients of the image displayed on the screen. However, SIFT’s keypoints are inconsistent and unstable under different screen-shooting attacks. Ahmadi et al. [3] presented a diffusion watermarking scheme that can embed a watermark in the DCT transform domain via learning. Li et al. [4] proposed a watermarking scheme using learned invariant keypoints, quaternion discrete Fourier transform (QDFT), and tensor decomposition (TD), which hybrid machine learning and traditional watermarking algorithms use to achieve screen-shooting-resilient watermarking.
At present, the applicability of deep learning technology in the field of watermarking algorithm is growing rapidly, because this technology can effectively solve the process of watermark embedding and extraction [3,4,5,6,7,8,9,10,11,12,13]. Tancik et al. [11] proposed a screen-shooting-resilient watermarking technique that jointly trains the scale distortion, blur, noise attack, JPEG compression, and color distortion attacks, while training the deep learning network to resist common printing, remakes, screen photography and other attacks. Wengrowski and Dana [5] put forward an end-to-end watermarking strategy for the screen-shooting process. Datasets [6] were introduced by the framework when the camera and display are at high-perspective angles to guarantee significantly better results than existing schemes. Zhang et al. [7] offered a watermarking scheme with universal deep hiding (UDH) and proved that only adding a perspective transformation layer between the encoder and decoder can resist distortions introduced by the screen-shooting process to a certain extent. Compared with these schemes, a strategy was proposed as the template generation scheme for creating the watermark-embedded template and then embedding the watermark into the red and blue of the carrier image [8]. The trained deep learning network was applied to extract the watermark in the watermark extraction stage. Fang et al. [9] subsequently established a watermarking scheme named “TERA”. Similar to [8], the process of watermark embedding depends on a manual design named the superposition-based embedding scheme, and watermark extraction is designed using a deep neural network (DNN). The extraction algorithm is composed of enhancing, attention-guided [10] and regression networks. HiDDeN [13] applied neural networks are used to realize robust watermarking. Specifically, noise distortion layers are added to enhance their framework.
We choose deep learning to construct an image watermarking framework that considers screen-shooting resilient watermarking complexity because of its recent outstanding performance on image watermarking tasks. Ahmadi et al. [3] proposed a robust watermarking technique via deep learning in the transform domain, but the method fails to resist screen-shooting attacks. Tancik et al. [11] put forward a screen-shooting resilient watermarking scheme in the spatial domain with minimal robustness. Spatial-domain-based watermarking is simple to implement, but it has less information embedded and poor robustness. Transform-domain-based watermarking can provide better embedding capacity and robustness under various attacks. We propose a screen-shooting-resilient watermarking algorithm based on [3,11] using deep networks in the DCT domain.
The main contributions of the paper are presented as follows:
  • The framework uses a differentiable noise layer for simulating screen shooting attacks between the embedding and extracting layers in the DCT domain to improve security and robustness.
  • The moiré phenomenon is a common screen-shooting attack. We trained a convolutional network to add moiré patterns to images automatically.
  • We introduce a loss function based on NC and SSIM to balance the robustness and imperceptibility of the algorithm.
  • We apply STN before extracting the watermark to correct the image with geometric attacks and enhance the precision of watermark extraction.
  • The proposed end-to-end blind robust watermarking framework reduces the artificial pretreatment and post-processing to ensure that the model can change the original input to the final output. Therefore, the model is given additional space to adjust automatically according to image features and improve its overall fitting.
The remainder of this paper is organized as follows. Relevant techniques are described in Section 2. Details of the proposed framework are provided in Section 3. The experimental results are discussed in Section 4. Finally, the conclusions of this study are summarized in Section 5.

2. Related Works

We summarize the key issues in designing the SSDeN framework in this section.

2.1. Watermarking Algorithm in the Transform Domain

The transform domain watermarking technique can help improve the imperceptibility and robustness of watermarking, and many watermarking schemes, such as DWT [14,15], DFT [16], DCT [17,18], quaternion discrete Fourier transform (QDFT) [19,20,21] and quaternion Hadamard transform (QHD) [22], have been proposed. Discrete cosine transform (DCT) is a widely used transform domain watermarking technique that modifies the image from the spatial domain to the frequency domain. DCT divides the image into different frequency bands, namely, high-, medium- and low-frequency bands, to select frequency bands easily when embedding the watermark. The middle-frequency band is mainly selected in this work because it can balance the relationship between robustness and imperceptibility and avoid diffusion to high-frequency components against compression and noise attacks to improve image robustness [2]. Watermarking algorithms for different carriers can be divided into two categories in the frequency domain, namely, single- (SCI) and multi-channel images (MCI).
Watermarking-algorithm-based SCI. Guan et al. [23] designed a robust digital image watermarking algorithm that can resist image attacks and distortions and embed the watermark into the two-level DCT coefficients of the image using certain rules. A robust watermarking scheme based on the wavelet transform domain was proposed to resist attacks for single-channel images effectively [24]. A novel image watermarking scheme based on singular value decomposition (SVD) was proposed [25] to overcome the FPP problem and meet the watermarking requirements using integer wavelet transform (IWT). Multi-objective ant colony optimization (MOACO) is utilized to balance robustness and imperceptibility and find the optimal scaling factor while completely avoiding the FPP problem. Rastegar et al. [26] embedded a gray-scale watermark into single-channel images by blending singular value decomposition and the steerable pyramid wavelet transform domain. A robust watermarking method with SVD and DWT [27] was put forward to embed the watermark into the singular values of DWT subbands of the gray image.
Watermarking algorithm-based MCI. Watermarking schemes for color image have continuously emerged because color images are widely used in network transmission and storage [15,19,21,28,29,30,31,32,33,34,35]. Chou and Liu [2] established a new color-image watermarking algorithm based on wavelet transform and significant difference, as well as embedded the maximum watermark information under imperceptible distortion. A link was established between the QDFT coefficient, and the color channel’s DFT and QDFT coefficients were modulated to embed the watermark in the transform domain [19]. Wang et al. [21] embedded a watermark into the cover image by adaptively modulating the real coefficients of the quaternion Fourier transform while using LS-SVM to correct the attacked images. Ma et al. [28] developed a local QDFT watermarking method for color images that utilize the advantage of QDFT to improve the robustness and imperceptibility of watermarking. Kais Rouis et al. [29] developed an image hashing algorithm dependent on the analysis of local geometric changes of multi-channel images via gradient measurement for image tempering detection. A geometrically corrected robust color image watermarking method based on the quaternion exponential moments was proposed [30]. Feng et al. [33] combined Tucker decomposition with A-LQIM for a robust image watermarking approach with satisfactory performance. Lorenzo et al. [35] proposed a watermarking method based on the Neyman–Pearson rule for multi-channel images to reduce the error rate when the watermark is nonexistent and to conduct blind image watermarking for gray-scale images in the DCT transform domain.

2.2. Watermarking Deep-Learning-Based Algorithm

Several deep learning (DL) watermarking algorithms have been proposed [3,11,13]. HiDDeN [13] uses an algorithm for digital image watermarking that applies a neural network to train an end-to-end model. The algorithmic network allows the adaptive balancing of watermark properties, such as capacity, confidentiality and robustness, to different types of noise by adjusting model parameters or noise layers during training. HiDDeN presents excellent advantages in robust digital image watermarking because new attacks can be directly added to the training process to improve the generalization of the algorithm. The model can robustly hide all kinds of image deformations by considering several different types of image attack layers, unchanged watermarked images, dropouts, crop-outs, cropping, Gaussian and JPEG.
StegaStamp [11] also presented an end-to-end algorithm based on deep learning. StegaStamp learns the watermark information embedding and extracts model resistant image perturbations to simulate the distorted space produced by real printing and photography, including changes in illumination, shading, movement, perspective, rotation, occlusion and viewing distance. Specifically, the researchers simulated five different types of attack layers, including perspective distortion, blurring, color manipulation, Gaussian noise and JPEG, and applied various kinds of image distortions. Finally, robust information extraction performance over multiple printers, screen-shooting and cameras was demonstrated in the experiments.
ReDMark [3] developed a deep end-to-end propagation watermarking algorithm that can flexibly transform the space and learn a new watermarking model. The model consists of CNN with residual blocks to achieve watermark embedding and blind extraction in the simulation environment. The whole network system consists of three main modules: embedding watermark convolution layers, simulating differentiable attack layers and blind extracting watermark convolutional layers. The developed model simulated multiple types of image attacks as a differentiable attack layers to benefit end-to-end training. The watermarking information can be spread in a wide area of the image to enhance the security and robustness of the watermarking algorithm.
StegaStamp used CNN networks that learn a robust encoding and decoding algorithm in the spatial domain and simulate the distortions caused by physically displaying and imaging and image perturbations to train the network for outstanding resistance to screen-shooting attacks. Many researchers proposed the watermarking-based spatial domain and presented weak robustness. ReDMark proposes a deep end-to-end diffusion watermarking framework, which can learn watermarking algorithms in the frequency domain space. However, ReDMark is unable to resist the screen-shooting attacks. The proposed hybrid ReDMark framework and screen-shooting attacks, similar to StegaStamp’s distortion [11], is used to train a robust screen-shooting-resilient watermarking.

2.3. Screen-Shooting Performance Influencing Factors and Capture Device Specifications

Screen-shooting attacks come from a variety of sources. We analyzed the impact of attacks on real shooting situations. We simulate attack layers on the basis of different cameras, displays and shooting environments using neural networks.
The attributes of display monitors exert a serious impact on the quality of the captured images. Some examples of monitor brands are, “Dell”, “Samsung”, “LG”, “ASUS”, “HP”, “Huawei”, “Acer”, “BenQ”, “MSI”, “Philips”, “Lenovo” and “AOC”. However, there are general attributes, such as luminosity, the color spectrum of the emitted light, the resolution of the screen and display equipment, that differ for each band.
Phone models also present an impact on the quality of the captured images. Many models, such as “SAMSUNG”, “APPLE”, “HUAWEI”, “NOKIA”, “SONY”, “LG”, “HTC”, “XIAOMI” and “LENOVO”, are available on the market. Different phone models use various sensors, and corresponding perspective and lens attacks may occur when the image displayed on the screen is captured by the camera equipment and then projected onto a different sensor. Furthermore, the camera device can generate nonlinear lens distortions when the sensor chip is not completely aligned with the imaging and cause radial and tangential lens attacks. The influence of perspective and lens attacks is very important because these attacks can modify the images and watermarks and further influence the extraction watermark due to the missing critical synchronization information. In addition, barrel attack was caused by the distance of the captured image and the angle of view of the image.
We create a dataset of moiré image pairs using many different combinations of video cameras and displaying devices to fine-tune the model and simulate other screen-shooting attacks in the framework of the watermarking algorithm to address the affect of device differences on screen-shooting resilient watermarking framework.

2.4. Multiple Real Leak Cases

Taking photos has become an efficient way of transmitting information with the prevalence of smart phones. Anyone can access files and steal information by taking photos without leaving any records. Several leaked cases, including photographs with the front view, vertical perspective angle and horizontal perspective angle, as well as photos taken from a far distance, are shown in Figure 1.

3. Method

We discuss screen-shooting-resilient watermarking via deep networks in the DCT domain in this section. The flowchart of the algorithm is shown in Figure 2. The proposed end-to-end framework can achieve blind robust watermarking for gray images in the DCT domain, and simultaneously simulate a differentiable attack layers as part of the framework to implement robust watermarking resistance from various screen-shooting attacks. We will describe the framework and its capabilities in detail.

3.1. Embedding Component

The embedding component can embed a watermark into the original image to minimize the perceptual difference between the original and the watermark images and improve the imperceptibility and security of the watermarked image. The component is composed of DCT transform layers, convolutional layers and IDCT transform layers, responsible for embedding the watermark into the transform coefficients of the images. We use a CNN architecture that receives a one-channel input of 512 × 512 pixels and outputs a gray watermarked image. The input watermark is represented as a 32 × 32 bit binary, and transformed image blocks are spliced with the watermark image. Specific procedures of watermark embedding is described in Algorithm 1.
Algorithm 1:Embedding algorithm.
Input: Image I o size of M × M , watermark w o size of h × g , strength factor α .
Output: Watermarked image I w .
1: Divide input image I o size of M × M into blocks of size h × g and each block will be embedding at least one watermark bit. Then we can obtain a tensor of size b × h × g .
2: Input the tensor into the DCT transform layer and output DCT coefficients tensor.
3: Concatenate the output of the DCT transform layer with the watermark, shaping the input tensor of size ( b + 1 ) × h × w as the input of the embedding network.
4: Get the output of embedding layer, that is watermarked DCT coefficient map.
5: Calculate the residual with watermarked DCT coefficient by the inverse transform layer.
6: Add the residual to the input image I o with a strength factor α weight, finally we can obtain watermarked image I w .

3.2. Transform Sub-Network

We insert the DCT transform and inverse DCT transform layers in our network framework to transform the image/coefficient and obtain the DCT coefficient feature map/image. f ( x , y ) is the M × M pixel matrix, and the specific practice details of DCT transform are presented as follows:    
D ( μ , ν ) = x = 0 M 1 y = 0 M 1 f ( x , y ) F ( x , y ) .
F ( x , y ) = α μ × cos ( 2 × y + 1 2 × M π × x ) .
where α μ = 1 M , μ = 0 ,   α μ = 2 M , μ 0 . D ( μ , ν ) is the discrete cosine transform coefficient corresponding to the pixel matrix f ( x , y ) .
Inverse DCT (IDCT) transform is as follows:
f ( x , y ) = x = 0 M 1 y = 0 M 1 D ( μ , ν ) F T ( x , y ) .
where f ( x , y ) is the pixel matrix of IDCT. The DCT transform layer is embedded in the network to realize the image space transform, and the watermark is embedded in the frequency domain to improve the security, robustness and ability to spread the watermark in a wide area of the image.

3.3. Screen-Shooting Attack Component

The screen-shooting operation can be perceived as a cross-media transmitting information procedure. This operation is defined as the process of displaying the watermarked image on the screen and then shooting the screen image with a camera phone. The watermarked image can be subjected to some attacks due to lens distortion, camera properties, display effects, the shooting scene, and desynchronization in this process. The image contains screen-shooting special attacks, three-dimensional distortion and other traditional deformations, including geometric deformation and image processing. Thus, the screen shooting process can be viewed as a collection of various attacks.

3.3.1. Perspective Transform

Screen-shooting photos are susceptible to distortion due to the shooting angle as well as the distance between a camera and a screen. This distortion results in the asynchrony of the watermark information, that is, the watermark’s position is shifted. Two images in the same plane are associated by homography to strengthen the robustness of the algorithm resisting perspective transformation attacks. Specifically, we randomly modify the four angular positions within a prescribed range and then solve for the homography that maps the original angle to the new position when sampling the homography.

3.3.2. Optical Distortion

Optical deformations are caused by the optical design of the lens. These deformations occur when special lens elements are used to reduce spherical aberrations and other aberrations. Briefly, optical deformations are the error of lens distortion, and the distortion causes the image to become concave or convex.

3.3.3. Light Distortion

Light distortion is caused by the deformation of light and solar eruptions. Light distortion will manifest itself when the camera takes a picture when a bright light or sunburst shines into the camera because the lighting of the shooting scene environment is difficult to control.

3.3.4. JPEG Distortion

The mobile shooting device will compress the image into a JPEG file during shooting to reduce the storage space that the image occupies. We consider the JPEG attack in the process of training the network to ensure that the compressed JPEG image can still extract the watermark.

3.3.5. Moiré Pattern

The moiré mode is rendered when the scene contains repeated details and exceeds the resolution of the sensor. As a result, photos that were taken demonstrate wavy patterns. Specific images are shown in Figure 3.

3.3.6. Screen Resolution

Different screen resolutions affect the quality of the shot image. Low resolutions will enhance the attack on the shooting image, whereas high resolutions exert a relatively lower impact on the shooting image.
These types of screen shooting attacks were simulated in the network. Although we attempt to reduce other attacks, except device and environment distortions, during the experiment, some effects, such as the special properties of display devices that we ignored in the simulation, were inevitable. In addition, illumination condition, camera positioning and focusing range were different.
In the attack layer, a variety of attacks can be combined to train a robust watermarking network to adapt to the mixture of multiple attacks. The training process is slightly different because in each training iteration, the network randomly selects an attack. Therefore, only one attack is allowed to affect the training loss in each training iteration.

3.4. STN

The spatial transformation network (STN) is an adaptive image transformation and alignment network that does not need to calibrate key points. Thus, STN can reduce the influence of geometric transformations, such as rotation, translation, scale, and distortion on the watermark extraction task. Joining an existing network can improve the learning ability of the network.

3.5. Extracting Component

We can map the distorted image with the corrected image and then extract the watermark from the corrected image after applying STN. This component extracts the watermark from the watermarked image with screen shooting attacks. The watermarked image is transformed into the DCT domain before the watermark is extracted given that the watermark is embedded in the DCT domain. Since the watermark is embedded in the transform domain, the extraction component contains a copy of the transform layer used to embed the module in order to represent the watermarked image on the same basis. Meanwhile, other network layers learn to extract watermarks in the dct domain. The specific steps of the watermark extraction procedure is described in Algorithm 2.
Algorithm 2:Extracting algorithm.
Input: Attacked image I a size of M × M .
Output: Watermark w e .
1: Apply spatial transformation network to the attacked image I a size of M × M and output corrected image.
2: Divide corrected image size of M × M into blocks of size h × g and obtain a tensor of size b × h × g .
3: Input the tensor into the DCT transform layer and output DCT coefficients tensor.
4: Fed the output of DCT transform layer into extracting layers, and then obtain watermark w e .

3.6. Estimation Criteria

The framework guarantees a watermarked image with maximum quality and minimum watermark extraction error rate to balance the imperceptibility and robustness of the algorithm. Peak signal-to-noise ratio ( P S N R ) [21], structural similarity index ( S S I M ) [3] and normalized cross-correlation ( N C ) [36] are used in this work to evaluate the capability of the SSDeN algorithm. P S N R and S S I M are applied to show the algorithm’s imperceptibility performance, and N C is utilized to demonstrate the robustness of watermarking. The mean square error ( M S E ) of the image is expressed as follows:
M S E = 1 M × M x = 0 M 1 y = 0 M 1 ( I o ( x , y ) I w ( x , y ) ) 2 .
The P S N R [37] is defined as follows:
P S N R = 10 log 10 M A X 2 M S E .
where I o ( x , y ) is the host image, I w ( x , y ) is the watermarked image, M A X is the max pixel value of the image:
L I = S S I M ( I o , I w ) = ( 2 μ I 0 μ I w + C 1 ) ( 2 σ I o I w + C 2 ) ( μ I o 2 + μ I w 2 + C 1 ) ( σ I o 2 + σ I w 2 + C 2 ) .
where I o and I w denote the original and watermarked images, respectively; μ I o and μ I w are the average values of I o and I w , respectively; σ I o and σ I w are the variances of I o and I w , respectively; σ I o I w is the covariance of I o and I w ; and C 1 and C 2 are two variables for stabilizing the division with the weak denominator.
Robustness is used to evaluate the ability of detecting or extracting the watermark from watermarked images after normal image processing or malicious attack. The robustness of the watermarking algorithm is high when the extracted watermark is close to the original watermark. Normalized cross-correlation ( N C ) can be adopted to measure the similarity between the original and the extracted watermarks given that the watermark is a binary image. The value range of N C is [0, 1]. The robustness of image watermarking improves as the value of N C approaches 1. We choose N C as part of the loss function. N C is defined as follows:
L w = N C = h = 1 H g = 1 G w e ( h , g ) × w o ( h , g ) ( h = 1 H g = 1 G ( w e ) 2 ) ( h = 1 H g = 1 G ( w o ) 2 )
where w 0 ( h , g ) is the original watermark, w e ( h , g ) is the extraction watermark and w * ( h , g ) is a watermark with size H × G .
Finally, the total training loss can be calculated as shown in Equation (8):
L t = α 1 L I + C 3 + β 1 L w + C 4
where α and β are the ratios of losses, α + β = 1 and C 3 and C 4 are two constants of the metric, which are set to 10 4 and 9 × 10 4 , respectively.

4. Experiments

The performance of the proposed algorithm is verified using series of experiments, Note that only some representative experimental results are presented. MATLAB R2019b and Python VScode are used in this study as the experimental platforms.
Imperceptibility and robustness are two important evaluation criteria of the watermarking algorithm. As mentioned in Section 3.6, PSNR and SSIM are used to evaluate the quality difference between the original and watermarked images. A large PSNR or SSIM corresponds to enhanced perception quality. The robustness of the algorithm is measured by calculating bit error rate (BER) and NC values. A small BER value or large NC value indicates high robustness. The size of the original image in our experiment is 512 × 512 , and the size of the watermark image is 32 × 32 , that is, 1024 bits of binary data.

4.1. Simulation Moiré Attack Experiments

The moiré phenomenon often occurs in photographs taken on a computer or television screen with a moving camera. The moiré pattern seriously damages the visual quality of photos and then affects the accuracy of watermark extraction. We train a network for simulating a moiré attack to enhance the robustness of the network. Sun [38] proposed an algorithm that can automatically remove moiré patterns from photographs and obtain pure images using a multi-resolution, fully convolutional network. This objective [38] actually works against our goals. We train a U-Net using a dataset containing 100,000 + images pairs in [38] to simulate the moiré attack layer. Detailed parameters and configurations of the U-Net are listed in Table 1, Table 2 and Table 3. The training process feeds a pair of images into the moiré pattern attack network, the input is the original image without the moiré pattern attack and the output is the image after the moiré pattern attack. The results of the test experiment are shown in Figure 4 and Figure 5.
Sun [38] built a base library of 135,000 image pairs, with each pair containing an image with the moiré pattern and its corresponding pure image. The image undergoes a series of transformations subjected to a combination of strong attacks, such as color transformations, optical distortion and moiré, when shooting the image displayed on the screen. Hence, the result fails to control a single variable in experiments. Furthermore, the gamut of the monitor is limited compared with the full RGB color space. Cameras use exposure settings, white balance and color correction matrices to modify their output. We demonstrated that the network learns these perturbations and generates images with color manipulation. Some test results are shown in Figure 5.
Figure 4 and Figure 5 show that the moiré pattern of the standard dataset is weak. Therefore, we build the moiré pattern data set to train the fine-tuned network. We positioned the camera on a tripod to facilitate alignment and steadying. Uncompromised reference images in our benchmark dataset are from the training and test images of the COCO 2014 dataset. We acquired sample pairs with different combinations of phone models and screens to ensure that the capture device shot the moiré pattern on different optical sensors and a diversity of screen resolutions. Phone models were “iPhone 8 Plus”, “(MI) Redmi K30” and “HUAWEI nova 8 Pro”, and display monitors were “MacBook Pro 14 (2021)”, “DELL E2216H” and “AOC E2270SWN5”. The training results of the U-Net are presented in Figure 6 and Figure 7. All images used in the experiment were colored, and the proposed watermarking algorithm uses gray-scale images as input images. Therefore, we converted the color image data set of COCO 2014 into gray scale, which was used as the input of the test, and fed it into the moiré network to obtain the texture map with moiré, as shown Figure 8.
The moiré network training criterion L m [38] is as follows:
L m = 1 m i = 1 m ( y i f ( x i ) ) 2
where m is the number of samples, y i is the ith ground truth, f ( x i ) is the ith predicted value and L m is the mean square deviation between the estimated and actual values. It is a positive value and decreases as the error approaches zero because the value is calculated from the square of the Euclidean distance. The detailed training loss values of the training and validation dataset of the moiré network are presented in Figure 9 and Figure 10, respectively. We also measure the PSNR of the test image pair experiments, as shown in Figure 11. PSNR fails to measure the effectiveness in the moiré pattern precisely, and an image corrupted by visually severe moiré patterns can demonstrate high PSNR. However, PSNR can measure the overall image quality and the moiré experimental results remain unaffected.
We insert the moiré attack layer between the embedding and extracting processes to improve the robustness of screen-shooting resilient watermarking and then add a variety of attack layers.

4.2. Optical Effects of Various Simulation Attacks

We show the results of simulation attacks in this part. Various attacks simulated for the attack layer are briefly explained. These distortions are generally classified into two categories: normal and special attacks.
Normal attacks (NA). (A). JPEG (parameters: 30, 50 and 60). Photographs are usually saved in a JPEG compression after being shot, and JPEG compression will cause image loss to the captured photographs. (B). Noise layer. (Gaussian noise: 1, 2, 3, , 10). Image noise represented as random blobs on a smooth surface can severely affect image quality. The intensity of the noise depends on exposure time, camera settings and the temperature. (C). Cropping attack (parameters: 0.06, 0.11, 0.16, 0.26 and 0.55). Only some images on the display will be captured as full images due to the shooting environment. (D). Blurring (parameters: 1, 1.2, 1.6, 2 and 2.8). The inconsistent bending of light leads to blurred images, and motion blur and defocus blurs are the common blur types of blur that degrade image quality. (E). Rotation (parameters: −5, −2, 5, 8 and 10), scaling (parameters: 0.25, 0.5, 1.25, 1.5 and 2) and translation (RST) distortions. RST distortions are common traditional geometric attacks. (F). Brightness and contrast (brightness = 0.3, contrast = 1.5). The whole range of tones in the image will rise or fall accordingly when the brightness is corrected. Midtones are eliminated and images will present a high percentage of dark or black and white or highlights with minimal midtones when the contrast is high. (G). Sharpening (1, 4, 12, 25 and 30). Sharpening is a distortion that increases the apparent sharpness of an image.
Special attacks (SA). The image suffers from space conversion processes and lead to a combination of strong attacks, such as perspective deformation, moiré (U-Net), lens distortion, visible watermark (parameters: 1, 3, 9, 12, 15 and 16), light (brightness and contrast), color changing (hue = 0.1, luminance = 0.5) and display distortion (scaling), when an image displayed on the screen is shot.
We provide attacked images, the specific results of the attack are illustrated in Figure 12.

4.3. Quantitative Analysis

We choose 25 images ( 512 × 512 ) from the Granada dataset for our test experiment. The fidelity of each watermarked image is evaluated from PSNR and SSIM criteria. Figure 13 shows the PSNR of watermarked images generated by all trained networks with fixed strength factors ( α = 1). The average of PSNR is 36.094852.
The robustness of a watermarking algorithm can be determined using two approaches. One approach is to enhance the watermark embedding strength. High-intensity embedding increases the robustness but reduces the PSNR. We balance the robustness and imperceptibility of the trained network by adjusting the watermark strength factor ( α = 1). The bit error rate of the watermark is calculated under a variety of attacks with fixed strength factors to prove the robustness of the proposed framework. The other approach is to increase the diversity of image attack types to enhance the robustness of the framework, as shown in Table 4. Our framework can effectively embed the watermark and avoid various image distortions. Figure 14 illustrates the original and watermarked images as well as the difference image between the original image and the watermarked images with an enhancement of 50×.

4.4. Qualitative Analysis

Watermarked DCT coefficients produced by the framework are illustrated in Figure 14. A random 1024-bit ( 32 × 32 ) watermark is embedded in the DCT domain of the ‘Cat’, ‘Barbara’ and ‘Baboon’ images from the Granada dataset using the proposed framework. The absolute difference between the original and the watermarked images is used to display the watermark distribution. This difference is multiplied by 10, 20, 50, 80 and 100 to improve the visualization. In addition, a small area of the image is enlarged to facilitate the presentation. We revealed that the amplitude of the resulting artifact varies in different regions of the original DCT coefficients in diverse images. The variation in the difference matrix indicates that the watermark symbol is adaptively embedded on the basis of local features of the image.
Analysis of the diffusion pattern of the DCT domain and its frequency energy map demonstrated that SSDeN adopts a distributed embedding policy concentrated in the middle- and low-frequency bands. As demonstrated in Figure 15 and Figure 16, Figure 16 presents the 3D plot of frequency energy. The frequency energy is calculated using the absolute values of DCT coefficients of an 8 × 8 block of the diffusion watermark. DCT coefficients are arranged on the horizontal axis in zigzag order of the DCT block. For example, we start from the DC coefficient DCT(0, 0) toward the maximum frequency DCT(7, 7). The framework learns to avoid embedding in the DC coefficient due to destructive effects on image quality and PSNR.
According to the Table 4, the proposed framework achieves an acceptable trade-off between imperceptibility and robustness. The experimental results showed that the proposed algorithm is superior to existing advanced schemes.

4.5. Comparison with State-of-the-Art

We exhibit and discuss the comparative experimental results and compare them with other schemes in this section [2,3,11,12,13,39]. The comparison of watermarked image quality between our scheme and existing schemes, including PSNR, network architecture, watermark embedding domain and published time, is presented in Table 5.
As Table 5 shows, in recent years, many screen-shooting watermarking methods have been presented [2,3,11,12,13,39]. Fang et al. [2] and Nakamura et al. [39] were using traditional methods for resistance to screen-shooting attacks; they have good image quality, but the registration point is unstable and may not be able to extract the watermark. As deep learning develops, deep-learning-based watermarking algorithms have been widely studied in various network structures [3,11,12,13], including U-Net, GAN, and CNN. For robust watermarking, HiDDeN [3] is, to our knowledge, the first end-to-end method using neural networks in the spatial domain, but it could not resist the screen-shooting attack. StegaStamp [11] was the first to propose a machine learning screen-shooting attack algorithm in the spatial domain. The algorithm has good robustness, but the image quality is slightly worse. ReDMark [3] was the first robust watermarking using deep learning in the frequency domain, and it has relatively good image quality; it also could not resist the screen-shooting attack.
Table 5 displays that there have been few frequency-domain watermarking algorithms based on machine learning proposed in recent years, except ReDMark [3]. Watermarking algorithms in the frequency domain can help improve the imperceptibility of watermarking. Frequency-domain watermarking algorithms can guarantee enhanced imperceptibility and robustness of embedding after various attacks. Therefore, we investigate watermarking in the frequency domain via deep learning to improve the performance and efficiency in balancing the relationship between robustness and imperceptibility. However, the framework is limited by its unsuitability for color images.
We also designed experiments to test the robustness of our networks and other existing schemes. The results are listed Table 6; (*) is the parameters used for each type of attack.
The performance of the proposed method is illustrated by comparing with other deep-learning-based watermarking methods [3,11,12,13], and the advantages of the proposed method are illustrated by experimental results.
StegaStamp [11] proposed an end-to-end watermarking scheme for the print-camera process in the spatial domain. This method demonstrated very high robustness to print-camera distortions. Table 6 shows that StegaStamp has a good performance against perspective and light. However, the method has a poor performance against cropping and JPEG. The overall robust performance is considered to be positive in the spatial domain.The watermark information in ReDMark [3] is spread in a wide area of the image. ReDMark performs well against cropping, scaling, noise and visible watermarking. Compared with the watermarking method based on the spatial domain, the watermarking method based on the transform domain is more robust. Although this method can resist traditional attacks robustly, it fails to resist screen-shooting attacks. We add the screen-shooting attack layer to train our model and improve the robustness. The results showed that our model presents strong resistance to screen-shooting attacks and traditional attacks. HiDDeN [13] fed augmentations into their training process to increase the robustness against several traditional attacks. The researchers also proposed the watermarking-based spatial domain and, compared with frequency-domain-based methods, presented weak robustness. From various attack results, this method has no outstanding advantage. However, HiDDeN [13] was the first end-to-end robust method using neural networks and provided lessons on deep-learning-based watermarking. The two-stage separable deep learning (TSDL) [12] scheme for robust watermarking is composed of noise-free end-to-end adversary training and noise-aware decoder-only training and robust to several black-box noises. TSDL excels in blurring and color changing for color image. The spatial-domain-based watermarking usually perform poorly in JPEG and cropping, so the same goes for TSDL, and it is not resistant to screen-shooting attacks.
Hence, these experiments demonstrated that the proposed SSDeN framework is robust not only to common traditional attacks but also to screen-shooting attacks.

5. Conclusions

We propose an end-to-end, screen-shooting-resilient image watermarking framework using convolution neural network in this study. Our framework consists of four components, namely, embedding, transform sub-network, simulating attack layers and extracting components. The embedding component is used to embed the watermark into the DCT coefficients of images. The transform sub-network modifies the image to obtain the DCT coefficient feature of the image. The simulating attack layer is designed to simulate a series of distortions, such as perspective transform, optical distortion, light distortion, JPEG distortion and moiré pattern, in the screen-shooting process. We design a moiré attack network for simulating the moiré phenomenon, which is the most common screen-shooting attack, to improve the resiliency of the watermarked image against distortions in real screen-shooting scenes. The extracting component extracts the watermark from the captured photos. In order to balance the robustness and imperceptibility of the algorithm, we introduce a loss function based on NC and SSIM. From the experiments, it can be concluded that SSDeN employs a distributed embedding strategy with greater concentration on low- and middle-frequency coefficients in keeping with the traditional method’s results. Simultaneously, the above shows the rationality of our proposed framework.
Our follow-up investigation will continue to improve and extend SSDeN. Specifically, we will simulate additional capture devices and increasingly complex shooting scenes to improve watermark detection in real-world scenarios. Screen-shooting-resilient watermarking can also be extended for color images while fully fitting all real screen-shooting scenarios.

Author Contributions

Conceptualization, S.Z. and R.B.; methodology, L.L.; software, R.B.; validation, R.B., L.L. and S.Z.; formal analysis, C.-C.C.; investigation, R.B.; resources, R.B.; data curation, R.B.; writing—original draft preparation, C.-C.C.; writing—review and editing, L.L.; visualization, J.L.; supervision, R.B.; project administration, R.B.; funding acquisition, L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62172132) and the Public Welfare Technology Research Project of Zhejiang Province (No. LGF21F020014).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Thank you to the reviewers who reviewed this paper and the MDPI editor who edited it professionally.

Conflicts of Interest

The authors declare no conflict of interest.

Sample Availability

Not applicable.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional neural network
DCTDiscrete cosine transform
IDCTInverse discrete cosine transform
QDFTQuaternion discrete Fourier transform
FRFSFeature regions filtering model to SuperPoint
CDTFCamera-display transfer function
UDHUniversal deep hiding
JPEGJoint photographic experts group
DNNDeep neural network
TERATransparency, efficiency, robustness and adaptability
NCNormalized correlation
SSIMStructural similarity index
DWTDiscrete wavelet transform
DFTDiscrete Fourier transformation
SCISingle-channel images
MCIMulti-channel images
SVDSingular value decomposition
FRATFinite radon transform
LS-SVMLeast squares support vector machines
QRQuadrature rectangle decomposition
DLDeep learning
PSNRPeak signal-to-noise ratio
MSEMean square error
NANormal attacks
RSTRotation, scaling and translation
SASpecial attacks
BERBit error rate
TSDLTwo-stage separable deep learning
TDTensor decomposition
SIFTScale invariant feature transform

References

  1. Baluja, S. Hiding Images in Plain Sight: Deep Steganography. Adv. Neural Inf. Process. Syst. 2007, 30. [Google Scholar]
  2. Fang, H.; Zhang, W.; Zhou, H.; Cui, H.; Yu, N. Screen-Shooting Resilient Watermarking. IEEE Trans. Inf. Forensics Secur. 2019, 14, 1403–1418. [Google Scholar] [CrossRef]
  3. Ahmadi, M.; Norouzi, A.; Karimi, N. ReDMark: Framework for Residual Diffusion Watermarking based on Deep Networks. Expert Syst. Appl. 2019, 146, 113157. [Google Scholar] [CrossRef]
  4. Li, L.; Bai, R.; Zhang, S.; Chang, C.C.; Shi, M. Screen-shooting resilient watermarking scheme via learned invariant keypoints and QT. Sensors 2021, 21, 6554. [Google Scholar] [CrossRef]
  5. Wengrowski, E.; Dana, K. Light field messaging with deep photographic steganography. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1515–1524. [Google Scholar]
  6. Zhang, C.; Lin, C.; Benz, P.; Chen, K.; Zhang, W.; Kweon, I.S. A brief survey on deep learning based data hiding, steganography and watermarking. arXiv 2021, arXiv:2103.01607. [Google Scholar]
  7. Zhang, C.; Benz, P.; Karjauv, A.; Sun, G.; Kweon, I.S. Udh: Universal deep hiding for steganography, watermarking, and light fieldmessaging. Adv. Neural Inf. Process. Syst. 2020, 33, 10223–10234. [Google Scholar]
  8. Fang, H.; Chen, D.; Huang, Q.; Zhang, J.; Ma, Z.; Zhang, W.; Yu, N. Deep template-based watermarking. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1436–1451. [Google Scholar] [CrossRef]
  9. Fang, H.; Chen, D.; Wang, F.; Ma, Z.; Liu, H.; Zhou, W.; Zhang, W.; Yu, N. Tera: Screen-to-camera image code with transparency, efficiency,robustnessand adaptability. IEEE Trans. Multimed. 2021, 24, 955–967. [Google Scholar] [CrossRef]
  10. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  11. Tancik, M.; Mildenhall, B.; Ren, N. StegaStamp: Invisible Hyperlinks in Physical Photographs. In Proceedings of the Computer Vision and Pattern Recognition, Glasgow, UK, 23–28 August 2020; pp. 2114–2123. [Google Scholar]
  12. Liu, Y.; Guo, M.; Zhang, J. A Novel Two-stage Separable Deep Learning Framework for Practical Blind Watermarking. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019. [Google Scholar]
  13. Zhu, J.; Kaplan, R.; Johnson, J.; Li, F.-F. Hidden: Hiding data with deep networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 657–672. [Google Scholar]
  14. Tan, Y.; Qin, J.; Xiang, X.; Ma, W.; Pan, W.; Xiong, N.N. A Robust Watermarking Scheme in YCbCr Color Space Based on Channel Coding. IEEE Access 2019, 7, 25026–25036. [Google Scholar] [CrossRef]
  15. Chou, C.; Liu, K. A Perceptually Tuned Watermarking Scheme for Color Images. IEEE Trans. Image Process 2010, 19, 2966–2982. [Google Scholar] [CrossRef] [PubMed]
  16. Tsui, T.K.; Zhang, X.; Androutsos, D. Color Image Watermarking Using the Spatio-Chromatic Fourier Transform. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006. [Google Scholar]
  17. Loan, N.A.; Hurrah, N.N.; Parah, S.A.; Lee, J.W.; Sheikh, J.A.; Bhat, G.M. Secure and Robust Digital Image Watermarking Using Coefficient Differencing and Chaotic Encryption. IEEE Access 2018, 6, 19876–19897. [Google Scholar] [CrossRef]
  18. Barni, M.; Bartolini, F.; Piva, A. Multichannel watermarking of color images. IEEE Trans. Circuits Syst. Video Technol. 2002, 12, 142–156. [Google Scholar] [CrossRef]
  19. Chen, B.; Coatrieux, G.; Chen, G.; Sun, X.; Coatrieux, J.L.; Shu, H. Full 4-D quaternion discrete Fourier transform based watermarking for color images. Digit. Signal Process 2014, 28, 106–119. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, C.; Wang, X.; Zhang, C.; Xia, Z. Geometric correction based color image watermarking using fuzzy least squares support vector machine and Bessel K form distribution. Signal Process 2017, 134, 197–208. [Google Scholar] [CrossRef]
  21. Wang, X.; Wang, C.; Yang, H.; Niu, P. A robust blind color image watermarking in quaternion Fourier transform domain. J. Syst. Softw. 2013, 86, 255–277. [Google Scholar] [CrossRef]
  22. Li, J.; Yu, C.; Gupta, B.B.; Ren, X. Color image watermarking scheme based on quaternion Hadamard transform and Schur decomposition. Multimed. Tools Appl. 2018, 77, 4545–4561. [Google Scholar] [CrossRef]
  23. Guan, H.; Zeng, Z.; Liu, J.; Zhang, S. A novel robust digital image watermarking algorithm based on two-level DCT. In Proceedings of the 2014 International Conference on Information Science, Electronics and Electrical Engineering, Sapporo, Japan, 26–28 April 2014; pp. 1804–1809. [Google Scholar]
  24. Li, C.; Song, X.; Liu, Z. A Robust Watermarking Scheme Based on Maximum Wavelet Coefficient Modification and Optimal Threshold Technique. J. Electr. Comput. Eng. 2015, 2015, 44. [Google Scholar] [CrossRef]
  25. Makbol, N.M.; Khoo, B.E.; Rassem, T.H.; Loukhaoukha, K. A new reliable optimized image watermarking scheme based on the integer wavelet transform and singular value decomposition for copyright protection. Inf. Sci. 2017, 417, 381–400. [Google Scholar] [CrossRef]
  26. Rastegar, S.; Namazi, F.; Yaghmaie, K.; Aliabadian, A. Hybrid watermarking algorithm based on Singular Value Decomposition and Radon transform. Int. J. Electron. Commun. 2011, 65, 658–663. [Google Scholar] [CrossRef]
  27. Lai, C.; Tsai, C. Digital image watermarking using discrete wavelet transform and singular value decomposition. IEEE Trans. Instrum. Meas. 2010, 59, 3060–3063. [Google Scholar] [CrossRef]
  28. Ma, X.; Xu, Y.; Song, L.; Yang, X. Color image watermarking using local quaternion Fourier spectral analysis. In Proceedings of the IEEE International Conference on Multimedia and Expo, Hannover, Germany, 23 June–26 April 2008; pp. 233–236. [Google Scholar]
  29. Rouis, K.; Gomez-Krämer, P.; Coustaty, M. Local Geometry Analysis For Image Tampering Detection. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 2551–2555. [Google Scholar]
  30. Yang, H.; Wang, Y.Z.P.; Wang, X.; Wang, C. A geometric correction based robust color image watermarking scheme using quaternion Exponent moments. Optik 2014, 125, 4456–4469. [Google Scholar] [CrossRef]
  31. Li, M.; Yuan, X.; Chen, H.; Li, J. Quaternion Discrete Fourier Transform-Based Color Image Watermarking Method Using Qurternion QR Decomposition. IEEE Access 2020, 8, 72308–72315. [Google Scholar] [CrossRef]
  32. Xu, H.; Jiang, G.; Yu, M.; Luo, T. A Color Image Watermarking Based on Tensor Analysis. IEEE Access 2018, 6, 51500–51514. [Google Scholar] [CrossRef]
  33. Feng, B.; Lu, W.; Sun, W.; Huang, J.; Shi, Y.Q. Robust image watermarking based on Tucker decomposition and Adaptive-Lattice Quantization Index Modulation. Signal Process. Image Commun. 2016, 41, 1–14. [Google Scholar] [CrossRef]
  34. Hai, F.; Quan, Z.; Kaijia, L. Robust Watermarking Scheme for Multispectral Images Using Discrete Wavelet Transform and Tucker Decomposition. J. Comput. 2013, 8, 2844–2850. [Google Scholar]
  35. Delgado-Guillen, L.A.; Garcia-Hern, J.J.; Torres-Huitzil, C. Digital watermarking of color images utilizing mobile platforms. In Proceedings of the 2013 IEEE 56th International Midwest Symposium on Circuits and Systems (MWSCAS), Columbus, OH, USA, 4–7 August 2013. [Google Scholar]
  36. Li, L.; Bai, R.; Lu, J.; Zhang, S.; Ching, C. A Watermarking Scheme for Color Image Using Quaternion Discrete Fourier Transform and Tensor Decomposition. Appl. Sci. 2021, 11, 5006. [Google Scholar] [CrossRef]
  37. Setiadi, D.R.I.M. PSNR vs. SSIM: Imperceptibility quality assessment for image steganography. Multimed. Tools Appl. 2021, 80, 8423–8444. [Google Scholar] [CrossRef]
  38. Sun, Y.; Yu, Y.; Wang, W. Moiré Photo Restoration Using Multiresolution Convolutional Neural Networks. IEEE Trans. Image Process. 2018, 27, 4160–4172. [Google Scholar] [CrossRef] [Green Version]
  39. Nakamura, T.; Katayama, A.; Yamamuro, M. Fast Watermark Detection Scheme from Analog Image for Camera-Equipped Cellular Phone. EICE Trans. Inf. Syst. 2004, 87, 2145–2155. [Google Scholar]
Figure 1. Real leak cases: (a) original image, (b) screen-shooting image with front view, (c) screen-shooting image with vertical perspective angle (up), (d) screen-shooting image with horizontal perspective angle (right), (e) screen-shooting image with vertical perspective angle (down), (f) screen-shooting image with horizontal perspective angle (left).
Figure 1. Real leak cases: (a) original image, (b) screen-shooting image with front view, (c) screen-shooting image with vertical perspective angle (up), (d) screen-shooting image with horizontal perspective angle (right), (e) screen-shooting image with vertical perspective angle (down), (f) screen-shooting image with horizontal perspective angle (left).
Applsci 12 09780 g001
Figure 2. The flowchart of the SSDeN framework.
Figure 2. The flowchart of the SSDeN framework.
Applsci 12 09780 g002
Figure 3. (Row 1): Pure images. (Row 2): Images with moiré pattern. (Row 3): A region of the image with moiré pattern and its amplified.
Figure 3. (Row 1): Pure images. (Row 2): Images with moiré pattern. (Row 3): A region of the image with moiré pattern and its amplified.
Applsci 12 09780 g003
Figure 4. Moiré experiment 1. Examples of image pairs from our test experiment: (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Figure 4. Moiré experiment 1. Examples of image pairs from our test experiment: (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Applsci 12 09780 g004
Figure 5. Moiré experimental 2. Examples of image pairs from our test experiment: (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Figure 5. Moiré experimental 2. Examples of image pairs from our test experiment: (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Applsci 12 09780 g005
Figure 6. Moiré pattern test results from COCO 2014 dataset by the fine-tune network. (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Figure 6. Moiré pattern test results from COCO 2014 dataset by the fine-tune network. (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Applsci 12 09780 g006
Figure 7. Moiré pattern test results from [38] dataset by the fine-tune network. (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Figure 7. Moiré pattern test results from [38] dataset by the fine-tune network. (a,c) pure images; (b,d) images ( 512 × 512 ) with moiré pattern.
Applsci 12 09780 g007
Figure 8. Moiré pattern test results from COCO 2014 gray dataset by the fine-tune network. (a) pure images; (b) images ( 512 × 512 ) with moiré pattern.
Figure 8. Moiré pattern test results from COCO 2014 gray dataset by the fine-tune network. (a) pure images; (b) images ( 512 × 512 ) with moiré pattern.
Applsci 12 09780 g008
Figure 9. Training loss of moiré network.
Figure 9. Training loss of moiré network.
Applsci 12 09780 g009
Figure 10. Validation loss of moiré network.
Figure 10. Validation loss of moiré network.
Applsci 12 09780 g010
Figure 11. PNSR of the moiré network.
Figure 11. PNSR of the moiré network.
Applsci 12 09780 g011
Figure 12. Simulating image distortions for screen-shooting.
Figure 12. Simulating image distortions for screen-shooting.
Applsci 12 09780 g012
Figure 13. Test images (25) PSNR of SSDeN.
Figure 13. Test images (25) PSNR of SSDeN.
Applsci 12 09780 g013
Figure 14. The results of images watermark embedding and the residual for original and watermarked images.
Figure 14. The results of images watermark embedding and the residual for original and watermarked images.
Applsci 12 09780 g014
Figure 15. The results of the DCT coefficient: (Row 1) original images; (Row 2) DCT coefficient of original images; (Row 3) DCT coefficient of watermarked images; (Row 4) the residual for original and watermarked DCT coefficients.
Figure 15. The results of the DCT coefficient: (Row 1) original images; (Row 2) DCT coefficient of original images; (Row 3) DCT coefficient of watermarked images; (Row 4) the residual for original and watermarked DCT coefficients.
Applsci 12 09780 g015
Figure 16. DCT energy 3D plot of the framework.
Figure 16. DCT energy 3D plot of the framework.
Applsci 12 09780 g016
Table 1. Moiré simulating network parameters.
Table 1. Moiré simulating network parameters.
EpochLearning RateLayersBest LossCriterionBatchsize
1000.000110100MSE8
Table 2. Down-sampling Layers.
Table 2. Down-sampling Layers.
ScaleKernelStrideChannels
1 3 × 3 1 × 1 32
1 3 × 3 1 × 1 32
2 3 × 3 1 × 1 32
2 3 × 3 1 × 1 64
3 3 × 3 1 × 1 64
3 3 × 3 1 × 1 128
4 3 × 3 1 × 1 128
4 3 × 3 1 × 1 256
5 3 × 3 1 × 1 256
5 3 × 3 1 × 1 256
Table 3. Up-sampling Layers.
Table 3. Up-sampling Layers.
ScaleKernelStrideChannels
1 3 × 3    conv 1 × 1 3
2 4 × 4    deconv 2 × 2 32
3 × 3    conv 1 × 1 3
3 4 × 4    deconv 2 × 2 64
4 × 4    deconv 2 × 2 32
3 × 3    conv 1 × 1 3
4 4 × 4    deconv 2 × 2 128
4 × 4    deconv 2 × 2 64
4 × 4    deconv 2 × 2 32
3 × 3    conv 1 × 1 3
5 4 × 4    deconv 2 × 2 256
4 × 4    deconv 2 × 2 128
4 × 4    deconv 2 × 2 64
4 × 4    deconv 2 × 2 32
3 × 3    conv 1 × 1 3
Table 4. For each attack, the attack parameter and the resulting extraction BER are shown for SSDeN.
Table 4. For each attack, the attack parameter and the resulting extraction BER are shown for SSDeN.
ImagesCatBarbaraBaboon
AttacksApplsci 12 09780 i001Applsci 12 09780 i002Applsci 12 09780 i003
Blurring (1.20)0.0150.0200.016
Cropping (0.06)0.0260.0200.021
Gaussian noise (8.00)0.0000.0020.001
Block noise (0.16)0.0110.0190.000
Scaling (2.00)0.00010.0040.001
Rotation (10.00)0.04830.05130.0480
JPEG (60.00)0.0200.0200.000
Sharpening (4.00)0.0000.0090.006
Moireé (U-Net)0.01790.02010.0103
Perspective ( θ = 5 )0.0360.0300.0298
Light (0.3, 1.5)0.00150.00260.0010
Visible watermark (16.00)0.0100.00910.0099
Table 5. Methods comparison.
Table 5. Methods comparison.
MethodsNetworkPSNRPublished TimeEmbedding Domain
StegaStamp [11]U-Net29.882020Spatial domain
HiDDeN [13]GAN33.552018Spatial domain
ReDMark [3]CNN35.932019Frequency domain
TSDL [12]GAN33.502019Spatial domain
Fang et al. [2]ISIFT and DCT42.302019Frequency domain
Nakamura et al. [39]Geometric distortion39.102004Spatial domain
The proposed (SSDeN)CNN36.092022Frequency domain
Table 6. Comparing the extracted watermark of the proposed algorithm and previous methods in terms of NC.
Table 6. Comparing the extracted watermark of the proposed algorithm and previous methods in terms of NC.
MethodsStegaStamp [11]HiDDeN [13]ReDMark [3]TSDL [12]The Proposed (SSDeN)
Cropping (10)0.76970.88230.98150.85010.9710
JPEG (50)0.79520.77310.97790.77700.9798
Blurring (1.2)0.91420.75530.78710.99560.9850
Scaling (2.0)0.9987NULL0.99950.92330.9800
Sharpening (4)NULLNULL0.99810.99001.0
Perspective0.9897NULLNULLNULL0.9640
Color changing0.9878NULLNULL0.9886NULL
Noise (10)0.95230.92120.98890.89951.0
Moiré (U-Net)0.8952NULLNULLNULL0.9926
Visible watermark (15)0.9700NULL0.99800.95670.9975
Light0.9994NULLNULL0.99290.9899
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bai, R.; Li, L.; Zhang, S.; Lu, J.; Chang, C.-C. SSDeN: Framework for Screen-Shooting Resilient Watermarking via Deep Networks in the Frequency Domain. Appl. Sci. 2022, 12, 9780. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199780

AMA Style

Bai R, Li L, Zhang S, Lu J, Chang C-C. SSDeN: Framework for Screen-Shooting Resilient Watermarking via Deep Networks in the Frequency Domain. Applied Sciences. 2022; 12(19):9780. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199780

Chicago/Turabian Style

Bai, Rui, Li Li, Shanqing Zhang, Jianfeng Lu, and Chin-Chen Chang. 2022. "SSDeN: Framework for Screen-Shooting Resilient Watermarking via Deep Networks in the Frequency Domain" Applied Sciences 12, no. 19: 9780. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199780

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop