Next Article in Journal
Design and Implementation of a Flexible Electromagnetic Actuator for Tunable Terahertz Metamaterials
Previous Article in Journal
Instant Candida albicans Detection Using Ultra-Stable Aptamer Conjugated Gold Nanoparticles
Previous Article in Special Issue
Coexisting Firing Patterns in an Improved Memristive Hindmarsh–Rose Neuron Model with Multi-Frequency Alternating Current Injection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MARR-GAN: Memristive Attention Recurrent Residual Generative Adversarial Network for Raindrop Removal

1
School of Electrical and Electronic Engineering, Changchun University of Technology, Changchun 130012, China
2
Changchun Changding Electronic Technology LLC, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Micromachines 2024, 15(2), 217; https://doi.org/10.3390/mi15020217
Submission received: 10 December 2023 / Revised: 22 January 2024 / Accepted: 29 January 2024 / Published: 31 January 2024

Abstract

:
Since machine learning techniques for raindrop removal have not been capable of completely removing raindrops and have failed to take into account the constraints of edge devices with limited resources, a novel software-hardware co-designed method with a memristor for raindrop removal, named memristive attention recurrent residual generative adversarial network (MARR-GAN), is introduced in this research. A novel raindrop-removal network is specifically designed based on attention gate connections and recurrent residual convolutional blocks. By replacing the basic convolution unit with recurrent residual convolution unit, improved capturing of the changes in raindrop appearance over time is achieved, while preserving the position and shape information in the image. Additionally, an attention gate is utilized instead of the original skip connection to enhance the overall structural understanding and local detail preservation, facilitating a more comprehensive removal of raindrops across various areas of the image. Furthermore, a hardware implementation scheme for MARR-GAN is presented in this paper, where deep learning algorithms are seamlessly integrated with neuro inspired computing chips, utilizing memristor crossbar arrays for accelerated real-time image-data processing. Compelling evidence of the efficacy and superiority of MARR-GAN in raindrop removal and image restoration is provided by the results of the empirical study.

1. Introduction

Adverse weather often results in a significant degradation of captured images [1,2,3,4,5]. As a result, it not only affects the quality of images but also hinders advanced vision tasks, such as semantic segmentation [6], object detection [7], etc. Recently, scholars have devoted considerable attention to the restoration of raindrop-laden images, given their importance in real-world applications [8,9,10]. However, removing raindrops from a single image remains a complex challenge due to the intricate physical structure of these droplets.
Currently, raindrop-removal techniques could be broadly classified into two categories: one is a traditional algorithm based on prior knowledge [11,12], while the other one is based on neural networks [13,14,15,16,17,18,19]. The first method requires a manual design or estimation of the raindrop degradation model of raindrops and relies on prior knowledge to model raindrops. This method, however, exhibits limitations in terms of its generalization ability and computational efficiency. For the past few years, the majority of raindrop-removal techniques have adopted the latter method. A generative adversarial network based on recurrent attention has been suggested by Qian et al. [14], which is used to precisely identify raindrop locations and efficiently remove them from images. To maximize the potential significance of raindrop images, Shao et al. [15] has explored the blurred attributes of raindrops and proposed an uncertainty-driven, multi-scale attention network for enhancing raindrop-affected images. The models generated by this approach exhibit commendable quality. However, they incur the tradeoff from increased model size and reduced inference speed. A rapid and efficient network model has been proposed by Guo et al. [13], wherein the key concept is to treat the rain-removal problem as a pixel-wise filtering task. By leveraging highly optimized filtering operations, accelerated inference speed on the GPU was achieved by this model, while significantly enhancing performance. Nevertheless, additional computational resources and memory consumption are still necessitated for deploying and utilizing this model. In addition, various multi-step structures have been created to improve the overall image restoration performance. These include the cascaded attention guidance network (CAG-Net) [16] and the multi-axis MLP (MAXIM) [17]. Compared to conventional methods, the aforementioned techniques have significantly improved the efficiency and quality of image processing. However, most of these systems rely largely on cyclic and progressive processing modes. As a result, system calculation, inference time, and memory usage all increase significantly. Meeting the requirements for computer vision applications that prioritize efficiency becomes challenging at that point. It also poses plenty of difficulties for the application and usage of rainfall algorithms on various platforms. Therefore, it is imperative to provide a reasoning method for embedded artificial intelligence applications that is more effective.
Deep learning has been greatly influenced by the remarkable contributions of deep convolutional neural networks (DCNNs). These well-constructed DCNNs demonstrated their efficient feature learning ability by successfully extracting abstract information from images of rain. There has been a noticeable enhancement in the quality of de-rain images [20,21,22]. However, the substantial computational and memory demands of DCNN arise from its extensive parameter count and calculations, presenting obstacles for hardware implementation. Additionally, the progress in integrated circuit development has faced limitations, making it challenging to overcome the memory constraints imposed by the von Neumann architecture. Therefore, there has been a significant focus on enhancing hardware implementation and expediting the performance of DCNNs. The memristor, renowned for its nanoscale dimensions, rapid operation, non-volatile data storage capability, low power consumption, programmability, and compatibility with complementary metal–oxide–semiconductor (CMOS) technology [23,24], offers innovative hardware solutions for large-scale integrated circuits as well as applications in artificial neural networks, pattern recognition, and image processing [25,26,27,28]. In recent developments, substantial progress has been made in the architecture of memristor-based neural network computing [29,30,31]. Wang et al. has developed a convolutional neural network with five levels that can adjust to the imperfections of the one-transistor–one-memristor array in order to accurately classify the MNIST dataset [32]. An experimental demonstration showing that a memristor crossbar array can be utilized to implement synaptic weights shared at different time steps in an LSTM [33]. This innovative approach not only minimizes circuit space requirements but also enables the storage of a vast number of parameters. Kong et al. introduced a novel 5D fractional order differentiation memristive Hopfield neural network (FOMHNN) [34]. Additionally, they presented a framework for constructing 2n + 1 dimensional simplest Hamiltonian conservative chaotic systems (HCCSs) to optimize encryption speed [35]. Furthermore, the capacity to do computations in memory is utilized, which successfully overcomes the constraints imposed by the ‘von Neumann bottleneck’. The neural signal analysis system’s high-efficiency capabilities allow for accurate neural signal identification and filtering associated with epilepsy. In addition, it shows a nearly 400% improvement in energy efficiency over CMOS technology. Different materials and structures have been utilized to develop a range of memristor devices following the example set by HP memristors [36,37,38], accompanied by the creation of numerous mathematical models for memristors. In particular, the VTEAM model (Voltage Threshold Adaptive Memristor) [39] is an appropriate option for memristor-based neuromorphic systems due to its fundamental physical device properties, experimental verification, and relatively straightforward and precise equations. Hence, our research is focused on utilizing VTEAM, which comprises two layers of TiO2-TiO2-x film positioned between independent Pt electrodes. It should be emphasized that there exist several other memristors and models demonstrating similar characteristics that can also be employed in this investigation. In particular, the utilization of a 1T1R configuration was proposed in this paper, which consists of a single transistor and memristor connected in series [40], revising the matter of leakage current that arises when applying a signal to an individual memristor. Furthermore, we incorporate image de-raining technology into the memristive neural network.
The main contributions of this study could be summarized as follows:
(1)
The proposed memristive attention recurrent residual generative adversarial network (MARR-GAN) presents a software and hardware co-design of the image raindrop removal network, which leverages a memristor crossbar array and deep learning technology.
(2)
A novel raindrop removal module is designed to construct a network for removing raindrops using the autoencoder architecture. Temporal changes in raindrops within images are effectively captured by this framework while preserving their location and shape information. Furthermore, an attention gate design is introduced to enhance the network’s understanding of both global structure and local details in order to achieve comprehensive raindrop removal across different regions of the image.
(3)
A hardware implementation scheme utilizing memristor crossbar arrays for MARR-GAN is presented, integrating deep learning algorithms with neuron computing chips to enhance the speed of processing image data in the real world. Finally, the efficacy and superiority of MARR-GAN have been demonstrated in removing raindrops and restoring fine-grained image details.

2. Related Works

2.1. Rain Streaks Removal

Rain streaks removal can be classified into two distinct categories: solutions based on data analysis and solutions based on mathematical models. Zhuang et al. has presented a novel approach for rain removal by using a deep convolutional neural network [41]. Data-driven approaches involve techniques such as sparse coding, low-rank approximations [42], and Gaussian mixture models (GMMs). A rain-removal technique based on directional gradient priors has been proposed by Huang et al. [43]. Its goal is to eliminate the rain streaks while preserving as much of the original rain image’s structural information as feasible. On the other hand, a wavelet-based sparse optimization approach has been offered by Sun et al. [44] as a solution to the problem of the rain removal process, followed by joint optimization to remove rain-line details from the former while removing the rain line from the latter. Model-based methods encompass convolutional neural networks, transformers, and diffusion modules. Wan et al. has developed a global–local transformer, which employs window-based local transformer blocks to process high-resolution feature maps in shallow layers of GLFormer [20].

2.2. Snow Removal

Snow removal techniques commonly employ prior models based on snow driving features such as a histogram of oriented gradient (HOG) [45], color assumptions [46], and frequency spatial separation. Furthermore, the application of deep convolutional neural networks (CNNS) [47] has been increasingly employed in the domain of removing snow from individual images. These methods generally exhibit commendable performance owing to the resemblance in the appearance of snow particles.

2.3. Raindrop Removal

Currently, deep learning methods are predominantly employed for image raindrop removal tasks owing to their robust feature-extraction capabilities.
Qian et al. has integrated attention mechanisms into a type of generative adversarial network, known as AGAN, for capturing details related to both raindrops and their surroundings [14]. In order to improve raindrop-removal performance, a modular dual residual connection network named DURN for achieving high-quality image restoration outcomes has been proposed [48]. It is important to consider that variations in blur and resolution occurring due to differences in the distance between rainfall and the camera when observing rainy scenes within images. To tackle this issue, a network effectively handles diverse raindrop appearances by independently processing different frequency bands was proposed by Zini et al. [9]. It is capable of performing raindrop removal on images of varying scales, resulting in rapid and high-quality restoration. However, it necessitates substantial computational resources during training and exhibits prolonged processing time.

3. Model Design of MARR-GAN

3.1. Comprehensive Network Architecture

A one-stage network is proposed in this study for effectively separating the observed raindrop image I into clear the background B and raindrop R. The limitations of current progressive processing schemas on edge devices are being addressed, and the application and deployment of rain removal technology are being facilitated by this network. The mathematical expression could be represented by Equation (1). The network architecture of MARR-GAN is illustrated in Figure 1, where an autoencoder framework and attention gate connections are utilized as the raindrop removal network.
I = ( 1 M ( x ) ) B + R
where M is the mask, and I is defined as the input image. In this mask, x is a pixel of input image, M(x) is dependent on the variable x, and M(x) = 1 indicates that x represents the raindrop region, M(x) = 0 indicates the background of the input image. Furthermore, B stands for the background information and R represents mixed raindrops, which symbolize an intricate blend of previous knowledge and reflected light from the environment.
From Figure 1, a generative adversarial network (GAN) is considered to be the main body of this framework. The architecture consists of two sub-networks, namely a generator and discriminator. First, the blue box is the raindrop feature extraction module, which is used to extract raindrop features and generate attention maps. Furthermore, attention residual recurrent gate (ARRG) is guided to generate a raindrop-free image based on the generated attention map. Finally, the discriminator determines whether the input image originates from the generative network or the input image (i.e., Ic).

3.1.1. Raindrop Feature Extraction Module

The visual attention model utilizes the attention recurrent network to enhance target recognition accuracy by identifying the target area [49]. Inspired by that, the network architecture employed in this paper bears resemblance to that designed for raindrop removal. The raindrop localization in an image is facilitated by the utilization of a visual attention guidance generative network and a distinguish network in our study. The generator section (see Figure 1) attention recurrent network comprises multiple circulation modules. There are four residual blocks, a convolution long short-term memory (Conv-LSTM) unit, and a convolution layer for each module. The residual block extracts raindrop feature information from both the input image and the previously generated attention map. Meanwhile, the Conv-LSTM unit and convolution layer generate a 2D attention map.
The generation of attention maps heavily relies on the mask, comprising only two distinct numerical values: 0 and 1. A value of 0 indicates the absence of raindrops in a pixel, while a value of 1 signifies the presence of raindrops in that pixel. The initial attention maps are generated by the first recurrent module of the attention cycle network, which takes both the raindrop image and the mask image as inputs. The mask map is obtained through the subtraction of the clear image from the raindrop-infused image, followed by the application of a specific threshold value for precise filtration. Despite its relatively coarse nature, the obtained mask map significantly contributes to generating a precise attention map. The main difference between the attention map and the mask map lies in what they contain: while the mask map solely consists of 0 and 1, the attention map values range from 0 to 1. The attention map’s higher median value indicates a stronger demand for pixel-focused attention, suggesting an increased probability of raindrops being present at that specific pixel. Even within identical areas affected by raindrops, variations could be observed in the attention map values due to differences in raindrop shape and thickness, thereby reflecting how raindrops impact different pixels within background images.
Conv-LSTM is a module that is part of the attention recurrent network. This module is composed of an input gate (named “it”), a forgetting gate (called “ft”), an output gate (called “Ot”), and a cell unit (called “Ct”). The definition of the state-gates temporal interaction is given as follows:
{ i t = σ ( W x i x t + W h i H t 1 + b i ) f t = σ ( W x f x t + W h f H t 1 + b f ) C t = f t C t 1 + i t t a n h ( W x c x t + W h c H t 1 + b c ) O t = σ ( W x o x t + W h o H t 1 + b o ) H t = O t t a n h ( C t )
where the convolution operation is denoted by ∗, while element-by-element multiplication is represented as . The Sigmoid function is defined as σ, and x t refers to input features. Weight matrix W corresponds to each convolution, and Tanh activation function is indicated by tanh. Bias vectors are represented as b i , b f , b o . By storing threshold weights, Conv-LSTM effectively preserves raindrop pattern features and utilizes Ht for hidden-layer representation. To perform computer vision tasks efficiently and adaptively, Conv-LSTM replaces fully connected weights in traditional LSTM with convolutional counterparts. Additionally, H t 1 provides insights into determining the learning states of raindrops.

3.1.2. Raindrop Removal Network

The raindrop removal network is an autoencoder, referred to as the attention residual recurrent gate (ARRG) module, illustrated in Figure 2.
Recurrent convolutional block combined with the residual unit is employed in both encoding and decoding paths. Additionally, attention gate (AG) is utilized instead of skipping connections to rectify low-resolution features by leveraging deep features. Moreover, batch normalization (BN) is incorporated to enhance the stability and accelerate the convergence speed throughout the neural network’s up-sampling procedure. The utilization of BN enables data standardization, facilitates smaller regularization, diminishes generalization errors, and enhances overall network performance [50].
The autoencoder receives as input the attention map that the raindrop picture and the attention recurrent network produce. The raindrop removal and background restoration processes are guided by the attention map. The autoencoder consists of nine residual recurrent convolution layer (RRCL). Both decoding and encoding have a symmetrical structure. To avoid blurring the image, an attention gate connection is made between the corresponding modules. The detailed configuration of RRCL can be observed in Figure 3.
The RRCL block performs recurrent convolutional layers based on the discrete time steps described in the recurrent convolutional neural network [51]. Let us consider an input sample ul in the lth layer of the RRCL block and a pixel located at (i, j) within an input sample on the kth feature map in the recurrent convolutional layers. The output at time step t, denoted as M i j k l ( t ) , could be represented as follows:
M i j k l ( t ) = ( w k f ) u l f ( i , j ) ( t ) + ( w k r ) T u l f ( i , j ) ( t - 1 ) + b k
In Equation (3), u l f ( i , j ) (t) and u l f ( i , j ) (t − 1) represent standard convolutional layers and the input sample of the lth recurrent convolutional layers, respectively. The standard convolutional layer and the recurrent convolutional layers of the kth feature maps are, respectively, weighted by w k f and w k r , and bk is the bias. The output of recurrent convolutional layers is activated by standard ReLU function f as follows:
F ( u l , w l ) = f ( M i j k l ( t ) ) = ( 0 , M i j k l ( t ) )
The output of the RRCL unit could be calculated as follows:
u l + 1 = u l + F ( u l , w l )
where ul is an RRCL layer input sample. The output of the down-sampling layer in the encoding path and the output of the up-sampling layer in the decoding path are both represented by ul+1, respectively. Figure 3a depicts the fundamental unit of autoencoder convolution, while Figure 3b shows the RRCL block’s structure.
Equations (4) and (5) depict the dynamic properties of recurrent convolutional layers. By extending the recurrent convolutional layers to T time steps, we can obtain a feedforward subnetwork with a depth of T + 1. In this study, we expand recurrent convolutional layers to two steps, denoted as T = 2. The structure of the expanded recurrent convolutional layers consists of a convolutional layer and two subsequent recurrent convolutional layers. Figure 4 shows the structure of the recurrent convolution layer.

3.2. Discrminator Network

The primary responsibility of the discriminative network is to discern between authentic and counterfeit data. In GAN, a global discriminator is typically employed by the discriminator to assess the disparity between the generated image and an authentic sample. However, relying solely on global information to evaluate the authenticity of an image prevents the producing network from effectively restoring local image details. In order to make following the target detection tasks easier, it is important to restore complex image elements as much as possible for raindrop removal. Consequently, existing discrimination networks cannot be directly utilized. Hence, both global and local discriminators are combined in this study to collectively ascertain whether output samples from the generator are true or false.
The local discriminator is employed by taking advantage of the information on the locations of the raindrops in the image. The attention map solves the problem of raindrop placements by being formed in the attention cycle network during the image restoration step. Consequently, the local discriminator could be guided to automatically recognize and pinpoint areas containing raindrops in a picture by adding the attention map into the discriminator network. Convolutional neural networks (CNNs) are used to extract features from the generator’s raindrop images as well as the discriminator’s inner layers. These extracted features are then combined with attention images to form a loss function for training the local discriminator. By introducing an attention map, it directs more focus towards discriminating raindrop regions within an image. In order to determine authenticity, a fully connected layer is utilized in conjunction with other layers in this discrimination network architecture as depicted in Figure 1’s lower section.

3.3. Loss Function

The generator and discriminator loss functions are contained in the MARR-GAN network. The generator loss function includes attention recurrent loss, raindrop-removal loss, and vgg perceptual loss. Both the global discriminator loss function and the attention map loss are included in the discriminator loss function.
Two images of the same backdrop scene, one with a raindrop and the other without, are fed into the generator. The loss function for each attention recurrent module is defined as the average squared difference between the output attention map and the binary mask, M. In the case of a recurrent network, a smaller weight is assigned to the loss function of the front module, while a larger weight is given to the back module’s loss function. The specific formulation of this loss function named La could be observed in Equation (6).
L a ( { A } , M ) = t = 1 N θ N t L M S E ( A t , M )
where At represents an attention graph produced by a cyclic network at time step t. A represents the output attention map. At = attentiont(Ft−1, Ht−1, Ct−1), Ft−1 denotes merging of raindrop-infused image and previous recurrent unit’s output attention map. Function attentiont represents the attentive-recurrent network at time step t. With this recurrent network architecture, increasing value of N leads to the generation of more detailed attention maps; however, it also requires greater memory capacity for storing intermediate parameters. Optimal network efficiency has been observed when N = 4 and θ = 0.8.
There are two types of loss functions employed in the raindrop-removal loss, namely multiscale loss and perceptual loss. The multiscale loss function effectively utilizes image information from various decoder layers to enhance the model optimization process and generate raindrop free images with improved clarity. The multiscale loss function is defined as:
L S ( { I c } , { S } ) = i = 1 M λ i L M S E ( I c i , S i )
where S represents the output of ARRG. Si represents the image characteristics that were extracted from the ARRG i-th layer, Ic represents the input raindrop free image, and Ici indicates the ground truth that has the same scale as that of Si. { λ i } i = 1 M is the weight of different scales. Large-scale picture feature extraction is given more consideration in the design of the loss function, whereas smaller images have less information and hence minimal bearing on model optimization. The last layer, third layer, and fifth layer of the decoder have output picture sizes of 1/4, 1/2, and 1 of the original sizes, respectively. The weights λ associated with these sizes are set to 0.6, 0.8, and 1.0.
In addition to the loss based on pixels, perceptual loss is also incorporated in this study, to assess the overall difference between the output of an automatic context encoder-decoder and its corresponding clear image. Perceptual loss evaluates how similar a raindrop-free image is to its real counterpart from a global perspective, thereby bringing it closer to reality. The VGG16 model could extract global information from images but requires pretraining on an ImageNet dataset beforehand. The definition of perceptual loss function is as follows:
L p ( I d , I c ) = L M S E ( V G G 16 ( I d ) , V G G 16 ( I c ) )
where the VGG16 model is a pre-trained convolutional neural network (CNN) that performs feature extraction on an input image. I d represents the output image generated by the automatic encoder, denoted as I d = G ( I r ) , while I r refers to a real-image sample devoid of raindrops. In summary, the generator network’s loss function could be explained as follows:
L G = 10 2 L G A N ( I d ) + L a ( { A } , M ) + L S ( { S } , { A } ) + L P ( I d , I c )
where L G A N ( I d ) = log ( 1 D ( I d ) ) .
The whole loss function of the discrimination network could be expressed as
L D ( I d , I r , A t ) = log ( D ( I r ) ) log ( D ( I D ) ) + α L m a p ( I d , I r , A t )
where α is 0.05, the global discriminator’s loss function is represented by the first two components of the formula, the local discriminator’s loss function is represented by Lmap, and the local discriminator’s loss function is displayed as follows:
L m a p ( I d , I r , A t ) = L M S E ( D m a p ( I d ) , A t ) + L M S E ( D m a p ( I r ) , 0 )
where Dmap is a 2D attention mask function generated by the discriminator, while I r , I d represents a pair of image samples extracted from the database. A value of 0 in the attention map indicates the absence of raindrops in the real image, hence no guidance is needed for feature extraction. The discriminator architecture in this study comprises eight convolution layers with a core size of (3, 3), followed by a fully connected layer with 1024 neurons that employ Sigmoid activation.

4. Memristive Circuit Design of MARR-GAN

In this section, a memristor circuit implementation scheme for MARR-GAN is proposed. To reduce training time costs, we pre-determine the network parameters using existing training and then map them onto the crossbar array. We utilize linear mapping to convert input data into an analog voltage signal that is scaled by an operational amplifier with a factor of 1000 to adjust the voltage from V to mV. This ensures that all feature information of the input circuit meets the memristor’s voltage threshold. Based on the VTEAM memristor’s threshold characteristic, when reading signals that are lower than writing-threshold voltage, there is minimal change in memristor resistance, thus, ensuring stable storage of network parameters in the crossbar array during the de-raining process. Finally, pixel-value information is recovered by reconstructing output results from the memristive neural network. In this section, we will provide further details regarding implementing three foundational components of MARR-GAN using a memristive circuit approach.

4.1. Convlution Calculation Implementation

In order to preserve the input edge information and avoid a decrease in output size, it is common practice in software programs to employ the same padding technique. This involves padding the input feature maps with zeros before performing convolution operations, ensuring that the output size remains unchanged. In terms of implementing crossbar arrays, this can be achieved by adding zero values to the input signals. As convolution kernel weights can encompass positive, negative, and zero values, whereas meminductance is restricted to non-negative values only, it is not feasible to directly represent negative values using a single memristor. Hence, two memristors are needed to store a single weight value. Assuming m and n represent the count of input and output channels, respectively, while k represents the dimension of convolution kernels. The corresponding memristive circuit configuration for this purpose is illustrated in Figure 5. The matrices of mk*1 and mk*2 are used to represent the inputs and convolution kernel matrices, respectively. The input voltage signals are transmitted horizontally and multiplied by the conductance. By leveraging Ohm’s law for multiplication and Kirchhoff’s current law for accumulation, the total currents are collected vertically. The two collected vertical currents are first passed through an amplifier to increase the amplitude of the signal, making the summed output more obvious, and then the two amplified currents are collected through the summing amplifier (blue block). The circuit in Figure 5 could be described using Equation (12). Each input-feature map undergoes n convolution operations, followed by m*n convolution operations on m maps. Consequently, n groups of m output maps need to be merged together to obtain the final n result maps. In other words, two columns perform single feature extraction.
Y N = R L [ m = 1 k = 1 x k , m ( σ k + k ( m 1 ) , 2 n 1 + σ k + k ( m 1 ) , 2 n ) ] ( n = 1 , 2 , 3 )
The input voltage signal is denoted as xk,m, while σ + and σ correspond to the conductance of a pair of memristors. It should be noted that all instances of σ have predetermined fixed values for their conductance. Therefore, by adjusting the value of σ + through calibration, we can effectively represent the convolution kernel weights associated with positive, negative, and zero.
The input to the network in this paper is an RGB, three-channel image with dimensions of W × H and a convolution kernel size of 3 × 3. Consequently, the image is transformed into an output vector of size W × H × 1, and each time a 3 × 3 field is added. This vector is then sent to the convolution kernel for the convolution operation. During this process, the convolution kernel continuously slides over the input image with a fixed step size. The convolution operation involves calculating the sum of products between shared weights associated with the convolution kernel and the generated input image. This operation could be decomposed into parallel multiply-accumulate operations, making it suitable for memristor-based in-memory computing architectures. To facilitate hardware implementation, the input image is expanded into a nine-dimensional vector, which drives a corresponding nine pulse channels to provide simultaneous access to nine-bit lines. The weights are represented by two differential conductances of 1T1R memristors, mapping positive and negative rows, respectively, for each weight value within a given convolutional kernel. By doing so, we calculate differences between cumulative flowing currents on two relevant source lines as weighted sums of both convoluted inputs and their respective kernel weights.

4.2. RRCL Network

MARR-GAN consists of the RRCL module and the global average pooling layer attention gate. Each RRCL module is followed by a BN processing by each convolution kernel, and the nonlinearity is introduced by ReLU activation function, after which it is passed to the next convolution layer.

4.2.1. Global Average Pooling Module

The global average pooling layer module (GAP) pools feature maps of different input sizes into 1 × 1 feature maps. The module will perform a global average pooling operation on images of size C*H*W. Therefore, instead of using memristors with varying resistance values, a resistor is used instead. Memristors are used in the summation circuit, and average pooling is implemented by a column parallel memristor cross array. All feature information within the pool window (one pool) is fed to the parallel cross array, which could be expressed as:
V A t = r ( V C , H W × R M t ) , t = 1 , 2 , 3 n
The schematic diagram of the GAP is illustrated in Figure 6. Assuming the pooling window size is k, The memristor’s significance within the module for an average pooling circuit M t = k × R , where R represents resistance. It should be noted that R is set to 1 kΩ.
Since the GAP layer does not require training weights in its computation, it takes a simplified approach to processing data. Specifically, it utilizes resistors instead of memristors that require training weights, which are determined during the manufacturing process and do not need to be trained or adjusted at runtime. This design greatly simplifies the complexity of the circuit and improves computational efficiency.
However, the summation circuit part is different. Here, the nonlinear nature of the memristor plays a key role. Due to this property of the amnesia, it can efficiently handle massively parallel computing tasks, which is crucial when dealing with large amounts of data. By applying the amnesia to the summation circuit, we can achieve fast and accurate data processing, which is crucial for the operation of neural networks.
In summary, the design of the GAP layer using a cross array of amnesia on the circuit as well as a simple summation circuit is an innovative approach. It skillfully combines the memory properties of the memristor and the stability of the resistor, making the whole circuit both simple and efficient.

4.2.2. Batch Normalization (BN)

The operation of the BN layer can be simplified as a linear transformation of the input x. The calculation method is as follows:
y = a x + b
A circuit for achieving a linear variation in input x can be realized by integrating an operational amplifier with two memristors.
The schematic diagram of the BN layer is illustrated in Figure 7. The op amp on the right functions as an inverting adder circuit, generating the output signal as follows:
y b n = R b n M b n 1 x j R b n M b n 2 x b
The primary function of the BN is to adjust the distribution of input feature data x. Therefore, it is essential for the transformation coefficient a of the BN layer to be positive. Rbn, Mbn1, and Mbn2 are greater than 0, if and only if xb > 0, b < 0.
Of the inputs xj and xb, xj is the input data, xb is the bias input, where xj and xb are input into two memristors, and then input into the inverting operational amplifier circuit to perform linear transformation to obtain the output ybn. The principle of this design is to use the memristors to store xi, and then through the inverting operational amplifier circuit to perform the linear transformation to realize the linear amplification or reduction in the input signal. The principle of this design is to use a memristor to store xj, and then perform a linear transformation through an inverting operational amplifier circuit to linearly amplify or reduce the input signal. By changing the conductance value of the memristor and adjusting the behavior of the BN layer, this design can improve the training efficiency and stability of the neural network and reduce the sensitivity of the model to the initial weights.

4.2.3. Sigmoid

Based on the exponential circuit, we have successfully developed circuits for implementing the activation functions sigmoid. The resulting outputs could be mathematically represented as follows:
V s = V b V b + exp ( V i )
where Vb represents a bias voltage with a fixed size of 1 V.
As the circuit shown in Figure 8, Vb represents the bias voltage, Vi represents the input voltage, and Vs represents the output voltage, which is realized by three inverting amplifier circuits, an adder, and a multiplier. A bias voltage of 1 V is added to the input of this circuit, as the sigmoid function changes its output slowly when the input is close to zero. By adding a bias voltage of 1 V, the input voltage can be made to vary over a wider range, which improves the linearization of the sigmoid function and extends the linearization range of the sigmoid function. In the neural network designed in this paper, this helps to improve the linear response of the network to the inputs, allowing the network to better learn and recognize the features of various raindrops.

5. Experiments

The performance and superiority of MARR-GAN are validated on various artificial datasets and real-world images affected by raindrops, with qualitative and quantitative comparisons conducted against different state-of-the-art (SOTA) techniques.

5.1. Implementation Details

5.1.1. Datasets

There are two types of datasets utilized in the experiment. One dataset was obtained by Qian et al. [14] through the manual application of water droplets on the camera lens followed by capturing photos. A total of 861 pairs of training data constitutes this dataset, comprising images with raindrop effects and their corresponding clean versions. The test dataset comprises Testa and Testb, containing a total of 58 and 249 samples, respectively. The other dataset that Quan et al. [52] constructed is a comprehensive dataset of real raindrop images, consisting of 1000 high-resolution photographs captured with a single-lens reflex (SLR) camera. A total of 250 pairs of raindrop images are encompassed by this dataset, covering diverse environmental settings such as parking lots, parks, and urban areas. Moreover, data collection were conducted at different times of the day morning, noon, and afternoon to capture raindrop images under varying lighting conditions in authentic scenes.

5.1.2. Training Details

The proposed network is trained and tested on Ubuntu 20.04, 128 GB host memory, 32 GB GPU, CUDA 11.6, CUDNN 8.5, Python 3.9, and NVIDIA RTX-3090. The experimental condition is based on Pytorch-GPU 1.12.0. The size of the input image pair is 720 × 480. We used the Adam optimizer with a learning rate of 0.0001 and a decay rate of 0.90. The model was trained for 500 epochs, and the entire training time of the proposed model was approximately 30 h.

5.2. Experimental Results

The Raindrop and RainDS datasets are utilized in this experiment to evaluate the performance of the network, which is then compared with state-of-the-art raindrop removal networks including AGAN [14], ATT [53], A2Net [54], DURN [48], and TMN [55]. By comparing the above networks, the raindrop removal performance is evaluated from SSIM, PSNR indicators, and visual effects.

5.2.1. Quantitative Evaluation

The evaluations on the Raindrop dataset of our approach are presented in Table 1 in comparison with AGAN [14], ATT [53], A2Net [54], DURN [48], and TMN [55].
From the comparison in Table 1, it can be seen that in contrast to the previous best approach, MARR-GAN improves 0.005 of SSIM on Testb; MARR-GAN improves 0.148 dB of PSNR on Testa and improves 0.397 dB of PSNR on Testb; MARR-GAN improves 1 of PHah on Testa.
The evaluations on the RainDS dataset of our approach are presented in Table 2 in comparison with AGAN [14], ATT [53], A2Net [54], DURN [48], and TMN [55].
From the comparison in Table 2, it can be seen that in contrast to the previous best approach, MARR-GAN improves 0.023 dB of PSNR and 0.023 of SSIM.

5.2.2. Qualitative Evaluation

The visual comparison results of various raindrop removal methods are depicted in Figure 9 and Figure 10. It can be observed that AGAN [14] has a color shift. The methods ATT [53] and DuRN [48] are associated with the presence of conspicuous distortions in terms of both artifacts and color accuracy. Although A2Net [54] and TMN [55] demonstrate satisfactory performance in overall perception, they lack consistency when it comes to restoring local structures from pristine samples. On the contrary, our approach guarantees the retention of authentic background details without any visually perceptible distortions. Moreover, Figure 11 presents a comparative analysis showcasing raindrops exhibiting reflected colors at different transparency levels. It becomes apparent that existing techniques yield restored images containing undesirable artifacts or incomplete semantic information. In contrast, our method enables more dependable image reconstruction while effectively accommodating various raindrop morphologies.

5.3. Ablation Study

This section develops an ablation study to investigate the importance of various MARR-GAN components. All of the approaches’ ablation experiments are carried out on Raindrop. In particular, we created the four variation experiments listed below:
Experiment I: In this experiment, we will exclude the recurrent convolutional unit and add a connection line in RRCL, instead utilizing the conventional convolution layer depicted in Figure 3b for image rain removal processing.
Experiment II: For this experiment, we will replace the attention gate connections in RRCL with skip connection blocks.
Experiment III: For this experiment, we will change the number of residual blocks to three in the recurrent attention network.
In experiments I to III, we examined the impact of incorporating recurrent convolutional units and adding connection lines, attention gate connections in RRCL, and varying the number of residual blocks in the recurrent attention network on image de-raining outcomes. To ensure reproducibility and minimize training randomness, we calculated average result values using Raindrop datasets. Table 3 presents a quantitative comparison of these three variations. It is evident that our MARR-GAN consistently achieves higher PSNR and SSIM scores compared to the other variants. Conversely, our overall model effectively enhances structural details in the restored images.

6. Conclusions

A novel image de-raining network, named MARR-GAN, which combines software and hardware design, is presented in this study. The proposed network utilizes an attention recurrent residual generative adversarial network to effectively remove raindrops from images. Additionally, we introduce a cyclic raindrop feature extraction module that facilitates the analysis and interpretation of pixel relationships within the raindrop feature image. Then, a raindrop removal module is designed to handle image areas with complex texture details. Based on this premise, a memristor-based hardware implementation scheme is put forward, offering an advantageous hardware resolution for accelerating DCNN and ensuring real-time and dependable processing of intricate images. Experimental results show that the MARR-GAN algorithm proposed in this article has good application prospects in removing raindrops and restoring image details.

Author Contributions

Conceptualization, Q.C. and Y.L.; methodology, Q.C. and Y.L.; software, Q.C. and Y.L.; validation, Q.C. and Y.L.; formal analysis, Q.C. and Y.L.; investigation, Q.C. and Y.L.; resources, Q.C. and Y.L.; data curation, Q.C. and Y.L.; writing—original draft preparation, Q.C. and Y.L.; writing—review and editing, Q.C. and Y.L.; visualization, Y.L.; supervision, Y.L.; project administration, Y.L.; funding acquisition, visualization, Q.C. and Y.L.; supervision, Q.C. and Y.L.; project administration, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the Science and Technology Project of Jilin Province under Grant 20220204090YY and Changchun Science and Technology Bureau Youth Team Innovation Project. Under Grant 23ZCX04.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

Author Liu, Y. was also employed by the company Changchun Changding Electronic Technology LLC. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Wu, W.; Liu, Y.; Li, Z. Subband Differentiated Learning Network for Rain Streak Removal. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4675–4688. [Google Scholar] [CrossRef]
  2. Jiang, N.; Luo, J.; Lin, J.; Chen, W.; Zhao, T. Lightweight Semi-Supervised Network for Single Image Rain Removal. Pattern Recognit. 2023, 137, 109277. [Google Scholar] [CrossRef]
  3. Cheng, B.; Li, J.; Chen, Y.; Zeng, T. Snow Mask Guided Adaptive Residual Network for Image Snow Removal. Comput. Vis. Image Underst. 2023, 236, 103819. [Google Scholar] [CrossRef]
  4. Wang, H.; Xie, Q.; Wu, Y.; Zhao, Q.; Meng, D. Single image rain streaks removal: A review and an exploration. Int. J. Mach. Learn. Cybern. 2020, 11, 853–872. [Google Scholar] [CrossRef]
  5. Yang, W.; Tan, R.T.; Feng, J.; Guo, Z.; Yan, S.; Liu, J. Joint Rain Detection and Removal from a Single Image with Contextualized Deep Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1377–1393. [Google Scholar] [CrossRef]
  6. Wang, F.; Zhang, Y. A De-raining semantic segmentation network for real-time foreground segmentation. J. Real-Time Image Process. 2021, 18, 873–887. [Google Scholar] [CrossRef]
  7. Huang, S.C.; Jaw, D.W.; Hoang, Q.V.; Le, T.H. 3FL-Net: An Efficient Approach for Improving Performance of Lightweight Detectors in Rainy Weather Conditions. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4293–4305. [Google Scholar] [CrossRef]
  8. Wu, D.; Guo, Z.; Li, A.; Yu, C.; Gao, C.; Sang, N. Conditional Boundary Loss for Semantic Segmentation. IEEE Trans. Image Process. 2023, 32, 3717–3731. [Google Scholar] [CrossRef]
  9. Zini, S.; Buzzelli, M. Laplacian encoder-decoder network for raindrop removal. Pattern Recogn. Lett. 2022, 158, 24–33. [Google Scholar] [CrossRef]
  10. Yan, W.; Xu, L.; Yang, W.; Tan, R.T. Feature-Aligned Video Raindrop Removal with Temporal Constraints. IEEE Trans. Image Process. 2022, 31, 3440–3448. [Google Scholar] [CrossRef]
  11. Luo, W.; Lai, J.; Xie, X. Weakly Supervised Learning for Raindrop Removal on a Single Image. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1673–1683. [Google Scholar] [CrossRef]
  12. Du, S.; Liu, Y.; Ye, M.; Zhao, M.; Li, Z. Rain streaks removal from single image based on texture constraint of background scene. Neurocomputing 2021, 419, 224–238. [Google Scholar] [CrossRef]
  13. Guo, Y.; Chen, J.; Ren, X.; Wang, A.; Wang, W. Joint Raindrop and Haze Removal from a Single Image. IEEE Trans. Image Process. 2020, 29, 9508–9519. [Google Scholar] [CrossRef]
  14. Qian, R.; Tan, R.T.; Yang, W.; Su, J.; Liu, J. Attentive Generative Adversarial Network for Raindrop Removal from A Single Image. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2482–2491. [Google Scholar]
  15. Shao, M.W.; Li, L.; Meng, D.Y.; Zuo, W.M. Uncertainty guided multi-scale attention network for raindrop removal from a single image. IEEE Trans. Image Process. 2021, 30, 4828–4839. [Google Scholar] [CrossRef]
  16. Wang, G.; Sun, C.; Sowmya, A. Cascaded attention guidance network for single rainy image restoration. IEEE Trans. Image Process. 2020, 29, 9190–9203. [Google Scholar] [CrossRef]
  17. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. MAXIM: Multi-Axis MLP for Image Processing. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5759–5770. [Google Scholar]
  18. Guo, Q.; Sun, J.Y.; Jue, F.; Ma, L.; Xie, X.F.; Liu, Y.; Zhao, J.J. Efficient Derain: Learning pixel-wise dilation filtering for high efficiency single-image deraining. In Proceedings of the 2021 AAAI Conference on Artificial Intelligence (AAAI), Vancouver, BC, Canada, 2–9 February 2021; pp. 1487–1495. [Google Scholar]
  19. Chen, M.; Wang, P.; Shang, D.; Wang, P. Cycle-attention-derain: Unsupervised rain removal with CycleGAN. Vis. Comput. 2023, 39, 3727–3739. [Google Scholar] [CrossRef]
  20. Wan, Y.; Shao, M.; Bao, Z.; Cheng, Y. Global–local transformer for single-image rain removal. Pattern Anal. Appl. 2023, 26, 1527–1538. [Google Scholar] [CrossRef]
  21. Xue, P.; He, H. Research of Single Image Rain Removal Algorithm Based on LBP-CGAN Rain Generation Method. Math. Probl. Eng. 2021, 10, 1155. [Google Scholar] [CrossRef]
  22. Hu, X.; Zhu, L.; Wang, T.; Fu, C.W.; Heng, P.A. Single-Image Real-Time Rain Removal Based on Depth-Guided Non-Local Features. IEEE Trans. Image Process. 2021, 30, 1759–1770. [Google Scholar] [CrossRef] [PubMed]
  23. Liu, Y.; Li, H.; Guo, S.-X.; Iu, H.H.C. Generation of Multi-Lobe Chua Corsage Memristor and Its Neural Oscillation. Micromachines 2022, 13, 1330. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, Y.; Iu, H.H.C. Novel Floating and Grounded Memory Interface Circuits for Constructing Mem-Elements and Their Applications. IEEE Access 2020, 8, 114761–114772. [Google Scholar] [CrossRef]
  25. Liu, Y.; Iu, H.H.C.; Guo, Z.; Si, G. The Simple Charge-Controlled Grounded/Floating Mem-Element Emulator. IEEE Trans. Circuits Syst. II Express Briefs 2021, 68, 2177–2181. [Google Scholar] [CrossRef]
  26. Xie, D.; Xiao, H.; Zhou, Y.; Duan, S.; Hu, X. MWA-MNN: Multi-patch Wavelet Attention Memristive Neural Network for image restoration. Expert Syst. Appl. 2024, 240, 122427. [Google Scholar] [CrossRef]
  27. Hong, Q.; Li, Y.; Wang, X. Memristive continuous Hopfield neural network circuit for image restoration. Neural Comput. Appl. 2020, 32, 8175–8185. [Google Scholar] [CrossRef]
  28. Yu, Y.-B.; Zhou, C.; Deng, Q.-X.; Zhong, Y.-J.-Y.; Cheng, M.; Kang, Z.-F. Memristor-based genetic algorithm for image restoration. J. Electron. 2022, 20, 100158. [Google Scholar] [CrossRef]
  29. Deng, Q.; Wang, C.; Lin, H. Memristive Hopfield neural network dynamics with heterogeneous activation functions and its application. Chaos Solitons Fractals 2024, 178, 114387. [Google Scholar] [CrossRef]
  30. Wang, C.; Tang, D.; Lin, H.; Yu, F.; Sun, Y. High-dimensional memristive neural network and its application in commercial data encryption communication. Expert Syst. Appl. 2024, 242, 122513. [Google Scholar] [CrossRef]
  31. Tang, D.; Wang, C.; Lin, H.; Yu, F. Dynamics analysis and hardware implementation of multi-scroll hyperchaotic hidden attractors based on locally active memristive Hopfield neural network. Nonlinear Dyn. 2024, 112, 1511–1527. [Google Scholar] [CrossRef]
  32. Wang, Z.; Li, C.; Lin, P.; Rao, M.; Nie, Y.; Song, W.; Qiu, Q.; Li, Y.; Yan, P.; Strachan, J.P. In situ training of feed-forward and recurrent convolutional memristor networks. Nat. Mach. Intell. 2019, 1, 434–442. [Google Scholar] [CrossRef]
  33. Li, C.; Wang, Z.; Rao, M.; Belkin, D.; Song, W.; Jiang, H.; Yan, P.; Li, Y.; Lin, P.; Hu, M. Long short-term memory networks in memristor crossbar arrays. Nat. Mach. Intell. 2019, 1, 49–57. [Google Scholar] [CrossRef]
  34. Kong, X.; Yu, F.; Yao, W.; Cai, S.; Zhang, J.; Lin, H. Memristor-induced hyperchaos, multiscroll and extreme multistability in fractional-order HNN: Image encryption and FPGA implementation. Neural Netw. 2024, 171, 85–103. [Google Scholar] [CrossRef] [PubMed]
  35. Kong, X.; Yu, F.; Yao, W.; Xu, C.; Zhang, J.; Cai, S.; Wang, C. A class of 2n+1 dimensional simplest Hamiltonian conservative chaotic systems and fast image encryption schemes. Appl. Math. Modell. 2024, 125, 351–374. [Google Scholar] [CrossRef]
  36. Isah, A.; Nguetcho, A.S.; Binczak, S.; Bilbault, J.-M. Comparison of the Performance of the Memristor Models in 2D Cellular Nonlinear Network. Electronics 2021, 10, 1577. [Google Scholar] [CrossRef]
  37. Hu, X.; Wang, W.; Sun, B.; Wang, Y.; Li, J.; Zhou, G. Refining the negative differential resistance effect in a TiOx-based memristor. J. Phys. Chem. Lett. 2021, 10, 1021. [Google Scholar] [CrossRef] [PubMed]
  38. Mladenov, V. A unified and open LTSPICE memristor model library. Electronics 2021, 10, 1594. [Google Scholar] [CrossRef]
  39. Ren, K.; Zhang, K.; Qin, X.; Yang, F.; Sun, B.; Zhao, Y.; Zhang, Y. VETAM-M: A General Model for Voltage-Controlled Memcapacitive-Coupled Memristors. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 1717–1721. [Google Scholar] [CrossRef]
  40. Cheng, L.; Li, J.C.; Zheng, H.X.; Yuan, P.; Yin, J.H.; Yang, L.; Luo, Q.; Li, Y.; Lv, H.B.; Chang, T.C.; et al. Memristors: In-Memory Hamming Weight Calculation in a 1T1R Memristive Array. Adv. Electron. Mater. 2020, 69, 4920–4923. [Google Scholar]
  41. Zhuang, J.H.; Luo, Y.S.; Zhao, X.L.; Jiang, T.X.; Chang, Y.; Liu, J. Unsupervised video rain streaks removal with deep foreground–background modeling. J. Comput. Appl. Math. 2024, 436, 115431. [Google Scholar] [CrossRef]
  42. Luo, Y.; Ling, J. Single-image de-raining using low-rank matrix approximation. Neural Comput. Appl. 2020, 32, 7503–7514. [Google Scholar] [CrossRef]
  43. Huang, S.; Xu, Y.; Ren, M.; Yang, Y.; Wan, W. Rain Removal of Single Image Based on Directional Gradient Priors. Appl. Sci. 2022, 12, 11628. [Google Scholar] [CrossRef]
  44. Sun, G.; Leng, J.; Cattani, C. A particular directional multilevel transform-based method for single-image rain removal. Knowl. Based Syst. 2020, 200, 106000. [Google Scholar] [CrossRef]
  45. Li, M.; Cao, X.; Zhao, Q.; Zhang, L.; Meng, D. Online Rain/Snow Removal from Surveillance Videos. IEEE Trans. Image Process. 2021, 30, 2029–2044. [Google Scholar] [CrossRef] [PubMed]
  46. Fazlali, H.; Shirani, S.; Bradford, M.; Kirubarajan, T. Single image rain/snow removal using distortion type information. Multimed. Tools Appl. 2022, 81, 14105–14131. [Google Scholar] [CrossRef]
  47. Li, P.; Yun, M.; Tian, J.; Tang, Y.; Wang, G.; Wu, C. Stacked dense networks for single-image snow removal. Neurocomputing 2019, 367, 152–163. [Google Scholar] [CrossRef]
  48. Liu, X.; Suganuma, M.; Sun, Z.; Okatani, T. Dual Residual Networks Leveraging the Potential of Paired Operations for Image Restoration. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7000–7009. [Google Scholar]
  49. Xu, C.; Gao, J.; Wen, Q.; Wang, B. Generative Adversarial Network for Image Raindrop Removal of Transmission Line Based on Unmanned Aerial Vehicle Inspection. Wirel. Commun. Mob. Comput. 2021, 10, 6668771. [Google Scholar] [CrossRef]
  50. Xu, Y.; Xie, L.; Xie, C.; Dai, W.; Mei, J.; Qiao, S.; Shen, W.; Xiong, H.; Yuille, A. BNET: Batch Normalization with Enhanced Linear Transformation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9225–9232. [Google Scholar] [CrossRef]
  51. Zuo, Q.; Chen, S.; Wang, Z. R2AU-Net: Attention Recurrent Residual Convolutional Neural Network for Multimodal Medical Image Segmentation. Secur. Commun. Netw. 2021, 10, 6625688. [Google Scholar] [CrossRef]
  52. Quan, R.; Yu, X.; Liang, Y.; Yang, Y. Removing Raindrops and Rain Streaks in One Go. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 9143–9152. [Google Scholar]
  53. Quan, Y.; Deng, S.; Chen, Y.; Ji, H. Deep Learning for Seeing through Window with Raindrops. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27–28 November 2019; pp. 2463–2471. [Google Scholar]
  54. Lin, H.; Jing, C.; Huang, Y.; Ding, X. A2Net: Adjacent Aggregation Networks for Image Raindrop Removal. IEEE Access 2020, 8, 60769–60779. [Google Scholar] [CrossRef]
  55. Kwon, H.-J.; Lee, S.-H. Raindrop-Removal Image Translation Using Target-Mask Network with Attention Module. Mathematics 2023, 11, 3318. [Google Scholar] [CrossRef]
Figure 1. The network overview of MARR-GAN.
Figure 1. The network overview of MARR-GAN.
Micromachines 15 00217 g001
Figure 2. The structure of ARRG.
Figure 2. The structure of ARRG.
Micromachines 15 00217 g002
Figure 3. (a) The basic unit of the auto encoder convolution. (b) The structure of RRCL.
Figure 3. (a) The basic unit of the auto encoder convolution. (b) The structure of RRCL.
Micromachines 15 00217 g003
Figure 4. Unfolded recurrent convolutional units for T = 2.
Figure 4. Unfolded recurrent convolutional units for T = 2.
Micromachines 15 00217 g004
Figure 5. Template for implementing a memristive circuit to perform convolution calculations.
Figure 5. Template for implementing a memristive circuit to perform convolution calculations.
Micromachines 15 00217 g005
Figure 6. Realization of the memristive circuit for calculating GAP.
Figure 6. Realization of the memristive circuit for calculating GAP.
Micromachines 15 00217 g006
Figure 7. Realization of the memristive circuit for calculating BN.
Figure 7. Realization of the memristive circuit for calculating BN.
Micromachines 15 00217 g007
Figure 8. Realization of the memristive circuit for calculating Sigmoid.
Figure 8. Realization of the memristive circuit for calculating Sigmoid.
Micromachines 15 00217 g008
Figure 9. Qualitative evaluation on Testa.
Figure 9. Qualitative evaluation on Testa.
Micromachines 15 00217 g009
Figure 10. Qualitative evaluation on Testb.
Figure 10. Qualitative evaluation on Testb.
Micromachines 15 00217 g010
Figure 11. Qualitative evaluation on RainDS.
Figure 11. Qualitative evaluation on RainDS.
Micromachines 15 00217 g011
Table 1. Results of Raindrop dataset provided by Qian et al. [14].
Table 1. Results of Raindrop dataset provided by Qian et al. [14].
DatasetMetricAGANATTA2NetDURNTMNMARR-GAN
TestaPSNR (dB)29.56930.80827.96130.100830.06230.956
SSIM0.9050.9170.9110.9120.9050.900
PHah314221
TestbPSNR (dB)24.228-- 125.08124.32224.31325.478
SSIM0.786-- 10.7930.8020.7870.807
PHah4-- 12342
1 “--”: results are not released.
Table 2. Results of RainDS dataset provided by Quan et al. [52].
Table 2. Results of RainDS dataset provided by Quan et al. [52].
DatasetMetricAGANATTA2NetDURNTMNMARR-GAN
RainDSPSNR (dB)24.298-- *24.46123.88224.15124.484
SSIM0.819-- *0.8420.8180.8280.847
PHah4-- *3443
* “--”: results are not released.
Table 3. Ablation results of Raindrop dataset provided by Qian et al. [14].
Table 3. Ablation results of Raindrop dataset provided by Qian et al. [14].
DatasetMetricExperiment IExperiment II Experiment IIIMARR-GAN
TestaPSNR (dB)29.35829.94128.69230.956
SSIM0.8960.8840.8650.900
PHah3341
TestbPSNR (dB)24.37625.31025.32725.478
SSIM0.7920.8010.7830.807
PHah2222
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chai, Q.; Liu, Y. MARR-GAN: Memristive Attention Recurrent Residual Generative Adversarial Network for Raindrop Removal. Micromachines 2024, 15, 217. https://doi.org/10.3390/mi15020217

AMA Style

Chai Q, Liu Y. MARR-GAN: Memristive Attention Recurrent Residual Generative Adversarial Network for Raindrop Removal. Micromachines. 2024; 15(2):217. https://doi.org/10.3390/mi15020217

Chicago/Turabian Style

Chai, Qiuyue, and Yue Liu. 2024. "MARR-GAN: Memristive Attention Recurrent Residual Generative Adversarial Network for Raindrop Removal" Micromachines 15, no. 2: 217. https://doi.org/10.3390/mi15020217

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop