1. Introduction
Multiple-input multiple-output (MIMO) technology has been widely applied in wireless communications to improve spectrum efficiency [
1]. The main advantage of MIMO systems is their ability to use spatial multiplexing techniques to greatly improve the spectral efficiency of communication systems with bandwidth and power constraints [
2]. Currently, modern communication systems often use MIMO and high-order quadrature amplitude modulation (QAM) to improve spectral efficiency.
However, the widespread deployment of high-dimensional MIMO systems is hampered by a number of limitations. In a practical scenario, communication systems are configured with a large number of radio frequency units, and the complexity of the receiver will be very high when performing signal processing [
3]. Moreover, the complexity of high-performance maximum likelihood (ML) detection or sphere decoding (SD) increases exponentially with the number of antennas and the modulation order [
4]. Therefore, low-complexity and high-performance MIMO detection techniques are of great research importance.
For MIMO communication systems, the ML detector needs to traverse all possible vectors, provided that the receiver obtains channel state information (CSI) and that the transmitted symbols obey a uniform distribution [
5]. The ML detector is an NP-hard problem [
6] and is too complex to be applied in practice when the number of transmitting antennas is large and high-order modulation is used. Therefore, how to simplify the algorithm of signal detection and approximate the performance of maximum likelihood detection has become a hot issue for the communication community.
The expectation propagation (EP) detection algorithm is a low-complexity algorithm with high performance. The EP algorithm is one of the Bayesian machine learning algorithms. It can efficiently handle continuous distributions through moment matching [
7]. Therefore, it has the ability to handle more complex and general approximation functions [
8].
The performance of EP detection is very close to ML detection. Even for very large MIMO orders such as 250 transmitting antennas, it only takes 7∼8 iterations to approximate the original distribution, and the complexity of each iteration depends on the inverse of the
matrix [
9]. Simulation results show that the number of iterations required by the EP algorithm is independent of the modulation order of the system, and the algorithm complexity is maintained at
[
10]. Moreover, EP detection is a soft output algorithm that provides a posterior probability estimate for each received symbol, which can be naturally performed for further channel decoding [
11].
The integration of artificial intelligence technology in communication will become a mainstream direction of the sixth-generation mobile communication technology (6G) development [
12]. Coupled with the improvement of hardware level and computing power, machine learning technology has been widely used.
Machine learning is a type of artificial intelligence that requires the support of large amounts of data to train models that ideally meet the needs [
13]. A neural network is an algorithm in machine learning that belongs to supervised learning. Supervised learning means that the output results of training data are real and certain, and can accurately give the training error and guide the model to train and learn [
14]. The neural network is composed of a large number of interconnected nodes. In the process of data transmission iteration, the parameter values of nodes will be gradually optimized and the loss function will be constantly reduced, so as to achieve the purpose of accurately mapping the input data to the output data [
15].
Deep learning is a branch of machine learning, which has attracted great attention in recent years, and a large number of experts have conducted in-depth research. Deep learning can be understood as a longitudinal extension of neural network algorithms [
16]. Based on the basic neural network, the number of network layers and network reasoning operation process is studied in detail, so that the network model can have more powerful learning ability and the network can learn extremely complex operation processes after a lot of iterative training [
17]. In recent years, deep learning has been used in various fields, such as natural language processing, computer vision, and recommendation algorithms.
In the past, most mainstream ideas relied on the powerful learning ability of neural networks to improve the performance of algorithms. The essence of the reasoning process of the network, namely the prediction of new data, is several matrix multiplication operations. Therefore, a novel idea is to learn and train some complex steps in the MIMO detection algorithm through the neural network, converting the complex calculation process into a few matrix multiplication operations, so as to greatly improve the efficiency of the detection algorithm. This strategy not only can effectively reduce the complexity of the detection algorithm in the form of a matrix product, but also can approximate the original detection performance by virtue of the fitting ability of the neural network.
The EP detection algorithm has good performance, but there are still some complicated operations in each iteration, such as matrix inversion, moment matching, etc. These calculations are not conducive to a low-complexity implementation in hardware. Therefore, this paper uses an artificial intelligence approach to optimize some of the complex steps of the EP algorithm to further reduce the complexity of the algorithm and approximate the performance of the EP algorithm. The following are the key contributions of this work:
This paper analyzes the steps with high complexity, and then uses the neural network model to learn and map some time-consuming steps in the EP detection algorithm, and convert the complex operation process into several matrix multiplication operations to reduce the complexity of the detection algorithm.
The improved method is tested with different MIMO channel models and verified to have strong robustness as well as adaptability. The performance and applicability of the method are verified in different scenarios.
Considering the problem of excessive differences in data with different SNRs, SNR is divided into different intervals, and different neural network models are trained on the data with different SNR intervals. The accuracy of the model mapping process is guaranteed.
A simple structure of 4 × 32 × 16 × 2 is chosen for the neural network model to minimize the complexity of neural network prediction and the space occupied while maintaining sufficient learning capability of the model.
The rest of this paper is organized as follows. In
Section 2, we show the MIMO system model and summarize and compare common detection algorithms briefly. In
Section 3, we detail the process of the EP detection algorithm. In
Section 4, we first analyze the steps with high complexity in the EP algorithm and determine the application strategy of the neural network, then detail the process of model building and training, and finally embed the neural network into the EP detection algorithm. Simulation results and corresponding analysis are shown in
Section 5. A conclusion is provided in
Section 6.
Notation: Column vectors and matrices are denoted by bold letters in lower case and in upper case, respectively. is an identity matrix. and are the conjugate transpose and transpose operator, respectively. denotes the complex matrix (or vector) with a scale of . and represent the expectation operator and covariance operator, respectively. denotes a circularly symmetric complex Gaussian (CSCG) distribution with a mean of 0 and a variance of .
2. MIMO Symbol Detection
In the MIMO system, an important problem is how to get the signal in the presence of noise, which is the process of MIMO detection. Assuming a MIMO communication system with
n transmitting antennas and
m receiving antennas, the transmitted symbol vector
passes through the complex channel matrix
and the Gaussian white noise
is added to get the output symbol vector
. It can be expressed as:
where
is
i.i.d transmitted symbol vector, where
.
is complex MIMO channel matrix.
is independent additive complex Gaussian white noise with a mean of 0 and variance of
, that is
.
is the output symbol vector.
The signal-to-noise ratio (SNR) is an important parameter of communication system [
18,
19]. A higher SNR means better communication conditions. The SNR of the MIMO communication system can be expressed as:
where
represents the average energy of transmitted symbols.
The signal detection is that the receiver gets the output symbol
and restores the transmitted symbol vector
from
. MIMO detection often adopts Bayesian estimation to estimate the expectation of a posteriori probability. It can be expressed as:
The posterior probability of transmitted symbol vector
can be expressed as:
Its a typical variation of Bayesian formula, where
represents a normal distribution with mean
and covariance matrix
.
means that if
, then this function value is 1; otherwise, it is 0. ML detection can be expressed as:
where
is the result of ML detection,
is the complex transmitted symbol vector taken from an
n-dimensional complex symbol set
,
is the
MIMO complex channel matrix, and
is the channel observation vector.
Equation (
1) is a complex system model. It can be converted to an equivalent real system model [
20]. After transforming the matrix and signal by
the equivalent real system model can be expressed as:
In the real system model, the noise power is
, the average energy of transmitted symbols is
. In the rest of this paper, we adopt the real-valued channel model formulation in (
10). We will use the EP detection algorithm to obtain the detected transmitted signal
from the received signal
. Then we use the neural network to optimize the EP detection algorithm and compare the performance of both in this real-valued channel model.
The detection algorithms commonly used in communication systems are the following categories: linear detection, nonlinear detection, optimal detection algorithm, and so on [
21]. The summary of common MIMO detection algorithms is shown in
Table 1.
3. Expectation Propagation Detection Algorithm
The EP detection algorithm uses Gaussian approximation
to approximate the signal posterior distribution
[
22]. Therefore, the optimal EP solution is that the mean
and variance
of Gaussian approximation
agree with those of the posteriori distribution
. The mean and variance can be expressed as:
It takes
operations to directly calculate the moments of the signal’s posteriori distribution
.
represents a constellation symbol set.
n is the number of transmitting antennas. EP iteration is a polynomial complexity with
n to appreciate (
11) and (
12). When the iteration is over, the EP detection will get the hard decision output. It can be expressed as:
The procedure of the EP detection algorithm is as follows:
First, each non-Gaussian factor is replaced with a non-normalized Gaussian factor
,
can be expressed as:
where
and
are real constants greater than 0.
is a Gaussian distribution with the mean vector
and the covariance matrix
.
and
can be expressed as:
where
,
.
represents the conjugate transpose operation.
represents that if the input is a square matrix, take its diagonal elements as a vector, and if the input is a vector, it is expanded as diagonal elements into a square matrix.
The EP detection updates
iteratively, and the mean
and variance
of the non-normalized Gaussian factor
will gradually approximate the mean and variance of the original signified posterior probability distribution
with polynomial complexity, that is, to get the optimal EP solution [
23]. All parameters related to the index
k of the sent symbol are first initialized before iterations. Initialize
and
.
is the symbol average energy (J/symbol). During each EP iteration,
will be updated in parallel, where
represents the number of rounds of the current EP iteration and
is the maximum number of internal iterations of EP. Given the marginal distribution of the distribution
, namely
, the pair
will be updated iteratively as follows [
24]:
- (1)
Obtain the cavity marginal distribution
based on
, namely approximate logarithmic likelihood function:
It follows a Gaussian distribution with mean and variance .
- (2)
Compute the mean
and variance
of the cavity marginal distribution
:
- (3)
Compute the mean
and variance
of the approximate posterior probability distribution
. The distribution
can be expressed as:
In order to ensure the stability of the algorithm, we set a constraint on
, namely
according to [
10], where
.
- (4)
Update the pair
:
According to the moment matching principle, the second moment of the cavity marginal distribution
is equal to the second moment of the approximate posterior probability distribution
. That is, make the unnormalized Gaussian distribution
as shown in (
23) have mean
and variance
.
During the update process, it may occur that is negative, however, this parameter is an accuracy term and should be positive. The reason for this situation is that there is no pair of can make the variance of approximate to the variance of the original distribution, that is, the moment matching condition cannot be satisfied. In this case, and will not be updated. Just keep the parameter values the same as before the iteration. That is , .
- (5)
In order to improve the robustness of the algorithm,
is smoothed. It can be expressed as:
where
is the smoothness coefficient.
When the change of the mean and covariance of the posterior probability distribution between two iterations is less than or the maximum number of iterations has been reached, the iteration is stopped.
4. Artificial Intelligence-Assisted Expectation Propagation Detection Algorithm
4.1. Application Strategy of Neural Network
In the EP detection algorithm, the most complex and time-consuming step is to calculate the mean vector
and covariance matrix
of Gaussian approximation
. In this process, there is an inversion operation, which will consume a lot of computation resources, as shown in (
15) and (
16). Secondly, in the process of EP inner loop iteration, it is relatively computation-consuming to calculate the mean
and variance
of the cavity marginal distribution
and update the pair
, as shown in (
18), (
19), (
21), and (
22).
Therefore, we try to simplify the above three steps by using a neural network. We take four variables, namely the mean and variance of , and the mean and variance of the cavity marginal distribution , as the input of the neural network, and the mean and variance of the cavity marginal distribution after the iteration as the output of the neural network. The relatively complicated steps in the EP detection process can be skipped, and the probability density function of the normal distribution can be generated directly according to the mean and variance of the output results of the neural network, so as to obtain the detection result.
The essence of a neural network is that the input data are calculated through network nodes to get the output and the intermediate calculation process can be equivalent to a matrix multiplication operation. So, the iteration process of EP detection can actually be equivalent to several matrix multiplication operations with the aid of neural networks, and the operation complexity can be greatly reduced compared with the complex formula calculation directly. Another advantage of neural networks is that no matter how big the data size is, the calculation process is still several steps of matrix operation, which can be effectively calculated in parallel. In the case of large amounts of data, the operation efficiency can be greatly improved.
4.2. Training Data Collection and Model Structure Design
The first step is to collect the required training data and design the model structure on the basis of comprehensive consideration of computational complexity and learning ability.
In the data collection stage, six arrays are initialized before the simulation to store the six variable data to be collected. In the simulation process, the data to be mapped during each EP iteration are saved. However, in the scenario with different SNRs, the magnitude of the data to be mapped by the neural network is quite different. For example, when SNR is 10, the order of magnitude of the variance of cavity marginal distribution , which neural network needs to output, is roughly . Correspondingly, when the SNR is 30, the order of magnitude of is roughly . In order to obtain better training results, SNR is divided into different intervals, and different neural network models are trained on the data with different SNR intervals. We train a total of four models and divide the SNR as follows: we train the first model on the data with , train the second model on the data with , train the third model on the data with , and train the last model on the data with .
Finally, there are 307,200 pieces of data in each SNR interval, and each piece of data has 6 features, that is, the data scale is 307,200 × 6. The first four features are inputs and the last two features are outputs.
The essence of forward propagation of the neural network model is matrix multiplication, and the parameters of the matrix are the weights that the model needs to be trained. The number of hidden layers determines the times of matrix multiplication required in the prediction, and to a large extent determines the algorithm complexity of the model prediction once. In addition, the number of nodes in each layer of the neural network will affect the learning ability of the model. The number of nodes of the model is more, the learning and fitting ability will be stronger, and the amount of data required will be larger; otherwise, the phenomenon of overfitting will occur. If the number of nodes is too small, the learning ability of the model is limited, it is not easy to learn the complex distribution, and it is prone to the underfitting phenomenon.
After many attempts and comprehensive consideration of computational complexity and learning ability, the final structure of the model has four nodes in the input layer, corresponding to four input variables , , and , two hidden layers of thirty-two and sixteen nodes, and two nodes in the output layer, corresponding to the two output variables and .
4.3. Model Training and Application
The neural network model designed above is adopted to learn and map the steps required for optimization, and the trained model parameters are obtained. The process is shown in
Figure 1.
First, the collected training data are imported and divided into input, output, training, and test data. The divided test set accounts for 20% of the total dataset to ensure the generalization ability of the model.
Then, we adopt the z-score method for data standardization to alleviate the impact of data dimension. Z-score standardization will normalize the data into a distribution with a mean of 0 and a standard deviation of 1. It can be expressed as:
where
and
are the mean and standard deviation of the input data of the training set, respectively. We use them to standardize the input data of the training and test set.
is the standardized result.
Then the data format is processed. The Pytorch implementation of the neural network requires the data in tensor format. Data with the tensor format can be computed faster in a neural network and can create higher-dimensional matrices and vectors. Therefore, the dataset needs to be converted to the tensor format. In addition, we convert the data format to float32. The default double data format cannot be trained in the Pytorch framework. Next, a data loader is created to train the data by passing it in batches in an iterative manner.
Finally, the neural network models are built. We adopt multilayer perceptron (MLP) with three fully connected layers, 4 × 32, 32 × 16, and 16 × 2, to realize the learning of data and ensure the simplicity of reasoning. We use ReLU() as the activation function to maximize the speed of computation and training convergence. It can be expressed as:
We use Adam optimizer and the initial learning rate is set to 0.001. The L1 loss function, namely the minimum absolute deviation, is adopted as the error loss function. It represents the average of the sum of the absolute values of the difference between the true value and the predicted value. The model trained with this loss function is more robust to some extent. It can be expressed as:
Environment parameters and neural network training parameters are shown in
Table 2.
After the trained model is obtained, it is necessary to further transform the complex calculation process of the EP detection algorithm into the forward propagation process of the neural network. According to SNR, the corresponding model parameters and the corresponding mean and standard deviation for standardization are imported. In the EP detection iteration process, when obtaining the mean
and variance
of the approximate posterior probability distribution
and the mean
and variance
of the last round cavity marginal distribution
, the format of these four variables is processed into the neural network input format, and then the standardization is carried out. After that, the matrix multiplication of input and the weight parameters of the neural network is performed, that is, the prediction process of the neural network is reproduced. The process of neural network prediction is as follows:
where
n is the num of transmitting antennas. Finally, the output results are the mean
and variance
of the cavity marginal distribution
after one iteration.
5. Simulation Results
In this section, we compare and analyze the proposed detection algorithm and the original EP detection algorithm in a Rayleigh fading MIMO channel.
In the scenario, there are eight transmitting antennas and eight receiving antennas, and the Rayleigh fading MIMO channel with a mean of 0 and variance of 1 is adopted. Cyclic redundancy check addition, code block segmentation, LDPC coding, and rate matching are performed on the transmitted signals in turn. The modulation modes are 64-QAM and 256-QAM for comparison. At the receiving end, rate matching, LDPC, block segmentation, and cyclic redundancy check are decoded in order. We use the original EP detection algorithm and neural network-assisted EP detection algorithm, respectively, to test and compare the simulation results of the bit error rate (BER) of the soft decision.
Figure 2 and
Figure 3 show the comparison of BER detection results after four iterations in the Rayleigh fading MIMO channel scenario with eight transmitting antennas and eight receiving antennas. 64-QAM and 256-QAM are adopted, respectively. “EP” represents the original EP detection. “MLP” represents the EP detection algorithm assisted by the neural network that we propose. When SNR is less than 20 dB, the EP detection algorithm assisted by the neural network can approximate the performance of the original EP algorithm. The results of the fourth iteration of the two algorithms are very close, indicating that the neural network can learn the core steps of the EP algorithm ideally. However, in the case of high SNR (
), the magnitude of the data that the neural network needs to learn is too little, as little as
order of magnitude size. It requires higher precision to fit the mapping data. In addition, the neural network needs to take into account the generalization ability of the model, and the characteristics of the model lead to its own errors. With the improvement of mapping data accuracy requirements, the tolerance of errors existing in neural networks will also decrease. Therefore, the error of the final detection performance is slightly larger, but the error is also within an acceptable range, and the effect of two iterations of the original EP algorithm can still be achieved. The results show that the EP algorithm assisted by neural networks is a feasible strategy.
Since the simulation data of the eight-input, eight-output Rayleigh fading MIMO channel scenario with 64-QAM constellation is used as the training data, in order to verify the applicability of the neural network-assisted EP detection algorithm in more different scenarios, we use a correlated Rayleigh fading MIMO channel scenario of four inputs and eight outputs, which is completely different from the training data, for testing. The correlation coefficient of adjacent antennas is 0.5. The 64-QAM and 256-QAM modulation modes are adopted, respectively. The simulation results are shown in
Figure 4 and
Figure 5.
When SNR is less than 20 dB, the method proposed in this paper can still approximate the original EP algorithm, and the results of the fourth iteration of the two algorithms are very close, indicating that the neural network has a good training effect on the core process of EP detection, and is applicable to different scenarios, with strong generalization ability. In the case of high SNR (), from the perspective of the iterative process, it is found that the performance of the fourth iteration is worse than that of the third iteration once in a while. The reason is that the third iteration has reached a good performance, and the next iteration requires more rigorous calculations to further improve the performance. However, the matrix multiplied by each iteration of the proposed algorithm is the same content, which lacks the effect of fine-tuning. In addition, in the scenario of high SNR, it is difficult for the accuracy of the neural network to meet the requirements, so further iteration may lead to performance degradation. From the perspective of performance, due to the model characteristics of the neural network itself and the improvement of data accuracy requirements, the performance error is slightly larger than that of the original EP algorithm, but still maintains a good detection effect, which fully verifies the applicability of the model for different scenarios.
Next, the complexity is analyzed and compared. For one iteration, the original EP algorithm firstly should calculate (
21) and (
22) to update
. Then, we calculate (
24) and (
25) to smooth
and deal with the case where
is negative. After that, (
15) and (
16) are calculated to find the mean vector
and covariance matrix
of the non-normalized Gaussian approximation
. This process involves matrix inversion and matrix multiplication, the complexity is
. Finally, according to (
18) and (
19), the mean
and variance
of the cavity marginal distribution
are obtained. Therefore, the computational complexity of one iteration of the original EP algorithm is
.
The iterative process of the neural network-assisted EP algorithm only needs to go through three matrix multiplication, matrix addition, and maximizing operations. According to (
29)–(
31), the main computational complexity of this process can be expressed as
. After simplification, the final computational complexity is
. The complexity is greatly reduced compared to MAP and the original EP algorithm. The computational complexity comparison of the three methods is shown in
Table 3.
However, the reduction in computational complexity inevitably takes up additional space. The main additional space occupied in our proposed algorithm comes from the parameters of the neural network models. The space required by a neural network model is three weight matrices with dimensions , , and , and three bias vectors with dimension , , and .
6. Conclusions
In this paper, an EP detection algorithm with outstanding performance and low complexity is optimized. We analyze the steps with high complexity, then use the neural network model to learn and map some time-consuming steps in the EP detection algorithm, and convert the complex operation process into several matrix multiplication operations to reduce the complexity of the detection algorithm. Finally, the performance and applicability of the method are verified in different scenarios. The simulation results show that the neural network-assisted EP detection algorithm can approximate the performance of the original EP algorithm when the SNR is less than 20 dB and can learn the core steps of the EP algorithm more ideally. However, in the case of high SNR (), the magnitude of the data that the neural network needs to learn is too small, as small as orders of magnitude. It requires higher precision to fit the mapping data. In addition, the neural network needs to take into account the generalization ability of the model, and the characteristics of the model lead to its own errors. With the improvement of mapping data accuracy requirements, the tolerance of errors existing in neural networks will also decrease. Therefore, the error of the final detection performance is slightly larger, but the error is also within an acceptable range, and the detection performance still maintains a good level. The results show that EP algorithm assisted by a neural network is a feasible strategy. It can guarantee that the performance is close to the original EP detection algorithm, and the computational complexity can be reduced from the original to .
However, performance can still be improved in scenarios with high SNRs. In the future, we can try to adopt other models to achieve a better mapping effect or use different models for training separately in each iteration so as to mitigate performance loss in scenarios with high SNRs.