Next Article in Journal
Cost-Sensitive Multigranulation Approximation in Decision-Making Applications
Previous Article in Journal
An Adaptive Offset Activation Function for CNN Image Classification Tasks
Previous Article in Special Issue
Content and Other Resources Recommendations for Individuals with Intellectual Disability: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals

1
College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China
2
Gansu Province Internet of Things Engineering Research Center, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(22), 3800; https://doi.org/10.3390/electronics11223800
Submission received: 14 September 2022 / Revised: 30 October 2022 / Accepted: 13 November 2022 / Published: 18 November 2022
(This article belongs to the Special Issue Recommender Systems and Technologies in Artificial Intelligence)

Abstract

:
Multimodal research is an emerging field of artificial intelligence, and the analysis of dangerous driving behavior is one of the main application scenarios in the field of multimodal fusion. Aiming at the problem of data heterogeneity in the process of behavior classification by multimodal fusion, this paper proposes a low-rank multimodal data fusion method, which utilizes the complementarity between data modalities of different dimensions in order to classify and identify dangerous driving behaviors. This method uses tensor difference matrix data to force low-rank fusion representation, improves the verification efficiency of dangerous driving behaviors through multi-level abstract tensor representation, and solves the problem of output data complexity. A recurrent network based on the attention mechanism, AR-GRU, updates the network input parameter state and learns the weight parameters through its gated structure. This model improves the dynamic connection between modalities on heterogeneous threads and reduces computational complexity. Under low-rank conditions, it can quickly and accurately classify and identify dangerous driving behaviors and give early warnings. Through a large number of experiments, the accuracy of this method is improved by an average of 1.76% compared with the BiLSTM method and the BiGRU-IAAN method in the training and verification of the self-built dataset.

1. Introduction

Hazardous driving is a major contributor to vehicle crashes and fatalities, including violent and emotional driving behaviors and a variety of other poor driving behaviors; according to the National Highway Traffic Safety Administration (NHTSA), traffic fatalities in 2021 increased by an estimated 10.5% over the previous year, with urban roadway fatalities rising by 16%. Therefore, hazardous driving behavior detection is essential for effective control strategies for drivers, and the improvement of mass transit safety is potentially valuable.
Typical sensor approaches for wireless perception include Wi-Fi, the Wi-Fi link selection continuous state decision process [1], and the Scalable Uplink Multiple-Access (SUMA) protocol for Wi-Fi backscatter systems [2]. The method improves the performance of the measurement action, in which the CSI (channel state information) [3,4] is an important part of the modeling state. Sensing information from sensors, such as body-worn sensors [5] and the Industrial Internet of Things [6], has led to the development of multi-information fusion algorithms and action-recognition technologies. Among these is video information, which contains multimodal information, such as reflected brightness information, distance, and heat distribution. However, in the complex environment of vehicle driving, deviations in behavior recognition may be caused by environmental brightness distortion and the misjudgment of distance heat information [7]. The traditional array radar can only monitor a single target object point-to-point, while the method based on the FM continuous wave encapsulates the millimeter-wave radar antenna array, which solves the problem of multi-path and complex environments caused by the narrow space inside the vehicle. This offers an excellent solution. Therefore, FMCW has a good prospect in human body pose estimation and microwave neighborhoods [8,9].
The purpose of multimodal fusion is to maintain the original method of the model in combination with multiple effective modalities and unified high-dimensional representation; therefore, the key problem in multimodal data is the existence of data heterogeneity. The normalized high-representative data can reflect the essential features of the original recognition, such as speech recognition fusion facial emotion recognition [10,11,12] and self-supervised learning (SSL), to extract multiple features for emotion recognition. In the process of FMCW radar frame signal preprocessing, features such as Doppler, velocity, scattering point difference, and amplitude are usually extracted separately, and two kinds of excellent features are randomly combined [13,14,15], which may not solve the problem of extracting features given the coupling relationship between each feature and the balance of feature space. In the literature [16,17,18], other methods explore the high-dimensional representation for multi-feature fusion, use the multi-feature fusion algorithm of HSV, LBP, and LSFP to improve high-dimensional fusion representation, and binarize the DM image constructed by multi-feature fusion. Human complex targets monitor small target features, perform algorithm fusion for local entropy, average gradient strength, and other features, and perform peak normalization processing in vector space (MFVS). Micro-action multimodal classification generally uses the Hidden Markov Model [19], more complex actions or gestures using a convolutional neural network (CNN) to verify the public NTU-Microsoft-Kinect through the DisturbIoU algorithm in [20,21], and the Gesture (NTU) dataset and the Vision for Intelligent Vehicles and Applications (VIVA) dataset.
In this paper, we use the FMCW radar-parsing raw data for DAM to fuse micro-Doppler, velocity Doppler, and amplitude features in low-rank tensor representation (LMF) and use the attention mechanism to assign weights in the gating system of the ReLU activation function. This method solves the problems of heterogeneous fusion confusion and feature space balance of multimodal fusion, and makes the following contributions: (1) It fuses the micro-Doppler spectrum with radial velocity and amplitude complex features in a dynamic sparse cascade, and couples and fuses low-dimensional action information features into a high-dimensional linear vector representation that can be represented dynamically, thus improving model stability as well as classification efficiency. (2) By context modeling and using a low-rank tensor network (LMF), the tensor representation is created by computing the outer product of three different single-peaked modes, and modeling the inter-interactions achieves end-to-end learning. (3) The proposed method in GRU-gated cells uses a self-attentive-rectified-linear-unit (AReLU)-based activation in GRU cells, and the performance of FM continuous-wave tensor representations can be improved by the activation function.
The rest of this paper is organized as follows. In the Section 2, an overview of the related methods is provided. In the Section 3, the related principles and technologies of radar information feature extraction and feature fusion are introduced. In addition, the Section 4 describes the implementation principle and parameters of the feature fusion method and classification model LMF-AR-GRU in detail. The Section 5 details the experimental setup and the analysis of the experimental results, and the overall efficiency and advantages of the system are compared and analyzed.

2. Related Work

2.1. Semantic Recognition Methods for Behavior Extraction

Dangerous driving behavior recognition is one of the most intuitive human–computer interaction methods. The behavior recognition methods include visual images, biological signals, inertial sensors, and radar signals. The image-based method captures RGB video information from multiple angles through the camera to identify and make the image sequence of the target heterogeneous. Khan, F.S. et al. used the bag-of-words image representation method in the study of static images [22]. Xin, M. et al. proposed a composite triple-loss function for high-similarity image information to solve the problem of intra- and inter-class confusion in action recognition [23]. Due to the poor light reflection effect in the car due to safety factors, the recognition accuracy is weakened from the input layer. Behavior recognition based on biological signals includes EEG signals, EMG signals, etc., and mainly uses the current information generated during muscle movement in the EMG for identification. Song, J.Y., and others still apply the wearable EMG acquisition system through the exoskeleton. LSTM was used to study the effect of different features of EMG signals on behavioral accuracy [24]. Based on sensors such as inertial sensors, gyroscopes, etc., Sun, P.S. et al. used a quadratic discriminant analysis (QDA) classifier and support vector machine (SVM) classifier through cascade classification using inertial sensors [25]. The degree of dependence is still a problem in in-vehicle perception.
Wu Zhu et al. used the channel state information of a non-wearable and low-cost Wi-Fi signal for feature extraction and classification of driver’s movements [26], but Wi-Fi R.F. signals are more difficult to use in the feature extraction of Doppler information. Therefore, Zhang, Z.L. et al. used a millimeter-wave radar to extract human motion Doppler features and proposed an adaptive extreme learning machine (ELM) method to improve the recognition rate in high-noise environments and obtained good results [27]. In this paper, a 77 GHz millimeter-wave radar was used to solve the limitations of other sensing input layers. The advantages of FMCW radar are simple, including an accurate range and speed measurement, independent of the environment, and a UHF high-bandwidth FM continuous wave, which lead to good speed resolution and distance resolution in in-vehicle target sensing. Figure 1 represents the millimeter-wave identification method equipment and environment.

2.2. Feature Fusion and Classification Methods

Feature fusion is not simply the process of stacking modal features, but also involves decomposing multidimensional patterns to generate a high-quality fusion feature for heterogeneous low-frequency fusion [28]. Dynamic cohesion within modalities is a key issue for classification. Zhang, Z.F. proposed a memory fusion network, a convolutional long short-term memory network with different attention (Conv-LSTM), and a fully connected LSTM (FC-LSTM) fusion proposed Spatial-Temporal Dual Attention Network (STDAN) to address the problem of human action recognition [29]. Du, P.F. proposed the fusion method of Dual Attention Modal Separation Fusion Network (BAMS), which fuses and separates the representations of each modality so that the independent modalities can be directly fused and represented [30]. Abavisani, M. proposed an adversarial encoder–decoder–classifier framework to perform multidimensional clustering of multimodal branches in a convolutional neural network, and introduced adversarial training to make the expression layers of each modality consistent [31]. Since differences can lead to data ambiguity and data loss during the normalization process, Li, M. proposed a tensor fusion network (DTFN), which introduced the proposed features into the tensor fusion layer to increase the spatial correlation. The real-time performance of dangerous driving warnings is extremely strong, and the high-dimensional tensor representation caused by the outer product leads to an exponential increase in the computational complexity of the system. Nie, W.Z. proposed a modality fusion network for individual local dependencies, which contains a bidirectional corpus multi-connection LSTM-FFN fusion network, thus improving the local dependency transfer between feature blocks [32] and avoiding the problem of exploding gradients. Z. Bai proposed a low-rank tensor fusion method whose tensor representation is based on a GRU-gated structure to fuse context-aware multimodal features [33]. Therefore, the context modeling of low-rank tensors can solve the situation of the high complexity of the model and has end-to-end fusion features, leading to a good performance in emotion recognition, action recognition, and gesture recognition [34,35,36,37]. Contextual multimodal integration of advantageous resources allows data flow to be effectively modeled dynamically based on feature modality in LSTM and GRU units. By adjusting the parameters of the fine-tuning layer gating unit, video features can be efficiently fused and identified [38].
Multimodal fusion, differentiated from dynamic connection, and the high complexity of multimodality cause ultra-high latency, which is a key problem. Therefore, this paper proposes a method for recognizing dangerous driving behavior based on millimeter-level frequency-modulated continuous-wave radar. In this paper, the RDM method is used to map the radar data in the velocity dimension to generate micro-Doppler eigenvectors, radial velocity eigenvectors, and amplitude eigenvectors through successive frames. The three vectors are represented by a low-rank tensor fusion representation through local dependency correlation, and a method of combining main features and effective features is adopted, and the high-dimensional data completely show the completeness of dangerous driving actions within the time window while reducing the complexity. The overall model uses the improved attention mechanism LMF-GRU model to improve the reliability of the system and uses the AReLU activation function to extract the fusion representation for target detection in the time series.

3. Driving Behavior Signal Processing

3.1. In-Vehicle Millimeter-Wave Radar Echo Detection

The method of using FMCW radar for human pose estimation involves firstly signal preprocessing and the optimal setting of radar parameters. Secondly, for the micro-motion feature extraction, the echo signal and the IF signal use fast Fourier transform for any one FM period, and the slow time distance Doppler feature of the time slice is obtained by determining the distance FFT of the IF signal in the frame period, and the radial velocity feature is obtained in the extraction of the specific distance and real-time window. The combined amplitude features are obtained by mapping in the horizontal direction, at the same distance and velocity index according to the stitching of the frame time [9].
Linear FM continuous-wave radar makes a fixed difference frequency signal between the echo signal and the transmit signal by modulating the frequency of the transmit signal, and uses the difference frequency signal to obtain the distance information and speed information of the target. The linear FM continuous-wave radar receiver is highly sensitive, and the transmitter power is low in the same noise environment with the same target detection capability. The target echo time delay is much smaller than the time width of the transmit signal, and unlike in the pulse radar, the radar transmitter and receiver are working at the same time, so the FMCW radar is not operating in a blind zone. For close-range detection of driving target behavior, millimeter-level error can be achieved. The basic principle of perception is to determine the distance in that direction from the transmitter to the receiver using the delay of the transmitting and returning waves, so the following expression represents the initial phase of the returning waves.
P r ( t ) = 2 π f 0 t τ + 1 2 T B t τ 2 + φ 0
The instantaneous phase of the differential frequency signal is the phase difference between the transmit phase and the return, expressed as the following expression.
P m ( t ) = 2 π f 0 τ + 2 π T B t τ + π T B τ 2
B is the maximum bandwidth, T is the period of the high-speed FM continuous wave, T c represents the pulse width of the linear FM wave, A T represents the amplitude of the transmit signal, and τ is the difference frequency signal after mixing the time delay of the transmit signal and the return signal. The phase noise of the mixed differential frequency signal in detecting human motion can be neglected, so the differential frequency signal is expressed as the following expression using millimeter-wave matrix sampling.
Y n , m = A R e j 2 π f b n T f + 4 π 1 λ R n T f + m T s

3.2. Feature Vector Extraction

Due to the scattering characteristics of the FM continuous wave, the human behavior facing the radar is reflected by the waveform to generate echoes, the reflected echoes are put into the frequency modulation of the radar and time-frequency analysis, and the kurtosis of the deviated signal distribution in the time domain can be expressed as F k u r = 1 T F v 4 T r t F m 4 d t , The skewness information for the signal symmetry measure in the time domain can be expressed as F s k e w = 1 T F v 3 T r t F m 3 d t , where r t denotes the original data of the received echo, and T denotes the length of the original data time slice. Performing multidimensional Fourier transform in the frequency domain data can prevent the time domain feature quantities from overlapping in the time slice, and feature extraction is performed in the frequency domain. The Fourier transform of the continuous frequency modulation signal of the radar echo is expressed as F ω = F f t = + f t e j ω t d t . In the unstable and irregular signal, the short-time Fourier transform is performed on the signal via the linear transformation method in the extraction process of the driver’s micro-Doppler feature, and the collected original echo signal is defined as x t L 2 R . The short-time Fourier is expressed as the following expression.
S T F T x t , Ω = + x τ g * τ t e j Ω τ d τ = x τ , g τ t e j Ω τ
The resolution in the time domain during processing is determined by the window function w t , and the frequency resolution in the frequency domain is determined by the window function W ω . The enantiomorphic relationship is that the longer the effective length of the window function w t is, the more concentrated the integration efficiency over the time series will be, leading to an increase in Doppler resolution. In this paper, micro-Doppler extraction is performed using pre-processed FM continuous waves by adding time function windows (STFT) for dangerous driving behavior, and micro-Doppler features are obtained by Fourier transform. The micro-Doppler features are used as the main features to identify the dangerous actions, and the micro-Doppler sparse matrix is obtained by mapping the velocity in a time slice sequence. The sensitivity accuracy of micro-Doppler for actions is higher than that of video frames and skeleton key points. The Doppler spectrum dataset at different distances responds to the tiny movements of different echo body points of the target, and the radial velocity features and amplitude information features of the driver at different distances continue to be extracted by different raw data mappings. Figure 2 represents the mapped micro-Doppler spectra of key features, combined into feature datasets at different distances.
In the effective range of radar detection, the motion of the target is described by the direction and velocity at the same distance from the radar. Two-dimensional FFT is carried out on the pre-processed data of the radar to extract the radial velocity of the radar, and the distance of the target point from the radar is determined by the CFAR method according to the fluctuation degree of the signal to decide the set threshold, and the radial velocity of the target point is measured according to the Doppler series information points to find the peak point. A single frame of the time series contains a 128 × 128 sparse matrix, and the original data size will be extracted first without compression. In the original data, the key frames are captured, and no obvious echo state frames or time redundant frames are lost, and the amplitude features are extracted from the original data, which are divided into real and imaginary parts, and the amplitude features will reflect the dense area of the action and have a predictive role for the start and end time series of the action.

3.3. Doppler Vector Fusion Representation

The key problem of multimodal fusion is the dynamic articulation between the features and the normalization problem, which is used to fuse the modalities for the classification normalization problem of the lower layer network, where the method is used to represent the tensor representation, but because of the large number of data collected in the paper to calculate the original multi-peak data of RDM mapping Doppler features, the individual parameters are redundant, and the relevant parameters are non-correlated with multiple data form inputs, so the fusion method of low-rank multimodal can be solved, and the tensor representation is expressed using the outer product between each modality and the interconnection between the modalities, and the data peak X of a single time slice is expressed by the following expression.
X = m m u , m u d u
where the outer product on the set u = 1 U is represented by the index vector u , and m u denotes the number plus one. The input tensor is represented as a linear layer vector g(·) through X d 1 × d 2 × × d u .The tensor is represented as the following expression.
h = g X ; W , b = W X + b   h , b d y
In the expression, W represents the weight of the layer and b represents the bias. When the number of modes is determined, X is represented as a tensor, and the weight of the W order is a U + 1-order tensor of d 1 × d 2 × × d u . In this method, high-efficiency low-rank multimodal fusion of parallel decomposition is used in vector fusion, and the original collected Doppler information is converted from tensor X to M u u = 1 U . This process is decomposed in parallel with the decomposition process of low-rank tensor factors. Tensor representations are expressed in parallel networks through simplification and factorization.
h = i = 1 r u = 1 U w u i X     = i = 1 r u = 1 U w u i X     = u = 1 U u = 1 U w u i m u
Equation (7) analyzes the complexity of the output tensor and the linear symmetric projection. Using this model prevents the disadvantage of inefficient system, and the complexity of the traditional method is O d y m = 1 M d m . This model reduces the complexity of the system to O d y × r × m = 1 M d m , where u = 1 U represents the product of the elements on the tensor sequence t = 1 3 x t = x 1 · x 2 · x 3 . This method represents the tensor fusion of three identical time slice single-peak modes of micro-Doppler features, radial velocity features, and amplitude features. The tensor fusion method for the two modes is represented by the following expressions.
h = i = 1 r w a i w v i X = i = 1 r w a i m a i = 1 r w v i m v

3.4. Recurrent Neural Network for Classification

GRU is a modified version of the recurrent neural network. GRU is a full-gated recurrent unit, and the recurrent neural network is represented using a t , where a is the memory unit, g   is the activation function, x is the input, b is the bias, and t is the time point. In GRU, the memory unit is c, but c ˜ is used instead of c.
a t = g W a a t 1 , x t + b a
c ˜ t = tan h w c c t 1 , x t + b c
A gate that plays an important role in iterative feature memory in the gating unit of GRU is the update gate Γ u . The odd number is 1 and the even number is 0, so the value of Γ u should be placed between 0 and 1. The applicable activation function is ReLU, which is in the ideal state when Γ u is equal to 0, making the t moment and t − 1 moment of the memory unit very similar. The update gate is expressed using the following expression.
Γ u = σ W u c t 1 , x t + b u
The input micro-Doppler features and the other two features are replaced by other parameters in order to approach 1 without affecting the cat, and to solve this problem, the next gate is generated by the following equation.
C t = Γ u C ˜ t + 1 Γ u C t 1

4. Methods and Models

4.1. Overall Methodology Process

As shown in Figure 3, dangerous driving is determined via the radial distance change of the action and signal generation peak, the input original echo with vector C, and the input original echo using the H, W, and T three-dimensional matrix representation X C × H × W × T . In the process of behavior signal acquisition, three kinds of action recognition features are used in information fusion to achieve the classification effect, three kinds of action features in the RDM radial mapping and Fourier transform are used to generate a new feature representation, and the micro-Doppler spectrum key features K C × H × W × T , distance time Doppler velocity spectrum hidden features, amplitude features of identification hidden features, and hidden features are uniformly expressed as Q C × H × W × T . This method firstly uses FMCW millimeter-wave radar to collect the original information on the dangerous people in the car, firstly by using an MTI filter for static clutter filtering, and then Fourier transform for data to accumulate velocity information for different distances. After that, Fourier transform is applied to the data so the velocity information of different distances can be accumulated in the distance–time spectrum, and the window functions of different time slices are STFT-transformed and modulated to obtain the key micro-Doppler features. The three different dimensional features are normalized to find the dynamic articulation between the modes expressed as a high-dimensional unified tensor, and the fused feature tensor is expressed as d 1 × d 2 × × d u . The low-rank multimodal fusion method for different dimensional features is reduced and normalized so as to reduce the model time complexity, which constitutes the temporal distance feature also known as the velocity feature through temporal processing, and each position of the target driving process body is relative to the radar. The distance of each position of the target’s body relative to the transmitting end of the radar will change, due to the distance Doppler generates for the Doppler spectrum peaks. According to the location of the peaks and the temporal sequence, the relative actions can be differentially classified, and the fused multi-level feature data d 1 × d 2 × × d u are input to the AR-GRU through the update gate and reset gate to update the state of the network input parameters, prompting the GRU to learn the weight parameters.
Since this method focuses on studying and explaining the methodical recognition of dangerous driving behavior, the method of testing is not described in too much detail. In the recognition of dangerous driving behavior under vehicle radar monitoring, the most important thing is the lightness of the system, as the system model must quickly detect dangerous driving actions that are happening or will happen soon, and so the system must have a low time complexity. Regarding the running time of the LMF low-rank multimodal model, it is important to talk about the experimentally collected self-built dataset, which is used in this method to evaluate the task. The “Verify” data subset and the “test-clean” dataset are used for the test data. The collected Doppler features are vector-decomposed and divided into multiple frames with a window size of 20 ms and a step size of 10 ms, and the corresponding features are extracted using a feature extractor, and a multidimensional filter is used to normalize the multimodal features to a mean of zero and perform mean variance analysis on the results. For the identification of all dangerous driving behaviors via the recognition of specific actions in experimental verification, the average time needed for the recognition of specific dangerous driving behaviors is 489 ms, which can fully meet the need for a timely warning.

4.2. AM-LMF and AR-GRU

4.2.1. Tensor Representations

The tensor is usually created by taking the outer product over the input modalities. Multimodal fusion is represented as the multilinear function f : V 1 × V 2 × × V M H , where V 1 , V 2 , …, V M is the vector space of input modalities, and the output vector space determines a set of vector representations { z m } m = 1 M that encode the single-peak information of different modalities, and the goal is to integrate the single-modal representations into a compact multimodal representation for downstream tasks. Tensor representation is a method of multimodal fusion, where the input representation is first transformed into a high-dimensional tensor and then mapped back to a low-dimensional output vector space. It is more effective at capturing multimodal interactions than simple concatenation or pooling. In addition, to be able to model interactions between any subset of modalities using a single tensor, this method proposes a single-peaked expanded representation that appends interaction information to single-peaked data. The input tensor Z formed by the single-peaked representation is calculated by the following equation.
Z = m = 1 M z m , z m R d m
The three unimodal two-by-two combinations are combined to form three multimodal features, and all three modalities are also stitched together to form a high-dimensional multimodal feature, and the modal combinations are used as inputs to the relational network to determine the relationship between the modalities. For the unimodal features, the relational network first performs nonlinear transformation operations on different dimensions, and the bimodal features that are combined two by two will represent the fusion results as expressions.
T b = F 4 f k e y f v e l + F 5 f k e y f a m p + F 6 f v e l f a m p
For the three modes of dangerous driving action recognition, all features are fused into high-dimensional modes and normalized using the connection relations of the relational network, and the process and results can be expressed as the following equation.
T c = F 7 f k e y f v e l f a m p
where F 1 , F 2 , F 7 denotes a series of different linear and nonlinear transformations; linear transformation is a fully connected operation, and nonlinear transformation uses the ReLu activation function; f k e y , f v e l , and f a m p denote the micro-Doppler feature, the radial velocity feature, and the amplitude feature, respectively; and represents the splicing operation. Figure 4 represents the whole process of linear transformation.

4.2.2. Low-Rank Tensor Fusion Method

In this paper, a novel low-rank multi-peak fusion model based on a self-attentive mechanism is proposed, which uses a low-rank weight tensor with an attention mechanism to make multi-peak fusion more effective and more globally relevant. The low-rank tensor representation solves the problem arising from feature fusion representation by introducing a weight factor W to decompose the low-rank factor W d 1 × d 2 × × d N to represent the tensor of order d_N. The principle is to perform weight decomposition into N-specific low-rank factors by weighting the weight factor W . The exact decomposition for the tensor is expressed as the following expression.
W m ˜ = i = 1 R n = 1 N w n , m i , w n , m i n d
The effective decomposition process R denotes the rank of this tensor and w n , m i n = 1 N i = 1 R denotes the decomposition weight factor of the original weight tensor based on the rank of the tensor. Giving a fixed value d to the rank R generates w n , m i n = 1 N i = 1 d , which can be decomposed by a specific rank, which can confer more tunability on the model. So, the representation of the multimodal Doppler feature tensor fusion in terms of the weight tensor can be expressed as W = i = 1 R n = 1 N w n i . Expanding the low-rank factor yields a set of decomposed low-rank factors, expressed using the following expression. Therefore, the corresponding low-rank factor is w n , m i n = 1 d .
w n i = w n , 1 i , w n , 2 i , w n , d h i
The unimodal represents Z a   , Z b   ,   Z c   , which is used as the input of the multimodal fusion to generate a new unimodal, which is multimodal-normalized by passing the unimodal input to the three sub-networks f d o p p l e r c o n t e x t , f v e l c o n t e x t , and f a m p c o n t e x t , respectively. Figure 5 represents the process of unimodal to multimodal normalization.

4.2.3. Compatibility Study of Attentional Mechanisms

The method proposed in this paper is an improvement of the low-rank multimodal fusion approach and is an effective integration of the input modality approach based on the low-rank multimodal fusion approach. The method proposes a novel self-attentive mechanism equipped with attentional neural networks with parallelizable computations, lightweight structures, and the ability to capture remote and local dependencies. The key to the attention mechanism approach is the measurement of the correlation between Z n   and q. The compatible function g Z n   , q generates a score, k, which reflects the dependencies between Z n   and q. The score is transformed into a probability by the function softmax, and finally, this probability is used as a weight and applied to the input modalities to improve the correlation and local dependence between the various modalities. The attention mechanism uses the dot product attention compatibility function, expressed as the following expression, where w d 1 and w d 2 are the learnable parameters.
g z n , q = w d 1 z n , w d 2 q
Compared with the model of the traditional attention mechanism, the present model does not introduce redundant parameters and the model has lower complexity and a faster running speed due to its ability to improve accuracy while maintaining low complexity. The tensor is highly expressive but does not scale well to a large number of modalities, so the weights are decomposed into low-rank factors, which reduces the number of parameter bases in the modalities. The low-rank decomposition computes tensor-based fusion by using a parallel decomposition of the low-rank weight tensor and the input tensor. The method in this paper can scale linearly with the number of modes.

4.2.4. AReLU-GRU

The common activation function of the GRU gating unit based on the attention mechanism is the hyperbolic tangent (tanh), although the effective activation function has a high degree of integration of the computational effect, and so the algorithm complexity is not suitable for use in the in-vehicle computing model. This model uses rectified linear units based on attention AReLU is the activation function and AReLU as the use of time-indexed element-by-element detection. Negative factors will be adaptive parameters for the target data to amplify by changing the weight factor method. Using this activation function can prevent the disappearance of Doppler data gradients in tensor processing, and there are also small learning samples with very few parameters and two parameters for fast network training. AReLU in multimodal fusion also helps to capture the interaction information and dependencies between the information, where the attention mechanism can be used to help GRU to focus on the most significant information for the tensor representation, and the use of the ReLU activation function in the GRU gating unit significantly improves the system performance because of the excellent performance that can be demonstrated for this data volume. Therefore, the model optimizes the alpha and beta values of AReLU by starting with alpha = 0.03 and beta = −3. The default values are suitable for tensor fusion and model efficiency, and the parameters are set so that AReLU is similar to the ReLU-like activation function. Since alpha is the weighting factor controlling negative values, a value of alpha = 0.03 with simultaneously negative values of the threshold function has minimal impact on the model. When the validation model of the activation function AReLU in GRU is feasible, the model turns the parameter beta = −3 into beta = 2, and by iteratively conducting experimental validation and using the formula derivation alpha = 0.57, the generalizability of the model in application to behavior recognition is not limited.

5. Experimental Design and Analysis

5.1. Experimental Setup and Environment Variables

The experimental equipment used in this paper included a millimeter-wave radar clock with an amplitude of 77–81 GHz, produced by Texas Instruments, corresponding to a wavelength of about 4 mm and a bandwidth of 4 GHz. The hardware had an ultra-high-frequency bandwidth that is superior to that of other wireless sensing devices, ensuring high-precision recognition of movements. The hardware experimental environment equipment used mainly included an IWR1642 radar sensor device and DCA1000EVM data capture card, sensor bracket, Lenovo R7000 i5-7300HQ GTX1050 802.11AC wireless network card laptop, and two cars and SUVs. In the experimental scenario, on sunny or rainy days, after experimental comparison of the same parameters of the state of the weather and environmental factors, if the millimeter-wave echoes are minimal, they do not merit discussion. An experiment was conducted for vehicles of different sizes, including a three-compartment sedan and off-road cars. The radar sensor was fixed to the B pillar of the vehicle, and at a height of 85 cm from the vehicle chassis, a radar transmitter and receiver were placed perpendicular to each other facing the target driver, and the driver location was just within the radar scanning range so that the target dangerous behavior characteristics could be obtained. The experimental setup is shown in Figure 6.
The experimental equipment and environment used for collecting information included a Windows 10 Intel Core i7-6700 CPU processor, with a main frequency of 3.4 GHz, a memory size of 32 GB, a GTX1080 graphics card, and PyCharm2018 software; the programming environment was Python 3.7. In this model, in the compiled environment Python 3.7, MemTotal indicated the actual amount of RAM occupied by the system, using physical memory minus some reserved bits and the size of the kernel’s binary code. MemTotal produced 816,136 kb in the environment command in the proc file using cat /proc/partitions. The partitions of the system itself were identified in the proc file as cat /proc/partitions. In the 10 partitions, from 0 to 9, the partition capacity was summed to obtain the ROM capacity of 1,563,645 kb. The data acquisition equipment used included a Logitech-wired Wi-Fi surveillance camera with 720P/1080P high-definition wide-angle images and was easy to install. After comparing and analyzing the data collected from multiple placement angles and positions, the camera installed in the front of the co-pilot position was found to have the best effect, and the sitting posture image data for the front-seat driver and occupant in the car were obtained.
Before setting the parameters, the speed and distance of the moving target in the car were first estimated, because an estimated maximum speed or distance that is too large will lead to a significant decrease in speed resolution. A suitable sampling time setting can ensure the completeness of an action sample and will not produce redundancy. The parameters of the millimeter-wave radar were set, as shown in Table 1. In the sampling process, to ensure the integrity of the behavior, the frame period was set to 50 ms, and the number of frames was set to 100 to achieve a sampling time of 5 s for each action, which ensured the integrity of the action and reduced the blank redundant time slices, so each unit was set to 256 samples and 128 Chirp.
In terms of the basic conditions, the experiment was conducted with 6 known experimental subjects, 3 males and 3 females with different heights and weights, with an age range of 23 to 55 years, a height range of 152 cm to 190 cm, and a weight range of 43 kg to 97 kg. Six random subjects with unknown information were used as comparisons. The sample set for this experiment consisted of 9 risk behavior datasets, with a sample acquisition time of 5 s and one specific action that was repeated 60 times. The dataset contained 6480 samples, and each data sample acquisition set was 255 × 255 × 3, the Doppler distance was set to 0.3–1.1 m, the unit was 0.1 m, and target sensing was performed at 9 distances.
Due to the special nature of in-vehicle recognition, radar recognition of dangerous movements can only be realized for the upper body due to obscurity, and leg movements could not be identified due to obscurity.

5.2. Experimental Analysis

5.2.1. Model Performance Analysis for Visual Verification

While capturing the driver’s wireless FM signal, the dual-channel processing of RGB image information followed by determining human skeletal key points is essential for describing human posture. Therefore, human skeletal key point detection is the basis for many computer vision tasks, such as motion classification, abnormal behavior detection, and autonomous driving, among others. This experiment validated the use of surveillance equipment to capture images of front-row drivers and passengers, and used the AlphaPose model to obtain the coordinates of 13 skeletal key points in the upper body: left ear, left eye, nose, right ear, right eye, left shoulder, right shoulder, left elbow, right elbow, left wrist, right wrist, left hip, and right hip. These skeletal key points and the line segments connected between these points were used to construct a human sitting posture, as shown in Figure. The dataset was divided into four parts: training (70%), validation (10%), test A (10%), and test B (10%). In the car scenario, the driver and passenger in the car were sitting most of the time, and the behavioral actions of the front passenger were composed of the movement of the upper body, while the movement of the lower body could be ignored. The motion and coordinate position information for the upper body skeleton key points was used to analyze and identify the abnormal behavior of the driver and passenger in the car. Figure 7 shows the predicted behavior of the skeleton key points for a specific action.
To extract the information of the key points of the skeleton, the change in distance between the key points is used to determine whether any of the nine dangerous actions can be observed, but the recognition rate is low, easily affected by other people’s shadows and poor lighting environment, and the average overall recognition rate for the nine behaviors was 73.14%; however, the millimeter-wave radar had the highest recognition rate at 95.74%. Therefore, the model has better recognition accuracy than other techniques used to recognize dangerous behavior within vehicles. Due to the larger internal space of the off-road vehicle, the noise rate of the recognition effect was slightly reduced, and the recognition accuracy inside the off-road vehicle according to experimental comparison was 0.641% higher than the accuracy obtained for the same individuals and the same parameters in the small car. Figure 8 presents a comparison of the experimental environment and equipment used in the car vs. the off-road vehicle. The off-road vehicle was obviously larger than the car, and there was a difference in the recognition effect.

5.2.2. Analysis of Cross-Person Results

The experimental setup included six people with different physical characteristics differing widely in height from 152 cm to 190 cm, including three males and three females. In FMCW-radar-based cross-person movement recognition, drivers with different physiques and heights may produce large differences in recognition accuracy because the radar is sensitive to distance and has different responses to driving targets regarding the three features. The six participants were given different letter labels, a, b, c, d, e, and f, to indicate different people. For each experiment, the dispersion of each set of data was demonstrated by visualizing the data nodes of the six targets to improve the generality of the experiment and determine a set of data for the six unknown participants for comparison, and a box plot graph was created to show each target median and magnitude. From the box plot diagram, it can be seen that good behavior recognition was achieved for every individual. Figure 9 presents the box plot diagram.

5.2.3. Analysis of Results at Different Distances

In behavior detection, the physiological characteristics of the tested participants will directly affect the recognition accuracy, including the weight and height of the participants, which will affect the distance of the target to the radar transmitter. The experiment aimed to reflect the millimeter-wave radar FM continuous-wave signal for different distances and determine the speed and the distance to the target. This method used nine time detection nodes, the unit of distance was 10 cm, and the resolved data for the distances of 0.3 m, 0.4 m, 0.5 m, 0.6 m, 0.7 m, 0.8 m, 0.9 m, 1.0 m, and 1.1 m were experimentally. The results obtained after collecting the Doppler spectrum of different distances show that the Doppler spectrum of the distance of 0.3–0.5 m had no obvious echo and radial velocity, indicating that there was no reflected signal of the target at this distance, and thus the actual action state of the driver cannot be reflected, so the data stream and tensor of this distance were eliminated through data screening. However, at the distance of 0.5–1.1 m from the radar transmitter, obvious behavioral Doppler feature data could be detected, and the recognition rate for the distance of 0.3–0.5 m did not exceed 20% as reflected by the phenomenon shown in Figure 10a, and the Doppler spectrum of this distance cannot be used as evaluation data; however, the recognition rate for driver behavior at the distance of 0.5–1.1 m reached an average of 92.1%. The CDF plot in Figure 10b reflects that under the same experimental conditions, three groups of targets with different physiological characteristics were set: target a, target b, and target c. Target a was taller and heavier than the other two targets, so it can be concluded that target a achieved a higher recognition accuracy at a smaller distance. However, target c was smaller than the other two targets and also lighter, so the best recognition accuracy was achieved at the sensing distance of 0.8–1.1 m. Therefore, cross-person hazardous driving perception must consider the uncertainty of target body characteristics.

5.2.4. Analysis of Classification Models under Different RNNs

Both GRU and LSTM are sub-branches of RNN. However, the fatal disadvantage of RNN is that the gradient disappears, and data explosion occurs in the process of processing data flow, so the introduction of a gating mechanism can effectively solve this problem. GRU controls the gradient of the input information of the upper layer and calculates the information flow of the hidden layer by updating the gate. In terms of parameters, the activation function that causes the gradient to disappear is replaced by a rectified linear (ReLU) function, which is initialized by an unbounded activation function and identities weights. LSTM is composed of two gate parameters (excluding gate and input gate), and the parameter control is completely different from that of GRU. In information processing, neither the information from the previous moment nor the information from the current hidden layer is selected, while GRU can choose. Additionally, the most important thing in the perception of human body posture in the car is a timely early warning to reduce the time complexity. Although GRU has one less gate than LSTM, GRU has faster convergence and fewer parameters. The accuracy of the same action of GRU reflected from the confusion matrix is slightly higher than that of LSTM parameters. The highest accuracy rate was 96%. The lowest accuracy rate was 88%, and the comparison of the confusion matrix shows that there were similar correlations between different actions, including similar misjudgment outputs, with the highest being 19%, and the output classification of the model may overlap. Additionally, multi-parameter LSTM resulted in higher-similarity output than GRU for the classification of adjacent actions. From the following matrix, it can be seen that the advantages of GRU are more obvious, and it is more suitable for the processing of radar information collected in the car. The confusion matrix is shown in Figure 11. The confusion matrix indicates the effect of the classifier on the test dataset, and the accuracy of the recognition effect for different behaviors is indicated using a gradient from white to blue. High accuracy for the same behavior is indicated using blue.

5.2.5. Analysis of Combined Modalities

A single mode cannot respond to the integrity movement of physiological action in the human body recognition process, so the feature input of the model in this paper was a three-channel input, with micro-Doppler spectral features as the main recognition input features and radial velocity and amplitude features as hidden features, and three single modes were combined two by two to form three multimodal peaks, and all three modes were also spliced together to form a high-dimensional multimodal feature, with the modal combination as the input to the relational network to determine the relationship between the modes. For the unimodal features, the relational network first transforms them nonlinearly before stitching them together. The model puts the single modality into the GRU model alone for the attention weight decomposition method for recognition, and the single modality is compared and analyzed overall with the two-by-two combined modalities and the fused multimodal, and it can be derived from the three line graphs in the figure that the single modality recognition was the worst, followed by the double dominant modal fusion, and finally the multimodal multi-peak representation. The lowest recognition rate was 76.2% and the highest recognition rate was 84.9% for unimodality, the lowest recognition rate was 84.6% and the highest recognition rate was 89.8% for double dominance modal fusion, and the lowest recognition rate was 91.7% and the highest recognition rate was 95.7% for multimodal multi-peak representation. The three-fold line data statistics indicate that there were no overlapping points, so the accuracy for the same action was in line with the overall modal accuracy, so the line graph indicates the quantitative relationship between the modal fusion method and the number, and the influence of the dependency between the modalities on the accuracy rate is shown in Figure 12.

5.2.6. Model and Related Model Performance Analysis

A common approach in multimodal corpus fusion is the use of BiLSTM models with different activation functions in frame-by-frame features using attentional operational layers and LSTMs with more parameters. The second most common interaction-aware attentional network is the IAAN contextual baseline-aware model BiGRU-IAAN; this model is based on the GRU network, and in the comparison of this model, using weighted accuracy (WA) and unweighted accuracy (UA) to compare the overall accuracy of each model, the AReLU proposed in this paper is significantly better than the tanh activation function in the absolute value of WA and the absolute value of UA, and it is clear from the analysis in Figure 12 that the accuracy of the LMF-AR-GRU model is higher than that of the BiGRU-IAAN and BiLSTM models both in terms of UA and WA. The method of integrating GRU units with a multimodal fusion of Attention ReLU activation proposed in this paper can improve the interaction between features, capture driver actions in remote interaction recognition, and solve the problem of BiLSTM gradient disappearance and information explosion in a series of experimental validation work to optimize the AReLU value so that the AReLU is similar to the ReLU. The AReLU activation function can control values below 0. AReLU effectively improves the absolute values of UA and WA relative to the unlearnable BiGRU-IAAN baseline approach, and there are clear experiments showing that suppressing negative values using the default AReLU weighting parameters negatively affects performance and brings them closer to 0. The control unit components of AR-GRU are AReLU ( α = 0.01 , β = 4 ), the specific parameter value that returns the AReLU method to its optimal state. The model uses this set parameter value, and the data representation can reach up to 84.6% unweighted accuracy (UA) and 89.7% weighted accuracy (WA). This variant of the model parameters improves UA and WA by only 0.6% and 0.7%, respectively, relative to the IAAN baseline. The results of this experiment suggest that the AReLU unit does help to improve the overall accuracy of the model, but requires empirical determination of the ideal parameter values. This ensures that negative values close to 0 do not adversely affect performance. The effectiveness of AR-GRU in a dangerous driving behavior data classification setting is demonstrated by the validation of the experimental accuracy rates. Figure 13 represents the visual representation of the accuracy of different technical methods, both underweighted and unweighted, and it can be seen that the proposed method in this paper demonstrates good recognition results.

6. Conclusions

In this paper, we explore the method of multi-featured low-rank multimodal fusion based on the use of FMCW radar, which decomposes Doppler information into different dimensions of information data by DAM, and the method generates multidimensional Doppler spectral information via Fourier change and then mapping to different vectors. Due to a large amount of information and the poor fit of the Doppler matrix, the information is normalized using specific low-rank multimodal (LMF) processing. The low-rank multimodality is linearly proportional to the number of modes and achieves competitive results in different multimodal tasks. The data of the tensor are fed into the proposed AR-GRU approach, and then the nonlinear fusion of the weight decomposition is performed by the attention unit. Based on multiple low-rank tensors to represent the dense weight matrix within the RNN layer, the present model is evaluated based on the millimeter-wave action self-built dataset, which significantly outperformed the conventional model in terms of dynamic articulation of low-rank fusion. The three Doppler features exhibited excellent weight cascades, the fusion tensor completely interpreted the complete limb echoes for the 5 s acquisition time from the beginning to the end of the action, and the redundant parameters could be reduced in the driving environment to reduce temporal complexity and achieve a timely warning and correction of dangerous driving behaviors. After this study, we will consider various variables to enrich the model input, and try out some new network architectures by manually adjusting the parameters and selecting the network architecture to improve the finer granularity of the recognition effect and reduce the overfitting effect of the recognition network.

Author Contributions

Conceptualization, Z.H. and Z.L.; methodology, Z.L.; software, Z.L.; validation, Z.L., X.D., Z.M. and G.L.; formal analysis, Z.L.; investigation, Z.L.; resources, Z.L.; data curation, Z.L.; writing—original draft preparation, Z.L.; writing—review and editing, Z.L.; visualization, Z.L.; supervision, Z.H.; project administration, Z.H.; funding acquisition, Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant 62262061, Grant 62162056), Key Science and Technology Support Program of Gansu Province (Grant 20YF8GA048), 2019 Chinese Academy of Sciences “Light of the West” Talent Program, Science and Technology Innovation Project of Gansu Province (Grant CX2JA037, 17CX2JA039), 2019 Lanzhou City Science and Technology Plan Project (2019-4-44), 2020 Lanzhou City Talent Innovation and Entrepreneurship Project (2020-RC-116), and Gansu Provincial Department of Education: Industry Support Program Project (2022CYZC-12).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shen, X.; Guo, L.; Lu, Z.; Wen, X.; Zhou, S. WiAgent: Link Selection for CSI-Based Activity Recognition in Densely Deployed Wi-Fi Environments. In Proceedings of the 2021 IEEE Wireless Communications and Networking Conference (WCNC), Nanjing, China, 29 March–1 April 2021; pp. 1–6. [Google Scholar]
  2. Kwon, J.H.; Zhang, X.; Kim, E.J. Scalable Wi-Fi Backscatter Uplink Multiple Access for Battery-Free Internet of Things. IEEE Access 2021, 9, 30929–30945. [Google Scholar] [CrossRef]
  3. Falling, C.; Stebbings, S.; Baxter, G.D.; Gearry, R.B.; Mani, R. Criterion validity and discriminatory ability of the central sensitization inventory short form in individuals with inflammatory bowel diseases. Scand. J. Pain 2021, 21, 577–585. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, Z.; Zhang, L.; Ding, Z. An efficient deep learning framework for low rate massive MIMO CSI reporting. IEEE Trans. Commun. 2020, 68, 4761–4772. [Google Scholar] [CrossRef]
  5. Zhang, X. Application of human motion recognition utilizing deep learning and smart wearable device in sports. Int. J. Syst. Assur. Eng. Manag. 2021, 12, 835–843. [Google Scholar] [CrossRef]
  6. Serror, M.; Hack, S.; Henze, M.; Schuba, M.; Wehrle, K. Challenges and opportunities in securing the industrial internet of things. IEEE Trans. Ind. Inform. 2020, 17, 2985–2996. [Google Scholar] [CrossRef]
  7. Cao, H.; Xu, Y.; Yang, J.; Mao, K.; Yin, J.; See, S. Effective action recognition with embedded key point shifts. Pattern Recognit. 2021, 120, 108172. [Google Scholar] [CrossRef]
  8. Chandra, A.; Rahman, A.U.; Ghosh, U.; Garcia-Naya, J.A.; Prokeš, A.; Blumenstein, J.; Mecklenbraeuker, C.F. 60-GHz millimeter-wave propagation inside bus: Measurement, modeling, simulation, and performance analysis. IEEE Access 2019, 7, 97815–97826. [Google Scholar] [CrossRef]
  9. Arsalan, M.; Santra, A.; Will, C. Improved contactless heartbeat estimation in FMCW radar via Kalman filter tracking. IEEE Sens. Lett. 2020, 4, 1–4. [Google Scholar] [CrossRef]
  10. Ma, J.; Sun, Y.; Zhang, X. Multimodal emotion recognition for the fusion of speech and EEG signals. Xi’an Dianzi Keji Daxue Xuebao/J. Xidian Univ. 2019, 46, 143–150. [Google Scholar]
  11. Siddiqui MF, H.; Javaid, A.Y. A multimodal facial emotion recognition framework through the fusion of speech with visible and infrared images. Multimodal Technol. Interact. 2020, 4, 46. [Google Scholar] [CrossRef]
  12. Siriwardhana, S.; Kaluarachchi, T.; Billinghurst, M.; Nanayakkara, S. Multimodal emotion recognition with transformer-based self supervised feature fusion. IEEE Access 2020, 8, 176274–176285. [Google Scholar] [CrossRef]
  13. Ding, W.; Guo, X.; Wang, G. Radar-based human activity recognition using hybrid neural network model with multidomain fusion. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 2889–2898. [Google Scholar] [CrossRef]
  14. Gao, Y.; Zhou, Y.; Wang, Y.; Zhuo, Z. Narrowband Radar Automatic Target Recognition Based on a Hierarchical Fusing Network With Multidomain Features. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1039–1043. [Google Scholar] [CrossRef]
  15. Li, H.; Liang, X.; Shrestha, A.; Liu, Y.; Heidari, H.; Le Kernec, J.; Fioranelli, F. Hierarchical sensor fusion for micro-gesture recognition with pressure sensor array and radar. IEEE J. Electromagn. RF Microw. Med. Biol. 2019, 4, 225–232. [Google Scholar] [CrossRef] [Green Version]
  16. Zhu, M.; Yu, X. Multi-feature fusion algorithm in VR panoramic image detail enhancement processing. IEEE Access 2020, 8, 142055–142064. [Google Scholar] [CrossRef]
  17. Ma, W.; Wang, X.; Xu, G.; Liu, Z.; Yin, Z.; Xu, Y.; Wu, H.; Baklaushev, V.P.; Peltzer, K.; Sun, H.; et al. Distant metastasis prediction via a multi-feature fusion model in breast cancer. Aging (Albany NY) 2020, 12, 18151. [Google Scholar] [CrossRef]
  18. Lou, Y.; Wang, Z.; Li, X. Detection of small infrared targets based on multi-feature fusion. Infrared Laser Eng. 2007, 36, 395. [Google Scholar]
  19. Bicego, M.; Murino, V. Investigating hidden Markov models’ capabilities in 2D shape classification. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 281–286. [Google Scholar] [CrossRef]
  20. Wu, X.; Zhang, J.; Xu, X. Hand gesture recognition algorithm based on faster R-CNN. J. Comput.-Aided Des. Comput. Graph. 2018, 30, 468–476. [Google Scholar] [CrossRef]
  21. Wang, Y.; Yang, Y.; Zhang, P. Gesture Feature Extraction and Recognition Based on Image Processing. Trait. Du Signal 2020, 37, 873–880. [Google Scholar] [CrossRef]
  22. Khan, F.S.; Muhammad Anwer, R.; Van De Weijer, J.; Bagdanov, A.D.; Lopez, A.M.; Felsberg, M. Coloring action recognition in still images. Int. J. Comput. Vis. 2013, 105, 205–221. [Google Scholar] [CrossRef] [Green Version]
  23. Xin, M.; Zhang, H.; Yuan, D.; Sun, M. Learning discriminative action and context representations for action recognition in still images. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 757–762. [Google Scholar]
  24. Song, J.; Zhu, A.; Tu, Y.; Huang, H.; Arif, M.A.; Shen, Z.; Zhang, X.; Cao, G. Effects of different feature parameters of semg on human motion pattern recognition using multilayer perceptrons and LSTM neural networks. Appl. Sci. 2020, 10, 3358. [Google Scholar] [CrossRef]
  25. Sun, P.S.; Xu, D.; Mai, J.; Zhou, Z.; Agrawal, S.; Wang, Q. Inertial sensors-based torso motion mode recognition for an active postural support brace. Adv. Robot. 2020, 34, 57–67. [Google Scholar] [CrossRef]
  26. Ma, Y.; Zhou, G.; Wang, S. WiFi sensing with channel state information: A survey. ACM Comput. Surv. (CSUR) 2019, 52, 1–36. [Google Scholar] [CrossRef] [Green Version]
  27. Zhang, Z.; Meng, W.; Song, M.; Liu, Y.; Zhao, Y.; Feng, X.; Li, F. Application of multi-angle millimeter-wave radar detection in human motion behavior and micro-action recognition. Meas. Sci. Technol. 2022, 33, 105107. [Google Scholar] [CrossRef]
  28. Zhang, L.; Wang, Z.; Qu, C.; Yang, F.; Lv, S. Research on hybrid fusion algorithm for multi-feature among heterogeneous image. Infrared Phys. Technol. 2020, 104, 103110. [Google Scholar] [CrossRef]
  29. Zhang, Z.; Lv, Z.; Gan, C.; Zhu, Q. Human action recognition using convolutional LSTM and fully-connected LSTM with different attentions. Neurocomputing 2020, 410, 304–316. [Google Scholar] [CrossRef]
  30. Du, P.; Gao, Y.; Li, X. Bi-attention Modal Separation Network for Multimodal Video Fusion. In Proceedings of the International Conference on Multimedia Modeling; Springer: Cham, Switzerland, 2022; pp. 585–598. [Google Scholar]
  31. Abavisani, M.; Patel, V.M. Deep multimodal subs44pace clustering networks. IEEE J. Sel. Top. Signal Process. 2018, 12, 1601–1614. [Google Scholar] [CrossRef] [Green Version]
  32. Nie, W.; Yan, Y.; Song, D.; Wang, K. Multi-modal feature fusion based on multi-layers LSTM for video emotion recognition. Multimed. Tools Appl. 2021, 80, 16205–16214. [Google Scholar] [CrossRef]
  33. Bai, Z.; Chen, X.; Zhou, M.; Yi, T.; Chien, W.C. Low-rank Multimodal Fusion Algorithm Based on Context Modeling. J. Internet Technol. 2021, 22, 913–921. [Google Scholar]
  34. Sahay, S.; Okur, E.; Kumar, S.H.; Nachman, L. Low rank fusion based transformers for multimodal sequences. arXiv 2020, arXiv:2007.02038. [Google Scholar]
  35. Pan, Y.; Xu, J.; Wang, M.; Ye, J.; Wang, F.; Bai, K.; Xu, Z. Compressing recurrent neural networks with tensor ring for action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4683–4690. [Google Scholar]
  36. Savvaki, S.; Tsagkatakis, G.; Panousopoulou, A.; Tsakalides, P. Matrix and tensor completion on a human activity recognition framework. IEEE J. Biomed. Health Inform. 2017, 21, 1554–1561. [Google Scholar] [CrossRef] [PubMed]
  37. Zhang, H.; Patel, V.M.; Chellappa, R. Robust multimodal recognition via multitask multivariate low-rank representations. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; Volume 1, pp. 1–8. [Google Scholar]
  38. Arachchi, S.K.; Shih, T.K.; Lin, C.Y.; Wijayarathna, G. Deep learning-based firework video pattern classification. J. Internet Technol. 2019, 20, 2033–2042. [Google Scholar]
Figure 1. Target driver radar echo model.
Figure 1. Target driver radar echo model.
Electronics 11 03800 g001
Figure 2. Distance target of 0.30 m to 1.10 m for the micro-Doppler characteristic spectrum.
Figure 2. Distance target of 0.30 m to 1.10 m for the micro-Doppler characteristic spectrum.
Electronics 11 03800 g002
Figure 3. Behavior recognition model with low-rank fusion weight analysis of multicascade Doppler features.
Figure 3. Behavior recognition model with low-rank fusion weight analysis of multicascade Doppler features.
Electronics 11 03800 g003
Figure 4. Multimodal attention mechanism for low−rank factor linear weight representation.
Figure 4. Multimodal attention mechanism for low−rank factor linear weight representation.
Electronics 11 03800 g004
Figure 5. Multimodal low-rank factor normalized representation process.
Figure 5. Multimodal low-rank factor normalized representation process.
Electronics 11 03800 g005
Figure 6. Multimodal low-rank factor normalized representation process.
Figure 6. Multimodal low-rank factor normalized representation process.
Electronics 11 03800 g006
Figure 7. (a) Right turn (H) key point marker movement prediction; (b) answering the phone (E) key point marker action prediction.
Figure 7. (a) Right turn (H) key point marker movement prediction; (b) answering the phone (E) key point marker action prediction.
Electronics 11 03800 g007
Figure 8. (a) Sedan experiment; (b) off-road vehicle experiment.
Figure 8. (a) Sedan experiment; (b) off-road vehicle experiment.
Electronics 11 03800 g008
Figure 9. (a) Box plot diagrams for six known experimental subjects; (b) box plots for the six unknown experimental subjects.
Figure 9. (a) Box plot diagrams for six known experimental subjects; (b) box plots for the six unknown experimental subjects.
Electronics 11 03800 g009
Figure 10. (a) Recognition effect at different distances; (b) cross-person identification CDF chart.
Figure 10. (a) Recognition effect at different distances; (b) cross-person identification CDF chart.
Electronics 11 03800 g010
Figure 11. (a) Accuracy under GRU parameters; (b) accuracy under LSTM parameters.
Figure 11. (a) Accuracy under GRU parameters; (b) accuracy under LSTM parameters.
Electronics 11 03800 g011
Figure 12. Dependencies of different modal combinations.
Figure 12. Dependencies of different modal combinations.
Electronics 11 03800 g012
Figure 13. (a) Model weighted accuracy (WA) histogram; (b) model unweighted accuracy (UA) histogram.
Figure 13. (a) Model weighted accuracy (WA) histogram; (b) model unweighted accuracy (UA) histogram.
Electronics 11 03800 g013
Table 1. Parameter setting for effective driving behavior perception.
Table 1. Parameter setting for effective driving behavior perception.
Radar ParameterUnitsNumerical Value
Effective bandwidthMHz3411.25
Maximum distance errorcm4
Maximum measurement m/s6
Speed resolutionm/s0.095
Frequency modulation slopeMHz/μs66.626
Frame periodms50
Number of samples1256
Chirp1128
The data in the table are the optimal parameters for the experiment.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hao, Z.; Li, Z.; Dang, X.; Ma, Z.; Liu, G. MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals. Electronics 2022, 11, 3800. https://doi.org/10.3390/electronics11223800

AMA Style

Hao Z, Li Z, Dang X, Ma Z, Liu G. MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals. Electronics. 2022; 11(22):3800. https://doi.org/10.3390/electronics11223800

Chicago/Turabian Style

Hao, Zhanjun, Zepei Li, Xiaochao Dang, Zhongyu Ma, and Gaoyuan Liu. 2022. "MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals" Electronics 11, no. 22: 3800. https://doi.org/10.3390/electronics11223800

APA Style

Hao, Z., Li, Z., Dang, X., Ma, Z., & Liu, G. (2022). MM-LMF: A Low-Rank Multimodal Fusion Dangerous Driving Behavior Recognition Method Based on FMCW Signals. Electronics, 11(22), 3800. https://doi.org/10.3390/electronics11223800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop