Next Article in Journal
A New Seminorm for d-Tuples of A-Bounded Operators and Their Applications
Next Article in Special Issue
Optimal Robust Tracking Control of Injection Velocity in an Injection Molding Machine
Previous Article in Journal
New Results on Finite-Time Synchronization Control of Chaotic Memristor-Based Inertial Neural Networks with Time-Varying Delays
Previous Article in Special Issue
A Novel Zeroing Neural Network for Solving Time-Varying Quadratic Matrix Equations against Linear Noises
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Appearance-Based Gaze Estimation Method Using Static Transformer Temporal Differential Network

1
Guangxi Colleges and Universities Key Laboratory of AI Algorithm Engineering, School of Artificial Intelligence, Guilin University of Electronic Technology, Jinji Road, Guilin 541004, China
2
School of Artificial Intelligence, Guilin University of Electronic Technology, Jinji Road, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(3), 686; https://doi.org/10.3390/math11030686
Submission received: 3 December 2022 / Revised: 19 January 2023 / Accepted: 25 January 2023 / Published: 29 January 2023

Abstract

:
Gaze behavior is important and non-invasive human–computer interaction information that plays an important role in many fields—including skills transfer, psychology, and human–computer interaction. Recently, improving the performance of appearance-based gaze estimation, using deep learning techniques, has attracted increasing attention: however, several key problems in these deep-learning-based gaze estimation methods remain. Firstly, the feature fusion stage is not fully considered: existing methods simply concatenate the different obtained features into one feature, without considering their internal relationship. Secondly, dynamic features can be difficult to learn, because of the unstable extraction process of ambiguously defined dynamic features. In this study, we propose a novel method to consider feature fusion and dynamic feature extraction problems. We propose the static transformer module (STM), which uses a multi-head self-attention mechanism to fuse fine-grained eye features and coarse-grained facial features. Additionally, we propose an innovative recurrent neural network (RNN) cell—that is, the temporal differential module (TDM)—which can be used to extract dynamic features. We integrated the STM and the TDM into the static transformer with a temporal differential network (STTDN). We evaluated the STTDN performance, using two publicly available datasets (MPIIFaceGaze and Eyediap), and demonstrated the effectiveness of the STM and the TDM. Our results show that the proposed STTDN outperformed state-of-the-art methods, including that of Eyediap (by 2.9%).

1. Introduction

Gaze direction is important non-invasive human behavioral information, and can be an important cue for understanding human intention, making it valuable in human-behavior-related fields, such as human–computer interaction [1,2,3], autonomous driving [4,5], and virtual reality [6,7]. Gaze estimation tasks can be divided into three main categories: (1) 3D gaze direction estimation [8,9]; (2) target estimation [10,11]; (3) tracking estimation [12,13]. In this paper, we focus on 3D gaze direction estimation.
3D gaze estimation methods can be divided into model-based gaze estimation methods and appearance-based gaze estimation methods, as shown in Figure 1. Model-based gaze estimation methods [14,15,16,17] usually consider the eye’s geometric features, such as the eyeball shape and the pupil position, using machine learning methods—for example, the support vector machine—to predict gaze direction; however, model-based gaze estimation methods usually require specific detection equipment, making them unpopular to use in real-world environments. Conversely, compared to model-based gaze estimation methods, appearance-based gaze estimation methods do not require specific detection equipment. Such methods can predict gaze direction, using face or eye images: however, these methods depend on large volumes of trainable data. Zhang et al. [18] first applied deep learning in the gaze estimation field: since then, many researchers [18,19,20,21] have proposed appearance-based deep learning gaze estimation methods and, more recently, a new pipeline of gaze estimation has been proposed [22]. The new pipeline of appearance-based gaze estimation includes static feature extraction, dynamic feature extraction, and gaze estimation: such methods are usually highly accurate, even when head pose and illumination change.
There are several significant problems with appearance-based deep learning gaze estimation methods: for example, the feature fusion stage has not been fully considered. Recent research [23,24] has proposed several feature fusion strategies to help settle this issue. In this paper, we considered the feature fusion stage, applying the self-attention mechanism [25] to fuse the coarse-grain face feature and fine-grain eye features. The self-attention mechanism was able to effectively learn interrelation among different features, and retain more important features. As far as we know, our study is the first time that the self-attention mechanism for feature fusion has been applied in appearance-based deep learning gaze estimation methods. The self-attention mechanism [25] was first proposed in natural language processing (NLP), the vision transformer (ViT) [26] making the connection between the self-attention mechanism and computer vision (CV). The ViT has two variants: the pure-ViT and the hybrid-ViT. The pure-ViT needs to split the original image, but the hybrid-ViT uses a convolutional network to extract the feature map from the original image, before splitting the feature map. Both the pure-ViT and the hybrid-ViT are extremely competitive state-of-the-art techniques used in the image classification field: Cheng et al. [27] applied them to gaze estimation, and achieved highly accurate results. Compared to the hybrid-ViT, the pure-ViT does not perform quite as well, because it needs to slice the facial image into several patches, which can destroy global information in the image, such as the head pose. To avoid the destruction of global information, we considered an independent but entire feature as a patch, independently. In particular, the face, the left eye image, and the right eye image were extracted, using a convolutional neural network, and flattened to a one-dimensional vector as one patch, independently. We implemented this operation in a static transformer module (STM): this module is described in detail in Section 3.1.
Another key problem in appearance-based deep learning gaze estimation methods is the efficient extraction of dynamic features. Unlike the static features extracted from an image, the dynamic features need to be extracted from an input video. Some research [22,28] has extended the basic static network, using a dynamic network—for example, using a long short-term memory (LSTM) approach [29]. However, these methods do not define dynamic features clearly—in other words, such methods obtain implicit dynamic features through RNNs. Liu et al. [30] proposed a differential network to predict differential information in solving personal-calibration problems in the gaze estimation field. The differential information was characterized as the difference between the gaze direction of two images, reflecting sight movements in continuous time. Inspired by this, we defined the dynamic feature to be sight movement, and we proposed a new RNN cell—the temporal differential module (TDM)—to obtain it in our work: this module can steadily extract effective dynamic features, using differential information. The TDM is described in detail in Section 3.2. The TDM is the core of the temporal differential network (TDN). We then combined the STM and the TDN into an end-to-end gaze estimation network—that is, a static transformer temporal differential network (STTDN)—to achieve better results.
To our knowledge, this is the first study to use self-attention mechanisms to fuse both coarse-grained facial features and fine-grained eye features. The contribution of our work can be summarized as follows:
(1)
We proposed a novel STTDN for gaze estimation, which could achieve better accuracy, compared to state-of-the-art algorithms;
(2)
We proposed the STM, to extract and fuse features. In the STM, we used a convolutional neural network to extract features from the face, left eye image, and right eye image. Then, we considered each feature as a patch, independently, to solve the key pure-ViT problem. Lastly, we used multi-head self-attention to fuse these patches.
(3)
The proposed TDM was used to obtain the dynamic information from video, and we clearly defined the dynamic feature to be sight movement.
The rest of the paper is organized as follows: we discuss the related work in Section 2; we describe our proposed STTDN method in Section 3; we summarize and analyze the experimental results in Section 4; finally, the study conclusions are discussed in Section 5.

2. Related Work

In recent years, appearance-based deep learning methods have been the mainstream methods used in the gaze estimation field. Compared to other traditional gaze estimation methods, appearance-based deep learning methods perform accurately when the head pose and illumination change. Zhang et al. [18] proposed the first deep learning model based on LeNet [31] for gaze estimation, which was used to predict the gaze direction from a grayscale image of the eye. Zhang et al. extended the convolutional neural network to 13 layers in their work [20], which achieved more accurate results, their appearance-based method using a VGG16-inherited network [32]. Zhang et al. also considered facial features as inputs in their work [21]. Other studies [32,33,34] further developed appearance-based deep learning gaze estimation methods. Fischer et al. [33] used two VGG16 networks simultaneously, to extract features from two images of the human eye, and connected the two-eye feature for regression. Cheng et al. [34] established a four-channel network for gaze estimation, in which two channels were used to extract features from the left eye and right eye images, the other two channels being used to extract facial features. Additionally, Chen et al. [19] applied a dilated convolution to extract high dimension features, the dilated convolution effectively increasing the receptive field while avoiding reduced image resolution. Krafka et al. [11] added face grid information to their model.
Although past appearance-based deep learning methods demonstrated high performance, they still needed to improve. For example, existing methods [33,34] simply concatenate the different obtained features (such as left eye image and right eye image) into one feature, without considering their internal relationship. The feature fusion stage is not fully considered in these existing methods. More recently, some research [23,24] has proposed the use of a feature fusion strategy. Bao et al. [23] applied the squeeze-and-excitation mechanism in the fusion of eye features, and used adaptive group normalization to correct fused eye features with facial features. Cheng et al. [24] proposed an attention module to fuse fine-grained eye features, with the guidance of coarse-grained facial features.
With the self-attention mechanism developing in CV, Cheng et al. [27] first applied the ViT in the gaze estimation field, and achieved an advanced level of performance. The self-attention mechanism [25] was proposed initially in the NLP field. Alexey Dosovitskiy et al. [26] proposed the ViT to integrate this method into the CV field. Liu et al. [35] further developed the ViT to the swin transformer (Swin), the sliding window technique being applied to reduce the computing resource overhead. At this stage, the Swin method remains at the forefront of several CV tasks—including image classification [36], semantic segmentation [37], and object detection [38], among others. In this paper, we further developed the self-attention mechanism application in the gaze estimation field, by specifically applying the self-attention mechanism to the feature fusion process.
Apart from the static features extracted from images, important features—namely, the dynamic features—can be obtained from video. Research [22,28,39,40] has proposed several gaze estimation methods that incorporate dynamic features. Kellnhofer et al. [22] proposed a video-based annotation tracking model, and used an LSTM method to obtain the dynamic features from a video. Zhou et al. [28] applied the Bi-LSTM to obtain dynamic features. Some research [22,28] has applied RNNs to implicitly obtaining dynamic features: at this stage, dynamic features are difficult to learn, because of the unstable extraction process of ambiguously defined dynamic features. More recently, some research [39,40] has defined dynamic features obtained clearly. Wang et al. [39] defined optical flow to be a dynamic feature, and then used the optical dynamic to reconstruct the three-dimensional face structure. Wang et al. [40] considered eye movement as a dynamic feature, and proposed a gaze tracking algorithm. Inspired by differential information, we also defined sight movement as a dynamic feature. Liu et al. [30] proposed a differential network to settle the personal calibration problem: in their work, the differential network predicted the differential information, which reflected eye movement. In this paper, we propose a new RNN cell—the TDM—to more efficiently obtain this dynamic feature.

3. Proposed Model and Algorithm

In this section, we describe in detail how we designed this end-to-end 3D gaze estimation network. The pipeline of the proposed STTDN is as shown in Figure 2. The STTDN consisted of two main components: (1) the STM, which was the static network; (2) the TDN, which was the dynamic network. The network used input video frames to predict the gaze direction corresponding to the last input frame. At this stage, the STM was responsible for extracting the static features from a single face image and the left eye and right eye images corresponding to it, while the TDN was responsible for extracting the dynamic features from the static features. The flow chart of the proposed model is shown in Figure 3. There were two basic states of our proposed model: back propagation and forward propagation. Forward propagation could be used to calculate the intermediate variables of each layer, and back propagation could be used to calculate the gradient of each layer.

3.1. The Design of the STM

The overall structure of the STM is as shown in Figure 2, where we integrated two convolutional neural networks and a multi-head self-attention fusion block. The STM module used a face image and its corresponding left eye and right eye images, before outputting static features.
Given a face image ( I f a c e ( i ) R H × W × C ) and its corresponding left eye ( I l e f t _ e y e ( i ) R H × W ) and right eye ( I r i g h t _ e y e ( i ) R H × W ) images—where (H, W) denotes the resolution of the original image, C denotes the number of channels, and i denotes the ith image in the input video frames—we used two independent convolutional neural networks for feature extraction. We differentiated left eye and right eye images in the feature extraction stage, because the left eye and right eye images contributed differently to the gaze direction, due to the headpose and illumination. The first convolutional neural network was used to extract the facial features; the second convolutional neural network extracted the left eye and right eye features simultaneously, before flattening them to one-dimensional vectors—that is, the f f a c e ( i ) R 1 × d , f l e f t _ e y e ( i ) R 1 × d , and f r i g h t _ e y e ( i ) R 1 × d , where d was the feature dimension and d = 32 in our experiments.
Similar to the position encoding and patch embedding process in [26], we created the feature matrix X : firstly, we created a learnable f t o k e n ( i ) R 1 × d ; secondly, we coded the feature position—specifically, f t o k e n ( i ) was coded as position 0, and f f a c e ( i ) , f l e f t _ e y e ( i ) , f r i g h t _ e y e ( i ) were coded as positions 1, 2, and 3, respectively; finally, we concatenated the four one-dimension vectors into a feature matrix X = f f a c e ( i ) ; f f a c e ( i ) ; f l e f t _ e y e ( i ) ; f r i g h t _ e y e ( i ) .
The feature matrix X could be further fused by using the multi-head self-attention fusion block. The core of the multi-head self-attention fusion block was the multi-head self-attention mechanism  [25], which was a derivative of the self-attention mechanism. The self-attention mechanism used a multi-layer perception (MLP) to map the feature matrix X to (Queries) Q R n × d k , ( Keys ) K R n × d k , and ( Values ) V R n × d v , where n was the batch size, d k was the dimension of Queries and Keys, and d v was the dimension of Values. For our experiment, d k = d v = 8 . The formulaic definition of self-attention can be expressed as follows:
A t t e n t i o n ( Q , K , V ) = S o f t M a x Q K T d k V
Unlike the self-attention mechanism, the multi-head self-attention mechanism used projection matrices W i Q R d × d k , W i K R d × d k , W i V R d × d v , projecting feature X R n × d into different representation subspaces. Moreover, a fusion matrix W O R h d v × d fused the information extracted from the different representation subspaces, where i denoted the i t h representation subspace, and h denoted the number of representation subspaces. In our experiment, we employed h = 4 . The definition of the multi-head self-attention mechanism calculation can be expressed as follows:
M u l t i H e a d ( Q , K , V ) = C o n c a t h e a d 1 , , h e a d h W O
where h e a d i = A t t e n t i o n ( Q W i Q , K W i K , V W i V ) , and C o n c a t represents concat operation.
We could stack multiple transformer layers [25], to implement the multi-head self-attention fusion block, the structure of which is shown in Figure 4. In Figure 4, L represents the number of transformer layers. An independent transformer layer comprises the MLP and a multi-head attention layer (MSA). The definition of the transformer calculation can be expressed as follows:
X = X + MSA ( LN ( X ) )
Y = X + MLP LN X
where X denotes the feature matrix received by each layer of the transformer, X denotes the intermediate variable, which is the feature output using a single layer transformer, and Y denotes the feature matrix output by each layer of the transformer.
Finally, g t o k e n ( i ) is the independent output from the STM, which represents the static feature of each frame.

3.2. Design of the TDN

In this section, we will take a closer look at the TDN in the STTDN. The structure of the TDN was as shown in Figure 5. In the TDN, we used five RNN cells (TDM) to obtain the dynamic features from five input video frames, and applied a fully connected layer to predict the gaze direction of the fifth input video frame.
The TDM was the core component in the TDN, its structure being as shown in Figure 6. We defined the dynamic feature as clear sight movement in the TDM. We introduced the differential information to define the dynamic feature clearly. Liu et al. [30] initially proposed the differential information concept, which represents the difference between the gaze direction of two images of the eye. We can generalize the definition of the dynamic feature as a sight movement using differential information, specifically information from the eye image to the face image, and from only these two images to the video frames. A generalized definition of the dynamic feature can be expressed as follows:
d i = g a z e i g a z e i 1
where d i denotes the sight movement from the ith frame to the ( i 1 ) th frame, g a z e i denotes the gaze direction of the ith frame, and g a z e i 1 denotes the gaze direction of the ( i 1 ) th frame.
We could apply d i in the TDM to extract the dynamic features from the video frames. Compared to the LSTM cell, we kept only two gates in the TDM: the forgotten gate ( f i ) and the output gate ( o i ) , the coefficient of both gates being determined by d i . The proposed algorithm of the TDM is shown in Algorithm 1:
Algorithm 1: TDM
Input: Feature vector of 5 frames g a z e t o k e n = g t o k e n ( 1 ) , g t o k e n ( 2 ) , g t o k e n ( 3 ) , g t o k e n ( 4 ) , g t o k e n ( 5 ) ,
Output: The state of hidden layer h 5
1:
procedureTDM( g a z e t o k e n )
2:
    Initialize two zero vectors h 0 R 1 × d , c 0 R 1 × d
3:
    repeat
4:
        for  i = 1 to 5  do
5:
           Calculate d i based on c i 1 and g t o k e n ( i ) using  (6)
6:
           Calculate f i and o i base on d i using  (7) and  (8)
7:
           Update c i based on c i 1 and g t o k e n ( i ) using  (9)
8:
           Update h i based on h i 1 and g t o k e n ( i ) using  (10)
9:
        end for
10:
    until eigenvector calculation completed
11:
    return  h 5
12:
end procedure
The definitions can be expressed as follows:
d i = tanh w 1 c i 1 + w 1 g t o k e n ( i ) + b 1
f i = σ w 2 d i + b 2
o i = σ w 3 d i + b 3
c i = c i 1 f i + tanh w 4 g t o k e n ( i ) + b 4 1 f i
h i = h i 1 + d i o i + tanh w 5 g t o k e n ( i ) + b 5 1 o i
where, w 1 , w 2 , w 3 , w 4 , and w 5 represent weight matrix, and b 1 , b 2 , b 3 , b 4 , and b 5 represent bias matrix.
Finally, h 5 was output by the last TDM in the TDN. A fully connected layer was implemented after the last TDM. The fully connected layer used h 5 to predict the gaze direction.

3.3. Topology of the STTDN Network

In this section, we supplied additional topology details. We integrated the convolutional neural network and the multi-head self-attention fusion block in the STM. We implemented ResNet18 [41] as the convolutional neural network to extract features, the number of feature maps being 32—two ResNet18 being used to extract features from eye and face images, respectively. An additional average pooling-down sampling layer followed multiple convolutional layers, to ensure consistent feature dimension. We then stacked six transformer encoders as the multi-head self-attention fusion block. The number of heads of each transformer encoder was 4, the input dimension of each transformer encoder was 32, and the inner dimension of each transformer encoder was 128. The TDM had fewer parameters—the hidden layer dimension was 32, the num layer was 1.

4. Experimental Results

In this section, we discuss the experimental performance of the proposed STTDN on two public datasets, MPIIFaceGaze [20] and Eyediap [42], and the effectiveness of the STM and the TDM. The remaining sections are as follows: first, we introduce the two public datasets used in this study (Section 4.1), and the evaluation metric (Section 4.2); then, we compare our proposed method with state-of-the-art methods, in Section 4.3; in Section 4.5 and Section 4.6, we evaluate the effectiveness of the two main ideas in the STTDN—that is, (1) the way to fuse features by STM, and (2) the way to extract dynamic features by DTN; in Section 4.7, we discuss an ablation study conducted for the STTDN.
To allow readers to reproduce our proposed architecture and conduct further research, we offer several important parameters related to this study. We implemented our model using Pytorch, and evaluated it on two TITAN RTX platforms. The optimizer used was the AdamW optimizer [43], the loss function was L1 loss, the epoch was set to 30, the initial learning rate was set to 5 × 10 4 , and the number of video frame inputs was set to 5.

4.1. Datasets

To better evaluate the performance of the STTDN, we used two public datasets: (i) MPIIFaceGaze [20] and (ii) Eyediap [42]. Figure 7 shows examples of face images on the two public datasets, and their corresponding left eye and right eye images.
The MPIIFaceGaze dataset [20] was proposed by Zhang et al., and is the most popular dataset used for appearance-based gaze estimation methods. The MPIIFaceGaze dataset contains a total of 213,659 images collected from 15 subjects during several months of daily life without head pose constraints. Because the images come from the real-world environment, the dataset has abundant illumination and head pose scenes. We considered the two face images with the shortest time interval as two adjacent frames.
The Eyediap [42] dataset contains 94 video clips from 16 participants in experimental scenes: it contains three visual target segments—that is, continuous moving targets, screen targets, and floating balls. Each participant was recorded with six static-head postures and six free-head postures. As the data were collected in a laboratory environment, the images lacked illumination variation. As 2 subjects lacked screen-target video, we obtained images from 14 subjects for our study.
Based on the dataset pre-processing process in [32], we cropped out RGB face images with resolutions of 224 × 224 pixels, as well as grayscale left eye and right eye images with resolutions of 36 × 60 pixels.

4.2. Evaluation Metric

We used the leave-one-person-out criterion as the experimental evaluation metric—a common choice in gaze estimation studies. We used 14 subjects as the training dataset, and 1 subject as the validation dataset, before selecting 15 objects as the validation dataset, in turn, and using the average error precision of the 15 experiments as the model performance. We used the angular error as the evaluation metric. The greater the angular error, the lower the accuracy of the model. The definition of the angular error can be expressed as follows:
L a n g u l a r = arccos ( g · g ^ g g ^ )
where g R 3 denotes the actual gaze direction, and g ^ R 3 denotes the estimated gaze direction.

4.3. Comparison with State-of-the-Art Methods

To evaluate the performance of the model, we selected several networks which had exhibited advanced performance, for comparison: Hybrid-ViT [27]; Gaze360 [22]; iTracker [11]; DilatedNet [19]; RTGene [33]; AFFNet [23]; and CANet [24]. We recorded the performance of the STTDN and the comparison methods, as shown in Table 1—a more intuitive representation being reflected in Figure 8.
The STTDN achieved an angular error of 3.73 on the MPIIFaceGaze dataset [20], which was highly competitive with the previous best method—that is, AFFNet [23], which also achieved an accuracy error of 3.73 on MPIIFaceGaze. The STTDN achieved an angular error of 5.02 on Eyediap [42], an improvement of 2.9% on the previous best method—that is, Hybrid-ViT [27], which achieved an angular error of 5.17 on the Eyediap dataset. It is evident that our model outperformed the other models, as shown in Figure 8.
We also analyzed the performance errors from two different perspective: (1) the angular error of the STTDN on different experiment participants; (2) the angular error distribution of the STTDN on different datasets.

4.4. Perfromance Analysis

4.4.1. The Angular Error of the STTDN on Different Experiment Participants

First, we analyzed the error of the STTDN on different experiment participants. We recorded the angular error of each experiment participant, as shown in Figure 9. As shown in Figure 9a, the STTDN performed best on person ID p0 (2.4 ), and worst on person ID p14 (4.74 ), the difference being 2.34 (4.74 –2.4 ). In Figure 9b, the STTDN performed best on person ID p14 (3.33 ), and worst on person ID p7 (7.57 ), the difference being 4.24 (7.57 –3.33 ).
There was still a large difference among different experiment participants, which prevented our model from performing better. This problem in the personal estimation field is called the personal calibration problem. The calibration problem can be considered as a domain adaption problem, where the training set is the source domain and the test set is the target domain. The proposed method did not use a calibration sample in the target domain: thus, this proposed method did not solve the personal calibration problem. For example, the facial contrast of person ID p8 in Figure 9b was quite different from the others, leading to a higher angular error. Moreover, the personal calibration problem exists in other gaze estimation methods, too. Compared to the Eyediap dataset [42], the difference was smaller using the MPIIFaceGaze dataset [20] (4.24 > 2.34 ). We analyzed the main reason, as follows. The MPIIFaceGaze had a larger data scale, and more trainable experiment participants. Additionally, the MPIIFaceGaze had richer illumination conditions. Improving the dataset could effectively alleviate the personal calibration problem, to a certain extent.

4.4.2. The Angular Error Distribution of the STTDN on Different Angles

We analyzed the angular error distribution of the STTDN at different gaze directions. We recorded the distribution of the gaze direction for the two datasets, as shown in Figure 10. Figure 11 shows the recorded distribution of the angular error of the STTDN in different gaze directions. The proposed method performed poorly at some extreme gaze directions, because the training dataset lacked data samples at extreme angles. Conversely, the more concentrated the gaze direction distribution, the better the proposed model’s performance. A limited number of data samples at extreme gaze directions resulted in the STTDN performing poorly at extreme gaze directions.

4.5. The Effectiveness of the STM

We applied the MSA method to the STM, to fuse the coarse-grained facial features and the fine-grained eye features. We evaluated the effectiveness of the fusion feature in this subsection. Specifically, we evaluated the effectiveness of: (1) adding fine-grained eye features; (2) fusing two differently-grained features, using the MSA method.
To evaluate the effectiveness of the fusion feature, we implemented one fully connected layer after the STM into an end-to-end gaze estimation network, called the transformer static network (STN). We evaluated the angular error of the STN, using two public datasets—MPIIFaceGaze [20] and Eyediap [42]—and recorded the results, as shown in Table 2. We also set up two comparison models—the STN-W/O self-attention and the STN-W/O eye patches—the angular errors of which were also recorded in Table 2. The specific implementation detail of the STN-W/O self-attention and STN-W/O eye patches was as follows:
(1)
STN-W/O self-attention: we removed the multi-head self-attention fusion block from the STN, to obtain the STN-W/O self-attention. At this stage, the fine-grained eye and coarse-grained facial features were connected and directly input to the fully connected layer, before the fully connected layer predicted the gaze direction.
(2)
STN-W/O eye patches: the model was approximately the same as the hybrid-ViT network architecture [27]. The hybrid-ViT [27] used only face image as its input, with its main structure also comprising two parts: Resnet18 and the transformer layers. The hybrid-ViT used Resnet18 to extract the feature map from the face image, before splitting the feature map into patches: finally, it input these patches into the transformer layer. The hyperparameters used in this model should remain consistent.
Table 2 shows that STN achieved the best performance on both datasets—that is, an angular error of 5.07 using the Eyediap dataset, and 3.75 using the MPIIFaceGaze dataset. After removing the multi-head self-attention fusion block from the STN, the angular error of the STN-W/O self-attention on the MPIIFaceGaze increased by 0.24 , and on the Eyediap dataset by 0.02 , demonstrating the effectiveness of the multi-head self-attention fusion block. Compared to the STN, the STN-W/O eye patches also exhibited different degrees of degradation on the two public datasets: specifically, the angular error of the STN-W/O eye patches increased by 0.25 on the MPIIFaceGaze dataset, and by 0.17 on the Eyediap dataset, demonstrating the effectiveness of adding fine-grained eye features.

4.6. The Effectiveness of the TDN

In this section, we explore the effectiveness of using dynamic features. We set up two comparison methods—that is STM–LSTM and STM–BiLSTM. Specifically, we replaced the TDM with LSTM and BiLSTM to get STM–LSTM and STM–BiLSTM, respectively. The number of input video frames was set up to five. We recorded the angular error of these models, as shown in Table 3. STTDN performed best among these models. Compared to STM–LSTM, STTDN had improved by 0.7% and 2.7% on the MpiiFaceGaze dataset [20] and the Eyediap dataset [42]. Compared to STM–BiLSTM, STTDN had improved by 4.6% and 4.1% on the MpiiFaceGaze dataset [20] and the Eyediap dataset [42]. This proves that the ability to extract dynamic features by TDN is better than the other common RNN.
We also discovered another critical issue: STM–LSTM and STM–BiLSTM showed degradation on both the MpiiFaceGaze dataset [20] and the Eyediap dataset [42], compared to TSN, which did not add dynamic features. Compared to TSN, STTDN/TSM–LSTM/TSM–BiLSTM all used RNN to extract dynamic features and predict gaze direction based on those dynamic features. In other words, STTDN, TSM–LSTM, and TSM–BiLSTM added the dynamic features in their model. Specifically, STN achieved the angle error of 3.75 on the MpiiFaceGaze dataset in Table 2, while STM–BiLSTM and STM–LSTM achieved 3.76 and 3.91 , respectively, on the MpiiFaceGaze dataset; STN achieved the angle error of 5.07 on the Eyediap dataset in Table 2, while STM–BiLSTM and STM–LSTM reached 5.16 , respectively, and 5.24 on the Eyediap dataset: however, this does not prove that adding dynamic features is ineffective, because the proposed STTDN performed better than STN on two datasets. A reliable explanation for this degradation is that RNN was difficult to be trained when RNN was used to extract ambiguous dynamic features. This phenomenon was more obvious in non light-weight models. Thus, the proposed TDN can extract better dynamic features and predict more accurate gaze direction.

4.7. Ablation Study

We conducted ablation experiments to evaluate the effectiveness of the main modules—(1) the STM and (2) the TDN—in the STTDN. For this purpose, we set up two variant models—namely, the STTDN-W/O STM and STTDN-W/O TDN:
(1)
The STTDN-W/O STM: unlike the STTDN, we replaced the Resnet18 structure in the STM with four 3 × 3 convolutional layers and a global average down-sampling layer, and replaced the multi-head self-attention fusion block with a fully connected layer.
(2)
The STTDN-W/O TDN: unlike the STTDN, we removed the TDN from the STTDN. With the removal of the TDN, an external fully connected layer was implemented after the STM, to predict the gaze direction.
We recorded the angular error of these variants and the STN on two public datasets, as shown in Table 4. The angular error of the STTDN reached 3.73 on the MPIIFaceGaze dataset [20], and 5.02 on the Eyediap dataset [42], respectively, when the number of input frames was set to five. After removing the STM, the angular error of the STTDN-W/O STM reached 4.67 on the MPIIFaceGaze dataset [20], and 5.85 on the Eyediap dataset [42], respectively. The STTDN-W/O STM and STTDN-W/O TDM methods exhibit more reduction than the STTDN, demonstrating the effectiveness of the STM and TDN in the STTDN.

4.8. Computational Complexity

Our proposed model included convolution structure, transformer structure, and RNN structure. In order to compute the whole model complexity, we defined the complexity calculation formulas of three main structures. The definition of the complexity computing of convolution structure can be expressed as follows:
T i m e O ( l = 1 D M l 2 × K l 2 × C l 1 × C l )
where D is network depth, l is the lth convolution layer, M l is the feature map side output by the lth convolution layer, M l is the kernel side of the lth convolution layer, C l is the channel output by the lth convolution layer, and C l 1 is the channel output by the l 1 th convolution layer.
The definition of the complexity computing of RNN structure can be expressed as follows:
T i m e O ( l = 1 D n l × d l 2 )
where D is network depth, l is the lth RNN layer, n is the RNN cells number of the lth RNN layers, and d is the input feature dimension of the lth RNN layers.
The definition of the complexity computing of transformer structure can be expressed as follows:
T i m e O ( l = 1 D n l 2 × d l )
where D is network depth, l is the lth transformer layer, n is the RNN cells number of the lth RNN layers, and d is the input feature dimension of the lth RNN layers.
We could compute the time complexity of the convolution part in our proposed model using (12), the multi-head self-attention fusion module using (14), and TDN using (13). The overall time complexity of our proposed model was O ( N 2 ) , where N was the input frame number. Compared to the convolution network (e.g., STN-W/O Self-Attention), the computational complexity of the proposed model only additionally increased when (1) using the multi-head self-attention mechanism to fuse the different-grain feature, and (2) extracting dynamic features. Fortunately, its complexity was not N times when using N frames as input. We will save the next N-1 frames fusion features during operation. The previous N-1 frames fusion features did not need to be recomputed in the next operation.

5. Conclusions

This study proposed a novel gaze estimation network, the STTDN, which integrated the STM and the TDM. We provided a multi-head self-attention fusion strategy STM for fusing fine-grained eye features and coarse-grained facial features. Additionally, we defined a new dynamic feature (sight movement) and proposed an innovative RNN cell-TDM to obtain it. Through experimental evaluation, the STTDN demonstrated great competitiveness, compared to the state-of-the-art methods, on two publicly available datasets: MPIIFaceGaze and Eyediap. In future work, we will apply contrastive learning in gaze estimation, to solve the personal calibration challenge, and to increase performance in extreme angle environments. In addition, we will apply the proposed STTDN to cognitive workload estimation, which presents the occupancy rate of human mental resources under working conditions.

Author Contributions

Y.L. was responsible for methodology, original draft preparation, review, and editing; L.H. was responsible for methodology, validation, review, and editing; J.C. was responsible for validation, review, and editing; X.W. was responsible for validation, review, and editing; B.T. was responsible for methodology, review, and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Guangxi Science and Technology Major Project (AA22068057), the Guangxi Natural Science Foundation (2022GXNSFBA035644, 2021GXNSFBA220039), and the National Natural Science Foundation of China (61903090).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, P.; Hou, X.; Duan, X.; Yip, H.; Song, G.; Liu, Y. Appearance-based gaze estimator for natural interaction control of surgical robots. IEEE Access 2019, 7, 25095–25110. [Google Scholar] [CrossRef]
  2. Mohammad, Y.; Nishida, T. Controlling gaze with an embodied interactive control architecture. Appl. Intell. 2010, 32, 148–163. [Google Scholar] [CrossRef]
  3. Vanneste, P.; Oramas, J.; Verelst, T.; Tuytelaars, T.; Raes, A.; Depaepe, F.; Van den Noortgate, W. Computer vision and human behaviour, emotion and cognition detection: A use case on student engagement. Mathematics 2021, 9, 287. [Google Scholar] [CrossRef]
  4. Fridman, L.; Reimer, B.; Mehler, B.; Freeman, W.T. Cognitive load estimation in the wild. In Proceedings of the 2018 Chi Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–9. [Google Scholar]
  5. Ma, H.; Pei, W.; Zhang, Q. Research on Path Planning Algorithm for Driverless Vehicles. Mathematics 2022, 10, 2555. [Google Scholar] [CrossRef]
  6. Patney, A.; Kim, J.; Salvi, M.; Kaplanyan, A.; Wyman, C.; Benty, N.; Lefohn, A.; Luebke, D. Perceptually-based foveated virtual reality. In Proceedings of the ACM SIGGRAPH 2016 Emerging Technologies, Anaheim, CA, USA, 24–28 July 2016; pp. 1–2. [Google Scholar]
  7. Moral-Sánchez, S.N.; Sánchez-Compaña, M.T.; Romero, I. Geometry with a STEM and Gamification Approach: A Didactic Experience in Secondary Education. Mathematics 2022, 10, 3252. [Google Scholar] [CrossRef]
  8. Funes-Mora, K.A.; Odobez, J.M. Gaze estimation in the 3d space using rgb-d sensors. Int. J. Comput. Vis. 2016, 118, 194–216. [Google Scholar] [CrossRef]
  9. Huang, L.; Li, Y.; Wang, X.; Wang, H.; Bouridane, A.; Chaddad, A. Gaze Estimation Approach Using Deep Differential Residual Network. Sensors 2022, 22, 5462. [Google Scholar] [CrossRef]
  10. Li, Y.; Tan, B.; Akaho, S.; Asoh, H.; Ding, S. Gaze prediction for first-person videos based on inverse non-negative sparse coding with determinant sparse measure. J. Vis. Commun. Image Represent. 2021, 81, 103367. [Google Scholar] [CrossRef]
  11. Krafka, K.; Khosla, A.; Kellnhofer, P.; Kannan, H.; Bhandarkar, S.; Matusik, W.; Torralba, A. Eye tracking for everyone. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2176–2184. [Google Scholar]
  12. Recasens, A.; Khosla, A.; Vondrick, C.; Torralba, A. Where are they looking? Adv. Neural Inf. Process. Syst. 2015, 28, 199–207. [Google Scholar] [CrossRef]
  13. Xu, B.; Li, W.; Liu, D.; Zhang, K.; Miao, M.; Xu, G.; Song, A. Continuous Hybrid BCI Control for Robotic Arm Using Noninvasive Electroencephalogram, Computer Vision, and Eye Tracking. Mathematics 2022, 10, 618. [Google Scholar] [CrossRef]
  14. Guestrin, E.D.; Eizenman, M. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 2006, 53, 1124–1133. [Google Scholar] [CrossRef] [PubMed]
  15. Zhu, Z.; Ji, Q. Novel eye gaze tracking techniques under natural head movement. IEEE Trans. Biomed. Eng. 2007, 54, 2246–2260. [Google Scholar] [PubMed]
  16. Valenti, R.; Sebe, N.; Gevers, T. Combining head pose and eye location information for gaze estimation. IEEE Trans. Image Process. 2011, 21, 802–815. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Alberto Funes Mora, K.; Odobez, J.M. Geometric generative gaze estimation (g3e) for remote rgb-d cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1773–1780. [Google Scholar]
  18. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. Appearance-based gaze estimation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4511–4520. [Google Scholar]
  19. Chen, Z.; Shi, B.E. Appearance-based gaze estimation using dilated-convolutions. In Proceedings of the Asian Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2018; pp. 309–324. [Google Scholar]
  20. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. Mpiigaze: Real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 162–175. [Google Scholar] [CrossRef] [Green Version]
  21. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. It’s written all over your face: Full-face appearance-based gaze estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 51–60. [Google Scholar]
  22. Kellnhofer, P.; Recasens, A.; Stent, S.; Matusik, W.; Torralba, A. Gaze360: Physically unconstrained gaze estimation in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 6912–6921. [Google Scholar]
  23. Bao, Y.; Cheng, Y.; Liu, Y.; Lu, F. Adaptive feature fusion network for gaze tracking in mobile tablets. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 9936–9943. [Google Scholar]
  24. Cheng, Y.; Huang, S.; Wang, F.; Qian, C.; Lu, F. A coarse-to-fine adaptive network for appearance-based gaze estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10623–10630. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  26. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  27. Cheng, Y.; Lu, F. Gaze estimation using transformer. arXiv 2021, arXiv:2105.14424. [Google Scholar]
  28. Zhou, X.; Lin, J.; Jiang, J.; Chen, S. Learning a 3D gaze estimator with improved Itracker combined with bidirectional LSTM. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 850–855. [Google Scholar]
  29. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  30. Liu, G.; Yu, Y.; Mora, K.A.F.; Odobez, J.M. A differential approach for gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1092–1099. [Google Scholar] [CrossRef] [Green Version]
  31. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Fischer, T.; Chang, H.J.; Demiris, Y. Rt-gene: Real-time eye gaze estimation in natural environments. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 334–352. [Google Scholar]
  34. Cheng, Y.; Lu, F.; Zhang, X. Appearance-based gaze estimation via evaluation-guided asymmetric regression. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 100–115. [Google Scholar]
  35. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  36. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  37. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  38. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  39. Wang, Z.; Chai, J.; Xia, S. Realtime and accurate 3D eye gaze capture with DCNN-based iris and pupil segmentation. IEEE Trans. Vis. Comput. Graph. 2019, 27, 190–203. [Google Scholar] [CrossRef] [PubMed]
  40. Wang, K.; Su, H.; Ji, Q. Neuro-inspired eye tracking with eye movement dynamics. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 9831–9840. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Funes Mora, K.A.; Monay, F.; Odobez, J.M. Eyediap: A database for the development and evaluation of gaze estimation algorithms from rgb and rgb-d cameras. In Proceedings of the Symposium on Eye Tracking Research and Applications, San Antonio, TX, USA, 22–24 March 2004; pp. 255–258. [Google Scholar]
  43. Loshchilov, I.; Hutter, F. Fixing weight decay regularization in adam. arXiv 2018, arXiv:1711.05101. [Google Scholar]
Figure 1. Two categories of gaze estimation methods: (1) appearance-based gaze estimation method; (2) model-based gaze estimation method.
Figure 1. Two categories of gaze estimation methods: (1) appearance-based gaze estimation method; (2) model-based gaze estimation method.
Mathematics 11 00686 g001
Figure 2. The pipeline of the proposed static transformer temporal differential network.
Figure 2. The pipeline of the proposed static transformer temporal differential network.
Mathematics 11 00686 g002
Figure 3. The flow chart of the proposed static transformer temporal differential network.
Figure 3. The flow chart of the proposed static transformer temporal differential network.
Mathematics 11 00686 g003
Figure 4. The structure of the multi-head self-attention fusion block.
Figure 4. The structure of the multi-head self-attention fusion block.
Mathematics 11 00686 g004
Figure 5. The structure of the temporal dynamic network.
Figure 5. The structure of the temporal dynamic network.
Mathematics 11 00686 g005
Figure 6. The structure of the TDM.
Figure 6. The structure of the TDM.
Mathematics 11 00686 g006
Figure 7. Example face images and their corresponding left eye and right eye images on two public datasets.
Figure 7. Example face images and their corresponding left eye and right eye images on two public datasets.
Mathematics 11 00686 g007
Figure 8. Angular error of the STTDN, and comparison methods on the MPIIFaceGaze [20] and Eyediap [42] datasets. We recorded the angular error of the MPIIFaceGaze on the horizontal axis, and the angular error of Eyediap on the vertical axis. The performance was better when the position was closer to the bottom left corner.
Figure 8. Angular error of the STTDN, and comparison methods on the MPIIFaceGaze [20] and Eyediap [42] datasets. We recorded the angular error of the MPIIFaceGaze on the horizontal axis, and the angular error of Eyediap on the vertical axis. The performance was better when the position was closer to the bottom left corner.
Mathematics 11 00686 g008
Figure 9. The angular error of the STTDN among the different experiment participants: (a) records the angular error of the different experiment participants (p0–p14) on the MPIIFaceGaze dataset [20]; (b) records the angular error of the different experiment participants (p1–p16) on the Eyediap dataset [42]. The angular error is recorded on the vertical axis. Each experiment participant ID and its corresponding face image is recorded on the horizontal axis.
Figure 9. The angular error of the STTDN among the different experiment participants: (a) records the angular error of the different experiment participants (p0–p14) on the MPIIFaceGaze dataset [20]; (b) records the angular error of the different experiment participants (p1–p16) on the Eyediap dataset [42]. The angular error is recorded on the vertical axis. Each experiment participant ID and its corresponding face image is recorded on the horizontal axis.
Mathematics 11 00686 g009
Figure 10. The data distribution of gaze direction on the two datasets: (a) MPIIFaceGaze [20]; (b) Eyediap [42]. The yaw of the gaze direction is recorded on the horizontal axis, and the pitch of the gaze direction on the vertical axis. The brighter the color of the heatmap, the more concentrated the gaze direction.
Figure 10. The data distribution of gaze direction on the two datasets: (a) MPIIFaceGaze [20]; (b) Eyediap [42]. The yaw of the gaze direction is recorded on the horizontal axis, and the pitch of the gaze direction on the vertical axis. The brighter the color of the heatmap, the more concentrated the gaze direction.
Mathematics 11 00686 g010
Figure 11. The angular error distribution of the STTDN on two datasets: (a) MPIIFaceGaze [20]; (b) Eyediap [42]. The yaw of the gaze direction is recorded on the horizontal axis, and the pitch of the gaze direction on the vertical axis. The brighter the color of the heatmap, the bigger the angular error.
Figure 11. The angular error distribution of the STTDN on two datasets: (a) MPIIFaceGaze [20]; (b) Eyediap [42]. The yaw of the gaze direction is recorded on the horizontal axis, and the pitch of the gaze direction on the vertical axis. The brighter the color of the heatmap, the bigger the angular error.
Mathematics 11 00686 g011
Table 1. Angular error of STTDN and comparison methods on the MpiiFaceGaze [20] and Eyediap [42] datasets.
Table 1. Angular error of STTDN and comparison methods on the MpiiFaceGaze [20] and Eyediap [42] datasets.
DatasetMpiiFaceGaze [20]Eyediap [42]
Method
Hybrid-VIT [16]4.00 5.17
Gaze360 [22]4.06 5.36
iTracker [11]7.33 7.13
DilatedNet [19]4.42 6.19
RTGene [33]4.66 6.02
AFFNet [23]3.73 6.41
CANet [24]4.27 5.27
STTDN (ours)3.73 5.02
Table 2. The angular error of STN and its variants, STN-W/O Self Attention and STN-W/O Eye Patches, on MpiiFaceGaze [20] and Eyediap [42].
Table 2. The angular error of STN and its variants, STN-W/O Self Attention and STN-W/O Eye Patches, on MpiiFaceGaze [20] and Eyediap [42].
DatasetMpiiFaceGaze [20]Eyediap [42]
Method
STN-W/O Self-Attention3.99 5.09
STN-W/O Eye Patches4.00 5.24
STN3.75 5.07
Table 3. The angular error of STM–LSTM, STM–BiLSTM, STTDN.
Table 3. The angular error of STM–LSTM, STM–BiLSTM, STTDN.
DatasetMpiiFaceGaze [20]Eyediap [42]
Method
STM–BiLSTM3.76 5.16
STM–LSTM3.91 5.24
STTDN (ours)3.73 5.02
Table 4. Angular errors of STTDN, STTDN-W/O STM, STTDN-W/O TDM on two public datasets: MpiiFaceGaze [20] and Eyediap [42].
Table 4. Angular errors of STTDN, STTDN-W/O STM, STTDN-W/O TDM on two public datasets: MpiiFaceGaze [20] and Eyediap [42].
MethodSTMTDNMpiiFace Gaze [20]Eyediap [42]
STTDN3.73 5.02
STTDN-W/O STM×4.67 5.85
STTDN-W/O TDN×3.75 5.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Huang, L.; Chen, J.; Wang, X.; Tan, B. Appearance-Based Gaze Estimation Method Using Static Transformer Temporal Differential Network. Mathematics 2023, 11, 686. https://doi.org/10.3390/math11030686

AMA Style

Li Y, Huang L, Chen J, Wang X, Tan B. Appearance-Based Gaze Estimation Method Using Static Transformer Temporal Differential Network. Mathematics. 2023; 11(3):686. https://doi.org/10.3390/math11030686

Chicago/Turabian Style

Li, Yujie, Longzhao Huang, Jiahui Chen, Xiwen Wang, and Benying Tan. 2023. "Appearance-Based Gaze Estimation Method Using Static Transformer Temporal Differential Network" Mathematics 11, no. 3: 686. https://doi.org/10.3390/math11030686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop