Next Article in Journal
EAIA: An Efficient and Anonymous Identity-Authentication Scheme in 5G-V2V
Previous Article in Journal
One-Channel Wearable Mental Stress State Monitoring System
Previous Article in Special Issue
Analysis of the Generalization Ability of Defogging Algorithms on RICE Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Decoupled Cross-Modal Transformer for Referring Video Object Segmentation

1
School of Information and Cyber Security, People’s Public Security University of China, Beijing 100038, China
2
Key Laboratory of Security Prevention Technology and Risk Assessment of Ministry of Public Security, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5375; https://doi.org/10.3390/s24165375
Submission received: 31 May 2024 / Revised: 13 August 2024 / Accepted: 19 August 2024 / Published: 20 August 2024
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
Referring video object segmentation (R-VOS) is a fundamental vision-language task which aims to segment the target referred by language expression in all video frames. Existing query-based R-VOS methods have conducted in-depth exploration of the interaction and alignment between visual and linguistic features but fail to transfer the information of the two modalities to the query vector with balanced intensities. Furthermore, most of the traditional approaches suffer from severe information loss in the process of multi-scale feature fusion, resulting in inaccurate segmentation. In this paper, we propose DCT, an end-to-end decoupled cross-modal transformer for referring video object segmentation, to better utilize multi-modal and multi-scale information. Specifically, we first design a Language-Guided Visual Enhancement Module (LGVE) to transmit discriminative linguistic information to visual features of all levels, performing an initial filtering of irrelevant background regions. Then, we propose a decoupled transformer decoder, using a set of object queries to gather entity-related information from both visual and linguistic features independently, mitigating the attention bias caused by feature size differences. Finally, the Cross-layer Feature Pyramid Network (CFPN) is introduced to preserve more visual details by establishing direct cross-layer communication. Extensive experiments have been carried out on A2D-Sentences, JHMDB-Sentences and Ref-Youtube-VOS. The results show that DCT achieves competitive segmentation accuracy compared with the state-of-the-art methods.

1. Introduction

Referring video object segmentation (R-VOS) is an emerging subtask in the field of video segmentation. Its objective is to segment the regions of interest within video frames based on provided natural language expressions. In contrast to traditional semi-supervised video object segmentation methods [1,2,3], R-VOS only requires a simple textual description of the target object instead of manually annotating it in the first video frame (or a few frames). Consequently, R-VOS avoids intricate and costly annotation procedures, enhancing its user-friendliness. The academic community has recently exhibited great interest in this task due to its promising potential in applications such as intelligent surveillance video processing, video editing, and human–computer interaction.
Referring video segmentation addresses two crucial issues: how to make the model understand which object is the referred one, and how to accurately segment the target from the background.
The key to addressing the former issue lies in accomplishing fine-grained cross-modal information interaction and aggregation. Most of the early methods involve direct fusion of the visual and linguistic features. Common strategies include concatenation–convolution [4,5], recurrent LSTM [6], dynamic filter [7,8] and attention mechanism [9,10,11,12]. In contrast, some recent approaches [13,14,15,16] introduce additional variables as a bridge for aggregation. They employ a set of object queries to collect and store entity-related information from multi-modal features. Subsequently, these queries are linked across frames to achieve the tracking effect. Methods of this type can better model the cross-modal dependencies, but often overlook the scale disparity between visual and linguistic features. Since visual features are often tens of times longer than linguistic features, the object queries will be overly biased to focus on visual content.
The latter issue, which is to precisely identify the boundary between the reference object and background, is also important for improving segmentation accuracy. Most of the existing R-VOS methods [11,13,17] use a conventional Feature Pyramid Network (FPN) [18] to fuse features of multiple levels. Despite showing good performance, there is still a lot of room for improvement because of the information loss in this progressive fusion process.
To address the above problems, we propose DCT, an end-to-end decoupled cross-modal transformer for referring video object segmentation, which adopts a DETR-like [19] framework. In particular, in order to precisely distinguish the referred target, we firstly propose a Language-Guided Visual Enhancement Module (LGVE). It uses cross-modal attention operations to transmit discriminative linguistic information to visual features of all levels, so as to strengthen the response of the referred area and preliminarily filter out irrelevant background. In addition, a decoupled transformer decoder is designed. In this module, object queries interact with visual and linguistic features parallelly to reduce the attention bias caused by feature size differences. For accurate boundary locating, we leverage the existing Cross-layer Feature Pyramid Network (CFPN) [20] structure to conduct direct communication across multiple layers, which reduces the information loss during the commonly used stage-wise fusion process.
In summary, the main contributions of this paper are as follows:
  • An end-to-end unified network termed DCT is proposed to tackle referring video object segmentation, which sufficiently utilizes multi-modal information and aggregates multi-scale visual features.
  • The Language-Guided Visual Enhancement Module (LGVE) and the decoupled transformer decoder are constructed to establish coordinated information interactions among object queries, visual features and linguistic features.
  • Cross-layer Feature Pyramid Network (CFPN) is brought in to reduce the information loss in the progressive fusion process.
  • Experiments on four benchmarks demonstrate that our proposed method achieves competitive segmentation accuracy compared with the state-of-the-art methods.

2. Related Works

2.1. Referring Video Object Segmentation

Referring video object segmentation (R-VOS) is an important research area at the intersection of computer vision and natural language processing. Existing R-VOS methods can be broadly categorized into three types: propagation-based methods [5,12,21], matching-based methods [22,23], and query-based methods [13,14,16].
Propagation-based methods apply image-level referring object segmentation methods [9,24] on individual video frames and then acquire important temporal context through mask propagation. Seo et al. [12] designed a memory attention module based on self-attention architecture to propagate spatio- and temporal information from memory frames to the current frame, enhancing the temporal consistency of the segmentation results. Hui et al. [21] used textual information to guide the weighted combination of features from the current frame and reference frames, further optimizing the feature representation of the referred target. Propagation-based R-VOS methods are simple and fast, but prone to error accumulation, particularly when there are significant changes in the appearance of the target.
Matching-based methods divide the R-VOS task into two steps: trajectory generation and cross-modal matching. In the trajectory generation step, a model for instance segmentation or object detection is used to identify all the objects. Then, the objects are associated across the entire video to construct a collection of object trajectories. In the cross-modal matching step, the methods compute the relevance scores between each trajectory and the language description and select the pair which matches best. However, methods of this type [22,23] have a more complex training process as they require separate optimization for their multiple submodules.
Queried-based methods view R-VOS as a sequence prediction problem. They introduce a set of object queries to represent video entities and link them across frames to achieve natural tracking. Botach et al. [13] proposes the first query-based R-VOS method, which is named MTTR. The method includes no text-related inductive bias modules or post-processing operations, which greatly simplifies the segmentation process. However, unlike object detection and panoramic segmentation, object attributes in R-VOS are more random and difficult to accurately describe by fixed query vectors. In this regard, Wu et al. [14] take the given language expression as a constraint and generated query vectors online to make them more focused on the referred target.

2.2. Transformer

Transformer [25] is an attention-based encoder–decoder architecture which has a remarkable ability to capture long-term global dependencies. It was originally used for sequence modeling in machine translation and has been widely applied in natural language processing (NLP) [26,27] and computer vision (CV) [28,29] tasks. More recently, transformer has also been introduced to the highly regarded multi-modal domain, which provides valuable insights for R-VOS. For example, the large-scale pre-training model CLIP [30] uses transformers to extract visual and text features and accurately align them in the embedding space through contrastive learning. Ding et al. [31] exploits transformer as the cross-modal decoder for referring image segmentation and proposes a query generation module based on multi-head attention, which can comprehend the given language expression from different perspectives under the guidance of visual cues. The proposal of DETR [19], an end-to-end object detector, is a significant milestone in the development of transformer. It introduces the query-based paradigm and simplifies the conventional pipeline of object detection. MDETR [32] extends this idea to the field of referring expression comprehension, proposing an end-to-end modulated detector that detects objects in an image conditioned on a raw text query. VisTR [33] employs a non-auto-regressive transformer to parallelly supervise and segment the video instances at the sequence level. Considering the simplicity and efficiency of this DETR-like framework, our proposed method also adopts this architecture, but further addresses the undesirable attention bias in the interaction between object query and multi-modal features.

3. Method

3.1. Overall Pipeline

Given a T-frame video clip V = { v t } t = 1 T with the spatial resolution of H × W and a L-word text expression = { r l } l = 1 L , the aim of DCT is to generate a binary segmentation mask sequence = { m t } t = 1 T for the referred object. The overall pipeline of DCT is shown in Figure 1. It consists of four components: Feature Extraction and Enhancement, the decoupled transformer decoder, Instance Segmentation and Instance Sequence Matching process.
Feature Extraction and Enhancement. For the given video–text pair, we first use a visual encoder and a linguistic encoder for feature extraction, then use LGVE to achieve language-guided visual enhancement. Specifically, Video Swin transformer [34] is adopted to extract the multi-level visual features of the video frames, which are F v 1 T × C 1 × H 4 × W 4 , F v 2 T × C 2 × H 8 × W 8 and F v 3 T × C 3 × H 16 × W 16 . Meanwhile, a pretrained language model RoBERTa [35] is employed to extract the word-level linguistic feature F l C l × L . Considering that the visual features do not include a particular focus on the referred object, we propose the LGVE to highlight language-related visual regions, generating the enhanced visual features F e i , i = 1 , 2 , 3 .
Decoupled Transformer Decoder. In this module, we introduce a set of N object queries Q = { q t } t = 1 T ,   q t N × C q to represent the instances for each frame. Firstly, the top-level feature F e 3 output by LGVE is picked out and combined with fixed positional encodings. Then, we feed Q into stacked transformer decoder layers along with the features F l and F e 3 to collect and store entity-related information. Details on this can be found in Section 4.3.
Instance Segmentation. On top of the transformer decoder and LGVE, we build a CFPN spatial decoder and two prediction heads to obtain the mask sequences. In particular, the CFPN spatial decoder takes the multi-level enhanced visual features as inputs and outputs the segmentation feature map F s e g = { f s e g t } t = 1 T for each frame. The mask head comprises three stacked linear layers and produces dynamic kernels Ω from each object query, which is then convolved with F s e g to obtain N instance sequences. The class head is a single-layer perceptron. It predicts the binary confidence score of each sequence, which indicates whether it matches the referred target.
Instance Sequence Matching and Loss. Having the instance sequences and their class scores, we proceed to find the optimal assignment between ground-truth and the predictions using Hungarian matching [36]. The loss of DCT is same as the one in [13], which is composed of c l s and m a s k . Specifically, c l s is a cross-entropy loss while m a s k is a combination of the Dice [37] and binary Focal loss [38]. The whole loss function is as follows:
= λ c l s C E + λ d D i c e + λ f F o c a l
where λ c l s , λ d and λ f are three hyperparameters. More detailed settings can be seen in Section 4.2.

3.2. Language-Guided Visual Enhancement

Since the visual features initially extracted by the backbone contain no special concentration on the referred object, it is important to convey the target-related linguistic semantics to redistribute their attention. MTTR [13] use a standard self-attention operation to facilitate information exchange only between the linguistic feature and the top-level visual feature, which may widen the semantic gap within the visual features of different levels. ReferFormer [14] proposes a Cross-modal Feature Pyramid Network to perform multi-scale cross-modal fusion but places it behind the transformer. As a result, the transformer still encounters difficulty in clearly identifying the target. Therefore, we design the Language-Guided Visual Enhancement Module, which incorporates linguistic information into visual features of all levels and puts it ahead. This module acts as a coarse locater and performs an initial filtering of irrelevant background regions.
The structure of LGVE is shown in Figure 2. Here, we use f v to represent the multi-level visual features of a single frame, and f v i C i × W i × H i is the one of level-i. Firstly, our LGVE generates the query (Q) and the intermediate visual feature f m i d i from f v i with 1 × 1 point-wise convolutions followed by 3 × 3 depth-wise convolutions, aggregating pixel-wise cross-channel context and channel-wise spatial context. Simultaneously, the key (K) and value (V) are generated from the linguistic feature F l with two linear projections. Then, we use cross-attention operations to assemble word-level linguistic features at each spatial location and produce a vision–language correlation filter S , whose dimension is the same as f v i . The above process is shown in Equations (2) and (3).
( f m i d i , Q , K , V ) = ( W d m W p m f v i , W d q W p q f v i , W l k F l , W l v F l )
S = Softmax ( Q K T C i ) V
where W d ( ) , W p ( ) and W l ( ) are 1 × 1 point-wise convolution, 3 × 3 depth-wise convolution and linear projections, respectively.
Finally, the intermediate feature f m i d i is multiplied by the spatial filter S element-wisely to obtain the i-th level enhanced visual feature f e i :
f e i = S f m i d i

3.3. Decoupled Transformer Decoder

The transformer module is a key component in query-based R-VOS methods. Its objective is to gather entity-related information from the multi-modal features and store it in the object queries. However, the decoding process of existing works ignores the severe imbalance in the sizes of multi-modal features. As the length of the visual feature is much longer than that of the linguistic feature (often 20 times longer or more), the object queries may be greatly biased to the former, which is not conducive to a fine-grained understanding of the language expression.
To address this problem, we consider separately interacting the object queries with features of the two modalities and merging the results with adaptive weights. As is shown in Figure 3, the decoupled transformer decoder has N d layers. In each decoder layer, we first perform a self-attention operation upon the object queries q to model its inner relationship and output q s . Then, two cross-attention layers are constructed to collect information from the linguistic feature F l and enhanced visual feature f e 3 independently, generating the single-model subqueries q t e x t and q v i d e o . Finally, the updated query q is obtained by a weighted sum of the subqueries, where the weights are learned from the subquery embeddings with liner projections. The above process can be represented by Equations (5)–(7).
q s = LN ( MSA ( q ) + q )
q t e x t = LN ( MCA ( q s , F l ) + q s ) q v i d e o = LN ( MCA ( q s , f e 3 ) + q s )
q = q t e x t × θ l ( q t e x t ) + q v i d e o × θ v ( q t e x t )
where MSA ( ) and MCA ( ) are multi-head self-attention and multi-head cross-attention. LN represents layer normalization. θ l and θ v are fully connected layers.

3.4. Cross-Layer Feature Pyramid Network

As described earlier, FPN is a classical method for fusing multi-scale features in R-VOS methods and brings about obvious improvements in most cases. However, in this kind of progressive fusion process, the low-level visual information (e.g., object texture and boundary details) is accessed only once in the final fusion stage, resulting in a low-qualified segmentation of the edges. Additionally, as the fusion proceeds, high-level semantic cues are gradually diluted, which diminishes the model’s ability to recognize the referred target. In view of the above two issues, we replace standard FPN with the Cross-layer Feature Pyramid Network (CFPN) [20]. Compared with FPN, CFPN promotes the information exchange among visual features of all layers by aggregating them simultaneously, thus generating a segmentation feature map rich in both semantics and spatial details.
As shown in Figure 4, CFPN first performs global average pooling (GAP) and concatenation operations on the enhanced visual features, resulting in a 1d global representation, Z. Then, we transform Z into the layer-wise fusion weight ψ = { φ i } i = 1 3 with a two-layer perceptron, formulated as
ψ = θ 1 ( ReLU ( θ 2 ( Z ) ) )
where θ 1 D × Y and θ 2 Y × 3 are fully connected layers, D = C 1 + C 2 + C 3 is the channel number of Z, Y is set to 256 empirically, and ReLU refers to the ReLU activation function. Afterwards, the dynamical fusion weight is used to rescale the original visual features and form the aggregated visual representation f g
f g = ( f e 1 × φ 1 ) UP ( f e 2 × φ 2 ) UP ( f e 3 × φ 3 )
where denotes the concatenation operation and UP refers to upsampling. Considering that f g is a naïve concatenation of the multi-level features, CFPN further constructs a cross-layer feature distribution structure. To be more precise, f g is fed into a set of average pooling layers followed by 3 × 3 convolutions to generate the redistributed features f d i ,   i = 1 , 2 , 3 , which are subsequently merged in a top–down manner to yield the final segmentation feature map f s e g . In contrast to FPN, the feature maps in f d collect information from the full spectrum of multi-level representation. This enables the retention of more discriminative and complementary visual information during the fusion process.

4. Experiments

4.1. Datasets and Evaluation Metric

4.1.1. Datasets

To evaluate the proposed method, we experiment on three open-sourced R-VOS datasets: A2D-Sentences [7], JHMDB-Sentences [7] and Ref-Youtube-VOS [12].
A2D-Sentences is an extension of the A2D [39] dataset. It consists of 3782 YouTube videos and 6655 textual descriptions, covering eight types of actions performed by seven categories of objects.
JHMDB-Sentences is an extension of the J-HMDB [40] dataset. It comprises 928 video sequences showcasing 21 human actions, each accompanied by a corresponding textual description. Notably, each frame of the videos is labeled with a 2d puppet mask.
Refer-YouTube-VOS is built upon the large-scale video segmentation dataset YouTube-VOS [41]. It consists of a total of 3978 videos, of which 3471 are used for training, 202 for validation, and 305 for testing. Each video in this dataset is annotated with high-quality instance segmentation masks for every fifth frame. Since the test set is accessible only during the competition, our evaluation experiments are conducted on the validation set.

4.1.2. Evaluation Metric

For A2D-Sentences and JHMDB-Sentences, we adopt precision at thresholds of 0.5, 0.6, 0.7, 0.8, and 0.9(P@X), overall Intersection-over-Union (OIoU), mean Intersection-over-Union (MIoU) and mean average precision over 0.50:0.05:0.95 (mAP) as evaluation metrics.
For Refer-YouTube-VOS, the method is evaluated with the criteria of region similarity ( J ), contour accuracy ( ) and their average value ( J & ).

4.2. Inplementation Details

In accordance with previous works [13,14], we train DCT on A2D-Sentences and Refer-YouTube-VOS and use all three datasets for evaluation. For model settings, the decoupled transformer has 4 decoder layers ( N d = 4 ) and each layer is configured with 8 attention heads. The number of object queries is set to 50. The hyperparameters of the loss function are set as λ c l s = 2 , λ d = 5 and λ f = 2 .
During training, we use sliding windows to crop video clips and the default window size is set to eight. The resolution of the frames is adjusted to ensure that the shorter side is at least 360 pixels and the longer side is at most 640 pixels. Random horizontal flipping, random cropping, and photometric distortion are used for data augmentation. AdamW is used as the optimizer and the weight decay is set to 1 × 10−4. For A2D-Sentences, we train the model for 60 epochs with a batch size of 6 and a dynamic learning rate, which is as shown in Equation (10). For Refer-YouTube-VOS, the epoch and batch size are 30 and 4, respectively, and the learning rate is shown in Equation (11).
l r A 2 D ( epoch ) = { 0.0001   , 0 epoch < 40 0.00005   , 40 epoch < 60
l r Refer ( epoch ) = { 0.0001   , 0 epoch < 20 0.00005   , 20 epoch < 30
When inferencing, DCT predicts N instance sequences corresponding to the N queries. For each sequence, we sum the confidence scores output by the class head across frames. Finally, the sequence with the highest total score is identified as the referred object.

4.3. Ablation Study

4.3.1. Ablation Study on the Main Components

We conduct extensive experiments on A2D-Sentences to evaluate the effectiveness of the key components in our proposed method. The results are presented in Table 1. In the baseline model (as shown in the first line), the LGVE is removed while the decoupled transformer decoder and CFPN are replaced with a conventional transformer decoder and a common FPN similar to the ones used in MTTR [13]. In this case, the OIoU, MIoU and mAP drop significantly by 3.8, 4.1 and 5.4. In experiment No.2 to No.4, we introduce the three proposed components on top of the baseline separately. Obviously, each component can bring about an improvement in the segmentation accuracy. In experiment No.5, we achieve a higher accuracy by simultaneously introducing the LGVE and CFPN. The reason for this may be that the LGVE facilitates the semantic consistency of the visual features, thus reducing the misalignment caused by semantic discrepancies during the global fusion in CFPN.

4.3.2. Analysis of the Temporal Context Size

Modeling the temporal context plays a crucial role in the R-VOS task for which actions or behaviors in videos cannot be fully understood or derived by analyzing a single frame. In DCT, we use sliding windows to crop video clips and adopt Video Swin transformer as the spatial–temporal encoder. To study the effect of the temporal context size, we change the window size during training and evaluating on A2D-Sentences dataset. The results are presented in Table 2. When the window size is 1, the model essentially transforms into an image-level approach, with mAP being only 42.6. As the window size increases, the model captures more time clues, and the segmentation performance gradually improves. However, the metrics reach their peak at the window size of 8. One possible reason for this phenomenon could be that over a longer time span, the target’s behavior and spatial position undergo more pronounced changes, which consequently make the target more challenging for the model to comprehend.

4.3.3. Analysis of the Query Number

To investigate the impact of the number of object queries on model performance, we conduct ablation experiments on the A2D-Sentences dataset under the window size of 8, as shown in Table 3. During the training process, randomly initialized object queries gradually converge towards fixed regions or specific categories of targets. As a result, a low number of queries (e.g., 5) are insufficient to cover the complex distribution of objects in the dataset. On the other hand, when queries are too dense (e.g., 75), more similar mask sequences are generated, making it more challenging for the model to optimize during the one-to-one Hungarian matching. Therefore, the number of queries for the final model is set to 50.

4.4. Comparison with Existing Methods

Results on A2D-Sentences and JHMDB-Sentences. We compare our proposed method on A2D-Sentences and JHMDB-Sentences with the current state-of-the-art methods, including Hu et al. [4], Gavrilyuk et al. [7], CMSA + CFSA [42], ACAN [43], CMPC-V [44], ClawCraneNet [45], MTTR [13] and ReferFormer [14]. As shown in Table 4, our proposed method achieves 73.5 OIoU, 65.0 MIoU and 48.7 mAP on A2D-Sentences, which greatly outperforms previous methods using CNN-based visual encoders (e.g., CMPC-V and ClawCraneNet). Besides, when equally employing Video-Swin-T as the visual encoder, the proposed method also has certain improvements over the query-based methods MTTR and ReferFormer. Table 5 shows the performance comparison on JHMDB-Sentences dataset. It can be seen that DCT achieves OIoU, MIoU and mAP values of 70.7, 70.5 and 39.6, respectively, which are 0.7, 1.2 and 0.5 higher than the best method ReferFormer. As the results on JHMDB-Sentences are obtained by evaluating the model trained on A2D-Sentences without finetuning, DCT further proves its good generalization.
Results on Refer-YouTube-VOS. Table 6 displays the experimental results on the largest and challenging R-VOS dataset Refer-YouTube-VOS. For fair comparison, we report the performance of ReferFormer trained from scratch and without post-process operations. It can be observed that our DCT surpasses all the cutting-edge methods with a gain of 0.6 J and 0.5 and achieves the state of the art.

4.5. Visualization and Analysis

In Figure 5, we show some of the visualization results of DCT on Ref-Youtube-VOS. It can be observed that our DCT can accurately comprehend the given language expression and accomplish precise segmentation even in cases of severe object deformation (as shown in row 1), interference from similar objects (as shown in row 2) and partial disappearance of the target (as shown in row 3).

5. Discussion

In this paper, we optimize the modules of multi-modal feature interaction and decoding segmentation specially, which brings about a great improvement in the segmentation performance. However, the method still has some limitations. For example, we link the masks across frames according to the permutation of the queries, but do not design a memory module to store the historical features of the referred targets, resulting in an insufficient utilization of the temporal information. In addition, during the inference process, we select the mask sequence with the highest total confidence score as the final output. This approach may lead to false-positive results when the referred target does not appear in the video. In summary, how to collect and utilize temporal information from historical frames while maintaining the simplicity of the model framework, as well as improving the strategy for instance matching, remains worthy of further research.

6. Conclusions

This paper presents an end-to-end decoupled cross-modal transformer for referring video object segmentation. In the proposed model, a Language-Guided Visual Enhancement Module (LGVE) and a decoupled transformer decoder are designed to establish sufficient and balanced information interactions between the object queries and features of different modalities, so as to accurately identify the referred object. Then, we introduce the Cross-layer Feature Pyramid Network, which was originally used for salient object detection, as the spatial decoder, generating high-qualified object boundaries by making a better use of the visual semantics and details. Extensive experiments are carried out on three benchmark datasets. The experimental results fully demonstrate the effectiveness of the proposed method.

Author Contributions

Conceptualization: A.W.; methodology: A.W.; formal analysis and investigation: A.W. and R.W.; Writing—original draft preparation: A.W.; writing—review and editing: A.W. and Q.T.; funding acquisition: R.W.; supervision: R.W., A.W. and Z.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Double First-Class Innovation Research Project for the People’s Public Security University of China, grant number 2023SYL08.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available online. These datasets were derived from the following public resources: [A2D-Sentences, JHMDB-Sentences, Ref-Youtube-VOS].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Caelles, S.; Maninis, K.-K.; Pont-Tuset, J.; Leal-Taixé, L.; Cremers, D.; Van Gool, L. One-shot video object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 221–230. [Google Scholar]
  2. Maninis, K.-K.; Caelles, S.; Pont-Tuset, J.; Van Gool, L. Deep extreme cut: From extreme points to object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 616–625. [Google Scholar]
  3. Yang, Z.; Wei, Y.; Yang, Y. Collaborative video object segmentation by foreground-background integration. In Proceedings of the European Conference on Computer Vision, Online, 23–28 August 2020; pp. 332–348. [Google Scholar]
  4. Hu, R.; Rohrbach, M.; Darrell, T. Segmentation from natural language expressions. In Computer Vision–ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Part I 14; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 108–124. [Google Scholar]
  5. Bellver, M.; Ventura, C.; Silberer, C.; Kazakos, I.; Torres, J.; Giro-i-Nieto, X. Refvos: A closer look at referring expressions for video object segmentation. arXiv 2020, arXiv:2010.00263. [Google Scholar]
  6. Liu, C.; Lin, Z.; Shen, X.; Yang, J.; Lu, X.; Yuille, A. Recurrent multimodal interaction for referring image segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1271–1280. [Google Scholar]
  7. Gavrilyuk, K.; Ghodrati, A.; Li, Z.; Snoek, C.G. Actor and action video segmentation from a sentence. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5958–5966. [Google Scholar]
  8. Wang, H.; Deng, C.; Ma, F.; Yang, Y. Context modulated dynamic networks for actor and action video segmentation with language queries. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12152–12159. [Google Scholar]
  9. Ye, L.; Rochan, M.; Liu, Z.; Wang, Y. Cross-modal self-attention network for referring image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10502–10511. [Google Scholar]
  10. Hu, Z.; Feng, G.; Sun, J.; Zhang, L.; Lu, H. Bi-directional relationship inferring network for referring image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4424–4433. [Google Scholar]
  11. Yang, Z.; Wang, J.; Tang, Y.; Chen, K.; Zhao, H.; Torr, P.H. Lavt: Language-aware vision transformer for referring image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18155–18165. [Google Scholar]
  12. Seo, S.; Lee, J.-Y.; Han, B. Urvos: Unified referring video object segmentation network with a large-scale benchmark. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020, Part XV 16; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 208–223. [Google Scholar]
  13. Botach, A.; Zheltonozhskii, E.; Baskin, C. End-to-end referring video object segmentation with multimodal transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4985–4995. [Google Scholar]
  14. Wu, J.; Jiang, Y.; Sun, P.; Yuan, Z.; Luo, P. Language as queries for referring video object segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4974–4984. [Google Scholar]
  15. Li, X.; Wang, J.; Xu, X.; Li, X.; Lu, Y.; Raj, B. R^2VOS: Robust Referring Video Object Segmentation via Relational Multimodal Cycle Consistency. arXiv 2022, arXiv:2207.01203. [Google Scholar]
  16. Yuan, L.; Shi, M.; Yue, Z. LoSh: Long-Short Text Joint Prediction Network for Referring Video Object Segmentation. arXiv 2023, arXiv:2306.08736. [Google Scholar]
  17. Feng, G.; Zhang, L.; Hu, Z.; Lu, H. Deeply Interleaved Two-Stream Encoder for Referring Video Segmentation. arXiv 2022, arXiv:2203.15969. [Google Scholar]
  18. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  19. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Online, 23–28 August 2020; pp. 213–229. [Google Scholar]
  20. Li, Z.; Lang, C.; Liew, J.H.; Li, Y.; Hou, Q.; Feng, J. Cross-layer feature pyramid network for salient object detection. IEEE Trans. Image Process. 2021, 30, 4587–4598. [Google Scholar] [CrossRef] [PubMed]
  21. Hui, T.; Liu, S.; Huang, S.; Li, G.; Yu, S.; Zhang, F.; Han, J. Linguistic structure guided context modeling for referring image segmentation. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020, Part X 16; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 59–75. [Google Scholar]
  22. Liang, C.; Wu, Y.; Zhou, T.; Wang, W.; Yang, Z.; Wei, Y.; Yang, Y. Rethinking cross-modal interaction from a top-down perspective for referring video object segmentation. arXiv 2021, arXiv:2106.01061. [Google Scholar]
  23. Yang, J.; Huang, Y.; Niu, K.; Huang, L.; Ma, Z.; Wang, L. Actor and action modular network for text-based video segmentation. IEEE Trans. Image Process. 2022, 31, 4474–4489. [Google Scholar] [CrossRef] [PubMed]
  24. Yu, L.; Lin, Z.; Shen, X.; Yang, J.; Lu, X.; Bansal, M.; Berg, T.L. Mattnet: Modular attention network for referring expression comprehension. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1307–1315. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  26. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  27. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  28. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  29. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  30. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  31. Ding, H.; Liu, C.; Wang, S.; Jiang, X. Vision-language transformer and query generation for referring segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16321–16330. [Google Scholar]
  32. Kamath, A.; Singh, M.; LeCun, Y.; Synnaeve, G.; Misra, I.; Carion, N. Mdetr-modulated detection for end-to-end multi-modal understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1780–1790. [Google Scholar]
  33. Wang, Y.; Xu, Z.; Wang, X.; Shen, C.; Cheng, B.; Shen, H.; Xia, H. End-to-end video instance segmentation with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8741–8750. [Google Scholar]
  34. Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video swin transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3202–3211. [Google Scholar]
  35. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  36. Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
  37. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  38. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  39. Xu, C.; Xiong, C.; Corso, J.J. Action understanding with multiple classes of actors. arXiv 2017, arXiv:1704.08723. [Google Scholar]
  40. Jhuang, H.; Gall, J.; Zuffi, S.; Schmid, C.; Black, M.J. Towards understanding action recognition. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 3192–3199. [Google Scholar]
  41. Xu, N.; Yang, L.; Fan, Y.; Yang, J.; Yue, D.; Liang, Y.; Price, B.; Cohen, S.; Huang, T. Youtube-vos: Sequence-to-sequence video object segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 585–601. [Google Scholar]
  42. Ye, L.; Rochan, M.; Liu, Z.; Zhang, X.; Wang, Y. Referring segmentation in images and videos with cross-modal self-attention network. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3719–3732. [Google Scholar] [CrossRef] [PubMed]
  43. Wang, H.; Deng, C.; Yan, J.; Tao, D. Asymmetric cross-guided attention network for actor and action video segmentation from natural language query. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 3939–3948. [Google Scholar]
  44. Liu, S.; Hui, T.; Huang, S.; Wei, Y.; Li, B.; Li, G. Cross-modal progressive comprehension for referring segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4761–4775. [Google Scholar] [CrossRef]
  45. Liang, C.; Wu, Y.; Luo, Y.; Yang, Y. Clawcranenet: Leveraging object-level relation for text-based video segmentation. arXiv 2021, arXiv:2103.10702. [Google Scholar]
  46. Ding, Z.; Hui, T.; Huang, S.; Liu, S.; Luo, X.; Huang, J.; Wei, X. Progressive multimodal interaction network for referring video object segmentation. 3rd Large-Scale Video Object Segm. Chall. 2021, 8. [Google Scholar]
Figure 1. Overall architecture of the proposed method.
Figure 1. Overall architecture of the proposed method.
Sensors 24 05375 g001
Figure 2. Language-Guided Visual Enhancement Module.
Figure 2. Language-Guided Visual Enhancement Module.
Sensors 24 05375 g002
Figure 3. Decoupled transformer decoder.
Figure 3. Decoupled transformer decoder.
Sensors 24 05375 g003
Figure 4. Cross-layer Feature Pyramid Network.
Figure 4. Cross-layer Feature Pyramid Network.
Sensors 24 05375 g004
Figure 5. Visualization results on Ref-Youtube-VOS dataset.
Figure 5. Visualization results on Ref-Youtube-VOS dataset.
Sensors 24 05375 g005
Table 1. Ablation experiments on the key components.
Table 1. Ablation experiments on the key components.
No.LGVEDecoupled
Transformer
Decoder
CFPNIoUmAP
OverallMean
1---69.760.943.3
2--71.863.546.3
3--71.062.745.1
4--72.363.346.8
5-73.064.247.2
673.565.048.7
Table 2. Ablation experiments on the sliding-window size.
Table 2. Ablation experiments on the sliding-window size.
Window SizeIoUmAP
OverallMean
169.761.042.6
470.862.044.5
671.963.346.1
873.565.048.7
1072.163.846.5
Table 3. Ablation experiments on the query number.
Table 3. Ablation experiments on the query number.
Query NumberIoUmAP
OverallMean
571.863.045.8
3073.164.447.5
5073.565.048.7
7572.363.646.4
Table 4. Performance comparison on A2D-Sentences dataset.
Table 4. Performance comparison on A2D-Sentences dataset.
MethodBackbonePrecisionIoUmAP
[email protected][email protected][email protected][email protected][email protected]OverallMean
Hu et al. [4] VGG-1634.823.613.33.30.147.435.013.2
Gavrilyuk et al. [7] I3D47.534.721.18.00.253.642.119.8
CMSA + CFSA [42]ResNet-10148.743.135.823.15.261.843.2-
ACAN [43]I3D55.745.931.916.02.060.149.027.4
CMPC-V [44]I3D65.559.250.634.29.865.357.340.4
ClawCraneNet [45]ResNet10170.467.761.748.917.163.159.9-
MTTR [13]Video-Swin-T75.471.263.848.516.972.064.046.1
ReferFormer [14]Video-Swin-T76.072.265.449.817.972.364.148.6
DCT (ours)Video-Swin-T76.372.866.050.218.373.565.048.7
Table 5. Performance comparison on JHMDB-Sentences dataset.
Table 5. Performance comparison on JHMDB-Sentences dataset.
MethodBackbonePrecisionIoUmAP
[email protected][email protected][email protected][email protected][email protected]OverallMean
Hu et al. [4] VGG-1663.335.08.50.20.054.652.817.8
Gavrilyuk et al. [7] I3D69.946.017.31.40.054.154.223.3
CMSA + CFSA [42]ResNet-10176.462.538.99.00.162.858.1-
ACAN [43]I3D75.656.428.73.40.057.658.428.9
CMPC-V [44]I3D81.365.737.17.00.061.661.734.2
ClawCraneNet [45]ResNet10188.079.656.614.70.264.465.6-
MTTR [13]Video-Swin-T93.985.261.616.60.170.169.839.2
ReferFormer [14]Video-Swin-T93.384.261.416.40.370.069.339.1
DCT (ours)Video-Swin-T94.785.562.016.90.170.770.539.6
Table 6. Performance comparison on Refer-YouTube-VOS dataset.
Table 6. Performance comparison on Refer-YouTube-VOS dataset.
MethodBackbone J & J
CMSA [9]ResNet-5034.933.336.5
URVOS [12]ResNet-5047.345.349.2
PMINet [46]ResNeSt-10148.246.749.6
MTTR [13]Video-Swin-T55.354.056.6
ReferFormer [14]Video-Swin-T56.154.857.3
DCT (ours)Video-Swin-T56.655.457.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, A.; Wang, R.; Tan, Q.; Song, Z. Decoupled Cross-Modal Transformer for Referring Video Object Segmentation. Sensors 2024, 24, 5375. https://doi.org/10.3390/s24165375

AMA Style

Wu A, Wang R, Tan Q, Song Z. Decoupled Cross-Modal Transformer for Referring Video Object Segmentation. Sensors. 2024; 24(16):5375. https://doi.org/10.3390/s24165375

Chicago/Turabian Style

Wu, Ao, Rong Wang, Quange Tan, and Zhenfeng Song. 2024. "Decoupled Cross-Modal Transformer for Referring Video Object Segmentation" Sensors 24, no. 16: 5375. https://doi.org/10.3390/s24165375

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop