Next Article in Journal
Lithological Classification Using ZY1-02D Hyperspectral Data by Means of Machine Learning and Deep Learning Methods in the Kohat–Pothohar Plateau, Khyber Pakhtunkhwa, Pakistan
Previous Article in Journal
Unified Model and Survey on Modulation Schemes for Next-Generation Automotive Radar Systems
Previous Article in Special Issue
Global–Local Feature Fusion of Swin Kansformer Novel Network for Complex Scene Classification in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FreeMix: Open-Vocabulary Domain Generalization of Remote-Sensing Images for Semantic Segmentation

by
Jingyi Wu
1,2,
Jingye Shi
3,
Zeyong Zhao
1,2,
Ziyang Liu
1,2 and
Ruicong Zhi
1,2,*
1
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
Beijing Key Laboratory of Knowledge Engineering for Material Science, Beijing 100083, China
3
Beijing Key Laboratory of Security and Privacy in Intelligent Transportation, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(8), 1357; https://doi.org/10.3390/rs17081357
Submission received: 30 December 2024 / Revised: 14 February 2025 / Accepted: 17 February 2025 / Published: 11 April 2025

Abstract

:
In this study, we present a novel concept termed open-vocabulary domain generalization (OVDG), which we investigate within the context of semantic segmentation. OVDG presents greater difficulty compared to conventional domain generalization, yet it offers greater practicality. It jointly considers (1) recognizing both base and novel classes and (2) generalizing to unseen domains. In OVDG, only the labels of base classes and the images from source domains are available to learn a robust model. Then, the model could be generalized to images from novel classes and target domains directly. In this paper, we propose a dual-branch FreeMix module to implement the OVDG task effectively in a universal framework: the base segmentation branch (BSB) and the entity segmentation branch (ESB). First, the entity mask is introduced as a novel concept for segmentation generalization. Additionally, semantic logits are learned for both the base mask and the entity mask, enhancing the diversity and completeness of masks for both base classes and novel classes. Second, the FreeMix utilizes pretrained self-supervised learning on large-scale remote-sensing data (RS_SSL) to extract domain-agnostic visual features for decoding masks and semantic logits. Third, a training tactic called dataset-aware sampling (DAS) is introduced for multi-source domain learning, aimed at improving the overall performance. In summary, RS_SSL, ESB, and DAS can significantly improve the generalization ability of the model on both a class level and a domain level. Experiments demonstrate that our method produces state-of-the-art results on several remote-sensing semantic-segmentation datasets, including Potsdam, GID5, DeepGlobe, and URUR, for OVDG.

1. Introduction

Remote -sensing images (RSI) are typically obtained from satellites, aerial platforms, or drones and provide valuable information about the Earth’s surface and its features. Semantic segmentation of these images, where the goal is to classify each pixel in an image into predefined categories, is crucial for various applications, including urban planning, environmental monitoring, agriculture, disaster management, and military intelligence. However, classes that are not labeled and not visible during the training process cannot be recognized during the inference stage, which greatly limits the scope of the application. For instance, the Potsdam dataset [1], which is widely used for benchmarking semantic-segmentation algorithms, only includes five classes. We call them base classes. However, in reality, RSIs often have more than five different types of objects. The classes that fall outside the five classes are called novel classes, which are presented as clutter or background in the Potsdam dataset. To identify novel classes, researchers have introduced different concepts like open-set learning [2,3,4], open-world learning [5,6], out-of-distribution detection (OOD) [7,8], zero-shot learning (ZSL) [9,10,11,12,13,14,15,16], and open-vocabulary learning (OVL) [17,18,19,20,21,22,23,24,25,26,27]. The differences of these concepts are shown in Figure 1a. The open-set, open-world, and OOD tasks only need to identify novel classes and set them as one label named ‘unknown’. They do not need to identify the specific class categories in such settings. However, zero-shot must classify novel classes into specific categories using predefined word embeddings [28,29,30]. During training, the zero-shot learning model is strictly trained on base classes. In the open-vocabulary setting, the model can classify novel classes with the help of pretrained vision–language models, which have large language vocabulary knowledge but are not strictly required to contain base classes and novel classes. Compared with zero-shot learning, open-vocabulary learning can further extend models’ generalizability on a class level [31]. Despite the resounding success of open-vocabulary learning in the computer vision field, its potential application in the context of RSI remains relatively unexplored. In addition, open-vocabulary learning assumes that samples come from a single known domain, resulting in limited applications in the real world.
Due to the influence of spatial resolution, acquisition time, sensor type, and geographical location, RSIs have different styles. Therefore, there is a significant domain shift between different remote-sensing datasets. The model trained on a specific domain exhibits poor performance on the target domains, which have a different but related distribution. To this end, it is necessary to develop a model that can generalize to unseen distributions. As illustrated in Figure 1b, domain adaptation (DA) focuses on adapting a model from one domain to another, while domain generalization aims to develop models that can generalize well across multiple domains without explicit domain-specific data during training. Moreover, domain generalization (DG) involves addressing a difficult scenario where one or multiple distinct yet interconnected domains are provided, with the objective being to train a model capable of generalizing effectively to an unseen test domain. However, existing techniques for domain generalization presuppose the presence of identical classes across all domains, which limits the practical applicability of these methods [32,33]. Our endeavor seeks to alleviate this constraint by enabling unseen test domains to contain novel classes absent in any training domain. As indicated in Figure 1c, we introduce this more challenging setting as open-vocabulary domain generalization (OVDG) for semantic segmentation, which, to the best of our understanding, represents the inaugural attempt at tackling this issue. Compared with DG, our OVDG demands not only considering the generalization performance on unseen domains but also identifying novel classes. The closest match to our setting is universal domain adaptation [34,35,36], and the primary distinction of our OVDG lies in involving source domains and identifying novel classes rather than simply labeling them as ’unknown’. OVDG combines the challenges of open-vocabulary learning (generalizing to unseen classes) and domain generalization (generalizing to unseen domains) into a unified framework, addressing a more realistic and complex scenario. Unlike existing methods that handle these tasks independently, OVDG enables models to generalize to both unseen domains and classes simultaneously, which is critical for remote-sensing applications where data spans diverse domains (e.g., urban, rural) and includes novel classes. This combination is novel because it leverages external knowledge (e.g., language models) for open-vocabulary learning while ensuring robustness to domain shifts, making it uniquely suited for real-world scenarios where labeled data from all domains or classes is unavailable. OVDG thus advances the field by providing a flexible and powerful approach for handling the dynamic and diverse nature of RSI.
Existing semantic-segmentation models [37,38,39,40] lack the capacity to handle OVDG tasks. For example, conventional semantic-segmentation models are trained on closed sets, which limits their ability to generalize at the class and domain levels. While there are existing methods [41,42,43,44] that focus on the model’s capacity to generalize to new visual concepts, such as zero-shot learning (ZSL) or generalized zero-shot learning (GZSL). Existing open-vocabulary semantic segmentation (OVS) methods have achieved remarkable success in natural image domains, leveraging large-scale vision–language models like CLIP to generalize to unseen classes. However, these methods face notable performance limitations when applied to RSI, primarily due to the absence of tailored designs for the unique characteristics of RSIs. Rapidly changing orientations, significant scale variations, and domain-specific semantics pose significant challenges for existing OVS methods. As a result, their direct application to RSIs often leads to subpar performance, highlighting the need for tailored approaches for RSIs.
In OVDG, the source data and target data remain isolated from each other. In this setting, aligning the domain distributions becomes unfeasible, unlike traditional methods [9,45,46] that rely on the simultaneous presence of source and target data. This paper introduces FreeMix, an efficient framework for open-vocabulary domain generalization (OVDG). FreeMix includes a dual-branch segmentation module with a base segmentation branch (BSB) and an entity segmentation branch (ESB), along with a CLIP-based recognition module. The dual-branch design generates diverse masks and visual-semantic features for class- and domain-level generalization. The BSB integrates the CMID model [47], adapted for mask and semantic logit decoding as RS_SSL, while the ESB incorporates entity masks from a pretrained extractor and a custom feature extractor. Outputs from both branches are fused for open-category recognition using CLIP. By freezing key components, we reduce trainable parameters and computation. To enhance performance across domains, we propose dataset-aware sampling (DAS), ensuring uniform domain sampling in each mini-batch. Our contributions are as follows:
1.
We introduce a new setting for semantic segmentation, i.e., open-vocabulary domain generalization (OVDG), which is an important yet unstudied problem. In addition, we propose an effective framework, FreeMix, for solving OVDG, which focuses on learning a generalized model by integrating entity masks to enhance the diversity and completeness of masks for both base classes and novel classes.
2.
We propose a dual-branch universal segmentation module by unifying the base segmentation branch (BSB) and the entity segmentation branch (ESB) in an end-to-end trainable framework, where the BSB leverages a self-supervised pretrained model, CMID, to extract domain-agnostic visual features for decoding masks and semantic logits.
3.
To integrate and leverage information from various source domains, we propose a simple yet effective training strategy called dataset-aware sampling (DAS). Extensive experiments on four benchmark datasets reveal that our proposed method outperforms the state-of-the-art methods on the OVL and the OVDG benchmark.
In the following sections of this paper, we will first review the relevant literature in Section 2, which discusses related work in the field. Section 3 introduces our proposed FreeMix method, detailing its innovative approach and underlying principles. In Section 4 and Section 5, we outline the experimental setup and datasets used and present the results of our experiments, providing a comprehensive analysis of the performance of the FreeMix method. Finally, Section 6 concludes the paper, summarizing our findings and suggesting directions for future research.

2. Related Works

2.1. Open-Vocabulary Semantic Segmentation

With the development of visual language pretraining models (VLPMs), such as CLIP [48] and ALIGN [49], models can now localize and recognize classes beyond the annotated label space no longer confined to identifying predefined classes present in the training set. Many studies have successfully transferred their robust class generalization capabilities to pixel-level classification tasks, such as semantic segmentation. Based on a pretrained CLIP model, Chen et al. [50] use a conditional UNet model to predict segmentation masks and use text descriptions and annotations from OpenStreetMap as auxiliary supervision. Exploration of open-vocabulary semantic segmentation in the remote-sensing field is relatively limited, while it is more prevalent in natural images. ZegFormer [16], as a simple yet effective zero-shot semantic-segmentation model, decomposes the problem into a class-agnostic segmentation task and mask classification task. Furthermore, it transfers semantic knowledge from seen classes to unseen classes solely with the assistance of VLPMs. Similarly, ZSSeg [51] proposes a two-stage semantic-segmentation framework where the first stage extracts generic candidate masks, and the second stage utilizes the CLIP model for open-vocabulary classification of the mask images generated in the first stage. To circumvent the time-consuming process of clipping image patches and computing features from an external pretrained CLIP model, MaskCLIP [19] designs a Relative Mask Attention (RMA) module, treating segmentation as additional tokens for ViT CLIP models.
Unlike the two-stage segmentation models discussed earlier, SAN [52] attaches a lightweight side network to a pretrained VLM to predict candidate masks and classification outputs. To enhance the versatility of the framework, FreeSeg [24] jointly learns multiple related segmentation tasks, including open-vocabulary semantic (OVSS), instance (OVIS), and panoptic segmentation (OVPS). Open-vocabulary methods achieve impressive results on natural images but remain relatively unexplored in the field of remote sensing. Existing visual language pretraining models lack participation of remote-sensing images in training, resulting in limited generalization capabilities in the field of remote-sensing. To address this limitation, some recent studies have specifically constructed large-scale image-text pairs datasets in the remote-sensing domain for training visual language models. For example, RS5M [53], SkyScript [54], MMRS [55].
RemoteCLIP transforms heterogeneous annotations of detection boxes and segmentation masks into a unified image-caption data format through Box-to-Caption (B2C) and Mask-to-Box (M2B) strategies for training purposes. In addition, GeoRSCLIP [53] proposes an image-text paired dataset RS5M in the remote-sensing domain, consisting of 5 million RSI with English descriptions. Furthermore, building upon the alignment of ground images with text, GRAFT [56] constructs pairs of ground images and remote-sensing images to train VLP models, thereby enabling the training of remote-sensing image vision–language models without using any text annotations.

2.2. Domain Generalization

Domain generalization (DG) task is one of the key challenges for deep learning models, facing domain shift between training and testing distributions. When the source domain consists of only one dataset, it simplifies to single-source domain generalization [32,33]. Existing domain generalization methods can be categorized into three main types. The first type [32,33] involves utilizing data augmentation to assist the model in learning universal representations, such as randomizing, transforming, or generating diverse data inputs to enhance the dataset. The second type [57,58,59] focuses on representation learning, aiming to learn domain-invariant representations or decompose representations into domain-shared and domain-specific components for better generalization, often incorporating feature alignment techniques to minimize discrepancies between different domains. The third type of methods [60,61,62,63] leverage general learning strategies to improve generalization capabilities, including ensemble learning, meta-learning, gradient manipulation, distributed robust optimization, and self-supervised learning. These three types of methods can complement each other and may be combined to achieve higher performance. Specifically, in the first type of method, CCDR [32] performs texture and style randomization for simple yet effective auxiliary domain generation to improve the reliability of classification in arbitrary unseen target domains. Recently, several methods [64,65] have utilized Mixup [66] for domain generalization, generating new samples by applying Mixup directly in the original space.
The second category of methods has received considerable attention. A Maximum Mean Discrepancy Depth Reconstruction Classification Network (MMD-DRCN) [58] is proposed for detecting oil palm trees from multi-source high-resolution satellite images in a new environment. The core idea is to utilize the Maximum Mean Discrepancy (MMD) module to learn invariant features across different source domains. Moreover, to acquire domain-invariant features, the frequency-based optimal style mix (FOSMix) [57] model randomizes the styles of images in the source domain. Additionally, language-aware domain generalization network (LDGnet) [59] is proposed to learn cross-domain-invariant representation from cross-domain shared prior knowledge.
The third type of method leverages general learning strategies to improve generalization capabilities. The approach of leveraging general learning strategies to improve generalization capabilities is simple yet effective; hence, it has gained popularity. Segu et al. [67] maintain domain-specific batch normalization (BN) parameters for different source domains while sharing other parameters. Li et al. [60] proposes Meta-Learning for Domain Generalization (MLDG), applying meta-learning strategies to domain generalization. MLDG divides the data in the source domain into meta-training and meta-testing sets to simulate domain transfer scenarios for learning universal representations. Recently, Self-supervised Learning (SSL) has emerged as a popular learning paradigm, constructing self-supervised tasks from large-scale unlabeled data. SSL, as a universal paradigm, can be applied to any existing DG method, particularly for unsupervised domain generalization where labeled data are unavailable in the training domain. Bhattacharya et al. [68] proposes a self-supervised prompt learning approach for remote-sensing images, which preserves domain-invariant feature learning while enhancing the representation of visual features. Our approach falls into this category by introducing a self-supervised backbone network to achieve domain generalization.

2.3. Self-Supervised Learning in Remote Sensing

Self-supervised learning (SSL) has garnered significant attention in the remote-sensing community and has undergone initial exploration in this field. SSL methods leverage large amounts of unlabeled data to learn generic representations, which can enhance the performance of downstream tasks. Contrastive self-supervised learning involves constructing contrastive learning tasks to train models by leveraging the inherent positive and negative samples within remote-sensing data. For instance, SauMoCo [69] utilizes the semantic similarity between nearby geographical locations and the inherent diversity within land-cover concepts to train the model. Moreover, Kumar et al. [70] utilizes temporally aligned images as positive sample data and introduces an auxiliary task of predicting the source of the images to enhance pretraining effectiveness. Similarly, Oscar et al. [71] exploits seasonal information inherent in the data to construct tasks for SSL. In addition to temporal information, Dilxat et al. [72] proposes IndexNet, which learns spatiotemporal invariant features by combining image-level contrast and pixel-level contrast.
Compared to contrastive self-supervised learning, Masked Image Modeling (MIM) self-supervised learning has become more popular. Studies have focused on collecting data from various sources, such as satellites or aerial platforms, to build large-scale datasets covering multiple scenes worldwide for MIM self-supervised pretraining. RingMo [73] optimizes mask strategies for small objects in remote-sensing images and employs the MAE model for self-supervised representation learning on a dataset comprising 3 million unlabeled remote-sensing images. To better represent robust remote-sensing data across various spatial scales, Scale-MAE [74] explicitly learns the relationships between data at different known scales throughout the entire pretraining process. In addition, to handle large-sized images and objects of different orientations in RSI, RVSA [75] introduces a new Rotated Varied-Size Window Attention (RVSA) mechanism, which significantly reduces computational costs and memory usage. Recently, a Transformer-based geographic spatial foundation model named Prithvi [76] has been proposed, pretraining it on over 1TB of multispectral satellite images from the Harmonized Landsat-Sentinel 2 (HLS) dataset.
Despite achieving some success, these methods have prerequisites in terms of underlying architectures; namely, contrastive self-supervised learning relies on CNNs, while most MIM methods are restricted to ViTs. In contrast, the CMID [47] method is not only agnostic to architectures but also combines contrastive and masked generation approaches to learn representations with global semantic separability and local spatial awareness. This ensures that the learned representations have sufficient generalization to meet the requirements of various downstream tasks in remote sensing. Furthermore, SMLFR [77] has constructed a large dataset named GeoSense, comprising approximately 9 million diverse remote-sensing images, to enhance the robustness and generalization capabilities of foundation models during the pretraining phase. Additionally, they implemented masked image modeling (MIM) based on CNN architectures.

3. Proposed Method

3.1. Problem Definition

Let X represent visual space, and Y represent label space. We define a domain as the joint distribution of data space. It can be represented as: D = { ( x i , y i ) X J × Y J } i = 1 J P X Y , where x i represents the i-th sample, y i is the corresponding label and J is the size of the domain. We define Y B as label space with base classes, Y N as label space with novel classes, and Y B + N as label space with both base and novel classes. As illustrated in Figure 1c, we have m source domains D s r c = D i t r a i n , D i t e s t P X Y B + N i i = 1 m and a target domain D t a r = { ( x k , y k ) X K × Y K B t + N t } k = 1 K P X Y B t + N t t with K samples. The target domain can be more than one, but for the sake of simplicity in explanation, only one is presented here. Here D i t r a i n = ( x j , y j ) X J × Y J B i j = 1 J P X Y B + N i is the training set of the i-th source domain, D i t e s t = { ( x l , y l ) X L × Y L B i + N i } l = 1 L P X Y B + N i is the testing set of the i-th source domain, J is the size of the training set of the source domain, and L is the size of the testing set of the source domain. In the open-vocabulary domain generalization (OVDG) setting, the base class labels of the target domain are a subset of the aggregated base class labels from the m source domains: Y B t Y B 1 + Y B 2 + . . . + Y B m . Additionally, novel class labels in the target domain do not intersect with base class labels, formally expressed as Y N t Y B t = . The main objective of the OVDG setting is to train a model on the m source domain and perform well on both the target domain and source domains.

3.2. Overview

In this section, we describe our proposed framework for OVDG, named FreeMix. Figure 2 shows the overall architecture of our proposed FreeMix, which comprises two main modules: a universal segmentation module and an open-vocabulary recognition module. Furthermore, the universal segmentation module consists of a base segmentation branch (BSB) and an entity segmentation branch (ESB). In the OVDG task, our attention is required to encompass not just the OV aspect but also the DG aspect. When tackling domain generalization challenges, we incorporate a self-supervised base model, CMID, within the BSB branch and efficiently adapt it using mask2former, thereby achieving seamless preservation of pretrained knowledge. Additionally, to tackle the challenge of open vocabulary, in the ESB, we integrate a diverse range of entity masks and design an extractor specifically for extracting the visual features of these masks. This endeavor seeks to enhance the model’s ability to segment and recognize both base and novel classes while maintaining generalizability. Finally, we utilize an open-vocabulary recognition module to classify these masks based on the CLIP [48] model. The classification method entails comparing the semantic logits of the masks with the cosine similarity of text embeddings derived from a contrastive language-image pretrained model.
Next, we first introduce our universal segmentation module (Section 3.3) and then describe the proposed the dataset-aware sample-training tactic on a multi-source domain (Section 3.4) thoroughly.

3.3. Universal Segmentation Module

The proposed universal segmentation module comprises two branches: a base segmentation branch (BSB)and an entity segmentation branch (ESB). Firstly, the diversity and completeness of extracted masks are pivotal factors influencing model performance in semantic segmentation. However, segmentation models trained on base classes often exhibit category bias, as novel classes are typically perceived as background and thus suppressed. Consequently, the mask proposal generator fails to truly achieve class-agnostic mask extraction, tending instead to extract masks corresponding to base classes. Even when unmatched candidate masks do not contribute to loss computation during training, this tendency persists. Moreover, due to disparities in data style between the target and source domains used for training, segmentation models trained on a specific data domain may exhibit sensitivity to stylistic differences in the target domain, leading to suboptimal outcomes. Hence, we endeavor to enhance model robustness and reduce sensitivity to data styles by incorporating dual branches (BSB and ESB). The BSB excels at segmenting the base classes in supervised learning. However, the single-branch BSB extracts very few masks for novel classes, resulting in low recall rates. Therefore, integrating an ESB is necessary to bolster mask extraction capabilities, especially for novel classes, by generating more universal masks.

3.3.1. Base Segmentation Branch (BSB)

To make the model applicable across various data domains, we propose adapting a self-supervised pretrained backbone for remote-sensing semantic segmentation named RS_SSL. This backbone has been trained on large-scale remote-sensing data. Consequently, the architecture of the BSB comprises a self-supervised pretrained backbone network, a pixel decoder, and a Transformer decoder as illustrated in Figure 2. Given an image X R H × W × C , the BSB will output mask proposals M b R N × H × W and their semantic logits E b R N × d , where N is the number of mask proposals, d represents the dimensionality of semantic logits, and H, W and C correspond to image height, image width and number of channels. The backbone network of popular open-vocabulary segmentation models [16,23,24,51] is typically trained in a supervised manner on the ImageNet dataset [78] or trained from scratch. However, the major distinction lies in the fact that our backbone network is initialized using a self-supervised pretrained model CMID [47] and remains frozen during training, with only the decoders being trainable. CMID combines contrastive learning and masked image modeling to learn robust representations in a self-distillation way. Moreover, it is architecture-agnostic, and is compatible with both convolutional neural networks (CNN) and vision transformers (ViT), allowing it to be easily adapted to our BSB. Compared to the mask proposal generator of other models [52], we do not modify the foundational model, allowing us to leverage the powerful generalization capabilities of self-supervised learning in remote-sensing images to extract domain-agnostic universal features. The pixel decoder uses six multi-scale deformable attention Transformer layers applied to feature maps with resolutions of 1 / 8 , 1 / 16 , and 1 / 32 . It also employs a simple upsampling layer to generate the feature map at a resolution of 1/4 as the per-pixel features. One resolution of the multi-scale feature is fed to the Transformer decoder at a time. The features are gradually processed by the Transformer decoder layer with mask attention, starting from low-resolution features and progressing to high-resolution features.

3.3.2. Entity Segmentation Branch (ESB)

To achieve the elimination of sensitivity across disparate data domains, the key is extracting class-agnostic masks and their corresponding visual features from images. In order to achieve this goal, we first employ CropFormer [79], a tiny entity segmentation model, as an entity mask extractor to generate high-quality entity masks. We then establish a visual feature extractor for these masks, leveraging the Transformer architecture.
1.
Entity masks extractor. An entity refers to each semantically coherent region within an image. Entity segmentation is an emerging task that focuses on open-world, class-agnostic, dense image segmentation. It is designed to have superior generalization capabilities for segmenting novel classes [80]. CropFormer [79], trained on a large-scale, high-quality entity segmentation dataset that includes images from various domains, including remote-sensing images, is highly suitable for extracting class-agnostic masks. In entity masks extractor, we use N K-dimensional queries Q R N × K to generate entity masks M m R N × H × W .
2.
Visual feature extractor. The visual feature extractor is based on vision Transformer (ViT) architecture, which consists of 12 Transformer blocks. We denote these blocks as B = { b 1 , b 2 , , b i , , b 12 } . Each block comprises a multihead attention layer followed by two MLP layers with GELU [81] non-linearity. Layer normalization is applied before each layer, and residual connections are added after each layer.
In the first k Transformer blocks, the visual feature extractor initially encodes the entire image X R H × W × C to obtain a representation E R ( h w + 1 ) × d . Here, h and w represent the height and width of the attention map in the ViT, the additional 1 corresponds to the semantic logits for the entire image, and d denotes the dimensionality of the features. In the remaining 12 k Transformer blocks, to extract the semantic logits E m R N × d for entity masks M m R N × H × W , we assign independent classification queries to each entity mask by repeating the semantic logits of the entire image N times. We then utilize the entity masks as the attention bias (A) in the multihead attention mechanism:
E m b + 1 = softmax Q E m b K E b T d + A V E b
where b indicates the block number, K E b = W k E b and V E b = W v E b are the key and value embeddings of the representation, and Q E m b = W q E m b is the query embedding of entity masks. Here, W q , W k , and W v are the weights of the query, key, and value embedding layer, respectively. To ensure the updating of semantic logits E m b for entity masks, we exclusively consider the representation corresponding to the entity mask and its own semantic logits without referencing the semantic logits of other entity masks. We construct a self-attention bias matrix A R N × N + h w as follows:
A ( x , y ) = 0 , if M ¯ ( x , y ) = 1 , if M ¯ ( x , y ) = 0
M ¯ = concat [ I ( N , N ) , f ( M m ) ]
where ( x , y ) is the feature location, I ( N , N ) denotes the ( N , N ) identity matrix, f ( M m ) represents resizing M m to ( h , w ) and then flattening it, and M ¯ { 0 , 1 } N × ( N + h w ) is the binarized output (thresholded at 0.5) of the flattened mask.
In the open-vocabulary recognition module, given the base classes Y B during training, class prompts V y are generated using the template: {“semantic segmentation” + learnable vectors + c }, where c represents the filled-in class names. Then the text prompts are then embedded using the pretrained CLIP text encoder φ :
E t = φ ( V y ) , y Y B
To predict the class of masks P R C × N , we compare the similarity between the semantic logits E = concat [ ψ ( E b ) , ψ ( E m ) ] of mask groups M = concat [ M b , M m ] and the text embedding E t R C × r . Here ψ ( ) represents the normalization process, C is the number of classes, and r is the dimension of text embeddings. Finally, we compute the semantic-segmentation map S = M × P T . Here, P T denotes the transpose of matrix P.

3.4. Train Tactics: Dataset-Aware Sampling

In order to effectively improve the generalization ability of the model, it is necessary to train on multi-source datasets D = { D 1 , D 2 , . . . , D i , . . . } which exhibit different styles [82,83]. Each dataset has its own distinct label space Y = { Y 1 , Y 2 , . . . Y i , . . . } . A straightforward approach to train on multiple datasets is to combine all annotations from these datasets into a larger dataset D = D 1 D 2 . . . , and then relabel and merge their label spaces into Y = Y 1 Y 2 . . . . The model is optimized with the same loss on a larger dataset. Fortunately, the number of classes for labeling remote-sensing semantic-segmentation data is relatively small, making manual label mapping feasible. However, significant variations exist in dataset sizes: GID5 [84] contains 3 × more images than Potsdam [1], DeepGlobe [85] is 8 × larger than Potsdam [1], and URUR [86] is 79 × larger. This imbalance in class distributions and dataset sizes virtually ensures that a mere concatenation of datasets will not work.
To effectively utilize multi-source datasets, we propose a simple yet efficient training strategy: Dataset-Aware Sampling (DAS). Re-sampling is a widely used strategy in addressing class-imbalanced or long-tail learning scenarios [87]. Inspired by this, we extend the concept of sampling to the data domain level, introducing dataset-aware sampling. Specifically, we uniformly sample instances from each dataset within every mini-batch. For each dataset, we then compute the loss individually for its samples. Finally, we aggregate these per-dataset losses by averaging them before performing backpropagation:
L overall = 1 T i L i
where T is the number of datasets, and L i represents the loss from the i-th dataset within a mini-batch. During training, L i comprises both mask losses and classification losses. Specifically, the mask losses include Dice loss L d i c and binary cross-entropy loss L b c e , while the classification loss L c l s uses cross-entropy loss.
L i = λ 1 L dic + λ 2 L bce + λ 3 L cls
where λ 1 , λ 2 , and λ 3 are set to 20, 5, and 2, respectively, in our setting. In the proposed DAS, we do not need to design a separate head for each dataset because we are addressing a single task across multiple datasets [88]. Consequently, there is no need to design distinct loss functions for each dataset or employ complex algorithms to search for the optimal weights of different losses. The experimental results in Section 4 demonstrate that our method, which employs a single model and a unified type of loss function, is both simple and effective for the OVDG task.

4. Materials and Experimental Settings

4.1. Experimental Datasets and Processing

We evaluate the proposed methods on four well-known remote-sensing datasets: (1) Potsdam [1] dataset is a widely used benchmark in remote sensing, comprising 38 images of size 6000 × 6000 pixels. It includes six classes, with ’background’ as one of them. (2) GID5 [84] contains 150 pixel-level annotated GF-2 images of 6800 × 7200, sampled from various cities in China. It has a total of 5 classes. (3) DeepGlobe [85] is a large-scale land-cover dataset that contains 803 images (2448 × 2448 pixels). It includes seven classes of landscape regions, including “unknown” regions. (4) URUR [86] features a substantial number of high-resolution images (3008 images of size 5120 × 5120 pixels) covering a wide range of complex scenes from 63 cities. In the open-vocabulary setting, we first remove the “clutter” class from Potsdam, the “background” class from DeepGlobe, and the “other” category from URUR by setting the corresponding labels to 255, marking them as invalid classes. Next, we merge similar classes across datasets. For instance, categories such as “built up”, “building”, and “urban land” are unified and renamed as “building”. Finally, we randomly partition the classes into base and novel classes across all datasets. We combine the four datasets into a single dataset, referred to as GPDU, comprising a total of 12 classes. These include seven base classes (building, farmland, forest land, impervious surface, car, range land, greenhouse) and five novel classes (meadow, water, tree, bare land, road). The details of class partitioning and the class mapping relationships can be found in Table 1. The class frequency distributions on the training and testing sets of the four datasets are illustrated in Figure 3.

4.2. Implementation Details

In the open-vocabulary recognition module, we adopt the pretrained vision–language model CLIP, employing the ViT-B backbone as the text encoder. For the base segmentation branch, we use ResNet50 [89] as the backbone. In the entity segmentation branch (ESB), we employ CropFormer [79] with a Swin-Tiny backbone for entity mask extraction. To stabilize training and leverage pretrained knowledge, we freeze several components:
  • The text and image encoders of CLIP to preserve the learned multimodal representations;
  • The entity mask extractor in ESB to maintain the robustness of the extracted masks;
  • The backbone of the base segmentation branch after initializing it with self-supervised pretraining weights.
During training, we focus on fine-tuning the decoders of BSB and the visual feature extractor of ESB. We conduct training on an NVIDIA GTX 3090 GPU using a mini-batch size of 2 images. Please note that we use N = 100 queries in both branches and set k = 8 in the visual feature extractor. For a fair comparison, we adopt the same training settings as FreeSeg [24]: The optimizer AdamW is adopted with an initial learning rate of 10 4 and weight decay of 10 4 . To avoid overfitting on the training set, the learning rate of the image encoder is multiplied by a factor of λ = 0.01 . There are, in total, 40k training iterations. Each dataset is separately divided into training and testing sets. During training, input images are cropped to 512 × 512 pixels.

4.3. Evaluation Metrics

We evaluate our model and the baselines with mean Intersection over Union (mIoU), frequency-weighted IoU (fwIoU), pixel Accuracy (pACC), and mean pixel Accuracy (mAcc). First, mIoU is a widely used metric for evaluating semantic-segmentation models. It calculates the average Intersection over Union (IoU) across all classes, where IoU measures the overlap between predicted and ground-truth segmentation masks. This provides a comprehensive measure of segmentation accuracy by averaging the performance for each class. fwIoU adjusts for class imbalance by assigning more weight to frequently occurring classes during evaluation. pACC measures the percentage of correctly classified pixels in the entire image, while mACC calculates the average proportion of correctly classified pixels across all classes.
m I o U = 1 C c = 1 C T P c T P c + F P c + F N c
f w I o U = 1 c = 1 C ( T P c + F P c + F N c ) c = 1 C T P c T P c + F P c + F N c
p A C C = c = 1 C ( T P c + T N c ) c = 1 C ( T P c + F P c + F N c + T N c )
m A C C = 1 C c = 1 C T P c T P c + F N c
where C is the number of classes, and for each class c: T P c is the number of true positive pixels. F P c is the number of false positive pixels. F N c is the number of false negative pixels. T N c is the number of true negative pixels. Furthermore, we compute mIoU and pACC separately on the base and novel classes, denoted as follows: m I o U s : Mean Intersection over Union for base classes, m I o U u s : Mean Intersection over Union for novel classes, p A C C s : Pixel Accuracy for base classes, p A C C u s : Pixel Accuracy for novel classes.

5. Results and Discussion

5.1. Comparison with SOTA Methods

5.1.1. Results of Open-Vocabulary Semantic Segmentation

Our proposed FreeMix is compared against several existing open-vocabulary methods, including ZSSeg [51], ZegFormer [16], MaskCLIP [19], SAN [52], OVSeg [23], FC-CLIP [90] and FreeSeg [24]. In the open-vocabulary setting, the model is trained using only base classes and their corresponding masks. During testing, both base classes and novel classes are evaluated to assess the model’s generalization capability. Table 2 shows the results on the Potsdam dataset. For a fair comparison, all models use ResNet50 as the image encoder, with the exception of OVSeg, which uses ResNet101, and FC-CLIP, which employs a large version of ConvNeXt. The proposed FreeMix significantly outperforms other methods across multiple evaluation metrics. Specifically, it achieves the highest mean Intersection over Union (mIoU) of 63.44% and the highest mean pixel Accuracy (mACC) of 73.87%, both of which are the best among all compared methods. Notably, FreeMix excels in segmenting base classes, achieving an mIoU of 86.46%. This high score underscores its proficiency in accurately delineating familiar object categories. Additionally, for novel classes, FreeMix attains an mIoU of 28.92%. Although this is lower than the performance in base classes, it still represents a commendable achievement considering the inherent challenges of recognizing previously unseen categories. These results collectively demonstrate that FreeMix not only maintains high accuracy for known classes but also generalizes well to novel classes, highlighting its robustness and adaptability in open-vocabulary semantic-segmentation tasks.
Compared to the second-best method, ZSSeg, our FreeMix achieves significant improvements across all metrics. This is primarily because ZSSeg relies on existing mask proposal networks without any optimization, leading to lower-quality mask generation. In contrast, our approach enhances mask quality through tailored optimizations, including the use of a dual-branch architecture to extract diverse and complementary mask representations. Notably, as shown in Table 2, our FreeMix outperforms FC-CLIP, which uses a larger image encoder. While FC-CLIP achieves the highest pixel Accuracy (pACC) on base classes, this is due to its frozen CNN-based CLIP backbone and supervised fine-tuning specifically for base classes, causing it to be more biased toward these categories. Our FreeMix incorporates a frozen self-supervised backbone in the base segmentation branch, contributing to strong performance on both base and novel classes. This design, facilitated by the two-branch architecture of FreeMix, helps maintain robust segmentation capabilities across all categories. In terms of novel classes, FreeMix demonstrates substantial improvements. Specifically, It gains an 11.11% IoU on the tree class and a 46.73% IoU on the meadow class. These improvements are particularly noteworthy because certain models, such as ZSSeg and ZegFormer, struggle to distinguish between similar classes, like trees and meadows, in remote-sensing images. The entity segmentation branch (ESB) of FreeMix plays a crucial role by generating universal masks and extracting corresponding visual features, thereby enhancing the model’s ability to generalize to novel classes. Overall, the proposed FreeMix can generalize well to novel classes while maintaining strong performance on base classes. This balanced approach ensures that FreeMix has the best overall performance, as demonstrated by its superior results across multiple evaluation metrics.
In addition, The qualitative results on the Potsdam testing set are depicted in Figure 4. We compare the segmentation results of different methods on the Potsdam dataset, where meadow and tree are novel classes (represented in gray and blue-purple, respectively). For base classes such as impervious surfaces, buildings, and cars, FreeMix produces more precise and boundary-aligned segmentations compared to other methods. Notably, when segmenting novel classes, models often struggle due to confusion between similar categories. However, as shown in the top row and third row of Figure 4, our segmentation module excels at distinguishing between challenging classes such as “meadow” and “tree,” outperforming any off-the-shelf model. In the second row, while FreeSeg, ZegFormer, and ZZSeg manage to segment the meadow class to some extent, they struggle with the tree class, producing fragmented or incorrect results. This demonstrates the strong open-vocabulary segmentation ability of our model.

5.1.2. Results of OVDG

We also verify the generalization ability of our FreeMix across different datasets. Specifically, the model is trained on one dataset and directly evaluated on another dataset without fine-tuning. This setup presents a challenging task where the model must handle both novel classes and address the domain gap between different datasets. We report the mean Intersection over Union (mIoU) for both base and novel classes. As shown in Table 3, FreeMix consistently achieves the highest average mIoU and mACC across both the source domain and the three target domains, regardless of the training dataset. Specifically, when trained on the GID5 dataset, FreeMix achieved the highest average mIoU of 25.46%, and when trained on the URUR dataset, it attained the highest average mAcc of 39.98%. FreeMix demonstrates strong single-source domain generalization capabilities, with robust performance across different source domains. While the overall performance remains consistently high, we observe minor variations depending on the specific training dataset used. These variations likely stem from the differing characteristics of each domain, which can impact the ease of generalization. Overall, the GID5 dataset proves to be a particularly strong choice for achieving better generalization performance. The consistently high performance across multiple domains underscores the robustness and adaptability of FreeMix in open-vocabulary domain generalization tasks.
Notably, when trained on the Potsdam dataset, FreeMix generalizes well to GID5 but underperforms when tested on DeepGlobe and URUR. A similar trend can be observed in other models. This phenomenon may be attributed to the relatively smaller size of the Potsdam dataset compared to others, which renders it more susceptible to overfitting given an equivalent number of training iterations. In contrast, the GID5 dataset—closer in size to Potsdam—demonstrates superior performance relative to both DeepGlobe and URUR. However, this also presents a significant limitation. Training on datasets with a narrow range of classes, such as Potsdam—which predominantly contains urban categories—can lead to suboptimal performance when FreeMix is tested on datasets like DeepGlobe. DeepGlobe includes natural and semi-natural classes that are not present in Potsdam. This mismatch in class distributions underscores the critical importance of dataset composition in domain generalization tasks. Additionally, our FreeMix not only enhances generalization performance but also excels in open-vocabulary segmentation capabilities. When trained on GID5 and tested on another dataset, it attains m I o U u s of 15.83% on Potsdam, 12.38% on DeepGlobe, and 9.79% on URUR. Moreover, m I o U u s reaches 20.26% on GID5 and 6.49% on URUR when FreeMix only trained on the DeepGlobe dataset, which outperforms FreeSeg with +4.30% and +3.78%, respectively. These results demonstrate that our proposed FreeMix not only performs excellently within the domain but also exhibits strong generalization capabilities across other data domains. When trained and tested on the single-source domain URUR, FreeMix surpasses FreeSeg, achieving mIoU and mACC that are 8.73% and 7.74% higher, respectively. Moreover, FreeMix generalizes well to other target domain datasets, except for GID5. Through analysis, we found that the IoU for the easily confused classes, farmland from the base classes and meadow from the novel classes, is lower in FreeMix compared to FreeSeg. Therefore, enhancing the model’s ability to accurately recognize similar classes in cross-domain scenarios presents a more challenging and promising direction for future work. For more qualitative results on the GID5 dataset, see Appendix A, shown in Figure A1.
MaskCLIP shares some similarities with our ESB, as both methods utilize a class-agnostic mask proposal network and extract features from a pretrained CLIP ViT model. However, FreeMix’s design incorporates several enhancements that lead to higher-quality masks and richer semantic features: the Entity Mask Extractor and BSB generate more accurate masks compared to MaskCLIP. The ESB in FreeMix includes an independent visual feature extractor, which contributes to the richness of the extracted features. This analysis reveals that both mask quality and the semantic richness of the features play a critical role in determining the final segmentation performance. The approach most akin to ours BSB is FC-CLIP [24], which involves directly substituting the backbone network of the mask generator with CLIP’s image encoder and subsequently freezing it during training. This strategy enables maximal retention of the model’s original generalization capabilities while concurrently reducing the memory requirements during training. By leveraging CLIP’s pretrained image encoder, FC-CLIP capitalizes on the robust feature representations learned from large-scale datasets, which are inherently transferable across diverse visual tasks. Freezing the encoder during training not only preserves these generalized features but also significantly reduces the computational overhead, as it eliminates the need to update the weights of a large backbone network. Despite using a larger image encoder, FC-CLIP is outperformed by FreeMix across all datasets. Notably, FreeMix achieves these results without employing multi-scale test-time augmentation, which other models use. Intuitively, the improved modules in FreeMix can more effectively generate universal masks for both base and novel classes, thereby enhancing its segmentation generalization ability on both class and domain levels. These findings underscore the effectiveness of the proposed method in open domains.

5.2. Experiments on Multi-Source Domain

To evaluate the performance of FreeMix across multiple source domains, we train it on a combined dataset named GPDU. This setup allows us to compare FreeMix with other state-of-the-art models under fair conditions, where all models are trained for 40K iterations. It is important to note that existing models are typically limited to training on a single dataset at a time. To enable a comprehensive comparison, we created the GPDU dataset by performing label mapping for the training of these models. In contrast, FreeMix can be trained on multiple datasets simultaneously due to the proposed dataset-aware sampling (DAS). DAS enables the model to dynamically adapt to the label spaces and distributions of different datasets without requiring manual alignment or unification of labels. This flexibility allows FreeMix to leverage diverse data sources effectively, enhancing its generalization capabilities.
As depicted in Table 4, FreeMix consistently outperforms other models across all datasets. The implementation of the universal segmentation module with a dual-branch architecture and dataset-aware sampling (DAS) has helped FreeMix set a new benchmark, achieving the highest average mIoU. In comparison, FreeSeg and SAN exhibit significantly lower performance, underscoring the superiority of FreeMix in semantic-segmentation tasks. Specifically, the use of DAS contributes to an overall improvement in FreeMix’s average mIoU. However, this enhancement reveals a trade-off between base and novel class performance. Significant improvements are observed on datasets like Potsdam and URUR. There is a noticeable decrease, particularly on GID5 and DeepGlobe. This trade-off highlights the need for further optimization to balance the performance between base and novel classes when applying DAS. Addressing this issue could lead to more consistent and robust performance across all classes and datasets.

5.3. Ablation Experiments

In this section, we conduct ablation studies on our FreeMix using the joint GPDU dataset. All experiments utilize a ResNet50 backbone and the CLIP-B/16 VLM, with FreeSeg serving as our baseline for comparison. Table 5 summarizes the effectiveness of three key components: the proposed initialization method with self-supervised learning in remote sensing (RS_SSL), the entity segmentation branch (ESB) and the dataset-aware sampling (DAS) training tactic. These results clearly indicate that each proposed component consistently enhances the overall performance of FreeMix, highlighting their individual and combined contributions to improved semantic-segmentation outcomes. Using RS_SSL initialization provides a substantial improvement over the baseline, indicating the value of leveraging self-supervised learning specifically tailored for remote-sensing tasks. The introduction of ESB yields consistent performance boosts across all datasets, suggesting its critical role in capturing more accurate and detailed entity representations. DAS not only elevates mIoU and mACC but also demonstrates its ability to balance performance across diverse datasets, thereby ensuring robust generalization.
When using supervised pretraining on ImageNet1K to initialize the image encoder and fine-tune it on the GPDU dataset, we observe a degradation in performance. However, with our RS_SSL initialization method and freezing the image encoder during training, the model can extract domain-agnostic universal features, thereby accelerating decoder optimization and improving learning efficiency. Furthermore, incorporating the ESB branch to construct a dual-branch network further enhances performance. The ESB effectively extracts entity masks and captures accurate semantic features for classification, contributing significantly to better segmentation results. Additionally, employing DAS improves overall performance, especially on datasets with relatively small amounts of data. For instance, on the Potsdam dataset, mIoU increased from 19.98% to 47.03%, demonstrating substantial improvement. Even on larger datasets like URUR, there was an increase in mIoU from 29.30% to 35.72%. The above findings indicate that our proposed methods—RS_SSL initialization, ESB, and DAS—are highly effective in enhancing the performance of FreeMix. The qualitative results of FreeMix are visualized in Figure 5, providing a clear demonstration of its superior performance in semantic-segmentation tasks.

5.4. Additional Experimental Results

5.4.1. Performance on Different Image Encoders of BSB

To further demonstrate the adaptability of the proposed RS_SSL adaptation strategy, we conduct experiments on both CNN and ViT architectures. In FreeMix, the backbone of BSB is from CMID [47], which is pretrained in a self-supervised manner on the remote-sensing dataset MillionAID. In contrast, other methods are either trained in a supervised manner on the ImageNet1K dataset or train the backbone network from scratch. Furthermore, we report comparative results under different training tactics: random sampling and the proposed DAS. According to the Table 6, using ResNet50 as the backbone and employing RS_SSL initialization to BSB, our FreeMix achieves average mIoU scores of 34.64% and 40.25% when using random sampling and DAS training tactic, respectively. It should be noticed that, without introducing RS_SSL initialization to BSB, the performance will drop significantly to average mIoU scores of 29.17% (random sampling) and 23.71% (DAS). Moreover, applying RS_SSL initialization to the Transformer-based backbone network Swin-B, the model gains an average mACC of 61.00% and 64.29% when using random sampling and DAS, respectively. Compared to training BSB from scratch, RS_SSL initialization improved average mACC across the four datasets by 1.27% (random sampling) and 7.08% (DAS). These results indicate that RS_SSL initialization is effective for both CNN and ViT architectures. Meanwhile, it demonstrates that FreeMix is architecture-agnostic, allowing further performance improvement by leveraging superior pretrained models of both convolutional neural networks and vision transformers.

5.4.2. Performance on Scaling Model Size of ESB

To evaluate the impact of scaling the model size of the ESB, we train 6 FreeMix models, sweeping over the backbone of ESB (Swin-Tiny, Swin-Large, Hornet-Large) and training tactics (random sampling, DAS). The main results are illustrated in Table 7. Swin-Tiny (Swin-T) consistently shows strong performance across the datasets, particularly when paired with the DAS training tactic. It achieves the highest average mIoU (40.25%) and mACC (73.05%). This confirms that smaller backbones like Swin-T perform more efficiently in terms of both segmentation accuracy and generalization across domains. Swin-Large (Swin-L), while larger in size, shows a drop in where its average mIoU falls to 32.51% and mACC to 49.67%. This performance drop may be attributed to Swin-L requiring more data to train effectively, as larger models often need more data to fully capture patterns and avoid overfitting. Hornet-Large (Hornet-L) performs well in certain metrics, such as achieving the highest mACC of 67.61% with random sampling and 66.17% on GID5 with DAS, but it underperforms in terms of average mIoU, especially on DeepGlobe and URUR. The mixed performance indicates that while Hornet-L excels in certain conditions, it may not generalize as effectively across all domains. In conclusion, these results underscore that smaller backbones like Swin-T not only maintain better overall performance but also generalize more effectively across datasets, particularly when using the DAS training tactic. Conversely, larger models like Swin-L and Hornet-L show diminishing returns, especially in terms of mIoU, suggesting that adding a lightweight ESB branch can lead to greater efficiency without sacrificing accuracy. This conclusion is specific to the results obtained in our study and does not necessarily generalize to all CNN and Transformer networks.

5.4.3. Comparison of the Extracted Proposal Masks

We visualize and compare the proposal masks extracted by FreeSeg and our FreeMix, as shown in Figure 6. FreeSeg processes 100 queries to generate 100 proposal masks. For visualization purposes, we randomly select 10 of these masks. Moreover, FreeMix generates 100 proposal masks for each branch (BSB and ESB). We randomly select 10 masks from each branch for visualization and comparison, indicated in yellow (BSB) and purple (ESB) in Figure 6. Based on the results, it is evident that the proposal masks generated by FreeSeg exhibit low distinguishability and low predicted confidence, as indicated by their gray appearance. In contrast, the masks decoded by FreeMix’s BSB, with the assistance of RS_SSL, exhibit higher confidence and greater distinguishability. Additionally, the ESB of FreeMix offers a richer set of entity masks, capturing more detailed and accurate representations of entities within the images. By combining these high-quality universal masks from both branches, FreeMix demonstrates superior performance on OVDG tasks compared to other models.

6. Conclusions

In this work, we introduce a new setting for remote-sensing image semantic segmentation called open-vocabulary domain generalization (OVDG). This novel setting holds great potential in the remote-sensing community by addressing the challenge of generalizing models to unseen domains and classes without retraining. To tackle this challenging problem, we propose an effective framework to train a robust model. Firstly, the proposed universal segmentation module has a dual branch: base segmentation branch (BSE) and entity segmentation branch (ESB). Moreover, the remote-sensing self-supervised learning (RS_SSL) initialization and adaptation method is introduced to extract domain-agnostic visual features for decoding masks and semantic logits. Additionally, ESB is proposed to generate entity masks for enhancing the segmentation and recognition of both base and novel classes. Furthermore, a dataset-aware sampling (DAS) training tactic is designed for multi-source domain learning, aiming to enhance the overall performance of the model. Extensive experiments demonstrate the effectiveness of our proposed universal segmentation module, incorporating RS_SSL initialization, BSB, and ESB, as well as the DAS training tactic. Our FreeMix achieves state-of-the-art results on open-vocabulary benchmarks and OVDG tasks. Although good semantic-segmentation results have been achieved on OVDG, FreeMix relies on class names provided during testing. Future work will focus on designing a more effective training strategy for multi-data set learning and extending the FreeMix framework by incorporating multimodal large language models to generate class predictions and localization.

Author Contributions

Conceptualization, R.Z. and J.W.; Validation, J.W., J.S., Z.Z. and Z.L.; Investigation, J.S. and Z.Z.; Resources, R.Z.; Data curation, R.Z., J.W., and Z.Z.; Writing—original draft, J.W.; Writing—review and editing, R.Z. and J.W.; Methodology, R.Z. and J.W.; Supervision, R.Z.; Project administration, R.Z. and J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the Key Laboratory Funding of China, grant number 2022-JCJQ-LA-001-080, and in part by the National Key Research and Development Program of China, grant number 2018YFC0823002.

Data Availability Statement

See the source code for more detailed experimental settings: https://github.com/GoldfishFive/FreeMix (accessed on 30 December 2024).

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

As shown in Figure A1, the qualitative results on the GID5 dataset compare the performance of FreeMix, FreeSeg, ZegFormer, and ZZSeg for semantic segmentation, where meadow and water are novel classes. FreeMix demonstrates superior performance in segmenting the novel class water, producing more accurate and coherent masks compared to other methods, which often miss large portions or misclassify water. Additionally, FreeMix excels in identifying base classes like buildings, even correctly detecting buildings missing in the ground-truth annotations. However, its segmentation masks can appear coarse, indicating a trade-off between class recognition accuracy and mask refinement. In contrast, ZegFormer and ZZSeg struggle to generate diverse and accurate masks, with low recognition accuracy for both base and novel classes, while FreeSeg shows moderate performance but still underperforms in novel classes. These results highlight FreeMix’s advancements in open-vocabulary domain generalization while revealing areas for further improvement in mask granularity.
Figure A1. Qualitative results of FreeMix, FreeSeg, ZegFormer, and ZZSeg on semantic-segmentation GID5 dataset, where meadow and water are novel classes.
Figure A1. Qualitative results of FreeMix, FreeSeg, ZegFormer, and ZZSeg on semantic-segmentation GID5 dataset, where meadow and water are novel classes.
Remotesensing 17 01357 g0a1

References

  1. Rottensteiner, F.; Sohn, G.; Gerke, M.; Wegner, J.D. ISPRS Semantic Labeling Contest; ISPRS: Leopoldshöhe, Germany, 2014; Volume 1, p. 4. [Google Scholar]
  2. Hong, J.; Li, W.; Han, J.; Zheng, J.; Fang, P.; Harandi, M.; Petersson, L. Goss: Towards generalized open-set semantic segmentation. Vis. Comput. 2024, 40, 2391–2404. [Google Scholar] [CrossRef]
  3. Nunes, I.; Laranjeira, C.; Oliveira, H.; dos Santos, J.A. A systematic review on open-set segmentation. Comput. Graph. 2023, 115, 296–308. [Google Scholar] [CrossRef]
  4. Nunes, I.M.; Poggi, M.; Oliveira, H.; Pereira, M.B.; Dos Santos, J.A. Deep open-set segmentation in visual learning. In Proceedings of the 2022 35th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Natal, Brazil, 24–27 October 2022; Volume 1, pp. 314–319. [Google Scholar]
  5. Joseph, K.; Khan, S.; Khan, F.S.; Balasubramanian, V.N. Towards open world object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 5830–5840. [Google Scholar]
  6. Bendale, A.; Boult, T. Towards open world recognition. In Proceedings of the IEEE Conference on Computer vision And Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1893–1902. [Google Scholar]
  7. Yang, J.; Zhou, K.; Li, Y.; Liu, Z. Generalized out-of-distribution detection: A survey. arXiv 2021, arXiv:2110.11334. [Google Scholar] [CrossRef]
  8. Liu, J.; Shen, Z.; He, Y.; Zhang, X.; Xu, R.; Yu, H.; Cui, P. Towards out-of-distribution generalization: A survey. arXiv 2021, arXiv:2108.13624. [Google Scholar]
  9. Zhang, H.; Ding, H. Prototypical matching and open set rejection for zero-shot semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 6974–6983. [Google Scholar]
  10. He, S.; Ding, H.; Jiang, W. Semantic-promoted debiasing and background disambiguation for zero-shot instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 19498–19507. [Google Scholar]
  11. Baek, D.; Oh, Y.; Ham, B. Exploiting a joint embedding space for generalized zero-shot semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 9536–9545. [Google Scholar]
  12. Gu, Z.; Zhou, S.; Niu, L.; Zhao, Z.; Zhang, L. Context-aware feature generation for zero-shot semantic segmentation. In Proceedings of the 28th ACM International Conference on Multimedia, Washington, DC, USA, 12–16 October 2020; pp. 1921–1929. [Google Scholar]
  13. Zheng, Y.; Wu, J.; Qin, Y.; Zhang, F.; Cui, L. Zero-shot instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 2593–2602. [Google Scholar]
  14. He, S.; Ding, H.; Jiang, W. Primitive generation and semantic-related alignment for universal zero-shot segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 11238–11247. [Google Scholar]
  15. Bucher, M.; Vu, T.H.; Cord, M.; Pérez, P. Zero-shot semantic segmentation. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  16. Ding, J.; Xue, N.; Xia, G.S.; Dai, D. Decoupling zero-shot semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11583–11592. [Google Scholar]
  17. Ma, C.; Yang, Y.; Wang, Y.; Zhang, Y.; Xie, W. Open-vocabulary semantic segmentation with frozen vision-language models. arXiv 2022, arXiv:2210.15138. [Google Scholar]
  18. Chen, X.; Li, S.; Lim, S.N.; Torralba, A.; Zhao, H. Open-vocabulary panoptic segmentation with embedding modulation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 1141–1150. [Google Scholar]
  19. Ding, Z.; Wang, J.; Tu, Z. Open-Vocabulary Panoptic Segmentation MaskCLIP. arXiv 2022, arXiv:2208.08984. [Google Scholar]
  20. Ghiasi, G.; Gu, X.; Cui, Y.; Lin, T.Y. Scaling open-vocabulary image segmentation with image-level labels. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 540–557. [Google Scholar]
  21. Zhou, C.; Loy, C.C.; Dai, B. Extract free dense labels from clip. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 696–712. [Google Scholar]
  22. Huynh, D.; Kuen, J.; Lin, Z.; Gu, J.; Elhamifar, E. Open-vocabulary instance segmentation via robust cross-modal pseudo-labeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 7020–7031. [Google Scholar]
  23. Liang, F.; Wu, B.; Dai, X.; Li, K.; Zhao, Y.; Zhang, H.; Zhang, P.; Vajda, P.; Marculescu, D. Open-vocabulary semantic segmentation with mask-adapted clip. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7061–7070. [Google Scholar]
  24. Qin, J.; Wu, J.; Yan, P.; Li, M.; Yuxi, R.; Xiao, X.; Wang, Y.; Wang, R.; Wen, S.; Pan, X.; et al. Freeseg: Unified, universal and open-vocabulary image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 19446–19455. [Google Scholar]
  25. Ren, S.; Zhang, A.; Zhu, Y.; Zhang, S.; Zheng, S.; Li, M.; Smola, A.J.; Sun, X. Prompt pre-training with twenty-thousand classes for open-vocabulary visual recognition. In Proceedings of the Advances in Neural Information Processing Systems 37 (NeurIPS 2014), Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
  26. Zhang, H.; Li, F.; Zou, X.; Liu, S.; Li, C.; Yang, J.; Zhang, L. A simple framework for open-vocabulary segmentation and detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 1020–1031. [Google Scholar]
  27. Wu, J.; Li, X.; Xu, S.; Yuan, H.; Ding, H.; Yang, Y.; Li, X.; Zhang, J.; Tong, Y.; Jiang, X.; et al. Towards open vocabulary learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 5092–5113. [Google Scholar] [CrossRef]
  28. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  29. Ding, H.; Cohen, S.; Price, B.; Jiang, X. Phraseclick: Toward achieving flexible interactive segmentation by phrase and click. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28. 2020, Proceedings, Part III 16; Springer: Cham, Switzerland, 2020; pp. 417–435. [Google Scholar]
  30. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems 26 (NIPS 2013), Lake Tahoe, NV, USA, 5–8 December 2013. [Google Scholar]
  31. Zhu, C.; Chen, L. A survey on open-vocabulary detection and segmentation: Past, present, and future. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 8954–8975. [Google Scholar] [CrossRef]
  32. Liang, C.; Li, W.; Dong, Y.; Fu, W. Single Domain Generalization Method for Remote Sensing Image Segmentation via Category Consistency on Domain Randomization. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  33. Wang, M.; Liu, J.; Luo, G.; Wang, S.; Wang, W.; Lan, L.; Wang, Y.; Nie, F. Smooth-Guided Implicit Data Augmentation for Domain Generalization. In IEEE Transactions on Neural Networks and Learning Systems; IEEE: New York, NY, USA, 2024; pp. 1–12. [Google Scholar] [CrossRef] [PubMed]
  34. You, K.; Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Universal domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2720–2729. [Google Scholar]
  35. Saito, K.; Kim, D.; Sclaroff, S.; Saenko, K. Universal domain adaptation through self supervision. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Virtual, 6–12 December 2020; pp. 16282–16292. [Google Scholar]
  36. Kundu, J.N.; Venkat, N.; Babu, R.V. Universal source-free domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 4544–4553. [Google Scholar]
  37. Li, R.; Zheng, S.; Duan, C.; Su, J.; Zhang, C. Multistage attention ResU-Net for semantic segmentation of fine-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  38. Niu, X.; Zeng, Q.; Luo, X.; Chen, L. FCAU-Net for the Semantic Segmentation of Fine-Resolution Remotely Sensed Images. Remote Sens. 2022, 14, 215. [Google Scholar] [CrossRef]
  39. Wang, L.; Li, R.; Wang, D.; Duan, C.; Wang, T.; Meng, X. Transformer meets convolution: A bilateral awareness network for semantic segmentation of very fine resolution urban scene images. Remote Sens. 2021, 13, 3065. [Google Scholar] [CrossRef]
  40. He, X.; Zhou, Y.; Zhao, J.; Zhang, D.; Yao, R.; Xue, Y. Swin transformer embedding UNet for remote sensing image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  41. Gui, R.; Xu, X.; Wang, L.; Yang, R.; Pu, F. A generalized zero-shot learning framework for PolSAR land cover classification. Remote Sens. 2018, 10, 1307. [Google Scholar] [CrossRef]
  42. Jia, X.; Khandelwal, A.; Nayak, G.; Gerber, J.; Carlson, K.; West, P.; Kumar, V. Incremental dual-memory lstm in land cover prediction. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 867–876. [Google Scholar]
  43. Li, A.; Lu, Z.; Wang, L.; Xiang, T.; Wen, J.R. Zero-shot scene classification for high spatial resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4157–4167. [Google Scholar] [CrossRef]
  44. Sumbul, G.; Cinbis, R.G.; Aksoy, S. Fine-grained object recognition and zero-shot learning in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2017, 56, 770–779. [Google Scholar] [CrossRef]
  45. Tsai, Y.H.; Hung, W.C.; Schulter, S.; Sohn, K.; Yang, M.H.; Chandraker, M. Learning to adapt structured output space for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7472–7481. [Google Scholar]
  46. Zheng, Z.; Yang, Y. Rectifying pseudo label learning via uncertainty estimation for domain adaptive semantic segmentation. Int. J. Comput. Vis. 2021, 129, 1106–1120. [Google Scholar] [CrossRef]
  47. Muhtar, D.; Zhang, X.; Xiao, P.; Li, Z.; Gu, F. CMID: A Unified Self-Supervised Learning Framework for Remote Sensing Image Understanding. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–17. [Google Scholar] [CrossRef]
  48. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  49. Jia, C.; Yang, Y.; Xia, Y.; Chen, Y.T.; Parekh, Z.; Pham, H.; Le, Q.; Sung, Y.H.; Li, Z.; Duerig, T. Scaling up visual and vision-language representation learning with noisy text supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 4904–4916. [Google Scholar]
  50. Chen, Y.; Bruzzone, L. Toward Open-World Semantic Segmentation of Remote Sensing Images. In Proceedings of the IGARSS 2023–2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 5045–5048. [Google Scholar]
  51. Xu, M.; Zhang, Z.; Wei, F.; Lin, Y.; Cao, Y.; Hu, H.; Bai, X. A simple baseline for open-vocabulary semantic segmentation with pre-trained vision-language model. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 736–753. [Google Scholar]
  52. Xu, M.; Zhang, Z.; Wei, F.; Hu, H.; Bai, X. Side adapter network for open-vocabulary semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 2945–2954. [Google Scholar]
  53. Zhang, Z.; Zhao, T.; Guo, Y.; Yin, J. Rs5m: A large scale vision-language dataset for remote sensing vision-language foundation model. arXiv 2023, arXiv:2306.11300. [Google Scholar]
  54. Wang, Z.; Prabha, R.; Huang, T.; Wu, J.; Rajagopal, R. Skyscript: A large and semantically diverse vision-language dataset for remote sensing. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 5805–5813. [Google Scholar]
  55. Zhang, W.; Cai, M.; Zhang, T.; Zhuang, Y.; Mao, X. Earthgpt: A universal multi-modal large language model for multi-sensor image comprehension in remote sensing domain. arXiv 2024, arXiv:2401.16822. [Google Scholar]
  56. Mall, U.; Phoo, C.P.; Liu, M.K.; Vondrick, C.; Hariharan, B.; Bala, K. Remote sensing vision-language foundation models without annotations via ground remote alignment. arXiv 2023, arXiv:2312.06960. [Google Scholar]
  57. Iizuka, R.; Xia, J.; Yokoya, N. Frequency-based Optimal Style Mix for Domain Generalization in Semantic Segmentation of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 62, 1–14. [Google Scholar] [CrossRef]
  58. Zheng, J.; Wu, W.; Yuan, S.; Fu, H.; Li, W.; Yu, L. Multisource-domain generalization-based oil palm tree detection using very-high-resolution (vhr) satellite images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  59. Zhang, Y.; Zhang, M.; Li, W.; Wang, S.; Tao, R. Language-aware domain generalization network for cross-scene hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–12. [Google Scholar] [CrossRef]
  60. Li, D.; Yang, Y.; Song, Y.Z.; Hospedales, T. Learning to generalize: Meta-learning for domain generalization. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  61. Balaji, Y.; Sankaranarayanan, S.; Chellappa, R. Metareg: Towards domain generalization using meta-regularization. In Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  62. Li, Y.; Yang, Y.; Zhou, W.; Hospedales, T. Feature-critic networks for heterogeneous domain generalization. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 09–15 June 2019; pp. 3915–3924. [Google Scholar]
  63. Shankar, S.; Piratla, V.; Chakrabarti, S.; Chaudhuri, S.; Jyothi, P.; Sarawagi, S. Generalizing across domains via cross-gradient training. arXiv 2018, arXiv:1804.10745. [Google Scholar]
  64. Wang, Y.; Li, H.; Kot, A.C. Heterogeneous domain generalization via domain mixup. In Proceedings of the ICASSP 2020–2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3622–3626. [Google Scholar]
  65. Shu, Y.; Cao, Z.; Wang, C.; Wang, J.; Long, M. Open domain generalization with domain-augmented meta-learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 9624–9633. [Google Scholar]
  66. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  67. Segu, M.; Tonioni, A.; Tombari, F. Batch normalization embeddings for deep domain generalization. Pattern Recognit. 2023, 135, 9. [Google Scholar] [CrossRef]
  68. Bhattacharya, A.; Singha, M.; Jha, A.; Banerjee, B. C-SAW: Self-Supervised Prompt Learning for Image Generalization in Remote Sensing. In Proceedings of the Fourteenth Indian Conference on Computer Vision, Graphics and Image Processing, Rupnagar, India, 15–17 December 2023; pp. 1–10. [Google Scholar]
  69. Kang, J.; Fernandez-Beltran, R.; Duan, P.; Liu, S.; Plaza, A.J. Deep unsupervised embedding for remotely sensed images based on spatially augmented momentum contrast. IEEE Trans. Geosci. Remote Sens. 2020, 59, 2598–2610. [Google Scholar] [CrossRef]
  70. Ayush, K.; Uzkent, B.; Meng, C.; Tanmay, K.; Burke, M.; Lobell, D.; Ermon, S. Geography-aware self-supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 10181–10190. [Google Scholar]
  71. Manas, O.; Lacoste, A.; Giró-i Nieto, X.; Vazquez, D.; Rodriguez, P. Seasonal contrast: Unsupervised pre-training from uncurated remote sensing data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 9414–9423. [Google Scholar]
  72. Muhtar, D.; Zhang, X.; Xiao, P. Index your position: A novel self-supervised learning method for remote sensing images semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  73. Sun, X.; Wang, P.; Lu, W.; Zhu, Z.; Lu, X.; He, Q.; Li, J.; Rong, X.; Yang, Z.; Chang, H.; et al. RingMo: A remote sensing foundation model with masked image modeling. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–22. [Google Scholar] [CrossRef]
  74. Reed, C.J.; Gupta, R.; Li, S.; Brockman, S.; Funk, C.; Clipp, B.; Keutzer, K.; Candido, S.; Uyttendaele, M.; Darrell, T. Scale-mae: A scale-aware masked autoencoder for multiscale geospatial representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 4088–4099. [Google Scholar]
  75. Wang, D.; Zhang, Q.; Xu, Y.; Zhang, J.; Du, B.; Tao, D.; Zhang, L. Advancing plain vision transformer toward remote sensing foundation model. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–15. [Google Scholar] [CrossRef]
  76. Jakubik, J.; Roy, S.; Phillips, C.; Fraccaro, P.; Godwin, D.; Zadrozny, B.; Szwarcman, D.; Gomes, C.; Nyirjesy, G.; Edwards, B.; et al. Foundation models for generalist geospatial artificial intelligence. arXiv 2023, arXiv:2310.18660. [Google Scholar]
  77. Dong, Z.; Gu, Y.; Liu, T. Generative ConvNet Foundation Model with Sparse Modeling and Low-Frequency Reconstruction for Remote Sensing Image Interpretation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  78. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  79. Qi, L.; Kuen, J.; Shen, T.; Gu, J.; Guo, W.; Jia, J.; Lin, Z.; Yang, M.H. High Quality Entity Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 4024–4033. [Google Scholar]
  80. Qi, L.; Kuen, J.; Wang, Y.; Gu, J.; Zhao, H.; Torr, P.; Lin, Z.; Jia, J. Open world entity segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 8743–8756. [Google Scholar] [CrossRef]
  81. Hendrycks, D.; Gimpel, K. Gaussian error linear units (gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  82. Shi, B.; Zhang, X.; Xu, H.; Dai, W.; Zou, J.; Xiong, H.; Tian, Q. Multi-dataset pretraining: A unified model for semantic segmentation. arXiv 2021, arXiv:2106.04121. [Google Scholar]
  83. Chen, Y.; Wang, M.; Mittal, A.; Xu, Z.; Favaro, P.; Tighe, J.; Modolo, D. ScaleDet: A scalable multi-dataset object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7288–7297. [Google Scholar]
  84. Tong, X.Y.; Xia, G.S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Land-cover classification with high-resolution remote sensing images using transferable deep models. Remote Sens. Environ. 2020, 237, 111322. [Google Scholar] [CrossRef]
  85. Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. Deepglobe 2018: A challenge to parse the earth through satellite images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 172–181. [Google Scholar]
  86. Ji, D.; Zhao, F.; Lu, H.; Tao, M.; Ye, J. Ultra-high resolution segmentation with ultra-rich context: A novel benchmark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 23621–23630. [Google Scholar]
  87. Shi, J.X.; Wei, T.; Xiang, Y.; Li, Y.F. How Re-sampling Helps for Long-Tail Learning? In Proceedings of the Advances in Neural Information Processing Systems 36 (NeurIPS 2023), New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  88. Zhou, X.; Koltun, V.; Krähenbühl, P. Simple multi-dataset detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 7571–7580. [Google Scholar]
  89. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  90. Yu, Q.; He, J.; Deng, X.; Shen, X.; Chen, L.C. Convolutions die hard: Open-vocabulary segmentation with single frozen convolutional clip. In Proceedings of the Advances in Neural Information Processing Systems 37 (NeurIPS 2024), Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
Figure 1. Concept comparison between different settings. (a) Concept comparison on a class level. Single-task learning: Both training and test samples are drawn from the same distribution P X Y . The model is trained and evaluated on data with identical characteristics and class labels. Open-set/Open World/Out-of-Distribution (OOD) settings: The model is trained on known classes but must identify and mark novel classes as ’unknown’ during testing. This setting focuses on detecting out-of-distribution samples without requiring specific classification of novel classes. Zero-shot setting: The model is trained on base classes but must classify novel classes into specific categories during testing. This is achieved by leveraging predefined word embeddings that describe the novel classes. Open-vocabulary setting: The model is trained on base classes B but can classify novel classes N during testing by utilizing large-scale language vocabulary knowledge instead of predefined word embeddings. This setting emphasizes the use of external linguistic knowledge to generalize to unseen classes. (b) Concepts comparison on a domain level. Domain adaptation setting: The model is trained on a source data distribution P X Y 1 and adapts to an unseen target distribution P X Y 2 by accessing unlabeled or limited labeled target data during training. Domain generalization setting: The model is trained on one or more source distributions and evaluated directly on an unseen target distribution P X Y t without any adaptation. The goal is to achieve robust performance across diverse domains. (c) Open vocabulary domain generalization setting: The model is trained on base classes from one or more source distributions and must generalize to any unseen target distribution P X Y B + N 1 , where Y contains both base classes B t and novel classes N t . This setting combines the challenges of domain generalization and open-vocabulary learning.
Figure 1. Concept comparison between different settings. (a) Concept comparison on a class level. Single-task learning: Both training and test samples are drawn from the same distribution P X Y . The model is trained and evaluated on data with identical characteristics and class labels. Open-set/Open World/Out-of-Distribution (OOD) settings: The model is trained on known classes but must identify and mark novel classes as ’unknown’ during testing. This setting focuses on detecting out-of-distribution samples without requiring specific classification of novel classes. Zero-shot setting: The model is trained on base classes but must classify novel classes into specific categories during testing. This is achieved by leveraging predefined word embeddings that describe the novel classes. Open-vocabulary setting: The model is trained on base classes B but can classify novel classes N during testing by utilizing large-scale language vocabulary knowledge instead of predefined word embeddings. This setting emphasizes the use of external linguistic knowledge to generalize to unseen classes. (b) Concepts comparison on a domain level. Domain adaptation setting: The model is trained on a source data distribution P X Y 1 and adapts to an unseen target distribution P X Y 2 by accessing unlabeled or limited labeled target data during training. Domain generalization setting: The model is trained on one or more source distributions and evaluated directly on an unseen target distribution P X Y t without any adaptation. The goal is to achieve robust performance across diverse domains. (c) Open vocabulary domain generalization setting: The model is trained on base classes from one or more source distributions and must generalize to any unseen target distribution P X Y B + N 1 , where Y contains both base classes B t and novel classes N t . This setting combines the challenges of domain generalization and open-vocabulary learning.
Remotesensing 17 01357 g001
Figure 2. Overview of the proposed FreeMix. It consists of two branches: a base segmentation branch (BSB) and an entity segmentation branch (ESB). In order to enhance the extraction and recognition effectiveness of masks for both base and novel classes, we introduce the ESB to extract universal masks and their corresponding visual features. To maintain the model’s generalization performance, we employ self-supervised learning initialization (RS_SSL) to BSB and freeze the image encoder of the BSB and the mask extractor of the ESB.
Figure 2. Overview of the proposed FreeMix. It consists of two branches: a base segmentation branch (BSB) and an entity segmentation branch (ESB). In order to enhance the extraction and recognition effectiveness of masks for both base and novel classes, we introduce the ESB to extract universal masks and their corresponding visual features. To maintain the model’s generalization performance, we employ self-supervised learning initialization (RS_SSL) to BSB and freeze the image encoder of the BSB and the mask extractor of the ESB.
Remotesensing 17 01357 g002
Figure 3. The class frequency distributions on the training and testing sets of the four datasets.
Figure 3. The class frequency distributions on the training and testing sets of the four datasets.
Remotesensing 17 01357 g003
Figure 4. Qualitative results of semantic-segmentation Potsdam dataset. The images are taken from the Potsdam testing set.
Figure 4. Qualitative results of semantic-segmentation Potsdam dataset. The images are taken from the Potsdam testing set.
Remotesensing 17 01357 g004
Figure 5. Qualitative results on Potsdam, GID5, DeepGlobe and URUR dataset.
Figure 5. Qualitative results on Potsdam, GID5, DeepGlobe and URUR dataset.
Remotesensing 17 01357 g005
Figure 6. Comparison of proposal masks between FreeSeg and our FreeMix. In FreeMix, the base segmentation branch is highlighted in yellow, while the entity segmentation branch is depicted in purple. For visualization purposes, 10 masks are randomly selected from the 100 proposal masks generated by FreeSeg and each branch of FreeMix (BSB and ESB).
Figure 6. Comparison of proposal masks between FreeSeg and our FreeMix. In FreeMix, the base segmentation branch is highlighted in yellow, while the entity segmentation branch is depicted in purple. For visualization purposes, 10 masks are randomly selected from the 100 proposal masks generated by FreeSeg and each branch of FreeMix (BSB and ESB).
Remotesensing 17 01357 g006
Table 1. The mapping of classes for each dataset, as well as the partitioning into basic and novel classes. ’Original’ indicates the annotated labels from the original dataset. ’Mapped’ represents the labels after class mapping.
Table 1. The mapping of classes for each dataset, as well as the partitioning into basic and novel classes. ’Original’ indicates the annotated labels from the original dataset. ’Mapped’ represents the labels after class mapping.
DatasetTypeBase ClassesNovel Classes
PotsdamOriginalimpervious
surface
buildingcar low vegetationtree
Mappedimpervious
surface
buildingcar meadowtree
GID5Originalbuilt upfarmlandforest meadowwater
Mappedbuildingfarmlandforest land meadowwater
DeepGlobeOriginalurban landagriculture landrange landforest landwaterbarren land
Mappedbuildingfarmlandrange landforest landwaterbare land
URUROriginalbuildingfarmlandgreenhousewood landbare landwaterroad
Mappedbuildingfarmlandgreenhouseforest landbare landwaterroad
Table 2. Open-vocabulary semantic segmentation on the Potsdam dataset. The proposed FreeMix demonstrates better performance than prior arts. “imper.” indicates the class of impervious surface, and “VLM” stands for vision–language model. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Table 2. Open-vocabulary semantic segmentation on the Potsdam dataset. The proposed FreeMix demonstrates better performance than prior arts. “imper.” indicates the class of impervious surface, and “VLM” stands for vision–language model. The best results are marked in bold, while the second-best results are underlined. (Values in %).
MethodYearImage EncoderVLMPotsdamIoU of Base ClassesIoU of Novel Classes
mIoU mIoU s mIoU us fwIoU mACC pACC pACC s pACC us imper. Building Car Tree Meadow
ZSSeg2021ResNet50CLIP-B/1654.2778.4917.9451.0266.7166.9888.0534.7559.7485.3190.420.0035.88
ZegFormer2022ResNet50CLIP-B/1649.2071.9915.0145.2461.7362.2784.6727.9953.9375.5286.510.0030.03
MaskCLIP2023ResNet50CLIP-L/1615.5821.846.1921.5028.5439.4660.167.7832.2433.290.0011.231.16
SAN2023ResNet50CLIP-B/1638.5660.256.0238.8259.8060.7196.016.7052.9469.0458.772.129.92
OVSeg2023ResNet101CLIP-B/1631.5650.433.2535.0743.4954.2887.443.5441.7274.6234.960.216.28
FC-CLIP2023ConvNeXt_LCLIP-RN5044.7873.741.3239.0359.1259.7697.851.4848.1281.3291.790.052.60
FreeSeg2023ResNet50CLIP-B/1651.2575.8914.2946.5764.1265.1095.8918.0053.9981.8691.823.5425.05
FreeMix(ours)2024ResNet50CLIP-B/1663.4486.4628.9264.4573.8775.8789.9254.3783.8990.1685.3211.1146.73
Table 3. Open-vocabulary domain generalization performance for semantic segmentation on multiple benchmark datasets. All models are trained on a single dataset. The first column indicates the single dataset used for training each model. “MS” indicates multi-scale testing, and “SS” indicates single-scale testing. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Table 3. Open-vocabulary domain generalization performance for semantic segmentation on multiple benchmark datasets. All models are trained on a single dataset. The first column indicates the single dataset used for training each model. “MS” indicates multi-scale testing, and “SS” indicates single-scale testing. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Training DatasetModelTesting
Type
Testing Dataset: PotsdamTesting Dataset: GID5Testing Dataset: DeepGlobeTesting Dataset: URUR avg . mIoU avg . mIoU
mIoU s mIoU us mIoU mAcc mIoU s mIoU us mIoU mAcc mIoU s mIoU us mIoU mAcc mIoU s mIoU us mIoU mAcc
PotsdamZSSegMS78.4917.9454.2766.711.7313.666.5034.530.1510.833.7118.850.153.131.4314.9916.4733.77
ZegFormerMS71.9915.0149.2061.730.584.872.3020.3312.562.139.0815.9514.171.618.7913.8517.3427.96
MaskCLIPMS21.846.1915.5828.5413.850.358.4523.687.070.004.7114.687.080.004.0412.128.1919.75
FC-CLIPMS73.741.3244.7859.1212.510.007.5116.645.550.003.7013.447.870.004.508.3915.1224.39
FreeSegMS75.8914.2951.2564.122.9918.809.3133.352.9611.875.9322.651.5610.345.3221.5917.9535.42
FreeMix(ours)SS86.4628.9263.4473.8715.7316.3115.9643.903.418.935.2519.953.533.303.4315.7622.0238.37
GID5ZSSegMS0.0010.774.3120.0033.150.6320.1437.603.035.853.9718.619.453.136.7419.438.7923.91
ZegFormerMS6.3512.308.7323.9328.164.1818.5738.2128.460.2719.0629.756.3512.308.7323.9313.7728.95
MaskCLIPMS21.068.6016.0829.1916.450.6610.1320.7810.710.007.1416.499.890.005.659.989.7519.11
FC-CLIPMS22.7810.0017.6736.486.120.133.7219.663.590.002.4016.543.870.012.2110.556.5020.80
FreeSegMS3.3217.599.0223.4673.3622.2252.9161.8819.058.8115.6426.3015.861.729.8015.5121.8431.78
FreeMix(ours)SS8.3315.8311.3326.3676.4722.5554.9065.4423.0112.3819.4735.8120.959.7916.1726.4825.4638.52
DeepGlobeZSSegMS5.4311.517.8623.2914.599.8512.6932.530.855.602.4417.180.935.322.8120.316.4523.32
ZegFormerMS0.0012.274.9120.950.170.250.2020.107.205.706.7019.710.011.130.4914.283.0718.76
MaskCLIPMS16.437.5112.8623.856.510.003.9020.369.950.006.6326.265.590.003.1914.536.6421.25
FC-CLIPMS24.895.9217.3037.535.770.003.4619.722.710.001.8014.253.740.002.148.736.1720.05
FreeSegMS17.6222.6119.6137.1041.4015.9631.2244.049.447.038.6323.168.442.715.9917.8116.3630.52
FreeMix(ours)SS17.8917.1417.5939.3732.8920.2627.8449.3724.979.3519.7633.8819.126.4913.7124.9719.7236.89
URURZSSegMS2.9211.126.2021.647.538.367.8633.105.526.395.8120.305.181.873.7616.375.9022.85
ZegFormerMS0.595.342.4922.720.760.250.5620.4710.560.007.0422.310.021.130.5014.302.6419.95
MaskCLIPMS12.9412.6312.8228.1915.390.449.4121.4810.390.006.9317.0112.240.006.9913.179.0319.96
FC-CLIPMS26.378.8919.3835.175.740.003.4419.892.970.772.2416.363.780.002.1610.196.8020.40
FreeSegMS12.9522.5616.7932.2243.9221.9335.1257.8421.258.0516.8532.1221.005.8114.4924.7120.8136.72
FreeMix(ours)SS15.7021.7318.1136.1233.0616.9726.6254.0228.3913.0823.2837.3333.299.8023.2232.4522.8039.98
Table 4. Semantic-segmentation results on a testing set of joint dataset GPDU. “MS” indicates multi-scale testing, and “SS” indicates single-scale testing. FreeMix† indicates that DAS is not used during the training phase. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Table 4. Semantic-segmentation results on a testing set of joint dataset GPDU. “MS” indicates multi-scale testing, and “SS” indicates single-scale testing. FreeMix† indicates that DAS is not used during the training phase. The best results are marked in bold, while the second-best results are underlined. (Values in %).
ModelTraining datasetTesting typePotsdamGID5DeepGlobeURUR a v g . m I o U
m I o U m I o U s m I o U u s m I o U m I o U s m I o U u s m I o U m I o U s m I o U u s m I o U m I o U s m I o U u s
ZSSegGPDUMS9.8410.858.323.660.029.132.380.007.161.941.652.324.46
ZegFormer4.310.0010.770.500.830.008.1111.591.150.480.001.133.35
MaskCLIP11.7218.781.1313.6222.700.009.2013.800.006.8712.020.0010.35
SAN23.8423.9923.6135.1057.671.2529.3842.862.4130.4448.396.4929.69
OVSeg9.3214.960.8615.5822.155.7325.8737.572.4719.5631.863.1617.58
FC-CLIP21.0228.769.402.874.790.001.812.720.001.482.600.006.80
FreeSeg17.5814.7521.8325.2633.9412.2424.5531.5810.4923.7138.993.3422.78
FreeMix†(ours)GPDUSS19.9817.2524.0657.2675.9129.2732.0341.1713.7529.342.4111.8334.64
FreeMix(ours)GPDUSS47.0369.5413.2643.1367.915.9735.1452.690.0435.7260.852.2240.26
Table 5. Ablations on RS_SSL, entity segmentation branch, and dataset-aware sample-training tactic. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Table 5. Ablations on RS_SSL, entity segmentation branch, and dataset-aware sample-training tactic. The best results are marked in bold, while the second-best results are underlined. (Values in %).
Training DatasetRS_SSLESBDASPotsdamGID5DeepGlobeURUR avg . mIoU avg . mACC Δ mIoU Δ mACC
mIoU mACC mIoU mACC mIoU mACC mIoU mACC
GPDU 17.5832.5825.2632.9224.5539.7523.7130.9822.7834.05
19.7337.3238.5357.2231.9747.0626.4634.4929.1744.02+6.39+9.96
19.9835.9257.2671.1032.0346.4929.3037.8534.6447.84+5.47+3.81
47.0362.5643.1358.4435.1447.3735.7245.7540.2653.53+5.62+5.69
Table 6. Ablation study on different image encoders and pretraining methods. All methods utilize CLIP with a ViT-B backbone. The training dataset used is GPDU, which integrates four remote-sensing datasets. “In1K” denotes the ImageNet-1k dataset. The best results within each group are highlighted in bold. (Values in %).
Table 6. Ablation study on different image encoders and pretraining methods. All methods utilize CLIP with a ViT-B backbone. The training dataset used is GPDU, which integrates four remote-sensing datasets. “In1K” denotes the ImageNet-1k dataset. The best results within each group are highlighted in bold. (Values in %).
BackbonePre-Train
Type
Pre-Train
Dataset
Training
Tactic
PotsdamGID5DeepGlobeURUR avg . mIoU avg . mACC
mIoU mACC mIoU mACC mIoU mACC mIoU mACC
ResNet50SupervisedIn1Krandom19.7343.6638.5376.2331.9763.6626.4670.4529.1763.50
ResNet50Self-SupervisedMillionAIDrandom19.9840.4357.2687.6532.0363.2429.3071.7034.6465.75
ResNet50SupervisedIn1KDAS39.4961.3541.1878.047.8122.716.3919.0223.7145.28
ResNet50Self-SupervisedMillionAIDDAS47.0362.4743.1381.9035.1469.3835.7278.4540.2573.05
Swin-Brandom11.7930.9233.7673.9922.2062.0626.0571.9623.4559.73
Swin-BSelf-SupervisedMillionAIDrandom11.5431.0939.3178.6323.2261.9127.7172.3825.4461.00
Swin-BDAS36.2660.7034.2972.5819.7848.3915.0547.1826.3457.21
Swin-BSelf-SupervisedMillionAIDDAS43.8566.5740.1179.4923.8156.5918.4454.5331.5564.29
Table 7. Scalability of entity segmentation branch. All methods utilize CLIP with a ViT-B backbone. The training dataset is GPDU, which combines four remote-sensing datasets. The best results within each group are highlighted in bold. (Values in %).
Table 7. Scalability of entity segmentation branch. All methods utilize CLIP with a ViT-B backbone. The training dataset is GPDU, which combines four remote-sensing datasets. The best results within each group are highlighted in bold. (Values in %).
Backbone of ESBTraining TacticPotsdamGID5DeepGlobeURUR avg . mIoU avg . mACC
mIoU mACC mIoU mACC mIoU mACC mIoU mACC
Swin-Trandom19.9840.4357.2687.6532.0363.2429.3071.7034.6465.75
Swin-L23.3145.9055.0788.7124.8550.9223.4059.6731.6561.30
Hornet-L21.8645.6553.1488.6330.4064.1527.7572.0333.2867.61
Swin-TDAS47.0362.4743.1381.9035.1469.3835.7278.4540.2573.05
Swin-L47.3957.0957.1185.9010.8931.0014.6624.7032.5149.67
Hornet-L53.6866.1753.4082.5512.9136.8115.6730.1033.9153.90
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, J.; Shi, J.; Zhao, Z.; Liu, Z.; Zhi, R. FreeMix: Open-Vocabulary Domain Generalization of Remote-Sensing Images for Semantic Segmentation. Remote Sens. 2025, 17, 1357. https://doi.org/10.3390/rs17081357

AMA Style

Wu J, Shi J, Zhao Z, Liu Z, Zhi R. FreeMix: Open-Vocabulary Domain Generalization of Remote-Sensing Images for Semantic Segmentation. Remote Sensing. 2025; 17(8):1357. https://doi.org/10.3390/rs17081357

Chicago/Turabian Style

Wu, Jingyi, Jingye Shi, Zeyong Zhao, Ziyang Liu, and Ruicong Zhi. 2025. "FreeMix: Open-Vocabulary Domain Generalization of Remote-Sensing Images for Semantic Segmentation" Remote Sensing 17, no. 8: 1357. https://doi.org/10.3390/rs17081357

APA Style

Wu, J., Shi, J., Zhao, Z., Liu, Z., & Zhi, R. (2025). FreeMix: Open-Vocabulary Domain Generalization of Remote-Sensing Images for Semantic Segmentation. Remote Sensing, 17(8), 1357. https://doi.org/10.3390/rs17081357

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop