Next Article in Journal
AI-Based Detection of Optical Microscopic Images of Pseudomonas aeruginosa in Planktonic and Biofilm States
Previous Article in Journal
How Facebook Mediated COVID-19 Risk Communication: Evidence from Chinese External Media During the Winter Olympics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MultiLTR: Text Ranking with a Multi-Stage Learning-to-Rank Approach

1
School of Artificial Intelligence, Zhongyuan University of Technology, Zhengzhou 450007, China
2
VISTA Lab, Algoritmi Center, University of Évora, 7000-671 Évora, Portugal
3
Department of Computer Science, University of Évora, 7000-671 Évora, Portugal
*
Author to whom correspondence should be addressed.
Information 2025, 16(4), 308; https://doi.org/10.3390/info16040308
Submission received: 13 February 2025 / Revised: 4 April 2025 / Accepted: 11 April 2025 / Published: 13 April 2025

Abstract

:
The division of retrieval into multiple stages has evolved to balance efficiency and effectiveness among various ranking models. Faster but less accurate models are used to retrieve results from the entire corpus. Slower yet more precise models refine the ranking within the top candidate list. This study proposes a multi-stage learning-to-rank (MultiLTR) method. MultiLTR applies learning-to-rank techniques across multiple stages. It incorporates text from different fields such as titles, body content, and abstracts to produce a more comprehensive and accurate ranking. MultiLTR iteratively refines ranking accuracy through sequential processing phases. It dynamically selects top-performing rankers from a diverse candidate pool at each stage. Experiments were carried out on benchmark datasets, MQ2007 and MQ2008, using three categories of learning-to-rank algorithms. The results demonstrate that MultiLTR outperforms state-of-the-art ranking approaches, particularly in field-based ranking tasks. This study improves ranking accuracy and offers new insights into enhancing multi-stage ranking strategies.

1. Introduction

The effectiveness of learning-to-rank (LTR) techniques within a multi-stage retrieval framework remains underexplored. This paper introduces a multi-stage learning-to-rank (MultiLTR) approach. MultiLTR integrates LTR techniques into a multi-stage retrieval process. This approach leverages field-based information and addresses ranking challenges in text re-ranking tasks. While existing methods demonstrate the value of field-based feature aggregation [1], their static weighted-sum approach to combining field signals exhibits some limitations, such as rigid interaction modeling that cannot adapt to query-dependent feature importance; error propagation from separate field-level optimization to final ranking; and context blindness in handling multi-stage retrieval dependencies.
The methodology presented in this article builds on previous research introduced in [1]. In that work, we proposed the field learning-to-rank (fLTR) framework and trained rankers using field-based features; results from various text fields were aggregated. In the final phase of the fLTR method, the ranking problem was reformulated as an aggregation task, and the aggregation techniques significantly impacted the overall training outcomes. In contrast, this article defines the task as an ordinal LTR problem and incorporates multistage retrieval principles. We refer to this approach as the MultiLTR method.
In summary, this paper makes the following contributions:
  • We propose a multi-stage ranking approach. The initial stage constructs a set of retrieval models and utilizes features derived from a single text field. The subsequent re-ranking stages build local and global rankers using learning-to-rank techniques.
  • We introduce a mid-evaluation strategy to select the best-performing model. This strategy evaluates a set of candidate models before generating the final ranking list. The model achieving the highest performance acts as the final ranking model.
  • We conduct empirical experiments on two publicly available benchmark datasets and use three different LTR algorithms. We evaluate the results using widely adopted assessment metrics.
The remainder of this paper is organized as follows: Section 2 reviews related work, Section 3 describes the proposed approach, Section 4 presents the experiments and results, Section 5 discusses the key findings. Finally, Section 6 concludes this paper and outlines potential directions for future research.

2. Literature Review

This section reviews the multi-stage ranking approach, focusing on re-ranking techniques. Additionally, it surveys neural ranking models and provides an overview of the LTR techniques.

2.1. Multi-Stage Ranking Architectures

A multi-stage ranking architecture typically consists of an initial ranking phase and one or more re-ranking stages [2,3,4,5]. Traditionally, the initial ranking uses classical term-based retrieval models. The retrieved results are subsequently passed to the re-ranking stage [6,7].
Recent studies have explored various approaches to multi-stage ranking tasks. In [8], a hybrid re-ranking framework was introduced. It uses BM25 as the initial ranker and a cross-attention neural model as the re-ranker. They are evaluated on two datasets, namely, the MS MARCO passage retrieval dataset and the BEIR corpus for zero-shot retrieval. The results demonstrate its effectiveness compared to baseline models and other approaches.
Wang et al. [9] proposed a multi-stage information search architecture to generate semantic embeddings. It employs a post-fusion strategy to project items retrieved from different stages into a common space. It utilizes multi-grained learning objectives to preserve multi-level similarity. The authors conducted experiments on a billion-scale corpus from the JD E-commerce website. They compared the results with those from DSSM [10], DPSR [11], and RSR [12], demonstrating its effectiveness.
Hai et al. [13] developed a lightweight learning technique to train a first-stage ranker. The paper used T5Mono for the second-stage ranking. They trained the two-stage model on the CANARD dataset and evaluated the results on the TREC CAsT 2020–2021 datasets. This approach is effective in conversational search tasks. In [14], Yang et al. proposed a mixed-initiative query reformulation module within a two-stage retrieval pipeline. BM25 serves as the first-stage ranker. MonoT5 and DuoT5 are used as re-rankers. Experiments were carried out on the TREC CAsT 2021–2022 datasets. The results showed that this method achieved results comparable to two popular reformulators, namely, CANARD-T5 and the historical query reformulator.
Finally, the LCE (localized contrastive estimation) method proposed by Gao et al. [15] re-ranks documents in a deep two-stage model. It aims to enhance initial rankings and experiments on the MS MARCO dataset. The results demonstrate that LCE outperforms the Vanilla method. The method achieves even higher scores when combined with HDCT as the initial retriever. The findings suggest that LCE has strong potential for integrating learning techniques with existing language models for re-ranking tasks.

2.2. Neural Ranking Models

In recent years, neural ranking models have been increasingly used for document re-ranking [6,16,17,18]. In text re-ranking tasks, existing research demonstrates that pre-trained neural network models can effectively learn text representations and text–pair interactions for various applications. The scenario includes document ranking and re-ranking, document summarization, query suggestion, and query parsing [7,19,20,21].
Neural ranking models can be categorized in two ways. On the one hand, they can be classified based on whether they utilize transformer-based BERT models. And this classification results in two groups, namely, pre-BERT- and BERT-based neural ranking models [21]. BERT is one of the most significant transformer-based pre-trained models [22]. In text re-ranking, BERT and its variants have been widely adopted to estimate the relevance between queries and documents [7,21]. Instead of relying on a single input representation, dynamic memory networks [23] have been used to model sentence-level representations and aggregate BERT output. They carry out the experiments on three passage-ranking datasets, namely, ANTIQUE, InsuranceQA, and TREC-DL 2019. The results demonstrate that incorporating sentence-level representations improves re-ranking performance over fine-tuned baseline BERT models. Additionally, researchers found that training only the DMN layer could achieve comparable performance to the full model while significantly enhancing training efficiency [23]. Two BERT-based ranking variants, monoBERT, and duoBERT, are proposed to address ranking problems [6]. monoBERT employs a point-wise classification approach, while duoBERT follows a pairwise classification strategy. These models are incorporated into a multi-stage ranking framework and evaluated on TREC CAR and MS MARCO datasets. This method achieves comparable results when compared to the state-of-the-art techniques.
On the other hand, neural ranking models can also be classified based on neural network techniques’ applications: during the representation or interaction phase. This results in three categories, namely, representation-based models, interaction-based models, and hybrid models [7,21]. Representation-based models independently learn dense vector representations for queries and documents. These representations are then compared using similarity measures such as cosine similarity or inner products to determine relevance [7,21]. In contrast, interaction-based models compare the representations of individual query terms with document terms. The interaction-based models construct a similarity matrix that captures word interactions. This matrix further processes to derive a relevance score [7,21]. Hybrid models incorporate elements from both representation-based and interaction-based approaches. A well-known example is Duet [24]. It integrates a representation-learning component with an interaction-based component responsible for detecting exact term matches.
Furthermore, Fan et al. [7] classified neural ranking models into two broader generations. The first generation, emerging around 2010, focused on word embedding techniques for learning word representations. The second generation leveraged transformer-based methods to learn text representations and interactions. Consequently, research that uses classical word embedding methods—such as Word2Vec [25] and GloVe [26]—falls into the pre-BERT category [7].

2.3. Learning-to-Rank Techniques

Semantic retrieval models and neural ranking models demonstrate their effectiveness across various ranking tasks [27]. A key distinction between these models and LTR methods is that they typically do not rely on additional manually crafted features to estimate document relevance to a query [28]. However, their relative performance compared to feature-based LTR methods remains an open question [29]. Additionally, feature-based LTR methods continue to attract research interest, particularly in areas such as efficiency, diversification, and permutation-invariant models [29]. Notably, recent efforts research on bridging the performance gap between neural and ensemble-based models [30].
Learning-to-rank techniques and their corresponding models play a crucial role in re-ranking within multi-stage ranking systems. These techniques are typically classified into three categories, namely, pointwise, pairwise, and listwise approaches. Classical LTR models include the following: MART [31], RankNet [32], RankBoost [32], AdaRank [33], Coordinate ascent [34], LambdaMART [35], ListNet [36], and random forest [37]. In recent years, several new LTR models have emerged, including the following: DirectRanker [38], PairRank [39], DeepPLTR [40], DLCM [41], DeepQRank [42], SetRank [43], PiRank [44], PoolRank [45], and ListMAP [46].
A study by Chen et al. [47] explored the use of a neural-based multi-layer perceptron model in combination with the BM25 retrieval model for document ranking. Their experiments on LambdaMART demonstrated that supervised learning-based re-rankers could enhance ranking performance. Han et al. [48] introduced a machine learning algorithm for document re-ranking. A learning-to-rank model was applied after encoding queries and documents using BERT. Their approach led to a 4.3% improvement over their previous best-performing re-ranking method. Additionally, Awan et al. [49] proposed a deep learning approach to address named entity normalization as a pairwise learning-to-rank problem. Their method uses BM25 to generate candidate entities and then applies BERT representations for re-ranking. Experimental results across various species entity types have demonstrated the superiority of their method over existing state-of-the-art techniques. One study highlighted the limitations of the traditional probability ranking principle (PRP) in multi-stage information retrieval systems. This study proposes a generalized probability ranking principle (GPRP). It integrates both stage-specific selection bias and user interests. It implements GPRP through a full-stage LTR framework [4].

3. Methodology

This section introduces the proposed MultiLTR approach. It presents the overall architecture and explains the MultiLTR framework in detail.

3.1. Architecture of the Proposed MultiLTR Approach

Given a query q and a set of documents t 1 , t 2 , …, t n from a document collection T, the objective of a ranking task is to order these documents based on their relevance to the query. Documents with higher relevance scores appear at the top of the ranking list [8]. In a multi-stage ranking system, the initial retrieval stage identifies a set of candidate documents, subsequent re-ranking phases refine their order, and the final stage returns the ranked list to the user. Our primary goal is to enhance ranking performance through effective re-ranking techniques.
The initial ranking layer employs classical text retrieval models to rank documents. For example, if a user searches for “effective treatments for diabetes” in a medical database, the initial retrieval might return all documents containing the keywords “diabetes” and “treatments”. The re-ranking phase would then apply a model to refine the ranking. For example, a TF-IDF-based algorithm can be used to prioritize documents based on keyword frequency. However, this approach does not consider the context of the information. This approach potentially ranks documents that mention “diabetes” and “treatments”. However, it misses detailed or reliable information about effective treatments.
In contrast, our approach improves this by multiple re-ranking layers. The initial layer generates preliminary results using BM25. And the local re-ranking layer applies supervised LTR techniques. Each re-ranker focuses on a specific field, such as the title or abstract. The global re-ranking layer integrates information from different fields. It mitigates correlation issues by delaying feature combinations. Additionally, the mid-evaluation layer assesses the ranker performance using standard evaluation metrics. We select the best-performing model to construct the final re-ranked list. This ensures that the most relevant and reliable documents are ranked at the top. This can provide users with high-quality, contextually relevant information about effective treatments for diabetes.
Figure 1 illustrates the hierarchical structure of the MultiLTR and the flow of information from input features to the final re-ranked output. The system begins with an initial ranking stage. A series of re-ranking layers then refine and filter the results from the initial retrieval. At each stage, we process and refine a ranked list of candidates from the previous phase before passing to the next stage. Finally, we select the best-performing model for the final process and return the ultimate ranking list to the user.
To clarify the interactions between different ranking layers, we present the inputs and outputs of each layer in Figure 2. Refining ranking performance is crucial in each stage, ensuring an efficient and effective retrieval process. We define the key notations used in Table 1. Algorithm 1 provides the detailed pseudocode for the proposed MultiLTR approach.
Algorithm 1: Pseudocode for the MultiLTR Approach.
Information 16 00308 i001

3.2. Layers in the Proposed MultiLTR Approach

This section provides a detailed explanation of the components illustrated in Figure 2, including the input layer, feature clustering layer, initial ranking layer, local re-ranking layer, normalization layer, global re-ranking layer, mid-evaluation layer, final re-ranking layer, and output layer.

3.2.1. Input Layer

The dataset undergoes pre-processing through feature engineering techniques to ensure compatibility with the MultiLTR approach. The dataset can also be transformed using pre-trained language models like BERT to generate embeddings. These refined features are then used as input for the ranking architecture.
Formally, we extract a feature f and represent it as follows:
f = E ( t , q )
where E represents a feature extractor, q denotes the given queries, and t corresponds to the associated documents. As a result, the input layer receives an array of features { f 1 , f 2 ,…, f s }. These features serve as the foundation for subsequent ranking processes.

3.2.2. Feature Clustering Layer

When training ranking models, highly correlated features can negatively impact performance. Empirical studies [1] show that features originating from different domains or fields often exhibit significant correlations. Traditionally, ranking models are trained using a combined set of extracted features without considering their relationships. However, the model may inadvertently rely on redundant information if highly correlated features are not handled properly. And this can lead to suboptimal performance [50].
In machine learning, blindly aggregating strongly correlated attributes can degrade model effectiveness. Incorporating diverse and discriminative features with minimal inter-correlation can help to improve performance. Previous research demonstrates that using field-based features outperforms naive feature aggregation [1]. Following this principle, the proposed MultiLTR approach clusters feature into distinct groups. This ensures that each ranker is trained using field-specific data. The feature clustering layer organizes input features into groups based on their respective fields, such as the title, abstract, URL, or body.
As illustrated in Figure 2, the approach assumes we have s features extracted from d fields. The feature clustering layer classifies these features into d distinct groups. To maintain balanced contributions from each field, we retain an equal number of features for every field. Thus, each group consists of k field-specific features. They are represented as f 1 ( i ) , f 2 ( i ) , …, f k ( i ) , and i indicates the field from which the features are extracted.

3.2.3. Initial Ranking Layer

The initial ranking layer is responsible for ranking documents using traditional models. Its primary function is to generate preliminary ranking results from the dataset, which are then passed to subsequent stages for further refinement.
Each initial ranker, denoted as r 1 ( i ) , is trained exclusively using features { f 1 ( i ) , f 2 ( i ) , …, f k ( i ) } extracted from a specific field i. As a result, this process produces a collection of initial rankers, represented as { r 1 ( 1 ) , r 1 ( 2 ) , …, r 1 ( d ) }, each based on field-specific characteristics.

3.2.4. Local Re-Ranking Layer

After the initial ranking layer, the local re-ranking layer applies supervised LTR techniques to refine the ranking results. We term this layer as the “local” re-ranking layer because each ranker focuses exclusively on information from a single field in this phase. This ensures a more field-specific ranking process.
As outlined in [51], LTR techniques rely on training data that consists of the following: a set of queries { q 1 , q 2 , …, q m }; a corresponding collection of documents { x 1 ( 1 ) , x 2 ( 1 ) , …, x n ( 1 ) }; and their associated relevance judgments { y ( 1 ) , y ( 2 ) , …, y ( m ) }. These relevance judgments can be formulated in the following ways: (1) binary or graded relevance, indicating whether a document is relevant to a query or the degree of relevance; (2) pairwise preference, specifying the relative relevance between two documents for a given query; and (3) listwise ranking, determining the optimal order of documents for a query. Consequently, LTR algorithms are categorized into three types: pointwise, pairwise, and listwise approaches. When we input a test query q into the trained LTR re-ranker, the relevant documents { x 1 , x 2 , …, x n } are re-ranked, producing a refined ranking list h(x) [51].
In the proposed MultiLTR framework, a group of LTR rankers is constructed. The training dataset for these rankers is derived from the results of the initial ranking layer. This ensures that only candidates are retrieved for further ranking refinements.
We apply eight local LTR algorithms for each field and generate eight distinct local re-rankers per field. As illustrated in Figure 2, this process results in a collection of local re-rankers { r 2 ( i l 1 ) , r 2 ( i l 2 ) , …, r 2 ( i l 8 ) } for each field i. This ensures a structured and comprehensive ranking refinement process.

3.2.5. Normalization Layer

Each model is trained exclusively on field-specific features in previous layers. The characteristics of each field heavily influence the re-ranking outcomes. And this often leads to significant variations in scale. Features with larger value ranges tend to dominate the results. Standardizing the data through normalization ensures each field contributes equally to the final ranking.
Normalization mitigates the influence of varying feature magnitudes. Normalization plays a crucial role in standardizing results before passing to the global re-ranking layer. Specifically, we apply two normalization methods, namely, min–max normalization and Z-score normalization [52].
Min–max normalization. Min–max normalization typically adjusts numerical data to fit within a predefined range between 0 and 1. It is computed using the following formula:
X norm = X X min X max X min
where X is the original value, X norm is the normalized value, and X min and X max are the minimum and maximum values in the dataset or feature.
Z-score normalization. Z-score normalization is a technique used to scale numerical data and has a mean of 0 and a standard deviation of 1. It is computed using the following formula:
Z = X μ σ
where X is the original data value, Z is the standardized value, and μ and σ represent the mean and the standard deviation of the dataset or feature, respectively.
Min–max normalization ensures uniform scaling across all features but may struggle to handle outliers effectively. In contrast, Z-score normalization is well-suited for managing outliers but does not produce data with a consistent scale. To address these limitations, the MultiLTR method integrates the two normalization techniques to balance uniform scaling and robustness to outliers. As illustrated in Figure 2, a set of normalized results { r 3 ( i l 1 ) , r 3 ( i l 2 ) , …, r 3 ( i l 8 ) } is generated for each field i,

3.2.6. Local Re-Ranker Selection Layer

This layer aims to select the most effective local rankers. We use normalized discounted cumulative gain (NDCG) [53] to evaluate the normalized outputs. NDCG, a widely recognized ranking metric, was applied with cut-off values set at 5, 10, 15, 20, and 30.
Each field contains eight local LTR algorithms, resulting in eight corresponding local re-rankers. We chose the top-performing rankers to advance to the global re-ranking layer and discarded the remaining ones. The selection process ensures coverage across all three categories of LTR algorithms. For each category, the best-performing LTR algorithm is identified. And each field ultimately retains three top-performing local re-rankers. As illustrated in Figure 2, this layer selects a set of rankers { r 4 ( i b e s t p o i n t w i s e ) , r 4 ( i b e s t p a i r w i s e ) , …, r 4 ( i b e s t l i s t w i s e ) } for each field i.

3.2.7. Global Re-Ranking Layer

The preceding layers process the information independently for each field without considering their combined impact. Combining all features from different fields at the beginning of model training could introduce correlation issues and affect ranking performance. This layer is a delayed fusion mechanism and integrates contributions from different fields. By deferring the combination, this approach mitigates such risks and allows a more effective integration of field-specific insights.
Traditional fusion techniques are often used to merge ranking results from multiple sources. The MultiLTR approach introduces an additional re-ranking process to enhance ranking quality further. At this stage, the same LTR techniques applied in the local re-ranking layer are utilized to construct global rankers. The results from the previous layer are treated as features within the LTR algorithms. This enables the model to learn and refine the final ranking.
As illustrated in Figure 2, a set of global re-rankers { r 5 ( l g 1 ) , r 5 ( l g 2 ) , …, r 5 ( l g 8 ) } is generated. Each ranker is trained using evaluation metrics at multiple cut-off points (5, 10, 15, 20, and 30).

3.2.8. Mid-Evaluation Layer

In the preceding layer, each model is trained using a specific evaluation metric at different cut-off values. This layer acts as an intermediate evaluation stage. We assess the trained rankers using widely accepted ranking metrics. The best-performing model is identified and then used to construct the final re-ranking list. We employ the normalized discounted cumulative gain (NDCG) metric with various cut-off values (5, 10, 15, 20, and 30).
To maintain consistency, we apply the same evaluation metric for both training and assessment. For instance, if a model is trained using NDCG@5, its evaluation is also conducted using NDCG@5. We select the model that achieves the highest improvement over the baseline as the best-performing model, denoted as r b e s t . We then use this model to re-rank the documents and generate the final ranked list.

3.2.9. Final Re-Ranking Layer and Output

The top-performing models { r 5 ( l g 1 b e s t ) , r 5 ( l g 2 b e s t ) , …, r 5 ( l g 8 b e s t ) } are selected and utilized for re-ranking the final results. The output layer then delivers the optimized ranked list to the system users. This ensures that the most relevant documents are presented at the top.

4. Experiments and Results

This section first presents the datasets used in the experiments. Then, we discuss the evaluation metrics. Next, we provide a detailed description of the constructed rankers and conclude the analysis of the results.

4.1. Datasets

The MultiLTR technique implements and evaluates using benchmark datasets from Microsoft LETOR 4.0 [54]. Specifically, the MQ2007 and MQ2008 datasets are derived from the TREC 2007–2008 Million Query Track, which utilizes the Gov2 web page collection, containing approximately 25 million pages. MQ2007 consists of around 1700 queries with labeled documents, and MQ2008 includes approximately 800 queries. Relevance scores range from 0 to 2, where 2 signifies high relevance and 0 indicates low relevance.
Table 2 presents a summary of the key statistics for both datasets.
Each dataset includes a predefined set of 46 standard features [54]. We exclude certain features to better align with the objectives of our research experiments. We group the features into six categories based on the field from which they originate: anchor, url, title, body, whole document, and other place. The other place category is omitted since it lacks field-specific information. Consequently, features such as number of child pages, number of inlinks, number of outlinks, and PageRank are removed. To maintain an equal and consistent number of features across all fields, we also exclude the following features: length of URL and number of slashes in URL.
The final dataset consists of five fields, each containing eight different types of features. The experimental features, totaling 40, are summarized in Table 3.
The datasets are pre-split into five folds. Each fold consists of three subsets, namely, training, validation, and test sets. The data distribution typically follows a ratio of 60% for training, 20% for validation, and 20% for testing. This ensures a balanced and effective evaluation of the models.

4.2. Evaluation Measures

We evaluate the effectiveness of the proposed method using two widely adopted assessment metrics, namely, mean average precision (MAP) [55] and NDCG [53]. These metrics provide a comprehensive measure of ranking quality and relevance.
Mean average precision. In a text retrieval system, the average precision (AP) for a single query is computed as the mean precision obtained after retrieving each relevant document from a ranked list of top documents [56]. The mean average precision (MAP) is derived as the arithmetic mean of the AP values across all query topics. MAP has strong discriminative power and stability [57].
Formally, MAP is defined as follows:
MAP = q = 1 Q Average Precision Q
where each query, q, is part of Q, the total number of queries [58].
Normalized discounted cumulative gain. Another widely used metric for evaluating ranking performance is NDCG [57]. It penalizes highly relevant documents that appear lower in the ranking. Conceptually, NDCG is defined as follows:
NDCG @ n = DCG n IDCG n
where discounted cumulative gain (DCG) represents the accumulated gain from ranking results, with higher-ranked documents receiving greater weight, and ideal discounted cumulative gain (IDCG) refers to the document ordering based on their actual relevance (the most relevant documents appear first). The NDCG value at a specific rank position n is denoted as NDCG@n.

4.3. Built Rankers

The ranking models are constructed using the RankLib tool (https://sourceforge.net/p/lemur/wiki/RankLib/, accessed on 12 January 2024). A diverse set of LTR algorithms is employed. They include two pointwise algorithms (MART and random forest), three pairwise algorithms (RankBoost, RankNet, and LambdaMART), and three listwise algorithms (ListNet, coordinate ascent, and AdaRank). For ease of reference, we abbreviate the algorithms as follows: MART (MR), random forest (RF), RankBoost (RB), RankNet (RN), LambdaMART (LM), ListNet (LN), coordinate ascent (CA), and AdaRank (AR).
The experiments adhered to the LETOR benchmark dataset configuration [54], utilizing five-fold cross-validation. Hyperparameter tuning was conducted using the validation dataset. The final performance was evaluated based on average scores from the test dataset at cut-offs (5, 10, 15, 20, and 30) for the NDCG and MAP metrics.
Following the framework illustrated in Figure 1 and Figure 2, the experimental process begins with the construction of primary rankers in the initial ranking layer. Each ranker is trained exclusively on features from a specific field. The local re-ranking layer then applies LTR algorithms to refine the rankings of the retrieved candidates. Next, the normalization layer standardizes the ranking scores across different fields to ensure fair comparison and combination. The intermediate evaluation results, obtained after normalization, are reported in Table 4 and Table 5.
The optimal algorithm selection is based on the total number of rankers that outperform the baseline. This is referred to as the winning number [55], as shown in Table 6. We identify the top three performers for each dataset. In Table 6, RF emerges as the top performer on the MQ2007 dataset with 17 wins, followed by LN with 15. Additionally, RN, RB, and CA each achieve a count of 10. We observe a similar pattern in MQ2008, where RF again leads with 17 wins, LN follows with 15, and RN and CA each attain 10.
During the selection process, RF performs as the best-performing algorithm, representing the pointwise category. LN, as the second-highest performer, represents the listwise category. And RN represents the pairwise category.
In the global re-ranking layer, we only utilize these selected optimal LTR algorithms (RF, RN, and LN) to integrate results from the local re-ranking layers. For each selected algorithm, multiple rankers are trained on the training data. We select and retain the model performing best on the validation data. The training metric is NDCG at cutoff values. We chose the top-performing ranker as the final ranking model. It is then saved and used to generate the final ranking results.

4.4. Results

The MultiLTR retrieval effectiveness is evaluated using the NDCG@n and MAP@n metrics, as shown in Figure 3 and Figure 4.
As outlined in Section 4.3, we select three optimal LTR algorithms. LN represents the pointwise approach, RN represents the pairwise approach, and RF represents the listwise approach. These algorithms are used to construct the local re-ranking rankers. In the global re-ranking layer, all eight LTR algorithms are incorporated. The reported results reflect performance after the completion of all layers.
Analyzing the results in Figure 3, the MultiLTR approach consistently outperforms the baselines across all NDCG@n evaluations for the MQ2007 dataset, except for cases where LM_Global is used as the re-ranking algorithm. Among all three selected local re-ranker algorithms, the top-performing global re-rankers are RF_Global and MR_Global. They are both pointwise-based methods and consistently yield the best results:
  • When employing LN_Local as the local re-ranker, the most effective model utilizes RF_Global as the global re-ranker. This achieves a notable improvement of 8.62%. Following closely, MR_Global as the global re-ranker yields an improvement of 6.95%;
  • For RN_Local as the local re-ranker, the optimal model applies MR_Global as the global re-ranker. This leads to a significant improvement of 7.39%, with RF_Global following at 6.58%;
  • In the case of RF_Local, the best-performing model integrates MR_Global as the global re-ranker. This produces a substantial improvement of 8.84%, while RF_Global ranks second with a 6.69% improvement.
Similarly, for the MQ2008 dataset, models using LM_Global as the global re-ranker fail to surpass the baselines across all three selected local LTR algorithms:
  • With LN_Local as the local re-ranker, RF_Global emerges as the most effective global re-ranker. It delivers a substantial 13.39% improvement in NDCG@5, closely followed by MR_Global with a 12.11% increase;
  • For RN_Local, RF_Global remains the optimal choice. It achieves a remarkable 14.69% improvement, with MR_Global trailing slightly at 13.89%;
  • When applying RF_Local, RF_Global again leads with a 12.29% improvement, while MR_Global secures the second-best performance at 10.40%.
Figure 4 presents the MAP@n evaluation results. It exhibits a trend similar to that observed with the NDCG@n metric. This reinforces the consistency of the proposed MultiLTR approach in enhancing retrieval performance.

5. Discussion

To better understand the results, we first conduct a comparative analysis of the MQ2007 dataset using NDCG and MAP metrics. Next, we assess the impact of three key components in MultiLTR through extensive ablation experiments. Additionally, we evaluate the computational efficiency of the proposed method.

5.1. Comparative Analysis

Table 7 and Table 8 present a comparative analysis of the MQ2007 dataset, using the NDCG and MAP metrics, respectively. We highlight and evaluate the top-performing models for each approach.
To thoroughly assess the effectiveness of our models against existing ones, we include the performance of these models: the classic BM25 and three neural-based ranking models (KNRM, HiNT, and DeepTileBars). KNRM [59] is a neural retrieval model that captures term-level interactions via embedded representations. It utilizes kernel-based pooling to extract multi-level soft-matching features, and computes a final relevance score to align documents with queries. HiNT [60] is a hierarchical neural retrieval model and processes segment-level interaction matrices as input. It leverages a local matching layer and a global decision layer to determine document relevance. DeepTileBars [61] segments documents into topic-based sections and applies Convolutional Neural Networks of varying sizes to analyze these sections.
As shown in Table 7 and Table 8, the MultiLTR rankers consistently outperform KNRM and DeepTileBars across all evaluation metrics and surpass HiNT at the @30 metric. Furthermore, the results demonstrate that the MultiLTR approach provides additional improvements over the fLTR method introduced in our previous study [1]. Specifically, as Table 7 presents, even the modest improvement achieved by MultiLTR under the LN algorithm shows the superiority of MultiLTR compared to other approaches: MultiLTR outperforms both BM25 and fLTR across all cutoff points (@5–@30). For instance, with a cutoff @30, MultiLTR (0.5645) surpasses BM25 (0.5390) and fLTR (0.5413) by 2.55% and 2.32%, respectively; when compared with the neural baselines, MultiLTR (0.5645) substantially exceeds all neural models with cutoff @30, including HiNT (0.5358), KNRM (0.4324), and DeepTileBars (0.4651), achieving 2.87–13.21% improvements; MultiLTR still outperforms other neural baselines (KNRM, DeepTileBars) for early cutoffs (@5–@20) while slightly trailing HiNT. Similar results with the MAP evaluation can be observed in Table 8.
The fLTR method [1] is a two-stage re-ranking technique that employs aggregation algorithms like comSUM [62] to merge results from different fields. In contrast, the MultiLTR approach utilizes a multi-stage re-ranking strategy for enhanced flexibility and performance.
Aggregation methods such as comSUM effectively address field-based retrieval tasks. However, they are constrained by fixed formulas and ranking score calculations. And they often rely on pre-defined or equal-weighted scores that lack adaptability. Essentially, aggregation functions as a simple regression-based approach, and the scoring function depends on the specific aggregation technique used. For instance, comSUM calculates the final score by summing individual field-based scores.
Re-ranking with LTR can be seen as an extension of aggregation methods, as LTR algorithms refine results from the previous local ranking layers. Unlike traditional aggregation, LTR provides a broader selection of algorithms and greater flexibility. Its parameters are fine-tuned during model training, enabling more adaptive and effective ranking performance.

5.2. Ablation Results

We conducted extensive ablation experiments to evaluate the impact of three key components in MultiLTR. Specifically, MultiLTR w/o feature clustering removes the feature clustering module and groups similar features to enhance data representation. MultiLTR w/o local ranking eliminates the local ranking module and captures feature relationships to improve ranking performance. MultiLTR w/o normalization excludes the normalization module and standardizes feature scales to ensure equal contribution to the model’s performance.
The results in Table 9 show that these components contribute to positive improvements for MultiLTR. Feature clustering has a moderate impact on the performance of MultiLTR, with the most significant improvements observed in the RF setting. The local ranking has the most significant impact on the performance of MultiLTR, with noticeable improvements across all LTR algorithms and evaluation metrics. Compared to the other two components, normalization has a smaller impact on the performance of MultiLTR. However, it still contributes to the overall performance.

5.3. Time Consumption

Having established the effectiveness of the proposed method, it is equally important to evaluate its computational efficiency. We assess the time required to construct a ranking model for each method and present it in Figure 5. Compared to the fLTR method, the MultiLTR approach demands more computational time. We conducted all experiments and evaluations on a machine equipped with an Intel i7-10700K CPU, 64GB of memory, and an Nvidia A5000 GPU, DELL, USA.

6. Conclusions and Future Work

This research introduces a MultiLTR approach and evaluates its effectiveness in text ranking tasks. We performed experiments on the MQ2007 and MQ2008 benchmark datasets and utilized a diverse set of learning-to-rank algorithms. The results demonstrate that MultiLTR consistently outperforms state-of-the-art baselines. ListNet, RankNet, and random forest are selected to construct the local re-ranking rankers. For all three local re-ranker algorithms, the top-performing global re-rankers are random forest and MART. They are both pointwise-based methods and consistently yield the best results. The results highlight that incorporating multi-stage ranking significantly enhances overall ranking performance.
MultiLTR proves superior to single-stage field-based LTR methods, underscoring the benefits of multiple re-ranking stages. However, its increased computational cost presents an opportunity for further optimization. This work lays the foundation for integration with neural ranking architectures through a structured three-tier approach [63]. First, we will augment MultiLTR’s feature representation by hybridizing BERT-derived contextual embeddings with traditional features via attention-gated fusion modules, preserving backward compatibility while enhancing semantic expressiveness. Second, a reranking system will deploy efficiency-optimized BERT variants (e.g., distilled cross-encoders) on top-K candidates, coupled with asynchronous batch processing to balance latency and relevance. Finally, knowledge distillation will bridge BERT’s document interaction insights into MultiLTR’s core ranking model through attention pattern transfer and hybrid loss alignment. Additionally, to bridge the gap between LTR frameworks and neural ranking paradigms, we will adapt MultiLTR’s multi-granularity optimization framework to transformer-based models. A potential approach involves employing MultiLTR to dynamically aggregate BERT-derived relevance signals at varying semantic levels. This integration could leverage attention-based gating mechanisms to weigh the contributions of features at different granularities. Moreover, we will rigorously validate this hybrid paradigm on both standardized large-scale benchmarks (MS MARCO, TREC DL) and industrial search systems. This ensures generalizability across diverse real-world scenarios. This systematic validation will quantify robustness against dataset shifts and operational constraints, bridging theoretical advances with practical deployment requirements.

Author Contributions

Conceptualization, methodology, software, data curation, validation, writing—original draft, funding acquisition, project administration, H.Y.; conceptualization, formal analysis, resources, supervision, writing—review and editing, T.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by the Key Scientific Research Project of Higher Education Institutions in Henan Province, (grant No. 24A520060).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available upon request.

Acknowledgments

We are thankful to Shilong Li for his work in parts of the experiments in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LTRlearning-to-rank
MultiLTRmulti-stage learning-to-rank
fLTRfield learning-to-rank
DCGdiscounted cumulative gain
IDCGIdea discounted cumulative gain
NDCGnormalized discounted cumulative gain
APaverage precision
MAPmean average precision
MRMART
RFrandom forest
RBRankBoost as RB
RNRankNet
LMLambdaMART
LNListNet
CAcoordinate ascent
ARAdaRank

References

  1. Yang, H.; Gonçalves, T. Field features: The impact in learning to rank approaches. Appl. Soft Comput. 2023, 138, 110183. [Google Scholar] [CrossRef]
  2. Clarke, C.L.; Culpepper, J.S.; Moffat, A. Assessing efficiency–effectiveness tradeoffs in multi-stage retrieval systems without using relevance judgments. Inf. Retr. J. 2016, 19, 351–377. [Google Scholar] [CrossRef]
  3. Zhang, L.; Zhang, Y.; Long, D.; Xie, P.; Zhang, M.; Zhang, M. A two-stage adaptation of large language models for text ranking. In Proceedings of the Findings of the Association for Computational Linguistics ACL 2024, Bangkok, Thailand, 11–16 August 2024; pp. 11880–11891. [Google Scholar]
  4. Zheng, K.; Zhao, H.; Huang, R.; Zhang, B.; Mou, N.; Niu, Y.; Song, Y.; Wang, H.; Gai, K. Full stage learning to rank: A unified framework for multi-stage systems. In Proceedings of the ACM Web Conference 2024, Singapore, 13–17 May 2024; pp. 3621–3631. [Google Scholar]
  5. Liu, Z.; Li, C.; Xiao, S.; Li, C.; Lian, D.; Shao, Y. Matryoshka Re-Ranker: A Flexible Re-Ranking Architecture with Configurable Depth and Width. arXiv 2025, arXiv:2501.16302. [Google Scholar]
  6. Nogueira, R.; Yang, W.; Cho, K.; Lin, J. Multi-stage document ranking with BERT. arXiv 2019, arXiv:1910.14424. [Google Scholar]
  7. Fan, Y.; Xie, X.; Cai, Y.; Chen, J.; Ma, X.; Li, X.; Zhang, R.; Guo, J. Pre-training methods in information retrieval. Found. Trends® Inf. Retr. 2022, 16, 178–317. [Google Scholar] [CrossRef]
  8. Lu, J.; Hall, K.; Ma, J.; Ni, J. HYRR: Hybrid Infused Reranking for Passage Retrieval. arXiv 2022, arXiv:2212.10528. [Google Scholar]
  9. Wang, B.; Li, M.; Zeng, Z.; Zhuo, J.; Wang, S.; Xu, S.; Long, B.; Yan, W. Learning Multi-Stage Multi-Grained Semantic Embeddings for E-Commerce Search. arXiv 2023, arXiv:2303.11009. [Google Scholar]
  10. Huang, P.S.; He, X.; Gao, J.; Deng, L.; Acero, A.; Heck, L. Learning deep structured semantic models for web search using clickthrough data. In Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, San Francisco, CA, USA, 27 October–1 November 2013; pp. 2333–2338. [Google Scholar]
  11. Zhang, H.; Wang, S.; Zhang, K.; Tang, Z.; Jiang, Y.; Xiao, Y.; Yan, W.; Yang, W.Y. Towards personalized and semantic retrieval: An end-to-end solution for e-commerce search via embedding learning. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 2407–2416. [Google Scholar]
  12. Qiu, Y.; Zhao, C.; Zhang, H.; Zhuo, J.; Li, T.; Zhang, X.; Wang, S.; Xu, S.; Long, B.; Yang, W.Y. Pre-training Tasks for User Intent Detection and Embedding Retrieval in E-commerce Search. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 4424–4428. [Google Scholar]
  13. Hai Le, N.; Gerald, T.; Formal, T.; Nie, J.Y.; Piwowarski, B.; Soulier, L. CoSPLADE: Contextualizing SPLADE for Conversational Information Retrieval. In Proceedings of the European Conference on Information Retrieval, Dublin, Ireland, 2–6 April 2023; Springer: Cham, Switzerland, 2023; pp. 537–552. [Google Scholar]
  14. Yang, D.; Zhang, Y.; Fang, H. An exploration study of mixed-initiative query reformulation in conversational passage retrieval. arXiv 2023, arXiv:2307.08803. [Google Scholar]
  15. Gao, L.; Dai, Z.; Callan, J. Rethink training of BERT rerankers in multi-stage retrieval pipeline. In Advances in Information Retrieval: Proceedings of the 43rd European Conference on IR Research, ECIR 2021, Virtual Event, 28 March–1 April 2021, Proceedings, Part II 43; Springer: Cham, Switzerland, 2021; pp. 280–286. [Google Scholar]
  16. Nogueira, R.; Cho, K. Passage Re-ranking with BERT. arXiv 2019, arXiv:1901.04085. [Google Scholar]
  17. Yilmaz, Z.A.; Wang, S.; Yang, W.; Zhang, H.; Lin, J. Applying BERT to document retrieval with birch. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): System Demonstrations, Hong Kong, China, 3–7 November 2019; pp. 19–24. [Google Scholar]
  18. Lin, S.C.; Yang, J.H.; Nogueira, R.; Tsai, M.F.; Wang, C.J.; Lin, J. Multi-stage conversational passage retrieval: An approach to fusing term importance estimation and neural query rewriting. ACM Trans. Inf. Syst. (TOIS) 2021, 39, 48. [Google Scholar] [CrossRef]
  19. Guo, J.; Fan, Y.; Pang, L.; Yang, L.; Ai, Q.; Zamani, H.; Wu, C.; Croft, W.B.; Cheng, X. A deep look into neural ranking models for information retrieval. Inf. Process. Manag. 2020, 57, 102067. [Google Scholar] [CrossRef]
  20. Craswell, N.; Mitra, B.; Yilmaz, E.; Campos, D.; Lin, J. Ms marco: Benchmarking ranking models in the large-data regime. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 1566–1576. [Google Scholar]
  21. Yates, A.; Nogueira, R.; Lin, J. Pretrained transformers for text ranking: BERT and beyond. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, Virtual, 8–12 March 2021; pp. 1154–1156. [Google Scholar]
  22. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  23. Leonhardt, J.; Beringer, F.; Anand, A. Exploiting Sentence-Level Representations for Passage Ranking. arXiv 2021, arXiv:2106.07316. [Google Scholar]
  24. Mitra, B.; Diaz, F.; Craswell, N. Learning to match using local and distributed representations of text for web search. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 1291–1299. [Google Scholar]
  25. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  26. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  27. Ahmadi, K.; Gathwala, A.; Osajima, J.; Hsiao, D.; Das, P. SLLIM-Rank: A Multi-Stage Item-to-Item Recommendation Model using Learning-to-Rank. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington, DC, USA, 15–18 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 2264–2268. [Google Scholar]
  28. Lee, J.; Bernier-Colborne, G.; Maharaj, T.; Vajjala, S. Methods, Applications, and Directions of Learning-to-Rank in NLP Research. In Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico, 16–21 June 2024; pp. 1900–1917. [Google Scholar]
  29. Dato, D.; MacAvaney, S.; Nardini, F.M.; Perego, R.; Tonellotto, N. The Istella22 Dataset: Bridging Traditional and Neural Learning to Rank Evaluation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 3099–3107. [Google Scholar]
  30. Qin, Z.; Yan, L.; Zhuang, H.; Tay, Y.; Pasumarthi, R.K.; Wang, X.; Bendersky, M.; Najork, M. Are neural rankers still outperformed by gradient boosted decision trees? In Proceedings of the ICLR’2021, Virtual, 3–7 May 2021. [Google Scholar]
  31. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Statist. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  32. Burges, C.; Shaked, T.; Renshaw, E.; Lazier, A.; Deeds, M.; Hamilton, N.; Hullender, G. Learning to rank using gradient descent. In Proceedings of the 22nd International Conference on Machine Learning, Bonn, Germany, 7–11 August 2005; pp. 89–96. [Google Scholar]
  33. Xu, J.; Li, H. Adarank: A boosting algorithm for information retrieval. In Proceedings of the 30th annual international ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; ACM: New York, NY, USA, 2007; pp. 391–398. [Google Scholar]
  34. Metzler, D.; Croft, W.B. Linear feature-based models for information retrieval. Inf. Retr. 2007, 10, 257–274. [Google Scholar] [CrossRef]
  35. Wu, Q.; Burges, C.J.; Svore, K.M.; Gao, J. Adapting boosting for information retrieval measures. Inf. Retr. 2010, 13, 254–270. [Google Scholar] [CrossRef]
  36. Cao, Z.; Qin, T.; Liu, T.Y.; Tsai, M.F.; Li, H. Learning to rank: From pairwise approach to listwise approach. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; ACM: New York, NY, USA, 2007; pp. 129–136. [Google Scholar]
  37. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  38. Köppel, M.; Segner, A.; Wagener, M.; Pensel, L.; Karwath, A.; Kramer, S. Pairwise learning to rank by neural networks revisited: Reconstruction, theoretical analysis and practical performance. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Würzburg, Germany, 16–20 September 2019; Springer: Cham, Switzerland, 2019; pp. 237–252. [Google Scholar]
  39. Jia, Y.; Wang, H.; Guo, S.; Wang, H. Pairrank: Online pairwise learning to rank by divide-and-conquer. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 146–157. [Google Scholar]
  40. Yuan, K.; Kuang, D. Deep Pairwise Learning To Rank For Search Autocomplete. arXiv 2021, arXiv:2108.04976. [Google Scholar]
  41. Ai, Q.; Bi, K.; Guo, J.; Croft, W.B. Learning a deep listwise context model for ranking refinement. In Proceedings of the 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 135–144. [Google Scholar]
  42. Sharma, A. Listwise Learning to Rank with Deep Q-Networks. arXiv 2020, arXiv:2002.07651. [Google Scholar]
  43. Pang, L.; Xu, J.; Ai, Q.; Lan, Y.; Cheng, X.; Wen, J. Setrank: Learning a permutation-invariant ranking model for information retrieval. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 499–508. [Google Scholar]
  44. Swezey, R.; Grover, A.; Charron, B.; Ermon, S. PiRank: Scalable Learning To Rank via Differentiable Sorting. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2021; Volume 34. [Google Scholar]
  45. Chen, Z.; Eickhoff, C. PoolRank: Max/Min Pooling-based Ranking Loss for Listwise Learning & Ranking Balance. arXiv 2021, arXiv:2108.03586. [Google Scholar]
  46. Keshvari, S.; Ensan, F.; Yazdi, H.S. ListMAP: Listwise learning to rank as maximum a posteriori estimation. Inf. Process. Manag. 2022, 59, 102962. [Google Scholar] [CrossRef]
  47. Chen, F.; Fang, H. An Exploration of Learning-to-re-rank Using a Two-step Framework for Fair Ranking. In Proceedings of the TREC, Online, 15–19 November 2022. [Google Scholar]
  48. Han, S.; Wang, X.; Bendersky, M.; Najork, M. Learning-to-Rank with BERT in TF-Ranking. arXiv 2020, arXiv:2004.08476. [Google Scholar]
  49. Awan, Z.; Kahlke, T.; Ralph, P.; Kennedy, P. Bi-Encoders based Species Normalization–Pairwise Sentence Learning to Rank. arXiv 2023, arXiv:2310.14366. [Google Scholar]
  50. Richards, J.A. Feature reduction. In Remote Sensing Digital Image Analysis; Springer: Berlin/Heidelberg, Germany, 2022; pp. 403–446. [Google Scholar]
  51. Liu, T.Y. Learning to rank for information retrieval. Found. Trends® Inf. Retr. 2009, 3, 225–331. [Google Scholar] [CrossRef]
  52. Cabello-Solorzano, K.; Ortigosa de Araujo, I.; Pe na, M.; Correia, L.; Tallón-Ballesteros, A.J. The impact of data normalization on the accuracy of machine learning algorithms: A comparative analysis. In Proceedings of the International Conference on Soft Computing Models in Industrial and Environmental Applications, Salamanca, Spain, 5–7 September 2023; Springer: Cham, Switzerland, 2023; pp. 344–353. [Google Scholar]
  53. Järvelin, K.; Kekäläinen, J. Cumulated gain-based evaluation of IR techniques. ACM Trans. Inf. Syst. (TOIS) 2002, 20, 422–446. [Google Scholar] [CrossRef]
  54. Qin, T.; Liu, T.Y. Introducing LETOR 4.0 datasets. arXiv 2013, arXiv:1306.2597. [Google Scholar]
  55. Qin, T.; Liu, T.Y.; Xu, J.; Li, H. LETOR: A benchmark collection for research on learning to rank for information retrieval. Inf. Retr. 2010, 13, 346–374. [Google Scholar] [CrossRef]
  56. Zhang, E.; Zhang, Y. Average Precision. In Encyclopedia of Database Systems; Springer: Boston, MA, USA, 2009; pp. 192–193. [Google Scholar] [CrossRef]
  57. Sanderson, M. Christopher D. Manning, Prabhakar Raghavan, Hinrich Schütze, Introduction to Information Retrieval, Cambridge University Press. 2008. ISBN-13 978-0-521-86571-5, xxi+ 482 pages. Nat. Lang. Eng. 2010, 16, 100–103. [Google Scholar] [CrossRef]
  58. Beitzel, S.M.; Jensen, E.C.; Frieder, O. MAP. In Encyclopedia of Database Systems; Springer: Boston, MA, USA, 2009; pp. 1691–1692. [Google Scholar] [CrossRef]
  59. Xiong, C.; Dai, Z.; Callan, J.; Liu, Z.; Power, R. End-to-end neural ad-hoc ranking with kernel pooling. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 55–64. [Google Scholar]
  60. Fan, Y.; Guo, J.; Lan, Y.; Xu, J.; Zhai, C.; Cheng, X. Modeling diverse relevance patterns in ad-hoc retrieval. In Proceedings of the 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 375–384. [Google Scholar]
  61. Tang, Z.; Yang, G.H. Deeptilebars: Visualizing term distribution for neural information retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 289–296. [Google Scholar]
  62. Fox, E.A.; Shaw, J.A. Combination of multiple searches. In NIST Special Publications SP; National Institute of Standards and Technology: Gaithersburg, MD, USA, 1994; Volume 243. Available online: https://trec.nist.gov/pubs/trec2/t2_proceedings.html (accessed on 12 February 2025).
  63. Askari, A.; Abolghasemi, A.; Pasi, G.; Kraaij, W.; Verberne, S. Injecting the score of the first-stage retriever as text improves BERT-based re-rankers. Discov. Comput. 2024, 27, 15. [Google Scholar] [CrossRef]
Figure 1. MultiLTR architecture overview. Key stages: Input → Feature clustering → Initial ranking → Local re-ranking (noise filtering) → Normalization → Dynamic model selection → Global re-ranking → Mid-evaluation → Final re-ranking → Output.
Figure 1. MultiLTR architecture overview. Key stages: Input → Feature clustering → Initial ranking → Local re-ranking (noise filtering) → Normalization → Dynamic model selection → Global re-ranking → Mid-evaluation → Final re-ranking → Output.
Information 16 00308 g001
Figure 2. Data flow in the MultiLTR approach. Key stages: Input → Feature Clustering → Initial Ranking → Local Re-ranking → Normalization → Global Re-ranking → Mid-Evaluation → Final Re-ranking → Output.
Figure 2. Data flow in the MultiLTR approach. Key stages: Input → Feature Clustering → Initial Ranking → Local Re-ranking → Normalization → Global Re-ranking → Mid-Evaluation → Final Re-ranking → Output.
Information 16 00308 g002
Figure 3. NDCG performance of MultiLTR vs. baselines. The local stage employs three optimal LTR algorithms (LN: pointwise, RN: pairwise, RF: listwise). The global stage integrates eight diverse LTR algorithms. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Figure 3. NDCG performance of MultiLTR vs. baselines. The local stage employs three optimal LTR algorithms (LN: pointwise, RN: pairwise, RF: listwise). The global stage integrates eight diverse LTR algorithms. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Information 16 00308 g003
Figure 4. MAP Performance of MultiLTR vs. baselines. The local stage employs three optimal LTR algorithms (LN: pointwise, RN: pairwise, RF: listwise). The global stage integrates eight diverse LTR algorithms. Models are assessed using MAP@k at cutoffs of 5, 10, 15, 20, and 30.
Figure 4. MAP Performance of MultiLTR vs. baselines. The local stage employs three optimal LTR algorithms (LN: pointwise, RN: pairwise, RF: listwise). The global stage integrates eight diverse LTR algorithms. Models are assessed using MAP@k at cutoffs of 5, 10, 15, 20, and 30.
Information 16 00308 g004
Figure 5. Comparison of time consumption between the fLTR and MultiLTR methods.
Figure 5. Comparison of time consumption between the fLTR and MultiLTR methods.
Information 16 00308 g005
Table 1. Key notations for the MultiLTR approach.
Table 1. Key notations for the MultiLTR approach.
NotationMeaning
Tdata collection
ta document in D
qa query in D
foriginal and unclustered features provided or extracted from a data collection T
sthe total number of features extracted from a data collection T
dthe number of fields where features are extracted from
kthe number of features included in a field
f j ( i ) the feature j from the field i
r m the ranker of the mth layer
Hfinal result
Table 2. Statistics for the MQ2007 and MQ2008 datasets.
Table 2. Statistics for the MQ2007 and MQ2008 datasets.
DatasetSelected FeaturesQueriesLabeled Query–Document Pairs
MQ200740170069,623
MQ20084080015,211
Table 3. Selected and discarded fields and features. ∘: selected feature, ⊗: rejected features, −: not provided in the dataset.
Table 3. Selected and discarded fields and features. ∘: selected feature, ⊗: rejected features, −: not provided in the dataset.
FeatureField
BodyAnchorTitleURLWholedocOther Place
TF
IDF
TF × IDF
DL
BM25
LMIR.ABS
LMIR.DIR
LMIR.JM
Length of URL
Number of slash in URL
Number of child page
Number of inlinks
Number of outlinks
PageRank
Table 4. Normalization NDCG@k performance on MQ2007 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Table 4. Normalization NDCG@k performance on MQ2007 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Alg.Features@5@10@15@20@30Alg.Features@5@10@15@20@30
MRbaseline0.41500.44290.47210.50300.5604RNbaseline0.39510.42430.45600.48700.5457
body0.39580.42430.45510.48560.5436 body0.34160.37340.40530.43820.5030
anchor0.34770.37640.41090.44150.5064 anchor0.34570.37770.41220.44330.5094
title0.40730.43450.46250.49350.5533 title0.40540.43440.46640.49680.5545
url0.40550.43290.46390.49310.5500 url0.39880.42980.46300.49340.5504
wholedoc0.39770.42860.45760.48900.5495 wholedoc0.36180.39770.43210.46390.5254
RBbaseline0.40460.43330.46650.49680.5527ARbaseline0.40130.43000.46010.49090.5480
body0.36820.39910.42990.46080.5215 body0.34390.37170.40520.43860.5044
anchor0.34540.38230.41450.44760.5111 anchor0.33080.36400.39580.42680.4958
title0.40480.43530.46800.49800.5539 title0.30230.33530.36720.40040.4657
url0.40180.43270.46670.49630.5519 url0.30790.34750.38650.42130.4893
wholedoc0.34050.37960.41530.44850.5122 wholedoc0.30940.34510.38060.41570.4837
CAbaseline0.40870.43860.47040.49970.5560LMbaseline0.41970.44780.47810.50840.5642
body0.36560.39220.42220.45710.5194 body0.39190.42100.45460.48630.5440
anchor0.34440.37980.41450.44620.5105 anchor0.34680.37830.41310.44650.5095
title0.40710.43450.46690.49570.5531 title0.40620.43640.46550.49510.5530
url0.40590.43710.46880.49850.5545 url0.40460.43520.46670.49540.5523
wholedoc0.35340.38950.42580.45960.5223 wholedoc0.40300.43530.46610.49680.5537
LNbaseline0.38900.41850.44820.47950.5390RFbaseline0.41290.43890.46940.49950.5569
body0.34120.36500.39640.43040.4965 body0.39410.42120.45110.48180.5423
anchor0.33820.37330.40360.43490.5022 anchor0.34440.37250.40760.43850.5048
title0.39570.42520.45420.48490.5432 title0.40650.43420.46130.49010.5496
url0.38760.42020.45330.48540.5428 url0.40360.43270.46290.49210.5485
wholedoc0.39390.42090.45250.48460.5453 wholedoc0.33520.36870.40260.43600.5033
Table 5. Normalization NDCG@k performance on MQ2008 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Table 5. Normalization NDCG@k performance on MQ2008 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Alg.Features@5@10@15@20@30Alg.Features@5@10@15@20@30
MRbaseline0.45860.50360.51790.52570.5328RNbaseline0.43450.48430.50100.50720.5157
body0.44690.49430.50770.51650.5257 body0.40550.45750.47710.48580.4948
anchor0.42990.48120.49890.50610.5153 anchor0.41220.46880.48590.49400.5036
title0.44960.49770.50960.51670.5258 title0.44460.49310.50840.51610.5254
url0.46160.50860.52290.53080.5383 url0.46040.50760.52140.52990.5375
wholedoc0.45170.50050.51550.52300.5308 wholedoc0.42240.47260.49100.49880.5085
RBbaseline0.45500.50030.51570.52300.5301ARbaseline0.43640.48500.50040.50750.5161
body0.40550.45750.47710.48580.4948 body0.40850.46670.48390.49120.5011
anchor0.42510.47630.49330.50110.5118 anchor0.40990.46640.48420.49340.5032
title0.45210.50200.51530.52250.5307 title0.36180.41180.42910.43560.4470
url0.46810.51070.52390.53220.5403 url0.35500.41640.44090.45060.4636
wholedoc0.44810.49220.50890.51620.5255 wholedoc0.43430.48130.49570.50490.5140
CAbaseline0.45530.50160.51500.52240.5304LMbaseline0.46190.50500.51800.52670.5338
body0.42890.47920.49600.50290.5127 body0.44570.49440.51080.51840.5279
anchor0.43510.48510.50190.51010.5196 anchor0.42670.47790.49630.50440.5127
title0.46860.51420.52840.53500.5432 title0.45870.50140.51630.52400.5331
url0.46920.51340.52670.53480.5435 url0.46120.50630.52120.53080.5386
wholedoc0.45090.49700.51100.51820.5277 wholedoc0.45320.50020.51640.52480.5318
LNbaseline0.43410.48510.50110.50800.5156RFbaseline0.45130.49850.51270.52000.5282
body0.40800.45850.47780.48690.4957 body0.44590.49850.51230.52010.5298
anchor0.41280.46780.48550.49340.5030 anchor0.42770.47780.49660.50400.5137
title0.43460.48640.50290.51030.5189 title0.45560.50400.51700.52480.5340
url0.45450.50380.51820.52620.5348 url0.46300.50890.52410.53290.5401
wholedoc0.41770.47280.48980.49650.5067 wholedoc0.45870.50540.52120.52800.5365
Table 6. Normalization NDCG@k performance on MQ2008 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Table 6. Normalization NDCG@k performance on MQ2008 after normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
MRRFRNRBLMARCALN
MQ20075171010401015
MQ2008517107401015
Table 7. Comparison between MultiLTR and fLTR on MQ2007 with NDCG@k, with cutoffs at 5, 10, 15, 20, and 30.
Table 7. Comparison between MultiLTR and fLTR on MQ2007 with NDCG@k, with cutoffs at 5, 10, 15, 20, and 30.
LTR Alg.Method@5@10@15@20@30
Neural ModelKNRM0.37900.41200.42560.43090.4324
HiNT0.46300.49000.51020.52530.5358
DeepTileBars0.39800.43400.45070.46050.4651
LNBM250.38900.41850.44820.47950.5390
fLTR0.39570.41930.44820.47960.5413
MultiLTR0.42250.44600.47560.50630.5645
RNBM250.39510.42430.45600.48700.5457
fLTR0.40510.42980.46030.48930.5504
MultiLTR0.42430.44770.47790.50720.5649
RFBM250.41290.43890.46940.49950.5569
fLTR0.42860.44890.47750.50630.5657
MultiLTR0.44940.46470.49200.52070.5804
Table 8. Comparison between MultiLTR fLTR on MQ2007 using MAP@k, with cutoffs at 5, 10, 15, 20, and 30.
Table 8. Comparison between MultiLTR fLTR on MQ2007 using MAP@k, with cutoffs at 5, 10, 15, 20, and 30.
LTR Alg.Method@5@10@15@20@30
Neural ModelKNRM0.15670.22540.26830.29410.3193
HiNT0.19160.26830.32180.35830.3953
DeepTileBars0.16470.23770.28400.31400.3430
LNBM250.16100.22920.28240.32730.3981
fLTR0.15660.22320.27480.31870.3897
MultiLTR0.16980.24320.29960.34650.4168
RNBM250.14610.21370.26700.30990.3822
fLTR0.16500.23440.28870.33230.4039
MultiLTR0.17540.24760.30370.35010.4210
RFBM250.16720.24320.30060.34810.4202
fLTR0.17100.24550.30200.34890.4212
MultiLTR0.18290.25690.31480.36310.4340
Table 9. Ablation study on the impact of feature clustering, local ranking, and normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
Table 9. Ablation study on the impact of feature clustering, local ranking, and normalization. Models are assessed using NDCG@k at cutoffs of 5, 10, 15, 20, and 30.
LTR Alg.Model@5@10@15@20@30
LNMultiLTR0.42250.44600.47560.50630.5645
MultiLTR w/o feature clustering0.41810.44510.47220.53100.5600
MultiLTR w/o local ranking0.41120.43520.46500.49500.5550
MultiLTR w/o normalization0.42020.44410.47300.50300.5620
RNMultiLTR0.42430.44770.47790.50720.5649
MultiLTR w/o feature clustering0.42010.44300.47300.50300.5620
MultiLTR w/o local ranking0.41500.44010.47000.50120.5601
MultiLTR w/o normalization0.42200.44600.47610.50600.5630
RFMultiLTR0.44940.46470.49200.52070.5804
MultiLTR w/o feature clustering0.44510.46000.48700.51700.5770
MultiLTR w/o local ranking0.43500.45500.48010.51110.5703
MultiLTR w/o normalization0.44710.46200.48900.51900.5780
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, H.; Gonçalves, T. MultiLTR: Text Ranking with a Multi-Stage Learning-to-Rank Approach. Information 2025, 16, 308. https://doi.org/10.3390/info16040308

AMA Style

Yang H, Gonçalves T. MultiLTR: Text Ranking with a Multi-Stage Learning-to-Rank Approach. Information. 2025; 16(4):308. https://doi.org/10.3390/info16040308

Chicago/Turabian Style

Yang, Hua, and Teresa Gonçalves. 2025. "MultiLTR: Text Ranking with a Multi-Stage Learning-to-Rank Approach" Information 16, no. 4: 308. https://doi.org/10.3390/info16040308

APA Style

Yang, H., & Gonçalves, T. (2025). MultiLTR: Text Ranking with a Multi-Stage Learning-to-Rank Approach. Information, 16(4), 308. https://doi.org/10.3390/info16040308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop