This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Pseudo Relevance Feedback with Deep Language Models and Dense Retrievers: Successes and Pitfalls

Hang Li hang.li@uq.edu.au 0000-0002-5317-7227 Ahmed Mourad a.mourad@uq.edu.au 0000-0002-9423-9404 Shengyao Zhuang s.zhuang@uq.edu.au 0000-0002-6711-0955 IElab, The University of QueenslandSt. LuciaQueenslandAustralia Bevan Koopman bevan.koopman@csiro.au 0000-0001-5577-3391 Australian E-Health Research Centre, CSIROHerstonQueenslandAustralia  and  Guido Zuccon g.zuccon@uq.edu.au 0000-0003-0271-5563 IElab, The University of QueenslandSt. LuciaQueenslandAustralia
(2018)
Abstract.

Pseudo Relevance Feedback (PRF) is known to improve the effectiveness of bag-of-words retrievers. At the same time, deep language models have been shown to outperform traditional bag-of-words rerankers. However, it is unclear how to integrate PRF directly with emergent deep language models. This article addresses this gap by investigating methods for integrating PRF signals with rerankers and dense retrievers based on deep language models. We consider text-based, vector-based and hybrid PRF approaches and investigate different ways of combining and scoring relevance signals. An extensive empirical evaluation was conducted across four different datasets and two task settings (retrieval and ranking).

Text-based PRF results show that the use of PRF had a mixed effect on deep rerankers across different datasets. We found that the best effectiveness was achieved when (i) directly concatenating each PRF passage with the query, searching with the new set of queries, and then aggregating the scores; (ii) using Borda to aggregate scores from PRF runs.

Vector-based PRF results show that the use of PRF enhanced the effectiveness of deep rerankers and dense retrievers over several evaluation metrics. We found that higher effectiveness was achieved when (i) the query retains either the majority or the same weight within the PRF mechanism, and (ii) a shallower PRF signal (i.e., a smaller number of top-ranked passages) was employed, rather than a deeper signal. Our vector-based PRF method is computationally efficient; thus, this represents a general PRF method others can use with deep rerankers and dense retrievers.

Pseudo Relevance Feedback, Dense Retrievers, Pre-trained Language Models for Information Retrieval, BERT
copyright: acmcopyrightjournalyear: 2018doi: 10.1145/1122445.1122456ccs: Information systems Retrieval models and rankingccs: Information systems Query reformulationccs: Information systems Query representation

1. Introduction

Pseudo Relevance Feedback (PRF) assumes the top-ranked passages from any phase of retrieval contain relevant signals and thus modifies the query by exploiting these signals in a bid to reduce the effect of query-passage vocabulary mismatch and improve search effectiveness (Azad and Deepak, 2019). Previous research has considered PRF in the context of traditional bag-of-words retrieval models such as probabilistic (Robertson and Zaragoza, 2009), vector space (Rocchio, 1971), and language models (Lv and Zhai, 2014; Zhai and Lafferty, 2001; Lavrenko and Croft, 2017). PRF methods such as Rocchio (Rocchio, 1971), relevance models (Lavrenko and Croft, 2017), RM3 (Lv and Zhai, 2009), and KL expansion models (Zhai and Lafferty, 2001) analyse the top-ranked passages to expand the query or to modify the query weights. The query and the passage are represented as either text or vectors, hence the categorisation of text-based and vector-based PRF approaches hereafter. Empirically, these approaches improve the initial retrieval effectiveness (Azad and Deepak, 2019).

Recently, Transformer (Vaswani et al., 2017) based deep language models (Devlin et al., 2019; Yang et al., 2019a; Dai et al., 2019; Radford et al., 2019; Raffel et al., 2020) have been adopted with promising results in information retrieval (Yates et al., 2021; Guo et al., 2020). Seminal in this context is the work of Nogueira and Cho (2019) who fine tuned BERT (Devlin et al., 2019) as a reranker on top of BM25. In this article, we investigate how to integrate PRF signals, effective for bag-of-words models, with deep language model rerankers, e.g. BERT (other models such as RoBERTa (Liu et al., 2019), query likelihood models (Zhuang et al., 2021; Zhuang and Zuccon, 2021; dos Santos et al., 2020) can be applied as well), and dense retrievers, (specifically ANCE (Xiong et al., 2020), RepBERT (Zhan et al., 2020), TCT-ColBERT V1 (Lin et al., 2020), TCT-ColBERT V2 HN+ (Lin et al., 2021c), DistilBERT KD (Hofstätter et al., 2020), DistilBERT Balanced (Hofstätter et al., 2021), and SBERT (Reimers et al., 2019)); extensive evaluations are done towards the proposed PRF methods along the side.

Our experiments investigate two alternative paths to integrate PRF signals with deep language models: text-based and vector-based. The text-based PRF approach is an obvious direction as the concatenation of the query text and the PRF passages text is used as the new formulated query to feed into the deep language model (e.g., BERT). However, this approach has two significant impediments: (i) the lengthy concatenated text would often exceed the allowed input size (input vector length) of these deep language models (Yang et al., 2019b; Ding et al., 2020) and (ii) it is computationally expensive or infeasible as it requires additional deep language model inferences at query time (Jiao et al., 2020; Liu et al., 2020). To solve the first challenge, we propose three different text handling methods to generate text partitions from the full concatenated text such that each of the partitions is within the length limit of the deep language models. Furthermore, because we split the concatenated text into partitions, we also propose three different score aggregation methods (Average, Borda, and Max) to aggregate the scores from each partition to calculate the final scores for each passage.

To address the computational complexity challenge, we use model pre-generated embeddings to represent text (Khattab and Zaharia, 2020; Zhan et al., 2020; Xiong et al., 2020; Qu et al., 2020). Query latency is reduced to the time of generating the query embeddings because the passages embeddings are pre-generated. In the context of PRF, we further utilise these pre-generated passages embeddings to efficiently integrate the relevance signals while eliminating the input size limit of deep language models, which we refer to as vector-based PRF approach. Each feedback passage is pre-generated as embeddings (vectors) in this approach. We adopt two different vector fusion methods (Average and Rocchio) to integrate the feedback vectors into the query vectors. The Rocchio method has two parameters: the query vector and the feedback passage vector weights. We empirically investigate the influence of query and feedback passages through weighting within the Rocchio PRF approach.

To evaluate these PRF approaches, we use the TREC Deep Learning Track Passage Retrieval Task (TREC DL) 2019 (Craswell et al., 2020) and 2020 (Craswell et al., 2021), the TREC Conversational Assistance Track 2019 (TREC CAsT) (Dalton et al., 2020), the Web Answer Passages (WebAP) (Keikha et al., 2014), and the DL HARD (Mackie et al., 2021). TREC CAsT and WebAP are used for the passage retrieval task rather than their original tasks (e.g., for CAsT, we do not consider the multi-turn conversational relationship between queries).

For the text-based PRF approach, we find that our models significantly outperform the baselines across several evaluation metrics on TREC DL 2019 while having mixed results on TREC DL 2020, TREC CAsT and WebAP. For DL HARD, the proposed approach does not have any significant improvements. The results suggest that TREC DL 2019 queries are easier – the results from the initial ranking contain less noise – hence, the PRF can add more relevant information to the queries. On the other hand, the queries of TREC DL 2020, TREC CAsT, and WebAP are more challenging —the results from the initial ranking contain more noise—hence, adding these PRF signals into the query will cause query drift and lead to worse performance. DL HARD is created by selecting queries from TREC DL 2019 and TREC DL 2020 based on the performance systems at TREC had (i.e., select queries for which systems cannot perform well) and the characteristics of the queries (Mackie et al., 2021). Our results show that text-based PRF did not work on DL HARD, suggesting that the feedback passages do not contain relevant signals or more noise than valuable signals.

Another challenge for text-based PRF is its computational complexity for the full ranking pipeline. It requires at least two inferences, depending on the text partitioning method. At least, this doubles the total run time compared to that of deep language rerankers without PRF.

For the vector-based PRF approach, we find that our models improve the respective baselines (seven dense retrievers) across all evaluation metrics and all datasets for the retrieval task; the proposed approach also outperforms the strong BM25+BERT ranker across several metrics. This result suggests that encoding the PRF feedback passages into embedding vectors better models the relevance signals exploited by the PRF mechanism. Unlike text-based PRF, the passage vectors are pre-generated and indexed, so the inference steps on passages are not required at retrieval or rerank time. This makes vector-based PRF very efficient: it takes only 1/20th of the time of the BM25+BERT reranker and only about double the time of the simple bag-of-words BM25. In addition, since our proposed approach works directly with the vector embeddings of queries and passages, they can be applied on top of any choice of dense retriever. For the reranking task, we find that our models outperform the BM25 and BM25+RM3 baselines across all metrics and datasets, while they have only mixed improvements over the strong BM25+BERT reranker. Overall, the vector-based PRF approach for retrieval tends to improve deep metrics, while for reranking, they tend to improve shallow metrics.

To summarize in this article we make the following contributions:

  • We thoroughly investigate the PRF effectiveness under different conditions, in particular how sensitive the effectiveness is to PRF depth, text handling/vector fusion, and score estimation;

  • We conduct a thorough comparison of text-based and vector-based approaches within the same reranking task;

  • We conduct a thorough comparison of different vector-based approaches within the same retrieval task;

  • We study the efficiency of the proposed text-based and vector-based PRF approaches.

2. Related Work

Pseudo-Relevance Feedback (PRF) is a classic query expansion method that modifies the original query in an attempt to address the mismatch between the query intent and the query representation (Clinchant and Gaussier, 2013; Wang et al., 2020b). A typical PRF setting uses the top-ranked passages from a retrieval system as the relevant signal to select query terms to add to the original query or to set the weights for the query terms. PRF approaches including Rocchio (Rocchio, 1971), KL expansion (Lv and Zhai, 2014; Zhai and Lafferty, 2001), query-regularized mixture model (Tao and Zhai, 2006), RM3 (Lv and Zhai, 2009), relevance-feedback matrix factorization (Zamani et al., 2016), and relevance models (Lavrenko and Croft, 2017) have been well studied. The use of PRF on top of efficient bag-of-words retrieval models is common in information retrieval systems, and it is an effective strategy for improving retrieval effectiveness (Clinchant and Gaussier, 2013; Lavrenko and Croft, 2017). Traditional PRF approaches (Rocchio, 1971; Zhai and Lafferty, 2001; Lv and Zhai, 2009; Lavrenko and Croft, 2017) are simple, but more effective, robust and generalisable, in comparison to more complex models (Tao and Zhai, 2006; Zamani et al., 2016), which instead achieve marginal gains, may be harder to implement/reproduce or maybe problematic to instantiate across different datasets or domains from those in which they have been originally evaluated. This study employs the most popular PRF method in existing research (RM3) as a baseline. The RM3 considers the original query and the feedback passages when creating a new query by assigning different weights to the original query and feedback terms. RM3 is effective and robust compared to other query expansion methods (Lv and Zhai, 2009), and it is used as a baseline in several pseudo relevance feedback studies (Miao et al., 2012; Lv and Zhai, 2010; Cao et al., 2008).

Recent research has studied PRF in different settings. Lin (2019) considered document ranking as a binary classification problem, combining PRF with text classification by introducing positive and negative pseudo labels. The positive pseudo labels are obtained from the top-kk documents, while the negative labels are from the bottom-nn documents. The final score is a linear interpolation of the classifier and retriever scores. Li et al. (2018) proposed a neural PRF framework, which was further extended by Wang et al. (2020a), that utilises a feed-forward neural network to determine the target document’s relevance score by aggregating the target query and the target feedback relevance scores. However, these proposed models have achieved marginal improvements over the BM25 baseline. Furthermore, the efficiency (run time) of the proposed models have not been reported, and thus it is difficult to establish whether these marginal improvements in effectiveness may be at the cost of efficiency.

Deep language models based on transformers (Vaswani et al., 2017), such as BERT (Devlin et al., 2019), T5 (Raffel et al., 2020), and RoBERTa (Liu et al., 2019), has surpassed the existing state-of-the-art effectiveness in different search tasks. BERT, in specific, has shown to improve over previous state-of-the-art for ad hoc retrieval (Nogueira and Cho, 2019). Recent research has also considered integrating PRF with deep language models. Padaki et al. (2020) integrated RM3 with BERT. The results, however, showed that the selection of highly weighted terms from the feedback passages via RM3 to expand the original query could significantly hurt the ranking quality of a fine-tuned BERT reranker. Yu et al. (2021a) presented a framework that integrates PRF into a Graph-based Transformer (PGT). It represents each feedback passage as a node, and the PRF signals are captured using sparse attention between graph nodes. While this approach handles the input-size limit of deep language models, it achieves marginal improvements compared to the BERT reranker approach across most evaluation metrics at the cost of efficiency. Specifically, compared to our results, PGT achieves a lower nDCG@10 than our simplest text-based PRF reranking approach; it also achieves lower effectiveness in reranking and similar effectiveness in retrieval than our vector-based PRF with ANCE, but at a much higher computational cost.

Wang et al. (2020b) argued that existing PRF research mainly considers relevance matching where terms are used to sort feedback documents. On the contrary, they propose a model that considers both relevance and semantic matching. The relevance score is obtained using BM25. For semantic matching, they split the top-kk PRF documents into sentences. For each sentence, they use BERT to estimate the semantic similarity with the query. Scores from the top-mm sentences of each document are considered as the semantic score for this document. The final scores of each document are calculated from a linear interpolation of the relevance and semantic scores. The expansion terms are then extracted from the reranked top-kk PRF documents and added to the original query for a second retrieval stage. Although the improvements are marginal, they demonstrate that BERT can identify relevance signals from the feedback documents at the sentence level to enhance retrieval effectiveness. However, this marginal improvement is at the expense of efficiency because expansion terms are identified through BERT.

Zheng et al. (2020, 2021) presented a three-phase BERT-based query expansion model: BERT-QE. The first phase is a standard BERT reranking (Nogueira and Cho, 2019) step. In the second phase, the top-kk passages are selected as feedback passages, further split into overlapping partitions using a sliding window. Together with the original query, these partitions are fed into BERT to get the top-mm partitions with the highest scores per passage. The top-mm partitions and the candidate passage are fed into BERT in the third phase. The score of a candidate passage in this phase is calculated as a weighted sum, where the weight is the relevance score of each partition in the top-mm partitions from phase two, and the score is the relevance score between the top-mm partitions and the candidate passage. The final score of a candidate passage is calculated by linear interpolation of the first phase BERT relevance score between the query and the passage, and the third phase weighted sum score between the top-mm partitions and the candidate passages. Although BERT-QE achieves significant improvements in effectiveness over BERT reranker, it requires 11.01x more computations than BERT, making it computationally infeasible in many practical applications.

Recently, Yu et al. (2021b) proposed a PRF framework based on ANCE (Xiong et al., 2020), which trains a new query encoder from ANCE that takes in the top-kk passages from the first-round ANCE retrieval, then concatenate the passage texts with the original query text to form the new PRF query, without changing ANCE’s passage encoder. This newly formed PRF query is passed to the trained query encoder to produce the PRF query representation, and to retrieve the results from the original passage collection index. Although the improvements are significant across several datasets over different metrics, according to a recent reproducibility paper from Li et al. (2022c), the proposed model does not generalise well to other dense retriever models and the training process needs to be adjusted accordingly with different dense retriever models, which makes it difficult to achieve the same effectiveness as the one proposed in the original paper.

Integrating PRF signals with deep language models implies a trade-off between effectiveness and efficiency. While current approaches ignored efficiency, the majority still achieved marginal improvements in effectiveness. In this study, we propose three approaches to integrate PRF signals to improve effectiveness while maintaining efficiency: (i) by concatenating the feedback passages text with the original query to form the new queries that contain the relevant signals, (ii) by pre-generating passage collection embeddings and performing PRF in the vector space, because embeddings promise to capture the semantic similarity between terms (Dalton et al., 2019; Diaz et al., 2016; Kuzi et al., 2016; Naseri et al., 2018; Roy et al., 2016; Zamani and Croft, 2016, 2017; Naseri et al., 2021), which makes it feasible as a method for first stage retrieval as well, (iii) by combining the previous two approaches into a hybrid approach.

3. Methodology

3.1. Text-Based Pseudo-Relevance Feedback

BERT is computationally expensive to be applied as a first-stage retriever. Hence, it is commonly employed as a reranker that considers only a subset of the initial retrieval results (usually top 1000). In this approach, we integrate the text-based PRF signal with the BERT reranker. Padaki et al. (2020) demonstrated that the use of RM3 (Lv and Zhai, 2009) to select highly weighted terms from the feedback passages and construct the new PRF queries significantly hurts the ranking quality of a fine-tuned BERT reranker. Therefore, we use the full passages text to construct the new PRF queries. We address the challenge of the input size limit of BERT by employing three text-based PRF methods:

  1. (1)

    Concatenate and Truncate: append the query and the top-kk feedback passages, then truncate to the length of 256 tokens. BERT has an input size limit of 512 tokens; we allocate 256 tokens to the new query and the remaining tokens are left to concatenate the candidate passage.

  2. (2)

    Concatenate and Aggregate: append the query to each of the top-kk feedback passages to form kk new queries. For each new query, use BERT to perform another rerank, resulting in kk new ranked lists. The final scores for the candidate passages are generated using different score estimation methods that combine the kk ranked lists (but not the ranked list of the original query).

  3. (3)

    Sliding Window: concatenate the top-kk passages, then use a sliding window to split the aggregated text into overlapping partitions . Concatenate the query with each partition to create jj new queries, then follow the same steps as Concatenate and Aggregate.

Methods 2 and 3 require the aggregation of multiple ranker lists to estimate the scores and obtain the final ranked list. For this, we aggregate the scores of a candidate passage using several methods:

  1. (1)

    Average: calculate the average of all the scores per candidate passage.

  2. (2)

    Max: consider only the highest score per candidate passage.

  3. (3)

    Borda: employ the Borda voting rule (Aslam and Montague, 2001; Macdonald and Ounis, 2008) to calculate the score of each candidate passage.

Refer to caption
Figure 1. The proposed architecture for integrating Text-based Pseudo-Relevance Feedback with BERT reranker. The initial retriever is a traditional bag-of-words BM25.

Figure 1 depicts the proposed architecture for integrating text-based PRF signals with BERT reranker. The initial retriever is a traditional bag-of-words BM25 followed by BERT reranker. As shown in step 1, the query is passed to BM25 to retrieve the initial ranked results from the inverted index. Then the query text and initial retrieval results are passed to BERT for reranking ( 2). The top-kk feedback passages from the reranked list are used as PRF relevance signals ( 4), after mapping them back to their text representation ( 3). Then, the query and feedback passage texts are combined together to form new query texts ( 5) followed by another BERT-based scoring step ( 6), and finally the individual scores are aggregated per candidate passage to form the final ranking ( 7). The core components of this architecture, which are PRF with Text Handling and Score Estimation, are described in the next two sections.

3.1.1. Text-Based PRF with Text Handling

We consider three different approaches to handle the text length that exceeds the BERT input size limit:

Concatenate and Truncate (CT)

A new query text is generated by concatenating the original query text with the top-kk feedback passage texts, separated by a space (     ). If the length of the new query exceeds 256, it will be truncated to the first 256 tokens. Then, we run the new query through BERT reranker. The new query is constructed as follows:

(1) Qnew,l<=256=Qoriginal+     +p1++     +pk256Q_{new,l<=256}=\lceil Q_{original}+\mbox{\kern 0.59998pt\vrule height=1.29167pt}\vbox{\hrule width=8.00003pt}\hbox{\vrule height=1.29167pt}+p_{1}+...+\mbox{\kern 0.59998pt\vrule height=1.29167pt}\vbox{\hrule width=8.00003pt}\hbox{\vrule height=1.29167pt}+p_{k}\rceil_{256}

where QnewQ_{new} and QoriginalQ_{original} represent the new query text and the original query text, respectively. l<=256l<=256 represents the input size limit enforced, which is achieved by truncating the sequence (denoted with .256\lceil.\rceil_{256}). p1,,pkp_{1},...,p_{k} represent the top-kk feedback passages from the BERT reranker.       is the space in between.

Concatenate and Aggregate (CA)

This approach generates kk new queries by concatenating the original query text with each of the top-kk feedback passage texts, separated by a space (     ). Then, each of the new queries is run through another BERT reranking step resulting into kk scores per candidate passage, which will be aggregated later to estimate the final score. The new queries are generated as follows:

(2) Q1,new=Qoriginal+     +p1Qk,new=Qoriginal+     +pk\begin{split}Q_{1,new}=Q_{original}+\mbox{\kern 0.59998pt\vrule height=1.29167pt}\vbox{\hrule width=8.00003pt}\hbox{\vrule height=1.29167pt}+p_{1}\\ ...\\ Q_{k,new}=Q_{original}+\mbox{\kern 0.59998pt\vrule height=1.29167pt}\vbox{\hrule width=8.00003pt}\hbox{\vrule height=1.29167pt}+p_{k}\end{split}

where Q1,new,,Qk,newQ_{1,new},...,Q_{k,new} represent the kk new queries. QoriginalQ_{original} represents the original query text. p1,..,pkp_{1},..,p_{k} represent the top-kk feedback passage texts.       is the separation token in between.

Sliding Window (SW)

In this approach, the top-kk feedback passage texts are appended together, then a sliding window is applied to split the text into jj overlapping partitions with different window size and stride according to different datasets’ passage lengths (Dai and Callan, 2019), as below:

(3) p1++pkSWp1,,pjp_{1}+...+p_{k}\xrightarrow[]{SW}p_{1},...,p_{j}

where p1,,pkp_{1},...,p_{k} represent the top-kk feedback passage texts, SWSW represents the sliding window mechanism, p1,,pjp_{1},...,p_{j} represent the jj partitions. Similar to the CA approach, the set of jj new queries is generated using Eq. 2.

Note that after generating each new query, the query/passage pair may exceed the BERT input size limit for the CT approach. Under this situation, if the length of the new query exceeds 256, we truncate the new query down to be of length 256. For CA and SW approaches, we also applied the same methodology to guarantee all the new queries are below the length of 256.

3.1.2. Text-Based with Score Estimation

CA and SW text-handling approaches generate kk and jj scores per candidate passage, respectively. To estimate a final score for each candidate passage, we consider the following estimation methods.

Average

The final score is estimated by calculating the mean of all scores:

(4) Sfinal=Avg(S1+S2++Sk)S_{final}=Avg(S_{1}+S_{2}+...+S_{k})

where SfinalS_{final} represents the final ranking score for each candidate passage, and S1,,SkS_{1},...,S_{k} represent the kk ranking scores for each candidate passage based on each of the kk new queries. For the rest of this paper, we refer to this method as Text-Average, represented by T-A for brevity.

Max

The final score is estimated by taking the highest score per candidate passage:

(5) Sfinal=Max(S1,S2,,Sk)S_{final}=Max(S_{1},S_{2},...,S_{k})

where SfinalS_{final} represents the maximum score for each candidate passage, and S1,,SkS_{1},...,S_{k} represent the kk ranking scores for each candidate passage based on each of the kk new queries. For the rest of this paper, we refer to this method as Max, represented by M for brevity.

Borda

The final score is estimated by using the Borda voting algorithm. The score of a candidate passage w.r.t a ranked list is the number of candidate passages in the ranked list that are ranked lower. Scores are summed over ranked lists as follows:

(6) Sfinal=Li:pLinrLi(p)+1nS_{final}=\sum_{L_{i}:p\in L_{i}}\frac{n-r_{L_{i}}(p)+1}{n}

where LiL_{i} represents the ii-thth ranked list produced using the ii-thth new query, pp represents the candidate passage, rr is the rank of the candidate passage, and nn represents the number of candidate passages in the ranked list. For the rest of this paper, we refer to this method as Borda, represented by B for brevity.

3.2. Vector-Based Pseudo-Relevance Feedback

Using existing, efficient first stage dense retrievers(RepBERT (Zhan et al., 2020), ANCE (Xiong et al., 2020), TCT-ColBERT V1 (Lin et al., 2020), TCT-ColBERT V2 HN+ (Lin et al., 2021c), DistilBERT KD (Hofstätter et al., 2020), DistilBERT Balanced (Hofstätter et al., 2021), and SBERT (Reimers et al., 2019)), we employ two vector-based PRF methods for the retrieval task:

  1. (1)

    Average: the mean of the original query embeddings and the feedback passage embeddings are used to generate the new query representation.

  2. (2)

    Rocchio: different weights are assigned to the original query embeddings and the feedback passage embeddings following the intuition provided by the original Relevance Feedback mechanism proposed by Rocchio (Rocchio, 1971).

Refer to caption
Figure 2. The proposed architecture for integrating Vector-based Pseudo-Relevance Feedback with Deep Language Model dense retrievers for the retrieval task.

Figure 2 depicts the proposed architecture for integrating vector-based PRF signals with deep language model dense retrievers. A single deep language model is used to generate offline the embeddings for all passages, which are then stored in a Faiss index (Johnson et al., 2019). The deep language model is also used to generate the query embedding at inference time (step 1). The query embedding is then passed to the dense retriever that exploits the Faiss index to perform the first pass of retrieval to obtain the initial ranked list ( 2). The top-kk feedback passage embeddings from the initial ranked list are used as PRF relevance signals ( 3), using vector operations, and are then used to perform the subsequent retrieval to get the final ranked list ( 4).

We describe the two proposed vector-based PRF approaches in the next two sections.

3.2.1. Vector-Based PRF with Average

A new query embedding is generated by averaging the original query embedding and the top-kk feedback passage embeddings. The intuition is to treat the original query at par of the signal from the top-kk feedback passages (i.e., the query weights as much as each passage). The new query embedding is computed as follows:

(7) EQnew=Avg(E(Qoriginal),E(p1),,E(pk))E_{Q_{new}}=Avg(E(Q_{original}),E(p_{1}),...,E(p_{k}))

where EE represents the embeddings of either the query or the feedback passage, EQnewE_{Q_{new}} represents the newly formulated query embeddings. We do not generate an actual text query in the vector-based PRF approaches: only the embedding of the new query is generated. QoriginalQ_{original} represents the original query, p1,,pkp_{1},...,p_{k} represent the top-kk passages retrieved by the first stage ranker. In the remainder of the paper we refer to this method as Vector Average, represented by V-A for brevity.

3.2.2. Vector-Based PRF with Rocchio

This method is inspired by the original Rocchio method for relevance feedback (Rocchio, 1971) but adapted to deep language models. The intuition is to transform the original query embedding towards the average of the top-kk feedback passage embeddings by assigning different weights to query and (the combination of) feedback passages, thus controlling the contribution of each component toward the final score. Unlike in the original version of Rocchio, in this work we do not model the PRF with non-relevant passages: hence the negative portion of Rocchio is omitted. We note that this could be extended by identifying which passages in the initial ranked list could represent a negative relevance signal (e.g., the bottom passages) – however we leave this for future consideration.

Thus, our Rocchio PRF approach consists of interpolating the query embedding and the average PRF embedding:

(8) EQnew=αE(Qoriginal)+βAvg(E(p1),,E(pk))E_{Q_{new}}=\alpha*E(Q_{original})+\beta*Avg(E(p_{1}),...,E(p_{k}))

where α\alpha controls the weight assigned to the original query embedding and β\beta the weight assigned to the PRF signal. In the remainder of the paper we refer to this method as Rocchio, represented by 𝒞α\mathcal{RC}_{\alpha} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} for brevity.

3.3. Hybrid Pseudo-Relevance Feedback

Text-based PRF is a computationally expensive approach for the reranking task, in our experiments, the BERT inference step is executed twice: one before the PRF, one after the PRF. On the other hand, vector-based PRF is an efficient approach for the retrieval task because of the high efficiency of the dense retriever models. In this section, we investigate a hybrid approach where the architecture of vector-based PRF in Figure 2 is adapted to the reranking task. The main difference is that the initial ranking of passages is obtained from an inverted-index (Text-based) multi-stage pipeline such as BM25+BERT (as in Figure 1). In particular, the initial retrieval results obtained through steps 1 and 2 in Figure 2 are replaced by steps 1, 2, 3, and 4 in Figure 1. The ranked list of passages produced by the BERT reranker is mapped to embeddings using the Faiss index before applying the vector-based PRF methods.

4. Experimental Setup

4.1. Datasets

Our experiments use the TREC Deep Learning Track Passage Retrieval Task 2019 (Craswell et al., 2020) (DL 2019) and 2020 (Craswell et al., 2021) (DL 2020), DL HARD (Mackie et al., 2021), the TREC Conversational Assistance Track 2019 (Dalton et al., 2020) (CAsT 2019), and the Web Answer Passages (WebAP) (Keikha et al., 2014). The detailed statistics for each dataset are listed in Table 1.

Table 1. Statistics of the four datasets considered in our experiments. Where #Q represents the number of queries, #P represents the number of passages in the collection, Avg Len represents the average length of passages, Avg #J/Q represents the average number of judged passages per query, and #J represents the number of judged passages in total.
#Q #P Avg Len Avg #J/Q #J
TREC DL 2019 43 8,841,823 64.7 215.3 9,260
TREC DL 2020 54 8,841,823 64.7 210.9 11,386
TREC CAsT 2019 502 38,618,941 68.6 63.2 31,713
WebAP 80 1,959,777 74.5 11858.8 948,700
DL HARD 50 8,841,823 64.7 85.1 4,256

TREC DL 2019 and 2020 contain 200 queries each. However for 2019, only 43 queries have judgements; and thus the remaining 157 queries without judgements are discarded from our evaluation. In 2020, only 54 queries have judgements; and thus the remaining 146 queries are similarly discarded. The relevance judgements for both datasets range from 0 (not relevant) to 3 (highly relevant). The passage collection is the same as the MS MARCO passage ranking dataset (Nguyen et al., 2016), which is a benchmark English dataset for ad-hoc retrieval tasks with \approx8.8M passages. The difference between TREC DL and MS MARCO is that queries in TREC DL have several judgements per query (215.3/210.9 on average for 2019/2020), instead of an average of one judgement per query for MS MARCO. The very sparse relevance judgements of MS MARCO would not be able to provide detailed, reliable information on the behaviour of the PRF approaches and thus we do not report them in this article. However, we still tried to apply our vector-based PRF for the retrieval task on MS MARCO dev set, which consists of 6,980 queries. We refer the reader to our github page for the full results.111https://github.com/castorini/pyserini/blob/master/docs/experiments-vector-prf.md

DL HARD builds upon the TREC DL 2019/2020 queries: these queries are considered as hard queries on which previous methods do not perform well, and new judgements are provided for the added new queries (originally unjudged in TREC) (Mackie et al., 2021). While TREC CAsT 2019 is originally constructed for multi-turn conversational search, we treat each turn independently, and we use the manually rewritten topic utterances. WebAP is built from the TREC 2004 Terabyte Track collection, and it contains 80 queries222In addition to two queries without relevance judgements, which are excluded in our experiments and about 2 Million passages (1,1858.8 judged passages per query, on average). The relevance judgements for TREC CAsT 2019 and WebAP ranged from 0 (not relevant) to 4 (highly relevant).

4.2. Evaluation Metrics

We employ MAP, nDCG@{1, 3, 10}, and Reciprocal Rank (RR)333If for a query no relevant passage is retrieved up to the considered standard cut-off (1,000), then we assign RR=0. for the reranking task on both text-based PRF and vector-based PRF. We select these metrics because they are the common measures reported for BERT based models and these datasets – thus allowing cross-comparison with previous and future work. For the retrieval task on vector-based PRF, we also report Recall@{1000}, but it is not considered for text-based PRF approaches because they are built on top of the BERT reranker where the Recall is limited by the initial retriever (BM25) to the top 1,000 passages. We report Recall for its diagnostic ability in informing whether a gain in e.g., MAP is produced because of a higher number of retrieved relevant passages, or because of a better ranking (i.e. ordering of the same number of relevant passages). For the TREC DL 2020 dataset, we follow the instructions from the organisers and consider the label binarized at relevance level 2 for all evaluation metrics. For all results, statistical significance is performed using two-tailed paired t-test.

4.3. Baselines

We consider the following baselines:

  • BM25: traditional first stage retriever, implemented using the Anserini toolkit (Yang et al., 2018) with its default settings.

  • BM25+RM3: RM3 pseudo relevance feedback method (Abdul-Jaleel et al., 2004) on top of BM25, as implemented in Anserini. We use this approach as a representative bag-of-words PRF method, since previous research has found alternative bag-of-words PRF approaches achieve similar effectiveness (Miao et al., 2012). We note that BM25+RM3 is a standard baseline for MS MARCO and TREC DL.

  • RepBERT (R): first stage dense retriever (Zhan et al., 2020). We use the implementation made available by the authors.

  • ANCE (A): first stage dense retriever (Xiong et al., 2020). We use the scripts provided by the authors for both data pre-processing and model implementation.

  • TCT-ColBERT V1, TCT-ColBERT V2 HN+, DistilBERT KD, DistilBERT Balanced, and SBERT: first stage dense retrievers employed to evaluate the generalisability of our hypotheses. We use the implementations provided in the pyserini toolkit (Lin et al., 2021a).

  • RepBERT+BERT (R+B): first stage dense retriever with an additional BERT reranker to rerank the initial results provided by RepBERT.

  • ANCE+BERT (A+B): first stage dense retriever with an additional BERT reranker to rerank the initial results provided by ANCE.

  • BM25+BERT (BB): A common two-stage reranker pipeline, first proposed by Nogueira and Cho (2019), where the initial stage is BM25, and BERT is used to rerank the results from BM25. BERT is fine-tuned on MS MARCO Passage Retrieval Dataset (Nguyen et al., 2016). In all of our experiments, we use the 12 layer uncased BERT-Base provided by Nogueira and Cho (2019), unless stated otherwise, and we simply refer to it as BERT. In Section 5.5 we also use BERT-Large for the efficiency analysis.

4.4. Applying PRF to Rerankers

Text-Based Pseudo-Relevance Feedback for Reranking. We refer to this approach as BB+PRF, where BB represents BM25+BERT. For the Sliding Window approach, we use the average passage length as the window size, and half of the window size as the stride. Details of the Sliding Window parameters for each dataset are shown in Table 2. We experiment by using the top k=1,3,5,10,15,20k=1,3,5,10,15,20 passages as pseudo relevance feedback.

Table 2. Window size and stride size of the Sliding Window PRF approach for each dataset.
window size stride
TREC DL 2019 65 32
TREC DL 2020 65 32
TREC CAsT 2019 69 34
WebAP 75 37
DL HARD 65 32

Vector-Based Pseudo-Relevance Feedback for Reranking. We consider the vector representations (embeddings) generated by RepBERT and ANCE to apply PRF as a second stage ranker, represented as BB+PRF-R and BB+PRF-A, where BB represents BM25+BERT, R represents RepBERT, and A represents ANCE. To achieve this, the top-kk passages IDs from BERT are mapped to their vector representations before estimating the final scores. For the Rocchio method, we experiment by assigning weights to the query and the feedback passage within the range of 0.1–1 with a step of 0.1. We experiment by using the top k=1,3,5,10k=1,3,5,10 passages as pseudo relevance feedback.

4.5. Applying PRF to Retrievers

We choose RepBERT (Zhan et al., 2020), ANCE (Xiong et al., 2020), TCT-ColBERT V1 (Lin et al., 2020), TCT-ColBERT V2 HN+ (Lin et al., 2021c), DistilBERT KD (Hofstätter et al., 2020), DistilBERT Balanced (Hofstätter et al., 2021), and SBERT (Reimers et al., 2019) as representative first stage dense retrievers because they achieve state-of-the-art effectiveness in previous work on MS MARCO. We note that a host of alternative first stage dense retrievers have been recently proposed, including stronger ones like RocketQA (Qu et al., 2020) and RocketQAv2 (Ren et al., 2021), but most of these retrievers consider more complex training procedures than those selected in this study. We further note that the implementation of the current best first stage dense retriever, RocketQAv2, has only just been made available and is based on PaddlePaddle (Ma et al., 2019), thus uses a setup that differs from ours and is not selected for simplicity. We expect that findings that apply for the dense retrievers we chose are likely to translate to other dense retrievers, like RocketQA and RocketQAv2.

For the dense retrievers, we utilise the Faiss toolkit (Johnson et al., 2019) to build the index and perform retrieval. We develop our PRF approaches on top of these dense retrievers. To be consistent with the original dense retriever models, we truncate the query tokens and passage tokens according to the original settings in their papers. For simplicity, we mainly investigate our proposed vector-based PRF models on top of ANCE and RepBERT; the rest of the models are only shown in Table 4 for validation purposes as well as a demonstration of the generalisability of our proposed models. Therefore, in the following sections, vector-based PRF with RepBERT as base model is represented by R+PRF-R, and with ANCE as base model is represented by A+PRF-A for the retrieval task.

4.6. Efficiency experiments

To measure the runtime of each method, we run our experiments on a Unix-based server with the Intel(R) Xeon(R) Gold 6132 CPU @ 2.60GHz for BM25 and BM25 + RM3. For dense retrievers and our approaches, we use a Unix-based server equipped with a single Tesla V100 SMX2 32GB GPU.

5. Results

The overarching research question we seek to answer with our experiments is: What are the successes and pitfalls of integrating PRF into deep language model rerankers and dense retrievers in terms of effectiveness and efficiency? Each of the following subsections addresses a more specific sub-question.

5.1. PRF Depth

RQ1: What is the impact of PRF depth on the effectiveness of reranking and retrieval? To answer this question, we vary the number of top-kk passages while displaying the distribution of results over other parameters (text handling and score estimation).

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 3. Impact of PRF depth on the effectiveness (y-axis) of BM25+BERT+PRF(BB+PRF) for the task of reranking, kk represents different PRF depths. Baseline BM25+BERT(BB) is marked with a dashed red line. PRF depth impacts the effectiveness of text-based reranking models negatively.
Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 4. Impact of PRF depth on the effectiveness (y-axis) of BM25+BERT+PRF-RepBERT(BB+PRF-R) and BM25+BERT+PRF-ANCE(BB+PRF-A) for the task of reranking, kk represents the different PRF depths. Baseline BM25+BERT(BB) is marked with a dashed red line. Increasing PRF depth tends to enhance the effectiveness of hybrid models over shallow metrics (RR, nDCG@{1,3}) for reranking.
Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 5. Impact of PRF depth on the effectiveness (y-axis) of ANCE+PRF+BERT(A+PRF+B) and RepBERT+PRF+BERT(R+PRF+B) for the task of reranking, kk represents the different PRF depths. Baseline ANCE+BERT(A+B) and RepBERT+BERT(R+B) are marked with a dash-dot blue line and a dashed red line respectively. Vector-based PRF with BERT reranker does not seem to improve the metrics significantly except MAP, across datasets and PRF depths.

5.1.1. Reranking with Different PRF Depths

Results of text-based PRF (BB+PRF) for reranking are shown in Figure 3. For TREC DL 2019, increased PRF depth is associated with a marginal improvement in effectiveness across most of the evaluation metrics, except for nDCG@10 and, to a minor extent, nDCG@3. On the other hand, increasing PRF depth decreases the effectiveness across the remaining datasets, and none of the PRF configurations is substantially better than the BB baseline.

Results of hybrid PRF models (BB+PRF-R and BB+PRF-A) are shown in Figure 4. For TREC DL 2019, increased PRF depth is associated with substantial improvements in RR and nDCG@1 using both BB+PRF-R and BB+PRF-A, and marginal improvements in nDCG at depths 3–10. For TREC DL 2020 and TREC CAsT, increased PRF depth is associated with marginal improvements in nDCG@{1,3}. For the remaining datasets, increased PRF depth shows mixed results, but overall it appears to decrease the effectiveness over all metrics. In addition, we report the results of hybrid PRF dense retrievers with a BERT reranker (R+PRF+B and A+PRF+B) in Figure 5. We observer that vector-based PRF models with BERT reranker either hurts RR or marginally improves it across all datasets with all PRF depths. On the other hand, all PRF approaches improve MAP across all datasets with PRF depths 3–5, with ANCE-based slightly better than RepBERT-based on TREC DL 2019, TREC DL 2020, RepBERT-based is slightly better than ANCE-based on TREC CAsT. Other datasets show similar effectiveness between these two over different PRF depths. For all other metrics, most of the highest effectivenesses are achieved with PRF depths 3–5, although the improvements are mostly marginal.

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 6. Impact of PRF depth on the effectiveness (y-axis) of RepBERT+PRF-RepBERT(R+PRF-R) and ANCE+PRF-ANCE(A+PRF-A) for the task of retrieval, kk represents different PRF depths. Baseline RepBERT(R) is marked with a dashed red line, ANCE(A) is marked with a dash-dot blue line. Increasing PRF depth tends to enhance the effectiveness over deep metrics (R@1000, nDCG@10 and MAP) for retrieval.

5.1.2. Retrieval with Different PRF Depths

Results of vector-based PRF (R+PRF-R and A+PRF-A) for retrieval are shown in Figure 6. For deep evaluation metrics (MAP, nDCG@10 and R@1000), increased PRF depth is associated with significant improvements in effectiveness over the baseline dense retrievers across all datasets, with few exceptions for DL HARD. Increased PRF depth is associated with decreased RR values across all datasets, with few exceptions for A+PRF-A where PRF at depth of 10 is on par or marginally better. For shallow metrics such as nDCG@{1, 3}, mixed impact across datasets is witnessed with respect to changing PRF depths. For TREC DL 2019 and 2020, PRF of depth 1 is on par with the baselines. For TREC CAsT and WebAP, increased PRF depth is associated with significant increases of effectiveness of A+PRF-A, while PRF of depth 1 enhances the effectiveness of R+PRF-R. For DL Hard, all PRF depths perform on par with the ANCE(A) baseline, while PRF of depth 1 performs on par with the RepBERT(R) baseline.

5.1.3. Summary

To summarize, increasing PRF depth tends to enhance the effectiveness of hybrid models over shallow metrics (RR, nDCG@{1,3}) for reranking, and deep metrics (R@1000, nDCG@10 and MAP) for retrieval. On the other hand, PRF depth negatively impacts the effectiveness of text-based reranking models. Vector-based PRF with BERT reranker does not seem to improve the metrics significantly except MAP, across datasets and PRF depths.

5.2. Text Handling

RQ2: What is the impact of text handling techniques on the effectiveness of reranking and retrieval? To answer this question, we vary the text handling techniques while displaying the distribution of results over other parameters (PRF depth and score estimation). We analyze the effectiveness under three text handling techniques: Concatenate and Truncate (CT), Concatenate and Aggregation (CA), and Sliding Window (SW); and two dense representations for text: RepBERT(R+PRF-R) and ANCE(A+PRF-A).

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 7. Impact of text handling on the effectiveness (y-axis) of PRF approaches for the task of reranking, where CT, CA and SW represent the text handling methods Concatenate and Truncate, Concatenate and Aggregate and Sliding Window, respectively, while BM25+BERT+PRF-RepBERT(BB+PRF-R) and BM25+BERT+PRF-ANCE(BB+PRF-A) are the dense representations for text. Baseline BM25+BERT(BB) is marked with a dashed red line. CA tends to improve more on MAP, but all other improvements are marginal and some significant losses can be observed.
Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 8. Impact of dense representations on the effectiveness (y-axis) of PRF approaches for the task of reranking, where R+PRF+B and A+PRF+B represents RepBERT+PRF+BERT and ANCE+PRF+BERT, respectively. Baseline ANCE+BERT(A+B) and RepBERT+BERT(R+B) are marked with a dash-dot blue line and a dashed red line respectively. When applying the BERT reranker after vector-based PRF, both ANCE-based and RepBERT-based improve MAP on all datasets, but there are no improvements nor losses on the remaining metrics across all datasets.

5.2.1. Reranking with Different Text Handling

Results are shown in Figure 7. For TREC DL 2019, CA substantially improves MAP, RR, and nDCG@1, and marginally improves nDCG@3. BB+PRF-A and BB+PRF-R substantially improve RR, while BB+PRF-A also substantially improves nDCG@1. On the other hand, BB+PRF-R is on par with the baseline over nDCG@1nDCG@1, and BB+PRF-A is on par with nDCG@{3, 10}. All other methods do not improve effectiveness. For TREC DL 2020, SW marginally improves nDCG@{1, 3}. BB+PRF-R is on par with the baseline for nDCG@1. All other methods do not improve over the baseline, and all methods, including SW and BB+PRF-R, hurt MAP.

For TREC CAsT 2019, unlike the previous datasets, no improvements can be observed for MAP and nDCG@10 across all methods. CT is on par with the baseline in terms of RR, and marginal improvements are present for nDCG@1. BB+PRF-A is on par with the baseline for nDCG@1. All other metrics are not improved when employing different text handling methods.

For WebAP, no substantial improvements are found, regardless of the metric, with the exception of nDCG@3, for which BB+PRF-A is on par with the baseline.

For DL HARD, all methods hurt MAP and RR. BB+PRF-A is on par with the baseline for nDCG@1. No substantial improvements on other metrics can be observed for the remaining methods.

The results for vector-based PRF models with BERT reranker are shown in Figure 8. For TREC DL 2019, TREC DL 2020, ANCE-based is better than RepBERT-based PRF models over MAP, RR, nDCG@10. For other datasets except DL HARD, RepBERT-based is better than ANCE-based w.r.t all metrics except nDCG@10. However, the improvements only occur with MAP on all datasets, although marginal on DL HARD. Both ANCE-based and RepBERT-based either hurts or on par with baseline on all other metrics across all datasets.

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 9. Vector-based PRF retrieval effectiveness (y-axis) by using different dense retrieval models RepBERT(R+PRF-R) and ANCE(A+PRF-A). Baseline RepBERT(R) is marked with dashed red line, ANCE(A) is marked with dash-dot blue line. A+PRF-A is, overall, a better representation, as it improves all metrics and outperforms all baselines. R+PRF-R performs worse than A+PRF-A. This is because RepBERT(R) baseline is worse than ANCE(A) baseline across most metrics, causing the top ranked results to contain less relevant passages compared to A: hence, the PRF mechanism receives a noisier relevance signal from the feedback passages.

5.2.2. Retrieval with Different Text Handling

Results are shown in Figure 9. For TREC DL 2019, both methods substantially outperform their respective baselines in terms of MAP, R@1000, and nDCG@10. No improvement can be observed for RR and nDCG@1. A+PRF-A does not outperform the baseline in terms of nDCG@3, but R+PRF-R does. For TREC DL 2020, both methods A+PRF-A and R+PRF-R substantially improve MAP and R@1000. On the other hand, they do not improve RR and nDCG@1. R+PRF-R is on par with the baseline for nDCG@{3, 10}. Marginal improvements can be observed for A+PRF-A in terms of nDCG@10.

For TREC CAsT 2019, both methods substantially improve the baseline in terms of MAP, R@1000, and nDCG@{3, 10}. Both A+PRF-A and R+PRF-R improve over the baseline in terms of nDCG@1, but A+PRF-A does so substantially; in addition A+PRF-A is on par with the baseline for RR. Both methods do not improve the baselines for other metrics.

For WebAP, similar trends can be observed for MAP, RR, and R@1000. R+PRF-R hurts the effectiveness over nDCG@{1, 3}, while A+PRF-A marginally improves nDCG@1 and substantially improves nDCG@3 and nDCG@10.

For DL HARD, both methods substantially improve MAP; R+PRF-R also substantially improves R@1000. A+PRF-A is on par with the baseline for RR, R@1000, nDCG@1, and marginally improves nDCG@10. No improvements are observed for the remaining metrics for either method.

5.2.3. Summary

When used for reranking, CA tends to improve more on MAP, BB+PRF-R tends to have more improvements for RR, and BB+PRF-A tends to improve more on nDCG@{1, 3, 10}. In general, all methods tend to improve more nDCG than RR or MAP. When applying the BERT reranker after vector-based PRF, both ANCE-based and RepBERT-based improve MAP on all datasets, but there are no improvements nor losses on the remaining metrics across all datasets.

When used for retrieval, A+PRF-A is, overall, a better representation, as it improves all metrics and outperforms all baselines. R+PRF-R performs worse than A+PRF-A. This is because RepBERT(R) baseline is worse than ANCE(A) baseline across most metrics, causing the top ranked results to contain less relevant passages compared to A: hence, the PRF mechanism receives a noisier relevance signal from the feedback passages.

5.3. Score Estimation

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 10. Reranking effectiveness (y-axis) by using different score estimation methods. Where T-A is Text Average, B is Borda, M is Max, V-A is Vector Average, 𝒞β\mathcal{RC}_{\beta} is Rocchio with fixed α\alpha value, and 𝒞α,β\mathcal{RC}_{\alpha,\beta} is Rocchio with α\alpha and β\beta. Baseline BM25+BERT(BB) is marked with dashed red line. 𝒞α,β\mathcal{RC}_{\alpha,\beta} is found to perform considerably well across all the metrics and datasets. B, T-A, and 𝒞β\mathcal{RC}_{\beta} also perform well across several metrics and all datasets. M performs poorly across all metrics and datasets.
Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 11. Reranking effectiveness (y-axis) by using different score estimation methods. Where A is Vector Average, 𝒞β\mathcal{RC}_{\beta} is Rocchio with fixed α\alpha value, and 𝒞α,β\mathcal{RC}_{\alpha,\beta} is Rocchio with α\alpha and β\beta. Baseline ANCE+BERT(A+B) and RepBERT+BERT(R+B) are marked with a dash-dot blue line and a dashed red line respectively. Significant improvements can be observed for MAP, with Average performing slightly better than the other two methods. Average also performs the best across the majority of the datasets and metrics, although the improvements where present are marginal compared to the baselines.

RQ3: What is the impact of score estimation methods on the effectiveness of reranking and retrieval? To answer this question, we vary the score estimation methods while displaying the distribution of results over other parameters (PRF depth and text handling). We analyze the effectiveness under three text-based score aggregation methods: Average (T-A), Borda (B) and Max (M); and three vector-based score fusion methods: Average (V-A), Rocchio with fixed α\alpha and varying β\beta (𝒞β\mathcal{RC}_{\beta}), and Rocchio with varying α\alpha and β\beta (𝒞α,β\mathcal{RC}_{\alpha,\beta}).

5.3.1. Reranking with Text Score Estimation and Vector Fusion

Results are shown in Figure 10. For TREC DL 2019, T-A outperforms the baseline in terms of MAP, RR, and nDCG@1, while B is only on par with the baseline for RR, and MM hurts effectiveness across all metrics. BB+PRF-A with V-A is on par with the baseline across all metrics, except for marginal improvements found for RR. BB+PRF-R with V-A only improves RR marginally. Both BB+PRF-A and BB+PRF-R with 𝒞β\mathcal{RC}_{\beta} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} substantially improve RR, while only BB+PRF-A with 𝒞α,β\mathcal{RC}_{\alpha,\beta} substantially improves RR and nDCG@1, and is on par with the baseline A for nDCG@{3, 10}.

For TREC DL 2020, no score estimation method can outperform the baseline in terms of MAP. B, BB+PRF-R with 𝒞β\mathcal{RC}_{\beta} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} are on par with the baseline for nDCG@{1, 3}. M is the worst estimation method for this dataset, as it only outperforms the baseline for nDCG@1, and all remaining methods decrease effectiveness.

For TREC CAsT 2019, M does not perform well across any metric, while T-A and B are on par with the baseline for nDCG@1. On the other hand, BB+PRF-A with V-A, 𝒞β\mathcal{RC}_{\beta}, and 𝒞α,β\mathcal{RC}_{\alpha,\beta} is on par with the baseline for nDCG@1.

For WebAP, all methods are worse than the baseline in terms of MAP. For RR, only BB+PRF-A with 𝒞β\mathcal{RC}_{\beta} is on par with the baseline for nDCG@3.

For DL HARD, all methods are worse than the baseline for MAP, RR, and nDCG@{3, 10}, except BB+PRF-A with 𝒞β\mathcal{RC}_{\beta} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} , which is on par with the baseline for nDCG@1.

The results detailed where applying BERT reranker after the vector-based PRF approaches are shown in Figure 11. Significant improvements for MAP over TREC DL 2019, TREC DL 2020, TREC CAsT, and WebAP can be observed with 𝒞β\mathcal{RC}_{\beta}, and 𝒞α,β\mathcal{RC}_{\alpha,\beta}. Average approach performs exceptionally well on TREC DL 2019 and TREC DL 2020 datasets. For all other metrics on all datasets, the majority of the improvements are marginal, and some statistically significant losses are observed. On the other hand, ANCE-based PRF with Average approach improves nDCG@1 significantly on TREC CAsT and DL HARD.

Refer to caption
(a) TREC DL 2019
Refer to caption
(b) TREC DL 2020
Refer to caption
(c) TREC CAsT
Refer to caption
(d) WebAP
Refer to caption
(e) DL HARD
Figure 12. Vector-based PRF retrieval effectiveness (y-axis) by using different vector fusion methods. Where A is Vector Average, 𝒞β\mathcal{RC}_{\beta} is Rocchio with fixed α\alpha value, and 𝒞α,β\mathcal{RC}_{\alpha,\beta} is Rocchio with α\alpha and β\beta. Baseline RepBERT(R) is marked with dashed red line, ANCE(A) is marked with dash-dot blue line. 𝒞α,β\mathcal{RC}_{\alpha,\beta} and 𝒞β\mathcal{RC}_{\beta} perform the best in most circumstances. A+PRF-A with these methods is more likely to improve nDCG at early cut-offs, while R+PRF-R with these methods is more likely to improve deep recall.

5.3.2. Retrieval with Vector Fusion

Results are shown in Figure 12. For TREC DL 2019, overall, A+PRF-A outperforms the baseline across MAP, R@1000, and nDCG@10, while it is worse than the baseline for RR, and nDCG@{1, 3}. R+PRF-R also substantially improves MAP, R@1000, and nDCG@10, and it is on par with the baseline for nDCG@3 when 𝒞α,β\mathcal{RC}_{\alpha,\beta} is used.

For TREC DL 2020, R+PRF-R performs exceptionally well in terms of R@1000, but both A+PRF-A and R+PRF-R do not outperform the respective baselines in terms of RR. On the other hand, A+PRF-A also outperforms the A+PRF-A baseline for R@1000, although the improvement is smaller than it was for R+PRF-R. All methods with A+PRF-A are on par with the baseline in terms of nDCG@10; a similar result is obtained for R+PRF-R, except that marginal improvements can be observed with 𝒞β\mathcal{RC}_{\beta}.

For TREC CAsT 2019, both A+PRF-A and R+PRF-R substantially outperform the respective baselines in terms of MAP. A+PRF-A achieves substantial improvements in terms of nDCG@{1, 3, 10}. Both A+PRF-A and R+PRF-R, combined with any of V-A, 𝒞α\mathcal{RC}_{\alpha} or 𝒞α,β\mathcal{RC}_{\alpha,\beta}, are either on par or worse than other metrics of the baseline and the dense retrievers without PRF (R and A).

For WebAP, both base models substantially improve MAP and R@1000, except R+PRF-R with A, which is on par with the baseline in terms of MAP. Overall, A+PRF-A is either on par with or improves the baseline across all metrics. On the other hand, R+PRF-R instead exhibits losses in terms of nDCG@{1, 3, 10}.

For DL HARD, A+PRF-A substantially improves MAP and nDCG@10 with 𝒞β\mathcal{RC}_{\beta}. R+PRF-R substantially improves MAP and R@1000 with 𝒞β\mathcal{RC}_{\beta}. All other metrics are either on par or worse than the baselines.

5.3.3. Summary

When the reranking task is considered, 𝒞α,β\mathcal{RC}_{\alpha,\beta} is found to perform considerably well across all the metrics and datasets. B, T-A, and 𝒞β\mathcal{RC}_{\beta} also perform well across several metrics and all datasets. M performs poorly across all metrics and datasets.

When adding the BERT reranker after the Vector-based PRF, significant improvements can be observed for MAP, with average performing slightly better than the other two methods. Average also performs the best across the majority of the datasets and metrics, although the improvements where present are marginal compared to the baselines.

When the retrieval task is considered, 𝒞α,β\mathcal{RC}_{\alpha,\beta} and 𝒞β\mathcal{RC}_{\beta} perform the best in most circumstances. A+PRF-A with these methods is more likely to improve nDCG at early cut-offs, while R+PRF-R with these methods is more likely to improve deep recall.

5.4. Effectiveness of PRF

Table 3. Results of PRF approaches for the tasks of reranking and retrieval across different datasets. For each parametric method, the settings that achieve optimal effectiveness over all metrics are reported. Statistical significance (paired t-test) with p<0.05p<0.05 between PRF models and BM25 is marked with a, between PRF models and BM25+RM3 is marked with b, between PRF and the corresponding baseline is marked with c, between Vector-Based PRF+BERT and Dense Retriever+BERT is marked with d (For these two we do not compare with other baselines). Best results with respect to each dataset and each metric are highlighted in bold.
Model MAP RR nDCG@1 nDCG@3 nDCG@10 R@1000
Trec DL 2019 BM25 .3773 .8245 .5426 .5230 .5058 .7389
BM25+RM3 .4270 .8167 .5465 .5195 .5180 .7882
BM25+BERT (BB) .4827 .9240 .6977 .7203 .7061 .7389
RepBERT+BERT (R+B) .4258 .9388 .7209 .7317 .6960 .6689
ANCE+BERT (A+B) .4315 .9388 .7209 .7371 .6965 .6610
RepBERT (R) .3311 .9243 .6589 .6256 .6100 .6689
ANCE (A) .3611 .9201 .7209 .6765 .6452 .6610
BB+PRF(k=10k=10,CA,BORDA) .4947ab .9826abc .7946abc .7528abc .7178ab
BB+PRF-R(k=10,β=.3k=10,\beta=.3) .4705a .9793ab .7326ab .6963ab .6993ab
BB+PRF-A(k=10,α=.3,β=.7k=10,\alpha=.3,\beta=.7) .4955ab .9690ab .7519ab .7385ab .7210ab
R+PRF+B(k=5,β=.5k=5,\beta=.5) .4463d .9388 .7209 .7371 .6968 .7097d
A+PRF+B(k=5,α=.4,β=.6k=5,\alpha=.4,\beta=.6) .4514d .9388 .7209 .7390 .6953 .6997d
R+PRF-R(k=10,β=.3k=10,\beta=.3) .3669c .9368 .7054c .6559abc .6252ab .7012bc
A+PRF-A(k=10,α=.4,β=.6k=10,\alpha=.4,\beta=.6) .4151c .9440a .7403ab .6807ab .6629ab .6962bc
Trec DL 2020 BM25 .2856 .6585 .5772 .5021 .4796 .7863
BM25+RM3 .3019 .6360 .5648 .4740 .4821 .8217
BM25+BERT (BB) .4926 .8531 .7901 .7598 .7064 .7863
RepBERT+BERT (R+B) .4358 .9082 .7099 .7276 .6715 .6593
ANCE+BERT (A+B) .4470 .9082 .7037 .7243 .6768 .6819
RepBERT (R) .3733 .8109 .7315 .6572 .6047 .7888
ANCE (A) .4076 .7907 .7346 .7082 .6458 .7764
BB+PRF(k=3k=3,SW,BORDA) .4644ab .8575ab .8179ab .7798ab .6739ab
BB+PRF-R(k=5,α=.4,β=.6k=5,\alpha=.4,\beta=.6) .4778ab .8638ab .8333ab .7544ab .7111ab
BB+PRF-A(k=1,α=.5,β=.5k=1,\alpha=.5,\beta=.5) .4606abc .8476ab .7963ab .7691ab .6984ab
R+PRF+B(k=3,α=.4,β=.6k=3,\alpha=.4,\beta=.6) .4530d .9050 .7099 .7320 .6750 .7022d
A+PRF+B(k=3,α=.4,β=.6k=3,\alpha=.4,\beta=.6) .4584d .9097 .7037 .7297 .6791 .7019d
R+PRF-R(k=1,α=.6,β=.4k=1,\alpha=.6,\beta=.4) .4239abc .7951ab .7315ab .6991ab .6393ab .8159c
A+PRF-A(k=3,α=.4,β=.6k=3,\alpha=.4,\beta=.6) .4341abc .8079ab .7407ab .7117ab .6598ab .7948
Trec CAsT BM25 .2936 .6502 .3631 .3542 .3526 .8326
BM25+RM3 .3132 .6556 .3971 .3829 .3817 .8246
BM25+BERT (BB) .3762 .8108 .5425 .5366 .5269 .8326
RepBERT+BERT (R+B) .3036 .7741 .4953 .5002 .4901 .6284
ANCE+BERT (A+B) .3007 .7665 .4855 .4998 .4890 .6179
RepBERT (R) .1969 .6604 .4307 .4087 .3752 .6284
ANCE (A) .2081 .6819 .4396 .4246 .3823 .6179
BB+PRF(k=10k=10,CC) .3247ac .8106ab .5510abc .5140ab .4838abc
BB+PRF-R(k=3,α=.5,β=.5k=3,\alpha=.5,\beta=.5) .3372ac .7985ab .5480ab .5468ab .5067abc
BB+PRF-A(k=3,α=.3,β=.7k=3,\alpha=.3,\beta=.7) .3274ac .8093ab .5914ab .5583ab .5055abc
R+PRF+B(k=5,α=.4,β=.6k=5,\alpha=.4,\beta=.6) .3162d .7722 .4915 .5023 .4909 .6635d
A+PRF+B(k=3,α=.3,β=.7k=3,\alpha=.3,\beta=.7) .3153d .7725 .4991 .5043 .4939 .6432d
R+PRF-R(k=10,α=.8,β=.2k=10,\alpha=.8,\beta=.2) .2150abc .6618 .4498a .4146a .3844c .6566abc
A+PRF-A(k=3,β=.9k=3,\beta=.9) .2347abc .6826 .4626a .4434abc .4138ac .6508abc
WebAP BM25 .0436 .3099 .1667 .1604 .1404 .2944
BM25+RM3 .0536 .2767 .1344 .1316 .1376 .3472
BM25+BERT (BB) .0845 .5856 .4042 .3356 .2897 .2944
RepBERT+B (R+B) .1088 .5859 .4042 .3361 .2939 .4133
ANCE+BERT (A+B) .1090 .5846 .4010 .3315 .2973 .3956
RepBERT (R) .0867 .4653 .2875 .2580 .2419 .4133
ANCE (A) .0886 .5107 .3469 .2863 .2638 .3956
BB+PRF(k=15k=15,CA,AVG) .0855ab .5459ab .3271ab .3444ab .2980ab
BB+PRF-R(k=1,α=.7,β=.3k=1,\alpha=.7,\beta=.3) .0809ab .5866ab .4198ab .3418ab .2708ab
BB+PRF-A(k=3,β=.8k=3,\beta=.8) .0790abc .5502ab .4083ab .3394ab .2842ab
R+PRF+B(k=3,α=.3,β=.7k=3,\alpha=.3,\beta=.7) .1146d .5880 .4042 .3383 .2995d .4253
A+PRF+B(k=5,α=.4,β=.6k=5,\alpha=.4,\beta=.6) .1134d .5790 .3885 .3308 .2996 .4202
R+PRF-R(k=3,β=.1k=3,\beta=.1) .0887abc .4690ab .2969ab .2594ab .2433ab .4206abc
A+PRF-A(k=3,β=.9k=3,\beta=.9) .0953abc .5134ab .3563ab .2928ab .2710ab .4027ab
DL-Hard BM25 .1845 .5422 .3533 .3137 .2850 .6288
BM25+RM3 .1925 .4381 .2467 .2508 .2555 .6522
BM25+BERT (BB) .2521 .6139 .4133 .4012 .3962 .6288
RepBERT+B (R+B) .2401 .6393 .4433 .4095 .3929 .6797
ANCE+BERT (A+B) .2386 .6405 .4433 .4150 .3934 .6564
RepBERT (R) .1576 .5489 .3200 .3263 .2982 .6797
ANCE (A) .1803 .5382 .3733 .3450 .3339 .6564
BB+PRF(k=3k=3,CA,BORDA) .2380a .5937b .4333b .3944b .3550abc
BB+PRF-R(k=5,α=.8,β=.2k=5,\alpha=.8,\beta=.2) .2255c .5843ab .3867b .3861b .3646abc
BB+PRF-A(k=5,α=.4,β=.6k=5,\alpha=.4,\beta=.6) .2422a .5904ab .4333b .4267ab .3968ab
R+PRF+B(k=5,β=.2k=5,\beta=.2) .2439d .6394 .4433 .4095 .3926 .6968d
A+PRF+B(k=5,α=.8,β=.2k=5,\alpha=.8,\beta=.2) .2419d .6405 .4433 .4150 .3941 .6683d
R+PRF-R(k=5,α=.9,β=.1k=5,\alpha=.9,\beta=.1) .1654 .5504b .3333 .3368 .3030 .6929c
A+PRF-A(k=10,β=.4k=10,\beta=.4) .1865 .5426 .3933b .3453 .3380b .6681c

RQ4: What is the impact of PRF models on the effectiveness of reranking and retrieval? To answer this question, we consider only our best performing PRF models with the optimal values for all the parameters combined. Results are presented in Table 3. For each dataset, the middle three rows represent PRF rerankers(BB+PRF, BB+PRF-R and BB+PRF-A), and the last two rows represent PRF retrievers (R+PRF-R and A+PRF-A). BB, R, and A are abbreviations for BM25+BERT, RepBERT, and ANCE, respectively. R@1000 is considered only for the evaluation of retrieval, mainly where it is infeasible to employ BERT for retrieval.

5.4.1. Reranking with Text-Based PRF (BB+PRF)

For TREC DL 2019, our model improves effectiveness over all metrics, with statistical significance mainly for shallow metrics (RR, nDCG@{1,3}). For TREC DL 2020, we observe improvements over shallow metrics only, although statistically not significant.

For TREC CAsT 2019, BB+PRF does not improve the effectiveness of BM25+BERT, except for nDCG@1. We speculate this is because BM25+BERT is trained with short passages, so it performs the best on CAsT (which consists of short passages).

For WebAP, improvements are observed over MAP and nDCG@{3,10}. Again, we believe this to be associated with the length of the passages in the dataset: here passages are longer and thus BM25+BERT (trained/fine-tuned on short passages) does not perform well.

For DL HARD, the improvement is only on nDCG@1, yet not significant, while nDCG@10 is significantly worse than the baseline. We speculate this is due to the poor relevance signals received by the PRF mechanism. Note that shallow metrics values on DL HARD are far below those in TREC DL 2019 and 2020 (which share the same passages): this means that the passages used for PRF are likely not relevant, thus possibly causing query drift.

To summarize, our proposed BB+PRF approach achieves substantially better results than BM25 and BM25+RM3. However, the improvements over BM25+BERT are more patchy, and are mostly achieved for shallow metrics. We put this down to the length of the text passages formed by the PRF methods: these are substantially longer than the passages used to train/fine-tune the BERT reranker.

5.4.2. Reranking with Vector-Based PRF (BB+PRF-R, BB+PRF-A)

When RepBERT is used as the base model (BB+PRF-R), for TREC DL 2019, improvements are obtained for RR and nDCG@1, while no improvements are obtained on the remaining metrics. For TREC DL 2020, we observe improvements over shallow metrics only (RR, nDCG@{1, 10}). For TREC CAsT 2019, the improvements are observed at nDCG@{1, 3}, but nDCG@10 is significantly worse than the baseline, and so is MAP. For WebAP, we observe improvements in shallow metrics as well (RR, nDCG@{1, 3}). For DL HARD, there are no improvements over all reported metrics; on the contrary, it performs significantly worse than the baseline on MAP and nDCG@10.

With ANCE as the base model (BB+PRF-A), for TREC DL 2019, all shallow metrics (RR, nDCG@{1, 3, 10}) and MAP are improved. For TREC DL 2020 and DL HARD, improvements are found at nDCG@{1, 3, 10}. For TREC CAsT and WebAP, we observe improvements over nDCG@{1, 3} for both datasets.

To summarize, the proposed vector-based PRF as reranker (BB+PRF-R, BB+PRF-A): (1) it improves the effectiveness over BM25+BERT across several metrics and for all datasets, (2) it achieves substantially better results than BM25, BM25+RM3, and RepBERT/ANCE, except on DL HARD, (3) it provides mixed results when compared with BM25+BERT, with no clear pattern of improvements (or deficiencies) across measures and datasets.

5.4.3. Reranking with BERT on Top of Vector-Based PRF (R+PRF+B, A+PRF+B)

R+PRF+B significantly improves MAP and R@1000 on all datasets except WebAP. All other results are either on par or slightly better (not significant) than the baseline. Moreover, R+PRF+B significantly improves nDCG@10 on WebAP compared to the baseline. However, the gain from R@1000 is mainly due to the gain obtained where moving from R to R+PRF: the reranking step does not contribute to this gain.

The trend for A+PRF+B is similar to that of R+PRF+B: it significantly improves MAP and R@1000 across all datasets except WebAP. A+PRF+B achieves the best effectiveness for nDCG@10 on WebAP, although effectiveness are not significant. All other results are either on par or slightly better than the baseline.

To summarize, the use of the BERT reranker on top of vector-based PRF significantly improves MAP, but for other metrics, improvements are not statistically significant.

5.4.4. Retrieval with Vector-Based PRF (R+PRF-R, A+PRF-A)

For R+PRF-R, results show similar trends on all datasets: improvements can be observed on all reported metrics (MAP, RR, R@1000, and nDCG@{1, 3, 10}, except on TREC DL 2020, where PRF performs worse than the RepBERT baseline for RR.

For A+PRF-A, PRF performs better than ANCE baseline on all evaluation metrics and across all datasets. The improvements in MAP are significant in TREC DL 2019, TREC DL 2020, TREC CAsT, and WebAP; the improvements for R@1000 are significant in TREC DL 2019, TREC CAsT, and DL HARD. Significant improvements in nDCG@{3, 10} are found only in TREC CAsT. Overall, A+PRF-A achieve higher effectiveness than R+PRF-R: ANCE per se is a stronger model than RepBERT, thus encoding more relevant information from the text. Hence, when PRF uses ANCE, it can better encode the additional relevance signals, leading to enhanced effectiveness.

To summarize, our proposed A+PRF-A and R+PRF-R models work well across all datasets and metrics. They also achieve substantial improvements over BM25 and BM25+RM3 baselines across almost all metrics, and they outperform the BM25+BERT baseline on several metrics.

5.4.5. Generalizability to Other Dense Retrievers

The results shown in Table 4 demonstrate that vector-based PRF consistently improves the effectiveness even where dense retrievers more effective than ANCE and RepBERT are used. Vector-based PRF tends to improve nDCG@3,10 (for the majority of the dense retrievers), MAP, nDCG@100, and R@1000 for all models in both TREC DL 2019 and TREC DL 2020. However, vector-based PRF does not improve RR and nDCG@1 in a consistent manner.

Table 4. Results of vector-based PRF for the task of retrieval, using dense retrievers more effective than ANCE and RepBERT. We randomly choose and fix the parameters for Rocchio and Average in all the experiments in this table, where Average PRF depth is 33, Rocchio PRF depth is 55. We also fix α\alpha and β\beta for Rocchio to be 0.4 and 0.6 respectively. The best results for each model are marked in Bold.
Model Method MAP RR nDCG@1 nDCG@3 nDCG@10 nDCG@100 R@1000
TREC DL 2019 TCT-ColBERT V1 Original 0.3864 0.9512 0.7326 0.6874 0.6700 0.5730 0.7207
Average 0.4457 0.8999 0.6705 0.6779 0.6639 0.6119 0.7570
Rocchio 0.4479 0.9368 0.7093 0.7083 0.6875 0.6143 0.7720
TCT-ColBERT V2 HN+ Original 0.4626 0.9767 0.8023 0.7410 0.7204 0.6318 0.7603
Average 0.5123 0.9767 0.7713 0.7454 0.7312 0.6719 0.8115
Rocchio 0.5161 0.9244 0.7248 0.7129 0.7111 0.6684 0.8147
DistilBERT KD Original 0.3759 0.9306 0.7558 0.7370 0.6994 0.5765 0.6853
Average 0.4362 0.9253 0.7481 0.7241 0.7096 0.6217 0.7180
Rocchio 0.4378 0.9345 0.7442 0.7286 0.7052 0.6189 0.7291
DistilBERT Balanced Original 0.4761 0.9510 0.7558 0.7494 0.7210 0.6360 0.7826
Average 0.5057 0.9458 0.7364 0.7383 0.7190 0.6526 0.8054
Rocchio 0.5249 0.9359 0.7364 0.7386 0.7231 0.6684 0.8352
SBERT Original 0.4097 0.9767 0.8372 0.7642 0.6930 0.5985 0.7201
Average 0.4565 0.9413 0.7403 0.7326 0.7001 0.6149 0.7357
Rocchio 0.4578 0.9355 0.7558 0.7448 0.6952 0.6149 0.7405
TREC DL 2020 TCT-ColBERT V1 Original 0.4290 0.8183 0.7500 0.7245 0.6678 0.5826 0.8181
Average 0.4725 0.8220 0.7346 0.7253 0.6957 0.6101 0.8667
Rocchio 0.4625 0.8392 0.7840 0.7410 0.6945 0.6056 0.8576
TCT-ColBERT V2 HN+ Original 0.4754 0.8392 0.7932 0.7199 0.6882 0.6206 0.8429
Average 0.4811 0.8212 0.7870 0.7386 0.6836 0.6228 0.8579
Rocchio 0.4860 0.8154 0.7685 0.7273 0.6804 0.6254 0.8518
DistilBERT KD Original 0.4159 0.8215 0.7284 0.7113 0.6447 0.5728 0.7953
Average 0.4214 0.7715 0.7130 0.6911 0.6316 0.5755 0.8403
Rocchio 0.4145 0.7703 0.7037 0.6823 0.6289 0.5760 0.8433
DistilBERT Balanced Original 0.4698 0.8350 0.7593 0.7426 0.6854 0.6346 0.8727
Average 0.4887 0.8380 0.7809 0.7510 0.7086 0.6449 0.9030
Rocchio 0.4879 0.8641 0.8056 0.7564 0.7083 0.6470 0.8926
SBERT Original 0.4124 0.7995 0.7346 0.6870 0.6344 0.5734 0.7937
Average 0.4258 0.7619 0.6728 0.6723 0.6412 0.5781 0.8169
Rocchio 0.4342 0.7941 0.7160 0.7032 0.6559 0.5851 0.8226

5.4.6. Summary

To answer RQ4, our results suggest that PRF used for either reranking or retrieval can improve effectiveness as measured across several metrics. More specifically, compared to the respective baselines, R+PRF-R and A+PRF-A tend to deliver improvements across all metrics and datasets, except for RR on TREC DL 2020 with RepBERT as base model. On the other hand, BB+PRF tends to only improve shallow metrics, especially when compared to BM25, BM25+RM3, and dense retriever baselines; BB+PRF-R and BB+PRF-A exhibit similar trends. Applying the BERT reranker on top of vector-based PRF significantly improves MAP, and outperforms the baseline on all metrics, but not significantly for the remaining metrics on all datasets with both R+PRF+B, A+PRF+B, and nDCG@10 on WebAP with R+PRF+B.

5.5. Efficiency of PRF

RQ5: What is the impact of PRF models on the efficiency of reranking and retrieval?

Table 5. Query latency of the investigated methods on TREC DL 2019: the lower latency, the better (faster).
Models Latency (ms/q)
Baselines BM25 (Anserini) 81
BM25 + RM3 (Anserini) 140
RepBERT(R) 93
ANCE(A) 94
RepBERT+BERT(R+B) 3,324
ANCE+BERT(A+B) 3,327
Vector-based PRF Retriever R+PRF-R-Average 163
R+PRF-R-Rocchio 163
A+PRF-A-Average 173
A+PRF-A-Rocchio 174
Vector-based PRF Reranker BB+PRF-R-Average 3,411
BB+PRF-R-Rocchio 3,414
BB+PRF-A-Average 3,409
BB+PRF-A-Rocchio 3,414
Text-based PRF Reranker BB+PRF(k=5k=5)-CT 6,889
BB+PRF(k=5k=5)-CA 17,266
BB+PRF(k=5k=5)-SW 22,314
Vector-based PRF with BERT Reranker R+PRF(Average)+B 3,395
R+PRF(Rocchio)+B 3,397
A+PRF(Average)+B 3,419
A+PRF(Rocchio)+B 3,421
BERT Reranker BM25 + BERT(BB) 3,246
BM25 + BERT Large 9,209

To answer this question, we study the efficiency of the PRF approaches and the baseline models. Low query latency – the time required for a ranker to produce a ranking in answer to a query – is an essential feature for the deployment of retrieval methods into real-time search engines. The query latency of the investigated methods is summarised in Table 5 and Figure 13444We have produced a fully annotated version of this image that includes each model name and made it available for online consultation at: https://github.com/ielab/Neural-Relevance-Feedback-Public/blob/master/figures/trade-off-with-label.pdf.

Refer to caption
Figure 13. Trade-off between effectiveness and efficiency for all methods in our experiments. Effectiveness is measured using nDCG@10, and efficiency is measured using log(ms/q)\log(ms/q). The sparse baselines (BM25 and BM25+RM3) cluster on the left bottom corner (red box). Dense based approaches, including ANCE, RepBERT, and our Vector-based PRF approaches cluster on the center left (green box). All rerankers, i.e., BM25+BERT(base/large), Text-based PRF, BM25+BERT+Vector-based PRF, Vector-based PRF+BERT reranker, cluster on the top right side (blue box); these methods present the worst efficiency compared to others. The black line shows the trade-off trend between effectiveness and efficiency.

The dense retrievers studied in this work (R and A) have a comparable query latency to BM25, with the latter being 10ms faster. Applying vector-based PRF to dense retrievers (R+PRF-R and A+PRF-A) has a comparable impact on query latency to BM25+RM3, with the latter being 23–34ms faster. The latency values measured in our experiments are compatible with the requirements of real-time search engines. On the other hand, applying vector-based PRF to BM25+BERT, as a reranking stage(BB+PRF-R and BB+PRF-A), has a high query latency, similar to that of BM25+BERT. Lastly, we found the two-stage BM25+BERT-Large to be the least efficient (up to 2 orders of magnitude slower than other methods) except the text-based PRF approaches(BB+PRF). While BERT and BERT-Large reranking models consider only the top 1,000 passages from BM25, their query latency remains impractical for real-time search engines, the BB+PRF approaches actually creates more queries from one original query, hence leads to worse query latency overall.

In terms of applying BERT on top of Vector-based PRF approaches, the efficiency is similar to the Vector-based PRF reranker and the BERT reranker, which is much lower than the Vector-based PRF approaches because of the additional BERT inference time. As for BERT, this approach is also impractical for real-time search engines.

We also analysed the relationship between query length (either original query, or query plus PRF signal) and latency. For BM25 and BM25+RM3, query latency increases with the increase of query length: the longer the query (including the PRF component), in fact, the more posting lists need to be traversed. On the other hand, query length does not affect the query latency of ANCE or RepBERT555If not just noticeably because more tokens need to be passed through the tokenizer., because the query is converted to fixed length vectors: no matter how many words in the query, the generated query vector is always of the same length. This same reasoning applied for the vector-based PRF approaches: even when increasing the number of PRF passages considered (kk), the query latency remains unchanged. As for the text-based PRF, we cut-off the query (including the revised query after PRF) to the length of 256 tokens, and before the query/passage pair is passed to BERT, the pair is padded to be of a total length of 512 tokens. Thus, no matter how long the query is (including possibly PRF), the sequence passed to BERT is always 512 tokens long. The reason for padding the input for BERT is that for each batch of query-passage pairs passed to BERT, the pairs in a batch need to be of the same length. The batching mechanism is useful for efficiently exploiting the GPU processing. In such cases, the query latency of the text-based PRF does not change with the increase of query length. A factor instead that does greatly affect the efficiency of text-based PRF with CA or SW is the depth of PRF. The more PRF passages, in fact, the more BERT inferences are required at run time, and thus the higher the query latency. For example, one original query with PRF depth at 5, the CA will generate 5 new queries, each for one feedback passage. So for each original query, it only needs one inference for BM25+BERT, but with the new queries from CA, it needs 5 more inferences, combined with BM25+BERT, it requires 6 inferences in total, similar for SW. (Table 5)

6. Use-case Analysis

To complement our analysis of the results and separate from our core research questions, we further analysed the results for a subset of the queries to gain a qualitative understanding of when, and possibly why, the PRF methods work – or don’t. For this, we limit our analysis to the retrieval task and the ANCE and Vector-based PRF with Rocchio with k=5k=5.

Figure 14(top) presents a query-by-query analysis of the gains and losses of Vector-based PRF (Rocchio) with respect to ANCE on the combined set of TREC DL 2019 and 2020 queries. The effectiveness measure used is nDCG@5 and gains and losses are represented by bars in the figure. We used nDCG@5 for this analysis and not the rank cut-offs we have reported for other analyses to align the evaluation cut-off with the cut-off used for the feedback signal, i.e. the PRF depth kk, which was also 5. The observations that follow, based on nDCG@5, are also found when other rank cut-offs are considered.

As one would expect, there are a number of queries for which the PRF method experiences losses, and other queries for which gains are observed; interestingly there is a considerable amount of queries for which neither gains nor losses are found. We further verified whether the encoding of the query obtained via PRF is different from that of the original query. This is done especially for the cases in which neither gains nor losses are observed: does this happen because there is no difference between the representation of the original query and of the query after PRF? We performed this analysis by comparing the two dense representations using the inner product between the encoding vectors. We found that the similarity between the initial query encoding and the one obtained with PRF varies across queries and that queries for which neither gains nor losses are obtained, are no different in terms of difference between query representations compared to queries that exhibit gains or losses (figure not shown here, but available in the online appendix666 https://github.com/ielab/Neural-Relevance-Feedback-Public). In other words, the amount by which PRF changes the initial query representation is not directly associated with a gain or a loss, and queries for which no difference in effectiveness is found are often characterised by differences between the original and PRF representations of the query.

Refer to caption
Figure 14. The query-by-query analysis on the combined set of TREC DL 2019 and 2020 queries of (1) the gain/loss obtained by Vector-based PRF(Rocchio) with respect to ANCE, as represented by the barplot at the top, (2) the retrieved passages overlap, as represented by the red dots in the top plot, (3) and the analysis of the passage relevance for the feedback signal provided as input to the PRF (top k=5k=5 passages from ANCE – mid plot) and the top 5 passages retrieved by the PRF method (bottom plot).

For each query, Figure 14 (top) also reports the amount of overlap between the top kk passages retrieved by the first stage of retrieval (ANCE) – this is effectively the signal that is used as input to the PRF method – and the top kk passages retrieved by the PRF method. The overlap is reported in percentage of the top kk passages that are in common between the two sets; k=5k=5 and thus an overlap of 80%80\% means that 4 out of 5 passages are in common. This statistic helps us to analyse whether the PRF method ended up just re-ranking the top kk passages from the first stage, or it did push into the top kk ranks passages that were not there before. Losses in effectiveness for the PRF method tend to be associated with re-ranking of the top kk passages from the first stage of retrieval: the average overlap for these queries is 85.2%. And thus, these losses are not caused by retrieving less of the relevant passages, or passages that are less relevant (e.g. marginally relevant instead of highly relevant); they are instead caused by re-ordering the passages from the first stage in a less effective manner. On the other hand, gains instead have a higher tendency to be associated with the ability of the PRF method to bring into the top kk additional, new relevant passages compared to what the first stage of retrieval could do: the average overlap for these queries is 82.8%. Differences, however, are not statistically significant (Pearson’s Chi-squared test).

Figure 14 (mid) presents an analysis of the relevancy of the top kk passages in the first stage of retrieval, while Figure 14 (bottom) presents an analysis of the relevancy of the top kk passages after PRF. Remember that the relevancy of the top kk passages in the first stage of retrieval in all effects represents the quality of the feedback signal that is passed as input to the PRF method. Each stacked column refers to a specific query, and the column is aligned to the gain-loss plot (Figure 14 (top)) so that it is possible to directly compare the relevancy of the PRF signal and the gain or loss produced by the PRF method. No obvious pattern appears at first sight. PRF signals that contain non-relevant documents produce results that can go either way: at times losses are obtained, but it is also common to obtain gains (or even have no effect at all). However, a more scrupulous examination of this signal leads to an interesting observation. We do this by examining the five queries with largest losses, and then the five queries with the largest gains for which there is at least one non-relevant passage. We observe the following. Non-relevant passages for queries that display losses are largely off topic: they have a very weak relationship to the query, if at all. For example, for query 1110678: what is the un fao, the PRF signal includes 4 non-relevant passages, and these passages are related to the card game Uno; they have nothing to do with the United Nations FAO program. A subset of these passages (along with other examples) are presented in Table 6. Another similar example is query 169208: does mississippi have an income tax, for which the PRF signal includes 3 non-relevant passages and, while they are related to tax and the state of Mississipi, they are about unemployment or start-up tax and thus not relevant to the need of information about the more general income tax. When this type of feedback is used in the PRF mechanism, the new query representation displays query drift, and it does score higher passages that are non-relevant, but are related to those in the feedback signal (e.g., for query 1110678 many top retrieved passages by PRF are related to Uno, even beyond the top kk). This situation instead does not occur when we examine the PRF signal of queries for which PRF provides a gain. In particular, while this signal does still contain non-relevant passages, the non-relevancy nature of these passages is different from that of passages for which losses are found. Let us unwrap this with actual examples. The PRF method exhibits a large gain for query 1136047: difference between a company’s strategy and business model is, despite the PRF signal for this query displays several non-relevant passages. These non-relevant passages, shown in Table 6, however, do display a strong aboutness relation (Maron, 1977) with the query, for example by mentioning the key terms of the query – and when this is used in the PRF mechanism it would help reinforce the query representation. We also note that the non-relevant label assigned to some of these passages by the original TREC assessors may be somewhat questioned: for example passage 8724036 for query 1136047 appears to us as at least marginally relevant. Other example queries that display a pattern similar to that of query 1136047 are included in Table 6.

Table 6. Example queries for which the Vector-based PRF (Rocchio) method produces losses/gains, along with example non-relevant passages ranked in the top kk from the first stage ranker (and thus used as input to PRF).
Query ID Query Passage
Queries for which PRF produces a loss 169208 does mississippi have an income tax777Two other, similar non-relevant passages appearing in the top k=5k=5 for this topic are omitted from the table for space reasons. 2888361: Unemployment Tax Rates. Reporting and Filing. In Mississippi, the tax rate for a start-up business is 1.00% the first year of liability, 1.10% the second year of liability and 1.20% the third and subsequent years of liability until the employer is eligible for a modified rate
1037798 who is robert gray 2868740: (Redirected from Gary Leroi Gray) Gary LeRoi Gray (born February 12, 1987) is an actor and voice actor involved with movies, television, and animation. He is most recognized for his childhood role as Nelson Tibideaux, the son of Sondra Huxtable Tibideaux and Elvin Tibideaux on the NBC sitcom The Cosby Show. He appeared on the series during its eighth and final season (1991-1992).
2866248: Matthew Gray Gubler. Matthew Gray Gubler is an Emmy award winning actor, director, producer, painter, and voice over actor from Las Vegas Nevada. While studying film directing at NYU he interned for Wes Anderson who gave him his first feature film role as Bill Murray’s loyal intern Nico in The Life Aquatic with Steve Zissou (2004). For the past eleven years Gubler …
1110678 what is the un fao7 5253767: Uno (/EuEnoE/; from Italian and Spanish for ‘one’) (stylized as UNO) is an American card game that is played with a specially printed deck (see Mau Mau for an almost identical game played with normal playing cards). The game was originally developed in 1971 by Merle Robbins in Reading, Ohio, a suburb of Cincinnati.
3386130: Uno (card game) For the video game adaptation, see Uno (video game). Uno (/EuEnoE/; from Italian and Spanish for ‘one’) (stylized as UNO) is an American card game that is played with a specially printed deck (see Mau Mau for an almost identical game played with normal playing cards). The game was originally developed in 1971 by Merle Robbins in Reading, Ohio, a suburb of Cincinnati.
Queries for which PRF produces a gain 1129237 hydrogen is a liquid below what temperature 8588226: Hydrogen is a liquid below what temperature? was asked by Shelly Notetaker on May 31 2017. 426 students have viewed the answer on StudySoup. View the answer on StudySoup. Sign Up Login
8588222: Answer to: Hydrogen is a liquid below what temperature? By signing up, you’ll get thousands of step-by-step solutions to your homework questions…. for Teachers for Schools for Companies
87181 causes of left ventricular hypertrophy 47203: Causes of Right Ventricular Hypertrophy. There are four usual causes of right ventricular hypertrophy. The first one is pulmonary hypertension. As stated earlier, pulmonary hypertension is a condition wherein the blood pressure increases in the pulmonary artery. And this can lead to shortness of breath, dizziness and fainting.
1136047 difference between a company’s strategy and business model is 8724032: 6. The difference between a company’s business model and a company’s strategy is that: a. a company’s business model is - Answered by a verified Business Tutor
8724038: Now, we address our second question. What is the difference between a strategy and a business model? A strategy is about the external logic of a business. How are we going to compete? Check out my earlier post on the elements of a strategy for more about strategies. A business model is about the internal logicâ¦
8724036: Now, we address our second question. What is the difference between a strategy and a business model? A strategy is about the external logic of a business. How are we going to compete? Check out my earlier post on the elements of a strategy for more about strategies. A business model is about the internal logic of the business. See my post on the elements of a business model. Does it make operational and economic sense? Do all of the pieces fit together? A business model is a tool that complements both a business strategy and a business plan. It is an important tool because 1) it ensures that you understand the logic of your business; and, 2) it helps you communicate the logic of your business. Of course, this begs the question, what is a business plan?

7. Discussion

Integrating PRF with deep language models has two main challenges: computational cost and input size limit, which can be roughly mapped to efficiency and effectiveness, respectively. Previous research has focused mainly on effectiveness, and applying PRF as a second stage that considers a subset of the data collection, to mitigate the computational cost. For instance, Zheng et al. (2020) proposed a PRF framework which is divided into three phases, each involving BERT, to rerank passages and chunks of text after the initial retrieval. Similarly, Wang et al. (2020b) applied PRF to the second stage retrieval, utilising BERT for reranking sentences. Other work that applied PRF to the first stage retrieval has limited the second phase of retrieval to a subset of the collection, either to depth of 500 (Yu et al., 2021a) or 1000 (Li et al., 2018). We argue that the main limitation with these approaches is related to the first stage retrieval model they employed — if the first stage is basic bag-of-words then the overall method inherits many of this models limitations.

To address the applicability, effectiveness, and efficiency of text-based, vector-based and hybrid PRF approaches, we compared them over two different tasks, retrieval for vector-based PRF and reranking for both text-based and hybrid PRF. While the text-based and hybrid PRF approaches handle the input size limit, we showed that they enhance the ranking effectiveness only marginally. This finding aligns with previous research, but we demonstrated it happens at the cost of efficiency. The BERT model is slow (ranking 1,000 passages for each query takes 9,209 ms), rendering text-based and hybrid PRF approaches infeasible and inapplicable to real-time search engines for this task, as shown in Figure 13. On the other hand, our proposed vector-PRF approach is substantially more efficient: it only takes 163-174 ms to process one query. In addition, it does so while also improving the effectiveness of dense retrievers in terms of all metrics compared to ANCE and RepBERT (for retrieval), and in terms of shallow metrics compared to BERT (for reranking).

We also apply the BERT reranker on top of Vector-based PRF approaches. Although this indeed improves effectiveness across all evaluation metrics, only improvements on MAP are significant on all datasets, and R+PRF+B significantly improves nDCG@10 on WebAP only. On the other hand, R@1000 is significantly improved across several datasets, but this improvement comes from the gain produced by the Vector-based PRF compared to the respective dense retrievers.

The BERT reranker used as a strong baseline in our experiments is an off-the-shelf model, originally produced by Nogueira and Cho (2019). This model is fine-tuned on the training set of the MS MARCO Passage Retrieval Dataset. The query length in this training set is much shorter than the newly formed queries from the text-based PRF approaches. This causes a mismatch between training and testing data. Theoretically, if the testing data is significantly different from the trained data data used to fine-tune the model, the model itself will be less effective. Despite this, in our experiments we found that the formed long PRF queries still significantly improve results for some metrics across several datasets (although improvements are often only marginal) – but in the majority of cases their effectiveness is worse than that of BM25+BERT. It will then be interesting to investigate whether the effectiveness of text-based PRF can be further improved by fine-tuning the BERT model with long PRF queries, hence eliminating the training/testing data mismatch. On the other hand, vector-based PRF is not affected by this issue, because the query vector is of fixed length: no matter how many PRF passages are added to the original query vector, the query vector formed via PRF maintains the same length, and thus no training/testing data mismatch occurs. Another advantage of vector-based PRF over the text-based is that, although its effectiveness might be improved significantly after fine-tuning on long PRF queries, the text-based PRF approach is still inefficient and it is infeasible to apply it to real-time settings.

As complementary experiments, we also executed the retrieval task and re-ranking task on the MS MARCO dev queries. We omit the detailed results for these experiments from the manuscript in the interest of space and focus; however we make them available in the online appendix888https://github.com/ielab/Neural-Relevance-Feedback-Public/blob/master/Vector_Based_PRF/README.md and we provide a summary of the observations next.

For both retrieval and re-ranking tasks on the MS MARCO dev queries, we found no improvement over the non-PRF baselines. This is unlike in the TREC DL datasets, where improvements are observed. We believe that the reason for this difference is to be found on the largely sparse annotated nature of MS MARCO vs. the more complete relevance annotations in TREC DL. One may argue that other neural PRF methods, such as ANCE-PRF, have shown improvements on MS MARCO too, and not just on TREC DL. However, we stress an important difference. Both the baseline dense retrievers and the learnt PRF methods (i.e. ANCE-PRF and the likes) are trained so as to build representations of the single “gold” (i.e. relevant) MS MARCO passage that are close to the query (for ANCE-PRF: representations of the query after PRF that are close to the gold MS MARCO passage). Thus, in the context of trained PRF methods like ANCE-PRF, the query representation is changed to become closer to the single MS MARCO target gold passage. Our PRF method changes the query representation (and consequently perturbs the ranking) based on the representation of the top kk passages alone: there is no notion of a gold passage towards which such new representation should be moved. This means these non-learnt vector-based PRF methods are not suitable to situations in which a single golden passage exists, for example for the task of known item retrieval – and this is shown by the MS MARCO results too. However, this does not mean these methods are not suitable when there is an array of relevant passages to be retrieved; and this in fact is what the TREC DL results show.

8. Conclusion

This article investigated the integration of PRF within transformer-based deep language models and dense retrievers for retrieval and reranking. In this context, a text-based PRF approach, applicable to the reranking task only, and a vector-based PRF approach, applicable to both retrieval and reranking tasks, were proposed to leverage the relevance signals from the feedback passages. The proposed approaches are based on the BERT reranker and the dense retrievers ANCE and RepBERT. In this context, we studied the impact of vector representation (for vector-based PRF), PRF depth, text handling techniques (for text-based PRF), and different score estimation (vector fusion) methods (for vector-based PRF) .

When analysing results for the vector-based PRF approaches, we found they differed depending on whether we considered the task of retrieval or reranking.

In terms of which vector-based representation is most effective for retrieval, we empirically found that performing PRF using ANCE as a representation (A+PRF-A) is better than when using RepBERT (R+PRF-R) across metrics and datasets, with the exception of R@1000. Indeed, our results show that R+PRF-R tends to improve deep metrics, while A+PRF-A tends to improve shallow metrics. When considering the reranking task, substantial improvements only occurred on a few datasets and for a limited amount of metrics, and overall ANCE-based PRF performs better than RepBERT-based.

In terms of the impact of depth of PRF signal for vector-based PRF, we found that when retrieval is considered, substantial improvements are recorded when 3 to 5 passages are given as feedback, while deeper feedback (10 passages) hurts effectiveness most of the times. When reranking is considered, PRF depth is not a factor substantially influencing effectiveness in a consistent manner across datasets and metrics.

In terms of score estimation methods (vector fusion) for the vector-based PRF, we found that, for the retrieval task, 𝒞β\mathcal{RC}_{\beta} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} achieve the best effectiveness across all datasets. Moreover, R+PRF-R often performs best when using either approximately evenly distributed weights between query/feedback passages, or when the query retains most of the weight. This is similar for A+PRF-A, although for this method there are cases where the best effectiveness is achieved by giving a higher weight to feedback passages. When the task of reranking is considered, we found that 𝒞α\mathcal{RC}_{\alpha} and 𝒞α,β\mathcal{RC}_{\alpha,\beta} record the most improvements. Substantial improvements, however, only occur on TREC DL 2019 and for only MAP, RR, and nDCG@1.

To validate the generalisability of our proposed Vector-based PRF models, we considered five additional dense retrievers (TCT-ColBERT V1, TCT-ColBERT V2 HN+, DistilBERT KD, DistilBERT Balanced, and SBERT) other than ANCE and RepBERT. Experimental results showed that our vector-based PRF approaches consistently improved the effectiveness over deep evaluation metrics.

The text-based PRF approaches were only applicable to the reranking task. In terms of depth of PRF, we found that, in most cases, substantial improvements are achieved when using 3, 5, and 10 passages for PRF. Deeper PRF signals (15 and 20 passages) were found to hurt effectiveness: this is likely because the addition of more feedback passages substantially contributed to query drift. In terms of text handling methods, we found that CA performs best in most cases, but only marginal improvements over the baselines are recorded across datasets and metrics.

We also experimented with applying the BERT reranker on top of Vector-based PRF models: this improves the effectiveness for all evaluation metrics, but only significantly for MAP. The significant improvements on R@1000 are purely produced by the Vector-based PRF models: the reranking step does not contribute to these improvements.

Vector-based and text-based PRF approaches could be compared only in the reranking task. In terms of effectiveness, neither of these two types of approaches clearly and consistently outperformed the other across datasets and metrics. However, they greatly differed in terms of efficiency (query latency). When the vector-based PRF was used for retrieval, latency was close to that of the bag-of-words PRF method BM25+RM3, and about double that of BM25. When used for reranking, the latency of the vector-based PRF remained the same. However, to this latency one needs to add that of the preceding part of the retrieval/ranking pipeline: in the case of our experiments that was the latency of BM25+BERT, which was substantial. On the other hand, the latency of text-based PRF was substantially higher. To the latency of the initial retrieval/ranking pipeline, in fact, the text-based PRF added the latency associated with the text handling methods (e.g., Concatenate and Aggregate) – and this operation involved the execution of many more, costly BERT inferences, rendering the overall latency prohibitive for real-time search engines. The extensive experiments provided in this paper serve as a guide for using state-of-the-art neural retrieval models into low latency environments.

This work opens up a number of avenues for future research. First, we note that the Vector-based PRF method, and especially the Rocchio approach, has a number of limitations that future work could address. For example, Rocchio has parameters (α\alpha, β\beta) that need to be properly tuned to achieve the highest performance; we performed this tuning with respect to a validation set. How these parameters could instead be set on a per query basis (e.g., using some form of weight prediction) is still unclear.

Second, the use-case analysis in Section 6 revealed that the PRF methods are susceptible to the quality of the PRF signal (i.e. the passages retrieved by the first stage method and then passed as input to PRF). Recent work has made initial inroads to investigate this aspect. Specifically, Li et al. have studied the effect of signals of different qualities on the PRF process (Li et al., 2022a) – although they only explored a limited set of possible signals and PRF methods. Zhuang et al. have devised effective methods to de-noise the feedback signal in the case of implicit feedback (Zhuang et al., 2022). The extent to which different PRF methods differ in terms of susceptibility to PRF signals of different qualities is still however unclear, especially beyond the specific methods that we examined in the use-case analysis and the signals and settings explored in the aforementioned recent work.

Third, this paper only considered the passage retrieval and ranking tasks. PRF techniques have, however, long been explored for document retrieval tasks and one wonders how the methods designed here for passages, and that exploit pre-trained language models, generalise when considering documents. We note that adapting these specific passage PRF methods to documents is not straightforward. Documents are sensibly longer in length than passages, and most likely their text exceeds the maximum input length that pre-trained language models like BERT can consider. This is certainly a problem for models like ANCE-PRF and our text-based PRF methods, where the text of the PRF signal needs to be encoded. But, this is also a problem for our Vector-based PRF methods: while the text associated to the PRF signal is not required by the PRF process itself, these documents need still to be encoded so that their dense vector can be used (for the initial round of retrieval, and for the PRF). Tackling this limitation of pre-trained language models for effective document retrieval is not new (Yang et al., 2019b; Yilmaz et al., 2019a, b; Mass and Roitman, 2020)– progress so far in this space has not yet solved this problem (especially if also efficiency considerations are made) and PRF in this context is largely unexplored (Wang et al., 2020b; Xu et al., 2009; Wang et al., 2020a).

Fourth, PRF is typically integrated in wider pipelines, including multiple rankers and a multi-stage ranking architecture. A recent trend with neural rankers is to create hybrid systems that comprise a dense and a sparse rankers (Wang et al., 2021; Arabzadeh et al., 2021; Lin et al., 2021b). In this paper, we have not considered such architecture type, and how PRF could be integrated. We however refer the interested reader to the recent work of Li et al. that has extended our methods to hybrid rankers (Li et al., 2022b).

The implementations of our PRF methods are made publicly available at http://ielab.io/publications/li-tois-2022, along with the full empirical results.

Acknowledgements.
This research is partially funded by the Grain Research and Development Corporation project AgAsk (UOQ2003-009RTX). Dr Guido Zuccon is the recipient of an Australian Research Council DECRA Research Fellowship (DE180101579).

References

  • (1)
  • Abdul-Jaleel et al. (2004) Nasreen Abdul-Jaleel, James Allan, W Bruce Croft, Fernando Diaz, Leah Larkey, Xiaoyan Li, Mark D Smucker, and Courtney Wade. 2004. UMass at TREC 2004: Novelty and HARD. Computer Science Department Faculty Publication Series (2004), 189.
  • Arabzadeh et al. (2021) Negar Arabzadeh, Xinyi Yan, and Charles LA Clarke. 2021. Predicting Efficiency/Effectiveness Trade-offs for Dense vs. Sparse Retrieval Strategy Selection. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management. 2862–2866.
  • Aslam and Montague (2001) Javed A Aslam and Mark Montague. 2001. Models for metasearch. In Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval. 276–284.
  • Azad and Deepak (2019) Hiteshwar Kumar Azad and Akshay Deepak. 2019. Query expansion techniques for information retrieval: a survey. Information Processing & Management 56, 5 (2019), 1698–1735.
  • Cao et al. (2008) Guihong Cao, Jian-Yun Nie, Jianfeng Gao, and Stephen Robertson. 2008. Selecting good expansion terms for pseudo-relevance feedback. In Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval. 243–250.
  • Clinchant and Gaussier (2013) Stéphane Clinchant and Eric Gaussier. 2013. A Theoretical Analysis of Pseudo-Relevance Feedback Models. In Proceedings of the 2013 Conference on the Theory of Information Retrieval. 6–13.
  • Craswell et al. (2020) Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen M Voorhees. 2020. Overview of the TREC 2019 Deep Learning Track. In Text REtrieval Conference, TREC.
  • Craswell et al. (2021) Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen M Voorhees. 2021. Overview of the TREC 2020 Deep Learning Track. In Text REtrieval Conference, TREC.
  • Dai and Callan (2019) Zhuyun Dai and Jamie Callan. 2019. Deeper Text Understanding For IR with Contextual Neural Language Modeling. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval. 985–988.
  • Dai et al. (2019) Zihang Dai, Zhilin Yang, Yiming Yang, Jaime G Carbonell, Quoc Le, and Ruslan Salakhutdinov. 2019. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2978–2988.
  • Dalton et al. (2019) Jeffrey Dalton, Shahrzad Naseri, Laura Dietz, and James Allan. 2019. Local and Global Query Expansion for Hierarchical Complex Topics. In European Conference on Information Retrieval. Springer, 290–303.
  • Dalton et al. (2020) Jeffrey Dalton, Chenyan Xiong, and Jamie Callan. 2020. TREC CAsT 2019: The Conversational Assistance Track Overview. In Text REtrieval Conference, TREC.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 4171–4186.
  • Diaz et al. (2016) Fernando Diaz, Bhaskar Mitra, and Nick Craswell. 2016. Query Expansion with Locally-Trained Word Embeddings. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 367–377.
  • Ding et al. (2020) Ming Ding, Chang Zhou, Hongxia Yang, and Jie Tang. 2020. CogLTX: Applying BERT to Long Texts. Advances in Neural Information Processing Systems 33 (2020).
  • dos Santos et al. (2020) Cicero dos Santos, Xiaofei Ma, Ramesh Nallapati, Zhiheng Huang, and Bing Xiang. 2020. Beyond [CLS] Through Ranking by Generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 1722–1727.
  • Guo et al. (2020) Jiafeng Guo, Yixing Fan, Liang Pang, Liu Yang, Qingyao Ai, Hamed Zamani, Chen Wu, W Bruce Croft, and Xueqi Cheng. 2020. A deep look into neural ranking models for information retrieval. Information Processing & Management 57, 6 (2020), 102067.
  • Hofstätter et al. (2020) Sebastian Hofstätter, Sophia Althammer, Michael Schröder, Mete Sertkan, and Allan Hanbury. 2020. Improving efficient neural ranking models with cross-architecture knowledge distillation. arXiv preprint arXiv:2010.02666 (2020).
  • Hofstätter et al. (2021) Sebastian Hofstätter, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy Lin, and Allan Hanbury. 2021. Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling. arXiv preprint arXiv:2104.06967 (2021).
  • Jiao et al. (2020) Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. 2020. TinyBERT: Distilling BERT for Natural Language Understanding. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings. 4163–4174.
  • Johnson et al. (2019) Jeff Johnson, Matthijs Douze, and Hervé Jégou. 2019. Billion-scale similarity search with gpus. IEEE Transactions on Big Data (2019).
  • Keikha et al. (2014) Mostafa Keikha, Jae Hyun Park, and W Bruce Croft. 2014. Evaluating Answer Passages Using Summarization Measures. In Proceedings of the 37th international ACM SIGIR conference on Research & development in information retrieval. 963–966.
  • Khattab and Zaharia (2020) Omar Khattab and Matei Zaharia. 2020. Colbert: Efficient and Effective Passage Search via Contextualized Late Interaction Over BERT. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 39–48.
  • Kuzi et al. (2016) Saar Kuzi, Anna Shtok, and Oren Kurland. 2016. Query Expansion Using Word Embeddings. In Proceedings of the 25th ACM international on conference on information and knowledge management. 1929–1932.
  • Lavrenko and Croft (2017) Victor Lavrenko and W Bruce Croft. 2017. Relevance-Based Language Models. In ACM SIGIR Forum, Vol. 51. ACM New York, NY, USA, 260–267.
  • Li et al. (2018) Canjia Li, Yingfei Sun, Ben He, Le Wang, Kai Hui, Andrew Yates, Le Sun, and Jungang Xu. 2018. NPRF: A Neural Pseudo Relevance Feedback Framework for Ad-hoc Information Retrieval. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 4482–4491.
  • Li et al. (2022a) Hang Li, Ahmed Mourad, Bevan Koopman, and Guido Zuccon. 2022a. How does Feedback Signal Qality Impact Effectiveness of Pseudo Relevance Feedback for Passage Retrieval. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR).
  • Li et al. (2022b) Hang Li, Shuai Wang, Shengyao Zhuang, Ahmed Mourad, Xueguang Ma, Jimmy Lin, and Guido Zuccon. 2022b. To Interpolate or not to Interpolate: PRF, Dense and Sparse Retrievers. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR).
  • Li et al. (2022c) Hang Li, Shengyao Zhuang, Ahmed Mourad, Xueguang Ma, Jimmy Lin, and Guido Zuccon. 2022c. Improving Query Representations for Dense Retrieval with Pseudo Relevance Feedback: A Reproducibility Study. In Proceedings of the 44rd European Conference on Information Retrieval.
  • Lin (2019) Jimmy Lin. 2019. The simplest thing that can possibly work: pseudo-relevance feedback using text classification. arXiv preprint arXiv:1904.08861 (2019).
  • Lin et al. (2021a) Jimmy Lin, Xueguang Ma, Sheng-Chieh Lin, Jheng-Hong Yang, Ronak Pradeep, and Rodrigo Nogueira. 2021a. Pyserini: An easy-to-use Python toolkit to support replicable IR research with sparse and dense representations. arXiv preprint arXiv:2102.10073 (2021).
  • Lin et al. (2021b) Jimmy Lin, Rodrigo Nogueira, and Andrew Yates. 2021b. Pretrained transformers for text ranking: BERT and beyond. Synthesis Lectures on Human Language Technologies 14, 4 (2021), 1–325.
  • Lin et al. (2020) Sheng-Chieh Lin, Jheng-Hong Yang, and Jimmy Lin. 2020. Distilling dense representations for ranking using tightly-coupled teachers. arXiv preprint arXiv:2010.11386 (2020).
  • Lin et al. (2021c) Sheng-Chieh Lin, Jheng-Hong Yang, and Jimmy Lin. 2021c. In-batch negatives for knowledge distillation with tightly-coupled teachers for dense retrieval. In Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021). 163–173.
  • Liu et al. (2020) Weijie Liu, Peng Zhou, Zhiruo Wang, Zhe Zhao, Haotang Deng, and Qi Ju. 2020. FastBERT: a Self-distilling BERT with Adaptive Inference Time. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 6035–6044.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A Robustly Optimized BERT Pretraining Approach. arXiv preprint arXiv:1907.11692 (2019).
  • Lv and Zhai (2009) Yuanhua Lv and ChengXiang Zhai. 2009. A Comparative Study of Methods For Estimating Query Language Models with Pseudo Feedback. In Proceedings of the 18th ACM ACM International Conference on Information and Knowledge Management. 1895–1898.
  • Lv and Zhai (2010) Yuanhua Lv and ChengXiang Zhai. 2010. Positional relevance model for pseudo-relevance feedback. In Proceedings of the 33rd international ACM SIGIR conference on Research and development in information retrieval. 579–586.
  • Lv and Zhai (2014) Yuanhua Lv and ChengXiang Zhai. 2014. Revisiting the Divergence Minimization Feedback Model. In Proceedings of the 23rd ACM International Conference on Information and Knowledge Management. 1863–1866.
  • Ma et al. (2019) Yanjun Ma, Dianhai Yu, Tian Wu, and Haifeng Wang. 2019. PaddlePaddle: An open-source deep learning platform from industrial practice. Frontiers of Data and Domputing 1, 1 (2019), 105–115.
  • Macdonald and Ounis (2008) Craig Macdonald and Iadh Ounis. 2008. Voting techniques for expert search. Knowledge and information systems 16, 3 (2008), 259–280.
  • Mackie et al. (2021) Iain Mackie, Jeffrey Dalton, and Andrew Yates. 2021. How Deep is your Learning: the DL-HARD Annotated Deep Learning Dataset. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval.
  • Maron (1977) Melvin E Maron. 1977. On indexing, retrieval and the meaning of about. Journal of the american society for information science 28, 1 (1977), 38–43.
  • Mass and Roitman (2020) Yosi Mass and Haggai Roitman. 2020. Ad-hoc document retrieval using weak-supervision with BERT and GPT2. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 4191–4197.
  • Miao et al. (2012) Jun Miao, Jimmy Xiangji Huang, and Zheng Ye. 2012. Proximity-based rocchio’s model for pseudo relevance. In Proceedings of the 35th international ACM SIGIR conference on Research and development in information retrieval. 535–544.
  • Naseri et al. (2021) Shahrzad Naseri, Jeffrey Dalton, Andrew Yates, and James Allan. 2021. CEQE: Contextualized Embeddings for Query Expansion. In The 43rd European Conference On Information Retrieval. 467–482.
  • Naseri et al. (2018) Shahrzad Naseri, John Foley, James Allan, and Brendan T O’Connor. 2018. Exploring Summary-Expanded Entity Embeddings for Entity Retrieval.. In CIKM Workshops.
  • Nguyen et al. (2016) Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A Human Generated Machine Reading Comprehension Dataset. In Workshop on Cognitive Computing at NIPS.
  • Nogueira and Cho (2019) Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Re-ranking with BERT. arXiv preprint arXiv:1901.04085 (2019).
  • Padaki et al. (2020) Ramith Padaki, Zhuyun Dai, and Jamie Callan. 2020. Rethinking Query Expansion for BERT Reranking. In Advances in Information Retrieval-42nd European Conference on IR Research, ECIR 2020.
  • Qu et al. (2020) Yingqi Qu, Yuchen Ding, Jing Liu, Kai Liu, Ruiyang Ren, Xin Zhao, Daxiang Dong, Hua Wu, and Haifeng Wang. 2020. RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering. arXiv preprint arXiv:2010.08191 (2020).
  • Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language Models Are Unsupervised Multitask Learners. In OpenAI blog.
  • Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research 21 (2020), 1–67.
  • Reimers et al. (2019) Nils Reimers, Iryna Gurevych, Nils Reimers, Iryna Gurevych, Nandan Thakur, Nils Reimers, Johannes Daxenberger, Iryna Gurevych, Nils Reimers, Iryna Gurevych, et al. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
  • Ren et al. (2021) Ruiyang Ren, Yingqi Qu, Jing Liu, Wayne Xin Zhao, Qiaoqiao She, Hua Wu, Haifeng Wang, and Ji-Rong Wen. 2021. RocketQAv2: A Joint Training Method for Dense Passage Retrieval and Passage Re-ranking. arXiv preprint arXiv:2110.07367 (2021).
  • Robertson and Zaragoza (2009) Stephen Robertson and Hugo Zaragoza. 2009. The Probabilistic Relevance Framework: BM25 and Beyond. Foundations and Trends in Information Retrieval 3, 4 (2009), 333–389.
  • Rocchio (1971) J.J. Rocchio. 1971. Relevance Feedback in Information Retrieval. In The SMART Retrieval System - Experiments in Automatic Document Processing. 313–323.
  • Roy et al. (2016) Dwaipayan Roy, Debjyoti Paul, Mandar Mitra, and Utpal Garain. 2016. Using Word Embeddings for Automatic Query Expansion. Neu-IR: The SIGIR 2016 Workshop on Neural Information Retrieval (2016).
  • Tao and Zhai (2006) Tao Tao and ChengXiang Zhai. 2006. Regularized Estimation of Mixture Models for Robust Pseudo-Relevance Feedback. In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’06). Association for Computing Machinery, 162–169.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems. 6000–6010.
  • Wang et al. (2020b) Junmei Wang, Min Pan, Tingting He, Xiang Huang, Xueyan Wang, and Xinhui Tu. 2020b. A Pseudo-Relevance Feedback Framework Combining Relevance Matching and Semantic Matching for Information Retrieval. Information Processing & Management 57, 6 (2020), 102342.
  • Wang et al. (2020a) Le Wang, Ze Luo, Canjia Li, Ben He, Le Sun, Hao Yu, and Yingfei Sun. 2020a. An end-to-end pseudo relevance feedback framework for neural document retrieval. Information Processing & Management 57, 2 (2020), 102182.
  • Wang et al. (2021) Shuai Wang, Shengyao Zhuang, and Guido Zuccon. 2021. Bert-based dense retrievers require interpolation with bm25 for effective passage retrieval. In Proceedings of the 2021 ACM SIGIR International Conference on Theory of Information Retrieval. 317–324.
  • Xiong et al. (2020) Lee Xiong, Chenyan Xiong, Ye Li, Kwok-Fung Tang, Jialin Liu, Paul Bennett, Junaid Ahmed, and Arnold Overwijk. 2020. Approximate Nearest Neighbor Negative Contrastive Learning For Dense Text Retrieval. arXiv preprint arXiv:2007.00808 (2020).
  • Xu et al. (2009) Yang Xu, Gareth JF Jones, and Bin Wang. 2009. Query dependent pseudo-relevance feedback based on wikipedia. In Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval. 59–66.
  • Yang et al. (2018) Peilin Yang, Hui Fang, and Jimmy Lin. 2018. Anserini: Reproducible Ranking Baselines Using Lucene. Journal of Data and Information Quality (JDIQ) 10, 4 (2018), 1–20.
  • Yang et al. (2019b) Wei Yang, Haotian Zhang, and Jimmy Lin. 2019b. Simple applications of BERT for ad hoc document retrieval. arXiv preprint arXiv:1903.10972 (2019).
  • Yang et al. (2019a) Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. 2019a. XLNet: Generalized Autoregressive Pretraining for Language Understanding. Advances in Neural Information Processing Systems 32 (2019), 5753–5763.
  • Yates et al. (2021) Andrew Yates, Rodrigo Nogueira, and Jimmy Lin. 2021. Pretrained Transformers for Text Ranking: BERT and Beyond. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining. 1154–1156.
  • Yilmaz et al. (2019a) Zeynep Akkalyoncu Yilmaz, Shengjin Wang, Wei Yang, Haotian Zhang, and Jimmy Lin. 2019a. Applying BERT to document retrieval with birch. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): System Demonstrations. 19–24.
  • Yilmaz et al. (2019b) Zeynep Akkalyoncu Yilmaz, Wei Yang, Haotian Zhang, and Jimmy Lin. 2019b. Cross-domain modeling of sentence-level evidence for document retrieval. In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). 3490–3496.
  • Yu et al. (2021a) HongChien Yu, Zhuyun Dai, and Jamie Callan. 2021a. PGT: Pseudo Relevance Feedback Using a Graph-Based Transformer. In European Conference on Information Retrieval.
  • Yu et al. (2021b) HongChien Yu, Chenyan Xiong, and Jamie Callan. 2021b. Improving Query Representations for Dense Retrieval with Pseudo Relevance Feedback. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management.
  • Zamani and Croft (2016) Hamed Zamani and W Bruce Croft. 2016. Embedding-Based Query Language Models. In Proceedings of the 2016 ACM international conference on the theory of information retrieval. 147–156.
  • Zamani and Croft (2017) Hamed Zamani and W Bruce Croft. 2017. Relevance-Based Word Embedding. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval. 505–514.
  • Zamani et al. (2016) Hamed Zamani, Javid Dadashkarimi, Azadeh Shakery, and W Bruce Croft. 2016. Pseudo-Relevance Feedback Based on Matrix Factorization. In Proceedings of the 25th ACM international on conference on information and knowledge management. 1483–1492.
  • Zhai and Lafferty (2001) Chengxiang Zhai and John Lafferty. 2001. Model-Based Feedback in the Language Modeling Approach to Information Retrieval. In Proceedings of the 10th ACM International Conference on Information and Knowledge Management. 403–410.
  • Zhan et al. (2020) Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, and Shaoping Ma. 2020. RepBERT: Contextualized Text Embeddings for First-Stage Retrieval. arXiv preprint arXiv:2006.15498 (2020).
  • Zheng et al. (2020) Zhi Zheng, Kai Hui, Ben He, Xianpei Han, Le Sun, and Andrew Yates. 2020. BERT-QE: Contextualized Query Expansion for Document Re-ranking. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings. 4718–4728.
  • Zheng et al. (2021) Zhi Zheng, Kai Hui, Ben He, Xianpei Han, Le Sun, and Andrew Yates. 2021. Contextualized query expansion via unsupervised chunk selection for text retrieval. Information Processing & Management 58, 5 (2021), 102672.
  • Zhuang et al. (2021) Shengyao Zhuang, Hang Li, and Guido Zuccon. 2021. Deep Query Likelihood Model for Information Retrieval. In The 43rd European Conference On Information Retrieval.
  • Zhuang et al. (2022) Shengyao Zhuang, Hang Li, and Guido Zuccon. 2022. Implicit Feedback for Dense Passage Retrieval: A Counterfactual Approach. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR).
  • Zhuang and Zuccon (2021) Shengyao Zhuang and Guido Zuccon. 2021. TILDE: Term Independent Likelihood Model for Passage Re-ranking. In The 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR).