A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (2024)

Sunhao Dai1,Weihao Liu1,Yuqi Zhou1,Liang Pang2,Rongju Ruan3,Gang Wang3,
Zhenhua Dong3,Jun Xu1,Ji-Rong Wen1
1Gaoling School of Artificial Intelligence, Renmin University of China
2CAS Key Laboratory of AI Safety, Institute of Computing Technology, CAS
3Huawei Noah’s Ark Lab
{sunhaodai, junxu}@ruc.edu.cn, pangliang@ict.ac.cn
Corresponding author.

Abstract

The proliferation of Large Language Models (LLMs) has led to an influx of AI-generated content (AIGC) on the internet, transforming the corpus of Information Retrieval (IR) systems from solely human-written to a coexistence with LLM-generated content. The impact of this surge in AIGC on IR systems remains an open question, with the primary challenge being the lack of a dedicated benchmark for researchers. In this paper, we introduce co*cktail, a comprehensive benchmark tailored for evaluating IR models in this mixed-sourced data landscape of the LLM era. co*cktail consists of 16 diverse datasets with mixed human-written and LLM-generated corpora across various text retrieval tasks and domains. Additionally, to avoid the potential bias from previously included dataset information in LLMs, we also introduce an up-to-date dataset, named NQ-UTD, with queries derived from recent events. Through conducting over 1,000 experiments to assess state-of-the-art retrieval models against the benchmarked datasets in co*cktail, we uncover a clear trade-off between ranking performance and source bias in neural retrieval models, highlighting the necessity for a balanced approach in designing future IR systems. We hope co*cktail can serve as a foundational resource for IR research in the LLM era, with all data and code publicly available at https://github.com/KID-22/co*cktail.

1 Introduction

Information retrieval (IR) systems, as the keystone in overcoming information overload, have seen widespread application across various domains, including search enginesLi etal. (2014), question answeringKarpukhin etal. (2020), dialog systemsChen etal. (2017), etc.A typical IR system aims at finding relevant documents or passages from a specific corpus in response to user’s queriesLi (2022); Zhao etal. (2023a).Traditionally, IR benchmarks, notably MS MARCONguyen etal. (2016), TRECCraswell etal. (2020) and BEIRThakur etal. (2021), have exclusively utilized human-written content.However, the recent surge in Artificial Intelligence Generated Content (AIGC) facilitated by advanced Large Language Models (LLMs) has revolutionized the IR landscapeDai etal. (2024a).This evolution has broadened the scope of IR systems, which now encompass a hybrid corpus composed of both human-written and LLM-generated contentAi etal. (2023); Zhu etal. (2023); Dai etal. (2024a), presenting new challenges and opportunities of IR in the LLM era.

For instance, recent studiesDai etal. (2024b); Xu etal. (2024) have shed light on a critical issue within neural retrieval models in the LLM era: a pronounced “source bias”.This bias, characterized by the preferential ranking of LLM-generated content over semantically equivalent human-written content, poses a significant threat to the IR ecosystem.Hence, the need to comprehensively understand such impact of LLM-generated content across different IR models and diverse IR domains and tasks has become more pressing, especially with the escalating prevalence of LLM-generated contentHanley and Durumeric (2023); Bengio etal. (2023).However, existing IR benchmarks either fail to reflect the real-world IR scenarios of the LLM era, as they solely contain human-written texts in their corpus, or provide limited datasets for exploring source bias.These shortcomings highlight the need for a comprehensive benchmark that accurately mirrors the current IR landscape, characterized by the integration of both human-written and LLM-generated texts within the corpus, to facilitate new research questions in this LLM era.

To fill this gap, we present a comprehensive benchmark tailored for IR in the LLM era, namely co*cktail, where the corpus contains both human-written and LLM-generated texts.co*cktail encompasses 16 retrieval datasets spanning different domains and tasks, enabling both in-domain and out-of-domain evaluation settings.To construct these datasets, we first select 15 widely used public human-written corpora from MS MARCONguyen etal. (2016), TRECCraswell etal. (2020), and BEIRThakur etal. (2021).Then, based on these human-written corpora, we use Llama2Touvron etal. (2023) to rewrite each text to preserve semantic equivalence while introducing LLM-generated corpora.Finally, we mix the original human-written corpora and the LLM-generated corpora to get the final co*cktail corpora and assign the same relevancy label for the corresponding query-document pairs.Furthermore, to address the potential biases introduced by the inherent knowledge of LLMs during the rewriting process, we collect an additional new dataset, NQ-UTD. This new dataset comprises 80 queries and 800 documents from recent events. It serves as a critical component of co*cktail, offering an essential perspective for assessing the performance of IR systems in the context of both pre- and post-LLM era datasets.

With the benchmarked diverse datasets in co*cktail, we then conduct comprehensive evaluations of over ten state-of-the-art (SOTA) retrieval models through more than 1,000 experiments. Our analysis firstly reinforces previous findings by Dai etal. (2024b), highlighting a pervasive bias towards LLM-generated content across nearly all 16 datasets in co*cktail among all neural retrieval models. Furthermore, the results illustrated in Figure1 reveal a distinct trade-off between ranking performance and source bias within these SOTA neural models.This observation suggests that while striving for high performance, these models may rely on inherent shortcuts, failing to grasp true semantic relevance and resulting in severe source bias. Hence, future work requires better considering a balance between performance and bias mitigation in the design of next-generation IR models.

A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (1)

In summary, our contributions are as follows:

(1) To the best of our knowledge, co*cktail is the first comprehensive benchmark with 16 datasets from a variety of domains and tasks tailed for IR research in the LLM era, where the corpus of each dataset contains both human-written texts and corresponding LLM-generated counterparts.

(2) We conduct extensive evaluations of state-of-the-art retrieval models using the co*cktail benchmark, assessing both retrieval accuracy and source bias. The evaluation tool, along with the codes, is open-sourced, facilitating ease of adaptation for evaluating new models and datasets.

(3) Extensive empirical studies reveal a clear trade-off between ranking performance and source bias in neural retrieval models. This finding underscores the importance of achieving a suitable balance between performance improvement and bias mitigation in future IR model designs.

DatasetTrainDevTestAvg. Word Length
DomainTaskRelevancy# Pairs# Query# Query# CorpusAvg. D/QQueryHuman DocLLM Doc
Collected Before the Emergence of LLM (similar-to\sim - 2021/04)
MS MARCOMisc.Passage-RetrievalBinary532,663-6,979542,2031.16.058.155.1
DL19Misc.Passage-RetrievalBinary--43542,20395.45.458.155.1
DL20Misc.Passage-RetrievalBinary--54542,20366.86.058.155.1
TREC-COVIDBio-MedicalBio-Medical IR3-level--50128,585430.110.6197.6165.9
NFCorpusBio-MedicalBio-Medical IR3-level110,5753243233,63338.23.3221.0206.7
NQWikipediaQuestion AnsweringBinary--3,446104,1941.29.286.981.0
HotpotQAWikipediaQuestion AnsweringBinary169,96354477,405111,1072.017.767.966.6
FiQA-2018FinanceQuestion AnsweringBinary14,04549964857,4502.610.8133.2107.8
Touché-2020Misc.Argument Retrieval3-level--49101,92218.46.6165.4134.4
CQADupStackStackEx.Dup. Ques.-RetrievalBinary--1,56339,9622.48.577.272.0
DBPediaWikipediaEntity-Retrieval3-level-67400145,03737.35.453.154.0
SCIDOCSScientificCitation-PredictionBinary--1,00025,2594.79.4169.7161.8
FEVERWikipediaFact CheckingBinary140,07966666,666114,5291.28.1113.491.1
Climate-FEVERWikipediaFact CheckingBinary--1,535101,3393.020.299.481.3
SciFactScientificFact CheckingBinary919-3005,1831.112.4201.8192.7
Collected After the Emergence of LLM (2023/11 - 2024/01)
NQ-UTDMisc.Question Answering3-level--808003.712.1101.194.7

2 Related Work

IR meets Large Language Models. Information retrieval (IR), the keystone of information access, has now significantly been reshaped by the emergence of large language models (LLMs)Zhao etal. (2023b); Ai etal. (2023); Dai etal. (2024a). This intersection has manifested in two pivotal ways.On the one hand, much effort has been made to utilize the advanced capabilities of LLMs to refine the whole retrieval pipelineZhu etal. (2023), including the integration of LLMs across various IR components, such as query rewritersSrinivasan etal. (2022), retrieversWang etal. (2023), re-rankersSun etal. (2023b), and readersShi etal. (2023).On the other hand, the capacity of LLMs for generating human-like text at scale has shifted the landscape of searchable corpora, which now includes both human-written and LLM-generated texts. This evolution has introduced new challenges, most notably the emergence of source biasDai etal. (2024b); Xu etal. (2024); Dai etal. (2024a), where neural retrievers exhibit a preference for LLM-generated content, potentially compromising the fairness and accuracy of search results. Moreover, the inclusion of LLM-generated content also raises concerns about privacyYao etal. (2023) and the dissemination of misinformationPan etal. (2023).In this paper, we focus on the second line and aim to establish a comprehensive benchmark for evaluating IR models in the LLM era, which can help understand the impact of LLM-generated content on IR systems.

Related Benchmarks.Historically, before the proliferation of LLM-generated content on the internet, several benchmarks were established for the evaluation of IR models, primarily utilizing corpora composed of human-written documents or passages. Notably, MS MARCOONguyen etal. (2016) and TRECCraswell etal. (2020) are widely used for supervised evaluation in IR research. Similarly, BEIRThakur etal. (2021) presents a diverse benchmark incorporating 18 datasets from various IR domains and tasks, tailored to zero-shot evaluation. Despite their contributions to advancing IR systems, these benchmarks fall short of reflecting the real-world scenarios of the current LLM era due to the absence of LLM-generated content in their corporaDai etal. (2024b); Xu etal. (2024); Dai etal. (2024a). This gap underscores the necessity for new benchmarks that include both human-written and LLM-generated texts, offering a more comprehensive and realistic evaluation environment to navigate the challenges and opportunities presented by the integration of LLMs into IR.

A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (2)

3 Benchmarking Retrieval Datasets

co*cktail establishes a comprehensive benchmark tailored for the evaluation of IR models in the LLM era, characterized by corpora containing both human-written and LLM-generated content. This benchmark aims to access the performance and biases of existing retrieval models while encouraging the creation of future IR systems that excel in robustness and generalization across diverse scenarios in the LLM era. To achieve this, we collect and construct 16 IR datasets, incorporating 15 datasets from the pre-LLM era and one newly developed dataset to ensure a wide representation of domains and tasks.The statistics for the 16 datasets benchmarked in co*cktail are summarized in Table1. We also list the dataset website links and the corresponding licenses in Appendix Table5.

3.1 Dataset Construction

A key concern in constructing datasets with LLMs is the potential for LLMs to have prior knowledge about queries, which could lead to an unfair evaluation. To address this, following previous worksDai etal. (2024b), we choose rewriting documents without incorporating query-related information rather than a full generation from LLMs with given queries, ensuring that any detected source bias is a genuine reflection of model preferences. Moreover, by avoiding full document generation based on queries, this approach also simplifies controlling the text length across different sources, ensuring consistency and further reducing potential biases in the evaluation.Specifically, we construct each dataset in co*cktail following a three-step process: 1) collecting seed datasets with human-written corpus and alongside relevancy labels for given queries; 2) leveraging LLMs to generate corresponding LLM-generated corpus with the human-written corpus as inputs; and 3) assigning relevancy labels to the LLM-generated content and given queries, ensuring seamless integration into the benchmark. An overview of our construction pipeline is shown in Figure2.

Seed Human Datasets.To cover as diverse IR domains and tasks as possible, we select the widely-used 15 datasets from MS MARCONguyen etal. (2016), TRECCraswell etal. (2020), and BEIRThakur etal. (2021) benchmark as our human-written corpus. These datasets span six domains and eight distinct tasks, facilitating comprehensive evaluations under both in-domain and out-of-domain settings. Moreover, as suggested in Sun etal. (2023b), we also construct a new test dataset NQ-UTD to ensure that relevance annotations have not been learned by LLMs.NQ-UTD comprises 80 queries distributed across eight domains, sourcing from recent events (from Nov. 2023 to Jan. 2024). We verify that neither GPT-4 nor Gemini, two of the most advanced LLMs to date, can not answer the questions within NQ-UTD, validating no prior knowledge of these questions in LLM.For details about the collection and annotation processes of NQ-UTD, along with detailed descriptions of all 16 datasets, please refer to AppendixA.1.

LLM-Generated Corpus.To ensure a fair evaluation of the impact of incorporating LLM-generation content, we utilize the widely used llama-2-7b-chat to rewrite each human-written text without changing its semantic information, with the following instructions: “Original Text: {{human-written text}} Please rewrite the above given text. Your answer must be formatted as follows: Rewritten Text: <your rewritten text>.” This post-instruction strengthens the instruction-following capabilities of LLMs, proving particularly effective for processing and generating responses to long input textsLiu etal. (2023).

Relevancy Label Assignment.Upon the creation of the LLM-generated corpus for each dataset, we assign the relevancy labels from the original <query, human-written doc> pairs to the new <query, LLM-generated doc> pairs. This process is underpinned by the premise that both sources of the content preserve nearly identical semantic information, which will be further verified in the following section.

All the datasets in co*cktail are organized in a standard format (corpus, queries, qrels) akin to the BEIR benchmarkThakur etal. (2021), facilitating ease of use and comparison. Examples for each dataset are showcased in Appendix Table12. More details about the data processing and quality control are provided in AppendixA.2.

3.2 Dataset Statistics and Analysis

As shown in Table1, there are minimal differences in average word length between human-written and LLM-generated documents, with the latter being marginally shorter.The detailed text length distribution, depicted in Appendix Figure5, further confirms the negligible variance in text length.Additionally, term-based Jaccard similarity and overlap distributions between LLM-generated and human-written documents, visualized in Appendix Figure6, show a noticeable distinction in terms despite similar semantic content.

Following practices in Dai etal. (2024b), our dataset construction process involved feeding only the original human-written text to the LLM for rewriting, without inputting any query-specific information to avoid introducing additional query-related bias.Furthermore, we employed the OpenAI embedding model111text-embedding-ada-002: https://platform.openai.com/docs/guides/embeddings to obtain semantic embeddings for both text sources, comparing them through cosine similarity. These comparisons, presented in Appendix Figure7, demonstrate a high similarity, indicating successful semantic retention of the original human-written text in the LLM-generated texts. Additionally, evaluations of several retrieval models on sole human-written or LLM-generated corpora showed consistent performance, as seen in Appendix Table8.Moreover, we also conduct human evaluations for quality verification for each dataset in Appendix Table10.These observations reinforce confidence in the quality of our newly constructed datasets, suggesting that the LLM-generated content maintains semantics comparable to human-written texts for IR tasks.The detailed dataset analysis is provided in AppendixA.3.

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO38.851.651.952.854.455.655.355.759.157.856.253.655.0
DL1957.878.376.472.571.772.176.477.578.381.075.274.375.9
DL2055.379.976.973.272.573.879.376.582.473.577.874.676.6
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID67.067.062.068.065.064.070.072.075.077.085.070.270.5
NFCorpus45.439.236.435.341.542.643.541.642.750.249.442.542.2
NQ45.762.560.060.065.069.665.667.570.469.168.964.065.9
HotpotQA84.375.362.972.184.688.182.188.589.193.990.782.982.7
FiQA-201824.424.724.928.928.134.027.231.836.335.538.730.431.0
Touché-202057.139.842.944.944.940.836.744.952.056.155.146.845.8
CQADupStack28.626.626.931.323.434.233.231.336.034.736.531.231.4
DBPedia36.655.350.648.056.362.156.956.859.658.632.952.253.7
SCIDOCS16.212.311.214.516.016.614.516.217.919.019.515.815.8
FEVER65.680.675.980.483.186.475.887.688.090.243.477.979.1
Climate-FEVER25.226.427.428.332.731.428.132.334.035.235.330.631.1
SciFact55.738.040.340.753.755.050.052.355.356.365.051.150.7
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD76.974.466.978.881.975.673.178.876.989.488.878.378.5
Averaged Result
Supervised50.669.968.466.266.267.270.369.973.370.869.767.569.2
Zero-shot48.447.945.348.652.053.950.554.056.458.954.651.852.2
All48.852.049.651.954.756.454.257.059.661.157.454.855.4

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO72.2-13.2-18.91.1-29.0-10.85.1-20.5-18.2-26.3-7.8-6.0-13.8
DL19108.7-51.311.0-0.8-55.0-21.4-39.6-53.9-111.1-18.3-53.7-25.9-39.4
DL20101.6-76.5-63.5-2.5-11.0-10.8-38.1-43.7-59.2-25.3-11.3-21.8-34.2
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID32.8-62.7-64.5-58.8-95.4-87.5-68.6-66.7-45.3-64.9-63.5-58.6-67.8
NFCorpus-29.5-30.6-50.5-23.2-44.8-99.5-49.2-17.3-37.9-66.1-38.9-44.3-45.8
NQ-17.9-26.6-16.7-12.7-41.8-37.9-25.0-25.2-47.2-61.6-33.6-31.5-32.8
HotpotQA51.0-1.1-3.5-13.60.2-5.7-5.45.2-8.536.614.86.41.9
FiQA-2018-8.2-38.18.8-33.3-6.4-38.3-29.4-52.8-12.7-42.3-35.7-26.2-28.0
Touché-2020-21.4-25.6-76.0-36.1-36.1-29.810.9-9.4-66.4-127.3-66.4-44.0-46.2
CQADupStack22.4-45.1-39.4-19.8-10.3-22.2-6.6-67.1-24.4-30.5-8.7-22.9-27.4
DBPedia18.6-5.4-19.3-25.82.54.5-11.6-24.6-21.5-13.35.5-8.2-10.9
SCIDOCS2.521.1-21.4-9.7-20.00.0-26.2-22.2-16.8-27.4-35.9-14.2-15.8
FEVER-26.22.5-2.4-87.1-16.8-20.4-22.2-1.4-26.8-4.411.5-17.6-16.7
Climate-FEVER6.3-15.2-16.1-109.9-22.6-12.7-17.8-15.5-10.65.1-89.0-27.1-30.4
SciFact1.1-52.6-14.9-29.6-53.31.5-21.6-29.4-38.7-8.2-38.2-25.8-28.5
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD37.2-35.5-27.8-22.3-47.1-31.4-56.3-73.1-30.9-51.9-16.9-32.4-39.3
Averaged Result
Supervised94.2-47.0-23.8-0.7-31.7-14.3-24.2-39.4-62.8-23.3-24.3-17.9-29.2
Zero-shot5.3-24.2-26.4-37.1-30.1-29.2-25.3-30.7-29.8-35.1-30.4-26.6-29.8
All22.0-28.5-25.9-30.3-30.4-26.4-25.1-32.3-36.0-32.9-29.2-25.0-29.7

4 Benchmarking Evaluation Protocol

Evaluation Framework. To standardize the assessment of IR models within the LLM era, we also develop a Python framework that combines user-friendliness with comprehensive evaluation capabilities. Built on the foundation of the BEIRThakur etal. (2021), our framework inherits its best features, including the ability to easily replicate experiments from open-sourced repositories and incorporate new models and datasets. A key innovation of our framework is its ability to conduct evaluations using either individual or mixed corpora, accommodating the mixture of human-written and LLM-generated content characteristic of the LLM era. These features make our framework an invaluable tool for advancing IR research and application in both academia and industry.

Evaluation Metrics.Aligned with the BEIR benchmarkThakur etal. (2021), we select Normalized Discounted Cumulative Gain (NDCG@K𝐾Kitalic_K) as our primary metric to assess retrieval accuracy, given its robustness in capturing the effectiveness of IR systems across tasks with binary and graded relevance judgments.Following previous studiesDai etal. (2024b); Xu etal. (2024), we choose K=1𝐾1K=1italic_K = 1 since the top-1111 item in the retrieved list is most likely to be viewed and clicked by users. To provide a more comprehensive evaluation, we also report results on K=3𝐾3K=3italic_K = 3 and K=5𝐾5K=5italic_K = 5 in AppendixC.

Moreover, in the multi-sourced corpus shaped by LLMs, the evaluation of IR systems necessitates not only mere accuracy metrics but also a critical assessment of source biasDai etal. (2024b). This bias, as evidenced by ranking LLM-generated content higher than human-written content, poses a significant challenge in today’s IR ecosystem. To quantify and normalize source bias in different datasets, we follow previous worksDai etal. (2024b); Xu etal. (2024) and adopt the RelativeΔRelativeΔ\text{Relative}~{}~{}\DeltaRelative roman_Δ. This metric captures the relative percentage difference in NDCG scores between human-written and LLM-generated content, which is defined as:

RelativeΔ=NDCGHumanNDCGLLM(NDCGHuman+NDCGLLM)/2×100%,RelativeΔsubscriptNDCGHumansubscriptNDCGLLMsubscriptNDCGHumansubscriptNDCGLLM2percent100\text{Relative $\Delta$}=\frac{\texttt{NDCG}_{\text{Human}}-\texttt{NDCG}_{%\text{LLM}}}{(\texttt{NDCG}_{\text{Human}}+\texttt{NDCG}_{\text{LLM}})/2}%\times 100\%,Relative roman_Δ = divide start_ARG NDCG start_POSTSUBSCRIPT Human end_POSTSUBSCRIPT - NDCG start_POSTSUBSCRIPT LLM end_POSTSUBSCRIPT end_ARG start_ARG ( NDCG start_POSTSUBSCRIPT Human end_POSTSUBSCRIPT + NDCG start_POSTSUBSCRIPT LLM end_POSTSUBSCRIPT ) / 2 end_ARG × 100 % ,

where the NDCGHumansubscriptNDCGHuman\texttt{NDCG}_{\text{Human}}NDCG start_POSTSUBSCRIPT Human end_POSTSUBSCRIPT and NDCGLLMsubscriptNDCGLLM\texttt{NDCG}_{\text{LLM}}NDCG start_POSTSUBSCRIPT LLM end_POSTSUBSCRIPT denote the NDCG scores attributed to human-written and LLM-generated content, respectively. Note that RelativeΔ>0RelativeΔ0\text{Relative}~{}\Delta>0Relative roman_Δ > 0 indicates that IR models rank human-written content higher than LLM-generated content, and conversely, RelativeΔ<0RelativeΔ0\text{Relative}~{}\Delta<0Relative roman_Δ < 0 indicates the opposite tendency. The absolute value of RelativeΔRelativeΔ\text{Relative}~{}\DeltaRelative roman_Δ reflects the extent of source bias.

5 Benchmarking Retrieval Models

In this section, we delve into the evaluation and analysis of various retrieval models utilizing the constructed co*cktail benchmarked datasets.

5.1 Retrieval Models

Following the BEIR benchmarkThakur etal. (2021), we focus on evaluating the advanced state-of-the-art transformer-based neural retrieval models. Besides the widely used lexical retriever BM25Robertson etal. (2009), our experiments include two main types of neural retrieval models:

Neural Retriever. We utilize and fine-tune the following two most commonly used pre-trained language models on the MS MARCO datasetNguyen etal. (2016) using the official training script222https://github.com/beir-cellar/beir/blob/main/examples/retrieval/training/train_msmarco_v3.py from BEIR: 1) BERTDevlin etal. (2019); 2) RoBERTaLiu etal. (2019). Additionally, we evaluate the performance of state-of-the-art models also trained on MS MARCO, employing officially released checkpoints:3) ANCEXiong etal. (2020); 4) TAS-BHofstätter etal. (2021); 5) ContrieverIzacard etal. (2022); 6) coCondenserGao and Callan (2022); 7) RetroMAEXiao etal. (2022) ; 8) DRAGONLin etal. (2023).

Neural Re-ranker. For re-ranking, we employ two state-of-the-art models with their publicly available official pre-trained checkpoints: 1) CEWang etal. (2020); 2) monoT5Raffel etal. (2020).

In our experiments, unless specified otherwise, neural re-rankers re-rank the top-100 documents retrieved by BM25.Detailed information on the benchmarked models, including the publicly available official pre-trained checkpoints and implementation details, can be found in AppendixB.

5.2 Benchmarked Results

We conduct extensive evaluations with more than 1,000 experiments on the co*cktail benchmarked datasets.The results of retrieval accuracy and source bias across all benchmarked retrieval models and datasets are reported in Table2 and Table3333Note that SciFact was regenerated with our prompt, leading to slight result discrepancies from Dai etal. (2024b), respectively. Figure1 shows average results of neural retrieval models across all datasets. From the results, we have the following key observations:

Neural models consistently exhibit source bias towards LLM-generated content.This bias is evident across neural retrieval models, spanning both in-domain and, more significantly, out-of-distribution datasets. This trend persists in data from both pre-LLM and LLM eras.Remarkably, the average RelativeΔRelativeΔ\text{Relative}~{}\DeltaRelative roman_Δ across all neural models on the datasets surpasses 25%percent25-25\%- 25 %.These findings further support the findings of Dai etal. (2024b) and verify the widespread source bias in neural retrieval models, regardless of the domain or task, highlighting an urgent need to address this bias.

Stronger neural retrieval models exhibit more severe source bias.The results illustrated in Figure1 underscore a significant trade-off faced by neural retrieval models: advancements in ranking performance often come with an increase in source bias.This trend suggests that these SOTA neural models may not fully understand semantic relevance.Instead, these models tend to leverage inherent shortcuts to enhance performance, inadvertently leading to increased bias.This phenomenon suggests that attempts to boost model performance could unintentionally magnify source bias issues, underlining the challenge of advancing model capabilities without leading to severe source bias.

MetricBM25+ CE+ monoT5DRAGON+ CE+ monoT5
NDCG@148.861.157.459.6 59.8 58.5
RelativeΔRelativeΔ\text{Relative}~{}\DeltaRelative roman_Δ22.0-32.9-29.2-36.0 -36.5 -33.7

Neural re-rankers generalize better but are still biased towards LLM-generated content.Neural re-rankers, while achieving superior ranking performance on most datasets and showing enhanced generalization capabilities compared to neural retrievers, are not exempt from pervasive source bias.Specifically, re-ranking models like CE and monoT5 exhibit a significant bias towards LLM-generated content, sometimes even more pronounced than that observed in neural retrievers. This is particularly evident in datasets such as NQ and Touché-2020, further emphasizing the widespread nature of source bias in PLM-based neural retrieval models.

Bias in the first retrieval stage tends to propagate and even amplify during the re-ranking stage.This trend is particularly notable in datasets like NFCorpus and Touché-2020, where the source bias observed in the first retrieval stage persists and tends to intensify during the second-stage re-ranking.Moreover, as detailed in Table4, enhancing the efficiency of the first-stage retrieval by replacing BM25 with DRAGON does not necessarily improve performance in the subsequent re-ranking phase. However, the source bias inherent in the first-stage retriever significantly impacts and may even magnify in subsequent re-ranking. This observation underscores the critical need for developing holistic approaches to mitigate bias throughout the retrieval pipeline, ensuring fairness and accuracy in the whole IR system.

5.3 Further Analysis

Impact of Different Pooling Strategies.Pooling strategies in PLMs are critical for aggregating information from the token embeddings for downstream semantic matching.We explore the ranking performance and source bias on BERT and RoBERTa w.r.t. different pooling strategies, including CLS token, max token, last tokenMuennighoff (2022), meanReimers and Gurevych (2019), and weighted mean poolingMuennighoff (2022). The averaged results of all benchmarked datasets in co*cktail are shown in Figure3. As we can see, the ranking performance and degree of source bias vary significantly with the specific pooling strategy.

Weighted mean pooling demonstrates the most effectiveness for ranking, which can be attributed to the nuanced semantic understanding by positionally weighting tokens, thus enhancing document relevance matching. Yet, this strategy also incurs the most severe source bias, possibly because LLM-generated texts have distinctive structural or stylistic features that become more pronounced and amplified under weighted aggregation.

Conversely, max pooling, which selects the maximum value across each dimension from the token embeddings, appears to be the least effective. This could be due to its focus on the most dominant features within the text, potentially overlooking the broader contextual nuances captured by other strategies. The dominance of specific features might not always align with the relevance signals needed for accurate document ranking, explaining the lower performance and less bias.

Other strategies, such as mean, CLS token, and last token pooling, strike a balance by capturing overall text representations while still highlighting key tokens. However, the results show that while they are capable of supporting effective document ranking, they still suffer from severe source bias with Relative ΔΔ\Deltaroman_Δ far below 10%percent10-10\%- 10 %.

Overall, the variance in ranking performance and source bias across pooling strategies underscores the critical role of model architecture in retrieval model design. This analysis suggests that while weighted mean pooling offers enhanced ranking capabilities, it comes with a trade-off of increased source bias. Future work can explore hybrid or innovative pooling methods to balance performance with bias mitigation.

A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (5)

Impact of PLM Model Size.Our investigation extends to the influence of model size on ranking performance and source bias, utilizing BERT models of varying sizes: mini, small, base, and large. As illustrated in Figure4, an increase in model size leads to a notable enhancement in ranking performance. However, this improvement is paralleled by a rise in source bias (i.e., more negative).

The trend indicates that larger models, with their enhanced semantic understanding and processing capabilities, are more effective in judging document relevance, boosting their ranking performance. However, the increased capabilities may also make them more sensitive to the nuanced distinctions between human-written and LLM-generated texts, amplifying the source bias. Consequently, while the advanced performance of larger models is promising, it is coupled with a more severe bias risk, posing a significant challenge that merits more exploration in future research.

6 Conclusion and Future Work

This study proposes co*cktail, the first comprehensive benchmark consisting of 16 diverse datasets with mixed human-written and LLM-generated corpora.Alongside this, we present an evaluation tool equipped with standardized data formats and easily adaptable evaluation codes for a wide array of retrieval models.This tool is designed to systematically evaluate ranking performance and source bias in IR systems, both in supervised and zero-shot settings, paving the way for the development of more robust next-generation IR models.

Utilizing co*cktail, we conduct an extensive evaluation of over ten state-of-the-art neural retrieval models through more than 1,000 experiments.Our findings reveal a clear trade-off between ranking performance and source bias in neural retrieval models, underscoring the difficulty of enhancing model performance without increasing bias towards LLM-generated content.This challenge emphasizes the need for a suitable balance between performance enhancement and bias mitigation in the design of future IR models.

The newly collected and annotated NQ-UTD dataset comprises queries derived from recent events, featuring all content not yet incorporated into the pre-training knowledge of most LLMs.This characteristic renders it a valuable resource for fairly evaluating the effectiveness of LLMs in processing new, unseen data, especially for LLM-based retrieval or question-answering systems.

Limitations

We believe our proposed co*cktail benchmark is a foundational step toward advancing IR research in the LLM era. Nonetheless, our work still has several limitations for future research efforts. First, while our benchmarked 16 datasets encompass a broad range of IR domains and tasks, the IR field is continually evolving, with new areas of interest emerging regularly. Future updates to the co*cktail benchmark could benefit from including datasets from other IR domains, such as legal information retrievalSansone and Sperlí (2022). This expansion would not only diversify but also enhance the benchmark’s utility across more specialized areas. Second, the scale of our NQ-UTD dataset is currently limited to 800 query-passage pairs, primarily due to the high costs of human annotation. This encompasses financial costs, the time required to develop annotation guidelines, train annotators, and perform manual audits and validations. Future initiatives could focus on expanding the NQ-UTD dataset, possibly by employing LLMs to support and streamline the annotation processZhang etal. (2023). Such an approach could facilitate broader coverage and richer labeled query-passage pairs. Third, the construction of LLM-generated corpus in our benchmarked 16 datasets was significantly influenced by the inference costs of LLM, leading us to rely exclusively on Llama2Touvron etal. (2023) for generating LLM-based content. However, the real-world IR landscape is shaped by a variety of LLMs. Future research might include content generated by an array of LLMs, such as OpenAI ChatGPT and Google Gemini, to mirror the diversity of LLM-generated content more accurately. Despite these limitations, we envision the co*cktail benchmark as a valuable resource for IR research in the LLM era, offering a foundation upon which next-generation IR models can be built.

Ethics Statement

We commit to maintaining the highest ethical standards in our research, strictly adhering to ethical guidelines, applicable licenses, and laws.The 15 datasets utilized in this study are sourced from public repositories and have been employed within the bounds of their respective licenses and legal constraints.For the collected NQ-UTD dataset, we followed stringent ethical protocols, ensuring the corpus collected from online news sources was carefully screened and double-checked to remove any personally identifiable or sensitive information. Participants involved in data annotation were thoroughly informed about the process and provided informed consent before their participation.

We recognize the potential risks posed by large language models (LLMs), such as generating harmful, biased, or incorrect contentPan etal. (2023); Chen and Shu (2023). The Llama2 model, used in this study for corpus generation, is not immune to these issues and may inadvertently produce misleading content.We have taken measures to minimize these risks, ensuring our benchmark aligns with ethical standards.Despite the acknowledged risks associated with LLMs, we assert that the scientific benefits and insights derived from our benchmark far outweigh potential concerns. Our resources are intended solely for scientific research, designed to foster advancements in information retrieval within this LLM era.

Acknowledgements

Jun Xu is the corresponding author.This work was funded by the National Key R&D Program of China (2023YFA1008704), the National Natural Science Foundation of China (No. 62377044, No.62376275), Beijing Key Laboratory of Big Data Management and Analysis Methods, Major Innovation & Planning Interdisciplinary Platform for the “Double-First Class” Initiative, PCC@RUC, funds for building world-class universities (disciplines) of Renmin University of China.This work was partially done at Engineering Research Center of Next-Generation Intelligent Search and Recommendation, Ministry of Education.

References

  • Ai etal. (2023)Qingyao Ai, Ting Bai, Zhao Cao, YiChang, Jiawei Chen, Zhumin Chen, Zhiyong Cheng, Shoubin Dong, Zhicheng Dou, Fuli Feng, etal. 2023.Information retrieval meets large language models: A strategic report from chinese ir community.AI Open, 4:80–90.
  • Bengio etal. (2023)Yoshua Bengio, Geoffrey Hinton, Andrew Yao, Dawn Song, Pieter Abbeel, YuvalNoah Harari, Ya-Qin Zhang, Lan Xue, Shai Shalev-Shwartz, Gillian Hadfield, etal. 2023.Managing ai risks in an era of rapid progress.arXiv preprint arXiv:2310.17688.
  • Bondarenko etal. (2020)Alexander Bondarenko, Maik Fröbe, Meriem Beloucif, Lukas Gienapp, Yamen Ajjour, Alexander Panchenko, Chris Biemann, Benno Stein, Henning Wachsmuth, Martin Potthast, etal. 2020.Overview of touché 2020: argument retrieval.In Experimental IR Meets Multilinguality, Multimodality, and Interaction: 11th International Conference of the CLEF Association, CLEF 2020, Thessaloniki, Greece, September 22–25, 2020, Proceedings 11, pages 384–395. Springer.
  • Boteva etal. (2016)Vera Boteva, Demian Gholipour, Artem Sokolov, and Stefan Riezler. 2016.A full-text learning to rank dataset for medical information retrieval.In Advances in Information Retrieval: 38th European Conference on IR Research, ECIR 2016, Padua, Italy, March 20–23, 2016. Proceedings 38, pages 716–722. Springer.
  • Chen and Shu (2023)Canyu Chen and Kai Shu. 2023.Can llm-generated misinformation be detected?arXiv preprint arXiv:2309.13788.
  • Chen etal. (2017)Hongshen Chen, Xiaorui Liu, Dawei Yin, and Jiliang Tang. 2017.A survey on dialogue systems: Recent advances and new frontiers.Acm Sigkdd Explorations Newsletter, 19(2):25–35.
  • Cohan etal. (2020)Arman Cohan, Sergey Feldman, IzBeltagy, Doug Downey, and Daniel Weld. 2020.SPECTER: Document-level representation learning using citation-informed transformers.In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2270–2282, Online. Association for Computational Linguistics.
  • Craswell etal. (2021)Nick Craswell, Bhaskar Mitra, Emine Yilmaz, and Daniel Campos. 2021.Overview of the trec 2020 deep learning track.
  • Craswell etal. (2020)Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and EllenM Voorhees. 2020.Overview of the trec 2019 deep learning track.arXiv preprint arXiv:2003.07820.
  • Dai etal. (2024a)Sunhao Dai, Chen Xu, Shicheng Xu, Liang Pang, Zhenhua Dong, and Jun Xu. 2024a.Unifying bias and unfairness in information retrieval: A survey of challenges and opportunities with large language models.Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.
  • Dai etal. (2024b)Sunhao Dai, Yuqi Zhou, Liang Pang, Weihao Liu, Xiaolin Hu, Yong Liu, Xiao Zhang, Gang Wang, and Jun Xu. 2024b.Neural retrievers are biased towards llm-generated content.Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.
  • Devlin etal. (2019)Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019.BERT: Pre-training of deep bidirectional transformers for language understanding.In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4171–4186.
  • Diggelmann etal. (2020)Thomas Diggelmann, Jordan Boyd-Graber, Jannis Bulian, Massimiliano Ciaramita, and Markus Leippold. 2020.Climate-fever: A dataset for verification of real-world climate claims.arXiv preprint arXiv:2012.00614.
  • Gao and Callan (2022)Luyu Gao and Jamie Callan. 2022.Unsupervised corpus aware language model pre-training for dense passage retrieval.In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2843–2853.
  • Hanley and Durumeric (2023)HansWA Hanley and Zakir Durumeric. 2023.Machine-made media: Monitoring the mobilization of machine-generated articles on misinformation and mainstream news websites.arXiv preprint arXiv:2305.09820.
  • Hasibi etal. (2017)Faegheh Hasibi, Fedor Nikolaev, Chenyan Xiong, Krisztian Balog, SveinErik Bratsberg, Alexander Kotov, and Jamie Callan. 2017.Dbpedia-entity v2: a test collection for entity search.In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1265–1268.
  • Hofstätter etal. (2021)Sebastian Hofstätter, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy Lin, and Allan Hanbury. 2021.Efficiently teaching an effective dense retriever with balanced topic aware sampling.In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 113–122.
  • Hoogeveen etal. (2015)Doris Hoogeveen, KarinM Verspoor, and Timothy Baldwin. 2015.Cqadupstack: A benchmark data set for community question-answering research.In Proceedings of the 20th Australasian document computing symposium, pages 1–8.
  • Izacard etal. (2022)Gautier Izacard, Mathilde Caron, Lucas Hosseini, Sebastian Riedel, Piotr Bojanowski, Armand Joulin, and Edouard Grave. 2022.Unsupervised dense information retrieval with contrastive learning.Transactions on Machine Learning Research.
  • Karpukhin etal. (2020)Vladimir Karpukhin, Barlas Oguz, Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih. 2020.Dense passage retrieval for open-domain question answering.In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6769–6781.
  • Kwiatkowski etal. (2019)Tom Kwiatkowski, Jennimaria Palomaki, Olivia Redfield, Michael Collins, Ankur Parikh, Chris Alberti, Danielle Epstein, Illia Polosukhin, Jacob Devlin, Kenton Lee, etal. 2019.Natural questions: a benchmark for question answering research.Transactions of the Association for Computational Linguistics, 7:453–466.
  • Lan etal. (2019)Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019.Albert: A lite bert for self-supervised learning of language representations.arXiv preprint arXiv:1909.11942.
  • Li (2022)Hang Li. 2022.Learning to rank for information retrieval and natural language processing.Springer Nature.
  • Li etal. (2014)Hang Li, Jun Xu, etal. 2014.Semantic matching in search.Foundations and Trends® in Information Retrieval, 7(5):343–469.
  • Lin etal. (2023)Sheng-Chieh Lin, Akari Asai, Minghan Li, Barlas Oguz, Jimmy Lin, Yashar Mehdad, Wen-tau Yih, and Xilun Chen. 2023.How to train your dragon: Diverse augmentation towards generalizable dense retrieval.arXiv preprint arXiv:2302.07452.
  • Liu etal. (2023)Yijin Liu, Xianfeng Zeng, Fandong Meng, and Jie Zhou. 2023.Instruction position matters in sequence generation with large language models.arXiv preprint arXiv:2308.12097.
  • Liu etal. (2019)Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019.Roberta: A robustly optimized bert pretraining approach.arXiv preprint arXiv:1907.11692.
  • Maia etal. (2018)Macedo Maia, Siegfried Handschuh, André Freitas, Brian Davis, Ross McDermott, Manel Zarrouk, and Alexandra Balahur. 2018.Www’18 open challenge: financial opinion mining and question answering.In Companion proceedings of the the web conference 2018, pages 1941–1942.
  • Muennighoff (2022)Niklas Muennighoff. 2022.Sgpt: Gpt sentence embeddings for semantic search.arXiv preprint arXiv:2202.08904.
  • Nguyen etal. (2016)Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and LiDeng. 2016.Ms marco: A human generated machine reading comprehension dataset.choice, 2640:660.
  • Pan etal. (2023)Yikang Pan, Liangming Pan, Wenhu Chen, Preslav Nakov, Min-Yen Kan, and WilliamYang Wang. 2023.On the risk of misinformation pollution with large language models.arXiv preprint arXiv:2305.13661.
  • Raffel etal. (2020)Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and PeterJ Liu. 2020.Exploring the limits of transfer learning with a unified text-to-text transformer.The Journal of Machine Learning Research, 21(1):5485–5551.
  • Reimers and Gurevych (2019)Nils Reimers and Iryna Gurevych. 2019.Sentence-BERT: Sentence embeddings using Siamese BERT-networks.In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3982–3992.
  • Robertson etal. (2009)Stephen Robertson, Hugo Zaragoza, etal. 2009.The probabilistic relevance framework: Bm25 and beyond.Foundations and Trends® in Information Retrieval, 3(4):333–389.
  • Sansone and Sperlí (2022)Carlo Sansone and Giancarlo Sperlí. 2022.Legal information retrieval systems: State-of-the-art and open issues.Information Systems, 106:101967.
  • Shi etal. (2023)Weijia Shi, Sewon Min, Michihiro Yasunaga, Minjoon Seo, Rich James, Mike Lewis, Luke Zettlemoyer, and Wen-tau Yih. 2023.Replug: Retrieval-augmented black-box language models.arXiv preprint arXiv:2301.12652.
  • Srinivasan etal. (2022)Krishna Srinivasan, Karthik Raman, Anupam Samanta, Lingrui Liao, Luca Bertelli, and Michael Bendersky. 2022.Quill: Query intent with large language models using retrieval augmentation and multi-stage distillation.In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: Industry Track, pages 492–501.
  • Sun etal. (2023a)Weiwei Sun, Lingyong Yan, Zheng Chen, Shuaiqiang Wang, Haichao Zhu, Pengjie Ren, Zhumin Chen, Dawei Yin, Maarten deRijke, and Zhaochun Ren. 2023a.Learning to tokenize for generative retrieval.In Thirty-seventh Conference on Neural Information Processing Systems.
  • Sun etal. (2023b)Weiwei Sun, Lingyong Yan, Xinyu Ma, Shuaiqiang Wang, Pengjie Ren, Zhumin Chen, Dawei Yin, and Zhaochun Ren. 2023b.Is ChatGPT good at search? investigating large language models as re-ranking agents.In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 14918–14937.
  • Thakur etal. (2021)Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, and Iryna Gurevych. 2021.BEIR: A heterogeneous benchmark for zero-shot evaluation of information retrieval models.In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track.
  • Thorne etal. (2018)James Thorne, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. 2018.FEVER: a large-scale dataset for fact extraction and VERification.In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 809–819, New Orleans, Louisiana. Association for Computational Linguistics.
  • Touvron etal. (2023)Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, etal. 2023.Llama 2: Open foundation and fine-tuned chat models.arXiv preprint arXiv:2307.09288.
  • Voorhees etal. (2021)Ellen Voorhees, Tasmeer Alam, Steven Bedrick, Dina Demner-Fushman, WilliamR Hersh, Kyle Lo, Kirk Roberts, Ian Soboroff, and LucyLu Wang. 2021.Trec-covid: constructing a pandemic information retrieval test collection.In ACM SIGIR Forum, volume54, pages 1–12. ACM New York, NY, USA.
  • Wadden etal. (2020)David Wadden, Shanchuan Lin, Kyle Lo, LucyLu Wang, Madeleine van Zuylen, Arman Cohan, and Hannaneh Hajishirzi. 2020.Fact or fiction: Verifying scientific claims.In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7534–7550, Online. Association for Computational Linguistics.
  • Wang etal. (2023)Liang Wang, Nan Yang, and Furu Wei. 2023.Query2doc: Query expansion with large language models.arXiv preprint arXiv:2303.07678.
  • Wang etal. (2020)Wenhui Wang, Furu Wei, LiDong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020.Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers.Advances in Neural Information Processing Systems, 33:5776–5788.
  • Xiao etal. (2022)sh*tao Xiao, Zheng Liu, Yingxia Shao, and Zhao Cao. 2022.Retromae: Pre-training retrieval-oriented language models via masked auto-encoder.In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 538–548.
  • Xiong etal. (2020)Lee Xiong, Chenyan Xiong, YeLi, Kwok-Fung Tang, Jialin Liu, Paul Bennett, Junaid Ahmed, and Arnold Overwijk. 2020.Approximate nearest neighbor negative contrastive learning for dense text retrieval.arXiv preprint arXiv:2007.00808.
  • Xu etal. (2024)Shicheng Xu, Danyang Hou, Liang Pang, Jingcheng Deng, Jun Xu, Huawei Shen, and Xueqi Cheng. 2024.Invisible relevance bias: Text-image retrieval models prefer ai-generated images.Proceedings of the 47th international ACM SIGIR conference on research and development in information retrieval.
  • Yang etal. (2018)Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William Cohen, Ruslan Salakhutdinov, and ChristopherD. Manning. 2018.HotpotQA: A dataset for diverse, explainable multi-hop question answering.In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380, Brussels, Belgium. Association for Computational Linguistics.
  • Yao etal. (2023)Yifan Yao, Jinhao Duan, Kaidi Xu, Yuanfang Cai, Eric Sun, and Yue Zhang. 2023.A survey on large language model (llm) security and privacy: The good, the bad, and the ugly.arXiv preprint arXiv:2312.02003.
  • Zhang etal. (2023)Ruoyu Zhang, Yanzeng Li, Yongliang Ma, Ming Zhou, and Lei Zou. 2023.LLMaAA: Making large language models as active annotators.In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 13088–13103.
  • Zhao etal. (2023a)WayneXin Zhao, Jing Liu, Ruiyang Ren, and Ji-Rong Wen. 2023a.Dense text retrieval based on pretrained language models: A survey.ACM Trans. Inf. Syst.
  • Zhao etal. (2023b)WayneXin Zhao, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, Beichen Zhang, Junjie Zhang, Zican Dong, etal. 2023b.A survey of large language models.arXiv preprint arXiv:2303.18223.
  • Zhou etal. (2022)Yujia Zhou, Jing Yao, Zhicheng Dou, Ledell Wu, and Ji-Rong Wen. 2022.Dynamicretriever: A pre-training model-based ir system with neither sparse nor dense index.arXiv preprint arXiv:2203.00537.
  • Zhu etal. (2023)Yutao Zhu, Huaying Yuan, Shuting Wang, Jiongnan Liu, Wenhan Liu, Chenlong Deng, Zhicheng Dou, and Ji-Rong Wen. 2023.Large language models for information retrieval: A survey.arXiv preprint arXiv:2308.07107.

Appendix A Dataset Details

In this section, we provide details about the datasets in co*cktail.Table5 lists the dataset website links and the corresponding licenses.

A.1 Detailed Description of Datasets

DatasetOfficial Website LinkLicense
MS MARCOhttps://microsoft.github.io/msmarco/MIT License
DL19https://microsoft.github.io/msmarco/TREC-Deep-Learning-2019CC BY 4.0 license
DL20https://microsoft.github.io/msmarco/TREC-Deep-Learning-2020CC BY 4.0 license
TREC-COVIDhttps://ir.nist.gov/covidSubmit/index.htmlDataset License Agreement
NFCorpushttps://www.cl.uni-heidelberg.de/statnlpgroup/nfcorpus/
NQhttps://ai.google.com/research/NaturalQuestionsCC BY-SA 3.0 license
HotpotQAhttps://hotpotqa.github.ioCC BY-SA 4.0 license
FiQA-2018https://sites.google.com/view/fiqa/
Touché-2020https://webis.de/events/touche-20/shared-task-1.htmlCC BY 4.0 license
CQADupStackhttp://nlp.cis.unimelb.edu.au/resources/cqadupstack/License 2.0 license
DBPediahttps://github.com/iai-group/DBpedia-Entity/CC BY-SA 3.0 license
SCIDOCShttps://allenai.org/data/scidocsGNU General Public License v3.0 license
FEVERhttp://fever.aiCC BY-SA 3.0 license
Climate-FEVERhttp://climatefever.ai
SciFacthttps://github.com/allenai/scifactCC BY-NC 2.0 license
NQ-UTDhttps://github.com/KID-22/co*cktailMIT License

Domain# Queries# Passages# Relevancy Scores# Source
012GoogleXRedditWikipediaQuoraFacebook
Sports121206617376022101990
News1313078143868231113150
Scientific1010068122059108698
Technology12120921276464121321
Autos88050151551617114
Music1010060182267685140
Movies770418215323930
Games880489235639840
All80800503942034787854797833

DomainQuestionReference Answer
SportsWho is the MVP of the first NBA In-Season Tournament?LeBron James
SportsWho wins 2023 FIFA Club World Cup?Manchester City F.C.
NewsWhere was the 44th Gulf Cooperation Council (GCC) Summit held?Doha, Qatar
NewsWhich country will take over as the rotating chair of the APEC in 2024?Peru
ScientificWhich paper won NeurIPS2023 Test of Time Award?Distributed Representations of Words and Phrases and their Compositionality
ScientificWhich university designed AI Coscientist?Carnegie Mellon University
TechnologyWhich company released the machine learning framework MLX?Apple
TechnologyWhat is the maximum number of cores in Intel fifth-generation Xeon CPU?64
AutosWhen did Xiaomi announce SU7?Dec 2023
AutosThe country with the largest automobile export volume in 2023?China
MusicWhich singer was named the 2023 Person of the Year by Time magazine?Taylor Swift
MusicWhen did Les McCann pass away?December 29, 2023
MovesWhich movie was the highest-grossing film worldwide in 2023?Barbie
MovesWhich actress won the Best Actress award at the European Film Awards?Sandra Wheeler
GamesWhich team is the champion of the League of Legends S13 Finals?SKT T1
GamesWhich one is the best-selling games in the US 2023?Hogwarts Legacy

A.1.1 NQ-UTD

The Natural Question-Up To Date (NQ-UTD) dataset comprises 80 questions focusing on recent hotspot events from November 2023 to January 2024. These questions span eight domains: Sports, News, Science, Technology, Autos, Music, Movies, and Games, with approximately 10 queries per domain related to the latest events, questioning about their times, locations, or other specifics. To gather relevant passages for each query, we searched these questions on platforms like Google, X (formerly Twitter), Reddit, Wikipedia, Quora, and Facebook.Following the practice in Sun etal. (2023b), we also conducted searches using keywords to retrieve passages that are partially relevant but do not directly answer the questions. We also queried these questions with the latest state-of-the-art LLMs, gpt-4-1106-preview and gemini-pro, which have their knowledge bases updated only until April 2023. The results show their 0% accuracy in answering questions from our test set, confirming that these LLMs had no pre-knowledge about these questions. The collected 800 documents were then manually annotated for relevance by the authors and their highly educated colleagues.Each document received a relevance score: 0 for not relevant, 1 for partially relevant (mentioning some aspects of the query but not fully answering it), and 2 for relevant (providing a complete answer to the query). To ensure consistent and high-quality annotations, each document was reviewed by five individuals, with a majority vote determining the final label.

The statistics and examples of the NQ-UTD dataset are presented in Table6 and Table7, respectively. Note that this dataset also offers a fair evaluation of the current capabilities of the advanced LLM-based IR models. As NQ-UTD contains content not previously included in LLM training data, it can serve as a valuable resource for evaluating LLMs’ ability to process and answer queries on recent events.

A.1.2 MS MARCO

MS MARCONguyen etal. (2016), developed by Microsoft Research, is a cornerstone dataset in the fields of natural language processing (NLP) and information retrieval (IR).This dataset comprises over a million user-generated questions derived from Bing search logs, tailored to advance three primary NLP tasks: document ranking, passage retrieval, and question answering.Within our co*cktail benchmark, we specifically utilize the passage retrieval subset, which contains 532K labeled query-passage pairs. A majority of the open-sourced pre-trained transformer checkpoints have been trained using this dataset. Following previous studiesThakur etal. (2021); Lin etal. (2023), we employ the MS MARCO Dev set as the supervised evaluation test set.

A.1.3 TREC

DL19.The DL19 (TREC Deep Learning Track 2019) datasetCraswell etal. (2020), is a key resource for exploring ad hoc ranking across extensive datasets from various domains. Our study focuses on the passage retrieval task utilizing the DL19 dataset, which comprises 43 queries and uses the MS MARCO passage corpus. This allows for its application in supervised evaluation settings.

DL20.The DL20 datasetCraswell etal. (2021) comes from the second year of the TREC Deep Learning Track, and also focuses on ad hoc ranking through human-annotated training sets across diverse domains. We also focus on the passage ranking task featured in the DL20 dataset, which includes 54 queries and also leverages the MS MARCO passage corpus. This also allows for its application in supervised evaluation settings.

A.1.4 BEIR

If not specified, we utilize the preprocessed version of the following datasets as included in the BEIR benchmarkThakur etal. (2021) to serve as the human-written seed datasets in our co*cktail benchmark, ensuring consistency and comparability in our evaluations.

TREC-COVID.The TREC-COVID dataset is developed by the Text REtrieval Conference (TREC)Voorhees etal. (2021), specifically designed to address the challenges of biomedical information retrieval in the context of the COVID-19 pandemic.This dataset represents a concerted effort to harness the rapidly expanding collection of scholarly articles related to COVID-19, providing a focused resource for biomedical IR.

NFCorpus.The NFCorpus datasetBoteva etal. (2016) serves as a detailed resource tailored for ranking tasks within the biomedical field. A distinctive aspect of NFCorpus is its meticulously curated relevance links, connecting queries directly to pertinent research articles.

NQ.The NQ (Natural Questions) dataset aims to propel advancements in natural language understanding, particularly focusing on the question-answering taskKwiatkowski etal. (2019). This dataset is compiled from real questions submitted by users on Google search, each annotated with answers extracted from Wikipedia articles.

HotpotQA.HotpotQAYang etal. (2018) is a large-scale dataset that includes diverse questions posed by human participants. These questions necessitate a multi-step reasoning process, often requiring the extraction of answers from extensive Wikipedia texts and the identification of supporting evidence within them. HotpotQA is specifically crafted to evaluate the effectiveness of models in deciphering complex, multi-hop questions and delivering precise answers.

FiQA-2018.FiQA-2018 (Financial Opinion Mining and Question Answering-2018)Maia etal. (2018) presents a specialized question-answering challenge within the financial domain. The dataset benefits from a comprehensive corpus gathered from esteemed financial news platforms, analyst reports, and influential financial blogs.

Touché-2020.Touché-2020 datasetBondarenko etal. (2020) was created with the specific aim of advancing the field of argument retrieval. Its purpose is to identify argumentative content relevant to contentious queries. It contains argumentative passages from a broad spectrum of domains, which are carefully selected and annotated.

CQADupStack.CQADupStackHoogeveen etal. (2015) is a popular dataset for duplicate question retrieval, which aims to identify duplicate questions in community questionanswering (cQA) forums.This dataset includes carefully annotated and categorized questions and answers across various domains. To maintain focus without losing the essence of generalizability, our study incorporates a subset from the English domains as our selected CQADupStack dataset.

DBPedia.The DBPedia datasetHasibi etal. (2017) includes structured information about entities, concepts, categories, and their relationships, gathered from Wikipedia entries.This dataset focuses on the entity retrieval task, where queries aim to retrieve relevant entities from the English DBpedia corpus dated October 2015.

SCIDOCS.The SCIDOCS datasetCohan etal. (2020) is a vast collection of scholarly articles from the Semantic Scholar database, enhanced with detailed metadata extraction and annotations. It serves as a robust dataset for the citation prediction task, where the objective is to identify papers cited by a given query paper title.

FEVER.The FEVER (Fact Extraction and VERification) datasetThorne etal. (2018) comes from the fact-checking domain, offering a curated collection of human-labeled claims sourced from Wikipedia. Each claim is meticulously classified as Supported, Refuted, or NotEnoughInfo, setting the stage for a nuanced fact-checking task.

Climate-FEVER.Climate-FEVERDiggelmann etal. (2020) is a specialized dataset focused on the area of climate science, comprising real-world claims accompanied by evidence sentences from Wikipedia. Similar to the FEVER dataset, the primary task involves evaluating whether the provided evidence supports, refutes, or lacks sufficient information to adjudicate each claim.

SciFact.SciFactWadden etal. (2020) is designed for the verification of scientific claims sourced from peer-reviewed literature, with each claim meticulously matched with corroborative or refuting evidence from related studies. The dataset comprises expert-crafted scientific claims alongside abstracts containing evidence, each annotated with labels indicating support or contradiction, and rationales explaining the basis of the evidence.

A.2 Dataset Processing and Quality Control

Given the substantial computational cost associated with rewriting documents for the 16 datasets using Large Language Models (LLMs), we adopted sampling methods in line with practices from Zhou etal. (2022); Sun etal. (2023a). Specifically, for datasets with an original corpus size smaller than 200,000 documents, we retained the entire corpus without filtering. For datasets exceeding 100,000 documents, we retain all candidate documents that appear in the labeled data (including train, valid, and test set). If the number of documents post-filtering fell below 100,000, we supplemented the corpus with a random selection of up to 100,000 additional documents from the remaining corpus to ensure a challenging dataset size.

In refining each corpus, we only kept documents with text lengths between 10 and 2,000 words. This criterion was set because texts shorter than 10 words often consisted of empty texts or symbols, compromising data quality—this was particularly evident in the TREC-COVID dataset, which contained over 40,000 documents with empty text.Conversely, texts longer than 2,000 words were filtered out to accommodate the context length limitation of 4,096 tokens for rewriting with Llama2, impacting only 0.05% of the data. These filters did not significantly alter the overall data distribution, with a total of less than 1% data removed, and our evaluation results align closely with those reported in the BEIR benchmarkThakur etal. (2021) for the original datasets, as shown in Table2.

When LLM is used for rewriting, LLM will inevitably refuse to rewrite due to the safety constraints, especially for datasets like Touché-2020, which focus on retrieving contentious viewpoints. For such LLM-generated data, we keep the same content as the human-written counterparts. This approach allows for the easy removal of refused rewrites if necessary, ensuring both the integrity and the quality of the dataset processing and control measures implemented. The statistics of the final 16 benchmarked datasets in co*cktails are summarized in Table1.

A.3 Dataset Statistics and Analysis

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
MS MARCOHuman38.649.949.851.454.255.354.155.260.258.356.8
LLM34.349.550.150.151.853.552.453.056.255.253.7
DL19Human56.277.179.572.176.474.477.976.476.780.278.7
LLM51.279.873.668.273.370.979.576.079.181.080.6
DL20Human54.674.775.373.575.973.278.773.581.879.676.9
LLM45.779.373.876.271.975.677.579.378.775.376.5
TREC-COVIDHuman62.063.061.074.075.064.073.079.075.068.083.0
LLM63.063.060.067.063.060.070.073.065.075.086.0
NFCorpusHuman43.838.432.533.440.942.943.040.342.649.248.9
LLM45.837.536.134.141.042.343.541.843.249.748.4
NQHuman44.861.459.558.964.369.364.566.569.968.569.3
LLM43.560.557.858.164.068.163.866.368.968.068.0
HotpotQAHuman83.974.161.570.484.288.080.888.289.094.291.1
LLM80.773.762.370.883.487.381.187.288.292.689.6
FiQA-2018Human23.323.324.527.828.731.327.330.135.533.040.0
LLM22.123.321.526.225.331.224.729.233.333.237.0
Touché-2020Human52.046.940.849.055.148.038.844.943.951.053.1
LLM57.146.949.041.838.837.827.644.955.158.251.0
CQADupStackHuman28.124.324.329.923.133.632.930.536.033.935.7
LLM22.824.123.526.721.029.828.928.032.129.731.3
DBPediaHuman34.154.349.046.856.661.857.457.860.560.837.0
LLM33.955.349.448.555.862.157.856.559.659.033.0
SCIDOCSHuman16.013.511.314.415.816.714.216.117.218.819.6
LLM15.911.811.813.615.116.014.415.917.218.419.5
FEVERHuman65.279.074.479.582.385.472.287.286.889.825.5
LLM63.079.173.380.082.385.674.586.587.189.224.2
Climate-FEVERHuman25.926.827.229.532.132.828.133.634.336.035.9
LLM24.325.926.228.032.631.327.631.033.234.433.9
SciFactHuman53.035.338.338.752.754.346.353.053.754.364.0
LLM56.335.339.739.350.753.046.751.752.753.063.7
NQ-UTDHuman71.975.663.176.381.377.573.180.076.988.189.4
LLM73.175.068.877.581.376.371.376.978.889.488.1

DatasetMatching PairsRandom Pairs
MS MARCO0.98020.98020.98020.9802 (±0.0138plus-or-minus0.0138\pm 0.0138± 0.0138)0.67250.67250.67250.6725 (±0.0369plus-or-minus0.0369\pm 0.0369± 0.0369)
DL190.97980.97980.97980.9798 (±0.0145plus-or-minus0.0145\pm 0.0145± 0.0145)0.67330.67330.67330.6733 (±0.0373plus-or-minus0.0373\pm 0.0373± 0.0373)
DL200.98170.98170.98170.9817 (±0.0133plus-or-minus0.0133\pm 0.0133± 0.0133)0.67280.67280.67280.6728 (±0.0362plus-or-minus0.0362\pm 0.0362± 0.0362)
TREC-COVID0.98750.98750.98750.9875 (±0.0088plus-or-minus0.0088\pm 0.0088± 0.0088)0.75020.75020.75020.7502 (±0.0436plus-or-minus0.0436\pm 0.0436± 0.0436)
NFCorpus0.98560.98560.98560.9856 (±0.0096plus-or-minus0.0096\pm 0.0096± 0.0096)0.74900.74900.74900.7490 (±0.0373plus-or-minus0.0373\pm 0.0373± 0.0373)
NQ0.98470.98470.98470.9847 (±0.0150plus-or-minus0.0150\pm 0.0150± 0.0150)0.69960.69960.69960.6996 (±0.0348plus-or-minus0.0348\pm 0.0348± 0.0348)
HotpotQA0.99050.99050.99050.9905 (±0.0123plus-or-minus0.0123\pm 0.0123± 0.0123)0.70730.70730.70730.7073 (±0.0366plus-or-minus0.0366\pm 0.0366± 0.0366)
FiQA-20180.96570.96570.96570.9657 (±0.0243plus-or-minus0.0243\pm 0.0243± 0.0243)0.72160.72160.72160.7216 (±0.0372plus-or-minus0.0372\pm 0.0372± 0.0372)
Touché-20200.96930.96930.96930.9693 (±0.0262plus-or-minus0.0262\pm 0.0262± 0.0262)0.72740.72740.72740.7274 (±0.0359plus-or-minus0.0359\pm 0.0359± 0.0359)
CQADupStack0.95830.95830.95830.9583 (±0.0336plus-or-minus0.0336\pm 0.0336± 0.0336)0.73600.73600.73600.7360 (±0.0311plus-or-minus0.0311\pm 0.0311± 0.0311)
DBPedia0.99000.99000.99000.9900 (±0.0121plus-or-minus0.0121\pm 0.0121± 0.0121)0.70230.70230.70230.7023 (±0.0373plus-or-minus0.0373\pm 0.0373± 0.0373)
SCIDOCS0.98630.98630.98630.9863 (±0.0112plus-or-minus0.0112\pm 0.0112± 0.0112)0.74440.74440.74440.7444 (±0.0352plus-or-minus0.0352\pm 0.0352± 0.0352)
FEVER0.98660.98660.98660.9866 (±0.0110plus-or-minus0.0110\pm 0.0110± 0.0110)0.69830.69830.69830.6983 (±0.0363plus-or-minus0.0363\pm 0.0363± 0.0363)
Climate-FEVER0.98730.98730.98730.9873 (±0.0107plus-or-minus0.0107\pm 0.0107± 0.0107)0.69920.69920.69920.6992 (±0.0366plus-or-minus0.0366\pm 0.0366± 0.0366)
SciFact0.98810.98810.98810.9881 (±0.0084plus-or-minus0.0084\pm 0.0084± 0.0084)0.74840.74840.74840.7484 (±0.0445plus-or-minus0.0445\pm 0.0445± 0.0445)
NQ-UTD0.98330.98330.98330.9833 (±0.0122plus-or-minus0.0122\pm 0.0122± 0.0122)0.71380.71380.71380.7138 (±0.0443plus-or-minus0.0443\pm 0.0443± 0.0443)
Avg.0.98160.98160.98160.9816 (±0.0150plus-or-minus0.0150\pm 0.0150± 0.0150)0.71350.71350.71350.7135 (0.03760.03760.03760.0376)

DatasetHuman DocLLM DocEqual
MS MARCO0.0% (0.0%)10.0% (0.0%)90% (77.8%)
DL190.0% (0.0%)5.0% (0.0%)95.0% (94.7%)
DL205.0% (0.0%)0.0% (0.0%)95.0% (78.9%)
TREC-COVID5.0% (0.0%)5.0% (0.0%)90% (61.1%)
NFCorpus0.0% (0.0%)0.0% (0.0%)100.0% (75.0%)
NQ0.0% (0.0%)0.0% (0.0%)100.0% (100.0%)
HotpotQA0.0% (0.0%)0.0% (0.0%)100.0% (80.0%)
FiQA-20185.0% (0.0%)0.0% (0.0%)95.0% (68.4%)
Touché-20200.0% (0.0%)0.0% (0.0%)100.0% (70.0%)
CQADupStack5.0% (0.0%)5.0% (0.0%)90% (83.3%)
DBPedia0.0% (0.0%)0.0% (0.0%)100.0% (85.0%)
SCIDOCS0.0% (0.0%)0.0% (0.0%)100.0% (100.0%)
FEVER5.0% (0.0%)0.0% (0.0%)95.0% (84.2%)
Climate-FEVER0.0% (0.0%)0.0% (0.0%)100.0% (90.0%)
SciFact10.0% (0.0%)5.0% (0.0%)85.0% (64.7%)
NQ-UTD0.0% (0.0%)0.0% (0.0%)100.0% (90.0%)
Avg.2.2%1.9%95.9%

Term-based Statistics.Figure5 illustrates the distribution of text lengths, revealing minimal variation between the lengths of texts. We then calculated the Jaccard similarity and overlap between each LLM-generated document dGsuperscript𝑑𝐺d^{G}italic_d start_POSTSUPERSCRIPT italic_G end_POSTSUPERSCRIPT and the corresponding human-written document dHsuperscript𝑑𝐻d^{H}italic_d start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT, using the following two formulas:

Jaccard similarity=|dGdH||dGdH|,absentsuperscript𝑑𝐺superscript𝑑𝐻superscript𝑑𝐺superscript𝑑𝐻\displaystyle=\frac{|d^{G}\bigcap d^{H}|}{|d^{G}\bigcup d^{H}|},= divide start_ARG | italic_d start_POSTSUPERSCRIPT italic_G end_POSTSUPERSCRIPT ⋂ italic_d start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT | end_ARG start_ARG | italic_d start_POSTSUPERSCRIPT italic_G end_POSTSUPERSCRIPT ⋃ italic_d start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT | end_ARG ,
Overlap=|dGdH||dH|.absentsuperscript𝑑𝐺superscript𝑑𝐻superscript𝑑𝐻\displaystyle=\frac{|d^{G}\bigcap d^{H}|}{|d^{H}|}.= divide start_ARG | italic_d start_POSTSUPERSCRIPT italic_G end_POSTSUPERSCRIPT ⋂ italic_d start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT | end_ARG start_ARG | italic_d start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT | end_ARG .

The results presented in Figure6 highlight significant differences in terms despite their ostensibly similar semantic information.

Semantic Analysis.Cosine similarity between LLM-generated and human-written documents, calculated using the OpenAI embedding model444text-embedding-ada-002, is displayed in Figure7.The results, predominantly above 0.90.90.90.9, signify a high level of semantic preservation in the LLM-generated texts compared to the original human-written content.We also compare the semantics between randomly selected <human doc, LLM doc> pairs and matching pairs <human doc, LLM doc>, as shown in Table9. The average similarity score for matching pairs was 0.98160.98160.98160.9816, significantly higher than the 0.71350.71350.71350.7135 average for random pairs. This significant statistical difference supports the close semantic alignment between LLM-generated and human-authored documents, further validating the quality of our constructed dataset.Further, evaluations of various retrieval models on sole human-written or LLM-generated corpora, as detailed in Table8, demonstrate consistent performance across all datasets between two types of corpus. These observations across all datasets reinforce confidence in the quality of our newly constructed datasets, suggesting that the LLM-generated content maintains semantics comparable to human-written texts for IR tasks.

A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (6)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (7)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (8)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (9)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (10)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (11)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (12)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (13)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (14)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (15)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (16)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (17)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (18)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (19)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (20)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (21)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (22)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (23)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (24)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (25)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (26)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (27)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (28)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (29)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (30)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (31)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (32)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (33)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (34)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (35)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (36)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (37)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (38)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (39)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (40)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (41)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (42)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (43)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (44)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (45)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (46)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (47)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (48)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (49)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (50)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (51)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (52)
A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (53)

Model NamePublicly Available Link
BERT (mini)https://huggingface.co/prajjwal1/bert-mini
BERT (small)https://huggingface.co/prajjwal1/bert-small
BERT (base)https://huggingface.co/bert-base-uncased
BERT (large)https://huggingface.co/bert-large-uncased
RoBERTahttps://huggingface.co/FacebookAI/roberta-base
ANCEhttps://huggingface.co/sentence-transformers/msmarco-roberta-base-ance-firstp
TAS-Bhttps://huggingface.co/sentence-transformers/msmarco-distilbert-base-tas-b
Contrieverhttps://huggingface.co/nthakur/contriever-base-msmarco
coCondenserhttps://huggingface.co/sentence-transformers/msmarco-bert-co-condensor
RetroMAEhttps://huggingface.co/nthakur/RetroMAE_BEIR
DRAGON (Query)https://huggingface.co/nthakur/dragon-roberta-query-encoder
DRAGON (Doc)https://huggingface.co/nthakur/dragon-roberta-context-encoder
CEhttps://huggingface.co/cross-encoder/ms-marco-MiniLM-L-6-v2
monoT5https://huggingface.co/castorini/monot5-base-msmarco-10k

Quality Verification with Human Evaluation.To further validate the assigned relevancy label, we also conduct a human evaluation study. Due to cost constraints of human evaluation, we sample 20 <query, human doc, LLM doc> triples from each of the 16 datasets included in co*cktail. These triples were annotated by graduate students and our colleagues, who were asked to assess which document is more semantically relevant to the given query, with options being “human doc”, “LLM doc”, or “equal”. At least three different annotators evaluated each triple, with the majority vote deciding the final label. We summarized the human evaluation results in Table10, with numbers in parentheses indicating the percentage of agreement among all three evaluators for each option.The results show a comparable level of semantic relevance between human-written and LLM-generated texts for the given queries, ensuring the fairness of our analysis of source bias.

A.4 Dataset Examples

DatasetQueryRelevant Human-Written DocumentRelevant LLM-Generated Document
MS MARCOdo physicians pay for insurance from their salaries?Active-duty physicians and their dependents will receive free or discounted health care coverage and dental coverage. Physicians in the Reserve and Guard can also participate in TRICARE Reserve Select, which is part of the Military’s health care plan. Finally, physicians can receive up to $400,000 in term life insurance coverage for only $29 a month.Active-duty physicians and their dependents are eligible for free or discounted healthcare coverage and dental coverage. In addition, physicians in the Reserve and Guard can enroll in TRICARE Reserve Select, which is part of the Military’s healthcare plan. Furthermore, physicians can receive up to $400,000 in term life insurance coverage for just $29 per month.
DL19axon terminals or synaptic knob definitionApril 29, 2013. the terminal part of an axon from which a neural signal is rendered, via dispersion of a neurotransmitter, across a synapse to a nearby neuron. TERMINAL BUTTON: The terminal button is commonly referred to as the synaptic button, end button, button terminal, terminal bulb, and synaptic knob..April 29, 2013. The terminal end of an axon, where a neural signal is transmitted to a nearby neuron through the diffusion of a neurotransmitter across a synapse. TERMINAL BUTTON: Also known as the synaptic button, end button, button terminal, terminal bulb, or synaptic knob.
DL20average salary for dental hygienist in nebraskaIn Nebraska, as the number of dental hygienists is growing, the salaries earned by dental hygienists are increasing. Dental hygienists earned a yearly mean salary of $58,410 in 2006. They earned a yearly mean salary of $64,440 in 2010.In Nebraska, as the number of dental hygienists is increasing, so are their salaries. According to data from 2006, dental hygienists earned a yearly mean salary of $58,410. By 2010, this figure had risen to $64,440.
TREC-COVIDwill SARS-CoV2 infected people develop immunity? Is cross protection possible?Summary COVID-19 had a mild clinical course in patients with Agammaglobulinemia lacking B lymphocytes, whereas it developed aggressively in Common Variable Immune Deficiency. Our data offer mechanisms for possible therapeutic targets.COVID-19 had a mild clinical course in patients with Agammaglobulinemia, a condition characterized by a lack of B lymphocytes, while it developed aggressively in Common Variable Immune Deficiency. Our data provide insights into potential therapeutic targets.
NFCorpusWhat is Actually in Chicken Nuggets?To study the origin and spread of Yersinia enterocolitica among pigs, fecal and blood samples were repeatedly taken on a fattening farm. A few piglets were found to be already infected on breeding farms. After the piglets were mixed, the infection spread through the whole unit. Eventually, all the pigs excreted the pathogen.To investigate the origins and spread of Yersinia enterocolitica among pigs, fecal and blood samples were repeatedly collected from a fattening farm. Initially, a few piglets on breeding farms were found to be infected. Once these piglets were mixed, the infection rapidly spread throughout the entire unit, eventually infecting all the pigs.
NQhow many episodes are in chicago fire season 4The fourth season of Chicago Fire, an American drama television series with executive producer Dick Wolf, and producers Derek Haas, Michael Brandt, and Matt Olmstead, was ordered on February 5, 2015, by NBC,[1] and premiered on October 13, 2015 and concluded on May 17, 2016.[2] The season contained 23 episodes.[3]The fourth season of Chicago Fire, a US drama television series created by executive producer Dick Wolf and producers Derek Haas, Michael Brandt, and Matt Olmstead, was greenlit by NBC on February 5, 2015, and premiered on October 13, 2015, concluding on May 17, 2016. The season consisted of 23 episodes.
HotpotQAWhat government position was held by the woman who portrayed Corliss Archer in the film Kiss and Tell?Kiss and Tell is a 1945 American comedy film starring then 17-year-old Shirley Temple as Corliss Archer. In the film, two teenage girls cause their respective parents much concern when they start to become interested in boys. The parents’ bickering about which girl is the worse influence causes more problems than it solves.Kiss and Tell is a 1945 American comedy film starring then 17-year-old Shirley Temple as Corliss Archer. In the film, two teenage girls, Corliss and her friend, cause their respective parents much distress when they start to develop romantic interests in boys. The parents’ arguing over which girl is the worse influence creates more problems than it solves.
FiQA-2018What are the ins/outs of writing equipment purchases off as business expenses in a home based business?Keep this rather corny acronym in mind. Business expenses must be CORN: As other posters have already pointed out, certain expenses that are capital items (computers, furniture, etc.) must be depreciated over several years, but you have a certain amount of capital items that you can write off in the current tax year.Keep this straightforward acronym in mind. Business expenses must be CLEAR: As other posters have already noted, certain expenses that are capital items (computers, furniture, etc.) must be depreciated over several years, but you have a certain amount of capital items that you can write off in the current tax year.
Touché-2020Should teachers get tenure?Why should teachers be laid off because of seniority? Many electives teachers got laid off. Even a Music Teacher got laid off even though she won a award for the best teacher in the valley! Ironic isn’t it?Teachers should not be laid off solely based on seniority. It is unfair that many elective teachers were let go, including a talented Music Teacher who recently won an award for being the best teacher in the valley. It is ironic that she was laid off despite her achievement.
CQADupStackIs "a wide range of features" singular or plural?I hope you can enlighten me. I get varying answers in Google and I need to find out which is the correct grammatical structure for these sentences. > The rest of the staff is/are on leave at the moment. > > The rest of my family is/are arriving late.I hope you can enlighten me. I’ve been getting conflicting answers in Google, and I need to determine the correct grammatical structure for these sentences.The rest of the staff are on leave at the moment.The rest of my family are arriving late.
DBPediasocial network group selectionCluster analysis or clustering is the task of grouping a set of objects in such a way that objects in the same group (called a cluster) are more similar (in some sense or another) to each other than to those in other groups (clusters).Clustering is the process of grouping a set of objects together based on their similarities, where objects within the same group (cluster) are more alike than those in other groups.
SCIDOCSDeltaCFS: Boosting Delta Sync for Cloud Storage Services by Learning from NFSThe state machine approach is a general method for implementing fault-tolerant services in distributed systems. This paper reviews the approach and describes protocols for two different failure models—Byzantine and fail stop. Systems reconfiguration techniques for removing faulty components and integrating repaired components are also discussed.The state machine approach is a versatile method for building fault-tolerant services in distributed systems. This paper examines the approach and outlines protocols for two distinct failure models—Byzantine and fail-stop. Additionally, the paper discusses techniques for reconfiguring systems to remove faulty components and integrate repaired ones.
FEVERKonidela Production Company was established.Konidela Production Company is an Indian film production company established by actor Ram Charan , son of Chiranjeevi .Konidela Production Company is a renowned Indian film production house founded by the talented actor Ram Charan, the son of the legendary actor Chiranjeevi.
Climate-FEVEREach of the six major past ice ages began when the atmospheric carbon dioxide content was far higher than at present.The Geologic temperature record are changes in Earth ’s environment as determined from geologic evidence on multi-million to billion ( 109 ) year time scales . The study of past temperatures provides an important paleoenvironmental insight because it is a crucial component of the climate and oceanography of the time .The geologic temperature record reflects changes in Earth’s environment as revealed through geologic evidence spanning millions to billions of years (109 years). The study of past temperatures offers valuable paleoenvironmental insights since it is a fundamental component of Earth’s climate and oceanography during those time periods.
SciFactCataract and trachoma are the primary cause of blindness in Southern Sudan.Background Blindness and low vision are thought to be common in southern Sudan. However, the magnitude and geographical distribution are largely unknown. We aimed to estimate the prevalence of blindness and low vision, identify the main causes of blindness and low vision, and estimate targets for blindness prevention programs in Mankien payam (district), southern Sudan.Background blindness and low vision are believed to be prevalent in southern Sudan, but the scope and geographical distribution of these issues remain largely unexplored. Our study aimed to determine the prevalence of blindness and low vision, identify the primary causes of blindness and low vision, and estimate targets for blindness prevention programs in Mankien payam (district), southern Sudan.
NQ-UTDWho won women’s doubles champion in the 2023 Badminton World Tour Finals?The women’s doubles final of the BWF World Tour Finals 2023 concluded at the Hangzhou Olympic Sports Center on Sunday. China’s top doubles pair, Chen Qingchen and Jia Yifan, defeated South Korea’s Baek Ha Na and Lee So Hee 2-0 to successfully defend their title and win $210,000 in prize money.The women’s doubles final of the BWF World Tour Finals 2023 came to a close at the Hangzhou Olympic Sports Center on Sunday. China’s premier doubles pair, Chen Qingchen and Jia Yifan, outplayed South Korea’s Baek Ha Na and Lee So Hee, winning the match 2-0 and successfully defending their title to collect a prize money of $210,000.

For a better understanding of the datasets used in our co*cktail benchmark, we offer examples for each dataset in Table12, showcasing a given query along with the corresponding relevant human-written document and LLM-generated document.

Appendix B Model Details

B.1 Detailed Description of Models

We select the widely used approach BM25Robertson etal. (2009) as the representation of lexical retrieval models in our benchmark.

For neural retrieval models that leverage pre-trained language models to acquire semantic relationships between queries and documents, we select the following representative and state-of-the-art models:

BERTDevlin etal. (2019), a foundational pre-trained language model, is commonly used as an encoder in dense retrieval systems. It was fine-tuned on the MSMARCO dataset using the official BEIR benchmark training script.

RoBERTaLiu etal. (2019) builds on BERT’s success with more data and refined training techniques to enhance performance. RoBERTa was fine-tuned in the same manner to BERT, using the MSMARCO dataset.

ANCEXiong etal. (2020) improves dense retrieval by selecting challenging negatives across the corpus and updating the Approximate Nearest Neighbor (ANN) index asynchronously with each training iteration.

TAS-BHofstätter etal. (2021), a bi-encoder model, leverages balanced margin sampling for efficient query selection and dual supervision from a cross-encoder and a ColBERT model for enhanced learning.

ContrieverIzacard etal. (2022) employs contrastive learning with positive samples generated through cropping and token sampling, using MoCo to incorporate negatives from a queue of previous batches for self-supervised training.

coCondenserGao and Callan (2022) undertakes a two-stage process beginning with pretraining and unsupervised contrastive loss for embedding generation, followed by supervised training with the pre-trained encoder.

RetroMAEXiao etal. (2022) utilizes a Masked Auto-Encoder approach for pretraining, enhancing sentence reconstruction from masked inputs to refine language modeling.

DRAGONLin etal. (2023) applies progressive supervision and query augmentation for effective dense retrieval, suitable for both supervised and zero-shot settings.

Additionally, the following two widely used neural re-ranking models are also adopted in our evaluation framework:

CEWang etal. (2020) is a cross-encoder model pre-trained on MS MARCO dataset Nguyen etal. (2016) through knowledge distillation from three teacher models: BERT-base Devlin etal. (2019), BERT-large Devlin etal. (2019), and ALBERT-large Lan etal. (2019).

monoT5Raffel etal. (2020) is a sequence-to-sequence re-ranker based on T5Raffel etal. (2020), which is also pre-trained on the MS MARCO dataset.

B.2 Checkpoints and Implementation Details

For the fine-tuning of BERT and RoBERTa models, we utilized the official training script555https://github.com/beir-cellar/beir/blob/main/examples/retrieval/training/train_msmarco_v3.py provided in the BEIR benchmark on the MSMARCO training dataset. Unless otherwise noted, mean pooling was employed as the pooling strategy, alongside cosine similarity as the scoring function. Training parameters for each model included a duration of 10101010 epochs, utilizing a batch size of 75757575, a learning rate set at 2e-52𝑒-52e\text{-}52 italic_e - 5, and the AdamW optimizer for efficient optimization. Our software framework utilizes PyTorch version 2.0.0 and the HuggingFace Transformers library version 4.31.0. All experiments were performed using approximately 100 hours on four NVIDIA RTX A6000 (48G) GPUs.

For the other benchmarked neural retrieval models, we use the publicly available official pre-trained checkpoints, which are listed in Table11. For all the neural models, only the first 512 word tokens of all documents are inputted.

Appendix C More Experimental Results

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO39.753.553.754.657.158.357.458.161.959.058.055.657.2
DL1951.775.775.569.475.072.675.575.576.678.875.872.975.0
DL2050.775.575.572.272.971.276.577.478.076.975.472.975.2
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID62.466.460.071.064.063.270.974.368.371.982.168.669.2
NFCorpus44.236.332.333.039.142.141.139.341.148.248.040.440.1
NQ46.262.560.060.366.370.666.067.971.268.669.164.466.3
HotpotQA72.362.351.759.272.376.369.476.577.083.581.471.171.0
FiQA-201822.221.922.226.226.129.925.328.633.531.634.927.528.0
Touché-202051.447.042.146.844.746.334.347.153.452.852.547.146.7
CQADupStack26.424.825.029.523.332.631.729.534.832.634.429.529.8
DBPedia34.252.647.345.854.559.354.855.157.757.433.250.251.8
SCIDOCS15.411.710.512.715.015.513.914.816.117.418.314.714.6
FEVER67.580.575.880.783.886.575.587.787.989.144.678.179.2
Climate-FEVER23.425.125.225.930.329.625.429.931.032.733.028.328.8
SciFact58.337.940.940.854.056.449.954.556.757.165.452.051.4
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD70.572.466.774.378.775.469.375.576.686.686.275.776.2
Averaged Result
Supervised47.468.268.265.468.367.469.870.372.271.669.767.169.1
Zero-shot45.746.343.146.650.252.648.352.454.356.152.549.850.2
All46.050.447.850.253.655.452.355.757.659.055.853.153.8

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO35.9-4.2-8.30.5-8.5-1.71.9-5.3-3.2-4.00.90.4-3.2
DL1981.6-50.6-17.2-21.0-49.3-21.9-21.1-47.6-45.4-19.5-25.4-21.6-31.9
DL2091.3-59.3-37.8-7.7-18.7-10.6-17.0-31.1-36.2-21.3-13.9-14.8-25.4
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID25.6-51.0-40.4-33.8-73.1-62.0-68.0-25.9-17.9-66.5-46.0-41.7-48.5
NFCorpus-16.2-16.8-29.7-19.1-22.7-43.2-24.9-18.9-22.9-38.4-18.5-24.7-25.5
NQ-3.9-8.0-3.8-4.3-12.4-10.2-8.5-7.9-14.7-16.6-8.7-9.0-9.5
HotpotQA21.0-0.5-1.8-5.70.2-1.6-2.91.4-1.814.14.92.50.6
FiQA-2018-3.6-14.0-5.8-20.3-10.7-28.9-19.0-20.0-15.1-25.4-17.2-16.4-17.6
Touché-2020-32.1-41.4-40.2-12.3-9.7-46.76.4-18.2-53.0-69.9-53.8-33.7-33.9
CQADupStack17.8-20.4-15.7-0.9-1.7-4.11.3-20.7-4.7-7.53.5-4.8-7.1
DBPedia21.3-9.9-14.1-17.7-5.1-5.3-10.8-13.1-10.9-2.610.9-5.2-7.9
SCIDOCS1.3-1.7-21.0-1.6-1.3-6.5-5.7-10.8-18.6-12.7-23.0-9.2-10.3
FEVER-5.20.2-0.2-25.8-4.5-6.3-8.00.6-7.0-0.76.2-4.6-4.5
Climate-FEVER10.3-7.4-8.1-38.4-10.2-6.9-9.3-4.6-5.03.2-37.9-10.4-12.5
SciFact0.6-12.5-2.5-5.9-16.42.2-10.3-6.2-7.51.3-8.4-6.0-6.6
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD9.0-13.3-14.2-13.9-9.7-17.5-10.7-18.7-20.6-21.8-9.0-12.8-14.9
Averaged Result
Supervised69.6-38.0-21.1-9.4-25.5-11.4-12.1-28.0-28.3-14.9-12.8-12.0-20.2
Zero-shot3.5-15.1-15.2-15.4-13.6-18.2-13.1-12.5-15.4-18.7-15.2-13.5-15.2
All15.9-19.4-16.3-14.2-15.9-16.9-12.9-15.4-17.8-18.0-14.7-13.2-16.2

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO43.558.058.159.061.763.062.062.666.662.761.759.961.5
DL1949.975.373.469.574.371.875.073.976.776.774.671.974.1
DL2048.374.973.971.673.770.775.076.077.974.672.071.774.0
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID61.363.257.768.164.462.070.774.468.470.980.167.468.0
NFCorpus41.534.330.030.137.539.738.237.339.745.445.638.137.8
NQ49.364.862.362.868.973.568.770.473.770.270.966.968.6
HotpotQA67.856.646.853.767.471.464.071.672.079.077.866.266.0
FiQA-201821.621.122.125.525.729.224.427.832.330.533.726.727.2
Touché-202050.445.640.948.742.943.732.644.450.150.252.045.645.1
CQADupStack27.325.425.730.124.533.532.630.235.733.134.830.330.6
DBPedia33.349.244.443.551.855.552.652.654.055.133.347.849.2
SCIDOCS14.110.79.511.413.314.212.413.614.815.816.513.313.2
FEVER71.482.578.182.485.888.178.589.289.589.948.180.381.2
Climate-FEVER22.123.823.924.028.628.023.727.929.031.331.526.727.2
SciFact62.041.043.042.956.459.653.057.159.961.268.254.954.2
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD68.771.363.972.476.574.269.474.374.683.984.374.074.5
Averaged Result
Supervised47.269.468.566.769.968.570.770.873.771.369.467.869.9
Zero-shot45.445.342.245.849.551.747.851.653.455.152.149.149.4
All45.849.947.149.753.354.952.155.257.258.255.352.653.3

Model (\rightarrow)LexicalNeural RetrieversNeural Re-rankers
BM25BERTRoBERTaANCETAS-BContrievercoCondenserRetroMAEDRAGONCEmonoT5
PLM-BERTRoBERTaRoBERTaDistilBERTBERTBERTBERTBERTMiniLMT5Average
# Paras-110M125M125M66M110M110M110M110M66M220MAllNeural
Supervised Evaluation (In-Domain Datasets Collected in Pre-LLM Era)
MS MARCO30.5-4.0-6.61.0-7.8-1.02.2-4.7-2.6-2.52.00.6-2.4
DL1959.5-37.9-27.7-17.0-39.3-20.7-16.5-38.6-48.1-21.7-23.0-21.0-29.0
DL2074.1-31.8-34.1-5.6-13.7-10.6-14.0-16.8-27.8-28.9-14.3-11.2-19.8
Zero-shot Evaluation (Out-of-Domain Datasets Collected in Pre-LLM Era)
TREC-COVID16.6-42.4-37.1-33.4-62.7-58.7-56.9-28.5-23.4-66.3-32.7-38.7-44.2
NFCorpus-15.8-15.1-19.4-10.4-13.6-34.9-13.8-17.6-16.4-26.1-12.3-17.8-18.0
NQ-2.3-6.7-3.2-2.2-9.7-8.8-7.0-5.8-11.3-14.8-7.5-7.2-7.7
HotpotQA16.5-0.4-1.4-4.50.8-0.7-1.41.8-1.110.24.62.20.8
FiQA-2018-5.2-14.1-5.1-12.2-9.2-18.8-14.5-18.2-12.2-17.8-14.1-12.9-13.6
Touché-2020-32.2-35.3-42.9-19.8-32.0-34.1-8.5-19.1-45.2-55.1-55.0-34.5-34.7
CQADupStack15.5-15.8-11.0-0.40.5-3.83.6-17.9-2.5-4.73.3-3.0-4.9
DBPedia18.3-11.8-13.1-17.7-4.4-2.2-11.1-9.0-9.4-3.913.9-4.6-6.9
SCIDOCS1.4-1.8-16.3-3.4-4.4-8.2-7.9-10.1-13.2-12.3-17.8-8.5-9.5
FEVER-4.20.30.2-22.3-3.9-5.0-6.10.4-6.1-0.84.9-3.9-3.8
Climate-FEVER7.2-4.8-5.3-37.2-10.1-4.6-7.2-4.6-3.52.3-28.3-8.7-10.3
SciFact0.4-10.9-3.2-4.0-11.61.2-8.7-5.2-7.0-0.6-6.4-5.1-5.6
Zero-shot Evaluation (Out-of-Domain Datasets Collected in the LLM Era)
NQ-UTD4.4-12.2-5.0-7.7-8.1-9.0-5.7-16.4-13.0-16.3-4.1-8.5-9.7
Averaged Result
Supervised54.7-24.6-22.8-7.2-20.3-10.8-9.4-20.0-26.2-17.7-11.8-10.5-17.1
Zero-shot1.6-13.2-12.5-13.5-13.0-14.4-11.2-11.6-12.6-15.9-11.7-11.6-12.9
All11.5-15.3-14.4-12.3-14.3-13.7-10.8-13.1-15.2-16.2-11.7-11.4-13.7

In addition to the findings presented through NDCG@1 in Table2 and Table3, we further extend our analysis to include the results for ranking performance and source bias for NDCG@3 and NDCG@5. These results are detailed in Table13 and Table14 for NDCG@3, and Table15 and Table16 for NDCG@5, respectively. Consistent with our earlier observations in Section5.2, results on these metrics also reveal a similar trend, underscoring the robustness of our findings.

A Comprehensive Information Retrieval Benchmark with LLM-Generated Documents Integration (2024)
Top Articles
Latest Posts
Article information

Author: The Hon. Margery Christiansen

Last Updated:

Views: 6125

Rating: 5 / 5 (50 voted)

Reviews: 81% of readers found this page helpful

Author information

Name: The Hon. Margery Christiansen

Birthday: 2000-07-07

Address: 5050 Breitenberg Knoll, New Robert, MI 45409

Phone: +2556892639372

Job: Investor Mining Engineer

Hobby: Sketching, Cosplaying, Glassblowing, Genealogy, Crocheting, Archery, Skateboarding

Introduction: My name is The Hon. Margery Christiansen, I am a bright, adorable, precious, inexpensive, gorgeous, comfortable, happy person who loves writing and wants to share my knowledge and understanding with you.