A bias–variance evaluation framework for information retrieval systems

Peng Zhang*, Hui Gao, Zeting Hu, Meng Yang, Dawei Song, Jun Wang, Yuexian Hou, Bin Hu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)

Abstract

In information retrieval (IR), the improvement of the effectiveness often sacrifices the stability of an IR system. To evaluate the stability, many risk-sensitive metrics have been proposed. Since the theoretical limitations, the current works study the effectiveness and stability separately, and have not explored the effectiveness–stability tradeoff. In this paper, we propose a Bias–Variance Tradeoff Evaluation (BV-Test) framework, based on the bias–variance decomposition of the mean squared error, to measure the overall performance (considering both effectiveness and stability) and the tradeoff between effectiveness and stability of a system. In this framework, we define generalized bias–variance metrics, based on the Cranfield-style experiment set-up where the document collection is fixed (across topics) or the set-up where document collection is a sample (per-topic). Compared with risk-sensitive evaluation methods, our work not only measures the effectiveness–stability tradeoff of a system, but also effectively tracks the source of system instability. Experiments on TREC Ad-hoc track (1993–1999) and Web track (2010–2014) show a clear effectiveness–stability tradeoff across topics and per-topic, and topic grouping and max–min normalization can effectively reduce the bias–variance tradeoff. Experimental results on TREC Session track (2010–2012) also show that the query reformulation and increase of user data are beneficial to both effectiveness and stability simultaneously.

Original languageEnglish
Article number102747
JournalInformation Processing and Management
Volume59
Issue number1
DOIs
Publication statusPublished - Jan 2022

Keywords

  • Effectiveness–stability tradeoff
  • Evaluation metrics
  • Information retrieval

Fingerprint

Dive into the research topics of 'A bias–variance evaluation framework for information retrieval systems'. Together they form a unique fingerprint.

Cite this