Question Answering (QA) on narrative text poses a unique challenge to current systems, requiring a deep understanding of long, complex documents. However, the reliability of NarrativeQA, the most widely used benchmark in this domain, is hindered by noisy documents and flawed QA pairs. In this work, we introduce LiteraryQA, a high-quality subset of NarrativeQA focused on literary works. Using a human- and LLM-validated pipeline, we identify and correct low-quality QA samples while removing extraneous text from source documents. We then carry out a meta-evaluation of automatic metrics to clarify how systems should be evaluated on LiteraryQA.This analysis reveals that all n-gram-based metrics have a low system-level correlation to human judgment, while LLM-as-a-Judge evaluations, even with small open-weight models, can strongly agree with the ranking identified by humans.Finally, we benchmark a set of long-context LLMs on LiteraryQA. We release our code and data at https://github.com/sapienzaNLP/LiteraryQA.

LiteraryQA: Towards Effective Evaluation of Long-document Narrative QA / Bonomo, Tommaso; Gioffre, Luca; Navigli, Roberto. - 1:(2025), pp. 34074-34095. ( Empirical Methods in Natural Language Processing Suzhou; China ) [10.18653/v1/2025.emnlp-main.1729].

LiteraryQA: Towards Effective Evaluation of Long-document Narrative QA

Tommaso Bonomo
;
Luca Gioffre;Roberto Navigli
2025

Abstract

Question Answering (QA) on narrative text poses a unique challenge to current systems, requiring a deep understanding of long, complex documents. However, the reliability of NarrativeQA, the most widely used benchmark in this domain, is hindered by noisy documents and flawed QA pairs. In this work, we introduce LiteraryQA, a high-quality subset of NarrativeQA focused on literary works. Using a human- and LLM-validated pipeline, we identify and correct low-quality QA samples while removing extraneous text from source documents. We then carry out a meta-evaluation of automatic metrics to clarify how systems should be evaluated on LiteraryQA.This analysis reveals that all n-gram-based metrics have a low system-level correlation to human judgment, while LLM-as-a-Judge evaluations, even with small open-weight models, can strongly agree with the ranking identified by humans.Finally, we benchmark a set of long-context LLMs on LiteraryQA. We release our code and data at https://github.com/sapienzaNLP/LiteraryQA.
2025
Empirical Methods in Natural Language Processing
question answering; benchmark; narrative; evaluation methodologies
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
LiteraryQA: Towards Effective Evaluation of Long-document Narrative QA / Bonomo, Tommaso; Gioffre, Luca; Navigli, Roberto. - 1:(2025), pp. 34074-34095. ( Empirical Methods in Natural Language Processing Suzhou; China ) [10.18653/v1/2025.emnlp-main.1729].
File allegati a questo prodotto
File Dimensione Formato  
Bonomo_LiteraryQA_2025.pdf

accesso aperto

Note: URL: https://aclanthology.org/2025.emnlp-main.1729/
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 540.51 kB
Formato Adobe PDF
540.51 kB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1755709
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact