Research Library

open-access-imgOpen AccessAdapting Standard Retrieval Benchmarks to Evaluate Generated Answers
Author(s)
Negar Arabzadeh,
Amin Bigdeli,
Charles L. A. Clarke
Publication year2024
Large language models can now directly generate answers to many factualquestions without referencing external sources. Unfortunately, relativelylittle attention has been paid to methods for evaluating the quality andcorrectness of these answers, for comparing the performance of one model toanother, or for comparing one prompt to another. In addition, the quality ofgenerated answers are rarely directly compared to the quality of retrievedanswers. As models evolve and prompts are modified, we have no systematic wayto measure improvements without resorting to expensive human judgments. Toaddress this problem we adapt standard retrieval benchmarks to evaluate answersgenerated by large language models. Inspired by the BERTScore metric forsummarization, we explore two approaches. In the first, we base our evaluationon the benchmark relevance judgments. We empirically run experiments on howinformation retrieval relevance judgments can be utilized as an anchor toevaluating the generated answers. In the second, we compare generated answersto the top results retrieved by a diverse set of retrieval models, ranging fromtraditional approaches to advanced methods, allowing us to measure improvementswithout human judgments. In both cases, we measure the similarity between anembedded representation of the generated answer and an embedded representationof a known, or assumed, relevant passage from the retrieval benchmark.
Language(s)English

Seeing content that should not be on Zendy? Contact us.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here