A Model-Based Evaluation Metric for Question Answering Systems

Loading...
Publication Logo

Date

2025

Journal Title

Journal ISSN

Volume Title

Publisher

World Scientific

Open Access Color

Green Open Access

No

OpenAIRE Downloads

OpenAIRE Views

Publicly Funded

No
Impulse
Average
Influence
Average
Popularity
Average

Research Projects

Journal Issue

Abstract

The paper addresses the limitations of traditional evaluation metrics for Question Answering (QA) systems that primarily focus on syntax and n-gram similarity. We propose a novel model-based evaluation metric, MQA-metric, and create a human-judgment-based dataset, squad-qametric and marco-qametric, to validate our approach. The research aims to solve several key problems: the objectivity in dataset labeling, the effectiveness of metrics when there is no syntax similarity, the impact of answer length on metric performance, and the influence of real answer quality on metric results. To tackle these challenges, we designed an interface for dataset labeling and conducted extensive experiments with human reviewers. Our analysis shows that the MQA-metric outperforms traditional metrics like BLEU, ROUGE and METEOR. Unlike existing metrics, MQA-metric leverages semantic comprehension through large language models (LLMs), enabling it to capture contextual nuances and synonymous expressions more effectively. This approach sets a standard for evaluating QA systems by prioritizing semantic accuracy over surface-level similarities. The proposed metric correlates better with human judgment, making it a more reliable tool for evaluating QA systems. Our contributions include the development of a robust evaluation workflow, creation of high-quality datasets, and an extensive comparison with existing evaluation methods. The results indicate that our model-based approach provides a significant improvement in assessing the quality of QA systems, which is crucial for their practical application and trustworthiness. © 2025 World Scientific Publishing Company.

Description

Keywords

Evaluation Metric, Generative Model, Large Language Model, Natural Language Processing, Question Answering, Transformer Models

Fields of Science

Citation

WoS Q

Q4

Scopus Q

Q3
OpenCitations Logo
OpenCitations Citation Count
N/A

Source

International Journal of Software Engineering and Knowledge Engineering

Volume

35

Issue

2

Start Page

243

End Page

262

Collections

PlumX Metrics
Citations

Scopus : 0

Captures

Mendeley Readers : 4

Page Views

1

checked on Apr 13, 2026

Google Scholar Logo
Google Scholar™
OpenAlex Logo
OpenAlex FWCI
0.7252

Sustainable Development Goals

SDG data is not available