·厦门大学信息学院超智医疗创新研究中心

当前位置:

中文主页 >> 中心概览 >> 论文成果

OVQA: A Clinically Generated Visual Question Answering Dataset

发布时间:2022-07-11 点击次数:
DOI码:
10.1145/3477495.3531724
所属单位:
School of Informatics, Xiamen University,
发表刊物:
SIGIR 2022
关键字:
Medical visual question answering; Benchmarking dataset; Semiautomatic data generation
摘要:
Medical visual question answering (Med-VQA) is a challenging problem that aims to take a medical image and a clinical question about the image as input and output a correct answer in natural language. Current medical systems often require large-scale and high-quality labeled data for training and evaluation. To address the challenge, we present a new dataset, denoted by OVQA, which is generated from electronic medical records. We develop a semiautomatic data generation tool for constructing the dataset. First, medical entities are automatically extracted from medical records and filled into predefined templates for generating question and answer pairs. These pairs are then combined with medical images extracted from corresponding medical records, to generate candidates for visual question answering (VQA). The candidates are finally verified with high-quality labels annotated by experienced physicians. To evaluate the quality of OVQA, we conduct comprehensive experiments on state-of-the-art methods for the Med-VQA task to our dataset. The results show that our OVQA can be used as a benchmarking dataset for evaluating existing Med-VQA systems. The dataset can be downloaded from http://47.94.174.82/.
第一作者:
Yefan Huang
通讯作者:
Xiaoli Wang
合写作者:
Feiyan Liu,Guofeng Huang
论文类型:
Original Research
文献类型:
Conference Paper
页面范围:
2924 - 2938
是否译文:
发表时间:
2022-07-11