Research output: Chapter in Book/Report/Conference proceeding › Conference paper › peer-review
What Would it Take to get Biomedical QA Systems into Practice? / Kell, Gregory; Marshall, Iain J.; Wallace, Byron C. et al.
Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021. ed. / Adam Fisch; Alon Talmor; Danqi Chen; Eunsol Choi; Minjoon Seo; Patrick Lewis; Robin Jia; Sewon Min. Association for Computational Linguistics (ACL), 2021. p. 28-41 (Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021).Research output: Chapter in Book/Report/Conference proceeding › Conference paper › peer-review
}
TY - CHAP
T1 - What Would it Take to get Biomedical QA Systems into Practice?
AU - Kell, Gregory
AU - Marshall, Iain J.
AU - Wallace, Byron C.
AU - Jaun, André
N1 - Funding Information: This work was supported in part by the National Institutes of Health (NIH), grant R01-LM012086. GK holds a doctoral studentship co-sponsored by Metadvice and the Guy's and St Thomas' Biomedical Research Centre. Funding Information: This work was supported in part by the National Institutes of Health (NIH), grant R01-LM012086. Publisher Copyright: © 2021 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - Medical question answering (QA) systems have the potential to answer clinicians' uncertainties about treatment and diagnosis on-demand, informed by the latest evidence. However, despite the significant progress in general QA made by the NLP community, medical QA systems are still not widely used in clinical environments. One likely reason for this is that clinicians may not readily trust QA system outputs, in part because transparency, trustworthiness, and provenance have not been key considerations in the design of such models. In this paper we discuss a set of criteria that, if met, we argue would likely increase the utility of biomedical QA systems, which may in turn lead to adoption of such systems in practice. We assess existing models, tasks, and datasets with respect to these criteria, highlighting shortcomings of previously proposed approaches and pointing toward what might be more usable QA systems.
AB - Medical question answering (QA) systems have the potential to answer clinicians' uncertainties about treatment and diagnosis on-demand, informed by the latest evidence. However, despite the significant progress in general QA made by the NLP community, medical QA systems are still not widely used in clinical environments. One likely reason for this is that clinicians may not readily trust QA system outputs, in part because transparency, trustworthiness, and provenance have not been key considerations in the design of such models. In this paper we discuss a set of criteria that, if met, we argue would likely increase the utility of biomedical QA systems, which may in turn lead to adoption of such systems in practice. We assess existing models, tasks, and datasets with respect to these criteria, highlighting shortcomings of previously proposed approaches and pointing toward what might be more usable QA systems.
UR - http://www.scopus.com/inward/record.url?scp=85132162300&partnerID=8YFLogxK
M3 - Conference paper
AN - SCOPUS:85132162300
T3 - Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021
SP - 28
EP - 41
BT - Proceedings of the 3rd Workshop on Machine Reading for Question Answering, MRQA 2021
A2 - Fisch, Adam
A2 - Talmor, Alon
A2 - Chen, Danqi
A2 - Choi, Eunsol
A2 - Seo, Minjoon
A2 - Lewis, Patrick
A2 - Jia, Robin
A2 - Min, Sewon
PB - Association for Computational Linguistics (ACL)
T2 - 3rd Workshop on Machine Reading for Question Answering, MRQA 2021
Y2 - 10 November 2021
ER -
King's College London - Homepage
© 2020 King's College London | Strand | London WC2R 2LS | England | United Kingdom | Tel +44 (0)20 7836 5454