Overview of the Medical Question Answering Task at TREC 2017 LiveQA

Asma Ben Abacha, Eugene Agichtein, Yuval Pinter, Dina Demner-Fushman

Research output: Contribution to conferencePaperpeer-review

Abstract

We present an overview of the medical question answering task organized at the TREC 2017 LiveQA track. The task addresses the automatic answering of consumer health questions received by the U.S. National Library of Medicine. We provided both training question-answer pairs, and test questions with reference answers1. All questions were manually annotated with the main entities (foci) and question types. The medical task received eight runs from five participating teams. Different approaches have been applied, including classical answer retrieval based on question analysis and similar question retrieval. In particular, several deep learning approaches were tested, including attentional encoder-decoder networks, long short-term memory networks and convolutional neural networks. The training datasets were both from the open domain and the medical domain. We discuss the obtained results and give some insights for future research in medical question answering.

Original languageAmerican English
StatePublished - 1 Jan 2017
Externally publishedYes
Event26th Text REtrieval Conference, TREC 2017 - Gaithersburg, United States
Duration: 15 Nov 201717 Nov 2017

Conference

Conference26th Text REtrieval Conference, TREC 2017
Country/TerritoryUnited States
CityGaithersburg
Period15/11/1717/11/17

All Science Journal Classification (ASJC) codes

  • Language and Linguistics
  • Computer Science Applications
  • Linguistics and Language

Fingerprint

Dive into the research topics of 'Overview of the Medical Question Answering Task at TREC 2017 LiveQA'. Together they form a unique fingerprint.

Cite this