Please use this identifier to cite or link to this item:
Full metadata record
DC FieldValueLanguage
dc.contributor.authorHuang, X-
dc.contributor.authorFang, Y-
dc.contributor.authorLu, M-
dc.contributor.authorYao, Y-
dc.contributor.authorLi, M-
dc.identifier.citationIEEE Access, 2019, 7 pp. 65757 - 65765en_US
dc.description.abstractAnnotating radiographic images with tags is an indispensable preliminary work in computer-aided medical research, which requires professional physician participated in and is quite timeconsuming. Therefore, how to automatically annotate radiographic images has become the focus of researchers. However, image report texts, containing crucial radiologic information, have not to be given enough attention for images annotation. In this paper, we propose a neural sequence-to-sequence annotation model. Especially, in the decoding phase, a probability is first learned to copy existing words from report texts or generate new words. Second, to incorporate the patient’s background information, ‘‘indication’’ section of the report is encoded as a sentence embedding, and concatenated with the decoder neural unit input. What’s more, we devise a more reasonable evaluation metric for this annotation task, aiming at assessing the importance of different words. On the Open-i dataset, our model outperforms existing non-neural and neural baselines under the BLEU-4 metrics. To our best knowledge, we are the first to use sequence-to-sequence model for radiographic image annotation.en_US
dc.format.extent65757 - 65765-
dc.subjectchest radiology reporten_US
dc.subjectdeep learningen_US
dc.subjectend-to-end modelen_US
dc.titleAn Annotation Model on End-to-End Chest Radiology Reportsen_US
dc.relation.isPartOfIEEE Access-
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdf6.22 MBAdobe PDFView/Open

Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.