Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/27482
Full metadata record
DC FieldValueLanguage
dc.contributor.authorNaseem, U-
dc.contributor.authorKhushi, M-
dc.contributor.authorKim, J-
dc.date.accessioned2023-10-31T12:53:37Z-
dc.date.available2023-10-31T12:53:37Z-
dc.date.issued2022-03-31-
dc.identifierORCID iD: Usman Naseem https://orcid.org/0000-0003-0191-7171-
dc.identifierORCID iD: Matloob Khushi https://orcid.org/0000-0001-7792-2327-
dc.identifierORCID iD: Jinman Kim https://orcid.org/0000-0001-5960-1060-
dc.identifier.citationNaseem, U., Khushi, M. and Kim, J. (2022) 'Vision-Language Transformer for Interpretable Pathology Visual Question Answering', IEEE Journal of Biomedical and Health Informatics, 27 (4), pp. 1681 - 1690. doi: 10.1109/JBHI.2022.3163751.en_US
dc.identifier.issn2168-2194-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/27482-
dc.description.abstractPathology visual question answering (PathVQA) attempts to answer a medical question posed by pathology images. Despite its great potential in healthcare, it is not widely adopted because it requires interactions on both the image (vision) and question (language) to generate an answer. Existing methods focused on treating vision and language features independently, which were unable to capture the high and low-level interactions that are required for VQA. Further, these methods failed to offer capabilities to interpret the retrieved answers, which are obscure to humans where the models’ interpretability to justify the retrieved answers has remained largely unexplored. Motivated by these limitations, we introduce a vision-language transformer that embeds vision (images) and language (questions) features for an interpretable PathVQA. We present an interpretable tra nsformer-based P ath- VQA (TraP-VQA), where we embed transformers’ encoder layers with vision and language features extracted using pre-trained CNN and domain-specific language model (LM), respectively. A decoder layer is then embedded to upsample the encoded features for the final prediction for PathVQA. Our experiments showed that our TraP-VQA outperformed the state-of-the-art comparative methods with public PathVQA dataset. Our experiments validated the robustness of our model on another medical VQA dataset, and the ablation study demonstrated the capability of our integrated transformer-based vision-language model for PathVQA. Finally, we present the visualization results of both text and images, which explain the reason for a retrieved answer in PathVQA.en_US
dc.description.sponsorshipARC (Grant Number: DP200103748).en_US
dc.format.extent1681 - 1690-
dc.format.mediumPrint-Electronic-
dc.languageEnglish-
dc.language.isoen_USen_US
dc.publisherIEEEen_US
dc.rightsCopyright © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/).-
dc.rights.urihttps://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.subjectpathology imagesen_US
dc.subjectinterpretabilityen_US
dc.subjectvisual question answeringen_US
dc.subjectvision-languageen_US
dc.titleVision-Language Transformer for Interpretable Pathology Visual Question Answeringen_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1109/JBHI.2022.3163751-
dc.relation.isPartOfIEEE Journal of Biomedical and Health Informatics-
pubs.issue4-
pubs.publication-statusPublished-
pubs.volume27-
dc.identifier.eissn2168-2208-
dc.rights.holderIEEE-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/).16.94 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.