Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/28249
Full metadata record
DC FieldValueLanguage
dc.contributor.authorZhang, S-
dc.contributor.authorChen, Y-
dc.contributor.authorSun, Y-
dc.contributor.authorWang, F-
dc.contributor.authorShi, H-
dc.contributor.authorWang, H-
dc.date.accessioned2024-02-07T18:22:47Z-
dc.date.available2024-02-07T18:22:47Z-
dc.date.issued2023-12-26-
dc.identifierORCID ID: Siyu Zhang https://orcid.org/0000-0002-0001-0204-
dc.identifierORCID iD: Yaoru Sun https://orcid.org/0000-0002-2179-0713-
dc.identifierORCID iD: Fang Wang https://orcid.org/0000-0003-1987-9150-
dc.identifierarXiv:2307.14142v1 [cs.CV]-
dc.identifier.citationZhang, S. et al. (2023) 'LOIS: Looking Out of Instance Semantics for Visual Question Answering', IEEE Transactions on Multimedia, 0 (early access), pp. 1 - 13. doi: 10.1109/TMM.2023.3347093.en_US
dc.identifier.issn1520-9210-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/28249-
dc.descriptionThe fiel archived on this institutional repository is a preprint available at arXiv:2307.14142v1 [cs.CV] (https://doi.org/10.48550/arXiv.2307.14142). It has not been certified by peer review. You are advised to use the peer reviewed version published by IEEE at https://doi.org/10.1109/TMM.2023.3347093 .-
dc.description.abstractVisual question answering (VQA) has been intensively studied as a multimodal task, requiring efforts to bridge vision and language for correct answer inference. Recent attempts have developed various attention-based modules for solving VQA tasks. However, the performance of model inference is largely bottlenecked by visual semantic comprehension. Most existing detection methods rely on bounding boxes, remaining a serious challenge for VQA models to comprehend and correctly infer the causal nexus of contextual object semantics in images. To this end, we propose a finer model framework without bounding boxes in this work, termed <italic>Looking Out of Instance Semantics (LOIS)</italic> to address this crucial issue. LOIS can achieve more fine-grained feature descriptions to generate visual facts. Furthermore, to overcome the label ambiguity caused by instance masks, two types of relation attention modules: 1) intra-modality and 2) inter-modality, are devised to infer the correct answers from different visual features. Specifically, we implement a mutual relation attention module to model sophisticated and deeper visual semantic relations between instance objects and background information. In addition, our proposed attention model can further analyze salient image regions by focusing on important word-related questions. Experimental results on four benchmark VQA datasets prove that our proposed method has favorable performance in improving visual reasoning capability. Our code is available on GitHub (<uri>https://github.com/ArcherCYM/LOIS</uri>).en_US
dc.format.extent1 - 13-
dc.format.mediumPrint-Electronic-
dc.language.isoen_USen_US
dc.publisherIEEEen_US
dc.relation.urihttps://arxiv.org/abs/2307.14142-
dc.rightsCopyright © 2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/)..-
dc.rights.urihttps://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.subjectvisual question answering (VQA)en_US
dc.subjectinstance semanticsen_US
dc.subjectvisual featuresen_US
dc.subjectmultimodal relation attentionen_US
dc.titleLOIS: Looking Out of Instance Semantics for Visual Question Answeringen_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1109/TMM.2023.3347093-
dc.relation.isPartOfIEEE Transactions on Multimedia-
pubs.publication-statusPublished-
pubs.volume0-
dc.identifier.eissn1941-0077-
dc.rights.holderIEEE-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdfCopyright © 2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/)..5.03 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.