Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/29101
Full metadata record
DC FieldValueLanguage
dc.contributor.authorChen, Z-
dc.contributor.authorWang, W-
dc.contributor.authorZhao, Z-
dc.contributor.authorSu, F-
dc.contributor.authorMen, A-
dc.contributor.authorMeng, H-
dc.date.accessioned2024-06-02T09:23:31Z-
dc.date.available2024-06-02T09:23:31Z-
dc.date.issued2024-06-17-
dc.identifierORCiD: Hongying Meng https://orcid.org/0000-0002-8836-1382-
dc.identifierarXiv:2404.09011v1 [cs.CV]-
dc.identifier.citationChen, Z. et al. (2024) 'PracticalDG: Perturbation Distillation on Vision-Language Models for Hybrid Domain Generalization.', Proceedings of The IEEE/CVF Conference on Computer Vision and Pattern Recognition 2024, Seattle, WA, USA, 17-21 June, pp. 1 - 11. [arXiv doi: 10.48550/arXiv.2404.09011].en_US
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/29101-
dc.descriptionA preprint version of the conference paper is available under a CC BY-NC-SA license at arXiv:2404.09011v1 [cs.CV], https://arxiv.org/abs/2404.09011 . It may not have been certified by peer review. Comments: Accepted to CVPR2024. You are advised to consult the final version to be published by IEEE in due course.en_US
dc.descriptionThis CVPR paper is the Open Access version, provided by the Computer Vision Foundation. Except for this watermark, it is identical to the accepted version; the final published version of the proceedings is available on IEEE Xplore. This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright.-
dc.description.abstractDomain Generalization (DG) aims to resolve distribution shifts between source and target domains and current DG methods are default to the setting that data from source and target domains share identical categories. Nevertheless there exists unseen classes from target domains in practical scenarios. To address this issue Open Set Domain Generalization (OSDG) has emerged and several methods have been exclusively proposed. However most existing methods adopt complex architectures with slight improvement compared with DG methods. Recently vision-language models (VLMs) have been introduced in DG following the fine-tuning paradigm but consume huge training overhead with large vision models. Therefore in this paper we innovate to transfer knowledge from VLMs to lightweight vision models and improve the robustness by introducing Perturbation Distillation (PD) from three perspectives including Score Class and Instance (SCI) named SCI-PD. Moreover previous methods are oriented by the benchmarks with identical and fixed splits ignoring the divergence between source domains. These methods are revealed to suffer from sharp performance decay with our proposed new benchmark Hybrid Domain Generalization (HDG) and a novel metric H^ 2 -CV which construct various splits to comprehensively assess the robustness of algorithms. Extensive experiments demonstrate that our method outperforms state-of-the-art algorithms on multiple datasets especially improving the robustness when confronting data scarcity.en_US
dc.description.sponsorshipChinese National Natural Science Foundation under Grants (62076033).en_US
dc.format.extent1 - 11-
dc.format.mediumElectronic-
dc.language.isoen_USen_US
dc.publisherIEEEen_US
dc.relation.urihttps://arxiv.org/abs/2404.09011-
dc.relation.urihttps://openaccess.thecvf.com/content/CVPR2024/html/Chen_PracticalDG_Perturbation_Distillation_on_Vision-Language_Models_for_Hybrid_Domain_Generalization_CVPR_2024_paper.html-
dc.rightsCopyright © 2024 The Author(s). This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (https://creativecommons.org/licenses/by-nc-sa/4.0/).-
dc.rightsCopyright © 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/).-
dc.rights.urihttps://creativecommons.org/licenses/by-nc-sa/4.0/-
dc.rights.urihttps://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.subjectcomputer vision and pattern recognition (cs.CV)en_US
dc.titlePracticalDG: Perturbation Distillation on Vision-Language Models for Hybrid Domain Generalizationen_US
dc.typeConference Paperen_US
dc.date.dateAccepted2024-03-06-
dc.relation.isPartOfProceedings of The IEEE/CVF Conference on Computer Vision and Pattern Recognition 2024-
pubs.volumeabs/2404.09011-
dc.rights.licensehttps://creativecommons.org/licenses/by-nc-sa/4.0/legalcode.en-
dc.rights.holderThe Author(s)-
Appears in Collections:Dept of Electronic and Electrical Engineering Research Papers

Files in This Item:
File Description SizeFormat 
Preprint.pdfCopyright © 2024 The Author(s). This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (https://creativecommons.org/licenses/by-nc-sa/4.0/).1.73 MBAdobe PDFView/Open
FullText.pdfThis CVPR paper is the Open Access version, provided by the Computer Vision Foundation. Except for this watermark, it is identical to the accepted version; the final published version of the proceedings is available on IEEE Xplore. Copyright © 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works (see: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/).1.83 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons