Select and Distill: Selective Dual-Teacher Knowledge Transfer for Continual Learning on Vision-Language Models

Type: Preprint

Publication Date: 2024-03-14

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2403.09296

Abstract

Large-scale vision-language models (VLMs) have shown a strong zero-shot generalization capability on unseen-domain data. However, when adapting pre-trained VLMs to a sequence of downstream tasks, they are prone to forgetting previously learned knowledge and degrade their zero-shot classification capability. To tackle this problem, we propose a unique Selective Dual-Teacher Knowledge Transfer framework that leverages the most recent fine-tuned and the original pre-trained VLMs as dual teachers to preserve the previously learned knowledge and zero-shot capabilities, respectively. With only access to an unlabeled reference dataset, our proposed framework performs a selective knowledge distillation mechanism by measuring the feature discrepancy from the dual teacher VLMs. Consequently, our selective dual-teacher knowledge distillation would mitigate catastrophic forgetting of previously learned knowledge while preserving the zero-shot capabilities from pre-trained VLMs. Through extensive experiments on benchmark datasets, we show that our proposed framework is favorable against state-of-the-art continual learning approaches for preventing catastrophic forgetting and zero-shot degradation.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Preventing Zero-Shot Transfer Degradation in Continual Learning of Vision-Language Models 2023 Zangwei Zheng
Mingyuan Ma
Kai Wang
Ziheng Qin
Xiangyu Yue
Yang You
+ PDF Chat Preventing Zero-Shot Transfer Degradation in Continual Learning of Vision-Language Models 2023 Zangwei Zheng
Mingyuan Ma
Kai Wang
Ziheng Qin
Xiangyu Yue
Yang You
+ PDF Chat Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models 2024 Longxiang Tang
Zhuotao Tian
Kai Li
Chunming He
Hantao Zhou
Hengshuang Zhao
Xiu Li
Jiaya Jia
+ PDF Chat Multi-Stage Knowledge Integration of Vision-Language Models for Continual Learning 2024 Hongsheng Zhang
Zhong Ji
Jingren Liu
Yanwei Pang
Jungong Han
+ PDF Chat Pre-trained Vision and Language Transformers Are Few-Shot Incremental Learners 2024 Keon-Hee Park
Kyungwoo Song
Gyeong-Moon Park
+ PDF Chat Improving Zero-shot Generalization of Learned Prompts via Unsupervised Knowledge Distillation 2024 Marco Mistretta
Alberto Baldrati
Marco Bertini
Andrew D. Bagdanov
+ PDF Chat Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters 2024 Jiazuo Yu
Yunzhi Zhuge
Lu Zhang
Dong Wang
Huchuan Lu
You He
+ Robust Fine-Tuning of Vision-Language Models for Domain Generalization 2023 Kevin Vogt-Lowell
Noah Lee
Theodoros Tsiligkaridis
Marc Vaillant
+ PDF Chat Robust Fine-Tuning of Vision-Language Models for Domain Generalization 2023 Kevin Vogt-Lowell
Noah Lee
Theodoros Tsiligkaridis
Marc Vaillant
+ PDF Chat Retaining and Enhancing Pre-trained Knowledge in Vision-Language Models with Prompt Ensembling 2024 Donggeun Kim
Yujin Jo
M.M. Lee
Taesup Kim
+ PDF Chat Advancing Cross-domain Discriminability in Continual Learning of Vison-Language Models 2024 Yicheng Xu
Yuxin Chen
Jiahao Nie
Yusong Wang
Huiping Zhuang
Manabu Okumura
+ PDF Chat Negative Yields Positive: Unified Dual-Path Adapter for Vision-Language Models 2024 Ce Zhang
Simon Stepputtis
Katia Sycara
Yaqi Xie
+ Task-Attentive Transformer Architecture for Continual Learning of Vision-and-Language Tasks Using Knowledge Distillation 2023 Yuliang Cai
Jesse Thomason
Mohammad Rostami
+ Task-Attentive Transformer Architecture for Continual Learning of Vision-and-Language Tasks Using Knowledge Distillation 2023 Yuliang Cai
Jesse Thomason
Mohammad Rostami
+ PDF Chat Fully Fine-tuned CLIP Models are Efficient Few-Shot Learners 2024 Mushui Liu
Bozheng Li
Yunlong Yu
+ Combining inherent knowledge of vision-language models with unsupervised domain adaptation through self-knowledge distillation 2023 Thomas Westfechtel
Dexuan Zhang
Tatsuya Harada
+ PDF Chat Task Residual for Tuning Vision-Language Models 2023 Tao Yu
Zhihe Lu
Xin Jin
Zhibo Chen
Xinchao Wang
+ Test-Time Adaptation with CLIP Reward for Zero-Shot Generalization in Vision-Language Models 2023 Shuai Zhao
Xiaohan Wang
Linchao Zhu
Yi Yang
+ PDF Chat Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models 2024 Jianyuan Guo
Hanting Chen
Chengcheng Wang
Kai Han
Chang Xu
Yunhe Wang
+ Rethinking Task Sampling for Few-shot Vision-Language Transfer Learning 2022 Zhenhailong Wang
Hang Yu
Manling Li
Han Zhao
Heng Ji

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors