PromptKD: Unsupervised Prompt Distillation for Vision-Language Models

Zheng Li1, Xiang Li 1*, Xinyi Fu2, Xin Zhang1, Weiqiang Wang2, Shuo Chen3, Jian Yang1*

1 PCALab, VCIP, College of Computer Science, Nankai University,
2Tiansuan Lab, Ant Group, 3 RIKEN
CVPR 2024
*Indicates Corresponding Author

zhengli97@mail.nankai.edu.com

Abstract

In this paper, we introduce an unsupervised domain prompt distillation framework, which aims to transfer the knowledge of a larger teacher model to a lightweight target model through prompt-driven imitation using unlabeled domain images.

Specifically, our framework consists of two distinct stages. In the initial stage, we pre-train a large CLIP teacher model using domain (few-shot) labels. After pre-training, we leverage the unique decoupled-modality characteristics of CLIP by pre-computing and storing the text features as class vectors only once through the teacher text encoder.
In the subsequent stage, the stored class vectors are shared across teacher and student image encoders for calculating the predicted logits. Further, we align the logits of both the teacher and student models via KL divergence loss, encouraging the student image encoder to generate similar probability distributions to the teacher through the learnable prompts.

The proposed prompt distillation process eliminates the reliance on labeled data, enabling the algorithm to leverage a vast amount of unlabeled images within the domain. Finally, the well-trained student image encoders and pre-stored text features (class vectors) are utilized for inference.

To our best knowledge, we are the first to (1) perform unsupervised domain-specific prompt-driven knowledge distillation for CLIP, and (2) establish a practical pre-storing mechanism of text features as shared class vectors between teacher and student. Extensive experiments on 11 recognition datasets demonstrate the effectiveness of our method.

Advantages

(1). We provide an efficient and simple CLIP distillation framework that can compress the knowledge of the ViT-L/14 CLIP model into the smaller ViT-B/16 CLIP model through prompt distillation. Notably, after distillation, the smaller ViT-B/16 CLIP model achieves better performance than the vanilla ViT-L/14 CLIP on ImageNet-1K (77.62 HM vs. 76.52 HM).

(2). Due to the characteristics of CLIP, the text encoder only requires a single forward calculation for all classes. Inspired by this, we propose to reuse the existing higher-quality teacher text features instead of training the student's own text encoder. This approach not only maintains the quality of the text features but also significantly reduces computational costs and memory usage during training.

(3). The existence of the teacher CLIP model liberates us from the need for labeled training samples. Learning from large amounts of unlabeled domain images, which is easily accessible, allows the prompt to learn richer and more generalized domain representations. This significantly enhances student performance and makes our method easier to apply in real world scenarios.

Experimental Results

Base-to-Novel Experiments
fail
Table 1. Comparison with existing state-of-the-art methods on base-to-novel generalization. Our PromptKD demonstrates strong generalization ability and achieves significant improvements on 11 recognition datasets given the ViT-B/16 image encoder of the CLIP model. The symbol △ denotes the performance improvement compared to the previous SOTA method.

fail
Figure 3. Harmonic mean (HM) comparison on base-to-novel generalization.

Cross Dataset Experiments
fail
Table 2. Comparison of PromptKD with existing advanced approaches on cross-dataset benchmark evaluation. Based on our pipeline, we perform unsupervised prompt distillation using the unlabeled domain data respectively (i.e., the transductive setting). The source model is trained on ImageNet. "ZSL" denotes the setting type for Zero-Shot Learning.

Comparison with Other Methods
fail
Table 3. Comparison with existing works using unlabeled data on the Flowers102 dataset. Our method performs better than previous methods.

Teacher Pre-training Methods
fail
Table 4. Comparison of different pre-training methods. Teacher pre-training with PromptSRC brings the best student performance. Notably, any type of teacher model can enhance the student model with a non-trivial improvement.

Distillation with Different Teachers
fail
Figure 4. Comparison of distillation results for pre-trained teachers with different capacities. Better teachers lead to better distillation performance.

Video Demonstration

Video demonstration coming soon.

BibTeX

If you find our paper is helpful for your research, please consider citing our paper:

@inproceedings{li2024promptkd,
      title={PromptKD: Unsupervised Prompt Distillation for Vision-Language Models},
      author={Li, Zheng and Li, Xiang and Fu, Xinyi and Zhang, Xin and Wang, Weiqiang and Chen, Shuo and Yang, Jian},
      booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
      year={2024}
    }