Diswot: Student architecture search for distillation without training

P Dong, L Li, Z Wei - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Abstract Knowledge distillation (KD) is an effective training strategy to improve the
lightweight student models under the guidance of cumbersome teachers. However, the large …

Automated knowledge distillation via monte carlo tree search

L Li, P Dong, Z Wei, Y Yang - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
In this paper, we present Auto-KD, the first automated search framework for optimal
knowledge distillation design. Traditional distillation techniques typically require handcrafted …

C2kd: Bridging the modality gap for cross-modal knowledge distillation

F Huo, W Xu, J Guo, H Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Existing Knowledge Distillation (KD) methods typically focus on transferring
knowledge from a large-capacity teacher to a low-capacity student model achieving …

Kd-zero: Evolving knowledge distiller for any teacher-student pairs

L Li, P Dong, A Li, Z Wei… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Knowledge distillation (KD) has emerged as an effective technique for compressing
models that can enhance the lightweight model. Conventional KD methods propose various …

Saswot: Real-time semantic segmentation architecture search without training

C Zhu, L Li, Y Wu, Z Sun - Proceedings of the AAAI Conference on …, 2024 - ojs.aaai.org
In this paper, we present SasWOT, the first training-free Semantic segmentation Architecture
Search (SAS) framework via an auto-discovery proxy. Semantic segmentation is widely used …

Pruner-zero: Evolving symbolic pruning metric from scratch for large language models

P Dong, L Li, Z Tang, X Liu, X Pan, Q Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
Despite the remarkable capabilities, Large Language Models (LLMs) face deployment
challenges due to their extensive size. Pruning methods drop a subset of weights to …

Auto-prox: Training-free vision transformer architecture search via automatic proxy discovery

Z Wei, P Dong, Z Hui, A Li, L Li, M Lu, H Pan… - Proceedings of the AAAI …, 2024 - ojs.aaai.org
The substantial success of Vision Transformer (ViT) in computer vision tasks is largely
attributed to the architecture design. This underscores the necessity of efficient architecture …

Applications of knowledge distillation in remote sensing: A survey

Y Himeur, N Aburaed, O Elharrouss, I Varlamis… - Information …, 2024 - Elsevier
With the ever-growing complexity of models in the field of remote sensing (RS), there is an
increasing demand for solutions that balance model accuracy with computational efficiency …

Detkds: Knowledge distillation search for object detectors

L Li, Y Bao, P Dong, C Yang, A Li, W Luo… - … on Machine Learning, 2024 - openreview.net
In this paper, we present DetKDS, the first framework that searches for optimal detection
distillation policies. Manual design of detection distillers becomes challenging and time …

Parameter-efficient and student-friendly knowledge distillation

J Rao, X Meng, L Ding, S Qi, X Liu… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Pre-trained models are frequently employed in multimodal learning. However, these models
have too many parameters and need too much effort to fine-tune the downstream tasks …