AI meets physics: a comprehensive survey

L Jiao, X Song, C You, X Liu, L Li, P Chen… - Artificial Intelligence …, 2024 - Springer
Uncovering the mechanisms of physics is driving a new paradigm in artificial intelligence
(AI) discovery. Today, physics has enabled us to understand the AI paradigm in a wide …

Foundation models for music: A survey

Y Ma, A Øland, A Ragni, BMS Del Sette, C Saitis… - arxiv preprint arxiv …, 2024 - arxiv.org
In recent years, foundation models (FMs) such as large language models (LLMs) and latent
diffusion models (LDMs) have profoundly impacted diverse sectors, including music. This …

C2kd: Bridging the modality gap for cross-modal knowledge distillation

F Huo, W Xu, J Guo, H Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Existing Knowledge Distillation (KD) methods typically focus on transferring
knowledge from a large-capacity teacher to a low-capacity student model achieving …

Stable diffusion is unstable

C Du, Y Li, Z Qiu, C Xu - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Recently, text-to-image models have been thriving. Despite their powerful generative
capacity, our research has uncovered a lack of robustness in this generation process …

Cloud-device collaborative learning for multimodal large language models

G Wang, J Liu, C Li, Y Zhang, J Ma… - Proceedings of the …, 2024 - openaccess.thecvf.com
The burgeoning field of Multimodal Large Language Models (MLLMs) has exhibited
remarkable performance in diverse tasks such as captioning commonsense reasoning and …

Detkds: Knowledge distillation search for object detectors

L Li, Y Bao, P Dong, C Yang, A Li, W Luo… - … on Machine Learning, 2024 - openreview.net
In this paper, we present DetKDS, the first framework that searches for optimal detection
distillation policies. Manual design of detection distillers becomes challenging and time …

FreeKD: Knowledge distillation via semantic frequency prompt

Y Zhang, T Huang, J Liu, T Jiang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Knowledge distillation (KD) has been applied to various tasks successfully and
mainstream methods typically boost the student model via spatial imitation losses. However …

Revisit the power of vanilla knowledge distillation: from small scale to large scale

Z Hao, J Guo, K Han, H Hu, C Xu… - Advances in Neural …, 2023 - proceedings.neurips.cc
The tremendous success of large models trained on extensive datasets demonstrates that
scale is a key ingredient in achieving superior results. Therefore, the reflection on the …

Attention-guided feature distillation for semantic segmentation

AM Mansourian, A Jalali, R Ahmadi… - arxiv preprint arxiv …, 2024 - arxiv.org
In contrast to existing complex methodologies commonly employed for distilling knowledge
from a teacher to a student, this paper showcases the efficacy of a simple yet powerful …

[HTML][HTML] Computer vision model compression techniques for embedded systems: A survey

A Lopes, FP dos Santos, D de Oliveira, M Schiezaro… - Computers & …, 2024 - Elsevier
Deep neural networks have consistently represented the state of the art in most computer
vision problems. In these scenarios, larger and more complex models have demonstrated …