[HTML][HTML] A survey of transformers
Transformers have achieved great success in many artificial intelligence fields, such as
natural language processing, computer vision, and audio processing. Therefore, it is natural …
natural language processing, computer vision, and audio processing. Therefore, it is natural …
[HTML][HTML] A review of green artificial intelligence: Towards a more sustainable future
Green artificial intelligence (AI) is more environmentally friendly and inclusive than
conventional AI, as it not only produces accurate results without increasing the …
conventional AI, as it not only produces accurate results without increasing the …
Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting
Long-term time series forecasting is challenging since prediction accuracy tends to
decrease dramatically with the increasing horizon. Although Transformer-based methods …
decrease dramatically with the increasing horizon. Although Transformer-based methods …
Efficient large language models: A survey
Large Language Models (LLMs) have demonstrated remarkable capabilities in important
tasks such as natural language understanding and language generation, and thus have the …
tasks such as natural language understanding and language generation, and thus have the …
Cswin transformer: A general vision transformer backbone with cross-shaped windows
Abstract We present CSWin Transformer, an efficient and effective Transformer-based
backbone for general-purpose vision tasks. A challenging issue in Transformer design is …
backbone for general-purpose vision tasks. A challenging issue in Transformer design is …
Multiscale vision transformers
Abstract We present Multiscale Vision Transformers (MViT) for video and image recognition,
by connecting the seminal idea of multiscale feature hierarchies with transformer models …
by connecting the seminal idea of multiscale feature hierarchies with transformer models …
Fnet: Mixing tokens with fourier transforms
We show that Transformer encoder architectures can be sped up, with limited accuracy
costs, by replacing the self-attention sublayers with simple linear transformations that" mix" …
costs, by replacing the self-attention sublayers with simple linear transformations that" mix" …
A-vit: Adaptive tokens for efficient vision transformer
We introduce A-ViT, a method that adaptively adjusts the inference cost of vision transformer
ViT for images of different complexity. A-ViT achieves this by automatically reducing the …
ViT for images of different complexity. A-ViT achieves this by automatically reducing the …
Transformers in vision: A survey
Astounding results from Transformer models on natural language tasks have intrigued the
vision community to study their application to computer vision problems. Among their salient …
vision community to study their application to computer vision problems. Among their salient …
Diffusion models without attention
In recent advancements in high-fidelity image generation Denoising Diffusion Probabilistic
Models (DDPMs) have emerged as a key player. However their application at high …
Models (DDPMs) have emerged as a key player. However their application at high …