Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Agent attention: On the integration of softmax and linear attention
The attention module is the key component in Transformers. While the global attention
mechanism offers high expressiveness, its excessive computational cost restricts its …
mechanism offers high expressiveness, its excessive computational cost restricts its …
CATNet: Cascaded attention transformer network for marine species image classification
Complex physicochemical environmental effects result in the underwater species images'
highly intricate and diverse backgrounds, which poses significant challenges for identifying …
highly intricate and diverse backgrounds, which poses significant challenges for identifying …
A survey on transformer compression
Large models based on the Transformer architecture play increasingly vital roles in artificial
intelligence, particularly within the realms of natural language processing (NLP) and …
intelligence, particularly within the realms of natural language processing (NLP) and …
Sam-6d: Segment anything model meets zero-shot 6d object pose estimation
Zero-shot 6D object pose estimation involves the detection of novel objects with their 6D
poses in cluttered scenes presenting significant challenges for model generalizability …
poses in cluttered scenes presenting significant challenges for model generalizability …
Efficient diffusion transformer with step-wise dynamic attention mediators
This paper identifies significant redundancy in the query-key interactions within self-attention
mechanisms of diffusion transformer models, particularly during the early stages of …
mechanisms of diffusion transformer models, particularly during the early stages of …
Vit-comer: Vision transformer with convolutional multi-scale feature interaction for dense predictions
Abstract Although Vision Transformer (ViT) has achieved significant success in computer
vision it does not perform well in dense prediction tasks due to the lack of inner-patch …
vision it does not perform well in dense prediction tasks due to the lack of inner-patch …
Hypformer: Exploring efficient transformer fully in hyperbolic space
Hyperbolic geometry have shown significant potential in modeling complex structured data,
particularly those with underlying tree-like and hierarchical structures. Despite the …
particularly those with underlying tree-like and hierarchical structures. Despite the …
P-mamba: Marrying perona malik diffusion with mamba for efficient pediatric echocardiographic left ventricular segmentation
In pediatric cardiology, the accurate and immediate assessment of cardiac function through
echocardiography is crucial since it can determine whether urgent intervention is required in …
echocardiography is crucial since it can determine whether urgent intervention is required in …
A novel state space model with local enhancement and state sharing for image fusion
In image fusion tasks, images from different sources possess distinct characteristics. This
has driven the development of numerous methods to explore better ways of fusing them …
has driven the development of numerous methods to explore better ways of fusing them …
DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution
Abstract Multimodal Large Language Models (MLLMs) have demonstrated remarkable
comprehension and reasoning capabilities with complex language and visual data. These …
comprehension and reasoning capabilities with complex language and visual data. These …