Overview of the 8th workshop on Asian translation
This paper presents the results of the shared tasks from the 8th workshop on Asian
translation (WAT2021). For the WAT2021, 28 teams participated in the shared tasks and 24 …
translation (WAT2021). For the WAT2021, 28 teams participated in the shared tasks and 24 …
Exploring better text image translation with multimodal codebook
Text image translation (TIT) aims to translate the source texts embedded in the image to
target translations, which has a wide range of applications and thus has important research …
target translations, which has a wide range of applications and thus has important research …
Video-Helpful Multimodal Machine Translation
Existing multimodal machine translation (MMT) datasets consist of images and video
captions or instructional video subtitles, which rarely contain linguistic ambiguity, making …
captions or instructional video subtitles, which rarely contain linguistic ambiguity, making …
Bigvideo: A large-scale video subtitle translation dataset for multimodal machine translation
We present a large-scale video subtitle translation dataset, BigVideo, to facilitate the study of
multi-modality machine translation. Compared with the widely used How2 and VaTeX …
multi-modality machine translation. Compared with the widely used How2 and VaTeX …
Vision talks: Visual relationship-enhanced transformer for video-guided machine translation
S Chen, Y Zeng, D Cao, S Lu - Expert Systems with Applications, 2022 - Elsevier
Video-guided machine translation is a promising task which aims to translate a source
language description into a target language utilizing the video information as supplementary …
language description into a target language utilizing the video information as supplementary …
Visa: An ambiguous subtitles dataset for visual scene-aware machine translation
Existing multimodal machine translation (MMT) datasets consist of images and video
captions or general subtitles, which rarely contain linguistic ambiguity, making visual …
captions or general subtitles, which rarely contain linguistic ambiguity, making visual …
TriFine: A Large-Scale Dataset of Vision-Audio-Subtitle for Tri-Modal Machine Translation and Benchmark with Fine-Grained Annotated Tags
Current video-guided machine translation (VMT) approaches primarily use coarse-grained
visual information, resulting in information redundancy, high computational overhead, and …
visual information, resulting in information redundancy, high computational overhead, and …
Overview of the 10th Workshop on Asian Translation
This paper presents the results of the shared tasks from the 10th workshop on Asian
translation (WAT2023). For the WAT2023, 2 teams submitted their translation results for the …
translation (WAT2023). For the WAT2023, 2 teams submitted their translation results for the …
A Survey on Multi-modal Machine Translation: Tasks, Methods and Challenges
In recent years, multi-modal machine translation has attracted significant interest in both
academia and industry due to its superior performance. It takes both textual and visual …
academia and industry due to its superior performance. It takes both textual and visual …
[PDF][PDF] Studies on Subword-based Low-Resource Neural Machine Translation: Segmentation, Encoding, and Decoding
S Haiyue - 2024 - repository.kulib.kyoto-u.ac.jp
In a world rich with diverse ideas and cultures, humans are isolated into islands of distinct
languages. Machine translation (MT) serves as a bridge, facilitating information access and …
languages. Machine translation (MT) serves as a bridge, facilitating information access and …