Momentdiff: Generative video moment retrieval from random to real

P Li, CW **e, H **e, L Zhao, L Zhang… - Advances in neural …, 2024 - proceedings.neurips.cc
Video moment retrieval pursues an efficient and generalized solution to identify the specific
temporal segments within an untrimmed video that correspond to a given language …

Temporal sentence grounding in videos: A survey and future directions

H Zhang, A Sun, W **g, JT Zhou - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Temporal sentence grounding in videos (TSGV), aka, natural language video localization
(NLVL) or video moment retrieval (VMR), aims to retrieve a temporal moment that …

Mad: A scalable dataset for language grounding in videos from movie audio descriptions

M Soldan, A Pardo, JL Alcázar… - Proceedings of the …, 2022 - openaccess.thecvf.com
The recent and increasing interest in video-language research has driven the development
of large-scale datasets that enable data-intensive machine learning techniques. In …

Knowing where to focus: Event-aware transformer for video grounding

J Jang, J Park, J Kim, H Kwon… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Recent DETR-based video grounding models have made the model directly predict moment
timestamps without any hand-crafted components, such as a pre-defined proposal or non …

Curriculum multi-negative augmentation for debiased video grounding

X Lan, Y Yuan, H Chen, X Wang, Z Jie, L Ma… - Proceedings of the …, 2023 - ojs.aaai.org
Video Grounding (VG) aims to locate the desired segment from a video given a sentence
query. Recent studies have found that current VG models are prone to over-rely the …

[PDF][PDF] The elements of temporal sentence grounding in videos: A survey and future directions

H Zhang, A Sun, W **g, JT Zhou - arxiv preprint arxiv …, 2022 - researchgate.net
Temporal sentence grounding in videos (TSGV), aka, natural language video localization
(NLVL) or video moment retrieval (VMR), aims to retrieve a temporal moment that …

Collaborative debias strategy for temporal sentence grounding in video

Z Qi, Y Yuan, X Ruan, S Wang… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Temporal sentence grounding in video has witnessed significant advancements, but suffers
from substantial dataset bias, which undermines its generalization ability. Existing debias …

[PDF][PDF] Overcoming weak visual-textual alignment for video moment retrieval

M Jung, Y Jang, S Choi, J Kim, JH Kim… - arxiv preprint arxiv …, 2023 - researchgate.net
Video moment retrieval (VMR) aims to identify the specific moment in an untrimmed video for
a given natural language query. However, this task is prone to suffer the weak visual-textual …

Self-supervised learning for semi-supervised temporal language grounding

F Luo, S Chen, J Chen, Z Wu… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Given a text description, Temporal Language Grounding (TLG) aims to localize temporal
boundaries of the segments that contain the specified semantics in an untrimmed video. TLG …

Transform-Equivariant Consistency Learning for Temporal Sentence Grounding

D Liu, X Qu, J Dong, P Zhou, Z Xu, H Wang… - ACM Transactions on …, 2024 - dl.acm.org
This paper addresses the temporal sentence grounding (TSG). Although existing methods
have made decent achievements in this task, they not only severely rely on abundant video …