TASED-net: Temporally-aggregating spatial encoder-decoder network for video saliency detection K Min, JJ Corso Proceedings of the IEEE International Conference on Computer Vision (ICCV …, 2019 | 199 | 2019 |
Adversarial Background-Aware Loss for Weakly-supervised Temporal Activity Localization K Min, JJ Corso Proceedings of the European Conference on Computer Vision (ECCV), 2020 | 117 | 2020 |
Hierarchical novelty detection for visual object recognition K Lee, K Lee, K Min, Y Zhang, J Shin, H Lee Proceedings of the IEEE Conference on Computer Vision and Pattern …, 2018 | 90 | 2018 |
Integrating Human Gaze into Attention for Egocentric Activity Recognition K Min, JJ Corso Proceedings of the IEEE Winter Conference on Applications of Computer Vision …, 2020 | 59 | 2020 |
Learning long-term spatial-temporal graphs for active speaker detection K Min, S Roy, S Tripathi, T Guha, S Majumdar European Conference on Computer Vision, 371-387, 2022 | 34* | 2022 |
WOUAF: Weight Modulation for User Attribution and Fingerprinting in Text-to-Image Diffusion Models C Kim*, K Min*, M Patel, S Cheng, Y Yang arXiv preprint arXiv:2306.04744, 2023 | 29 | 2023 |
Unbiased scene graph generation in videos S Nag, K Min, S Tripathi, AK Roy-Chowdhury Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023 | 27 | 2023 |
Sourya Roy, Subarna Tripathi, Tanaya Guha, and Somdeb Majumdar. Learning long-term spatialtemporal graphs for active speaker detection K Min arXiv preprint arXiv:2207.07783 2 (3), 2022 | 21 | 2022 |
Svitt: Temporal learning of sparse video-text transformers Y Li, K Min, S Tripathi, N Vasconcelos Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023 | 15 | 2023 |
RACE: Robust Adversarial Concept Erasure for Secure Text-to-Image Diffusion Model C Kim*, K Min*, Y Yang arXiv preprint arXiv:2405.16341, 2024 | 10 | 2024 |
Intel Labs at Ego4D Challenge 2022: A Better Baseline for Audio-Visual Diarization K Min 2nd International Ego4D Workshop @ ECCV 2022, 2022 | 10 | 2022 |
Action Scene Graphs for Long-Form Understanding of Egocentric Videos I Rodin*, A Furnari*, K Min*, S Tripathi, GM Farinella arXiv preprint arXiv:2312.03391, 2023 | 7 | 2023 |
STHG: Spatial-Temporal Heterogeneous Graph Learning for Advanced Audio-Visual Diarization K Min 3rd International Ego4D Workshop @ CVPR 2023, 2023 | 6 | 2023 |
Contrastive Language Video Time Pre-training H Liu, K Min, HA Valdez, S Tripathi arXiv preprint arXiv:2406.02631, 2024 | 2 | 2024 |
Graph-Based Multimodal and Multi-view Alignment for Keystep Recognition JL Romero, K Min, S Tripathi, M Karimzadeh arXiv preprint arXiv:2501.04121, 2025 | | 2025 |
Deep Geometric Moments Promote Shape Consistency in Text-to-3D Generation U Nath, R Goel, ES Jeon, C Kim, K Min, Y Yang, Y Yang, P Turaga arXiv preprint arXiv:2408.05938, 2024 | | 2024 |
Ego-VPA: Egocentric Video Understanding with Parameter-efficient Adaptation TY Wu, K Min, S Tripathi, N Vasconcelos arXiv preprint arXiv:2407.19520, 2024 | | 2024 |
SViTT-Ego: A Sparse Video-Text Transformer for Egocentric Video HA Valdez, K Min, S Tripathi arXiv preprint arXiv:2406.09462, 2024 | | 2024 |
Long duration structured video action segmentation AD Rhodes, K Min, S Tripathi, G Raffa, S Biswas US Patent App. 18/459,824, 2024 | | 2024 |
Intel Labs at ActivityNet Challenge 2022: SPELL for Long-Term Active Speaker Detection K Min, S Roy, S Tripathi, T Guha, S Majumdar International Challenge on Activity Recognition (ActivityNet), 2022 | | 2022 |