Volgen
Shengbang Tong
Shengbang Tong
NYU Courant
Geverifieerd e-mailadres voor berkeley.edu - Homepage
Titel
Geciteerd door
Geciteerd door
Jaar
Eyes wide shut? exploring the visual shortcomings of multimodal llms
S Tong, Z Liu, Y Zhai, Y Ma, Y LeCun, S Xie
CVPR 2024 (Oral), 2024
2462024
Cambrian-1: A Fully Open, Vision-Centric Exploration of Multimodal LLMs
S Tong*, E Brown*, P Wu*, S Woo, M Middepogu, SC Akula, J Yang, ...
NeurIPS 2024 (Oral), 2024
2042024
Investigating the catastrophic forgetting in multimodal large language models
Y Zhai, S Tong, X Li, M Cai, Q Qu, YJ Lee, Y Ma
CPAL 2024, 2023
134*2023
White-box transformers via sparse rate reduction
Y Yu, S Buchanan, D Pai, T Chu, Z Wu, S Tong, B Haeffele, Y Ma
NeurIPS 2023, 2023
752023
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning
Y Zhai, H Bai*, Z Lin*, J Pan*, S Tong*, Y Zhou*, A Suhr, S Xie, Y LeCun, ...
NeurIPS 2024, 2024
402024
Mass-producing failures of multimodal systems with language models
S Tong, E Jones, J Steinhardt
NeurIPS 2023, 2023
362023
Ctrl: Closed-loop transcription to an ldr via minimaxing rate reduction
X Dai*, S Tong*, M Li*, Z Wu*, M Psenka, KHR Chan, P Zhai, Y Yu, ...
Entropy 24 (4), 456, 2022
36*2022
MMMU-Pro: A More Robust Multi-discipline Multimodal Understanding Benchmark
X Yue, T Zheng, Y Ni, Y Wang, K Zhang, S Tong, Y Sun, M Yin, B Yu, ...
arXiv preprint arXiv:2409.02813, 2024
332024
Emp-ssl: Towards self-supervised learning in one training epoch
S Tong*, Y Chen*, Y Ma, Y LeCun
arXiv preprint arXiv:2304.03977, 2023
302023
Revisiting sparse convolutional model for visual recognition
X Dai, M Li, P Zhai, S Tong, X Gao, SL Huang, Z Zhu, C You, Y Ma
NeurIPS 2022, 2022
292022
Incremental learning of structured memory via closed-loop transcription
S Tong, X Dai, Z Wu, M Li, B Yi, Y Ma
ICLR 2023, 2022
262022
Emergence of segmentation with minimalistic white-box transformers
Y Yu*, T Chu*, S Tong, Z Wu, D Pai, S Buchanan, Y Ma
CPAL 2024, 2023
212023
Image clustering via the principle of rate reduction in the age of pretrained models
T Chu*, S Tong*, T Ding*, X Dai, B Haeffele, R Vidal, Y Ma
ICLR 2024, 2024
122024
Unsupervised manifold linearizing and clustering
T Ding, S Tong, KHR Chan, X Dai, Y Ma, BD Haeffele
ICCV 2023, 2023
122023
White-Box Transformers via Sparse Rate Reduction: Compression Is All There Is?
Y Yu, S Buchanan, D Pai, T Chu, Z Wu, S Tong, H Bai, Y Zhai, ...
JMLR, 2023
11*2023
Metamorph: Multimodal understanding and generation via instruction tuning
S Tong, D Fan, J Zhu, Y Xiong, X Chen, K Sinha, M Rabbat, Y LeCun, ...
arXiv preprint arXiv:2412.14164, 2024
7*2024
Closed-loop transcription via convolutional sparse coding
X Dai, K Chen, S Tong, J Zhang, X Gao, M Li, D Pai, Y Zhai, XI Yuan, ...
CPAL 2024, 2023
72023
Unsupervised learning of structured representations via closed-loop transcription
S Tong*, X Dai*, Y Chen, M Li, Z Li, B Yi, Y LeCun, Y Ma
CPAL 2024, 2022
62022
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training
T Chu*, Y Zhai*, J Yang, S Tong, S Xie, D Schuurmans, QV Le, S Levine, ...
arXiv preprint arXiv:2501.17161, 2025
32025
Connecting Joint-Embedding Predictive Architecture with Contrastive Self-supervised Learning
S Mo, S Tong
NeurIPS 2024 (Spotlight), 2024
2024
Het systeem kan de bewerking nu niet uitvoeren. Probeer het later opnieuw.
Artikelen 1–20