Folgen
Zhenmei Shi
Zhenmei Shi
Research Scientist at Voyage AI; PhD from University of Wisconsin–Madison
Bestätigte E-Mail-Adresse bei cs.wisc.edu - Startseite
Titel
Zitiert von
Zitiert von
Jahr
SF-Net: Structured feature network for continuous sign language recognition
Z Yang*, Z Shi*, X Shen, YW Tai
arXiv preprint arXiv:1908.01341, 2019
832019
A Theoretical Analysis on Feature Learning in Neural Networks: Emergence from Inputs and Advantage over Fixed Features
Z Shi*, J Wei*, Y Liang
ICLR 2022: International Conference on Learning Representations, 2022
632022
Deep Online Fused Video Stabilization
Z Shi, F Shi, WS Lai, CK Liang, Y Liang
WACV 2022: Winter Conference on Applications of Computer Vision, 2022
362022
Do Large Language Models Have Compositional Ability? An Investigation into Limitations and Scalability
Z Xu*, Z Shi*, Y Liang
COLM 2024: Conference on Language Modeling, 2024
352024
The Trade-off between Universality and Label Efficiency of Representations from Contrastive Learning
Z Shi*, J Chen*, K Li, J Raghuram, X Wu, Y Liang, S Jha
ICLR 2023 (Spotlight): International Conference on Learning Representations, 2023
302023
Towards Few-Shot Adaptation of Foundation Models via Multitask Finetuning
Z Xu, Z Shi, J Wei, F Mu, Y Li, Y Liang
ICLR 2024: International Conference on Learning Representations, 2024
282024
Tensor Attention Training: Provably Efficient Learning of Higher-order Transformers
Y Liang*, Z Shi*, Z Song*, Y Zhou*
AFM Workshop @ NeurIPS 2024, 2024
272024
When and How Does Known Class Help Discover Unknown Ones? Provable Understandings Through Spectral Analysis
Y Sun, Z Shi, Y Liang, Y Li
ICML 2023: International Conference on Machine Learning, 2023
252023
Fourier Circuits in Neural Networks and Transformers: A Case Study of Modular Arithmetic with Multiple Inputs
C Li*, Y Liang*, Z Shi*, Z Song*, T Zhou*
AIStats 2025: International Conference on Artificial Intelligence and Statistics, 2025
24*2025
Multi-Layer Transformers Gradient Can be Approximated in Almost Linear Time
Y Liang*, Z Sha*, Z Shi*, Z Song*, Y Zhou*
OPT Workshop @ NeurIPS 2024, 2024
242024
Conv-basis: A new paradigm for efficient attention inference and gradient computation in transformers
Y Liang*, H Liu*, Z Shi*, Z Song*, Z Xu*, J Yin*
arXiv preprint arXiv:2405.05219, 2024
242024
Discovering the gems in early layers: Accelerating long-context llms with 1000x input token reduction
Z Shi, Y Ming, XP Nguyen, Y Liang, S Joty
arXiv preprint arXiv:2409.17422, 2024
232024
A Graph-Theoretic Framework for Understanding Open-World Semi-Supervised Learning
Y Sun, Z Shi, Y Li
NeurIPS 2023 (Spotlight): Neural Information Processing Systems, 2023
232023
Beyond Linear Approximations: A Novel Pruning Approach for Attention Matrix
Y Liang*, J Long*, Z Shi*, Z Song*, Y Zhou*
ICLR 2025: International Conference on Learning Representations, 2025
22*2025
A Tighter Complexity Analysis of SparseGPT
X Li*, Y Liang*, Z Shi*, Z Song*
Compression Workshop @ NeurIPS 2024, 2024
202024
HSR-Enhanced Sparse Attention Acceleration
B Chen*, Y Liang*, Z Sha*, Z Shi*, Z Song*
arXiv preprint arXiv:2410.10165, 2024
182024
Toward Infinite-Long Prefix in Transformer
Y Liang*, Z Shi*, Z Song*, C Yang*
arXiv preprint arXiv:2406.14036, 2024
182024
Domain generalization via nuclear norm regularization
Z Shi, Y Ming, Y Fan, F Sala, Y Liang
CPAL 2024: Conference on Parsimony and Learning, 179-201, 2024
182024
Looped ReLU MLPs May Be All You Need as Practical Programmable Computers
Y Liang*, Z Sha*, Z Shi*, Z Song*, Y Zhou*
AIStats 2025: International Conference on Artificial Intelligence and Statistics, 2025
172025
Differential Privacy of Cross-Attention with Provable Guarantee
Y Liang*, Z Shi*, Z Song*, Y Zhou*
SafeGenAi Workshop @ NeurIPS 2024, 2024
172024
Das System kann den Vorgang jetzt nicht ausführen. Versuchen Sie es später erneut.
Artikel 1–20