Підписатись
Druv Pai
Druv Pai
PhD Student, UC Berkeley
Підтверджена електронна адреса в berkeley.edu - Домашня сторінка
Назва
Посилання
Посилання
Рік
White-box transformers via sparse rate reduction
Y Yu, S Buchanan, D Pai, T Chu, Z Wu, S Tong, B Haeffele, Y Ma
Advances in Neural Information Processing Systems 36, 9422-9457, 2023
80*2023
Emergence of segmentation with minimalistic white-box transformers
Y Yu, T Chu, S Tong, Z Wu, D Pai, S Buchanan, Y Ma
arXiv preprint arXiv:2308.16271, 2023
212023
Independent and decentralized learning in markov potential games
C Maheshwari, M Wu, D Pai, S Sastry
arXiv preprint arXiv:2205.14590, 2022
212022
Masked completion via structured diffusion with white-box transformers
D Pai, ZW Wu, S Buchanan, Y Yu, Y Ma
International Conference on Learning Representations, 2023
102023
Pursuit of a discriminative representation for multiple subspaces via sequential games
D Pai, M Psenka, CY Chiu, M Wu, E Dobriban, Y Ma
arXiv preprint arXiv:2206.09120, 2022
92022
Representation learning via manifold flattening and reconstruction
M Psenka, D Pai, V Raman, S Sastry, Y Ma
Journal of Machine Learning Research 25 (132), 1-47, 2024
72024
Closed-loop transcription via convolutional sparse coding
X Dai, K Chen, S Tong, J Zhang, X Gao, M Li, D Pai, Y Zhai, XI Yuan, ...
arXiv preprint arXiv:2302.09347, 2023
72023
Scaling white-box transformers for vision
J Yang, X Li, D Pai, Y Zhou, Y Ma, Y Yu, C Xie
Advances in Neural Information Processing Systems 37, 36995-37019, 2025
52025
Congestion pricing for efficiency and equity: Theory and applications to the san francisco bay area
C Maheshwari, K Kulkarni, D Pai, J Yang, M Wu, S Sastry
arXiv preprint arXiv:2401.16844, 2024
52024
A global geometric analysis of maximal coding rate reduction
P Wang, H Liu, D Pai, Y Yu, Z Zhu, Q Qu, Y Ma
arXiv preprint arXiv:2406.01909, 2024
42024
Active-dormant attention heads: Mechanistically demystifying extreme-token phenomena in llms
T Guo, D Pai, Y Bai, J Jiao, MI Jordan, S Mei
arXiv preprint arXiv:2410.13835, 2024
32024
Simplifying DINO via Coding Rate Regularization
Z Wu, J Zhang, D Pai, XD Wang, C Singh, J Yang, J Gao, Y Ma
arXiv preprint arXiv:2502.10385, 2025
2025
Token Statistics Transformer: Linear-Time Attention via Variational Rate Reduction
Z Wu, T Ding, Y Lu, D Pai, J Zhang, W Wang, Y Yu, Y Ma, BD Haeffele
arXiv preprint arXiv:2412.17810, 2024
2024
Learning Low-Dimensional Structure via Closed-Loop Transcription: Equilibria and Optimization
D Pai
2023
Attention-Only Transformers via Unrolled Subspace Denoising
P Wang, Y Lu, Y Yu, D Pai, Q Qu, Y Ma
Scaling White-Box Transformers for Vision-Appendix
J Yang, X Li, D Pai, C Xie11
У даний момент система не може виконати операцію. Спробуйте пізніше.
Статті 1–16