팔로우
Noam Brown
Noam Brown
Research Scientist, OpenAI
cs.cmu.edu의 이메일 확인됨 - 홈페이지
제목
인용
인용
연도
Superhuman AI for multiplayer poker
N Brown, T Sandholm
Science 365 (6456), 885-890, 2019
10112019
Superhuman AI for heads-up no-limit poker: Libratus beats top professionals
N Brown, T Sandholm
Science 359 (6374), 418-424, 2018
9602018
Deep counterfactual regret minimization
N Brown, A Lerer, S Gross, T Sandholm
International Conference on Machine Learning, 2019
2922019
Human-level play in the game of Diplomacy by combining language models with strategic reasoning
Meta Fundamental AI Research Diplomacy Team (FAIR)†, A Bakhtin, ...
Science 378 (6624), 1067-1074, 2022
2432022
Safe and nested subgame solving for imperfect-information games
N Brown, T Sandholm
Neural Information Processing Systems, 2017
236*2017
Solving imperfect-information games via discounted regret minimization
N Brown, T Sandholm
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 1829-1836, 2019
1892019
Combining deep reinforcement learning and search for imperfect-information games
N Brown, A Bakhtin, A Lerer, Q Gong
Advances in Neural Information Processing Systems 33, 17057-17069, 2020
1712020
Libratus: The Superhuman AI for No-Limit Poker.
N Brown, T Sandholm
IJCAI, 5226-5228, 2017
1552017
Depth-limited solving for imperfect-information games
N Brown, T Sandholm, B Amos
Advances in neural information processing systems 31, 2018
1012018
Improving Policies via Search in Cooperative Partially Observable Games
A Lerer, H Hu, J Foerster, N Brown
AAAI Conference on Artificial Intelligence, 2020
872020
Hierarchical abstraction, distributed equilibrium computation, and post-processing, with application to a champion no-limit Texas Hold'em agent
N Brown, S Ganzfried, T Sandholm
Workshops at the twenty-ninth AAAI conference on artificial intelligence, 2015
872015
Off-belief learning
H Hu, A Lerer, B Cui, L Pineda, N Brown, J Foerster
International Conference on Machine Learning, 4369-4379, 2021
702021
A unified approach to reinforcement learning, quantal response equilibria, and two-player zero-sum games
S Sokota, R D'Orazio, JZ Kolter, N Loizou, M Lanctot, I Mitliagkas, ...
arXiv preprint arXiv:2206.05825, 2022
612022
Modeling strong and human-like gameplay with KL-regularized search
AP Jacob, DJ Wu, G Farina, A Lerer, H Hu, A Bakhtin, J Andreas, N Brown
International Conference on Machine Learning, 9695-9728, 2022
592022
Dream: Deep regret minimization with advantage baselines and model-free learning
E Steinberger, A Lerer, N Brown
arXiv preprint arXiv:2006.10410, 2020
592020
Human-level performance in no-press diplomacy via equilibrium search
J Gray, A Lerer, A Bakhtin, N Brown
arXiv preprint arXiv:2010.02923, 2020
562020
Dynamic thresholding and pruning for regret minimization
N Brown, C Kroer, T Sandholm
Proceedings of the AAAI conference on artificial intelligence 31 (1), 2017
542017
No-press diplomacy from scratch
A Bakhtin, D Wu, A Lerer, N Brown
Advances in Neural Information Processing Systems 34, 18063-18074, 2021
462021
Mastering the game of no-press Diplomacy via human-regularized reinforcement learning and planning
A Bakhtin, DJ Wu, A Lerer, J Gray, AP Jacob, G Farina, AH Miller, ...
arXiv preprint arXiv:2210.05492, 2022
452022
Stable-predictive optimistic counterfactual regret minimization
G Farina, C Kroer, N Brown, T Sandholm
International conference on machine learning, 1853-1862, 2019
432019
현재 시스템이 작동되지 않습니다. 나중에 다시 시도해 주세요.
학술자료 1–20