Representational strengths and limitations of transformers

C Sanford, DJ Hsu, M Telgarsky - Advances in Neural …, 2023 - proceedings.neurips.cc
Attention layers, as commonly used in transformers, form the backbone of modern deep
learning, yet there is no mathematical description of their benefits and deficiencies as …

On the connection between mpnn and graph transformer

C Cai, TS Hy, R Yu, Y Wang - International conference on …, 2023 - proceedings.mlr.press
Graph Transformer (GT) recently has emerged as a new paradigm of graph learning
algorithms, outperforming the previously popular Message Passing Neural Network (MPNN) …

Advances in Set Function Learning: A Survey of Techniques and Applications

J **e, G Tong - ACM Computing Surveys, 2025 - dl.acm.org
Set function learning has emerged as a crucial area in machine learning, addressing the
challenge of modeling functions that take sets as inputs. Unlike traditional machine learning …

Universal representation of permutation-invariant functions on vectors and tensors

P Tabaghi, Y Wang - International Conference on …, 2024 - proceedings.mlr.press
A main object of our study is multiset functions—that is, permutation-invariant functions over
inputs of varying sizes. Deep Sets, proposed by Zaheer et al.(2017), provides a universal …

Polynomial width is sufficient for set representation with high-dimensional features

P Wang, S Yang, S Li, Z Wang, P Li - arxiv preprint arxiv:2307.04001, 2023 - arxiv.org
Set representation has become ubiquitous in deep learning for modeling the inductive bias
of neural networks that are insensitive to the input order. DeepSets is the most widely used …

Symmetric single index learning

A Zweig, J Bruna - arxiv preprint arxiv:2310.02117, 2023 - arxiv.org
Few neural architectures lend themselves to provable learning with gradient based
methods. One popular model is the single-index model, in which labels are produced by …

Towards antisymmetric neural ansatz separation

A Zweig, J Bruna - arxiv preprint arxiv:2208.03264, 2022 - arxiv.org
We study separations between two fundamental models (or\emph {Ans\" atze}) of
antisymmetric functions, that is, functions $ f $ of the form $ f (x_ {\sigma (1)},\ldots, x …

[หนังสือ][B] Local-to-global perspectives on graph neural networks

C Cai - 2023 - search.proquest.com
Abstract Message Passing Neural Networks (MPNN) has been the leading architecture for
machine learning on graphs. Its theoretical study focuses on increasing expressive power …

[หนังสือ][B] Theory of Symmetric Neural Networks

A Zweig - 2024 - search.proquest.com
Symmetric functions, which take as input an unordered, fixed-size set, find practical
application in myriad physical settings based on indistinguishable points or particles, and …

[หนังสือ][B] Representational Capabilities of Feed-Forward and Sequential Neural Architectures

CH Sanford - 2024 - search.proquest.com
Despite the widespread empirical success of deep neural networks over the past decade, a
comprehensive understanding of their mathematical properties remains elusive, which limits …