When Gaussian process meets big data: A review of scalable GPs

H Liu, YS Ong, X Shen, J Cai - IEEE transactions on neural …, 2020 - ieeexplore.ieee.org
The vast quantity of information brought by big data as well as the evolving computer
hardware encourages success stories in the machine learning community. In the …

Recent advances in data-driven wireless communication using Gaussian processes: A comprehensive survey

K Chen, Q Kong, Y Dai, Y Xu, F Yin, L Xu… - China …, 2022 - ieeexplore.ieee.org
Data-driven paradigms are well-known and salient demands of future wireless
communication. Empowered by big data and machine learning techniques, next-generation …

Cocktailsgd: Fine-tuning foundation models over 500mbps networks

J Wang, Y Lu, B Yuan, B Chen… - International …, 2023 - proceedings.mlr.press
Distributed training of foundation models, especially large language models (LLMs), is
communication-intensive and so has heavily relied on centralized data centers with fast …

Kernel methods through the roof: handling billions of points efficiently

G Meanti, L Carratino, L Rosasco… - Advances in Neural …, 2020 - proceedings.neurips.cc
Kernel methods provide an elegant and principled approach to nonparametric learning, but
so far could hardly be used in large scale problems, since naïve implementations scale …

Generalized robust Bayesian committee machine for large-scale Gaussian process regression

H Liu, J Cai, Y Wang, YS Ong - International Conference on …, 2018 - proceedings.mlr.press
In order to scale standard Gaussian process (GP) regression to large-scale datasets,
aggregation models employ factorized training process and then combine predictions from …

Distributed learning systems with first-order methods

J Liu, C Zhang - Foundations and Trends® in Databases, 2020 - nowpublishers.com
Scalable and efficient distributed learning is one of the main driving forces behind the recent
rapid advancement of machine learning and artificial intelligence. One prominent feature of …

Bagua: scaling up distributed learning with system relaxations

S Gan, X Lian, R Wang, J Chang, C Liu, H Shi… - arxiv preprint arxiv …, 2021 - arxiv.org
Recent years have witnessed a growing list of systems for distributed data-parallel training.
Existing systems largely fit into two paradigms, ie, parameter server and MPI-style collective …

Asynchronous parallel large-scale Gaussian process regression

Z Dang, B Gu, C Deng, H Huang - IEEE Transactions on Neural …, 2024 - ieeexplore.ieee.org
Gaussian process regression (GPR) is an important nonparametric learning method in
machine learning research with many real-world applications. It is well known that training …

Exact gaussian process regression with distributed computations

DT Nguyen, M Filippone, P Michiardi - Proceedings of the 34th ACM …, 2019 - dl.acm.org
Gaussian Processes (GPs) are powerful non-parametric Bayesian models for function
estimation, but suffer from high complexity in terms of both computation and storage. To …

[PDF][PDF] Persia: a hybrid system scaling deep learning based recommenders up to 100 trillion parameters

X Lian, B Yuan, X Zhu, Y Wang, Y He… - arxiv preprint arxiv …, 2021 - ask.qcloudimg.com
Deep learning based models have dominated the current landscape of production
recommender systems. Furthermore, recent years have witnessed an exponential growth of …