Instance-conditional timescales of decay for non-stationary learning

N Jain, P Shenoy - Proceedings of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
Slow concept drift is a ubiquitous, yet under-studied problem in practical machine learning
systems. Although recent data is more indicative of future data in these settings, naively …

Multitask learning via interleaving: A neural network investigation

D Mayo, TR Scott, M Ren, G Elsayed… - Proceedings of the …, 2023 - escholarship.org
The most common settings in machine learning to study multitask learning assume either
that a random task is selected on each training trial, or that one task is trained to mastery and …

LoKO: Low-Rank Kalman Optimizer for Online Fine-Tuning of Large Models

H Abdi, M Sun, A Zhang, S Kaski, W Pan - arxiv preprint arxiv:2410.11551, 2024 - arxiv.org
Training large models with millions or even billions of parameters from scratch incurs
substantial computational costs. Parameter Efficient Fine-Tuning (PEFT) methods …

[PDF][PDF] Neural network online training with sensitivity to multiscale temporal structure

M Jones, D Mayo, T Scott, M Ren… - NeurIPS workshop on …, 2022 - 128.138.210.16
Many online-learning domains in artificial intelligence involve data with nonstationarities
spanning a wide range of timescales. Heuristic approaches to combat nonstationarity …

Hidden Markov Neural Networks

L Rimella, N Whiteley - Entropy, 2025 - mdpi.com
We define an evolving in-time Bayesian neural network called a Hidden Markov Neural
Network, which addresses the crucial challenge in time-series forecasting and continual …