Randomness in neural networks: an overview

S Scardapane, D Wang - Wiley Interdisciplinary Reviews: Data …, 2017 - Wiley Online Library
Neural networks, as powerful tools for data mining and knowledge engineering, can learn
from data to build feature‐based classifiers and nonlinear predictive models. Training neural …

Random features for kernel approximation: A survey on algorithms, theory, and beyond

F Liu, X Huang, Y Chen… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
The class of random features is one of the most popular techniques to speed up kernel
methods in large-scale problems. Related works have been recognized by the NeurIPS Test …

Unsupervised deep generative adversarial hashing network

KG Dizaji, F Zheng, N Sadoughi… - Proceedings of the …, 2018 - openaccess.thecvf.com
Unsupervised deep hash functions have not shown satisfactory improvements against the
shallow alternatives, and usually, require supervised pretraining to avoid getting stuck in …

Kernel continual learning

MM Derakhshani, X Zhen, L Shao… - … on Machine Learning, 2021 - proceedings.mlr.press
This paper introduces kernel continual learning, a simple but effective variant of continual
learning that leverages the non-parametric nature of kernel methods to tackle catastrophic …

Direct shape regression networks for end-to-end face alignment

X Miao, X Zhen, X Liu, C Deng… - Proceedings of the …, 2018 - openaccess.thecvf.com
Face alignment has been extensively studied in computer vision community due to its
fundamental role in facial analysis, but it remains an unsolved problem. The major …

Non-stationary spectral kernels

S Remes, M Heinonen, S Kaski - Advances in neural …, 2017 - proceedings.neurips.cc
We propose non-stationary spectral kernels for Gaussian process regression by modelling
the spectral density of a non-stationary kernel function as a mixture of input-dependent …

Coded machine unlearning

N Aldaghri, H Mahdavifar, A Beirami - IEEE Access, 2021 - ieeexplore.ieee.org
There are applications that may require removing the trace of a sample from the system, eg,
a user requests their data to be deleted, or corrupted data is discovered. Simply removing a …

Sketching for large-scale learning of mixture models

N Keriven, A Bourrier, R Gribonval… - … and Inference: A …, 2018 - academic.oup.com
Learning parameters from voluminous data can be prohibitive in terms of memory and
computational requirements. We propose a 'compressive learning'framework, where we …

Recasting self-attention with holographic reduced representations

MM Alam, E Raff, S Biderman… - … on Machine Learning, 2023 - proceedings.mlr.press
In recent years, self-attention has become the dominant paradigm for sequence modeling in
a variety of domains. However, in domains with very long sequence lengths the $\mathcal …

Implicit kernel learning

CL Li, WC Chang, Y Mroueh, Y Yang… - The 22nd …, 2019 - proceedings.mlr.press
Kernels are powerful and versatile tools in machine learning and statistics. Although the
notion of universal kernels and characteristic kernels has been studied, kernel selection still …