Model zoos: A dataset of diverse populations of neural network models

K Schürholt, D Taskiran, B Knyazev… - Advances in …, 2022 - proceedings.neurips.cc
In the last years, neural networks (NN) have evolved from laboratory environments to the
state-of-the-art for many real-world problems. It was shown that NN models (ie, their weights …

NeRN--Learning Neural Representations for Neural Networks

M Ashkenazi, Z Rimon, R Vainshtein, S Levi… - arxiv preprint arxiv …, 2022 - arxiv.org
Neural Representations have recently been shown to effectively reconstruct a wide range of
signals from 3D meshes and shapes to images and videos. We show that, when adapted …

Towards scalable and versatile weight space learning

K Schürholt, MW Mahoney, D Borth - arxiv preprint arxiv:2406.09997, 2024 - arxiv.org
Learning representations of well-trained neural network models holds the promise to
provide an understanding of the inner workings of those models. However, previous work …

Adaptive Fine-Tuning in Degradation-Time-Series Forecasting via Generating Source Domain

J Pan, B **, S Wang, X Yuwen, X Jiao - IEEE Access, 2023 - ieeexplore.ieee.org
Parameter-Efficient Fine-Tuning is widely used to transfer models between different
domains. However, for some high-reliability-equipment, the degradation is at a slow rate and …

Sparsified Model Zoo Twins: Investigating Populations of Sparsified Neural Network Models

D Honegger, K Schürholt, D Borth - arxiv preprint arxiv:2304.13718, 2023 - arxiv.org
With growing size of Neural Networks (NNs), model sparsification to reduce the
computational cost and memory demand for model inference has become of vital interest for …

Recurrent Diffusion for Large-Scale Parameter Generation

K Wang, D Tang, W Zhao, Y You - arxiv preprint arxiv:2501.11587, 2025 - arxiv.org
Parameter generation has struggled to scale up for a long time, significantly limiting its range
of applications. In this study, we introduce\textbf {R} ecurrent diffusion for large-scale\textbf …

Hyper-Representations: Learning from Populations of Neural Networks

K Schürholt - arxiv preprint arxiv:2410.05107, 2024 - arxiv.org
This thesis addresses the challenge of understanding Neural Networks through the lens of
their most fundamental component: the weights, which encapsulate the learned information …

Eurosat Model Zoo: A Dataset and Benchmark on Populations of Neural Networks and Its Sparsified Model Twins

D Honegger, K Schürholt… - IGARSS 2023-2023 …, 2023 - ieeexplore.ieee.org
The availability of large-scale labeled datasets in remote sensing and Earth observation
accelerated the use of deep neural networks in this domain. In the standard workflow, data is …

[PDF][PDF] Pre-training Meta-models for Interpretability

E Dordevic - mlmi.eng.cam.ac.uk
Mechanistic interpretability is a field that aims to explain the behaviour of trained neural
networks by studying their learnt parameters. As most of this line of work is laborious and …

3.2 Self-supervised Learning, Foundation Models, and ModelZoos

D Borth - Space and Artificial Intelligence - scholar.archive.org
Self-supervised learning allowed us to train large task agnostic backbones, which can be
successfully finetuned for specialized downstream tasks with only little supervision. This …