The evolution of distributed systems for graph neural networks and their origin in graph processing and deep learning: A survey
Graph neural networks (GNNs) are an emerging research field. This specialized deep
neural network architecture is capable of processing graph structured data and bridges the …
neural network architecture is capable of processing graph structured data and bridges the …
A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
Foundational challenges in assuring alignment and safety of large language models
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …
language models (LLMs). These challenges are organized into three different categories …
Dataset distillation via factorization
In this paper, we study dataset distillation (DD), from a novel perspective and introduce
a\emph {dataset factorization} approach, termed\emph {HaBa}, which is a plug-and-play …
a\emph {dataset factorization} approach, termed\emph {HaBa}, which is a plug-and-play …
Robust fine-tuning of zero-shot models
Large pre-trained models such as CLIP or ALIGN offer consistent accuracy across a range of
data distributions when performing zero-shot inference (ie, without fine-tuning on a specific …
data distributions when performing zero-shot inference (ie, without fine-tuning on a specific …
Towards out-of-distribution generalization: A survey
Traditional machine learning paradigms are based on the assumption that both training and
test data follow the same statistical pattern, which is mathematically referred to as …
test data follow the same statistical pattern, which is mathematically referred to as …
Dataset distillation: A comprehensive review
Recent success of deep learning is largely attributed to the sheer amount of data used for
training deep neural networks. Despite the unprecedented success, the massive data …
training deep neural networks. Despite the unprecedented success, the massive data …
Neural fields in visual computing and beyond
Recent advances in machine learning have led to increased interest in solving visual
computing problems using methods that employ coordinate‐based neural networks. These …
computing problems using methods that employ coordinate‐based neural networks. These …
Physics-informed machine learning
Despite great progress in simulating multiphysics problems using the numerical
discretization of partial differential equations (PDEs), one still cannot seamlessly incorporate …
discretization of partial differential equations (PDEs), one still cannot seamlessly incorporate …
Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields
Neural Radiance Fields (NeRF) are able to reconstruct scenes with unprecedented fidelity,
and various recent works have extended NeRF to handle dynamic scenes. A common …
and various recent works have extended NeRF to handle dynamic scenes. A common …