Causal discovery from temporal data: An overview and new perspectives

C Gong, C Zhang, D Yao, J Bi, W Li, YJ Xu - ACM Computing Surveys, 2024 - dl.acm.org
Temporal data, representing chronological observations of complex systems, has always
been a typical data structure that can be widely generated by many domains, such as …

Interventional causal representation learning

K Ahuja, D Mahajan, Y Wang… - … conference on machine …, 2023 - proceedings.mlr.press
Causal representation learning seeks to extract high-level latent factors from low-level
sensory data. Most existing methods rely on observational data and structural assumptions …

Weakly supervised causal representation learning

J Brehmer, P De Haan, P Lippe… - Advances in Neural …, 2022 - proceedings.neurips.cc
Learning high-level causal representations together with a causal model from unstructured
low-level data such as pixels is impossible from observational data alone. We prove under …

Nonparametric identifiability of causal representations from unknown interventions

J von Kügelgen, M Besserve… - Advances in …, 2024 - proceedings.neurips.cc
We study causal representation learning, the task of inferring latent causal variables and
their causal relations from high-dimensional functions (“mixtures”) of the variables. Prior …

Citris: Causal identifiability from temporal intervened sequences

P Lippe, S Magliacane, S Löwe… - International …, 2022 - proceedings.mlr.press
Understanding the latent causal factors of a dynamical system from visual observations is
considered a crucial step towards agents reasoning in complex environments. In this paper …

Causal component analysis

L Wendong, A Kekić, J von Kügelgen… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Independent Component Analysis (ICA) aims to recover independent latent
variables from observed mixtures thereof. Causal Representation Learning (CRL) aims …

Weakly supervised representation learning with sparse perturbations

K Ahuja, JS Hartford, Y Bengio - Advances in Neural …, 2022 - proceedings.neurips.cc
The theory of representation learning aims to build methods that provably invert the data
generating process with minimal domain knowledge or any source of supervision. Most prior …

Temporally disentangled representation learning

W Yao, G Chen, K Zhang - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Recently in the field of unsupervised representation learning, strong identifiability results for
disentanglement of causally-related latent variables have been established by exploiting …

Synergies between disentanglement and sparsity: Generalization and identifiability in multi-task learning

S Lachapelle, T Deleu, D Mahajan… - International …, 2023 - proceedings.mlr.press
Although disentangled representations are often said to be beneficial for downstream tasks,
current empirical and theoretical understanding is limited. In this work, we provide evidence …

Subspace identification for multi-source domain adaptation

Z Li, R Cai, G Chen, B Sun, Z Hao… - Advances in Neural …, 2024 - proceedings.neurips.cc
Multi-source domain adaptation (MSDA) methods aim to transfer knowledge from multiple
labeled source domains to an unlabeled target domain. Although current methods achieve …