Deep learning for plant stress phenoty**: trends and future perspectives

AK Singh, B Ganapathysubramanian, S Sarkar… - Trends in plant …, 2018 - cell.com
Deep learning (DL), a subset of machine learning approaches, has emerged as a versatile
tool to assimilate large amounts of heterogeneous data and provide reliable predictions of …

Dags with no tears: Continuous optimization for structure learning

X Zheng, B Aragam, PK Ravikumar… - Advances in neural …, 2018 - proceedings.neurips.cc
Estimating the structure of directed acyclic graphs (DAGs, also known as Bayesian
networks) is a challenging problem since the search space of DAGs is combinatorial and …

You only propagate once: Accelerating adversarial training via maximal principle

D Zhang, T Zhang, Y Lu, Z Zhu… - Advances in neural …, 2019 - proceedings.neurips.cc
Deep learning achieves state-of-the-art results in many tasks in computer vision and natural
language processing. However, recent works have shown that deep networks can be …

Advancements in federated learning: Models, methods, and privacy

H Chen, H Wang, Q Long, D **, Y Li - ACM Computing Surveys, 2024 - dl.acm.org
Federated learning (FL) is a promising technique for resolving the rising privacy and security
concerns. Its main ingredient is to cooperatively learn the model among the distributed …

Which neural net architectures give rise to exploding and vanishing gradients?

B Hanin - Advances in neural information processing …, 2018 - proceedings.neurips.cc
We give a rigorous analysis of the statistical behavior of gradients in a randomly initialized
fully connected network N with ReLU activations. Our results show that the empirical …

Cooperative-game-based day-ahead scheduling of local integrated energy systems with shared energy storage

C Chen, Y Li, W Qiu, C Liu, Q Zhang… - IEEE Transactions …, 2022 - ieeexplore.ieee.org
In the context of the current sharing economy, the application of shared energy storage
(SES) among local integrated energy systems (LIESs) is underexplored. There is an urgent …

Decoupled neural interfaces using synthetic gradients

M Jaderberg, WM Czarnecki… - International …, 2017 - proceedings.mlr.press
Training directed neural networks typically requires forward-propagating data through a
computation graph, followed by backpropagating error signal, to produce weight updates. All …

A theory-guided deep-learning formulation and optimization of seismic waveform inversion

J Sun, Z Niu, KA Innanen, J Li, DO Trad - Geophysics, 2020 - library.seg.org
Deep-learning techniques appear to be poised to play very important roles in our processing
flows for inversion and interpretation of seismic data. The most successful seismic …

Offsite-tuning: Transfer learning without full model

G **ao, J Lin, S Han - arxiv preprint arxiv:2302.04870, 2023 - arxiv.org
Transfer learning is important for foundation models to adapt to downstream tasks. However,
many foundation models are proprietary, so users must share their data with model owners …

Implicit deep learning

L El Ghaoui, F Gu, B Travacca, A Askari, A Tsai - SIAM Journal on …, 2021 - SIAM
Implicit deep learning prediction rules generalize the recursive rules of feedforward neural
networks. Such rules are based on the solution of a fixed-point equation involving a single …