Complex-valued neural networks: A comprehensive survey

CY Lee, H Hasegawa, S Gao - IEEE/CAA Journal of …, 2022 - ieeexplore.ieee.org
Complex-valued neural networks (CVNNs) have shown their excellent efficiency compared
to their real counter-parts in speech enhancement, image and signal processing …

Normalization techniques in training dnns: Methodology, analysis and application

L Huang, J Qin, Y Zhou, F Zhu, L Liu… - IEEE transactions on …, 2023 - ieeexplore.ieee.org
Normalization techniques are essential for accelerating the training and improving the
generalization of deep neural networks (DNNs), and have successfully been used in various …

Resurrecting recurrent neural networks for long sequences

A Orvieto, SL Smith, A Gu, A Fernando… - International …, 2023 - proceedings.mlr.press
Abstract Recurrent Neural Networks (RNNs) offer fast inference on long sequences but are
hard to optimize and slow to train. Deep state-space models (SSMs) have recently been …

Parameterized quantum circuits as machine learning models

M Benedetti, E Lloyd, S Sack… - Quantum Science and …, 2019 - iopscience.iop.org
Hybrid quantum–classical systems make it possible to utilize existing quantum computers to
their fullest extent. Within this framework, parameterized quantum circuits can be regarded …

An empirical evaluation of generic convolutional and recurrent networks for sequence modeling

S Bai, JZ Kolter, V Koltun - arxiv preprint arxiv:1803.01271, 2018 - arxiv.org
For most deep learning practitioners, sequence modeling is synonymous with recurrent
networks. Yet recent results indicate that convolutional architectures can outperform …

Concept whitening for interpretable image recognition

Z Chen, Y Bei, C Rudin - Nature Machine Intelligence, 2020 - nature.com
What does a neural network encode about a concept as we traverse through the layers?
Interpretability in machine learning is undoubtedly important, but the calculations of neural …

Independently recurrent neural network (indrnn): Building a longer and deeper rnn

S Li, W Li, C Cook, C Zhu… - Proceedings of the IEEE …, 2018 - openaccess.thecvf.com
Recurrent neural networks (RNNs) have been widely used for processing sequential data.
However, RNNs are commonly difficult to train due to the well-known gradient vanishing and …

Regularizing and optimizing LSTM language models

S Merity, NS Keskar, R Socher - arxiv preprint arxiv:1708.02182, 2017 - arxiv.org
Recurrent neural networks (RNNs), such as long short-term memory networks (LSTMs),
serve as a fundamental building block for many sequence learning tasks, including machine …

Deep complex networks

C Trabelsi, O Bilaniuk, Y Zhang, D Serdyuk… - arxiv preprint arxiv …, 2017 - arxiv.org
At present, the vast majority of building blocks, techniques, and architectures for deep
learning are based on real-valued operations and representations. However, recent work on …

Phase-aware speech enhancement with deep complex u-net

HS Choi, JH Kim, J Huh, A Kim, JW Ha… - … Conference on Learning …, 2018 - openreview.net
Most deep learning-based models for speech enhancement have mainly focused on
estimating the magnitude of spectrogram while reusing the phase from noisy speech for …