Automatic music transcription: An overview

E Benetos, S Dixon, Z Duan… - IEEE Signal Processing …, 2018 - ieeexplore.ieee.org
The capability of transcribing music audio into music notation is a fascinating example of
human intelligence. It involves perception (analyzing complex auditory scenes), cognition …

A tutorial on deep learning for music information retrieval

K Choi, G Fazekas, K Cho, M Sandler - arxiv preprint arxiv:1709.04396, 2017 - arxiv.org
Following their success in Computer Vision and other areas, deep learning techniques have
recently become widely adopted in Music Information Retrieval (MIR) research. However …

Heart sound classification based on improved MFCC features and convolutional recurrent neural networks

M Deng, T Meng, J Cao, S Wang, J Zhang, H Fan - Neural Networks, 2020 - Elsevier
Heart sound classification plays a vital role in the early detection of cardiovascular disorders,
especially for small primary health care clinics. Despite that much progress has been made …

Convolutional recurrent neural networks for polyphonic sound event detection

E Cakır, G Parascandolo, T Heittola… - … on Audio, Speech …, 2017 - ieeexplore.ieee.org
Sound events often occur in unstructured environments where they exhibit wide variations in
their frequency content and temporal structure. Convolutional neural networks (CNNs) are …

Adieu features? end-to-end speech emotion recognition using a deep convolutional recurrent network

G Trigeorgis, F Ringeval, R Brueckner… - … on acoustics, speech …, 2016 - ieeexplore.ieee.org
The automatic recognition of spontaneous emotions from speech is a challenging task. On
the one hand, acoustic features need to be robust enough to capture the emotional content …

Convolutional recurrent neural networks for music classification

K Choi, G Fazekas, M Sandler… - 2017 IEEE International …, 2017 - ieeexplore.ieee.org
We introduce a convolutional recurrent neural network (CRNN) for music tagging. CRNNs
take advantage of convolutional neural networks (CNNs) for local feature extraction and …

Onsets and frames: Dual-objective piano transcription

C Hawthorne, E Elsen, J Song, A Roberts… - arxiv preprint arxiv …, 2017 - arxiv.org
We advance the state of the art in polyphonic piano music transcription by using a deep
convolutional and recurrent neural network which is trained to jointly predict onsets and …

Automatic tagging using deep convolutional neural networks

K Choi, G Fazekas, M Sandler - arxiv preprint arxiv:1606.00298, 2016 - arxiv.org
We present a content-based automatic music tagging algorithm using fully convolutional
neural networks (FCNs). We evaluate different architectures consisting of 2D convolutional …

[KSIĄŻKA][B] Learning-based methods for comparing sequences, with applications to audio-to-midi alignment and matching

C Raffel - 2016 - search.proquest.com
Sequences of feature vectors are a natural way of representing temporal data. Given a
database of sequences, a fundamental task is to find the database entry which is the most …

CNN-RNN and data augmentation using deep convolutional generative adversarial network for environmental sound classification

B Bahmei, E Birmingham… - IEEE Signal Processing …, 2022 - ieeexplore.ieee.org
Deep neural networks in deep learning have been widely demonstrated to have higher
accuracy and distinct advantages over traditional machine learning methods in extracting …