Deep face recognition: A survey

M Wang, W Deng - Neurocomputing, 2021 - Elsevier
Deep learning applies multiple processing layers to learn representations of data with
multiple levels of feature extraction. This emerging technique has reshaped the research …

Network representation learning: A survey

D Zhang, J Yin, X Zhu, C Zhang - IEEE transactions on Big Data, 2018 - ieeexplore.ieee.org
With the widespread use of information technologies, information networks are becoming
increasingly popular to capture complex relationships across various disciplines, such as …

Large language models are zero-shot time series forecasters

N Gruver, M Finzi, S Qiu… - Advances in Neural …, 2023 - proceedings.neurips.cc
By encoding time series as a string of numerical digits, we can frame time series forecasting
as next-token prediction in text. Develo** this approach, we find that large language …

Transformers are rnns: Fast autoregressive transformers with linear attention

A Katharopoulos, A Vyas, N Pappas… - … on machine learning, 2020 - proceedings.mlr.press
Transformers achieve remarkable performance in several tasks but due to their quadratic
complexity, with respect to the input's length, they are prohibitively slow for very long …

[PDF][PDF] Bert: Pre-training of deep bidirectional transformers for language understanding

JDMWC Kenton, LK Toutanova - Proceedings of naacL-HLT, 2019 - au1206.github.io
We introduce a new language representation model called BERT, which stands for
Bidirectional Encoder Representations from Transformers. Unlike recent language …

Bert: Pre-training of deep bidirectional transformers for language understanding

J Devlin, MW Chang, K Lee… - Proceedings of the 2019 …, 2019 - aclanthology.org
We introduce a new language representation model called BERT, which stands for
Bidirectional Encoder Representations from Transformers. Unlike recent language …

Neural attentive session-based recommendation

J Li, P Ren, Z Chen, Z Ren, T Lian, J Ma - Proceedings of the 2017 ACM …, 2017 - dl.acm.org
Given e-commerce scenarios that user profiles are invisible, session-based
recommendation is proposed to generate recommendation results from short sessions …

Word embeddings: A survey

F Almeida, G Xexéo - arxiv preprint arxiv:1901.09069, 2019 - arxiv.org
This work lists and describes the main recent strategies for building fixed-length, dense and
distributed representations for words, based on the distributional hypothesis. These …

An efficient framework for learning sentence representations

L Logeswaran, H Lee - arxiv preprint arxiv:1803.02893, 2018 - arxiv.org
In this work we propose a simple and efficient framework for learning sentence
representations from unlabelled data. Drawing inspiration from the distributional hypothesis …

[KNJIGA][B] Deep learning

I Goodfellow, Y Bengio, A Courville, Y Bengio - 2016 - synapse.koreamed.org
Kwang Gi Kim https://doi. org/10.4258/hir. 2016.22. 4.351 ing those who are beginning their
careers in deep learning and artificial intelligence research. The other target audience …