A detailed review on word embedding techniques with emphasis on word2vec

SJ Johnson, MR Murty, I Navakanth - Multimedia Tools and Applications, 2024 - Springer
Text data has been growing drastically in the present day because of digitalization. The
Internet, being flooded with millions of documents every day, makes the task of text …

An overview of word and sense similarity

R Navigli, F Martelli - Natural Language Engineering, 2019 - cambridge.org
Over the last two decades, determining the similarity between words as well as between
their meanings, that is, word senses, has been proven to be of vital importance in the field of …

From word to sense embeddings: A survey on vector representations of meaning

J Camacho-Collados, MT Pilehvar - Journal of Artificial Intelligence …, 2018 - jair.org
Over the past years, distributed semantic representations have proved to be effective and
flexible keepers of prior knowledge to be integrated into downstream applications. This …

Topical word embeddings

Y Liu, Z Liu, TS Chua, M Sun - Proceedings of the AAAI Conference on …, 2015 - ojs.aaai.org
Most word embedding models typically represent each word using a single vector, which
makes these models indiscriminative for ubiquitous homonymy and polysemy. In order to …

[PDF][PDF] A unified model for word sense representation and disambiguation

X Chen, Z Liu, M Sun - Proceedings of the 2014 conference on …, 2014 - aclanthology.org
Most word representation methods assume that each word owns a single semantic vector.
This is usually problematic because lexical ambiguity is ubiquitous, which is also the …

Autoextend: Extending word embeddings to embeddings for synsets and lexemes

S Rothe, H Schütze - arxiv preprint arxiv:1507.01127, 2015 - arxiv.org
We present\textit {AutoExtend}, a system to learn embeddings for synsets and lexemes. It is
flexible in that it can take any word embeddings as input and does not need an additional …

Probabilistic fasttext for multi-sense word embeddings

B Athiwaratkun, AG Wilson, A Anandkumar - arxiv preprint arxiv …, 2018 - arxiv.org
We introduce Probabilistic FastText, a new model for word embeddings that can capture
multiple word senses, sub-word structure, and uncertainty information. In particular, we …

Inter-block GPU communication via fast barrier synchronization

S **ao, W Feng - … IEEE International Symposium on Parallel & …, 2010 - ieeexplore.ieee.org
While GPGPU stands for general-purpose computation on graphics processing units, the
lack of explicit support for inter-block communication on the GPU arguably hampers its …

Breaking sticks and ambiguities with adaptive skip-gram

S Bartunov, D Kondrashkin… - artificial intelligence …, 2016 - proceedings.mlr.press
The recently proposed Skip-gram model is a powerful method for learning high-dimensional
word representations that capture rich semantic relationships between words. However …

Improved word representation learning with sememes

Y Niu, R **e, Z Liu, M Sun - … of the 55th Annual Meeting of the …, 2017 - aclanthology.org
Sememes are minimum semantic units of word meanings, and the meaning of each word
sense is typically composed by several sememes. Since sememes are not explicit for each …