[BOOK][B] Exploiting cross-lingual representations for natural language processing

S Upadhyay - 2019 - search.proquest.com
Traditional approaches to supervised learning require a generous amount of labeled data
for good generalization. While such annotation-heavy approaches have proven useful for …

[PDF][PDF] Constructing high quality sense-specific corpus and word embedding via unsupervised elimination of pseudo multi-sense

H Shi, X Wang, Y Sun, J Hu - Proceedings of the eleventh …, 2018 - aclanthology.org
Multi-sense word embedding is an important extension of neural word embeddings. By
leveraging context of each word instance, multi-prototype version of word embeddings were …

Understanding and improving multi-sense word embeddings via extended robust principal component analysis

H Shi, Y Sun, J Hu - arxiv preprint arxiv:1803.01255, 2018 - arxiv.org
Unsupervised learned representations of polysemous words generate a large of pseudo
multi senses since unsupervised methods are overly sensitive to contextual variations. In …

[PDF][PDF] 基于历时语料库的在线词典编纂系统设计

吴先, 胡俊峰 - 中文信息学报, 2020 - cips-cl.org
语料库语言学是借助大规模语料库对语言现象进行发现, 挖掘的研究学科,
目前已经存在很多在线语料库辅助语言学家的研究. 该文提供了一个按时间分片进行管理的 …

Implicit Subjective and Sentimental Usages in Multi-sense Word Embeddings

Y Sun, H Shi, J Hu - Proceedings of the 9th Workshop on …, 2018 - aclanthology.org
In multi-sense word embeddings, contextual variations in corpus may cause a univocal word
to be embedded into different sense vectors. Shi et al.(2016) show that this kind of pseudo …