Why concepts are (probably) vectors
For decades, cognitive scientists have debated what kind of representation might
characterize human concepts. Whatever the format of the representation, it must allow for the …
characterize human concepts. Whatever the format of the representation, it must allow for the …
Evaluating word embedding models: Methods and experimental results
Extensive evaluation on a large number of word embedding models for language
processing applications is conducted in this work. First, we introduce popular word …
processing applications is conducted in this work. First, we introduce popular word …
Inducing relational knowledge from BERT
One of the most remarkable properties of word embeddings is the fact that they capture
certain types of semantic and syntactic relationships. Recently, pre-trained language models …
certain types of semantic and syntactic relationships. Recently, pre-trained language models …
Fair is better than sensational: Man is to doctor as woman is to doctor
Analogies such as man is to king as woman is to X are often used to illustrate the amazing
power of word embeddings. Concurrently, they have also been used to expose how strongly …
power of word embeddings. Concurrently, they have also been used to expose how strongly …
A comparative evaluation and analysis of three generations of Distributional Semantic Models
Distributional semantics has deeply changed in the last decades. First, predict models stole
the thunder from traditional count ones, and more recently both of them were replaced in …
the thunder from traditional count ones, and more recently both of them were replaced in …
Transfer learning and analogical inference: A critical comparison of algorithms, methods, and applications
Artificial intelligence and machine learning (AI/ML) research has aimed to achieve human-
level performance in tasks that require understanding and decision making. Although major …
level performance in tasks that require understanding and decision making. Although major …
Approaches to improve preprocessing for Latent Dirichlet Allocation topic modeling
As a part of natural language processing (NLP), the intent of topic modeling is to identify
topics in textual corpora with limited human input. Current topic modeling techniques, like …
topics in textual corpora with limited human input. Current topic modeling techniques, like …
Does ChatGPT have semantic understanding? A problem with the statistics-of-occurrence strategy
LM Titus - Cognitive Systems Research, 2024 - Elsevier
Over the last decade, AI models of language and word meaning have been dominated by
what we might call a statistics-of-occurrence, strategy: these models are deep neural net …
what we might call a statistics-of-occurrence, strategy: these models are deep neural net …
From word types to tokens and back: A survey of approaches to word meaning representation and interpretation
M Apidianaki - Computational Linguistics, 2023 - direct.mit.edu
Vector-based word representation paradigms situate lexical meaning at different levels of
abstraction. Distributional and static embedding models generate a single vector per word …
abstraction. Distributional and static embedding models generate a single vector per word …
Scientific and creative analogies in pretrained language models
This paper examines the encoding of analogy in large-scale pretrained language models,
such as BERT and GPT-2. Existing analogy datasets typically focus on a limited set of …
such as BERT and GPT-2. Existing analogy datasets typically focus on a limited set of …