[HTML][HTML] Extracting sentence embeddings from pretrained transformer models

L Stankevičius, M Lukoševičius - Applied Sciences, 2024 - mdpi.com
Pre-trained transformer models shine in many natural language processing tasks and
therefore are expected to bear the representation of the input sentence or text meaning …

Generate, discriminate and contrast: A semi-supervised sentence representation learning framework

Y Chen, Y Zhang, B Wang, Z Liu, H Li - arxiv preprint arxiv:2210.16798, 2022 - arxiv.org
Most sentence embedding techniques heavily rely on expensive human-annotated
sentence pairs as the supervised signals. Despite the use of large-scale unlabeled data, the …

Contrastive Learning of Sentence Embeddings from Scratch

J Zhang, Z Lan, J He - arxiv preprint arxiv:2305.15077, 2023 - arxiv.org
Contrastive learning has been the dominant approach to train state-of-the-art sentence
embeddings. Previous studies have typically learned sentence embeddings either through …

What makes sentences semantically related: A textual relatedness dataset and empirical study

M Abdalla, K Vishnubhotla, SM Mohammad - arxiv preprint arxiv …, 2021 - arxiv.org
The degree of semantic relatedness of two units of language has long been considered
fundamental to understanding meaning. Additionally, automatically determining relatedness …

Compositionality and Sentence Meaning: Comparing Semantic Parsing and Transformers on a Challenging Sentence Similarity Dataset

J Fodor, S De Deyne, S Suzuki - Computational Linguistics, 2024 - direct.mit.edu
One of the major outstanding questions in computational semantics is how humans integrate
the meaning of individual words into a sentence in a way that enables understanding of …

Semantic-Aware Contrastive Sentence Representation Learning with Large Language Models

H Wang, L Cheng, Z Li, DW Soh, L Bing - arxiv preprint arxiv:2310.10962, 2023 - arxiv.org
Contrastive learning has been proven to be effective in learning better sentence
representations. However, to train a contrastive learning model, large numbers of labeled …

SimCSE++: Improving contrastive learning for sentence embeddings from two perspectives

J Xu, W Shao, L Chen, L Liu - arxiv preprint arxiv:2305.13192, 2023 - arxiv.org
This paper improves contrastive learning for sentence embeddings from two perspectives:
handling dropout noise and addressing feature corruption. Specifically, for the first …

Synwmd: Syntax-aware word mover's distance for sentence similarity evaluation

C Wei, B Wang, CCJ Kuo - Pattern Recognition Letters, 2023 - Elsevier
Abstract Word Mover's Distance (WMD) computes the distance between words and models
text similarity with the moving cost between words in two text sequences. Yet, it does not …

[PDF][PDF] Why is sentence similarity benchmark not predictive of application-oriented task performance?

K Abe, S Yokoi, T Kajiwara, K Inui - Proceedings of the 3rd …, 2022 - aclanthology.org
Computing the semantic similarity between two texts is crucial in various NLP tasks. For
more than a decade, a framework, known as Semantic Textual Similarity (STS) has been …

Large Language Models can Contrastively Refine their Generation for Better Sentence Representation Learning

H Wang, Z Li, L Cheng, L Bing - … of the 2024 Conference of the …, 2024 - aclanthology.org
Recently, large language models (LLMs) have emerged as a groundbreaking technology
and their unparalleled text generation capabilities have sparked interest in their application …