Abstract
Unsupervised learning of word representations involves capturing the contextual information surrounding word occurrences, which can be grounded in the observation that word form is largely disconnected from word meaning. While there are fewer reasons to believe that the same holds for sentences, learning through context has been carried over to learning representations of word sequences. However, this work pays minimal to no attention to the role of context in inferring sentence representations. In this article, we present a dialog act tag probing task designed to explicitly compare content-, and context-oriented sentence representations inferred on utterances of telephone conversations (SwDA). Our results suggest that there is no clear benefit of context-based sentence representations over content-based sentence representations. However, there is a very clear benefit of increasing the dimensionality of the sentence vectors in nearly all approaches.
| Original language | English |
|---|---|
| Pages | 8784-8792 |
| Number of pages | 9 |
| DOIs | |
| Publication status | Published - 2023 |
| Event | The 2023 Conference on Empirical Methods in Natural Language Processing - Singapore, Resort World Convention Centre, Singapore, Singapore Duration: 6 Dec 2023 → 10 Dec 2023 https://2023.emnlp.org/ |
Conference
| Conference | The 2023 Conference on Empirical Methods in Natural Language Processing |
|---|---|
| Abbreviated title | EMNLP 2023 |
| Country/Territory | Singapore |
| City | Singapore |
| Period | 6/12/23 → 10/12/23 |
| Internet address |