Hostname: page-component-89b8bd64d-72crv Total loading time: 0 Render date: 2026-05-07T20:59:23.180Z Has data issue: false hasContentIssue false

Compositional matrix-space models of language: Definitions, properties, and learning methods

Published online by Cambridge University Press:  09 August 2021

Shima Asaadi
Affiliation:
Technische Universität Dresden, Dresden, Germany
Eugenie Giesbrecht
Affiliation:
IBM Deutschland GmbH, Ehningen, Germany
Sebastian Rudolph*
Affiliation:
Technische Universität Dresden, Dresden, Germany
*
*Corresponding author. E-mail: sebastian.rudolph@tu-dresden.de
Rights & Permissions [Opens in a new window]

Abstract

We give an in-depth account of compositional matrix-space models (CMSMs), a type of generic models for natural language, wherein compositionality is realized via matrix multiplication. We argue for the structural plausibility of this model and show that it is able to cover and combine various common compositional natural language processing approaches. Then, we consider efficient task-specific learning methods for training CMSMs and evaluate their performance in compositionality prediction and sentiment analysis.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Author(s), 2021. Published by Cambridge University Press
Figure 0

Table 1. Summary of the literature review in semantic compositionality

Figure 1

Table 2. Summary of the literature review in compositionality prediction

Figure 2

Table 3. Summary of the literature review in compositional sentiment analysis. SST denotes Stanford Sentiment Treebank dataset

Figure 3

Figure 1. Semantic mapping as homomorphism.

Figure 4

Figure 2. Simulating compositional VSM via CMSMs.

Figure 5

Table 4. Pearson value r for compositionality prediction using word2vec

Figure 6

Table 5. Pearson value r for compositionality prediction using fastText

Figure 7

Table 6. Average number of training iterations for each supervised model trained using word2vec and fastText

Figure 8

Figure 3. Sample compounds from Reddy++ with predicted average compositionality scores by different models and gold standard scores. Results of fastText embeddings are reported. Gold standard scores are between 0 (non-compositional) and 1 (fully compositional). A, adjective; N, noun; FF, Feedforward.

Figure 9

Figure 4. Sample compounds from Farahmand15 with predicted average compositionality scores by different models and gold standard scores. Results of fastText embeddings are reported. Gold standard scores are between 0 (non-compositional) and 1 (fully compositional). FF, Feedforward.

Figure 10

Figure 5. Sentiment composition of a short phrase with matrix multiplication as a composition operation in CMSMs.

Figure 11

Table 7. Phrase polarities and their occurrence frequencies in the SCL-OPP dataset

Figure 12

Table 8. Phrase polarities and intensities in the MPQA corpus, their translation into sentiment scores and their occurrence frequency

Figure 13

Table 9. Performance comparison for different methods in SCL-OPP dataset considering only trigram phrases

Figure 14

Table 10. Example phrases with average sentiment scores on 10-fold cross-validation and different POS tags. A, adjective; N, noun; V, verb; &, and; D, determiner

Figure 15

Table 11. Performance comparison for different dimensions of matrices in the complete SCL-OPP dataset (i.e., considering bigrams and trigrams for the experiment)

Figure 16

Table 12. Ranking loss of compared methods

Figure 17

Table 13. Frequent phrases with average sentiment scores

Figure 18

Figure 6. The order of sentiment scores for sample phrases (trained on MPQA corpus).

Figure 19

Table 14. Time cost for training CMSMs with different dimensionality and datasets. Time is reported in minutes

Figure 20

Figure A1. Circular convolution operation on two three-dimensional vectors ${{\textbf{v}}_1}$ and ${{\textbf{v}}_2}$. Illustration adapted from Plate (1995).

Figure 21

Figure C1. Matrices as cognitive state transformations.