A Neurobiologically Motivated Analysis of Distributional Semantic Models
Abstract
The pervasive use of distributional semantic models or word embeddings in a variety of research fields is due to their remarkable ability to represent the meanings of words for both practical application and cognitive modeling. However, little has been known about what kind of information is encoded in text-based word vectors. This lack of understanding is particularly problematic when word vectors are regarded as a model of semantic representation for abstract concepts. This paper attempts to reveal the internal information of distributional word vectors by the analysis using Binder \BOthers.’s \APACyear2016 brain-based vectors, explicitly structured conceptual representations based on neurobiologically motivated attributes. In the analysis, the mapping from text-based vectors to brain-based vectors is trained and prediction performance is evaluated by comparing the estimated and original brain-based vectors. The analysis demonstrates that social and cognitive information is better encoded in text-based word vectors, but emotional information is not. This result is discussed in terms of embodied theories for abstract concepts.
Keywords: Distributional semantic models; Word vectors; Brain-based representation; Embodied cognition; Emotional and social information; Abstract concepts
Introduction
One of the most important advances in the study of semantic processing is the development of distributional semantic models for representing word meanings. In the distributional semantic model, words are represented as high-dimensional vectors, which can be learned from the distributional statistics of word occurrence in large collections of text. Any words that occur in the corpus can be learned regardless of their part-of-speech class, abstractness, novelty and familiarity. This is an important advantage of text-based distributional semantic models over other spatial models of semantic representation such as feature-based vectors Andrews \BOthers. (\APACyear2009) and image-based vectors Silberer \BOthers. (\APACyear2017).
Word vectors have been employed in a variety of research fields and many successful results have been obtained. In the field of natural language processing (NLP), deep learning has recently been applied to a number of NLP tasks such as machine translation and automatic summarization, and achieved the impressive performance as compared to the traditional statistical methods. One of the reasons for the successful results is the use of word vectors as semantic representations for the input and output of recurrent neural networks Goldberg (\APACyear2017). Research on cognitive science also benefits greatly from distributional semantic models Jones \BOthers. (\APACyear2015). Word vectors have been demonstrated to explain a number of cognitive phenomena relevant to semantic memory or mental lexicon, such as word association Jones \BOthers. (\APACyear2017); Utsumi (\APACyear2015), semantic priming Mandera \BOthers. (\APACyear2017), semantic transparency Marelli \BBA Baroni (\APACyear2015) and conceptual combination Vecchi \BOthers. (\APACyear2017). Furthermore, recent brain imaging studies have demonstrated that distributional word vectors have a powerful ability to predict the neural brain activity evoked by lexical processing Mitchell \BOthers. (\APACyear2008); Huth \BOthers. (\APACyear2016); Güçlü \BBA van Gerven (\APACyear2015). These voxel-wise modeling by word vectors is expected to open a door for brain-machine interfaces.
Despite the fact that successful results are obtained in many research fields, little has been known about what kind of information or knowledge is encoded in word vectors. This lack of understanding makes distributional semantic models unable to predict human language behavior and performance at the same level of detail and precision of other cognitive models. It also limits further improvements on the practical performance of word vectors for many NLP tasks.
In this paper, therefore, we attempt to reveal the internal information (or knowledge) encoded in text-based word vectors generated by distributional semantic models. Our approach to this problem is to simulate a brain-based semantic representation proposed by \citeABinder16 using text-based vectors. This semantic representation comprises 65 attributes based entirely on functional divisions in the human brain. Each word is represented as a 65-dimensional vector and each dimension represents the salience of the corresponding attribute, namely the degree to which the concept referred to by that word is related to that attribute. Because these attributes are based on not only sensorimotor experiences but also affective, social, and cognitive experiences, we can analyze distributional word vectors considering a wide variety of information. In the analysis, we trained the mapping from the text-based vectors to the brain-based vectors, by which brain-based vectors of untrained words are predicted. Prediction accuracy was measured for each attribute and word using a leave-one-out cross-validation.
The secondary purpose of this paper is to discuss the relationship between the embodied theory for abstract words and distributional semantic models from the results of the analysis. Recently it has been accepted that language or linguistic experience is much more important for abstract concepts than for concrete concepts, because abstract words are unlikely to be grounded in perceptual and sensorimotor experiences, in which concrete concepts are grounded Borghi \BOthers. (\APACyear2017) A number of approaches have been proposed to explain the role of language as a simple shortcut Barsalou \BOthers. (\APACyear2008) or indirect grounding in perceptual or sensorimotor experiences Louwerse (\APACyear2011); Dove (\APACyear2014), and the need for other information such as emotional Kousta \BOthers. (\APACyear2011) and social information Borghi \BBA Binkofski (\APACyear2014). The analysis of information encoded in text-based word vectors, which can be regarded as realizations of linguistic experiences, is expected to provide some implications for recent embodied approached to abstract concepts.
Method
In order to examine what kind of information is encoded in distributional word vectors, we evaluated how accurately they can simulate Binder \BOthers.’s \APACyear2016 brain-based vectors. The simulation was performed by training the mapping from text-based vectors to brain-based vectors and applying the trained mapping to the text-based vectors of untrained words. Prediction performance was evaluated by comparing the estimated brain-based vectors with the original brain-based vectors.
Category | Word | Category | Word |
---|---|---|---|
plant | apricot, rose, tree | human | actor, girl, parent |
vehicle | car, subway, boat | social action | celebrate, help |
place | airport, lake, lab | visual property | black, new, dark |
Domain | Attributes |
---|---|
Vision | Vision, Bright, Dark, Color, Pattern, Large, Small, Motion, Biomotion, Fast, Slow, Shape, Complexity, Face, Body |
Somatic | Touch, Temperature, Texture, Weight, Pain |
Audition | Audition, Loud, Low, High, Sound, Music, Speech |
Gustation | Taste |
Olfaction | Smell |
Motor | Head, UpperLimb, LowerLimb, Practice |
Spatial | Landmark, Path, Scene, Near, Toward, Away, Number |
Temporal | Time, Duration, Long, Short |
Causal | Caused, Consequential |
Social | Social, Human, Communication, Self |
Cognition | Cognition |
Emotion | Benefit, Harm, Pleasant, Unpleasant, Happy, Sad, Angry, Disgusted, Fearful, Surprised |
Drive | Drive, Needs |
Attention | Attention, Arousal |
Brain-based Vectors
As mentioned above, we used Binder \BOthers.’s \APACyear2016 brain-based componential representation of words as a gold standard. They provided 65-dimensional vectors of 535 words comprising 434 nouns, 62 verbs and 39 adjectives, some of which are listed in Table 1. The dimensions correspond to neurobiologically plausible attributes whose neural correlates have been well described. Table 2 lists 65 attributes (and 14 domains) used in Binder \BOthers.’s \APACyear2016 brain-based vectors.
Word Vectors
In order to ensure the generality of the findings obtained through the analysis, we constructed six semantic spaces, which were obtained from the combinations of three distributional semantic models (SGNS, GloVe, PPMI) and two corpora (COCA and Wikipedia). As a distributional semantic model, we used three representative models, namely skip-gram with negative sampling <SGNS;¿Mikolov13a, GloVe Pennington \BOthers. (\APACyear2015) and positive pointwise mutual information (PPMI) with SVD Bullinaria \BBA Levy (\APACyear2007). SGNS and GloVe are prediction-based models that train word vectors by predicting context words on either side of a target word, while PPMI is a counting-based model that trains word vectors by counting and weighting word occurrences. We set a vector dimension and a window size for all semantic spaces.
Two corpora used in the analysis were English Wikipedia dump of enwiki-20160601 (Wiki) and Corpus of Contemporary American English (COCA). The Wiki and COCA corpora include 1.89G and 0.56G word tokens, respectively. We built a vocabulary from frequent words that occur 50 times or more in Wiki corpus 111Out of 535 words for brain-based vectors, only one word “joviality” was not selected as frequent words for Wiki corpus. Hence, we added it to the vocabulary for Wiki corpus. or 30 times of more in COCA corpus. As a result, the vocabulary of Wiki and COCA contained 291,769 words and 108,230 words, respectively. These two corpora differ in that Wiki is a raw text corpus that is untagged and unlemmatized, while COCA is a fully tagged and lemmatized corpus. For Wiki corpus, raw texts were extracted from the dump files using WikiExtractor.py 222http://medialab.di.unipi.it/wiki/Wikipedia_Extractor and no other preprocessing, such as lemmatization, was applied.
Training the Mapping from Text-based Vectors to Brain-based Vectors
We used two learning methods, namely linear transformation (LT) and multi-layer perceptron (MLP). LT trains a mapping matrix M such that where B is the matrix with brain-based word vectors as rows and W is a matrix with text-based word vectors as rows. MLP trains a neural network with one hidden layer comprising 150 sigmoid units and a linear output layer. In both methods, the mapping was trained by minimizing the mean squared error, and gradient descent with AdaGrad was used as an optimization method.
Estimation of brain-based vectors from text-based vectors was performed by a leave-one-out cross validation procedure. For each of the 535 words, we trained the mapping between brain-based and text-based vectors of the remaining 534 words and estimated a brain-based vector for the target word using the trained mapping. By repeating this procedure for all words as a target, we obtained with estimated brain-based vectors as rows.
Performance Measure
Prediction performance of the estimated vectors was measured using Spearman’s rank correlation between the estimated brain-based matrix and the original matrix B. 333Mean squared error can also be a measure for prediction performance. However, we are interested in the similarity of order, rather than of absolute value, between the original and estimated vectors, and thus we used rank correlations in this paper. We performed two analyses: column-wise and row-wise matrix correlation. The column-wise matrix correlation indicates the estimation accuracy for each attribute, while the row-wise correlation indicates the accuracy for each word.
In addition, we performed a k-means clustering analysis in which 535 words were grouped into 28 clusters using the estimated brain-based vectors, and the obtained clustering result was compared with the 28-cluster solution computed using the original brain-based vectors by \citeABinder16. The clustering result was evaluated for each gold-standard cluster by the normalized entropy as follows:
(1) |
where is the -th gold-standard cluster and denotes the number of words in that were assigned to the -th estimated cluster. The normalized entropy represents how diversely words in a word category are clustered by the estimated vectors. A lower entropy implies that more words in are grouped into the same cluster. If and only if all words in are grouped into one cluster, .
correlation_attribute.eps
correlation_domain.eps
SGNS | GloVe | PPMI | ||
---|---|---|---|---|
Wikipedia | MLP | 0.576 | 0.522 | 0.483 |
LT | 0.549 | 0.450 | 0.429 | |
COCA | MLP | 0.634 | 0.554 | 0.440 |
LT | 0.598 | 0.494 | 0.454 |
Result
Correlation Analysis by Attribute
We evaluated the prediction accuracy for attributes by computing column-wise matrix correlations between the estimated and original brain-based vector spaces. Figure 1 shows correlation coefficients for 65 attributes. In addition, these results are summarized in Figure 2, which depicts mean correlations averaged over attributes of the same domain.
Although in this paper we are not concerned with the overall performance of word vectors, Table 3 shows that SGNS achieved the best prediction performance, and word vectors trained using the COCA corpus were superior to those of the Wiki corpus. In addition, as expected, MLP trained better mappings than LT. Despite these differences of overall performance, Figures 1 and 2 demonstrate that relative performance among attributes did not significantly differ, regardless of distributional model, corpus and training method.
Attributes in causal, cognitive, social, and attentional domains were generally predicted with higher accuracy (i.e., their rank correlations of SGNS+COCA+MLP exceeded 0.7). In other words, the information of these attributes, which characterize abstract concepts, is likely to be encoded in text-based word vectors. It seems to suggest that abstract concepts can be largely acquired through linguistic experiences. On the other hand, sensorimotor and spatiotemporal attributes were relatively more difficult to predict from text-based word vectors. This result is consistent with the embodied view of cognition that perceptual or sensorimotor information for grounding concrete concepts cannot be acquired through linguistic experiences. Note that some perceptual attributes such as vision, pattern, shape, texture and sound were predicted as accurately as abstract attributes, suggesting that text-based word vectors can encode these kinds of information.
A somewhat surprising result was that emotional attributes were not predicted as accurately as social and cognitive ones, although a large number of NLP studies have demonstrated successful results of sentiment analysis Taboada (\APACyear2016). From a cognitive science (or embodied cognition) perspective, however, this result suggests that emotional information is more likely to be acquired from direct emotional experiences than from linguistic ones, and it is consistent with the view that emotional experiences are required for grounding abstract concepts Kousta \BOthers. (\APACyear2011); Vigliocco \BOthers. (\APACyear2014).
correlation_word_category_sorted_ave.eps
Correlation Analysis by Word
We computed row-wise matrix correlations between the estimated and original brain-based vector spaces, and then averaged these 535 correlations according to 47 word categories. These word categories are provided a priori by \citeABinder16 and reflect grammatical classes (i.e., noun, verb, adjective) and semantic classes. 444Note that word categories provided online slightly differ from those shown in Binder \BOthers.’s \APACyear2016 article. In this paper, we used the online version of word categories. Figure 3 shows mean correlations per word category. As in the case of the attribute analysis, there were no crucial differences among semantic spaces and among training methods.
The overall result was that brain-based vectors for human-related categories such as mental action, social action, human and social event were relatively better predicted from text-based word vectors. Emotional and cognitive categories such as emotion and cognitive property were predicted well, but with lower accuracy than human-related categories. These results are consistent with the findings obtained by the attribute analysis. On the other hand, other abstract concepts, in particular many categories of action and property, were difficult to predict from text-based word vectors. Distributional semantic models may be insufficient for representing some kinds of abstract concepts, and other experiences than linguistic one would be required <e.g.,¿Borghi17.
Interestingly, many artifact categories such as instruments, food, and vehicle, and some natural objects such as plant and animal showed higher prediction performance. There is no doubt that, as the embodied theory of language argues, these concrete words or concepts are grounded in perceptual and sensorimotor experiences, but some kinds of concrete concepts, in particular artifacts, may be able to be represented (or indirectly grounded) by text-based word vectors.
kmeans_entropy.eps
Cluster Analysis
We performed a cluster analysis in which 535 words were clustered into 28 clusters by their estimated brain-based vectors. In the cluster analysis, k-means clustering was used with k-means++ initialization. Because a k-means algorithm is nondeterministic owing to random initialization, we repeated k-means clustering 10 times and averaged mean entropy over these 10 trials. The gold-standard set of 28 word categories is provided through Binder \BOthers.’s \APACyear2016 data-driven cluster analysis of the original brain-based vectors. The data-driven clustering revealed several novel distinctions not considered in the predefined category, such as the distinction of positive/negative and social/non-social categories.
Figure 4 shows the result of cluster analysis, i.e., mean normalized entropies for 28 gold-standard categories, using the predicted vectors by SGNS+COCA+MLP (i.e., MLP learning for SGNS+COCA vectors). The overall mean entropy averaged across categories was 0.434 for the estimated vectors (SGNS+COCA+MLP) and 0.459 for the original 300-dimensional SGNS+COCA vectors, indicating that some semantic information can be better represented by mapping the original text-based vectors into the brain-based vector space.
Figure 4 demonstrates that words of human-related categories such as Human Types, Neutral Human Roles, and Negative Human Roles were more likely to be grouped into the same clusters by the estimated vectors, but words in emotional categories were less likely to belong in the same cluster. This result is fully consistent with the results of correlation analysis. In addition, some natural categories such as Animals and Plants and Foods, and artifacts such as Musical Instruments and Quiet Vehicles achieved very low entropy values. Again, this result suggests the possibility that some kinds of concrete concepts can be represented by text-based word vectors without using multimodal information.
Discussion
In this paper, we have demonstrated that text-based distributional word vectors can predict social and cognitive information quite accurately, but the accuracy of emotional information is not so high. Given the existing empirical findings on the importance of emotion for abstract concepts Vigliocco \BOthers. (\APACyear2014); Buccino \BOthers. (\APACyear2016), this result suggests that direct emotional experiences are necessary for grounding abstract concepts, and thus may lend support to some embodied theories Kousta \BOthers. (\APACyear2011); Vigliocco \BOthers. (\APACyear2014). On the other hand, some other embodied theories such as WAT theory Borghi \BBA Binkofski (\APACyear2014) have argued that social experiences also play an important role in representation of abstract concepts. However, the result of our analysis that social information can be conveyed by language may diminish the importance of social experiences for concrete concepts. Furthermore, the need of social-cognitive ability is not specific to abstract concepts; concrete concepts are acquired and processed through social abilities such as a Theory of Mind <e.g.,¿Bloom00.
It was also found from the analysis that perceptual, sensorimotor and spatiotemporal information is less likely to be encoded in word vectors. This is what is expected from a number of studies claiming that distributional semantic models learn only from co-occurrences of amodal symbols that are not grounded in the real world Glenberg \BBA Robertson (\APACyear2000). It is also consistent with the findings of multimodal distributional semantics that inclusion of visual information improves semantic representation for concrete words <e.g.,¿Kiela14. At the same time, the analysis also suggested the possibility that some perceptual information can be derived from distributional semantic models. This result does not deny the embodied account that grounding in perceptual and sensorimotor experiences is necessary for representing and acquiring concrete concepts. For practical applications to NLP and AI, however, text-based word vectors can possibly provide enough information without considering the embodied nature of word meanings.
Of course, the analysis presented in this paper is not comprehensive and has some limitations. One important limitation is that the brain-based vectors represent the salience of attributes that characterize concepts, but do not necessarily represent the value of salient attributes. For some attributes such as Bright and Happy, their value is indistinguishable from their salience, but many other attributes such as Color and Human have distinct values independent of their salience. Hence, the analysis in this paper cannot examine the representational power of attribute values. Our analysis is also limited within a small set of vocabulary words. To generalize and refine the findings presented in this paper, we have to evaluate a much larger set of vocabulary words that are not included in Binder \BOthers.’s \APACyear2016 dataset. It would be interesting and vital for further work to extend the analysis and to develop a novel analysis method so as to overcome these limitations.
Acknowledgments
This research was supported by JSPS KAKENHI Grant Numbers JP15H02713 and SCAT Research Grant.
=3ex
References
- Andrews \BOthers. (\APACyear2009) \APACinsertmetastarAndrews09{APACrefauthors}Andrews, M., Vigliocco, G.\BCBL \BBA Vinson, D. \APACrefYearMonthDay2009. \BBOQ\APACrefatitleIntegrating experiential and distributional data to learn semantic representations Integrating experiential and distributional data to learn semantic representations.\BBCQ \APACjournalVolNumPagesPsychological Review116463–498. \PrintBackRefs\CurrentBib
- Barsalou \BOthers. (\APACyear2008) \APACinsertmetastarBarsalou08b{APACrefauthors}Barsalou, L\BPBIW., Santos, A., Simmons, W\BPBIK.\BCBL \BBA Wilson, C\BPBID. \APACrefYearMonthDay2008. \BBOQ\APACrefatitleLanguage and simulation in conceptual processing Language and simulation in conceptual processing.\BBCQ \BIn M. de Vega, A. Glenberg\BCBL \BBA A. Graesser (\BEDS), \APACrefbtitleSymbols and Embodiment: Debates on Meaning and Cognition Symbols and embodiment: Debates on meaning and cognition (\BPGS 245–283). \APACaddressPublisherNew YorkOxford University Press. \PrintBackRefs\CurrentBib
- Binder \BOthers. (\APACyear2016) \APACinsertmetastarBinder16{APACrefauthors}Binder, J\BPBIR., Conant, L\BPBIL., Humphries, C\BPBIJ., Fernandino, L., Simons, S\BPBIB., Aguilar, M.\BCBL \BBA Desai, R\BPBIH. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleToward a brain-based componential semantic representation Toward a brain-based componential semantic representation.\BBCQ \APACjournalVolNumPagesCognitive Neuropsychology333–4130–174. \PrintBackRefs\CurrentBib
- Bloom (\APACyear2000) \APACinsertmetastarBloom00{APACrefauthors}Bloom, P. \APACrefYear2000. \APACrefbtitleHow Children Learn the Meanings of Words How children learn the meanings of words. \APACaddressPublisherMIT Press. \PrintBackRefs\CurrentBib
- Borghi \BBA Binkofski (\APACyear2014) \APACinsertmetastarBorghi14b{APACrefauthors}Borghi, A\BPBIM.\BCBT \BBA Binkofski, F. \APACrefYear2014. \APACrefbtitleWords as Social Tools: An Embodied View on Abstract Concepts Words as social tools: An embodied view on abstract concepts. \APACaddressPublisherNew YorkSpringer. \PrintBackRefs\CurrentBib
- Borghi \BOthers. (\APACyear2017) \APACinsertmetastarBorghi17{APACrefauthors}Borghi, A\BPBIM., Binkofski, F., Castelfranchi, C., Cimatti, F., Scorolli, C.\BCBL \BBA Tummolini, L. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleThe challenge of abstract concepts The challenge of abstract concepts.\BBCQ \APACjournalVolNumPagesPsychological Bulletin1433263–292. \PrintBackRefs\CurrentBib
- Buccino \BOthers. (\APACyear2016) \APACinsertmetastarBuccino16{APACrefauthors}Buccino, G., Colagè, I., Gobbi, N.\BCBL \BBA Bonaccorso, G. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleGrounding meaning in experience: A broad perspective on embodied language Grounding meaning in experience: A broad perspective on embodied language.\BBCQ \APACjournalVolNumPagesNeuroscience & Biobehavioral Reviews6969–78. \PrintBackRefs\CurrentBib
- Bullinaria \BBA Levy (\APACyear2007) \APACinsertmetastarBullinaria07{APACrefauthors}Bullinaria, J\BPBIA.\BCBT \BBA Levy, J\BPBIP. \APACrefYearMonthDay2007. \BBOQ\APACrefatitleExtracting semantic representations from word co-occurrence statistics: A computational study Extracting semantic representations from word co-occurrence statistics: A computational study.\BBCQ \APACjournalVolNumPagesBehavior Research Methods393510–526. \PrintBackRefs\CurrentBib
- Dove (\APACyear2014) \APACinsertmetastarDove14{APACrefauthors}Dove, G. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleThinking in words: Language as an embodied medium of thought Thinking in words: Language as an embodied medium of thought.\BBCQ \APACjournalVolNumPagesTopics in Cognitive Science6371–389. \PrintBackRefs\CurrentBib
- Glenberg \BBA Robertson (\APACyear2000) \APACinsertmetastarGlenberg00{APACrefauthors}Glenberg, A.\BCBT \BBA Robertson, D. \APACrefYearMonthDay2000. \BBOQ\APACrefatitleSymbol grounding and meaning: A comparison of high-dimensional and embodied theories of meaning Symbol grounding and meaning: A comparison of high-dimensional and embodied theories of meaning.\BBCQ \APACjournalVolNumPagesJournal of Memory and Language43379–401. \PrintBackRefs\CurrentBib
- Goldberg (\APACyear2017) \APACinsertmetastarGoldberg17{APACrefauthors}Goldberg, Y. \APACrefYear2017. \APACrefbtitleNeural Network Methods for Natural Language Processing Neural network methods for natural language processing. \APACaddressPublisherMorgan & Claypool Publishers. \PrintBackRefs\CurrentBib
- Güçlü \BBA van Gerven (\APACyear2015) \APACinsertmetastarGuclu15{APACrefauthors}Güçlü, U.\BCBT \BBA van Gerven, M\BPBIA\BPBIJ. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleSemantic vector space models predict neural responses to complex visual stimuli Semantic vector space models predict neural responses to complex visual stimuli.\BBCQ \APACjournalVolNumPagesarXiv:1510.04738 [q-bio.NC]. \PrintBackRefs\CurrentBib
- Huth \BOthers. (\APACyear2016) \APACinsertmetastarHuth16{APACrefauthors}Huth, A\BPBIG., de Heer, W\BPBIA., Griffiths, T\BPBIL., Theunissen, F\BPBIE.\BCBL \BBA Gallant, J\BPBIL. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleNatural speech reveals the semantic maps that tile human cerebral cortex Natural speech reveals the semantic maps that tile human cerebral cortex.\BBCQ \APACjournalVolNumPagesNature532453–458. \PrintBackRefs\CurrentBib
- Jones \BOthers. (\APACyear2017) \APACinsertmetastarJones17{APACrefauthors}Jones, M\BPBIN., Gruenenfelder, T\BPBIM.\BCBL \BBA Recchia, G. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleIn defense of spatial models of semantic representation In defense of spatial models of semantic representation.\BBCQ \APACjournalVolNumPagesNew Ideas in Psychologyin press. \PrintBackRefs\CurrentBib
- Jones \BOthers. (\APACyear2015) \APACinsertmetastarJones15{APACrefauthors}Jones, M\BPBIN., Willits, J.\BCBL \BBA Dennis, S. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleModels of semantic memory Models of semantic memory.\BBCQ \BIn J\BPBIR. Busemeyer, Z. Wang, J\BPBIT. Townsend\BCBL \BBA A. Eidels (\BEDS), \APACrefbtitleOxford Handbook of Mathematical and Computational Psychology Oxford handbook of mathematical and computational psychology (\BPGS 232–254). \APACaddressPublisherNew York, NYOxford University Press. \PrintBackRefs\CurrentBib
- Kiela \BOthers. (\APACyear2014) \APACinsertmetastarKiela14{APACrefauthors}Kiela, D., Hill, F., Korhonen, A.\BCBL \BBA Clark, S. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleImproving multi-modal representations using image dispersion: Why less is sometimes more Improving multi-modal representations using image dispersion: Why less is sometimes more.\BBCQ \BIn \APACrefbtitleProceedings of the 52nd Annual Meeting of the Association for Computational Linguistics Proceedings of the 52nd annual meeting of the association for computational linguistics (\BPGS 835–841). \PrintBackRefs\CurrentBib
- Kousta \BOthers. (\APACyear2011) \APACinsertmetastarKousta11{APACrefauthors}Kousta, S\BHBIT., Vigliocco, G., Vinson, D\BPBIP., Andrews, M.\BCBL \BBA Del Campo, E. \APACrefYearMonthDay2011. \BBOQ\APACrefatitleThe representation of abstract words: Why emotion matters The representation of abstract words: Why emotion matters.\BBCQ \APACjournalVolNumPagesJournal of Experimental Psychology: General140114–34. \PrintBackRefs\CurrentBib
- Louwerse (\APACyear2011) \APACinsertmetastarLouwerse11{APACrefauthors}Louwerse, M\BPBIM. \APACrefYearMonthDay2011. \BBOQ\APACrefatitleSymbol interdependency in symbolic and embodied cognition Symbol interdependency in symbolic and embodied cognition.\BBCQ \APACjournalVolNumPagesTopics in Cognitive Science3273–302. \PrintBackRefs\CurrentBib
- Mandera \BOthers. (\APACyear2017) \APACinsertmetastarMandera17{APACrefauthors}Mandera, P., Keuleers, E.\BCBL \BBA Brysbaert, M. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleExplaining human performance in psycholinguistic tasks with models of semantic similarity based on prediction and counting: A review and empirical validation Explaining human performance in psycholinguistic tasks with models of semantic similarity based on prediction and counting: A review and empirical validation.\BBCQ \APACjournalVolNumPagesJournal of Memory and Language9257–78. \PrintBackRefs\CurrentBib
- Marelli \BBA Baroni (\APACyear2015) \APACinsertmetastarMarelli15{APACrefauthors}Marelli, M.\BCBT \BBA Baroni, M. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleAffixation in semantic space: Modeling morpheme meanings with compositional distributional semantics Affixation in semantic space: Modeling morpheme meanings with compositional distributional semantics.\BBCQ \APACjournalVolNumPagesPsychological Review1223485–515. \PrintBackRefs\CurrentBib
- Mikolov \BOthers. (\APACyear2013) \APACinsertmetastarMikolov13a{APACrefauthors}Mikolov, T., Chen, K., Corrado, G.\BCBL \BBA Dean, J. \APACrefYearMonthDay2013. \BBOQ\APACrefatitleEfficient estimation of word representations in vector space Efficient estimation of word representations in vector space.\BBCQ \BIn \APACrefbtitleProceedings of Workshop at the International Conference on Learning Representation (ICLR). Proceedings of workshop at the international conference on learning representation (iclr). \PrintBackRefs\CurrentBib
- Mitchell \BOthers. (\APACyear2008) \APACinsertmetastarMitchell08{APACrefauthors}Mitchell, T\BPBIM., Shinkareva, S\BPBIV., Carlson, A., Chang, K\BHBIM., Malave, V\BPBIL., Mason, R\BPBIA.\BCBL \BBA Just, M\BPBIA. \APACrefYearMonthDay2008. \BBOQ\APACrefatitlePredicting human brain activity associated with the meanings of nouns Predicting human brain activity associated with the meanings of nouns.\BBCQ \APACjournalVolNumPagesScience3201191–1195. \PrintBackRefs\CurrentBib
- Pennington \BOthers. (\APACyear2015) \APACinsertmetastarPennington14{APACrefauthors}Pennington, J., Socher, R.\BCBL \BBA Manning, C\BPBID. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleGloVe: Global vectors for word representation GloVe: Global vectors for word representation.\BBCQ \BIn \APACrefbtitleProceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) Proceedings of the 2014 conference on empirical methods in natural language processing (emnlp) (\BPGS 1532–1543). \PrintBackRefs\CurrentBib
- Silberer \BOthers. (\APACyear2017) \APACinsertmetastarSilberer17{APACrefauthors}Silberer, C., Ferrari, V.\BCBL \BBA Lapata, M. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleVisually grounded meaning representations Visually grounded meaning representations.\BBCQ \APACjournalVolNumPagesIEEE Transactions on Pattern Recognition and Machine Intelligence39112284–2297. \PrintBackRefs\CurrentBib
- Taboada (\APACyear2016) \APACinsertmetastarTaboada16{APACrefauthors}Taboada, M. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleSentiment analysis: An overview from linguistics Sentiment analysis: An overview from linguistics.\BBCQ \APACjournalVolNumPagesAnnual Review of Linguistics2325–347. \PrintBackRefs\CurrentBib
- Utsumi (\APACyear2015) \APACinsertmetastarUtsumi:15:paper{APACrefauthors}Utsumi, A. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleA complex network approach to distributional semantic models A complex network approach to distributional semantic models.\BBCQ \APACjournalVolNumPagesPLoS ONE108e0136277. \PrintBackRefs\CurrentBib
- Vecchi \BOthers. (\APACyear2017) \APACinsertmetastarVecchi17{APACrefauthors}Vecchi, E\BPBIM., Marelli, M., Zamparelli, R.\BCBL \BBA Baroni, M. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleSpicy adjectives and nominal donkeys: Capturing semantic deviance using compositionality in distributional spaces Spicy adjectives and nominal donkeys: Capturing semantic deviance using compositionality in distributional spaces.\BBCQ \APACjournalVolNumPagesCognitive Science41102–136. \PrintBackRefs\CurrentBib
- Vigliocco \BOthers. (\APACyear2014) \APACinsertmetastarVigliocco14{APACrefauthors}Vigliocco, G., Kousta, S\BHBIT., Rosa, P\BPBIA\BPBID., Vinson, D\BPBIP., Tettamanti, M., Devlin, J\BPBIT.\BCBL \BBA Cappa, S\BPBIF. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleThe neural representation of abstract words: The role of emotion The neural representation of abstract words: The role of emotion.\BBCQ \APACjournalVolNumPagesCerebral Cortex241767–1777. \PrintBackRefs\CurrentBib