Understanding the Semantic Content of Sparse Word Embeddings Using a Commonsense Knowledge Base
|Title||Understanding the Semantic Content of Sparse Word Embeddings Using a Commonsense Knowledge Base|
|Publication Type||Conference Paper|
|Year of Publication||2020|
|Authors||Balogh V, Berend G, Diochnos DI, Turán G|
|Conference Name||Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20)|
Word embeddings have developed into a major NLP toolwith broad applicability. Understanding the semantic content of word embeddings remains an important challenge for additional applications. One aspect of this issue is to explore the interpretability of word embeddings. Sparse word embeddings have been proposed as models with improved interpretability. Continuing this line of research, we investigate the extent to which human interpretable semantic concepts emerge along the bases of sparse word representations. In order to have a broad framework for evaluation, we consider three general approaches for constructing sparse word representations, which are then evaluated in multiple ways. We propose a novel methodology to evaluate the semantic content of word embeddings using a commonsense knowledgebase, applied here to the sparse case. This methodology isillustrated by two techniques using the ConceptNet knowl-edge base. The first approach assigns a commonsense concept label to the individual dimensions of the embedding space. The second approach uses a metric, derived by spreading activation, to quantify the coherence of coordinates along the individual axes. We also provide results on the relationship between the two approaches. The results show, for example, that in the individual dimensions of sparse word embeddings, words having high coefficients are more semantically related in terms of path lengths in the knowledge base than the one shaving zero coefficients.