Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1041
|View full text |Cite
|
Sign up to set email alerts
|

Rotated Word Vector Representations and their Interpretability

Abstract: Vector representation of words improves performance in various NLP tasks, but the high-dimensional word vectors are very difficult to interpret. We apply several rotation algorithms to the vector representation of words to improve the interpretability. Unlike previous approaches that induce sparsity, the rotated vectors are interpretable while preserving the expressive performance of the original vectors. Furthermore, any pre-built word vector representation can be rotated for improved interpretability. We app… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
44
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 42 publications
(49 citation statements)
references
References 36 publications
(30 reference statements)
0
44
0
Order By: Relevance
“…16 In this kind of work, a word embedding model may be deemed more interpretable if humans are better able to identify the intruding words. Since the evaluation is costly for high-dimensional representations, alternative automatic metrics were considered (Park et al, 2017;Senel et al, 2018).…”
Section: Other Methodsmentioning
confidence: 99%
“…16 In this kind of work, a word embedding model may be deemed more interpretable if humans are better able to identify the intruding words. Since the evaluation is costly for high-dimensional representations, alternative automatic metrics were considered (Park et al, 2017;Senel et al, 2018).…”
Section: Other Methodsmentioning
confidence: 99%
“…Note, however, that it was shown in [27] that total interpretability of an embedding is constant under any orthogonal transformation and it can only be redistributed across the dimensions. With a similar motivation to [27], [28] proposed rotation algorithms based on exploratory factor analysis (EFA) to preserve the expressive performance of the original word embeddings while improving their interpretability. In [28], interpretability was calculated using a distance ratio (DR) metric that is effectively proportional to the metric used in [27].…”
Section: Related Workmentioning
confidence: 99%
“…With a similar motivation to [27], [28] proposed rotation algorithms based on exploratory factor analysis (EFA) to preserve the expressive performance of the original word embeddings while improving their interpretability. In [28], interpretability was calculated using a distance ratio (DR) metric that is effectively proportional to the metric used in [27]. Although interpretability evaluations used in [27] and [28] are free of human effort, they do not necessarily reflect human interpretations since they are directly calculated from the embeddings.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Though powerful and fairly easy to implement with specialized packages (e.g., the Gensim library; Rehurek & Sojka, 2010), these new methods still suffer in part from a crucial drawback shared with LSA, in that the embeddings used to assess semantic similarity are highdimensional mathematical spaces whose intrinsic meaning can be challenging to apprehend (Smalheiser & Bonifield, 2018). Though there has been research into techniques that attempt to address this issue (e.g., Luo, Liu, Luan, & Sun, 2015;Park, Bak, & Oh, 2017), generally these approaches make both the interpretation of the dimensions of the semantic space and understanding of the influence of specific keywords difficult. Further, though some of the simplicity of using word2vec comes from using pretrained embeddings, these spaces may not be optimal for particular applications, and training new embeddings can present several challenges (Smalheiser & Bonifield, 2018).…”
Section: Quantifying Semantic Contentmentioning
confidence: 99%