2017
DOI: 10.1609/icwsm.v11i1.14963
|View full text |Cite
|
Sign up to set email alerts
|

25 Tweets to Know You: A New Model to Predict Personality with Social Media

Abstract: Predicting personality is essential for social applications supporting human-centered activities, yet prior modeling methods with users’ written text require too much input data to be realistically used in the context of social media. In this work, we aim to drastically reduce the data requirement for personality modeling and develop a model that is applicable to most users on Twitter. Our model integrates Word Embedding features with Gaussian Processes regression. Based on the evaluation of over 1.3K users on… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 60 publications
(19 citation statements)
references
References 10 publications
0
15
0
Order By: Relevance
“…, 2011). Further, Arnoux et al. (2017) studied the accuracy of prior work on big five personality and the dependence on the size of the input text, and introduced a method using word embedding and Gaussian process.…”
Section: Theoretical Backgroundmentioning
confidence: 99%
“…, 2011). Further, Arnoux et al. (2017) studied the accuracy of prior work on big five personality and the dependence on the size of the input text, and introduced a method using word embedding and Gaussian process.…”
Section: Theoretical Backgroundmentioning
confidence: 99%
“…GloVe uses a count-based model, which learns embeddings by looking at how often a word appears in the context of another word within the corpus, focusing on the co-occurrence probabilities of words within a large training corpus of documents such as Wikipedia. Studies of personality inferences that use neural word embeddings include (Kamijo et al, 2016;Arnoux et al, 2017;Majumder et al, 2017;Jayaratne and Jayatilleke, 2020). Though pre-trained neural word embeddings are widely used, they assume that a word's meaning is relatively stable and does not change across different sentences.…”
Section: Word and Document Representationsmentioning
confidence: 99%
“…Based on the test results, we concluded that the CNN model trained using GloVe weighting could produce higher test data accuracy than the CNN model trained using random weighting or weighting without GloVe. Because GloVe word embedding had previously trained with large corpus from Wikipedia and Gigaword 5 (a collection of Englishlanguage news source networks), its vector representation brings external knowledge to our classification task [16]. The MBTI model only needs a slight update of the embedding weight value to reach the convergence point.…”
Section: Random and Glove Weighting Testmentioning
confidence: 99%