Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017) 2017
DOI: 10.18653/v1/s17-2026
|View full text |Cite
|
Sign up to set email alerts
|

UMDeep at SemEval-2017 Task 1: End-to-End Shared Weight LSTM Model for Semantic Textual Similarity

Abstract: We describe a modified shared-LSTM network for the Semantic Textual Similarity (STS) task at SemEval-2017. The network builds on previously explored Siamese network architectures. We treat max sentence length as an additional hyperparameter to be tuned (beyond learning rate, regularization, and dropout). Our results demonstrate that hand-tuning max sentence training length significantly improves final accuracy. After optimizing hyperparameters, we train the network on the multilingual semantic similarity task … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 6 publications
0
2
0
Order By: Relevance
“…The architecture is abstractly similar to ECNU's deep learning models. UMDeep (Barrow and Peskov, 2017) took a similar approach using LSTMs rather than CNNs for the sentence embeddings.…”
Section: Methodsmentioning
confidence: 99%
“…The architecture is abstractly similar to ECNU's deep learning models. UMDeep (Barrow and Peskov, 2017) took a similar approach using LSTMs rather than CNNs for the sentence embeddings.…”
Section: Methodsmentioning
confidence: 99%
“…Tian and Lan [27] employed the sentence embedding and word-embedding to measure the semantic textual similarity. Some other works also used word-embedding for semantic similarity measurement such as [28], [29]. The mentioned related works did not consider the grammatical structure of the sentences to estimate the semantics similarity, but the grammatical structure holds different properties of the sentences that might be useful to compute the similarity.…”
Section: Related Workmentioning
confidence: 99%