Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017) 2017
DOI: 10.18653/v1/s17-2032
|View full text |Cite
|
Sign up to set email alerts
|

L2F/INESC-ID at SemEval-2017 Tasks 1 and 2: Lexical and semantic features in word and textual similarity

Abstract: This paper describes our approach to the SemEval-2017 "Semantic Textual Similarity" and "Multilingual Word Similarity" tasks. In the former, we test our approach in both English and Spanish, and use a linguistically-rich set of features. These move from lexical to semantic features. In particular, we try to take advantage of the recent Abstract Meaning Representation and SMATCH measure. Although without state of the art results, we introduce semantic structures in textual similarity and analyze their impact. R… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
5
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 13 publications
1
5
0
Order By: Relevance
“…The overall results obtained following this setting are promising (with an accuracy of 0.835 in the test set). Comparing our results with those of the participants in the ASSIN Challenge [14], our approach obtains an overall accuracy close to the results obtained by the best performing system (0.8385 of accuracy obtained by Fialho et al [33]) and outperforms all the proposed systems in the corresponding macro F1-score metric (the best performing system [33] reported 0.71 of macro F1-score). Additionally, we observe that the performance of our approach improved with semantic-based features, albeit not significantly.…”
Section: Discussionsupporting
confidence: 78%
See 1 more Smart Citation
“…The overall results obtained following this setting are promising (with an accuracy of 0.835 in the test set). Comparing our results with those of the participants in the ASSIN Challenge [14], our approach obtains an overall accuracy close to the results obtained by the best performing system (0.8385 of accuracy obtained by Fialho et al [33]) and outperforms all the proposed systems in the corresponding macro F1-score metric (the best performing system [33] reported 0.71 of macro F1-score). Additionally, we observe that the performance of our approach improved with semantic-based features, albeit not significantly.…”
Section: Discussionsupporting
confidence: 78%
“…Fialho et al [33] extracted several metrics for each pair of sentences, namely edit distance, words overlap, BLEU [28] and ROUGE [34], among others. They reported several experiments considering different preprocessing steps in the NLP pipeline, namely: original sentences (baseline), removing stop-words, lower-case words and clusters of words.…”
Section: Feature-engineered Machine Learning Modelsmentioning
confidence: 99%
“…15 Select methods are highlighted below. 83.02• 15.50 compiLIG 76.84 14.64 compiLIG 79.10 14.94 DT TEAM (Maharjan et al, 2017) 85.36 DT TEAM (Maharjan et al, 2017) 83.60 DT TEAM (Maharjan et al, 2017) 83.29 FCICU (Hassan et al, 2017) 82.17 ITNLPAiKF (Liu et al, 2017) 82.31 ITNLPAiKF (Liu et al, 2017) 82.31 ITNLPAiKF (Liu et al, 2017) 81.59 L2F/INESC-ID (Fialho et al, 2017 10: STS 2017 rankings ordered by average correlation across tracks 1-5. Performance is reported by convention as Pearson's r × 100.…”
Section: Methodsmentioning
confidence: 99%
“…In the end we received a wide variety of participants: proposing distributional semantic models learnt directly from raw corpora, using syntactic features, exploiting knowledge from lexical resources, and hybrid approaches combining corpus-based and knowledge-based clues. Due to lack of space we cannot describe all the systems in detail, but we recommend the reader to refer to the system description papers for more information about the individual systems: HCCL (He et al, 2017), Citius (Gamallo, 2017), jmp8 (Melka and Bernard, 2017), l2f (Fialho et al, 2017), QLUT (Meng et al, 2017), RUFINO (Jimenez et al, 2017), MERALI (Mensa et al, 2017), Luminoso (Speer and Lowry-Duda, 2017), hhu (Qasem-iZadeh and Kallmeyer, 2017), Mahtab (Ranjbar et al, 2017), SEW (Delli Bovi and Raganato, 2017) and Wild Devs (Rotari et al, 2017), and OoO.…”
Section: Participating Systemsmentioning
confidence: 99%