2022
DOI: 10.1016/j.neucom.2022.05.103
|View full text |Cite
|
Sign up to set email alerts
|

Enhanced distance-aware self-attention and multi-level match for sentence semantic matching

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 16 publications
0
4
0
Order By: Relevance
“…Wan et al put forward a model to match two sentences with multiple positional sentence representations, by the aggregation over interactions between different positional sentence representations and the employment of K-Max pooling, the semantic relationship between sentences can be well captured by its model design [30]. In order for more effectual extraction of sentence interactive features, in many recent studies on sentence matching, the cross-attention mechanism has been largely applied, intended for a more accurate sentence alignment, which is usually implemented by stacking multiple of them so that more in-depth interaction information can be possibly exploited [15], [31]- [33]. For instance, Chen et al utilized two Bi-LSTM (Bi-directional Long Short-Term Memory) in its model with the first to encode sentence semantic information, and the second to aggregate semantic information and sentence alignment information extracted by attention mechanism, so that more effective sentence representation vectors can be obtained for final semantic matching [3].…”
Section: B Matching-aggregation Based Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Wan et al put forward a model to match two sentences with multiple positional sentence representations, by the aggregation over interactions between different positional sentence representations and the employment of K-Max pooling, the semantic relationship between sentences can be well captured by its model design [30]. In order for more effectual extraction of sentence interactive features, in many recent studies on sentence matching, the cross-attention mechanism has been largely applied, intended for a more accurate sentence alignment, which is usually implemented by stacking multiple of them so that more in-depth interaction information can be possibly exploited [15], [31]- [33]. For instance, Chen et al utilized two Bi-LSTM (Bi-directional Long Short-Term Memory) in its model with the first to encode sentence semantic information, and the second to aggregate semantic information and sentence alignment information extracted by attention mechanism, so that more effective sentence representation vectors can be obtained for final semantic matching [3].…”
Section: B Matching-aggregation Based Methodsmentioning
confidence: 99%
“…Secondly, in recent years, due to wide application of at-tention mechanism in sentence matching models, it's not uncommon to find studies trying to address this issue by stacking multiple cross-attention layers so that implicit interactive features can be more profoundly extracted [15], [16]. However, for this interaction-layer-stacking strategy, there may exist some drawbacks.…”
Section: B Research Challengesmentioning
confidence: 99%
See 2 more Smart Citations