2019
DOI: 10.48550/arxiv.1906.08401
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Hierarchical Document Encoder for Parallel Corpus Mining

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…To understand how its representation quality changes with input length, we report its alignment performance of encoding whole documents and encoding only the first 512 tokens respectively. The average of sentence embedding has been shown to be a strong approach to derive document representation (Guo et al, 2019). We include it as another strong baseline in our experiments, with which we can analyze the gains brought by sentence weighting.…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…To understand how its representation quality changes with input length, we report its alignment performance of encoding whole documents and encoding only the first 512 tokens respectively. The average of sentence embedding has been shown to be a strong approach to derive document representation (Guo et al, 2019). We include it as another strong baseline in our experiments, with which we can analyze the gains brought by sentence weighting.…”
Section: Resultsmentioning
confidence: 99%
“…Averaging is a simple yet strong approach to represent compositional semantics. Strong performance comparable to supervised neural models has been achieved by sentence embeddings derived from average word embeddings (Arora et al, 2017) and document embeddings from average sentence embeddings (Guo et al, 2019) respectively.…”
Section: Weighted Document Representationmentioning
confidence: 99%
See 2 more Smart Citations