Proceedings of the 2016 ACM Symposium on Document Engineering 2016
DOI: 10.1145/2960811.2960815
|View full text |Cite
|
Sign up to set email alerts
|

A Multimodal Crowdsourcing Framework for Transcribing Historical Handwritten Documents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0
1

Year Published

2016
2016
2020
2020

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 10 publications
(10 citation statements)
references
References 14 publications
0
9
0
1
Order By: Relevance
“…Regarding the crowdsourcing framework adjustment, in a previous work, we observed as in this system the speaker order and the reliability verification did not show a significant impact on the results. Besides, the highest reliability for this test set is obtained when the multimodal combination is a bit balanced to the speech output ( α =0.6) and the LM interpolation to the original LM ( λ =0.4).…”
Section: Experimental Conditionsmentioning
confidence: 49%
See 2 more Smart Citations
“…Regarding the crowdsourcing framework adjustment, in a previous work, we observed as in this system the speaker order and the reliability verification did not show a significant impact on the results. Besides, the highest reliability for this test set is obtained when the multimodal combination is a bit balanced to the speech output ( α =0.6) and the LM interpolation to the original LM ( λ =0.4).…”
Section: Experimental Conditionsmentioning
confidence: 49%
“…The initial multimodal crowdsourcing framework allowed us to use speech utterances for improving the transcription of historical manuscripts. This system was improved by adding a line selection module that allows us to optimize the collaboration effort (CE) .…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In a previous work [9] we observed that this crowdsourcing framework presents the highest reliability (for this corpus) when the multimodal combination is a bit balanced to the speech output (α = 0.6, with Θ = 10 −4 ), and the language model interpolation to the original model (λ = 0.4). We also noted that the speaker ordering and the reliability verification did not show a significant impact on the results.…”
Section: A Baseline and Framework Adjustmentmentioning
confidence: 73%
“…Works such as that of [4] reveal the feasibility of the acquisition of speech corpora by using mobile devices and the capacity of the crowdsourcing framework to obtain annotated speech corpora at several levels. In [9], a first step on the incorporation of multimodality in crowdsourcing is shown, by presenting a framework where the acquired modality (speech) is not the one to be transcribed (handwritten text).…”
Section: Related Workmentioning
confidence: 99%