2014
DOI: 10.1162/coli_a_00198
|View full text |Cite
|
Sign up to set email alerts
|

Applications of Lexicographic Semirings to Problems in Speech and Language Processing

Abstract: This article explores lexicographic semirings and their application to problems in speech and language processing. Specifically, we present two instantiations of binary lexicographic semirings, one involving a pair of tropical weights, and the other a tropical weight paired with a novel string semiring we term the categorial semiring. The first of these is used to yield an exact encoding of backoff models with epsilon transitions. This lexicographic language model semiring allows for off-line optimization of e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
6
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 24 publications
0
6
0
Order By: Relevance
“…Russian, Spanish, Italian and Portuguese saw error rate reductions in the 50-60% range. While accuracy in Russian is slightly lower than Sproat et al [16] showed, our approach does not require any language-specific feature engineering.…”
Section: Grapheme Features In Stress Predictionmentioning
confidence: 69%
See 1 more Smart Citation
“…Russian, Spanish, Italian and Portuguese saw error rate reductions in the 50-60% range. While accuracy in Russian is slightly lower than Sproat et al [16] showed, our approach does not require any language-specific feature engineering.…”
Section: Grapheme Features In Stress Predictionmentioning
confidence: 69%
“…Sproat et al [16] show that models can be trained to predict stress placement given spelling, following Dou et al [17] who report numbers on both stand-alone stress prediction as well as joint phoneme and stress prediction accuracy. Dou et al [17] report numbers over four European languages and use the CELEX2 [18] lexicon.…”
Section: Grapheme-to-phoneme Predictionmentioning
confidence: 99%
“…As a consequence, much of the subsequent work on applying machine learning to text normalization for speech applications focuses on specific semiotic classes, like letter sequences(Sproat and Hall 2014), abbreviations(Roark and Sproat 2014), or cardinal numbers(Gorman and Sproat 2016). 4 In fact, Kestrel(Ebden and Sproat 2014) uses a machine-learned morphosyntactic tagger for Russian.…”
mentioning
confidence: 99%
“…Pruning could still be done for acronym pronunciation variants. A classification algorithm based on Maximum Entropy-based rankers has previously been implemented to determine whether acronyms are pronounced as a word, letter sequence, or a mix of both [23].…”
Section: Conclusion and Discussionmentioning
confidence: 99%