2004
DOI: 10.1023/b:jiis.0000039534.65423.00
|View full text |Cite
|
Sign up to set email alerts
|

A Dynamic Programming Algorithm for Linear Text Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
48
0

Year Published

2005
2005
2021
2021

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 64 publications
(48 citation statements)
references
References 22 publications
0
48
0
Order By: Relevance
“…The studies using text were pioneered by the TextTiling approach [2], where adjacent sentence blocks were compared using a similarity measure based on bag-of-words (BOW) features, such as term frequency -inverted document frequency (tf-idf). Later studies indicated that globally optimized segmentation methodssuch as dynamic programming (DP) and the hidden Markov model (HMM) [3,4,14] -can improve the performance, and usage of probabilistic topic modeling such as probabilistic latent semantic analysis (pLSA) [15,7] and latent Dirichlet allocation (LDA) [16,17] can further increase the accuracy. Analogous to approaches used in automatic speech recognition (ASR), deep neural networks have been combined with HMMs (DNN-HMM) and successfully applied to story segmentation, using BOW features of text data, with significant improvement in performance [18].…”
Section: Introductionmentioning
confidence: 99%
“…The studies using text were pioneered by the TextTiling approach [2], where adjacent sentence blocks were compared using a similarity measure based on bag-of-words (BOW) features, such as term frequency -inverted document frequency (tf-idf). Later studies indicated that globally optimized segmentation methodssuch as dynamic programming (DP) and the hidden Markov model (HMM) [3,4,14] -can improve the performance, and usage of probabilistic topic modeling such as probabilistic latent semantic analysis (pLSA) [15,7] and latent Dirichlet allocation (LDA) [16,17] can further increase the accuracy. Analogous to approaches used in automatic speech recognition (ASR), deep neural networks have been combined with HMMs (DNN-HMM) and successfully applied to story segmentation, using BOW features of text data, with significant improvement in performance [18].…”
Section: Introductionmentioning
confidence: 99%
“…Choi (2000) introduced the probabilistic algorithm using matrix-based ranking and clustering to determine similarities between segments. Galley et al (2003) combined contentbased information with acoustic cues in order to detect discourse shifts whereas Utiyama and Isahara (2001) and Fragkou et al (2004) minimized different segmentation cost functions with dynamic programming.…”
Section: Related Workmentioning
confidence: 99%
“…An extensive discussion of precisely the same problem addressed here, but with a different approach to its solution, is in [3], [4]. Work by Hubert [10], [11], with applications to meteorology, influenced Kehagias and co-workers [8], [15], [16], [17], [18], [19], who developed a dynamic programming algorithm much like ours, for applications such as text segmentation (see also [9]), where the raw data are provided in the form of a similarity matrix. [22] gives an O(kN 2 ) dynamic programming algorithm for finding the optimal partition of an interval into k blocks, for a given k. See also [20], [2] for related work.…”
Section: Introduction: the Problemmentioning
confidence: 99%