Proceedings of the Sixteenth ACM Conference on Conference on Information and Knowledge Management 2007
DOI: 10.1145/1321440.1321459
|View full text |Cite
|
Sign up to set email alerts
|

Automatic call section segmentation for contact-center calls

Abstract: This paper presents a SVM (Support Vector Machine) classification system which divides contact-center call transcripts into "Greeting", "Question", "Refine", "Research", "Resolution", "Closing" and "Out-of-topic" sections. This call section segmentation is useful to improve search and retrieval functions and to provide more detailed statistics on calls. We use an off-the-shelf automatic speech recognition (ASR) system to generate call transcripts from recorded calls between customers and service representative… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2008
2008
2021
2021

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 16 publications
0
10
0
Order By: Relevance
“…Note that in these approaches we assume utterances as basic units instead of the words. To convert the word sequence in the transcript into utterances we use utterance boundary detection algorithm described in [14]. The method is tuned to generate short utterances such that an utterance is unlikely to span two or more topics.…”
Section: Topic Segmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that in these approaches we assume utterances as basic units instead of the words. To convert the word sequence in the transcript into utterances we use utterance boundary detection algorithm described in [14]. The method is tuned to generate short utterances such that an utterance is unlikely to span two or more topics.…”
Section: Topic Segmentationmentioning
confidence: 99%
“…The data set used (same as in [14]) for experimental evaluation consists of automatic speech recognition transcripts of 100 calls that are conversations between agents and customers in a help-desk scenario. This data set comprises 13.2 hours of calls consisting of 5350 utterances.…”
Section: Databasementioning
confidence: 99%
“…Domain-specific words and acronyms are rendered with the expected capitalization patterns (e.g., "ABS"), as is the pronoun "I." Sentence boundaries are detected, using the methods described in [10], and the resulting sentences are given initial capitals and punctuated with periods.…”
Section: Token and Sentence Normalizationmentioning
confidence: 99%
“…An utterance boundary detector [10] divides the continuous stream of words from the recognition engine into normal sentences. The algorithm, based on Maximum Entropy classification, uses linguistic and prosodic features such as the probabilities of unigrams and bi-grams occurring as the first or last unigram (or bi-gram) in utterances, and the length of pauses between two words.…”
Section: Sentence and Segment Boundary Detectionmentioning
confidence: 99%
“…In such a scenario, summarization would play the role of a feature selection module to choose essential intent conveying part useful for further processing, discarding the irrelevant parts. A similar approach of using shorter segments of the conversations has been shown to be beneficial in a call classification task [4] where simply the initial part of the conversation is used. However, it still used large initial parts of the conversations assuming the relevant part is captured there.…”
Section: Introductionmentioning
confidence: 99%