2005
DOI: 10.1016/j.ijmedinf.2005.03.013
|View full text |Cite
|
Sign up to set email alerts
|

Effects of information and machine learning algorithms on word sense disambiguation with small datasets

Abstract: Summary Current approaches to word sense disambiguation use (and often combine) various machine learning techniques. Most refer to characteristics of the ambiguity and its surrounding words and are based on thousands of examples. Unfortunately, developing large training sets is burdensome, and in response to this challenge, we investigate the use of symbolic knowledge for small datasets. A naïve Bayes classifier was trained for 15 words with 100 examples for each. Unified Medical Language System (UMLS) semanti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
39
0

Year Published

2009
2009
2023
2023

Publication Types

Select...
4
3
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 43 publications
(40 citation statements)
references
References 13 publications
1
39
0
Order By: Relevance
“…This has already been shown in the literature [14-16]. The MFS indicates that usually one sense of the ambiguous word is highly represented compared to the rest of the senses.…”
Section: Resultssupporting
confidence: 65%
See 1 more Smart Citation
“…This has already been shown in the literature [14-16]. The MFS indicates that usually one sense of the ambiguous word is highly represented compared to the rest of the senses.…”
Section: Resultssupporting
confidence: 65%
“…Among the knowledge-based methods we find the Journal Descriptor Indexing method [12] and several based on graph algorithms [13]. Machine learning algorithms have been explored in several studies where alternative combinations of features are compared [14-16]; these studies obtain a performance of over 0.86 in terms of accuracy using the collection prepared by Weeber et al [17]. …”
Section: Introductionmentioning
confidence: 99%
“…We noticed such inconsistencies in prior work by us [30] and by others [19]. In the present testbed, we encountered a few omissions i.e., annotations that were identified by our parser but not by the human annotators.…”
Section: Parser Evaluationmentioning
confidence: 59%
“…In order to compare our method with other methods in verifying the efficiency of extracting coordinate relationship, we adopt the following two kinds of methods, Naïve Bayesian (NB) [20][21][22][23] and Support Vector Machine (SVM) [24][25][26][27]. The detailed descriptions for these two methods are as follows.…”
Section: B Comparison Methodsmentioning
confidence: 99%