2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07 2007
DOI: 10.1109/icassp.2007.366172
|View full text |Cite
|
Sign up to set email alerts
|

High Throughput Systolic SOM IP Core for Fpgas

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
14
0

Year Published

2010
2010
2020
2020

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(17 citation statements)
references
References 8 publications
0
14
0
Order By: Relevance
“…The Euclidean distance metric is usually employed in this calculation, however, as with other hardware classifier designs, the Manhattan (city block) distance may be used to save hardware [Manolakos and Logaras 2007]. Finally, the test vector is classified by assigning to it the label that is encountered most frequently among the labels of its k nearest neighbors.…”
Section: Architecture Designmentioning
confidence: 99%
“…The Euclidean distance metric is usually employed in this calculation, however, as with other hardware classifier designs, the Manhattan (city block) distance may be used to save hardware [Manolakos and Logaras 2007]. Finally, the test vector is classified by assigning to it the label that is encountered most frequently among the labels of its k nearest neighbors.…”
Section: Architecture Designmentioning
confidence: 99%
“…Specifically, if a stream of m query vectors is presented to the system for classification, element-by-element the one after the other, it takes overall T m cycles (1) to produce the results and the performance is P m operations/sec given by (2). As the number of input vectors that are pipelined increases, the performance P m becomes in the limit P b (3).…”
Section: Introductionmentioning
confidence: 99%
“…Afterwards, when a query vector Y is presented to the system for classification, its distance to each one of the reference vectors is calculated and the "k" smallest distances are found. The Euclidean distance metric is usually employed, but for a hardware implementation, as with other classifier designs, the Manhattan distance is often preferred, since it is simpler to realize [2]. Finally, the class label of the query vector is determined as the label that is encountered most times among its k nearest neighbor labels.…”
Section: Introductionmentioning
confidence: 99%
“…When an unseen test vector Y is presented for classification, its distance to each one of the stored training vectors is calculated and the "k" smallest distances are retained. The Euclidean distance metric is usually employed, however for a hardware implementation, as with other classified designs, the Manhattan distance is often preferred, since it is simpler to realize in hardware [7]. Finally, the class label of the test vector is determined as the one that is encountered most times among its k nearest neighbor labels.…”
Section: Introductionmentioning
confidence: 99%