2006 1st International Conference on Communication Systems Software &Amp; Middleware
DOI: 10.1109/comswa.2006.1665153
|View full text |Cite
|
Sign up to set email alerts
|

A Measurement Study of the Linux TCP/IP Stack Performance and Scalability on SMP systems

Abstract: Abstract-The performance of the protocol stack implementation of an operating system can greatly impact the performance of networked applications that run on it. In this paper, we present a thorough measurement study and comparison of the network stack performance of the two popular Linux kernels: 2.4 and 2.6, with a special focus on their performance on SMP architectures. Our findings reveal that interrupt processing costs, device driver overheads, checksumming and buffer copying are dominant overheads of pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 15 publications
0
5
0
Order By: Relevance
“…Data destined for a thread running on CPU0 may be received by the kernel on CPU1 causing cache related slowdowns. Therefore, the dilation in TCP processing times seen in 64x2 run is very likely cache related ( [24] also found evidence of TCP/IP cache problems on SMP).…”
Section: Chiba Experimentsmentioning
confidence: 90%
“…Data destined for a thread running on CPU0 may be received by the kernel on CPU1 causing cache related slowdowns. Therefore, the dilation in TCP processing times seen in 64x2 run is very likely cache related ( [24] also found evidence of TCP/IP cache problems on SMP).…”
Section: Chiba Experimentsmentioning
confidence: 90%
“…iperf is a TCP/UDP-based network bandwidth measurement application that prints network bandwidth of the system. The major network overhead is interrupt processing cost, device driver overhead, checksumming, and buffer copying (the overhead of buffer copying is about 23%) [11]. Exploiting DMA reduces CPU overhead for memory copy, but it still suffers from throughput limitation originating from CPU-side datapath [50].…”
Section: Experimental Methodologymentioning
confidence: 99%
“…Data destined for a thread running on CPU0 may be received by the kernel on CPU1 causing cache related slowdowns. Therefore, the dilation in TCP processing times seen in 64x2 run is very likely cache related ( [19] also found TCP/IP cache problems on SMP).…”
Section: Chiba Experimentsmentioning
confidence: 95%