1: IntroductionHigh-performance servers are often implemented as shared-memory multiprocessor machines because of their flexibility in handling numerous data transactions simultaneously when transactions make use of data in orthogonal address spaces, while also efficiently supporting a coherency protocol when data is shared [1]. The performance of a shared-memory multiprocessor system often depends on the characteristics of the bus interconnect used to share data among the processors and memory modules in the system [1][2]. The ideal case is a fully-interconnected set of processors and memory modules. Interconnection between processors allows cache-to-cache transfers ("interventions") of modified data without waiting for memory access.However, the number of links in a fully-interconnected system grows proportionally to the square of the number of nodes [2]. In a VLSI environment using a synchronous bus interconnect, the number of pins and wires necessary to support a fully-connected bus structure becomes prohibitive. In addition, the increased number of connections lowers the overall yield of the interconnection and on-chip bus selection logic, thereby decreasing the overall faulttolerance of the system [3][4].When the network becomes something less than fullyinterconnected, such as in a ring or cube interconnect, the performance degrades because of the increasing pathlength of the communication between nodes [1]. Contention for fewer available communication points causes longer queueing delays. The bus utilization increases, and throughput suffers.Contention and queueing delays associated with processor interventions and memory access may prohibit the scalability of a multiprocessor system in which processors and memory share a common bus [1], generating diminishing performance returns as more processors are connected to the same bus structure. At the same time, the growing gap between processor speeds of over a gigahertz and synchronous interconnect speeds of a couple of hundred megahertz or less causes a larger percentage of system performance loss to be attributed to the memory subsystem and interconnect.There has been some recent interest in using high-bandwidth communications protocols to meet the interconnect needs of processing systems. One such technique is CodeDivision Multiple-Access (CDMA). CDMA is a spreadspectrum technique which encodes information prior to transmission onto a communications medium, permitting simultaneous use of the medium by separate information streams. In [5-6], a multiple-valued CDMA scheme is examined as an interconnect for parallel processing systems. In [7], CDMA on a fiber-optic interconnect operating at 10 gigahertz speeds is proposed as a local area network interconnect. In [8], CDMA is investigated as a means of more-fully interconnecting artificial neural network circuits since those circuits are known to suffer from a combinational explosion of nodes. The basic idea of these works is that pin counts and interconnect wiring can be drastically reduced by using ...