GB2409073A - Dedicated connection between CPU and network interface in multi-processor systems - Google Patents
Dedicated connection between CPU and network interface in multi-processor systems Download PDFInfo
- Publication number
- GB2409073A GB2409073A GB0427107A GB0427107A GB2409073A GB 2409073 A GB2409073 A GB 2409073A GB 0427107 A GB0427107 A GB 0427107A GB 0427107 A GB0427107 A GB 0427107A GB 2409073 A GB2409073 A GB 2409073A
- Authority
- GB
- United Kingdom
- Prior art keywords
- data
- network interface
- cpu
- network
- compute node
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000004891 communication Methods 0.000 claims abstract description 81
- 239000000872 buffer Substances 0.000 claims description 31
- 238000000034 method Methods 0.000 claims description 27
- 230000008878 coupling Effects 0.000 claims 1
- 238000010168 coupling process Methods 0.000 claims 1
- 238000005859 coupling reaction Methods 0.000 claims 1
- 238000012546 transfer Methods 0.000 abstract description 13
- 238000005516 engineering process Methods 0.000 description 16
- 230000008569 process Effects 0.000 description 14
- 230000002093 peripheral effect Effects 0.000 description 13
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 8
- 229910052802 copper Inorganic materials 0.000 description 8
- 239000010949 copper Substances 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 5
- 239000013307 optical fiber Substances 0.000 description 5
- 238000013461 design Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 3
- 238000005457 optimization Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000011664 signaling Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000007723 transport mechanism Effects 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F13/00—Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
- G06F13/38—Information transfer, e.g. on bus
- G06F13/382—Information transfer, e.g. on bus using universal interface adapter
- G06F13/385—Information transfer, e.g. on bus using universal interface adapter for adaptation of a particular data processing system to different peripheral devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
- G06F15/163—Interprocessor communication
- G06F15/173—Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star, snowflake
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
Compute nodes in a high performance computer system are interconnected by an inter-node communication network. Each compute node has a network interface coupled directly to a CPU by a dedicated full-duplex packetized interconnect. Data may be exchanged between compute nodes using eager or rendezvous protocols. The network interfaces may include facilities to manage data transfer between computer nodes.
Description
) 2409073
DIRECTLY CONNECTED LOW LATENCY NETWORK AND
INTERFACE
Reference to Related Applications
1] This application claims priority from U.S. patent application Nos. 60/528,774 entitled "DIRECTLY CONNECTED LOW LATENCY NETWORK", filed l2 December 2003; 60/531,999 entitled "LOW
LATENCY NETWORK WITH DIRECTLY CONNECTED
INTERFACE", filed 24 December 2003; and 10/788,455 entitled "DIRECTLY CONNECTED LOW LATENCY NETWORK AND INTERFACE", filed l March 2004.
Technical Field
l0002l This invention relates to multiprocessor computer systems.
:.: . l 5 In particular, the invention relates to communication networks for as..
c. exchanging data between processors in multiprocessor computer systems. a..e c
:. Background
: 10003] Multiprocessor, high performance computers (e.g. .e 1 . 20 supercomputers) are often used to solve large complex problems. Figure l shows schematically a multiprocessor computer 10 having compute nodes 20 connected by a communication network 30. Software applications running on such computers split large problems up into smaller sub- problems. Each sub-problem is assigned to one of compute nodes 20. A program is executed on one or more CPUs of each compute node 20 to solve the sub-problem assigned to that compute node 20. The program run on each compute node has one or more processes.
Executing each process involves executing a sequence of software - 2 instructions. All of the processes execute concurrently and may communicate with each other.
[00041 Some problems cannot be split up into sub-problems which are independent of other sub-problems. In such cases, to solve at least some of the sub-problems, an application process must communicate with other application processes that are solving related sub-problems to exchange intermediate results. The application processes cooperate with each other to obtain a solution to the problem.
10005] Communication between processes solving related sub-problems often requires the repeated exchange of data. Such data exchanges occur frequently in high performance computers. a.
:.: ' Communication performance in terms of bandwidth, and especially . ee ëa 15 latency, are a concern. Overall application performance is, in many me cases, strongly dependent on communication latency. e
e 10006] Communication latency has three major components: ease eerie the latency to transfer a data packet from a CPU or other device in a sending compute node to a communication network; the latency to transfer a data packet across the communication network; and, the latency to transfer a data packet from the communication network to a device such as a CPU in a receiving compute node.
10007] In order to reduce latency, various topologies (e.g. hypercube, mesh, toroid, fat tree) have been proposed and/or used for - 3 interconnecting compute nodes in computer systems. These topologies may be selected to take advantage of communication patterns expected for certain types of high performance applications. These topologies often require that individual compute nodes be directly connected to multiple other compute nodes.
100081 Continuous advances have been made over the years in communication network technology. State of the art communication networks have extremely high bandwidth and very low latency. The l O inventors have determined that available communication network technology is not necessarily a limiting factor in improving the performance of high performance computers as it once was. Instead, the performance of such computers is often limited by currently accepted techniques used to transfer data between CPUs and associated network l 5 interfaces and the network interfaces themselves. The following description explains various existing computer architectures and provides i. the inventors' comments on some of their shortcomings for use in high A..
performance computing.
e...e [00091 Figure 2 shows how early computers, and even some modern computers, support data communication. A CPU 100 is connected to memory and peripherals using an address and data bus 160.
Address and data bus l 60 combines a parallel address bus and a parallel data bus. Memory 110, video display interface 120, disk interface 130, network interface 140, keyboard interface 150, and any other peripherals are each connected to address and data bus 160. Bus 160 is shared for all communication between CPU Leo and all other devices in the computer. to 4
Bandwidth and latency between CPU 100 and network interface 140 are degraded because network interface 140 must compete with memory and all the other peripherals for use of bus 160. Further, hardware design considerations limit the rate at which data can be carried over an address and data bus.
10010] CPU speeds have increased over the years. It is increasingly difficult to directly interface high-speed CPUs to low-speed peripherals.
This led to the computer architecture shown in Figure 3 in which CPU 200 is connected by a high-speed front side bus (FSB) 240 to north bridge chip 230. North bridge 230 provides an interface to memory 210 and to high-speed peripherals such as video display interface 220. In modern personal computers, an AGP interface is used between north bridge 230 and video display interface 220. A variety of interfaces (e.g. SDRAM, DDR, RAMBUS_) have been used to interface memory 210 to north bridge 230. . .e A..
[00111 Low-speed peripherals such as keyboard 250, mouse 260, : . and disk 270 are connected to south bridge chip 280. South bridge 280 is : 20 connected to north bridge 230 via a medium- to high-speed bus 290.
South bridge 280 will often support an I/O bus 310 (e.g. ISA, PCI, : PCI-X) to which peripheral cards can be connected. Network interfaces (e.g. 300) are connected to I/O bus 310.
2] Some vendors have implemented I/O bus 310 in north bridge 230 instead of south bridge 280 and some have used I/O bus technology for both bus 310 and bus 290. - s -
[0013l Modern designs involving north bridges and south bridges are still very poor for high performance data communication. While the north bridge can accommodate higher speed FSB 240, network interface 300 shares FSB 240 with memory 210 and all other peripherals. In addition, network traffic must now traverse both north bridge 230 and south bridge 280. I/O bus 310 is still shared between network interface 300 and any other add-in peripheral cards.
100141 Some designs exacerbate the above problems. These l O designs connect more than one CPU 200 (e.g. two or four) to FSB 240 to create twoway or four-way shared memory processors (SMPs). All of the CPUs must contend for FSB 240 in order to access shared memory 210 and other peripherals.
l 5 lO015] Another limitation of existing architectures is that there are technical impediments to significantly increasing the speed at which . front side buses operate. These buses typically include address and data buses each consisting of many signal lines operating in parallel. As : speed increases, signal skew and crosstalk reduce the distance that these buses can traverse to a few inches. Signal reflections from terminations on multiple CPUs and the north bridge adversely affect bus signal :: quality.
6] A few vendors (e.g. AMD and Motorola) have started to make CPUs having parallel interconnects which have a reduced number of signal lines (reduced-parallel interconnects) or serial system interconnects. These interconnects use fewer signal lines than parallel o - 6 address and data buses, careful matching of signal line lengths, and other improvements to drive signals further at higher speeds than can be readily provided using traditional FSB architectures. Current high performance interconnects typically use Low Voltage Differential Signaling (LVDS) to achieve higher data rates and reduced electromagnetic interference (EMI). These interconnects are configured as properly terminated point-to-point links and are not shared in order to avoid signal reflections. Such serial and reduced-parallel interconnects typically operate at data rates that exceed 300 MBps (megabytes per 1 0 second).
10017] Examples of such interconnects include HyperTransport_, RapidIO_, and PCI Express. Information about these interconnects can be found at various sources including the following: HyperTransport I/O Link Specification, HyperTransport Consortium, lttp://www.hypertransport.org/ . :.: . RapidIO Interconnect Specification, RapidIO Trade Association, e..
http://www.rapidio.org/ RapidIO Interconnect GSM Logical Specification, RapidIO Trade Association, http://www.rapidio.org/ RapidIO Serial Physical Layer Specification, RapidIO Trade Association, http://www.rapidio.org/ RapidIO System and Device Inter-operability Specification, RapidIO Trade Association, http://www.rapidio.org/ PCI Express Base Specification, PCI-SIG, http://www.pcisig.com/ PCI Express Card Electromechanical Specification, PCI-SIG, http://www. pcisig.com/ 7
PCI Express Mini Card Specification, PCI-SIG,
http://www.pcisig.com/ 10018] Because the number of signal lines in a serial or reduced parallel interconnect is less than the width of data being transferred, it is not possible to transfer data over such interconnects in a single clock cycle. Instead, both serial and reduced-parallel interconnects package and transfer data in the form of packets.
10019] These interconnects can be operated using protocols which use memory access semantics. Memory access semantics associate a source or destination of data with an address which can be included in a packet. Read request packets contain an address and number of bytes to be fetched. Read response packets return the requested data. Write request packets contain an address and data to be written. Write confirmation packets optionally acknowledge the completion of a write. ..
:.: . The internal structure of individual packets, the protocols for exchanging ..
packets and the terminology used to describe packets differ between the A. various packetized interconnect technologies.
: [00201 Interconnects which use memory access semantics including packetized parallel interconnects having a number of signal lines which is smaller than a width of data words being transferred and packetized serial interconnects are referred to collectively herein as "packetized interconnects". The term "packetized interconnects" has been coined specifically for use in this disclosure and is not defined by any existing usage in the field of this invention. For example, packetized interconnect - 8 is not used herein to refer to packet-based data communication protocols (e.g. TCP/IP) that do not use memory access semantics.
[00211 An important side effect of using an interconnect which has a reduced number of signal lines is that it is possible to connect multiple packetized interconnects to one CPU. For example, one model of AMD Opteron_ CPU terminates three instances of a packetized interconnect (i.e. HyperTransport_). A few CPUs (e.g. the AMD Opteron_) combine the use of packetized interconnects with a traditional address O and data bus which is used for access to main memory.
10022] The computer architecture of Figure 4 uses a CPU which connects to peripherals by a packetized interconnect. CPU 420 is directly connected to memory 400 by a traditional, parallel, address and data bus 410. CPU 420 is directly connected to a video display interface 430, a south bridge 440, and an I/O interface 450 via packetized ea interconnects 460. Keyboard 480 and mouse 490 are connected to south e..
bridge 440. I/O interface 450 connects packetized interconnect 460 to a he traditional I/O bus 510 (e.g. PCI, PCI-X). Network interface 500 is .
connected to I/O bus 510. : .e
: : [00231 The architecture of Figure 4 provides some benefits relative to earlier architectures. Peripheral cards such as network interface 500 no longer have to share a FSB with memory. They have exclusive use of one instance of packetized interconnect 460 to communicate with CPU 420. The inventors have recognized that the architecture of Figure 4 still has the following problems: - 9 - Network interface 500 must share I/O bus 510 with all other add-in peripheral cards; and, Latency is increased because data passing in either direction between CPU 420 and network interface 500 must traverse I/O interface 450.
[00241 Despite the various architectural improvements, the aforementioned architectures still have a serious problem with regards to the high bandwidth, low latency data communication that is required by high performance computer systems. Data packets are forced to traverse a traditional I/O bus 510 in the process of being transferred between CPU 420 and network interface 500. Because bus 510 uses a common address and data bus to transfer data back and forth between devices, bus 510 operates in half duplex mode. Only one device can transfer data at a time (e.g. network interface 500 to I/O interface 450 or I/O interface 450 to network interface 500). In contrast, packetized interconnects and most . .
modern communication network data links operate in full duplex mode I..
with separate transmit and receive signal lines. e.e
* 20 10025] In Figure 5, which corresponds to the architecture shown in Figure 4, it can be seen that I/O interface 450 must convert between full : ': duplex packetized interconnect 460 and half-duplex I/O bus 510.
Similarly, network interface 500 must convert between half-duplex I/O bus 510 and full-duplex communication data link 520. Converting between halfduplex and full-duplex transmission decreases communication performance. Unless the half-duplex bandwidth of bus 510 is equal to or greater than the sum of the bandwidth in each direction - 10 on interconnect 460, the full bandwidth of interconnect 460 cannot be utilized. Similar reasoning shows that the full bandwidth of communication link 520 cannot be exploited unless the half-duplex bandwidth of bus 510 is equal to or greater than the sum of the bandwidth in each direction on communication link 520.
[00261 As an example, if HyperTransport_ is used to implement packetized interconnect 460, it can be operated at a rate of 25.6 Gbps (Gigabits per second) in each direction for an aggregate bi-directional l O bandwidth of 51.2 Gbps. Similarly, if InfniBand_ 4X or l OGigE technology were used to implement data link 520, the data link could support a bandwidth of l O Gbps in each direction for an aggregate bi- directional bandwidth of 20 Gbps. In contrast, 64 bit wide PCI-X operating at 133 MHz can only support a half duplex bandwidth of 8.5 Gbps. In this example the PCI-X I/O bus provides a bottleneck.
100271 Because I/O bus 510 can only transmit in one direction at a time, packets may have to be queued in either I/O interface 450 or : : network interface 500 until bus 510 can be reversed to support communication in the desired direction. This can increase latency : unacceptably for some applications. For example, consider a packet with : : a size of l OOO bytes that is being transferred from network interface 500 over a PCI-X bus 510 having the aforementioned characteristics to I/O interface 450. If a packet arrives at 1/0 interface 450 from CPU 420, it may be necessary to queue the packet at I/O interface 450 for up to 0.94 microseconds. - 11
10028] High performance computers can ideally transfer a data packet from a CPU in one compute node to a CPU in another compute node in 3 microseconds or less. Where a 1000 byte packet has to be queued to use the half duplex I/O bus in each compute node, it is conceivable that as much as 1.88 microseconds might be spent waiting.
This leaves very little time for any other communication delays. Moving beyond the status quo, high performance computing would benefit greatly if communication latencies could be reduced from 3 microseconds to l microsecond or better.
10029] Network interfaces present other problem areas. As speeds of data communication networks have increased there has been a trend to move away from copper-based cabling to optical fibers. For example, copper-based cabling is used for 10 Mbps (megabits per second), 100 Mbps, and 1 Gbps Ethernet. In contrast, 10 Gbps Ethernet currently requires optical fiberbased cabling. A single high performance computer . :: system may require a large number of cables. As an example, a product .e under development by the inventors terminates up to 24 data links. The : . product can be configured in various ways. For example, the product may : 20 used to construct a 1000 compute node high performance computer with : a fat tree topology communication network. Some configurations use up .
to 48,000 connections between different compute nodes. If a separate cable were used for each connection then 48,000 cables would be required. The cost of cables alone can be significant.
[00301 Optical fiber-based cabling is currently significantly more expensive than copper-based cabling. Network interface terminations for 12 optical fiber-based cabling are currently significantly more expensive than terminations for copper-based cabling. As mentioned previously, high performance computers often have to terminate multiple communication network data links. Providing cables and terminations for large numbers of optical fiber-based data links can be undesirably expensive.
10031] Of the few high speed communication network technologies that use copper-based cabling, most are undesirably complicated for high performance computing. These technologies have been implemented to satisfy the wide variety of requirements imposed by enterprise data centers.
10032] One such communication network technology is InfiniBand_. InfiniBand_ was developed for use in connecting computers to storage devices. Since then it has evolved, and its feature ë set has expanded. InfiniBand_ is now a very complicated, feature rich -..
technology. Unfortunately, InfiniBand_ technology is so complex that A--: it is ill suited for use in communication networks in high performance : computingOhio State University discovered that a test communication .
: network based on InfiniBand_ had a latency of 7 microseconds. While technical improvements can reduce this latency, it is too large for use in high performance computing.
[00331 There remains a need in the supercomputing field for a cost effective and practical communication network technology that provides dedicated high bandwidth, and low latency. of - 13
Brief Description of the Drawings
100341 In drawings which illustrate non-limiting embodiments of the invention: Figure 1 is a schematic illustration of the architecture of a prior art multiprocessor computer system; Figure 2 is a block diagram illustrating the architecture of early
and certain modern prior art personal computers;
Figure 3 is a block diagram illustrating the architecture of most
modern prior art personal computers;
Figure 4 is a block diagram illustrating an architecture of a state ofthe-art personal computer having CPUs connected to other devices by packetized interconnects; Figure 5 is a block diagram illustrating a data communication path in a state of the art computer system having a CPU connected to other devices by a packetized interconnect; Figure 6 is a block diagram illustrating a computer system ë- according to an embodiment of the invention having a network interface -- directly connected to a CPU via a packetized interconnect dedicated to .
data communication; . . : Figure 7 is a block diagram illustrating a data communication path : in a compute node that implements the invention; Figure 8 is a diagram illustrating layers in a communication protocol; and, Figures 9 and 10 are block diagrams illustrating data communication paths in a computer system according to the invention; and, - 14 Figures 1 1 to 13 illustrate a network interface.
5] Various aspects of the invention and features of specific embodiments of the invention are described below.
Description
10036] Throughout the following description, specific details are set forth in order to provide a more thorough understanding of the invention. However, the invention may be practiced without these particulars. In other instances, well known elements have not been shown or described in detail to avoid unnecessarily obscuring the invention. Accordingly, the specification and drawings are to be regarded in an illustrative, rather than a restrictive, sense.
[00371 This invention exploits the packetized interconnects as provided, for example, by certain state of the art CPUs to achieve low ..
latency data communication between CPUs in different compute nodes of A. a computer system. A CPU has at least one packetized interconnect I..
dedicated to data communication. This provides guaranteed bandwidth :.
for data communication. A network interface is attached directly to the .
CPU via the dedicated packetized interconnect. Preferably the .
packetized interconnect and a communication data link to which the network interface couples the packetized interconnect both operate in a full-duplex mode.
10038] In some embodiments of the invention the communication network uses a communication protocol based on InfiniBand_. In some - 15 cases the communication protocol is a simplified communication protocol which uses standard InfiniBand_ layers l and 2. A high- performance computing-specific protocol replaces InfiniBand_ layers 3 and above.
10039] A computer system according to a preferred embodiment of the invention is shown in Figure 6. Figure 6 shows only 2 compute nodes 20A and 20B (collectively compute nodes 20) for simplicity. A computer system according to the invention may have more than two compute l O nodes. Computer systems according to some embodiments of the invention have l DO or more compute nodes. Computer systems according to the invention may have 500 or more, 1000 or more, or 5,000 or more compute nodes. Some advantages of the invention are fully realized in computer systems having many (i.e. l OO or more) interconnected compute nodes. ë . .
0] CPU 610 is connected to memory 600 using interconnect ë 620. Interconnect 620 may comprise a traditional parallel address and . ee.
data bus, a packetized interconnect or any other suitable data path which :e allows CPU 610 to send data to or receive data from memory 600. .
Memory 600 may include a separate memory controller or may be ë controlled by a controller which is integrated with CPU 610. A packetized interconnect 640 attached to CPU 610 is dedicated to data communication between CPU 610 and a network interface 630. Apart from CPU 610 and network interface 630, no device which consumes a significant share of the bandwidth of packetized interconnect 640 or injects traffic sufficient to increase latency of interconnect 640 to any o - 16 significant degree shares packetized interconnect 640. In this case, a significant share of bandwidth is 5% or more and a significant increase in latency is 5% or more. In preferred embodiments of the invention no other device shares packetized interconnect 640.
lO041] Network interface 630 is directly attached to CPU 610 via packetized interconnect 640. No gateway or bridge chips are interposed between CPU 610 and network interface 630. The lack of any gateway or bridge chips reduces latency since such chips, when present, take time l O to transfer packets and to convert the packets between protocols.
2] Packetized interconnect 640 extends the address space of CPU 610 out to network interface 630. CPU 610 uses memory access semantics to interact with network interface 630. This provides an efficient mechanism for CPU 610 to interact with network interface 630. . 8'
00431 Referring now to Figure 7 which corresponds to the architecture shown in Figure 6, full duplex packetized interconnect 640 868..
A. is directly interfaced to full duplex communication data link 650. The receive signal lines of interconnect 640 (relative to network interface 630) are interfaced to the transmit signal lines of data link 650.
8 (8 Similarly, the receive signal lines of data link 650 are interfaced to the transmit signal lines of interconnect 640.
4] Since network interface 630 directly connects the two full duplex links 640 and 650 together, interface 630 can be constructed so that there is no bandwidth bottleneck. If communication data link 650 is o - 17 slower than packetized interconnect 640, the full bandwidth of link 650 can be utilized. If packetized interconnect 640 were slower instead, the full bandwidth of interconnect 640 could be utilized.
5] Directly connecting full duplex links 640 and 650 together also eliminates queuing points as would be required at a transition between full duplex and half duplex technologies. This eliminates a major source of latency. The only queuing point that remains is the transition from the faster technology to the slower technology. For lO example, if packetized interconnect 640 is faster than communication data link 650, a queuing point is provided in the direction and at the location in network interface 630 where outgoing data packets are transferred from packetized interconnect 640 to data link 650. Such a queuing point handles the different speeds and bursts of data packets. If the two technologies implement flow control, packets will not normally . queue at this queuing point. . see. S..
6] In embodiments of the invention wherein packetized e. ease interconnect 640 and communication data link 650 are both full-duplex network interface 630 can be simplified. In such embodiments network interface 630 need only transform packets from the packetized Seneca e interconnect protocol to the communication data link protocol and vice versa in the other direction. No functionality need be included to handle access contention for a half duplex bus. As mentioned above, queuing can be removed in one direction. Simple protocols may be used to manage the flow of data between CPU 610 and communication network 30. The result of these simplifications is that network interface 630 is - 18 less expensive to implement and both latency and bandwidth can be further improved.
10047] A single CPU can be connected to multiple network S interfaces 630. If multiple packetized interconnects 640 are terminated on a single CPU and are available, each such packetized interconnect 640 may be dedicated to a different network interface 630. A compute node may include multiple CPUs which may each be connected to one or more network interfaces by one or more packetized interconnects. If network l O interface 630 is capable of handling the capacity of multiple packetized interconnects, it may terminate multiple packetized interconnects 640 originating from one or more CPUs.
8] It will usually be the case that packetized interconnect 640 is faster than communication data link 650. The shorter distances traversed , by packetized interconnects allow higher clock speeds to be achieved. If - . . . the speed of a packetized interconnect 640 is at least some multiple N. of the speed of a data link 650 (where N is an integer and N> 1), network interface 630 can terminate up to N communication data links. Even if a packetized interconnect 640 is somewhat less than N times faster than a communication data link 650, network interface 630 could still terminate ..e * N communication data links with little risk that packetized interconnect 640 will be unable to handle all of the traffic to and from the N communication data links. There is a high degree of probability that not all of the communication data links will be simultaneously fully utilized. to - 19
[0049l Network interface 630 preferably interfaces packetized interconnect 640 to a communication protocol on data link 650 that is well adapted for high performance computing (HPC). Preferred embodiments of the inventionuse a communication protocol that supports copper-based cabling to lower the cost of implementation.
100501 Figure 8 shows a protocol stack for a HPC communication protocol that is used in some embodiments of the invention. The communication protocol uses the physical layer and link layer from InfiniBand_. The complex upper layers of InfiniBand_ are replaced by a special-purpose protocol layer designated as the HPC layer. The HPC layer supports an HPC protocol. One or more application protocols use the HPC protocol. Examples of application protocols include MPl, PVM, SHMEM, and global arrays.
0051 l The InfiniBand_ physical layer supports copper-based cabling. Optical fiber-based cabling may also be supported. Full duplex . . transmission separates transmit data from receive data. LVDS and a .
limited number of signaling lines (to improve skew, etc.) provide high speed communication. :. ë
: 10052] The InfiniBand_ link layer supports packetization of data, source and destination addressing, and switching. Where communication links 650 implement the standard InfiniBand_ link layer, commercially available InfiniBand_ switches may be used in communication network 30. In some embodiments of the invention the link layer supports packet corruption detection using cyclic redundancy checks (CRCs). The link of - 20 layer supports some capability to prioritize packets. The link layer provides flow control to throttle the packet sending rate of a sender.
3] The HPC protocol layer is supported in an InfiniBand_ standardcompliant manner by encapsulating HPC protocol layer packets within link layer packet headers. The HPC protocol layer packets may, for example, comprise raw ethertype datagrams, raw IPv6 datagrams, or any other suitable arrangement of data capable of being carried within a link layer packet and of communicating HPC protocol layer information.
4] The HPC protocol layer supports messages (application protocol layer packets) of varying lengths. Messages may fit entirely within a single link layer packet. Longer messages may be split across two or more link layer packets. The HPC protocol layer automatically segments messages into link layer packets in order to adhere to the Maximum Transmission Unit (MTU) size of the link layer. . . .
100551 The HPC protocol layer directly implements eager and ë rendezvous protocols for exchanging messages between sender and receiver. Uses of eager and rendezvous protocols in other contexts are .
known to those skilled in the art. Therefore, only summary explanations . . : of these protocols are provided here. ë
6] The eager protocol is used for short messages and the rendezvous protocol is used for longer messages. Use of the eager or rendezvous protocol is not necessarily related to whether a message will fit in a single link layer packet. By implementing eager and rendevous Go) - 21 protocols in the HPC protocol layer, a higher degree of optimization can be achieved. Some embodiments of the invention provide hardware acceleration of the eager and/or rendevous protocols.
[00571 Figure 9 shows the flow of messages in an eager protocol transaction. A sender launches a message toward a receiver without waiting to see if a receiving application process has a buffer to receive the message. The receiving network interface receives the message and directs the message to a separate set of buffers reserved for the eager protocol. These are referred to herein as eager protocol buffers. When the receiving application process indicates it is ready to receive a message and supplies a buffer, the previously-received message is copied from the eager protocol buffer to the supplied application buffer.
I S 10058] As an optimization, the receiving network interface may send the received message directly to a supplied application buffer, bypassing the eager protocol buffers, if the receiving application has ë previously indicated that it is ready to receive a message. The eager . e.
protocol has the disadvantage of requiring a memory-to-memory copy for e at least some messages. This is compensated for by the fact that no ë overhead is incurred in maintaining coordination between sender and . . receiver. a:
9] Figure 10 shows how the rendezvous protocol is used to transmit a long message directly between buffers of the sending and receiving application processes. A sending application running on CPU 610 instructs network interface 630 to send a message and provides the - 22 size of the message and its location in memory 600. Network interface 630 sends a short Ready-To-Send (RTS) message to network interface 730 indicating it wants to send a message. When the receiving application process running on CPU 710 is ready to receive a message, it informs network interface 730 that it is ready to receive a message. In response, network interface 730 processes the Ready-To-Send message and returns a short Ready-To-Receive (RTR) message indicating that network interface 630 can proceed to send the message. The RTR message provides the location and the size of an empty message buffer in l O memory 700. Network interface 630 reads the long message from memory 600 and transmits the message to network interface 730.
Network interface 730 transfers the received long message to memory 700 directly into the application buffer supplied by the receiving application.
0] When network interface 630 has completed sending the long message, it sends a short Sending-Complete (SC) message to network . interface 730. Network interface 730 indicates that a message has been received to the receiving application running in CPU 710. The Ready-To-Send, Ready-To-Receive, and Sending-Complete messages may be transferred using the eager protocol and are preferably generated A. automatically and processed by network interfaces 630 and 730. As a less preferable alternative, software running on CPUs 610 and 710 can control the generation and processing of these messages. The rendezvous protocol has the disadvantage of requiring three extra short messages to be sent, but it avoids the memory-to-memory copying of messages. go - 23
1] HPC communication should ideally be readily scalable to tens of thousands of CPUs engaged in all-to-all communication patterns.
Conventional transport layer protocols (e.g. the InfiniBand_ transport layer) do not scale well to the number of connections desired in high performance computer systems. In such transport layer protocols, each connection has an elaborate state. Each message must pass through work queues (queue pairs in InfiniBand_). Elaborate processing is required to advance the connection state. This leads to excessive memory and l O CPU time consumption.
[00621 The HPC protocol layer may use a simplified connection management scheme that takes advantage of direct support for the eager and rendezvous protocols. Each receiver allocates a set of eager protocol buffers. During connection establishment, a reference to the allocated set of eager protocol buffers is provided by the receiver to the sender. The sender references these buffers in any eager protocol messages in order . e.
to direct the message to the correct receiving application process. Since A. the eager protocol is also used to coordinate the transfer of messages by - the rendezvous protocol, it is unnecessary for the connection to be used . . to manage the large rendezvous protocol messages. :
- .e .e [0063] As a variant, it is possible for a single larger set of eager protocol buffers to be shared by a single receiving application amongst multiple connections. In such embodiments each connection would require a control data structure to record the identities of the buffers of - 24 associated with the connection. This variant reduces the memory usage further at the receiver, but incurs extra processing overhead.
[00641 Conventional transport layer protocols support reliable transport of messages separately for each connection. This adds to the connection state information. In contrast, the HPC protocol layer supports reliable transport between pairs of CPUs. All connections between a given pair of CPUs share the same reliable transport mechanism and state information. Like conventional transport layer protocols, the HPC reliable transport mechanism is based on acknowledgment of successfully received messages and retransmission of lost or damaged messages.
100651 Memory protection keys may be used to protect the I S receiver's memory from being overwritten by an erroneous or malicious sender. The memory protection key incorporates a binary value that is associated with that part of the receiver's memory which contains ë.
message buffers for received messages. During connection setup, a memory protection key corresponding to the set of eager protocol buffers .....
is provided to the sender. Memory protection keys may thereafter be :.
provided to the sender for the message buffers supplied by the receiving . . application for rendezvous protocol long messages. A sender must ë -e provide a memory protection key with each message. The receiving network interface verifies the memory protection key against the targeted message buffers before writing the message into the buffer(s). The generation and verification of memory protection keys may be performed automatically. - 25
6] Network interface 630 implements the functions of terminating a packetized interconnect, terminating a communication protocol, and converting packets between the packetized interconnect and communication network technologies.
[00671 For example, in a specific embodiment, network interface 630 implements the physical layer of InfiniBand_ (see Figure 1 1) by terminating an InfiniBand_ 1X, 4X, or 1 2X data link. For copper-based cabling, the data link carries data respectively over sets of 1, 4, or 12 sets (lanes) of four wires. Within a set of four wires, two wires form a transmit LVDS pair and two wires form a receive LVDS pair.
10068] Network interface 630 may also byte stripe all data to be transmitted across the available lanes, pass data through an encoder (e.g. an 8 bit to 10 bit (8b/lOb) encoder), serialize the data, and transmit the A. data by the differential transmitter using suitable encoding (e.g. NRZ ë encoding). All data is received by a differential receiver, deserialized, passed through a 10 bit to 8 bit decoder, and un-striped from the A..
available data lanes. :. :
9] Network interface 630 implements the link layer of ...e InfniBand_ (see Figure 12). Network interface 630 may prioritize, packets prior to transmission. Plow control prevents packets from overflowing the buffers of receiving network interfaces. A CRC is generated prior to transmission and verified upon receipt. - 26
10070] Network interface 630 implements the HPC protocol layer (see Figure 13). Amongst other functions performed by the network interface, memory protection keys are generated for memory buffers that are to be exposed by receivers to senders. Memory protection keys are verified on receipt of messages. The network interface automatically selects and manages the eager and rendezvous protocols based on message size. Packets are fragmented and defragmented as needed to ensure that they fit within the link layer MTU size. The network interface ensures that messages are reliably transmitted and received.
[00711 As will be apparent to those skilled in the art, Figures 1 1, 12, and 13 are illustrative in nature. There are many different ways in which the functions of a network interface can be organized in order to get an equivalent result. Network interfaces according to the invention may not provide all of these functions or may provide additional functions. ë .. .
10072] In a preferred embodiment of the invention, network e interface 630 is implemented as an integrated circuit (e.g. ASIC, FPGA) Bee for maximum throughput and minimum latency. Network interface 630 directly implements a subset or all of the protocols of packetized : interconnect 640 in hardware for maximum performance. Network B...
interface 630 directly implements a subset or all of the protocols of communication data link 650 in hardware for maximum performance.
Network interface 630 may implement the InfiniBand_ physical layer, the InfiniBand_ link layer, and the HPC protocol in hardware. o - 27
Application level protocols are typically implemented in software but may be implemented in hardware in appropriate cases.
[00731 CPUs 610 and 710 use memory access semantics to interact with network interfaces 630 and 730. CPU 610 can send a message in one of two ways. It can either write the message directly to address space that is dedicated to network interface 630. This will direct the message over packetized interconnect 640 to network interface 630 where it can be transmitted over communication network 30.
4] In the alternative a message may be stored in memory 600.
CPU 610 can cause network interface 630 to send the message by writing the address of the message in memory 600 and the length of the message to network interface 630. Network interface 630 can use DMA techniques to retrieve the message from memory 600 for sending at the same time as CPU 610 proceeds to do something else. For receipt of long messages under the rendezvous protocol, CPU 710 writes the address and length of application buffers to network interface 730. Both CPUs 610 and 710 write directly to network interfaces 630 and 730 to initialize end configure them. : A..
5] Where a component (e.g. a software module, CPU, interface, node, processor, assembly, device, circuit, etc.) is referred to above, unless otherwise indicated, reference to that component (including a reference to a "means") should be interpreted as including as equivalents of that component any component which performs the function of the described component (i.e., that is functionally equivalent), including - 28 components which are not structurally equivalent to the disclosed structure which performs the function in the illustrated exemplary embodiments of the invention.
6] As will be apparent to those skilled in the art in the light of the foregoing disclosure, many alterations and modifications are possible in the practice of this invention without departing from the spirit or scope thereof. .. . . A.
- .e e :e ë A.
A ( - J - 29
Claims (2)
- WHAT IS CLAIMED IS: 1. A method for communicating data from a firstcompute node of a computer system comprising multiple compute nodes interconnected by an inter-node communication network to a second one of the multiple compute nodes, the method comprising: placing the data on a full-duplex packetized interconnect directly connecting a CPU of the first compute node to a network interface connected to the inter-node communication network; receiving the data at the network interface; and, transmitting the data to a network interface of the second compute node by way of the inter-node communication network.
- 2. network providing at least one full-duplex data link to the network interface of each of the nodes. a.... I:eeece2. A method according to claim 1 wherein the network interface and the CPU are the only devices configured to place data on the packetized interconnect.3. A method according to claim I comprising transmitting the data . . from the network interface to the second computer node by way of ë a full-duplex communication link of the inter-node communication ...network. .. :4. A method according to claim 3 comprising passing the data through a buffer at the network interface before transmitting the data.5. A method according to claim 1 comprising, at the network interface, determining a size of the data and, based upon the size of - 30 the data, selecting among two or more protocols for transmitting the data.6. A method according to claim 5 wherein the two or more protocols comprise an eager protocol and a rendezvous protocol.7. A method according to claim 6 comprising, upon selecting the rendezvous protocol, automatically generating a Ready To Send message at the network interface of the first compute node.8. A method according to claim l wherein the data comprises a raw ethertype datagram and transmitting the data comprises encapsulating the raw ethertype datagram within one or more link layer packet headers.9. A method according to claim 8 wherein the link layer packet headers comprise InfiniBand_ link layer packet headers. . r ël O. A method according to claim I wherein the data comprises a raw . internet protocol datagram and transmitting the data comprises encapsulating the internet protocol datagram within one or more link layer packet headers.11. A compute node for use in a multi-compute-node computer system; the compute node comprising: a CPU; a network interface; and, o - 31 a dedicated full-duplex packetized interconnect directly coupling the CPU to the network interface.12. A compute node according to claim 1 1 wherein the dedicated packetized full-duplex interconnect is not shared by any devices other than the CPU and the network interface.13. A compute node according to claim 1 1 comprising a memory, and a facility configured to allocate eager protocol buffers in the memory and to automatically signal to one or more other compute nodes that the eager protocol buffers have been allocated.14. A compute node according to claim 13 comprising a facility configured to automatically associate memory protection keys with the eager protocol buffers and a facility configured to verify memory protection keys in incoming eager protocol messages : ,, before writing the incoming eager protocol messages to the eager ë. e " .. protocol buffers. se.eeeee.A. 20 15. A compute node according to claim 1 1 wherein the network interface comprises a hardware facility at the interface configured : to encapsulate data received on the packetized interconnect in link acme e layer packet headers.16. A compute node according to claim l I wherein the network interface comprises a buffer connected to buffer outgoing data. to - 3217. A compute node according to claim 1 1 comprising a plurality of CPUs each connected to the interface by a separate dedicated full- duplex packetized interconnect.18. A compute node according to claim 1 1 wherein the CPU is connected to each of a plurality of network interfaces by a plurality of dedicated full-duplex packetized interconnects.19. A compute node according to claim 1 1 wherein the network interface comprises a facility configured to determine a size of data to be transmitted to another compute node and, based upon the size, to select among two or more protocols for transmitting the data to the other compute node.20. A computer system comprising a plurality of compute nodes according to claim 11 interconnected by an inter-node data : . .. communication network, the inter-node data communication ea.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US52877403P | 2003-12-12 | 2003-12-12 | |
US53199903P | 2003-12-24 | 2003-12-24 | |
US10/788,455 US20050132089A1 (en) | 2003-12-12 | 2004-03-01 | Directly connected low latency network and interface |
Publications (3)
Publication Number | Publication Date |
---|---|
GB0427107D0 GB0427107D0 (en) | 2005-01-12 |
GB2409073A true GB2409073A (en) | 2005-06-15 |
GB2409073B GB2409073B (en) | 2007-03-28 |
Family
ID=34084541
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB0427107A Expired - Fee Related GB2409073B (en) | 2003-12-12 | 2004-12-10 | Directly connected low latency network and interface |
Country Status (2)
Country | Link |
---|---|
US (1) | US20050132089A1 (en) |
GB (1) | GB2409073B (en) |
Families Citing this family (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6747384B2 (en) * | 2002-01-24 | 2004-06-08 | Visteon Global Technologies, Inc. | Alternator hybrid magnet rotor design |
US7664110B1 (en) | 2004-02-07 | 2010-02-16 | Habanero Holdings, Inc. | Input/output controller for coupling the processor-memory complex to the fabric in fabric-backplane interprise servers |
US7757033B1 (en) | 2004-02-13 | 2010-07-13 | Habanero Holdings, Inc. | Data exchanges among SMP physical partitions and I/O interfaces enterprise servers |
US7953903B1 (en) | 2004-02-13 | 2011-05-31 | Habanero Holdings, Inc. | Real time detection of changed resources for provisioning and management of fabric-backplane enterprise servers |
US7843906B1 (en) | 2004-02-13 | 2010-11-30 | Habanero Holdings, Inc. | Storage gateway initiator for fabric-backplane enterprise servers |
US7860097B1 (en) | 2004-02-13 | 2010-12-28 | Habanero Holdings, Inc. | Fabric-backplane enterprise servers with VNICs and VLANs |
US8868790B2 (en) | 2004-02-13 | 2014-10-21 | Oracle International Corporation | Processor-memory module performance acceleration in fabric-backplane enterprise servers |
US7685281B1 (en) | 2004-02-13 | 2010-03-23 | Habanero Holdings, Inc. | Programmatic instantiation, provisioning and management of fabric-backplane enterprise servers |
US7860961B1 (en) | 2004-02-13 | 2010-12-28 | Habanero Holdings, Inc. | Real time notice of new resources for provisioning and management of fabric-backplane enterprise servers |
US7633955B1 (en) | 2004-02-13 | 2009-12-15 | Habanero Holdings, Inc. | SCSI transport for fabric-backplane enterprise servers |
US7873693B1 (en) | 2004-02-13 | 2011-01-18 | Habanero Holdings, Inc. | Multi-chassis fabric-backplane enterprise servers |
US8145785B1 (en) | 2004-02-13 | 2012-03-27 | Habanero Holdings, Inc. | Unused resource recognition in real time for provisioning and management of fabric-backplane enterprise servers |
US7843907B1 (en) | 2004-02-13 | 2010-11-30 | Habanero Holdings, Inc. | Storage gateway target for fabric-backplane enterprise servers |
US7990994B1 (en) | 2004-02-13 | 2011-08-02 | Habanero Holdings, Inc. | Storage gateway provisioning and configuring |
US7561571B1 (en) | 2004-02-13 | 2009-07-14 | Habanero Holdings, Inc. | Fabric address and sub-address resolution in fabric-backplane enterprise servers |
US8713295B2 (en) | 2004-07-12 | 2014-04-29 | Oracle International Corporation | Fabric-backplane enterprise servers with pluggable I/O sub-system |
US7496797B2 (en) * | 2005-03-31 | 2009-02-24 | Intel Corporation | Advanced switching lost packet and event detection and handling |
DE102005025582B4 (en) * | 2005-06-01 | 2011-08-18 | Phoenix Contact GmbH & Co. KG, 32825 | Device and method for the combined transmission of input / output data in automation bus systems |
US20070033247A1 (en) * | 2005-08-02 | 2007-02-08 | The Mathworks, Inc. | Methods and system for distributing data to technical computing workers |
US8924590B2 (en) * | 2006-02-14 | 2014-12-30 | Hewlett-Packard Development Company, L.P. | System and method for communicating in a networked system |
JP2011055044A (en) * | 2009-08-31 | 2011-03-17 | Fujitsu Ltd | Communication system |
JP2012065281A (en) * | 2010-09-17 | 2012-03-29 | Fujitsu Ltd | Communication program, communication apparatus, communication method, and communication system |
CN102932099B (en) * | 2012-10-11 | 2015-10-21 | 三维通信股份有限公司 | Data transmission method between a kind of RMII and multiple CPRI |
US9544261B2 (en) | 2013-08-27 | 2017-01-10 | International Business Machines Corporation | Data communications in a distributed computing environment |
US10277547B2 (en) * | 2013-08-27 | 2019-04-30 | International Business Machines Corporation | Data communications in a distributed computing environment |
US9753790B2 (en) | 2015-08-18 | 2017-09-05 | Nxp Usa, Inc. | Data processing system having messaging |
US10222992B2 (en) | 2016-01-30 | 2019-03-05 | Western Digital Technologies, Inc. | Synchronization method and apparatus for an interconnection network using parallel-headerless TDMA routing |
US10644958B2 (en) | 2016-01-30 | 2020-05-05 | Western Digital Technologies, Inc. | All-connected by virtual wires network of data processing nodes |
US10034407B2 (en) * | 2016-07-22 | 2018-07-24 | Intel Corporation | Storage sled for a data center |
US20220019960A1 (en) * | 2020-07-17 | 2022-01-20 | Custody Chain LLC | UniTOS Universal Transportation Operating System |
US12026091B2 (en) | 2021-12-16 | 2024-07-02 | Hewlett Packard Enterprise Development Lp | Orthogonal memory lane-based memory management |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020002443A1 (en) * | 1998-10-10 | 2002-01-03 | Ronald M. Ames | Multi-level architecture for monitoring and controlling a functional system |
US6643764B1 (en) * | 2000-07-20 | 2003-11-04 | Silicon Graphics, Inc. | Multiprocessor system utilizing multiple links to improve point to point bandwidth |
GB2393536A (en) * | 2002-08-02 | 2004-03-31 | Hewlett Packard Development Co | Back plane for modular cells in which the configuration of the backplane determines the performance of the system |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5996020A (en) * | 1995-07-21 | 1999-11-30 | National Security Agency | Multiple level minimum logic network |
US6178174B1 (en) * | 1997-08-26 | 2001-01-23 | International Business Machines Corporation | Optimistic, eager rendezvous transmission mode and combined rendezvous modes for message processing systems |
CN1311376C (en) * | 2001-02-24 | 2007-04-18 | 国际商业机器公司 | Novel massively parallel super computer |
US7401126B2 (en) * | 2001-03-23 | 2008-07-15 | Neteffect, Inc. | Transaction switch and network interface adapter incorporating same |
US7190667B2 (en) * | 2001-04-26 | 2007-03-13 | Intel Corporation | Link level packet flow control mechanism |
US20020159456A1 (en) * | 2001-04-27 | 2002-10-31 | Foster Michael S. | Method and system for multicasting in a routing device |
US20020165978A1 (en) * | 2001-05-07 | 2002-11-07 | Terence Chui | Multi-service optical infiniband router |
US7093024B2 (en) * | 2001-09-27 | 2006-08-15 | International Business Machines Corporation | End node partitioning using virtualization |
US7155537B1 (en) * | 2001-09-27 | 2006-12-26 | Lsi Logic Corporation | Infiniband isolation bridge merged with architecture of an infiniband translation bridge |
US7447778B2 (en) * | 2002-05-06 | 2008-11-04 | Qlogic, Corporation | System and method for a shared I/O subsystem |
US6944719B2 (en) * | 2002-05-15 | 2005-09-13 | Broadcom Corp. | Scalable cache coherent distributed shared memory processing system |
US7062610B2 (en) * | 2002-09-30 | 2006-06-13 | Advanced Micro Devices, Inc. | Method and apparatus for reducing overhead in a data processing system with a cache |
US8037224B2 (en) * | 2002-10-08 | 2011-10-11 | Netlogic Microsystems, Inc. | Delegating network processor operations to star topology serial bus interfaces |
US7613900B2 (en) * | 2003-03-31 | 2009-11-03 | Stretch, Inc. | Systems and methods for selecting input/output configuration in an integrated circuit |
US7334102B1 (en) * | 2003-05-09 | 2008-02-19 | Advanced Micro Devices, Inc. | Apparatus and method for balanced spinlock support in NUMA systems |
US20040233934A1 (en) * | 2003-05-23 | 2004-11-25 | Hooper Donald F. | Controlling access to sections of instructions |
US7386626B2 (en) * | 2003-06-23 | 2008-06-10 | Newisys, Inc. | Bandwidth, framing and error detection in communications between multi-processor clusters of multi-cluster computer systems |
US20050018669A1 (en) * | 2003-07-25 | 2005-01-27 | International Business Machines Corporation | Infiniband subnet management queue pair emulation for multiple logical ports on a single physical port |
US20050044301A1 (en) * | 2003-08-20 | 2005-02-24 | Vasilevsky Alexander David | Method and apparatus for providing virtual computing services |
US7443860B2 (en) * | 2004-06-08 | 2008-10-28 | Sun Microsystems, Inc. | Method and apparatus for source authentication in a communications network |
-
2004
- 2004-03-01 US US10/788,455 patent/US20050132089A1/en not_active Abandoned
- 2004-12-10 GB GB0427107A patent/GB2409073B/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020002443A1 (en) * | 1998-10-10 | 2002-01-03 | Ronald M. Ames | Multi-level architecture for monitoring and controlling a functional system |
US6643764B1 (en) * | 2000-07-20 | 2003-11-04 | Silicon Graphics, Inc. | Multiprocessor system utilizing multiple links to improve point to point bandwidth |
GB2393536A (en) * | 2002-08-02 | 2004-03-31 | Hewlett Packard Development Co | Back plane for modular cells in which the configuration of the backplane determines the performance of the system |
Also Published As
Publication number | Publication date |
---|---|
GB0427107D0 (en) | 2005-01-12 |
US20050132089A1 (en) | 2005-06-16 |
GB2409073B (en) | 2007-03-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20050132089A1 (en) | Directly connected low latency network and interface | |
US6993611B2 (en) | Enhanced general input/output architecture and related methods for establishing virtual channels therein | |
KR100666515B1 (en) | Storage and forward switch device, system and method | |
KR100687659B1 (en) | A network interface for controlling lock operation according to the AI protocol, a packet data communication on-chip interconnect system including the network interface, and a method of operating the network interface. | |
US9602408B2 (en) | General input/output architecture, protocol and related methods to implement flow control | |
US5247626A (en) | Fddi controller having flexible buffer management | |
US9430432B2 (en) | Optimized multi-root input output virtualization aware switch | |
US8571033B2 (en) | Smart routing between peers in a point-to-point link based system | |
US7133943B2 (en) | Method and apparatus for implementing receive queue for packet-based communications | |
US20160283422A1 (en) | Network interface controller with direct connection to host memory | |
CN109710548A (en) | A kind of DMA control data transmission method, system and equipment | |
WO2001018988A1 (en) | Bridge between parallel buses over a packet-switched network | |
CN1279785A (en) | Intelligent data bus inferface using multi-port memory | |
GB2226738A (en) | Cluster link interface for a local area network | |
US7596148B2 (en) | Receiving data from virtual channels | |
US6856619B1 (en) | Computer network controller | |
US20040019704A1 (en) | Multiple processor integrated circuit having configurable packet-based interfaces | |
CN1143775A (en) | Apparatus and method for inter-module communication using a system bus controller | |
US7218638B2 (en) | Switch operation scheduling mechanism with concurrent connection and queue scheduling | |
US7313146B2 (en) | Transparent data format within host device supporting differing transaction types | |
US20040017813A1 (en) | Transmitting data from a plurality of virtual channels via a multiple processor device | |
KR100324281B1 (en) | Centralized High Speed Data Processing Module | |
US20040030799A1 (en) | Bandwidth allocation fairness within a processing system of a plurality of processing devices | |
Chen | PCI Express-based Ethernet Switch | |
IL148263A (en) | Bridge between parallel buses over a packet-switched network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
732E | Amendments to the register in respect of changes of name or changes affecting rights (sect. 32/1977) |
Free format text: REGISTERED BETWEEN 20091001 AND 20091007 |
|
732E | Amendments to the register in respect of changes of name or changes affecting rights (sect. 32/1977) |
Free format text: REGISTERED BETWEEN 20120913 AND 20120919 |
|
PCNP | Patent ceased through non-payment of renewal fee |
Effective date: 20151210 |