US20080250181A1 - Server - Google Patents

Server Download PDF

Info

Publication number
US20080250181A1
US20080250181A1 US12/130,742 US13074208A US2008250181A1 US 20080250181 A1 US20080250181 A1 US 20080250181A1 US 13074208 A US13074208 A US 13074208A US 2008250181 A1 US2008250181 A1 US 2008250181A1
Authority
US
United States
Prior art keywords
interface
blade
slots
backboard
server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/130,742
Other versions
US7865655B2 (en
Inventor
Minqiu Li
Feng Hong
Chunming Sheng
Tinghong Wang
Xing Rao
Jin Yu
Shaolin Zhang
Hansi Wang
Dingliang Gan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Assigned to HUAWEI TECHNOLOGIES CO., LTD. reassignment HUAWEI TECHNOLOGIES CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SHENG, CHUNMING, YU, JIN, GAN, DINGLIANG, LI, MINQIU, RAO, XING, WANG, HANSI, ZHANG, SHAOLIN, HONG, FENG, WANG, TINGHONG
Publication of US20080250181A1 publication Critical patent/US20080250181A1/en
Application granted granted Critical
Publication of US7865655B2 publication Critical patent/US7865655B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/18Packaging or power distribution
    • G06F1/183Internal mounting support structures, e.g. for printed circuit boards, internal connecting means
    • G06F1/184Mounting of motherboards

Definitions

  • the present invention relates to the field of communications, in particular, to a server.
  • the server is being developed in two directions, one is the direction of a normal server, the other is the direction of a high-end server utilizing a high-performance multi-processing architecture.
  • the high-end server includes a Symmetrical Multi-Processing (SMP) server utilizing an SMP architecture and a Cache Coherence-Non Uniform Memory Access (CC-Numa) server utilizing a CC-Numa architecture.
  • SMP Symmetrical Multi-Processing
  • CC-Numa Cache Coherence-Non Uniform Memory Access
  • a plurality of Central Processing Units operate symmetrically without primary and secondary relationships or dependent relationships among them.
  • Each CPU shares the same physical Random Access Memory (RAM), and it takes each CPU the same time period to access the RAM at any address.
  • RAM Random Access Memory
  • Methods for extending an SMP server include adding RAM, using faster CPUs, adding CPUs, expanding I/O capabilities, and employing more external devices, such as disk storages, etc.
  • the characteristics of a CC-Numa server are similar to those of an SMP server except that each CPU takes different periods to access local RAM and remote RAM.
  • a blade server is a low cost server platform with High Availability High Density (HAHD) designed specially for specific applications and high-density computer environments.
  • Each of the blades in the blade server is a system mainboard which may start its own operating system, such as Windows NT/2000, Linux, Solaris, etc., through a local hard disk, with each of the blades serving as an independent server. In such an operating mode, each of the mainboards operates its own system and provides services for a designated user group without relationships among users.
  • These blades may also be combined into a server cluster using system software. In the server cluster mode, all of the mainboards may be interconnected to provide a high-speed network environment, and share the resources to provide their respective services for the same user group.
  • the performance of such a system may be improved by inserting new blades into the cluster. Since each blade is hot swappable, the system may be replaced easily and maintenance time may be minimized.
  • CPUs may be configured into a variety of subsystems.
  • the blade servers on the same rack may share the same set of devices including a CD driver, a floppy driver, a keyboard, a display, and a mouse via a new type of intelligent KVM switch board, so that a plurality of servers can be accessed, thereby facilitating the upgrade and maintenance of the system as well as the access of files on the blade servers.
  • both a normal blade server and a multi-processing blade server in the prior art can meet the requirement for a high-performance and cost-effective server.
  • interfaces and wiring at the slots on the backboard of a normal blade server which are used for the insertion and interconnection of normal blades, are of normal blade-specific designs.
  • the interfaces and wiring of a normal blade and those of a multiple-processing blade are not compatible with each other, as a result, it is difficult to flexibly configure the system by selecting the respective blades for various levels of services, and thus it is difficult to meet the requirement for a cost-effective server.
  • An embodiment of the present invention provides a server for solving the problem related to the incompatibility between normal blades and multi-processing blades in a conventional server.
  • a first server includes a backboard, on which backboard wiring and at least one first slot are disposed. Both a number of Cluster Switch interfaces and a number of SMP Switch interfaces are disposed on the at least one first slot. At least two second slots are further disposed on the backboard. Both a first interface configured to be connected to a normal blade and a second interface configured to be connected to a multi-processing blade are disposed on each of the second slots, the first interface being connected to a corresponding Cluster Switch interface via the backboard wiring, and the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
  • the second interface may be further connected to a corresponding Cluster Switch interface.
  • a third interface configured to be connected to the multi-processing blade is further disposed on each of the second slots, the third interface being connected to a corresponding Cluster Switch interface via the backboard wiring.
  • a second server includes a backboard, on which backboard wiring and at least one first slot are disposed.
  • a number of Cluster Switch interfaces are disposed on the at least one first slot.
  • At least two second slots are further disposed on the backboard.
  • Both a first interface configured to be connected to a normal blade and a fourth interface configured to be connected to a multi-processing blade are disposed on each of the second slots, the first interface being connected to a corresponding Cluster Switch interface via the backboard wiring, and the fourth interfaces being connected with each other via the backboard wiring.
  • a number of Cluster Switch interfaces are further disposed on the at least one first slot, and a second interface configured to be connected to a multi-processing blade is further disposed on each of the second slots, the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
  • the multi-processing blades and the normal blades may coexist in the above-described servers. Moreover, the combination of blades may be adjusted dynamically in order to meet the requirement for flexible configurations of the system depending on various services, thereby minimizing user costs while ensuring that the system has a relatively high degree of extensibility.
  • FIG. 1 is a schematic diagram showing a combination structure of slots in a server according to an embodiment of the invention
  • FIG. 2 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention
  • FIG. 3 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention
  • FIG. 4 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention
  • FIG. 5 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention
  • FIG. 6 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention
  • FIG. 7 is a schematic diagram showing a structure of a server according to an embodiment of the invention.
  • FIG. 8 is an exemplary application of a server according to an embodiment of the invention.
  • FIG. 9 is an exemplary application of a server according to an embodiment of the invention.
  • a server includes a backboard, on which backboard wiring and a plurality of slots, i.e. first slots 10 - 1 and 10 - 2 , referred to collectively as first slots 10 , and second slots 20 - 1 , 20 - 2 . . . , and 20 - n, referred to collectively as second slots 20 , as shown in FIG. 1 , are disposed.
  • Two mutual backup first slots 10 - 1 , 10 - 2 are generally disposed on the backboard for connecting to a Switch blade, and a plurality of second slots 20 parallel to each other are disposed on the backboard for connecting to a normal blade or a multi-processing blade.
  • Both a Cluster Switch and an SMP Switch can be included on a Switch blade.
  • Cluster Switch interfaces and SMP Switch interfaces having the number as that of the second slots 20 are disposed on the first slots 10 .
  • a first group of interfaces i.e. a number of Cluster Switch interfaces configured to be connected to the Cluster Switch on the Switch blade
  • a second group of interfaces i.e. a number of SMP Switch interfaces configured to be connected to the SMP Switch on the Switch blade
  • the interface disposed on the second slot 20 includes a first interface 201 configured to be connected to a normal blade.
  • the first interface 201 is connected to one of the Cluster Switch interfaces via the backboard wiring.
  • the first interface 201 enables the normal blades connected to the second slots 20 to be interconnected via a Cluster Switch on a Switch blade, thereby forming a normal blade server subsystem.
  • a second interface 202 is configured to be connected to a multi-processing blade.
  • the second interface 202 is connected to one of the SMP Switch interfaces via the backboard wiring.
  • the second interface 202 enables the multi-processing blades to be interconnected via an SMP Switch on a Switch blade, thereby forming an SMP server subsystem.
  • a third group of interfaces i.e. a number of Cluster Switch interfaces may also be disposed on the first slot 10 , and the second interface 202 on the second slot 20 is further connected to one of the third group of interfaces, i.e. Cluster Switch interfaces, via the backboard wiring.
  • a SMP blade and a normal blade constitute a blade server via a Cluster Switch.
  • an SMP blade may be connected to an SMP Switch or a Cluster Switch on a Switch blade via the shared second interface 202 .
  • the interface on an SMP blade configured to be connected to the Cluster Switch on the Switch blade may be independently disposed.
  • a third interface 203 on the second slot 20 is configured to be connected to a multi-processing blade.
  • the third interface 203 is connected to one of a third group of interfaces, i.e. Cluster Switch interfaces, via the backboard wiring.
  • the second interface 202 enables the multi-processing blades to be interconnected via an SMP Switch on a Switch blade
  • the third interface 203 enables the multi-processing blades to be interconnected via a Cluster Switch on a Switch blade.
  • an SMP blade and a normal blade constitute a blade server via a Cluster Switch.
  • connection between a Cluster Switch and an SMP Switch on the same Switch blade may be enabled or disabled by enabling or disabling the SMP Switch.
  • the multi-processing blades may also be interconnected directly via a high-speed bus.
  • a fourth interface 204 configured to be connected to a multi-processing blade may further be disposed on the second slot 20 .
  • the fourth interface 204 is connected with another fourth interface via the backboard wiring. All of the multi-processing blades inserted into the second slots 20 are interconnected with each other through the fourth interface 204 via the backboard wiring.
  • the normal blades may be interconnected by using a Switch blade provided only with a Cluster Switch.
  • one or both of the second interface 202 and the third interface 203 may be disposed on each of the second slots 20 .
  • the SMP blades may be interconnected via an SMP Switch disposed on a Switch blade, or may be interconnected directly via the backboard wiring.
  • the fourth interface 204 , the second interface 202 and the third interface 203 are each disposed on the second slot 20 .
  • the normal blades may also be interconnected via an external switching device. Accordingly, a fifth interface 205 configured to be connected to a multi-processing blade is further disposed on each of the second slots 20 . At least two normal blades inserted into the second slots may be interconnected through the fifth interface 205 using the external switching device. Further, a maintenance signal interface 206 shared by the normal blades and/or the multi-processing blades may be further disposed on each of the second slots 20 .
  • the backboard wiring may be in the form of a serial high-speed bus or parallel high-speed bus.
  • a multi-processing blade includes an SMP blade and a CC-Numa blade.
  • the first interface, the second interface or the third interface for interconnecting blades may be a 100-Megabit Ethernet interface, a 1000-Megabit Ethernet interface, or another high-speed interface, such as an Infiniband interface, a Myrinet interface, etc.
  • the architecture of an SMP server (or a CC-Numa server) is combined with that of a blade server, so that an independent SMP server (or a CC-Numa server), an independent normal blade server, or a server with a hybrid architecture based on “a SMP server (or a CC-Numa server)+a normal blade server” may be constituted according to various user requirements.
  • the combination of the blades may be determined by user application requirements, thereby minimizing and protecting users' device investment.
  • FIG. 7 is a schematic diagram showing a structure of a server according to an embodiment of the invention.
  • Each of blades 1 , 2 , 3 and 4 is an SMP blade
  • each of blades 5 , 6 , 7 and 8 is a normal blade
  • each of blades 9 and 10 is a Switch blade.
  • the SMP blades 1 , 2 , 3 and 4 are each connected to the Switch blade 9 via a respective interface and backboard wiring, thereby constituting an SMP server subsystem.
  • the normal blades 5 , 6 , 7 and 8 are each connected to the Switch blade 9 via a respective interface and backboard wiring, thereby constituting a normal blade server subsystem.
  • the SMP server subsystem and the normal blade server subsystem may also constitute an extended server system with a hybrid architecture based on “an SMP server+a normal blade server”.
  • the Switch blade 10 is a redundant backup of the Switch blade 9 .
  • any connection to the Switch blade 10 is not shown if redundant design is employed.
  • the connections are the same as those of the Switch blade 9 .
  • high-speed bus interfaces are interconnected via backboard wiring in each slot.
  • the SMP blades are interconnected via the backboard.
  • the high-speed bus is not required to go through the Switch blade 9 and the Switch blade 10 .
  • a bus interface interconnected with a Switch blade may also be disposed in each slot, and the SMP blades may be interconnected via an SMP Switch disposed in a Switch blade. Therefore, based on user requirements, normal blades may be configured into independent blade servers, and SMP blades may be combined into an integral SMP server or may be configured into independent SMP blade servers.
  • the interface for interconnecting blades may be in the form of a 100-Megabit/1000-Megabit Ethernet interface or another high-speed interface, such as an Infiniband interface, a Myrinet interface, etc.
  • FIG. 8 shows an application embodiment of a server according to the invention.
  • SMP blades 1 , 2 , 3 and 4 constitute an SMP server through an SMP Switch, and each of the SMP blades is connected to the Cluster Switch.
  • One or more of the connections to the Cluster Switch may be enabled, thereby allowing the SMP server and normal blades 5 , 6 , 7 and 8 to constitute a blade server via the Cluster Switch; however, all of the connections may be disabled, thereby constituting an independent SMP server and an independent normal blade server.
  • Both the Cluster Switch and the SMP Switch shown in FIG. 8 are disposed on a Switch blade.
  • SMP blades 1 , 2 , 3 and 4 are directly interconnected via a high-speed bus on the backboard, thereby constituting an SMP server, and each of the SMP blades is connected to the Cluster Switch via the backboard wiring and an interface.
  • One or more of the connections to the Cluster Switch may be enabled, allowing the SMP server and normal blades 5 , 6 , 7 and 8 to constitute a normal blade server via the Cluster Switch; however, all of the interconnections may be disabled, thereby constituting an independent SMP server and an independent normal blade server.
  • the Cluster Switch shown in FIG. 9 is disposed on a Switch blade.
  • no connection between an SMP server and a normal blade server may exist, and a plurality of independent subsystems may be constituted, or a plurality of SMP server systems may be constituted, for example, SMP blades 1 and 2 may constitute an SMP system, while SMP blades 3 and 4 may constitute another SMP system, or they may be combined into a plurality of SMP blade server subsystems.

Abstract

The present invention relates to the field of communications, in particular, to a server for solving the problem related to the incompatibility between normal blades and multi-processing blades in a conventional server. The server according to an embodiment of the invention includes a backboard, on which backboard wiring and a first slot are disposed. At least two second slots are further disposed on the backboard. Both a first interface configured to be connected to a normal blade and a second interface configured to be connected to a multi-processing blade are disposed on each of the second slots, the first interface being connected to a corresponding Cluster Switch interface disposed on the first slot via the backboard wiring, and the second interface being interconnected directly via the backboard wiring or being connected to a corresponding Symmetrical Multi-Processing Switch interface disposed on the first slot via the backboard wiring.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application is a continuation application of commonly-assigned International Application No. PCT/CN2006/003236, filed Nov. 30, 2006, which claims the benefit of Chinese Patent Application No. 200520119927.3, filed on Dec. 1, 2005, both of which are herein incorporated by reference.
  • FIELD OF THE INVENTION
  • The present invention relates to the field of communications, in particular, to a server.
  • BACKGROUND OF THE INVENTION
  • At present, the server is being developed in two directions, one is the direction of a normal server, the other is the direction of a high-end server utilizing a high-performance multi-processing architecture. The high-end server includes a Symmetrical Multi-Processing (SMP) server utilizing an SMP architecture and a Cache Coherence-Non Uniform Memory Access (CC-Numa) server utilizing a CC-Numa architecture.
  • In an SMP server, a plurality of Central Processing Units (CPUs) operate symmetrically without primary and secondary relationships or dependent relationships among them. Each CPU shares the same physical Random Access Memory (RAM), and it takes each CPU the same time period to access the RAM at any address. Methods for extending an SMP server include adding RAM, using faster CPUs, adding CPUs, expanding I/O capabilities, and employing more external devices, such as disk storages, etc. The characteristics of a CC-Numa server are similar to those of an SMP server except that each CPU takes different periods to access local RAM and remote RAM.
  • A blade server is a low cost server platform with High Availability High Density (HAHD) designed specially for specific applications and high-density computer environments. Each of the blades in the blade server is a system mainboard which may start its own operating system, such as Windows NT/2000, Linux, Solaris, etc., through a local hard disk, with each of the blades serving as an independent server. In such an operating mode, each of the mainboards operates its own system and provides services for a designated user group without relationships among users. These blades may also be combined into a server cluster using system software. In the server cluster mode, all of the mainboards may be interconnected to provide a high-speed network environment, and share the resources to provide their respective services for the same user group. The performance of such a system may be improved by inserting new blades into the cluster. Since each blade is hot swappable, the system may be replaced easily and maintenance time may be minimized. In a blade server, CPUs may be configured into a variety of subsystems. The blade servers on the same rack may share the same set of devices including a CD driver, a floppy driver, a keyboard, a display, and a mouse via a new type of intelligent KVM switch board, so that a plurality of servers can be accessed, thereby facilitating the upgrade and maintenance of the system as well as the access of files on the blade servers.
  • At present, both a normal blade server and a multi-processing blade server in the prior art can meet the requirement for a high-performance and cost-effective server. However, interfaces and wiring at the slots on the backboard of a normal blade server, which are used for the insertion and interconnection of normal blades, are of normal blade-specific designs. The interfaces and wiring of a normal blade and those of a multiple-processing blade are not compatible with each other, as a result, it is difficult to flexibly configure the system by selecting the respective blades for various levels of services, and thus it is difficult to meet the requirement for a cost-effective server.
  • SUMMARY OF THE INVENTION
  • An embodiment of the present invention provides a server for solving the problem related to the incompatibility between normal blades and multi-processing blades in a conventional server.
  • A first server according to an embodiment of the invention includes a backboard, on which backboard wiring and at least one first slot are disposed. Both a number of Cluster Switch interfaces and a number of SMP Switch interfaces are disposed on the at least one first slot. At least two second slots are further disposed on the backboard. Both a first interface configured to be connected to a normal blade and a second interface configured to be connected to a multi-processing blade are disposed on each of the second slots, the first interface being connected to a corresponding Cluster Switch interface via the backboard wiring, and the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
  • The second interface may be further connected to a corresponding Cluster Switch interface. Alternatively, a third interface configured to be connected to the multi-processing blade is further disposed on each of the second slots, the third interface being connected to a corresponding Cluster Switch interface via the backboard wiring.
  • A second server according to an embodiment of the present invention includes a backboard, on which backboard wiring and at least one first slot are disposed. A number of Cluster Switch interfaces are disposed on the at least one first slot. At least two second slots are further disposed on the backboard. Both a first interface configured to be connected to a normal blade and a fourth interface configured to be connected to a multi-processing blade are disposed on each of the second slots, the first interface being connected to a corresponding Cluster Switch interface via the backboard wiring, and the fourth interfaces being connected with each other via the backboard wiring.
  • A number of Cluster Switch interfaces are further disposed on the at least one first slot, and a second interface configured to be connected to a multi-processing blade is further disposed on each of the second slots, the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
  • The multi-processing blades and the normal blades may coexist in the above-described servers. Moreover, the combination of blades may be adjusted dynamically in order to meet the requirement for flexible configurations of the system depending on various services, thereby minimizing user costs while ensuring that the system has a relatively high degree of extensibility.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic diagram showing a combination structure of slots in a server according to an embodiment of the invention;
  • FIG. 2 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention;
  • FIG. 3 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention;
  • FIG. 4 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention;
  • FIG. 5 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention;
  • FIG. 6 is a schematic diagram showing an interface structure and connections between a first slot and a second slot in a server according to an embodiment of the invention;
  • FIG. 7 is a schematic diagram showing a structure of a server according to an embodiment of the invention;
  • FIG. 8 is an exemplary application of a server according to an embodiment of the invention; and
  • FIG. 9 is an exemplary application of a server according to an embodiment of the invention.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • A server according to an embodiment of the invention includes a backboard, on which backboard wiring and a plurality of slots, i.e. first slots 10-1 and 10-2, referred to collectively as first slots 10, and second slots 20-1, 20-2 . . . , and 20-n, referred to collectively as second slots 20, as shown in FIG. 1, are disposed. Two mutual backup first slots 10-1, 10-2 are generally disposed on the backboard for connecting to a Switch blade, and a plurality of second slots 20 parallel to each other are disposed on the backboard for connecting to a normal blade or a multi-processing blade.
  • Both a Cluster Switch and an SMP Switch can be included on a Switch blade. Cluster Switch interfaces and SMP Switch interfaces having the number as that of the second slots 20 are disposed on the first slots 10.
  • Referring to FIG. 2, an interface structure and connections between an exemplary first slot 10 and an exemplary second slot 20 are shown. A first group of interfaces, i.e. a number of Cluster Switch interfaces configured to be connected to the Cluster Switch on the Switch blade, and a second group of interfaces, i.e. a number of SMP Switch interfaces configured to be connected to the SMP Switch on the Switch blade, are disposed on the first slot 10.
  • Again referring to FIG. 2, the interface disposed on the second slot 20 includes a first interface 201 configured to be connected to a normal blade. The first interface 201 is connected to one of the Cluster Switch interfaces via the backboard wiring. The first interface 201 enables the normal blades connected to the second slots 20 to be interconnected via a Cluster Switch on a Switch blade, thereby forming a normal blade server subsystem.
  • A second interface 202 is configured to be connected to a multi-processing blade. The second interface 202 is connected to one of the SMP Switch interfaces via the backboard wiring. The second interface 202 enables the multi-processing blades to be interconnected via an SMP Switch on a Switch blade, thereby forming an SMP server subsystem.
  • Referring to FIG. 3, a third group of interfaces, i.e. a number of Cluster Switch interfaces may also be disposed on the first slot 10, and the second interface 202 on the second slot 20 is further connected to one of the third group of interfaces, i.e. Cluster Switch interfaces, via the backboard wiring. Thus, a SMP blade and a normal blade constitute a blade server via a Cluster Switch. In such a case, an SMP blade may be connected to an SMP Switch or a Cluster Switch on a Switch blade via the shared second interface 202.
  • As shown in FIG. 4, the interface on an SMP blade configured to be connected to the Cluster Switch on the Switch blade may be independently disposed. For example, a third interface 203 on the second slot 20 is configured to be connected to a multi-processing blade. The third interface 203 is connected to one of a third group of interfaces, i.e. Cluster Switch interfaces, via the backboard wiring. As such, the second interface 202 enables the multi-processing blades to be interconnected via an SMP Switch on a Switch blade, and the third interface 203 enables the multi-processing blades to be interconnected via a Cluster Switch on a Switch blade. Thus, an SMP blade and a normal blade constitute a blade server via a Cluster Switch.
  • The connection between a Cluster Switch and an SMP Switch on the same Switch blade may be enabled or disabled by enabling or disabling the SMP Switch.
  • As shown in FIG. 5, the multi-processing blades may also be interconnected directly via a high-speed bus. A fourth interface 204 configured to be connected to a multi-processing blade may further be disposed on the second slot 20. The fourth interface 204 is connected with another fourth interface via the backboard wiring. All of the multi-processing blades inserted into the second slots 20 are interconnected with each other through the fourth interface 204 via the backboard wiring. In this case, the normal blades may be interconnected by using a Switch blade provided only with a Cluster Switch.
  • In addition to the fourth interface 204, one or both of the second interface 202 and the third interface 203 may be disposed on each of the second slots 20. In this case, the SMP blades may be interconnected via an SMP Switch disposed on a Switch blade, or may be interconnected directly via the backboard wiring. For example, as shown in FIG. 6, the fourth interface 204, the second interface 202 and the third interface 203 are each disposed on the second slot 20.
  • Referring to FIG. 2-6, the normal blades may also be interconnected via an external switching device. Accordingly, a fifth interface 205 configured to be connected to a multi-processing blade is further disposed on each of the second slots 20. At least two normal blades inserted into the second slots may be interconnected through the fifth interface 205 using the external switching device. Further, a maintenance signal interface 206 shared by the normal blades and/or the multi-processing blades may be further disposed on each of the second slots 20.
  • In an embodiment of the invention, the backboard wiring may be in the form of a serial high-speed bus or parallel high-speed bus.
  • In an embodiment of the invention, a multi-processing blade includes an SMP blade and a CC-Numa blade.
  • In an embodiment of the invention, the first interface, the second interface or the third interface for interconnecting blades may be a 100-Megabit Ethernet interface, a 1000-Megabit Ethernet interface, or another high-speed interface, such as an Infiniband interface, a Myrinet interface, etc.
  • In an embodiment of the invention, with the above-described combination structure of the slots and the backboard wiring designed for it, the architecture of an SMP server (or a CC-Numa server) is combined with that of a blade server, so that an independent SMP server (or a CC-Numa server), an independent normal blade server, or a server with a hybrid architecture based on “a SMP server (or a CC-Numa server)+a normal blade server” may be constituted according to various user requirements. The combination of the blades may be determined by user application requirements, thereby minimizing and protecting users' device investment.
  • FIG. 7 is a schematic diagram showing a structure of a server according to an embodiment of the invention. Each of blades 1, 2, 3 and 4 is an SMP blade, each of blades 5, 6, 7 and 8 is a normal blade, and each of blades 9 and 10 is a Switch blade. The SMP blades 1, 2, 3 and 4 are each connected to the Switch blade 9 via a respective interface and backboard wiring, thereby constituting an SMP server subsystem. The normal blades 5, 6, 7 and 8 are each connected to the Switch blade 9 via a respective interface and backboard wiring, thereby constituting a normal blade server subsystem. With the Switch blade 9, the SMP server subsystem and the normal blade server subsystem may also constitute an extended server system with a hybrid architecture based on “an SMP server+a normal blade server”. As shown in FIG. 7, the Switch blade 10 is a redundant backup of the Switch blade 9. For simplicity, any connection to the Switch blade 10 is not shown if redundant design is employed. The connections are the same as those of the Switch blade 9.
  • For the application of SMP blades, high-speed bus interfaces are interconnected via backboard wiring in each slot. The SMP blades are interconnected via the backboard. The high-speed bus is not required to go through the Switch blade 9 and the Switch blade 10. A bus interface interconnected with a Switch blade may also be disposed in each slot, and the SMP blades may be interconnected via an SMP Switch disposed in a Switch blade. Therefore, based on user requirements, normal blades may be configured into independent blade servers, and SMP blades may be combined into an integral SMP server or may be configured into independent SMP blade servers.
  • According to the application of blade servers, the interface for interconnecting blades may be in the form of a 100-Megabit/1000-Megabit Ethernet interface or another high-speed interface, such as an Infiniband interface, a Myrinet interface, etc.
  • FIG. 8 shows an application embodiment of a server according to the invention. SMP blades 1, 2, 3 and 4 constitute an SMP server through an SMP Switch, and each of the SMP blades is connected to the Cluster Switch. One or more of the connections to the Cluster Switch may be enabled, thereby allowing the SMP server and normal blades 5, 6, 7 and 8 to constitute a blade server via the Cluster Switch; however, all of the connections may be disabled, thereby constituting an independent SMP server and an independent normal blade server. Both the Cluster Switch and the SMP Switch shown in FIG. 8 are disposed on a Switch blade.
  • As shown in FIG. 9, in another application embodiment of a server according to the invention, SMP blades 1, 2, 3 and 4 are directly interconnected via a high-speed bus on the backboard, thereby constituting an SMP server, and each of the SMP blades is connected to the Cluster Switch via the backboard wiring and an interface. One or more of the connections to the Cluster Switch may be enabled, allowing the SMP server and normal blades 5, 6, 7 and 8 to constitute a normal blade server via the Cluster Switch; however, all of the interconnections may be disabled, thereby constituting an independent SMP server and an independent normal blade server. The Cluster Switch shown in FIG. 9 is disposed on a Switch blade.
  • In an embodiment of the invention, no connection between an SMP server and a normal blade server may exist, and a plurality of independent subsystems may be constituted, or a plurality of SMP server systems may be constituted, for example, SMP blades 1 and 2 may constitute an SMP system, while SMP blades 3 and 4 may constitute another SMP system, or they may be combined into a plurality of SMP blade server subsystems.
  • Additional advantages and modifications will readily occur to those skilled in the art. Therefore, the invention in its broader aspects is not limited to the specific details and representative embodiments shown and described herein. Accordingly, various variations and modifications may be made without departing from the spirit or scope of the invention as defined by the appended claims and their equivalents.

Claims (18)

1. A server, comprising:
a backboard on which backboard wiring, at least one first slot, and at least two second slots are disposed, the backboard having a number of Cluster Switch interfaces and a number of Symmetrical Multi-Processing (SMP) Switch interfaces disposed on the at least one first slot,
wherein each of the at least two second slots includes a first interface configured to be connected to a normal blade and a second interface configured to be connected to a multi-processing blade, wherein each of the first interfaces is connected to a corresponding Cluster Switch interface via the backboard wiring, and the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
2. The server according to claim 1, wherein the second interface is further connected to a corresponding Cluster Switch interface.
3. The server according to claim 1, wherein a third interface configured to be connected to the multi-processing blade is further disposed on each of the second slots, wherein the third interface is connected to a corresponding Cluster Switch interface via the backboard wiring.
4. The server according to claim 1, wherein a fourth interface configured to be connected to a multi-processing blade is disposed on each of the second slots, wherein the respective fourth interfaces are connected with each other via the backboard wiring.
5. The server according to claim 4, wherein the backboard wiring is in the form of a serial high-speed bus or parallel high-speed bus.
6. The server according to claim 1, wherein a fifth interface configured to be connected to a normal blade is further disposed on each of the second slots, wherein the normal blade is connected to an external switching device through the fifth interface.
7. The server according to claim 1, wherein the multi-processing blade includes an SMP blade and a Cache Coherence-Non Uniform Memory Access (CC-Numa) blade.
8. The server according to claim 1, wherein two mutual backup first slots are disposed in parallel.
9. The server according to claim 1, wherein a maintenance signal interface of at least one of the normal blades and the multi-processing blades is further disposed on each of the second slots.
10. A server, comprising:
a backboard on which backboard wiring, at least one first slot, and at least two second slots are disposed, the backboard having a number of Cluster Switch interfaces disposed on the at least one first slot,
wherein each of the at least two second slots includes a first interface configured to be connected to a normal blade and a fourth interface configured to be connected to a multi-processing blade, wherein each of the first interfaces is connected to a corresponding Cluster Switch interface via the backboard wiring, and the respective fourth interfaces are connected with each other via the backboard wiring.
11. The server according to claim 10, wherein a third interface configured to be connected to the multi-processing blade is further disposed on each of the second slots, wherein the third interface is connected to a corresponding Cluster Switch interface via the backboard wiring.
12. The server according to claim 11, wherein a number of Cluster Switch interfaces are further disposed on the at least one first slot, and a second interface configured to be connected to a multi-processing blade is further disposed on each of the second slots, wherein the second interface is connected to a corresponding Symmetrical Multi-Processing (SMP) Switch interface via the backboard wiring.
13. The server according to claim 12, wherein the backboard wiring is in the form of a serial high-speed bus or parallel high-speed bus.
14. The server according to claim 10, wherein a fifth interface configured to be connected to a normal blade is further disposed on each of the second slots, and the normal blade is connected to an external switching device through the fifth interface.
15. The server according to claim 10, wherein the multi-processing blade includes an SMP blade and a Cache Coherence-Non Uniform Memory Access (CC-Numa) blade.
16. The server according to claim 10, wherein two mutual backup first slots are disposed in parallel.
17. The server according to claim 10, wherein a maintenance signal interface for at least one of the normal blades and the multi-processing blades is further disposed on each of the second slots.
18. A backboard of a server, comprising:
at least one first slot, wherein the at least one first slot is coupled to a number of Cluster Switch interfaces and a number of Symmetrical Multi-Processing (SMP) Switch interfaces; and
at least two second slots, wherein each of the at least two second slots includes a first interface configured to be connected to a normal blade and a second interface configured to be connected to a multi-processing blade, wherein each of the first interfaces is connected to a corresponding Cluster Switch interface via backboard wiring, and the second interface being connected to a corresponding SMP Switch interface via the backboard wiring.
US12/130,742 2005-12-01 2008-05-30 Extended blade server Active 2027-12-22 US7865655B2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CNU2005201199273U CN2852260Y (en) 2005-12-01 2005-12-01 Server
CN200520119927.3 2005-12-01
CN200520119927U 2005-12-01
PCT/CN2006/003236 WO2007062593A1 (en) 2005-12-01 2006-11-30 A kind of server

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2006/003236 Continuation WO2007062593A1 (en) 2005-12-01 2006-11-30 A kind of server

Publications (2)

Publication Number Publication Date
US20080250181A1 true US20080250181A1 (en) 2008-10-09
US7865655B2 US7865655B2 (en) 2011-01-04

Family

ID=37586217

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/130,742 Active 2027-12-22 US7865655B2 (en) 2005-12-01 2008-05-30 Extended blade server

Country Status (3)

Country Link
US (1) US7865655B2 (en)
CN (2) CN2852260Y (en)
WO (1) WO2007062593A1 (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090156031A1 (en) * 2007-12-12 2009-06-18 International Business Machines Corporation Coupler Assembly for a Scalable Computer System and Scalable Computer System
US20120014390A1 (en) * 2009-06-18 2012-01-19 Martin Goldstein Processor topology switches
US20120221761A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation Shared system to operationally connect logic nodes
US20120221762A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation Logic node connection system
US20120221760A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation System to operationally connect logic nodes
US20130111230A1 (en) * 2011-10-31 2013-05-02 Calxeda, Inc. System board for system and method for modular compute provisioning in large scalable processor installations
US9008079B2 (en) 2009-10-30 2015-04-14 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric
US9054990B2 (en) 2009-10-30 2015-06-09 Iii Holdings 2, Llc System and method for data center security enhancements leveraging server SOCs or server fabrics
US9077654B2 (en) 2009-10-30 2015-07-07 Iii Holdings 2, Llc System and method for data center security enhancements leveraging managed server SOCs
US9311269B2 (en) 2009-10-30 2016-04-12 Iii Holdings 2, Llc Network proxy for high-performance, low-power data center interconnect fabric
CN105703943A (en) * 2015-12-31 2016-06-22 北京航管软件技术有限公司 Device for switching communication lines
US9465771B2 (en) 2009-09-24 2016-10-11 Iii Holdings 2, Llc Server on a chip and node cards comprising one or more of same
US9585281B2 (en) 2011-10-28 2017-02-28 Iii Holdings 2, Llc System and method for flexible storage and networking provisioning in large scalable processor installations
US9648102B1 (en) 2012-12-27 2017-05-09 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US9680770B2 (en) 2009-10-30 2017-06-13 Iii Holdings 2, Llc System and method for using a multi-protocol fabric module across a distributed server interconnect fabric
US9876735B2 (en) 2009-10-30 2018-01-23 Iii Holdings 2, Llc Performance and power optimized computer system architectures and methods leveraging power optimized tree fabric interconnect
US10140245B2 (en) 2009-10-30 2018-11-27 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US10877695B2 (en) 2009-10-30 2020-12-29 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
CN112650363A (en) * 2020-12-11 2021-04-13 北京四方继保工程技术有限公司 Distributed telemechanical system based on balanced communication CPU node group and design method
US11467883B2 (en) 2004-03-13 2022-10-11 Iii Holdings 12, Llc Co-allocating a reservation spanning different compute resources types
US11494235B2 (en) 2004-11-08 2022-11-08 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11496415B2 (en) 2005-04-07 2022-11-08 Iii Holdings 12, Llc On-demand access to compute resources
US11522952B2 (en) 2007-09-24 2022-12-06 The Research Foundation For The State University Of New York Automatic clustering for self-organizing grids
US11630704B2 (en) 2004-08-20 2023-04-18 Iii Holdings 12, Llc System and method for a workload management and scheduling module to manage access to a compute environment according to local and non-local user identity information
US11652706B2 (en) 2004-06-18 2023-05-16 Iii Holdings 12, Llc System and method for providing dynamic provisioning within a compute environment
US11650857B2 (en) 2006-03-16 2023-05-16 Iii Holdings 12, Llc System and method for managing a hybrid computer environment
US11658916B2 (en) 2005-03-16 2023-05-23 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US11720290B2 (en) 2009-10-30 2023-08-08 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US11960937B2 (en) 2022-03-17 2024-04-16 Iii Holdings 12, Llc System and method for an optimizing reservation in time of compute resources based on prioritization function and reservation policy parameter

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101150413B (en) * 2007-10-31 2010-06-02 中兴通讯股份有限公司 A multi-frame cascading system and method for ATCA knife server
CN101853185B (en) * 2009-03-30 2015-07-08 华为技术有限公司 Blade server and service dispatching method thereof
CN102955502B (en) * 2011-08-25 2017-05-17 赛恩倍吉科技顾问(深圳)有限公司 Backplane interface circuit, hard disk backplane and server system
US9104639B2 (en) * 2012-05-01 2015-08-11 SEAKR Engineering, Inc. Distributed mesh-based memory and computing architecture
CN102999801A (en) * 2012-09-11 2013-03-27 上海研庆电子有限公司 Municipal road and bridge common management platform
CN105229621B (en) 2013-05-16 2018-08-28 慧与发展有限责任合伙企业 Multi-mode proxy server
CN110456889A (en) * 2019-09-20 2019-11-15 广州市智谱科技有限公司 Edge calculations array server based on ARM framework
CN110968540A (en) * 2019-11-06 2020-04-07 北京计算机技术及应用研究所 Redundant high-speed backplate of two stars types based on VPX

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6370585B1 (en) * 1997-09-05 2002-04-09 Sun Microsystems, Inc. Multiprocessing computer system employing a cluster communication launching and addressing mechanism
US20040022022A1 (en) * 2002-08-02 2004-02-05 Voge Brendan A. Modular system customized by system backplane
US6693901B1 (en) * 2000-04-06 2004-02-17 Lucent Technologies Inc. Backplane configuration without common switch fabric
US6701404B1 (en) * 2000-05-05 2004-03-02 Storage Technology Corporation Method and system for transferring variable sized loop words between elements connected within serial loop through serial interconnect
US20070079046A1 (en) * 2005-09-30 2007-04-05 Tyan Computer Corp. Multiprocessor system
US20070110088A1 (en) * 2005-11-12 2007-05-17 Liquid Computing Corporation Methods and systems for scalable interconnect
US20080046705A1 (en) * 2006-08-15 2008-02-21 Tyan Computer Corporation System and Method for Flexible SMP Configuration
US20080043769A1 (en) * 2006-08-16 2008-02-21 Tyan Computer Corporation Clustering system and system management architecture thereof
US20090043937A1 (en) * 2007-08-08 2009-02-12 Mitac International Corp. Three-Dimensional Interconnection Architecture For Multiprocessor Computer
US20090156031A1 (en) * 2007-12-12 2009-06-18 International Business Machines Corporation Coupler Assembly for a Scalable Computer System and Scalable Computer System

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7339786B2 (en) * 2001-03-05 2008-03-04 Intel Corporation Modular server architecture with Ethernet routed across a backplane utilizing an integrated Ethernet switch module

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6370585B1 (en) * 1997-09-05 2002-04-09 Sun Microsystems, Inc. Multiprocessing computer system employing a cluster communication launching and addressing mechanism
US6693901B1 (en) * 2000-04-06 2004-02-17 Lucent Technologies Inc. Backplane configuration without common switch fabric
US6701404B1 (en) * 2000-05-05 2004-03-02 Storage Technology Corporation Method and system for transferring variable sized loop words between elements connected within serial loop through serial interconnect
US20040022022A1 (en) * 2002-08-02 2004-02-05 Voge Brendan A. Modular system customized by system backplane
US20070079046A1 (en) * 2005-09-30 2007-04-05 Tyan Computer Corp. Multiprocessor system
US20070110088A1 (en) * 2005-11-12 2007-05-17 Liquid Computing Corporation Methods and systems for scalable interconnect
US20080046705A1 (en) * 2006-08-15 2008-02-21 Tyan Computer Corporation System and Method for Flexible SMP Configuration
US20080043769A1 (en) * 2006-08-16 2008-02-21 Tyan Computer Corporation Clustering system and system management architecture thereof
US20090043937A1 (en) * 2007-08-08 2009-02-12 Mitac International Corp. Three-Dimensional Interconnection Architecture For Multiprocessor Computer
US20090156031A1 (en) * 2007-12-12 2009-06-18 International Business Machines Corporation Coupler Assembly for a Scalable Computer System and Scalable Computer System

Cited By (62)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11467883B2 (en) 2004-03-13 2022-10-11 Iii Holdings 12, Llc Co-allocating a reservation spanning different compute resources types
US11652706B2 (en) 2004-06-18 2023-05-16 Iii Holdings 12, Llc System and method for providing dynamic provisioning within a compute environment
US11630704B2 (en) 2004-08-20 2023-04-18 Iii Holdings 12, Llc System and method for a workload management and scheduling module to manage access to a compute environment according to local and non-local user identity information
US11537435B2 (en) 2004-11-08 2022-12-27 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11861404B2 (en) 2004-11-08 2024-01-02 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11762694B2 (en) 2004-11-08 2023-09-19 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11494235B2 (en) 2004-11-08 2022-11-08 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11709709B2 (en) 2004-11-08 2023-07-25 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11537434B2 (en) 2004-11-08 2022-12-27 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11656907B2 (en) 2004-11-08 2023-05-23 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11886915B2 (en) 2004-11-08 2024-01-30 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11658916B2 (en) 2005-03-16 2023-05-23 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US11831564B2 (en) 2005-04-07 2023-11-28 Iii Holdings 12, Llc On-demand access to compute resources
US11533274B2 (en) 2005-04-07 2022-12-20 Iii Holdings 12, Llc On-demand access to compute resources
US11522811B2 (en) 2005-04-07 2022-12-06 Iii Holdings 12, Llc On-demand access to compute resources
US11496415B2 (en) 2005-04-07 2022-11-08 Iii Holdings 12, Llc On-demand access to compute resources
US11765101B2 (en) 2005-04-07 2023-09-19 Iii Holdings 12, Llc On-demand access to compute resources
US11650857B2 (en) 2006-03-16 2023-05-16 Iii Holdings 12, Llc System and method for managing a hybrid computer environment
US11522952B2 (en) 2007-09-24 2022-12-06 The Research Foundation For The State University Of New York Automatic clustering for self-organizing grids
US20090156031A1 (en) * 2007-12-12 2009-06-18 International Business Machines Corporation Coupler Assembly for a Scalable Computer System and Scalable Computer System
US9094317B2 (en) * 2009-06-18 2015-07-28 Hewlett-Packard Development Company, L.P. Processor topology switches
US20120014390A1 (en) * 2009-06-18 2012-01-19 Martin Goldstein Processor topology switches
US9465771B2 (en) 2009-09-24 2016-10-11 Iii Holdings 2, Llc Server on a chip and node cards comprising one or more of same
US9262225B2 (en) 2009-10-30 2016-02-16 Iii Holdings 2, Llc Remote memory access functionality in a cluster of data processing nodes
US9454403B2 (en) 2009-10-30 2016-09-27 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric
US11720290B2 (en) 2009-10-30 2023-08-08 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US9680770B2 (en) 2009-10-30 2017-06-13 Iii Holdings 2, Llc System and method for using a multi-protocol fabric module across a distributed server interconnect fabric
US9749326B2 (en) 2009-10-30 2017-08-29 Iii Holdings 2, Llc System and method for data center security enhancements leveraging server SOCs or server fabrics
US9008079B2 (en) 2009-10-30 2015-04-14 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric
US9866477B2 (en) 2009-10-30 2018-01-09 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric
US9876735B2 (en) 2009-10-30 2018-01-23 Iii Holdings 2, Llc Performance and power optimized computer system architectures and methods leveraging power optimized tree fabric interconnect
US9929976B2 (en) 2009-10-30 2018-03-27 Iii Holdings 2, Llc System and method for data center security enhancements leveraging managed server SOCs
US9054990B2 (en) 2009-10-30 2015-06-09 Iii Holdings 2, Llc System and method for data center security enhancements leveraging server SOCs or server fabrics
US9977763B2 (en) 2009-10-30 2018-05-22 Iii Holdings 2, Llc Network proxy for high-performance, low-power data center interconnect fabric
US9075655B2 (en) 2009-10-30 2015-07-07 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric with broadcast or multicast addressing
US10050970B2 (en) 2009-10-30 2018-08-14 Iii Holdings 2, Llc System and method for data center security enhancements leveraging server SOCs or server fabrics
US10135731B2 (en) 2009-10-30 2018-11-20 Iii Holdings 2, Llc Remote memory access functionality in a cluster of data processing nodes
US10140245B2 (en) 2009-10-30 2018-11-27 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US10877695B2 (en) 2009-10-30 2020-12-29 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US9077654B2 (en) 2009-10-30 2015-07-07 Iii Holdings 2, Llc System and method for data center security enhancements leveraging managed server SOCs
US9509552B2 (en) 2009-10-30 2016-11-29 Iii Holdings 2, Llc System and method for data center security enhancements leveraging server SOCs or server fabrics
US9479463B2 (en) 2009-10-30 2016-10-25 Iii Holdings 2, Llc System and method for data center security enhancements leveraging managed server SOCs
US9311269B2 (en) 2009-10-30 2016-04-12 Iii Holdings 2, Llc Network proxy for high-performance, low-power data center interconnect fabric
US9405584B2 (en) 2009-10-30 2016-08-02 Iii Holdings 2, Llc System and method for high-performance, low-power data center interconnect fabric with addressing and unicast routing
US11526304B2 (en) 2009-10-30 2022-12-13 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US8738828B2 (en) * 2011-02-26 2014-05-27 International Business Machines Corporation System to operationally connect logic nodes
US20120221761A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation Shared system to operationally connect logic nodes
US20120221762A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation Logic node connection system
US20120221760A1 (en) * 2011-02-26 2012-08-30 International Business Machines Corporation System to operationally connect logic nodes
US8589608B2 (en) * 2011-02-26 2013-11-19 International Business Machines Corporation Logic node connection system
US8713228B2 (en) * 2011-02-26 2014-04-29 International Business Machines Corporation Shared system to operationally connect logic nodes
US10021806B2 (en) 2011-10-28 2018-07-10 Iii Holdings 2, Llc System and method for flexible storage and networking provisioning in large scalable processor installations
US9585281B2 (en) 2011-10-28 2017-02-28 Iii Holdings 2, Llc System and method for flexible storage and networking provisioning in large scalable processor installations
US9792249B2 (en) 2011-10-31 2017-10-17 Iii Holdings 2, Llc Node card utilizing a same connector to communicate pluralities of signals
US9965442B2 (en) 2011-10-31 2018-05-08 Iii Holdings 2, Llc Node card management in a modular and large scalable server system
US9069929B2 (en) 2011-10-31 2015-06-30 Iii Holdings 2, Llc Arbitrating usage of serial port in node card of scalable and modular servers
US20130111230A1 (en) * 2011-10-31 2013-05-02 Calxeda, Inc. System board for system and method for modular compute provisioning in large scalable processor installations
US9092594B2 (en) * 2011-10-31 2015-07-28 Iii Holdings 2, Llc Node card management in a modular and large scalable server system
US9648102B1 (en) 2012-12-27 2017-05-09 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
CN105703943A (en) * 2015-12-31 2016-06-22 北京航管软件技术有限公司 Device for switching communication lines
CN112650363A (en) * 2020-12-11 2021-04-13 北京四方继保工程技术有限公司 Distributed telemechanical system based on balanced communication CPU node group and design method
US11960937B2 (en) 2022-03-17 2024-04-16 Iii Holdings 12, Llc System and method for an optimizing reservation in time of compute resources based on prioritization function and reservation policy parameter

Also Published As

Publication number Publication date
US7865655B2 (en) 2011-01-04
WO2007062593A1 (en) 2007-06-07
WO2007062593A8 (en) 2007-11-22
CN201444276U (en) 2010-04-28
CN2852260Y (en) 2006-12-27

Similar Documents

Publication Publication Date Title
US7865655B2 (en) Extended blade server
US7782893B2 (en) Method and apparatus for shared I/O in a load/store fabric
US7046668B2 (en) Method and apparatus for shared I/O in a load/store fabric
EP1950666B1 (en) Method and apparatus for shared I/O in a load/store fabric
JP3783017B2 (en) End node classification using local identifiers
US7533210B2 (en) Virtual communication interfaces for a micro-controller
US6633946B1 (en) Flexible switch-based I/O system interconnect
US8626978B2 (en) Computer apparatus, computer system and adapter carry-over method
US20040024831A1 (en) Blade server management system
US6044207A (en) Enhanced dual port I/O bus bridge
US6917999B2 (en) Platform and method for initializing components within hot-plugged nodes
JPH11316749A (en) Method for identifying redundant data in data storage system
EP1708076B1 (en) Storage system and storage control method
US8074009B2 (en) Sharing of host bus adapter context
KR20020041281A (en) Network switch-integrated high-density multi-sever system
EP0855819A1 (en) Network switch stacking mechanism
US7543179B2 (en) Error management topologies
EP0858036A2 (en) Fibre channel attached storage architecture
US20060294317A1 (en) Symmetric multiprocessor architecture with interchangeable processor and IO modules
JP2923491B2 (en) Cluster system
US7069362B2 (en) Topology for shared memory computer system
Quintero et al. IBM EServer Certification Study Guide: Cluster 1600 Managed by PSSP

Legal Events

Date Code Title Description
AS Assignment

Owner name: HUAWEI TECHNOLOGIES CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LI, MINQIU;HONG, FENG;SHENG, CHUNMING;AND OTHERS;REEL/FRAME:021095/0332;SIGNING DATES FROM 20080521 TO 20080606

Owner name: HUAWEI TECHNOLOGIES CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LI, MINQIU;HONG, FENG;SHENG, CHUNMING;AND OTHERS;SIGNING DATES FROM 20080521 TO 20080606;REEL/FRAME:021095/0332

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552)

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12