[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN105117170A - Computer system architecture - Google Patents

Computer system architecture Download PDF

Info

Publication number
CN105117170A
CN105117170A CN201510524315.0A CN201510524315A CN105117170A CN 105117170 A CN105117170 A CN 105117170A CN 201510524315 A CN201510524315 A CN 201510524315A CN 105117170 A CN105117170 A CN 105117170A
Authority
CN
China
Prior art keywords
information processing
module
computer system
system architecture
link
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510524315.0A
Other languages
Chinese (zh)
Inventor
王磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Inspur Beijing Electronic Information Industry Co Ltd
Original Assignee
Inspur Beijing Electronic Information Industry Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Inspur Beijing Electronic Information Industry Co Ltd filed Critical Inspur Beijing Electronic Information Industry Co Ltd
Priority to CN201510524315.0A priority Critical patent/CN105117170A/en
Publication of CN105117170A publication Critical patent/CN105117170A/en
Pending legal-status Critical Current

Links

Landscapes

  • Information Transfer Systems (AREA)

Abstract

本发明公开了一种计算机系统架构,包括信息处理模块、PCIE交换模块及GPU计算模块,GPU计算模块包括多个GPU,任意两个GPU间均通过NVLink链路进行数据通信;信息处理模块通过PCIE交换模块与GPU计算模块建立PCIE链路连接,信息处理模块与GPU计算模块间通过PICE链路进行数据通信。本申请实现了基于信息处理模块作为管理的冗余设计,提高了系统的冗余度;同时,任意两个GPU间均可直接进行数据通信,且其进行数据通信基于NVLink链路,NVLink链路是每个GPU具备的专有的高速互联通道,其能够提供的带宽能够达到20GB/s,因此,有效提高了GPU间数据传输的效率。

The invention discloses a computer system architecture, including an information processing module, a PCIE exchange module and a GPU computing module, the GPU computing module includes a plurality of GPUs, and any two GPUs perform data communication through an NVLink link; the information processing module passes the PCIE The switching module establishes a PCIE link connection with the GPU computing module, and data communication is performed between the information processing module and the GPU computing module through the PICE link. This application realizes the redundant design based on the information processing module as management, which improves the redundancy of the system; at the same time, any two GPUs can directly perform data communication, and the data communication is based on the NVLink link, and the NVLink link It is a dedicated high-speed interconnect channel for each GPU, which can provide a bandwidth of up to 20GB/s, thus effectively improving the efficiency of data transmission between GPUs.

Description

一种计算机系统架构A computer system architecture

技术领域technical field

本发明涉及计算机技术领域,更具体地说,涉及一种计算机系统架构。The present invention relates to the field of computer technology, and more specifically, to a computer system architecture.

背景技术Background technique

随着目前网络及数据量的爆发式增长,对于数据的高性能计算变得越来越重要。With the explosive growth of the current network and data volume, high-performance computing for data is becoming more and more important.

由于目前的计算机系统架构中通用的处理器并不是专门为对数据的高性能计算设计的,用其来完成对数据的高性能计算存在运算效率低的问题,因此,通常采用处理器搭配GPU(GraphicsProcessingUnit,图形处理器)的方式来完成对于数据的高性能计算。但是,目前的计算机架构中的GPU之间无法直接进行数据传输,而是通过处理器将一个GPU的数据进行处理后传送至另一个GPU以完成两个GPU间的数据传输。由于GPU与处理器间的数据传输通常基于PCIE(PeripheralComponentInterfaceExpress,总线和接口标准)3.0x16链路,而其目前能提供的最大带宽仅仅为16GB/s,并且作为传输媒介的处理器存在运算效率低的问题,因此导致GPU间数据传输的效率较低。Since the general-purpose processors in the current computer system architecture are not specially designed for high-performance computing of data, there is a problem of low computing efficiency when using them to complete high-performance computing of data. Therefore, processors are usually used with GPUs ( GraphicsProcessingUnit, graphics processor) to complete high-performance computing for data. However, in the current computer architecture, data transmission cannot be directly performed between GPUs, but the data of one GPU is processed by the processor and sent to another GPU to complete the data transmission between the two GPUs. Since the data transmission between the GPU and the processor is usually based on the PCIE (Peripheral Component Interface Express, bus and interface standard) 3.0x16 link, the maximum bandwidth it can currently provide is only 16GB/s, and the processor used as the transmission medium has low computing efficiency Therefore, the efficiency of data transmission between GPUs is low.

综上所述,现有技术的计算机系统架构存在GPU间数据传输的效率较低的问题。To sum up, the computer system architecture in the prior art has the problem of low data transmission efficiency between GPUs.

发明内容Contents of the invention

本发明的目的是提供一种计算机系统架构,以解决现有技术的计算机系统架构中存在的GPU间数据传输的效率较低的问题。The purpose of the present invention is to provide a computer system architecture to solve the problem of low efficiency of data transmission between GPUs existing in the computer system architecture of the prior art.

为了实现上述目的,本发明提供如下技术方案:In order to achieve the above object, the present invention provides the following technical solutions:

一种计算机系统架构,包括信息处理模块、PCIE交换模块及GPU计算模块,所述GPU计算模块包括多个GPU,任意两个所述GPU间均通过NVLink链路进行数据通信;所述信息处理模块通过所述PCIE交换模块与所述GPU计算模块建立PCIE链路连接,所述信息处理模块与所述GPU计算模块间通过所述PICE链路进行数据通信。A kind of computer system framework, comprises information processing module, PCIE exchange module and GPU calculation module, described GPU calculation module comprises a plurality of GPUs, all carries out data communication by NVLink link between any two described GPUs; Described information processing module Establishing a PCIE link connection with the GPU computing module through the PCIE switching module, and performing data communication between the information processing module and the GPU computing module through the PICE link.

优选的,所述PCIE交换模块与所述信息处理模块之间通过PCIE3.0x8链路进行数据通信,所述PCIE交换模块与所述GPU计算模块之间通过PCIE3.0x16链路进行数据通信。Preferably, data communication is performed between the PCIE switching module and the information processing module through a PCIE3.0x8 link, and data communication is performed between the PCIE switching module and the GPU computing module through a PCIE3.0x16 link.

优选的,所述GPU计算模块包括L个计算增强单元,所述GPU的数量为4L个,每个所述计算增强单元均包括4个所述GPU,所述PCIE交换模块包括L个PCIE交换单元,所述PCIE交换单元与所述计算增强单元一一对应连接,L为正整数。Preferably, the GPU calculation module includes L calculation enhancement units, the number of GPUs is 4L, each of the calculation enhancement units includes 4 GPUs, and the PCIE switch module includes L PCIE switch units , the PCIE switching unit is connected to the computing enhancement unit in a one-to-one correspondence, and L is a positive integer.

优选的,所述信息处理模块包括信息处理单元,所述信息处理单元具体为X86处理器或者ARM处理器。Preferably, the information processing module includes an information processing unit, and the information processing unit is specifically an X86 processor or an ARM processor.

优选的,所述信息处理单元的数量为两个。Preferably, the number of the information processing units is two.

优选的,所述信息处理模块还包括集中管理单元,所述集中管理单元用于在所述信息处理单元的控制下对所述计算机系统架构中的I/O通讯设备进行集中管理。Preferably, the information processing module further includes a centralized management unit, and the centralized management unit is configured to centrally manage the I/O communication devices in the computer system architecture under the control of the information processing unit.

优选的,所述信息处理模块还包括信息采集单元,所述信息采集单元用于在所述集中管理单元的控制下对所述计算机系统架构的温度及电压进行实时监控。Preferably, the information processing module further includes an information collection unit configured to monitor the temperature and voltage of the computer system architecture in real time under the control of the centralized management unit.

优选的,所述集中管理单元与所述信息处理单元之间通过DMI总线连接,所述集中管理单元与所述信息采集单元之间通过PCIE3.0x1链路连接。Preferably, the centralized management unit is connected to the information processing unit through a DMI bus, and the centralized management unit is connected to the information collection unit through a PCIE3.0x1 link.

优选的,所述计算机系统架构还包括SAS扩展模块,所述SAS扩展模块包括M个物理磁盘,M为正整数;Preferably, the computer system architecture further includes a SAS extension module, the SAS extension module includes M physical disks, and M is a positive integer;

所述SAS扩展模块通过PCIE3.0x8链路与所述信息处理模块连接,用于在所述信息处理模块的控制下完成数据的存储与读取。The SAS expansion module is connected to the information processing module through a PCIE3.0x8 link, and is used to complete data storage and reading under the control of the information processing module.

优选的,所述信息处理模块还包括磁盘存储阵列,所述磁盘存储阵列包括N个物理磁盘,N为正整数;Preferably, the information processing module further includes a disk storage array, and the disk storage array includes N physical disks, where N is a positive integer;

所述磁盘存储阵列通过SAS信号与所述SAS扩展模块进行数据通信。The disk storage array performs data communication with the SAS expansion module through SAS signals.

本发明提供的一种计算机系统架构,包括信息处理模块、PCIE交换模块及GPU计算模块,GPU计算模块包括多个GPU,任意两个GPU间均通过NVLink链路进行数据通信;信息处理模块通过PCIE交换模块与GPU计算模块建立PCIE链路连接,信息处理模块与GPU计算模块间通过PICE链路进行数据通信。与现有技术相比,首先,本申请实现了基于信息处理模块作为管理的冗余设计,提高了系统的冗余度;其次,实现了GPU间的全互连,即任意两个GPU间均可直接进行数据通信,且其进行数据通信基于NVLink链路,NVLink链路是每个GPU具备的专有的高速互联通道,其能够提供的带宽能够达到20GB/s,比现有技术中GPU与处理器间进行数据通信基于的PCIE3.0x16链路的带宽高出25%,因此,实现了基于点对点模式的数据高速传输,即有效提高了GPU间数据传输的效率。A computer system architecture provided by the present invention includes an information processing module, a PCIE exchange module and a GPU computing module, the GPU computing module includes a plurality of GPUs, and any two GPUs perform data communication through an NVLink link; the information processing module passes the PCIE The switching module establishes a PCIE link connection with the GPU computing module, and data communication is performed between the information processing module and the GPU computing module through the PICE link. Compared with the prior art, firstly, this application realizes the redundant design based on the information processing module as management, which improves the redundancy of the system; secondly, realizes the full interconnection between GPUs, that is, any two GPUs are It can directly carry out data communication, and its data communication is based on the NVLink link. The NVLink link is a proprietary high-speed interconnection channel for each GPU, and the bandwidth it can provide can reach 20GB/s. The bandwidth of the PCIE3.0x16 link based on data communication between processors is 25% higher. Therefore, high-speed data transmission based on point-to-point mode is realized, which effectively improves the efficiency of data transmission between GPUs.

附图说明Description of drawings

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据提供的附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the accompanying drawings in the following description are only It is an embodiment of the present invention, and those skilled in the art can also obtain other drawings according to the provided drawings on the premise of not paying creative efforts.

图1为本发明实施例提供的一种计算机系统架构的结构示意图;FIG. 1 is a schematic structural diagram of a computer system architecture provided by an embodiment of the present invention;

图2为本发明实施例提供的一种计算机系统架构中当L为2时的结构示意图;FIG. 2 is a schematic structural diagram when L is 2 in a computer system architecture provided by an embodiment of the present invention;

图3为本发明实施例提供的一种计算机系统架构中当L为2时的模型图;FIG. 3 is a model diagram when L is 2 in a computer system architecture provided by an embodiment of the present invention;

图4为本发明实施例提供的一种计算机系统架构中信息处理模块的架构示意图。FIG. 4 is a schematic diagram of an information processing module in a computer system architecture provided by an embodiment of the present invention.

具体实施方式Detailed ways

下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

请参阅图1,其示出了本发明实施例提供的一种计算机系统架构的结构示意图,该计算机系统架构可以包括信息处理模块1、PCIE交换模块2及GPU计算模块3,GPU计算模块3包括多个GPU,任意两个GPU间均通过NVLink链路进行数据通信;信息处理模块1通过PCIE交换模块2与GPU计算模块3建立PCIE链路连接,信息处理模块1与GPU计算模块3间通过PICE链路进行数据通信。Please refer to Fig. 1, which shows a schematic structural diagram of a computer system architecture provided by an embodiment of the present invention, the computer system architecture can include an information processing module 1, a PCIE switching module 2 and a GPU computing module 3, and the GPU computing module 3 includes Multiple GPUs, data communication between any two GPUs through the NVLink link; the information processing module 1 establishes a PCIE link connection with the GPU computing module 3 through the PCIE switching module 2, and the information processing module 1 and the GPU computing module 3 pass the PICE link for data communication.

其中,GPU的数量可根据实际需要进行确定。Wherein, the number of GPUs may be determined according to actual needs.

首先,本申请实现了基于信息处理模块作为管理的冗余设计,提高了系统的冗余度;其次,实现了GPU间的全互连,即任意两个GPU间均可直接进行数据通信,且其进行数据通信基于NVLink链路,NVLink链路是每个GPU具备的专有的高速互联通道,其能够提供的带宽能够达到20GB/s,比现有技术中GPU与处理器间进行数据通信基于的PCIE3.0x16链路的带宽高出25%,因此,实现了基于点对点模式的数据高速传输,即有效提高了GPU间数据传输的效率。First of all, this application realizes the redundant design based on the information processing module as management, which improves the redundancy of the system; secondly, it realizes the full interconnection between GPUs, that is, any two GPUs can directly perform data communication, and Its data communication is based on the NVLink link. The NVLink link is a proprietary high-speed interconnection channel for each GPU. The bandwidth it can provide can reach 20GB/s, which is higher than the data communication between the GPU and the processor in the prior art. The bandwidth of the PCIE3.0x16 link is 25% higher. Therefore, high-speed data transmission based on point-to-point mode is realized, which effectively improves the efficiency of data transmission between GPUs.

其中,GPU计算模块中的GPU属于数据的高性能计算器件,其主要为系统提供高强度的浮点计算性能。信息处理模块是整个系统的信息处理核心,由其控制GPU计算模块中的GPU完成相关工作。而PCIE交换单元能够起到一个传输中介的作用,其能够使得信息处理模块与GPU计算模块之间通过PCIE交换模块进行间接连接,进而使得信息处理单元通过PCIE信息控制GPU计算模块完成相关工作。Among them, the GPU in the GPU computing module is a high-performance computing device for data, which mainly provides high-intensity floating-point computing performance for the system. The information processing module is the information processing core of the whole system, and it controls the GPU in the GPU computing module to complete related work. The PCIE switching unit can act as a transmission intermediary, which can make the information processing module and the GPU computing module indirectly connected through the PCIE switching module, and then make the information processing unit control the GPU computing module to complete related work through the PCIE information.

需要说明的是,深度学习是机器学习研究中的一个新的领域,其动机在于建立和模拟人脑进行分析学习的神经网络,它模仿人脑的机制来解释数据,例如图像,声音和文本。深度学习就包括对于数据的高性能计算,如对一张图片进行特征提取需要170G次浮点计算,120万张图片进行一次网络训练,则需要318P的浮点运算,如果由一个CPU(CentraMProcessingUnit,中央处理器)完成上述计算量需要22天的时间,而如果由一个GPU来完成上述工作量大概仅需要18个小时的时间,因此本发明中采用GPU进行数据的高性能计算具有计算效率高的优点。It should be noted that deep learning is a new field in machine learning research, and its motivation is to establish and simulate the neural network of the human brain for analysis and learning, which imitates the mechanism of the human brain to interpret data, such as images, sounds and texts. Deep learning includes high-performance computing for data. For example, feature extraction for a picture requires 170G floating-point calculations, and 1.2 million pictures for a network training requires 318P floating-point calculations. If a CPU (CentraMProcessingUnit, Central processing unit) needs 22 days to complete the above-mentioned workload, and if a GPU is used to complete the above-mentioned workload, it only needs about 18 hours. Therefore, the high-performance computing of data using GPU has the advantages of high computing efficiency in the present invention. advantage.

优选的,本发明实施例提供的一种计算机系统架构中,PCIE交换模块与信息处理模块之间可以通过PCIE3.0x8链路进行数据通信,PCIE交换模块与GPU计算模块之间可以通过PCIE3.0x16链路进行数据通信。Preferably, in a kind of computer system framework that the embodiment of the present invention provides, can carry out data communication through PCIE3.0x8 link between PCIE switch module and information processing module, can pass PCIE3.0x16 link for data communication.

需要说明的是,PCIE3.0x8链路和PCIE3.0x16链路均属于PCIE链路,而PCIE3.0x8链路比较符合PCIE交换模块与信息处理模块之间进行数据通信的特点,而PCIE3.0x16链路比较符合PCIE交换模块与GPU计算模块之间进行数据通信的特点,且,实验证明,采用上述链路进行相关模块间的通信,既节约了带宽资源,又能达到数据的正常通信。It should be noted that both the PCIE3.0x8 link and the PCIE3.0x16 link belong to the PCIE link, and the PCIE3.0x8 link is more in line with the characteristics of data communication between the PCIE switching module and the information processing module, while the PCIE3.0x16 link The path is more in line with the characteristics of data communication between the PCIE switching module and the GPU computing module, and the experiment proves that using the above link for communication between related modules not only saves bandwidth resources, but also achieves normal communication of data.

优选的,上述实施例提供的一种计算机系统架构,GPU计算模块可以包括L个计算增强单元,GPU的数量可以为4L个,每个计算增强单元均可以包括4个GPU,PCIE交换模块可以包括L个PCIE交换单元,PCIE交换单元与计算增强单元一一对应连接,L为正整数。Preferably, a kind of computer system framework that above-mentioned embodiment provides, GPU calculation module can comprise L calculation enhancement unit, the quantity of GPU can be 4L, and each calculation enhancement unit can comprise 4 GPUs, and PCIE switch module can comprise L PCIE switching units, the PCIE switching units are connected to the computing enhancement units in one-to-one correspondence, and L is a positive integer.

需要说明的是,L的具体数值可根据实际需要进行确定。当GPU的数量为4L时,为了在实际生产时合理分配各模块的排布,可设置成每4个GPU属于同一计算增强单元,由此,GPU计算模块则包括L个计算增强单元,每个计算增强单元均通过一个PCIE交换单元与信息处理模块连接。举例说明,当L为2时,本发明实施例提供的一种计算机系统架构的示意图如图2所示,计算机系统架构的模型图如图3所示,包括信息处理模块1、PCIE交换模块及GPU计算模块,其中,PCIE交换模块包括PCIE交换单元21,GPU计算模块包括计算增强单元31。It should be noted that the specific value of L can be determined according to actual needs. When the number of GPUs is 4L, in order to reasonably allocate the arrangement of each module in actual production, it can be set that every 4 GPUs belong to the same computing enhancement unit, thus, the GPU computing module includes L computing enhancement units, each The computing enhancement units are all connected to the information processing module through a PCIE switching unit. For example, when L is 2, a schematic diagram of a computer system architecture provided by an embodiment of the present invention is shown in Figure 2, and a model diagram of the computer system architecture is shown in Figure 3, including an information processing module 1, a PCIE switching module and A GPU computing module, wherein the PCIE switching module includes a PCIE switching unit 21, and the GPU computing module includes a computing enhancement unit 31.

由此,能够在实际生产时合理分配各模块的排布,避免因排布过于杂乱在系统检测出错误时无法准确定位错误的部件。As a result, the layout of each module can be reasonably allocated during actual production, and it is avoided that the wrong component cannot be accurately positioned when the system detects an error due to the messy layout.

优选的,上述实施例提供的一种计算机系统架构,信息处理模块可以包括信息处理单元,信息处理单元具体可以为X86处理器或者ARM(AdvancedRISC(ReducedInstructionSetComputer,精简指令集计算机)Machines,先进的RISC机器)处理器。Preferably, in the computer system architecture provided by the foregoing embodiments, the information processing module may include an information processing unit, and the information processing unit may specifically be an X86 processor or an ARM (Advanced RISC (Reduced Instruction Set Computer, reduced instruction set computer) Machines, an advanced RISC machine )processor.

信息处理单元是信息处理模块的核心,其完成信息处理单元的主要工作,而X86处理器或者ARM处理器,无论是数据处理能力还是工作效率方面都是比较优秀的,符合本发明实施例中对于信息处理单元的要求。The information processing unit is the core of the information processing module, which completes the main work of the information processing unit, and the X86 processor or ARM processor is relatively excellent in terms of data processing capability and work efficiency, which is consistent with the requirements of the embodiment of the present invention. Information processing unit requirements.

优选的,上述实施例提供的一种计算机系统架构,信息处理单元的数量可以为两个。Preferably, in the computer system architecture provided by the above embodiment, the number of information processing units may be two.

由此,当本发明实施例提供的一种计算机系统架构正常工作时,信息处理模块的工作能够由两个信息处理单元同时完成,避免信息处理单元因过度工作导致的损坏等;并且,当任一个信息处理单元发生故障时,信息处理模块的工作能够由另一个信息处理单元完成,以避免影响计算机系统架构的整体运行。Therefore, when the computer system architecture provided by the embodiment of the present invention works normally, the work of the information processing module can be completed by two information processing units at the same time, avoiding the damage of the information processing unit due to overwork; and, when any When an information processing unit fails, the work of the information processing module can be completed by another information processing unit, so as to avoid affecting the overall operation of the computer system architecture.

优选的,上述实施例提供的一种计算机系统架构,信息处理模块还可以包括集中管理单元,集中管理单元用于在信息处理单元的控制下对计算机系统架构中的I/O(input/output,输入/输出)通讯设备进行集中管理。Preferably, in the computer system architecture provided by the above-mentioned embodiments, the information processing module may also include a centralized management unit, and the centralized management unit is used to control the I/O (input/output, input/output) communication equipment for centralized management.

由此,集中管理单元在信息处理单元的控制下对I/O通讯设备进行集中管理,能够根据需要选择I/O通讯设备的工作状态,并且在任意I/O通讯设备发生故障时及时获知,以避免由于I/O通讯设备的长时间故障影响系统的整体运行。Thus, the centralized management unit can centrally manage the I/O communication equipment under the control of the information processing unit, can select the working status of the I/O communication equipment according to the needs, and know in time when any I/O communication equipment fails, In order to avoid affecting the overall operation of the system due to the long-term failure of I/O communication equipment.

优选的,上述实施例提供的一种计算机系统架构,信息处理模块还可以包括信息采集单元,信息采集单元用于在集中管理单元的控制下对计算机系统架构的温度及电压进行实时监控。Preferably, in the computer system architecture provided by the above embodiment, the information processing module may further include an information collection unit, which is used to monitor the temperature and voltage of the computer system architecture in real time under the control of the centralized management unit.

信息采集单元能够实时监控计算机系统架构的温度及电压,包括对于计算机系统架构中所有器件的监控,由此,当任意器件的温度不符合正常范围或者当任意器件的电压不符合正常范围时及时获知,以避免由于任意部件的温度或电压不符合正常范围而造成其成为故障电压,进而影响系统的整体运行。The information acquisition unit can monitor the temperature and voltage of the computer system architecture in real time, including the monitoring of all devices in the computer system architecture, so that when the temperature of any device does not meet the normal range or when the voltage of any device does not meet the normal range, it is promptly informed In order to avoid the temperature or voltage of any component not meeting the normal range and causing it to become a fault voltage, thereby affecting the overall operation of the system.

优选的,上述实施例提供的一种计算机系统架构,集中管理单元与信息处理单元之间可以通过DMI(DirectMediaInterface,直接媒体接口)总线连接,集中管理单元与信息采集单元之间可以通过PCIE3.0x1链路连接。Preferably, in the computer system architecture provided by the foregoing embodiments, the centralized management unit and the information processing unit can be connected through a DMI (DirectMediaInterface, direct media interface) bus, and the centralized management unit and the information collection unit can be connected through a PCIE3.0x1 link connection.

集中管理单元通过DMI总线接收信息处理单元发送的指令,以根据指令对I/O通讯设备进行集中管理,而信息采集单元通过PCIE3.0x1链路接收集中管理单元发送的指令,以根据指令实时监控计算机系统架构的温度及电压。其中,信息采集单元具体可以采用ASP2400系列芯片组。实验证明,上述连接方式符合相关部件之间进行数据通信的特点,在节约带宽资源的情况下能够正常发挥数据通信作用。The centralized management unit receives the instructions sent by the information processing unit through the DMI bus to centrally manage the I/O communication equipment according to the instructions, and the information acquisition unit receives the instructions sent by the centralized management unit through the PCIE3.0x1 link to monitor in real time according to the instructions Temperature and voltage of computer system architecture. Wherein, the information collection unit may specifically adopt an ASP2400 series chipset. Experiments have proved that the above connection mode conforms to the characteristics of data communication between related components, and can normally play a role in data communication while saving bandwidth resources.

当信息处理模块包括信息处理单元11、集中管理单元12及信息采集单元13时,其结构示意图请参阅图3所示。When the information processing module includes an information processing unit 11 , a centralized management unit 12 and an information collection unit 13 , its structural diagram is shown in FIG. 3 .

优选的,上述实施例提供的一种计算机系统架构,计算机系统架构还可以包括SAS(SerialconnectionSCSI(SmallComputerSystemInterface,小型计算机系统接口),串行连接SCSI)扩展模块,SAS扩展模块包括M个物理磁盘,M为正整数;Preferably, a kind of computer system framework that above-mentioned embodiment provides, computer system framework can also comprise SAS (SerialconnectionSCSI (SmallComputerSystemInterface, small computer system interface), serial connection SCSI) expansion module, SAS expansion module comprises M physical disks, M is a positive integer;

SAS扩展模块通过PCIE3.0x8链路与信息处理模块连接,用于在信息处理模块的控制下完成数据的存储与读取。The SAS expansion module is connected to the information processing module through a PCIE3.0x8 link, and is used to complete data storage and reading under the control of the information processing module.

需要说明的是,M的具体数值可根据实际需要进行确定。实际设置时,可以是SAS扩展模块的前端用于安装物理磁盘,后端通过硬盘背板连接至系统的背板上,再由系统的背板连接至信息处理模块上。信息处理模块还可以包括磁盘控制器,用于控制SAS扩展模块的物理磁盘中数据的存储与读写。It should be noted that the specific value of M can be determined according to actual needs. In actual setup, the front end of the SAS expansion module can be used to install physical disks, and the rear end is connected to the system backplane through the hard disk backplane, and then connected to the information processing module by the system backplane. The information processing module may also include a disk controller, which is used to control the storage and reading and writing of data in the physical disk of the SAS expansion module.

由此,通过为计算机系统架构增加SAS扩展模块,可以增强整个系统的数据存储能力,完善系统的功能。Therefore, by adding a SAS expansion module to the computer system architecture, the data storage capacity of the entire system can be enhanced and the system's functions can be improved.

优选的,上述实施例提供的一种计算机系统架构,信息处理模块还可以包括磁盘存储阵列,磁盘存储阵列包括N个物理磁盘,N为正整数;Preferably, in the computer system architecture provided in the foregoing embodiment, the information processing module may further include a disk storage array, and the disk storage array includes N physical disks, where N is a positive integer;

磁盘存储阵列通过SAS信号与SAS扩展模块进行数据通信。The disk storage array communicates with the SAS expansion module through SAS signals.

需要说明的是,N的具体数值可根据实际需要进行确定。例如,该磁盘存储阵列可以包括4个2.5寸的物理磁盘。It should be noted that the specific value of N can be determined according to actual needs. For example, the disk storage array may include four 2.5-inch physical disks.

磁盘存储阵列通过SAS信号与SAS扩展模块进行数据通信,SAS扩展模块接收信息处理模块发送的控制信息,并将其转换成基于SAS信号的数据后将其传送至磁盘存储阵列,以控制磁盘存储阵列中物理磁盘的数据的存储和读写。由此,通过磁盘存储阵列与SAS扩展模块相互配合,增强了系统的数据存储能力,并且在其中一方发生故障时,还可以由另一方完成数据的存储和读写,保证了系统的正常运行。The disk storage array communicates with the SAS expansion module through the SAS signal. The SAS expansion module receives the control information sent by the information processing module, converts it into data based on the SAS signal, and then transmits it to the disk storage array to control the disk storage array. Storage and reading and writing of data on physical disks. Therefore, the data storage capacity of the system is enhanced through the cooperation of the disk storage array and the SAS expansion module, and when one of them fails, the other can complete the data storage, reading and writing, ensuring the normal operation of the system.

另外,通过SAS扩展模块中的物理磁盘还可以组建RAID0或RAID5,以满足对于数据高速访问或冗余性的需求。In addition, RAID0 or RAID5 can also be formed through the physical disks in the SAS expansion module to meet the requirements for high-speed data access or redundancy.

对所公开的实施例的上述说明,使本领域技术人员能够实现或使用本发明。对这些实施例的多种修改对本领域技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本发明的精神或范围的情况下,在其它实施例中实现。因此,本发明将不会被限制于本文所示的这些实施例,而是要符合与本文所公开的原理和新颖特点相一致的最宽的范围。The above description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be implemented in other embodiments without departing from the spirit or scope of the invention. Therefore, the present invention will not be limited to the embodiments shown herein, but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1.一种计算机系统架构,其特征在于,包括信息处理模块、PCIE交换模块及GPU计算模块,所述GPU计算模块包括多个GPU,任意两个所述GPU间均通过NVLink链路进行数据通信;所述信息处理模块通过所述PCIE交换模块与所述GPU计算模块建立PCIE链路连接,所述信息处理模块与所述GPU计算模块间通过所述PICE链路进行数据通信。1. A kind of computer system framework, it is characterized in that, comprise information processing module, PCIE exchange module and GPU calculation module, described GPU calculation module comprises a plurality of GPUs, all carry out data communication by NVLink link between any two described GPUs The information processing module establishes a PCIE link connection with the GPU computing module through the PCIE switching module, and the information processing module and the GPU computing module perform data communication through the PICE link. 2.根据权利要求1所述的计算机系统架构,其特征在于,所述PCIE交换模块与所述信息处理模块之间通过PCIE3.0x8链路进行数据通信,所述PCIE交换模块与所述GPU计算模块之间通过PCIE3.0x16链路进行数据通信。2. The computer system architecture according to claim 1, wherein data communication is carried out between the PCIE switching module and the information processing module through a PCIE3.0x8 link, and the PCIE switching module and the GPU computing The data communication between the modules is carried out through the PCIE3.0x16 link. 3.根据权利要求2所述的计算机系统架构,其特征在于,所述GPU计算模块包括L个计算增强单元,所述GPU的数量为4L个,每个所述计算增强单元均包括4个所述GPU,所述PCIE交换模块包括L个PCIE交换单元,所述PCIE交换单元与所述计算增强单元一一对应连接,L为正整数。3. The computer system architecture according to claim 2, wherein the GPU calculation module includes L calculation enhancement units, the number of the GPUs is 4L, and each of the calculation enhancement units includes 4 calculation enhancement units. The GPU, the PCIE switching module includes L PCIE switching units, the PCIE switching units are connected to the computing enhancement unit in a one-to-one correspondence, and L is a positive integer. 4.根据权利要求1所述的计算机系统架构,其特征在于,所述信息处理模块包括信息处理单元,所述信息处理单元具体为X86处理器或者ARM处理器。4. The computer system architecture according to claim 1, wherein the information processing module comprises an information processing unit, and the information processing unit is specifically an X86 processor or an ARM processor. 5.根据权利要求4所述的计算机系统架构,其特征在于,所述信息处理单元的数量为两个。5. The computer system architecture according to claim 4, wherein the number of the information processing units is two. 6.根据权利要求4所述的计算机系统架构,其特征在于,所述信息处理模块还包括集中管理单元,所述集中管理单元用于在所述信息处理单元的控制下对所述计算机系统架构中的I/O通讯设备进行集中管理。6. The computer system architecture according to claim 4, wherein the information processing module further comprises a centralized management unit configured to control the computer system architecture under the control of the information processing unit Centralized management of the I/O communication devices in the system. 7.根据权利要求6所述的计算机系统架构,其特征在于,所述信息处理模块还包括信息采集单元,所述信息采集单元用于在所述集中管理单元的控制下对所述计算机系统架构的温度及电压进行实时监控。7. The computer system framework according to claim 6, wherein the information processing module further comprises an information collection unit, and the information collection unit is used to control the computer system framework under the control of the centralized management unit Real-time monitoring of temperature and voltage. 8.根据权利要求7所述的计算机系统架构,其特征在于,所述集中管理单元与所述信息处理单元之间通过DMI总线连接,所述集中管理单元与所述信息采集单元之间通过PCIE3.0x1链路连接。8. The computer system architecture according to claim 7, characterized in that, the centralized management unit and the information processing unit are connected by a DMI bus, and the centralized management unit and the information collection unit are connected by a PCIE3 .0x1 link connection. 9.根据权利要求1至8任一项所述的计算机系统架构,其特征在于,所述计算机系统架构还包括SAS扩展模块,所述SAS扩展模块包括M个物理磁盘,M为正整数;9. The computer system architecture according to any one of claims 1 to 8, wherein the computer system architecture also includes a SAS extension module, the SAS extension module includes M physical disks, and M is a positive integer; 所述SAS扩展模块通过PCIE3.0x8链路与所述信息处理模块连接,用于在所述信息处理模块的控制下完成数据的存储与读取。The SAS expansion module is connected to the information processing module through a PCIE3.0x8 link, and is used to complete data storage and reading under the control of the information processing module. 10.根据权利要求9所述的计算机系统架构,其特征在于,所述信息处理模块还包括磁盘存储阵列,所述磁盘存储阵列包括N个物理磁盘,N为正整数;10. The computer system architecture according to claim 9, wherein the information processing module further includes a disk storage array, and the disk storage array includes N physical disks, where N is a positive integer; 所述磁盘存储阵列通过SAS信号与所述SAS扩展模块进行数据通信。The disk storage array performs data communication with the SAS expansion module through SAS signals.
CN201510524315.0A 2015-08-24 2015-08-24 Computer system architecture Pending CN105117170A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510524315.0A CN105117170A (en) 2015-08-24 2015-08-24 Computer system architecture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510524315.0A CN105117170A (en) 2015-08-24 2015-08-24 Computer system architecture

Publications (1)

Publication Number Publication Date
CN105117170A true CN105117170A (en) 2015-12-02

Family

ID=54665173

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510524315.0A Pending CN105117170A (en) 2015-08-24 2015-08-24 Computer system architecture

Country Status (1)

Country Link
CN (1) CN105117170A (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106201983A (en) * 2016-07-15 2016-12-07 浪潮(北京)电子信息产业有限公司 A kind of computer system architecture
CN108052482A (en) * 2017-11-13 2018-05-18 华中科技大学 A kind of method and system to communicate between GPU
CN109032976A (en) * 2018-07-25 2018-12-18 郑州云海信息技术有限公司 A kind for the treatment of method and apparatus of physical channel
CN109342930A (en) * 2018-11-12 2019-02-15 郑州云海信息技术有限公司 A GPU test verification board
TWI658365B (en) * 2017-10-30 2019-05-01 緯創資通股份有限公司 Connecting module
WO2019116116A1 (en) * 2017-12-15 2019-06-20 Ati Technologies Ulc Parallel data transfer to increase bandwidth for accelerated processing devices
CN109933430A (en) * 2019-03-08 2019-06-25 北京百度网讯科技有限公司 The method and apparatus for distributing graphics processor
CN110166341A (en) * 2019-07-01 2019-08-23 联想(北京)有限公司 A kind of server, server cluster and communication means
US10423429B2 (en) 2018-01-02 2019-09-24 International Business Machines Corporation Reconfiguring processing groups for cascading data workloads
US10747280B2 (en) 2018-11-27 2020-08-18 International Business Machines Corporation Reconfigurble CPU/GPU interconnect to mitigate power/thermal throttling
WO2020199560A1 (en) * 2019-04-03 2020-10-08 华为技术有限公司 Ai training network and method
US10831688B2 (en) 2018-08-21 2020-11-10 International Business Machines Corporation Reconfigurable network infrastructure
CN113672541A (en) * 2021-07-13 2021-11-19 中国人民解放军战略支援部队航天工程大学 A GPU-based implementation method for incoherent demodulation of PCM/FM telemetry signals
CN115801582A (en) * 2022-09-27 2023-03-14 海光信息技术股份有限公司 System link bandwidth improving method, related device and computer equipment
CN107391428B (en) * 2017-07-31 2024-02-13 郑州云海信息技术有限公司 Novel framework four-way server

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1983226A (en) * 2005-12-15 2007-06-20 威盛电子股份有限公司 Method and system for supporting multiple graphic processing units
CN104813309A (en) * 2012-11-28 2015-07-29 辉达公司 Method and system for cloud based virtualized graphics processing for remote displays

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1983226A (en) * 2005-12-15 2007-06-20 威盛电子股份有限公司 Method and system for supporting multiple graphic processing units
CN104813309A (en) * 2012-11-28 2015-07-29 辉达公司 Method and system for cloud based virtualized graphics processing for remote displays

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
D.FOLEY: "NVLink,Pascal and Stacked Memory:Feeding the Appetite for Big Data", 《HTTP://DEVBLOGS.NVIDIA.COM/PARALLELFORALL/NVLINK-PASCAL-STACKED-MEMORY-FEEDING-APPETITE-BIG-DATA》 *
路松行主编: "《简明微机组装与维护教程》", 31 August 2013, 西安:西安电子科技大学出版社 *

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106201983A (en) * 2016-07-15 2016-12-07 浪潮(北京)电子信息产业有限公司 A kind of computer system architecture
CN106201983B (en) * 2016-07-15 2019-04-23 浪潮(北京)电子信息产业有限公司 a computer system
CN107391428B (en) * 2017-07-31 2024-02-13 郑州云海信息技术有限公司 Novel framework four-way server
TWI658365B (en) * 2017-10-30 2019-05-01 緯創資通股份有限公司 Connecting module
CN108052482B (en) * 2017-11-13 2020-05-19 华中科技大学 Method and system for communication between GPUs
CN108052482A (en) * 2017-11-13 2018-05-18 华中科技大学 A kind of method and system to communicate between GPU
CN111417936A (en) * 2017-12-15 2020-07-14 Ati科技无限责任公司 Parallel data transfer for increasing bandwidth of accelerated processing device
US11276135B2 (en) 2017-12-15 2022-03-15 Ati Technologies Ulc Parallel data transfer to increase bandwidth for accelerated processing devices
CN111417936B (en) * 2017-12-15 2024-06-04 Ati科技无限责任公司 Parallel data transfer for increasing bandwidth of an accelerated processing device
US10672095B2 (en) 2017-12-15 2020-06-02 Ati Technologies Ulc Parallel data transfer to increase bandwidth for accelerated processing devices
WO2019116116A1 (en) * 2017-12-15 2019-06-20 Ati Technologies Ulc Parallel data transfer to increase bandwidth for accelerated processing devices
US10423429B2 (en) 2018-01-02 2019-09-24 International Business Machines Corporation Reconfiguring processing groups for cascading data workloads
US11132231B2 (en) 2018-01-02 2021-09-28 International Business Machines Corporation Reconfiguring processing groups for cascading data workloads
CN109032976A (en) * 2018-07-25 2018-12-18 郑州云海信息技术有限公司 A kind for the treatment of method and apparatus of physical channel
US10831688B2 (en) 2018-08-21 2020-11-10 International Business Machines Corporation Reconfigurable network infrastructure
CN109342930A (en) * 2018-11-12 2019-02-15 郑州云海信息技术有限公司 A GPU test verification board
US10747280B2 (en) 2018-11-27 2020-08-18 International Business Machines Corporation Reconfigurble CPU/GPU interconnect to mitigate power/thermal throttling
CN109933430A (en) * 2019-03-08 2019-06-25 北京百度网讯科技有限公司 The method and apparatus for distributing graphics processor
WO2020199560A1 (en) * 2019-04-03 2020-10-08 华为技术有限公司 Ai training network and method
CN110166341B (en) * 2019-07-01 2021-08-17 联想(北京)有限公司 Server, server cluster and communication method
CN110166341A (en) * 2019-07-01 2019-08-23 联想(北京)有限公司 A kind of server, server cluster and communication means
CN113672541A (en) * 2021-07-13 2021-11-19 中国人民解放军战略支援部队航天工程大学 A GPU-based implementation method for incoherent demodulation of PCM/FM telemetry signals
CN115801582A (en) * 2022-09-27 2023-03-14 海光信息技术股份有限公司 System link bandwidth improving method, related device and computer equipment
CN115801582B (en) * 2022-09-27 2024-05-31 海光信息技术股份有限公司 System link bandwidth improvement method, related devices and computer equipment

Similar Documents

Publication Publication Date Title
CN105117170A (en) Computer system architecture
CN105955898A (en) Hard disk backboard compatible with SAS hard disk and NVMe hard disk
CN107688376A (en) A kind of hard disk backboard for supporting adaptive hard-disk interface
CN111901164A (en) Adaptive control method, device, equipment and system for OCP NIC network card
WO2024078223A1 (en) Electronic device, and pcie topology configuration method and apparatus thereof
CN204833236U (en) Support memory system of hybrid storage
CN107092570A (en) The adaptive configuring method and system of a kind of onboard M.2 hard disk of server
CN205507751U (en) Storage mainboard
CN207408936U (en) A kind of multiplex roles PCIE device adapter
CN207232854U (en) A kind of hard disk backboard for supporting adaptive hard-disk interface
CN206696771U (en) One kind is based on the road server master board of Purley platforms four
CN105138494B (en) A kind of multichannel computer system
CN108280036A (en) A kind of storage hard disk backboard for supporting NVME
CN211628241U (en) A PCIE bus structure with software switching topology
CN115757236A (en) Storage expansion pooling equipment and system
CN204705946U (en) A kind of novel 4 road server systems based on fusion architecture
CN107748726A (en) A kind of GPU casees
CN106020737A (en) Globally-shared-disc high-density storage frame system
CN204557308U (en) Novel high-density blade server based on fusion framework
CN108121429A (en) A kind of storage biplane electric power system
US20220342835A1 (en) Method and apparatus for disaggregation of computing resources
CN204166522U (en) A kind of high-speed high capacity FLASH veneer memory circuit plate
KR20120134918A (en) Electronic apparatus including a plurality of processors
CN107832006A (en) More main frames share the multi-functional fdisk storage device of same expanding storage apparatus
CN205050007U (en) Storage expansion card in blade server based on fusion framework

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20151202