CN104866457A - On-chip multi-core processor static architecture based on shared cache - Google Patents

On-chip multi-core processor static architecture based on shared cache Download PDF

Info

Publication number
CN104866457A
CN104866457A CN201510302580.4A CN201510302580A CN104866457A CN 104866457 A CN104866457 A CN 104866457A CN 201510302580 A CN201510302580 A CN 201510302580A CN 104866457 A CN104866457 A CN 104866457A
Authority
CN
China
Prior art keywords
node
category
buffer memory
nodes
cache
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510302580.4A
Other languages
Chinese (zh)
Other versions
CN104866457B (en
Inventor
李嵩
褚廷斌
黄乐天
袁正希
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201510302580.4A priority Critical patent/CN104866457B/en
Publication of CN104866457A publication Critical patent/CN104866457A/en
Application granted granted Critical
Publication of CN104866457B publication Critical patent/CN104866457B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Memory System Of A Hierarchy Structure (AREA)

Abstract

The invention discloses an on-chip multi-core processor static architecture based on a shared cache comprising n nodes, wherein n is an even number, which is more than 0, and the nodes comprise n/2 class-A nodes and n/2 class-B nodes; each class-A node comprises a processing core, a local privately-owned cache L1 and a router R; each class-B node comprises a processing core, a local privately-owned cache L1, a router R and a shared cache l2bank; the nodes communicate with each other via the routers; the class-A nodes and the class-B nodes are distributed at intervals; the capacity of the shared cache l2bank is twice as the shared cache in a universal framework. The on-chip multi-core processor static architecture based on the shared cache provided by the invention saves hardware expenditure and area of the last level cache on the basis of not adding access delay of a whole on-chip network and congestion extent of the whole on-chip network to save static power consumption of the cache.

Description

A kind of chip multi-core processor static framework based on shared buffer memory
Technical field
The present invention relates to a kind of chip multi-core processor static framework based on shared buffer memory.
Background technology
As shown in Figure 1, a kind of conventional network-on-chip polycaryon processor framework, we are with 16 conventional cores, second level cache is example, this conventional structure is made up of 16 nodes, each node includes the router R for communicating, the shared buffer memory l2bank of process core core and a local private cache L1 and larger area, because this structure is carried out data interaction based on the mechanism of shared buffer memory and communicate, and the area of shared buffer memory l2bank shared by whole network-on-chip is very large, therefore its power consumption impact brought is also very large, especially the ratio shared by quiescent dissipation.
From Fig. 1, we can see, each processor core core is connected with a shared buffer memory l2bank, due in the process of routine access, read and write data the data read and write possibly and store in the shared buffer memory l2bank at relative distance processor core core place far away, so when doing Static Design, the average number of hops mean distance of shared buffer memory l2bank (core access) and in the process of data interaction, the congestion problems (guaranteeing fairness) of whole network is all the factor needing to consider, the requirement that data store is increasing, the area of shared buffer memory l2bank is also increasing, thus its quiescent dissipation accounts for the ratio of whole network-on-chip also in the trend increasingly increased, therefore adopt this generic structure also to become a very important problem due to power problems that shared buffer memory l2bank area is excessive brought.
Summary of the invention
The object of the invention is to overcome the deficiencies in the prior art, a kind of chip multi-core processor static framework based on shared buffer memory is provided, do not increase whole network-on-chip access time delay, do not increase on the basis of Congestion Level SPCC of whole network-on-chip, save hardware spending and the area of afterbody high-speed cache, save the quiescent dissipation of high-speed cache.
The object of the invention is to be achieved through the following technical solutions: a kind of chip multi-core processor static framework based on shared buffer memory, it comprises n node, n be greater than 0 even number, described node comprises n/2 category-A node and n/2 category-B node; Described category-A node comprises process core core, local private cache L1 and router R; Described category-B node comprises process core core, local private cache L1, router R and shared buffer memory l2bank; Communicated by router between described node; Described category-A node becomes batch (-type) to distribute with category-B node; The capacity of described shared buffer memory l2bank is the twice of the shared buffer memory in generic structure.
Described node number is 16, comprises the category-A node described in 8 and the category-B node described in 8.
Described local private cache L1 comprises instruction buffer and data buffer storage.
The invention has the beneficial effects as follows: (1) can ensure that the mean distance of accessing shared buffer memory l2bank is the same with the mean distance of generic structure, thus additionally can not increase the access time delay of whole network-on-chip.
(2) the present invention is in design process of hardware, each shared buffer memory l2bank structure is identical with traditional, only storage space adds one times, and the decreased number of shared buffer memory l2bank is original half, the i.e. shared buffer memory capacity of whole afterbody high-speed cache and the identical of generic structure, is used for meeting the needs that data store; But decrease the l2bank of half, it also reduce the peripheral hardware expense of half, the number of such as amplifier and code translator we decrease half, therefore its hardware design area also can reduce to some extent, and the quiescent dissipation of high-speed cache also can reduce.
(3) node comprising shared buffer memory l2bank becomes batch (-type) to distribute with the node not comprising shared buffer memory l2bank, in the process of data interaction, can guarantee fairness, can not bring the problem of network congestion.
Accompanying drawing explanation
Fig. 1 is conventional network-on-chip polycaryon processor configuration diagram;
Fig. 2 is structural representation of the present invention.
Embodiment
Below in conjunction with accompanying drawing, technical scheme of the present invention is described in further detail, but protection scope of the present invention is not limited to the following stated.
As shown in Figure 2, a kind of chip multi-core processor static framework based on shared buffer memory, it comprises n node, n be greater than 0 even number, described node comprises n/2 category-A node and n/2 category-B node; Described category-A node comprises process core core, local private cache L1 and router R; Described category-B node comprises process core core, local private cache L1, router R and shared buffer memory l2bank; Communicated by router between described node; Described category-A node becomes batch (-type) to distribute with category-B node; The capacity of described shared buffer memory l2bank is the twice of the shared buffer memory in generic structure.
Described node number is 16, comprises the category-A node described in 8 and the category-B node described in 8.
Described local private cache L1 comprises instruction buffer and data buffer storage.
As can be seen from Figure 2, do not comprise the category-A node of shared buffer memory l2bank and comprise the distribution situation of shared buffer memory l2bankB category node, 16 Node distribution shown in the figure are four lines four row, no matter from laterally or longitudinally, all comprise a category-B node between every two category-A nodes, between every two category-B nodes, all comprise a category-A node; The batch (-type) distribution that Here it is mentions hereinbefore.

Claims (3)

1. based on a chip multi-core processor static framework for shared buffer memory, it is characterized in that: it comprises n node, n be greater than 0 even number, described node comprises n/2 category-A node and n/2 category-B node; Described category-A node comprises process core core, local private cache L1 and router R; Described category-B node comprises process core core, local private cache L1, router R and shared buffer memory l2bank; Communicated by router between described node; Described category-A node becomes batch (-type) to distribute with category-B node; The capacity of described shared buffer memory l2bank is the twice of the shared buffer memory in generic structure.
2. a kind of chip multi-core processor static framework based on shared buffer memory according to claim 1, is characterized in that: described node number is 16, comprises the category-A node described in 8 and the category-B node described in 8.
3. a kind of chip multi-core processor static framework based on shared buffer memory according to claim 1, is characterized in that: described local private cache L1 comprises instruction buffer and data buffer storage.
CN201510302580.4A 2015-06-04 2015-06-04 A kind of chip multi-core processor static framework based on shared buffer memory Expired - Fee Related CN104866457B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510302580.4A CN104866457B (en) 2015-06-04 2015-06-04 A kind of chip multi-core processor static framework based on shared buffer memory

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510302580.4A CN104866457B (en) 2015-06-04 2015-06-04 A kind of chip multi-core processor static framework based on shared buffer memory

Publications (2)

Publication Number Publication Date
CN104866457A true CN104866457A (en) 2015-08-26
CN104866457B CN104866457B (en) 2018-06-15

Family

ID=53912297

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510302580.4A Expired - Fee Related CN104866457B (en) 2015-06-04 2015-06-04 A kind of chip multi-core processor static framework based on shared buffer memory

Country Status (1)

Country Link
CN (1) CN104866457B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107870871A (en) * 2016-09-23 2018-04-03 华为技术有限公司 The method and apparatus for distributing caching

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101510191A (en) * 2009-03-26 2009-08-19 浙江大学 Multi-core system structure with buffer window and implementing method thereof
CN101706755A (en) * 2009-11-24 2010-05-12 中国科学技术大学苏州研究院 Caching collaboration system of on-chip multi-core processor and cooperative processing method thereof
CN102103568A (en) * 2011-01-30 2011-06-22 中国科学院计算技术研究所 Method for realizing cache coherence protocol of chip multiprocessor (CMP) system
CN102270180A (en) * 2011-08-09 2011-12-07 清华大学 Multicore processor cache and management method thereof
US20140280687A1 (en) * 2013-03-13 2014-09-18 Futurewei Technologies, Inc. Disaggregated Server Architecture for Data Centers

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101510191A (en) * 2009-03-26 2009-08-19 浙江大学 Multi-core system structure with buffer window and implementing method thereof
CN101706755A (en) * 2009-11-24 2010-05-12 中国科学技术大学苏州研究院 Caching collaboration system of on-chip multi-core processor and cooperative processing method thereof
CN102103568A (en) * 2011-01-30 2011-06-22 中国科学院计算技术研究所 Method for realizing cache coherence protocol of chip multiprocessor (CMP) system
CN102270180A (en) * 2011-08-09 2011-12-07 清华大学 Multicore processor cache and management method thereof
US20140280687A1 (en) * 2013-03-13 2014-09-18 Futurewei Technologies, Inc. Disaggregated Server Architecture for Data Centers

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHONGMIN LI等: "Fast Hierarchical Cache Directory:A Scalable Cache Organization for Large-scale CMP", 《2010 FIFTH IEEE INTERNATIONAL CONFERENCE ON NETWORKING,ARCHITECTURE,AND STORAGE》 *
杜建军: "共享高速缓存多核处理器的关键技术研究", 《中国博士学位论文全文数据库(信息科技辑)》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107870871A (en) * 2016-09-23 2018-04-03 华为技术有限公司 The method and apparatus for distributing caching
CN107870871B (en) * 2016-09-23 2021-08-20 华为技术有限公司 Method and device for allocating cache

Also Published As

Publication number Publication date
CN104866457B (en) 2018-06-15

Similar Documents

Publication Publication Date Title
US10409766B2 (en) Computer subsystem and computer system with composite nodes in an interconnection structure
US11748535B2 (en) System and method to generate a network-on-chip (NoC) description using incremental topology synthesis
CN102270180B (en) Multicore processor cache and management method thereof
US20150261698A1 (en) Memory system, memory module, memory module access method, and computer system
CN106936739B (en) Message forwarding method and device
US10282293B2 (en) Method, switch, and multiprocessor system using computations and local memory operations
EP3534579B1 (en) Packet transmission method, device and computer-readable medium
EP3327993B1 (en) Routing management
CN103092798B (en) The method of the access means under SOC(system on a chip) and bus
US20100274975A1 (en) Forming Multiprocessor Systems Using Dual Processors
US9812186B2 (en) Reducing latency in an expanded memory system
CN106951390B (en) NUMA system construction method capable of reducing cross-node memory access delay
EP3223162A1 (en) Configurable on-chip interconnection system and method and apparatus for implementing same, and storage medium
EP2652636B1 (en) Split traffic routing in a distributed shared memory multiprocessor
CN104035525A (en) Computational node
US9378167B2 (en) Enhanced data transfer in multi-CPU systems
CN104199521A (en) Blade node and extension method thereof
CN104102549A (en) Method, device and chip for realizing mutual exclusion operation of multiple threads
US20160077985A1 (en) Multi-mode agent
CN101178697B (en) PCIE apparatus satellite communication method and system
CN104778132A (en) Multi-core processor directory cache replacement method
CN107168810A (en) A kind of calculate node internal memory sharing system and reading and writing operation internal memory sharing method
CN104866457A (en) On-chip multi-core processor static architecture based on shared cache
WO2021120714A1 (en) Data exchange method, device, processor, and computer system
CN104834606A (en) Heterogeneous confusion hierarchical memory device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
EXSB Decision made by sipo to initiate substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20180615

CF01 Termination of patent right due to non-payment of annual fee