[go: up one dir, main page]

CN116009968A - Computer system, memory device and control method based on wafer stacking architecture - Google Patents

Computer system, memory device and control method based on wafer stacking architecture Download PDF

Info

Publication number
CN116009968A
CN116009968A CN202111219027.6A CN202111219027A CN116009968A CN 116009968 A CN116009968 A CN 116009968A CN 202111219027 A CN202111219027 A CN 202111219027A CN 116009968 A CN116009968 A CN 116009968A
Authority
CN
China
Prior art keywords
memory
line
array
common
address
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111219027.6A
Other languages
Chinese (zh)
Inventor
蔡昆华
严逸纬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Whalechain Technology Co ltd
Original Assignee
Whalechain Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Whalechain Technology Co ltd filed Critical Whalechain Technology Co ltd
Priority to CN202111219027.6A priority Critical patent/CN116009968A/en
Publication of CN116009968A publication Critical patent/CN116009968A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Static Random-Access Memory (AREA)

Abstract

本申请提出一种基于晶圆堆迭架构的计算机系统,包含内存装置和逻辑电路层结合成晶圆堆迭。所述内存装置中包含内存阵列和线路驱动器。所述内存阵列中包含共用线路以及多个内存单元,所述共用线路连接所述内存单元。所述线路驱动器连接所述共用线路,驱动所述内存单元。所述逻辑电路层包含多个用于传递信号的连接垫,以及延迟控制器,透过所述连接垫连接所述内存阵列调整所述共用线路上连接的内存单元个数,以动态改变所述内存阵列的延迟特性。本申请亦提出所述内存装置和内存控制方法。

Figure 202111219027

The present application proposes a computer system based on a wafer stacking architecture, which includes memory devices and logic circuit layers combined into a wafer stack. The memory device includes a memory array and a line driver. The memory array includes a shared line and a plurality of memory units, and the shared line connects the memory units. The line driver is connected to the shared line to drive the memory unit. The logic circuit layer includes a plurality of connection pads for transmitting signals, and a delay controller, through which the connection pads are connected to the memory array to adjust the number of memory units connected to the shared line to dynamically change the Latency characteristics of memory arrays. The application also proposes the memory device and memory control method.

Figure 202111219027

Description

基于晶圆堆迭架构的计算机系统,内存装置和控制方法Computer system based on wafer stacking architecture, memory device and control method

技术领域technical field

本申请是关于一种内存装置,尤其是关于可根据应用程序需求而改变延迟特性的内存架构,以及应用所述内存架构和晶圆堆迭技术实作而成的计算机系统。The present application relates to a memory device, in particular to a memory architecture capable of changing delay characteristics according to application requirements, and a computer system implemented by applying the memory architecture and wafer stacking technology.

背景技术Background technique

在这个年代,人工智能和区块链的应用成为一种新的商机。区块链可以广泛应用于智能合约,数字身份,共用经济等应用。In this era, the application of artificial intelligence and blockchain has become a new business opportunity. Blockchain can be widely used in smart contracts, digital identities, sharing economy and other applications.

然而一些区块链平台为了各种安全性考虑或是漏洞修补,经常会改变区块链的算法。除了增加运算难度之外,也经常刻意为了降低特定应用芯片(ASIC)的运算效率而做出特殊设计,例如增加内存吞吐量的要求,或是储存装置的容量要求。However, some blockchain platforms often change the algorithm of the blockchain for various security considerations or bug fixes. In addition to increasing the difficulty of computing, special designs are often deliberately made to reduce the computing efficiency of application-specific chips (ASICs), such as increasing the requirements of memory throughput or the capacity of storage devices.

因此,对于区块链服务器的开发者而言,也随着必须要改变硬件架构,来适应区块链算法的变化。然而,这些众筹平台提出的算法,有可能不断的更新。因此,如何使同一套区块链服务器的硬件能弹性地改变参数以适应新的算法,是有待开发的。Therefore, for developers of blockchain servers, it is also necessary to change the hardware architecture to adapt to changes in blockchain algorithms. However, the algorithms proposed by these crowdfunding platforms may be constantly updated. Therefore, how to make the hardware of the same block chain server elastically change parameters to adapt to new algorithms remains to be developed.

发明内容Contents of the invention

本申请提出一种计算机系统,可弹性地适应区块链算法的变化要求。在一计算机系统的实施例中,采用了晶圆堆迭(wafer on wafer)的技术,使内存装置所在的晶圆和内核逻辑电路的晶圆堆迭成立体结构。这个做法可使两片晶圆之间不需要多余的面积,直接以成千上万个连接垫做为信号传递的路径。由于传送线路的数量不再受到平面设计的限制,因此可以使用大量的专用接线来解决资料传递的效能问题。This application proposes a computer system that can flexibly adapt to changing requirements of blockchain algorithms. In an embodiment of a computer system, a wafer on wafer technology is adopted, so that the wafer on which the memory device is located and the wafer on which the core logic circuit is located are stacked into a three-dimensional structure. This method can make no extra area between the two wafers, and directly use thousands of connection pads as signal transmission paths. Since the number of transmission lines is no longer limited by the plane design, a large number of dedicated wiring can be used to solve the performance problem of data transmission.

本申请的内存装置,配置于内存专用的一层晶圆,其中可包含多个内存阵列(BANK)。每一内存阵列主要由一共用线路以及多个内存单元组成。所述共用线路,在本实施例中可以是数据线或地址线的代称,每条共用线路各对应地连接所述内存单元的其中一行或一列。内存单元指的是储存位信息的基本单位,通常受到地址信号的控制而开启,并受到数据信号的控制而读出或写入数据。The memory device of the present application is configured on a layer of wafer dedicated to memory, which may include multiple memory arrays (BANKs). Each memory array is mainly composed of a common circuit and a plurality of memory units. The common lines in this embodiment may be referred to as data lines or address lines, and each common line is correspondingly connected to one row or one column of the memory units. A memory cell refers to a basic unit for storing bit information, which is usually turned on under the control of an address signal, and read or written into data under the control of a data signal.

所述内存装置中还包含一线路驱动器,连接所述共用线路,用于驱动所述内存单元。所述线路驱动器可以是数据驱动器或地址译码器的代称。The memory device also includes a line driver connected to the shared line for driving the memory unit. The line driver may be a synonym for a data driver or an address decoder.

如上所述,所述计算机系统中包含一逻辑电路层,与所述内存晶体层结合成一晶圆堆迭(Wafer on Wafer)。其中包含多个连接垫,用于传递信号。As mentioned above, the computer system includes a logic circuit layer, which is combined with the memory crystal layer to form a wafer stack (Wafer on Wafer). It contains several connection pads for passing signals.

本申请在计算机系统的逻辑电路层中,配置了一个延迟控制器,透过所述连接垫连接所述内存阵列。其设计目的是弹性调整所述共用线路上连接的内存单元个数,以动态改变所述内存阵列的延迟特性。In the present application, a delay controller is configured in the logic circuit layer of the computer system, and the memory array is connected through the connection pad. Its design purpose is to flexibly adjust the number of memory units connected to the shared line, so as to dynamically change the delay characteristics of the memory array.

在进一步的实施例中,每一内存阵列中配置了多个多工器。每个多工器之间相隔特定数量的行数或列数。这些多工器将一内存阵列定义为多个内存区域,每一内存区域各包含特定行数或特定列数的内存单元。换言之,每两个内存区域相邻之处,就会配置有一个多工器,和所述线路驱动器以一专用线路相连。In a further embodiment, multiple multiplexers are configured in each memory array. Each multiplexer is separated by a specific number of rows or columns. These multiplexers define a memory array as a plurality of memory regions, and each memory region contains a specific number of rows or columns of memory cells. In other words, a multiplexer is configured where two memory areas are adjacent, and is connected to the line driver with a dedicated line.

当所述延迟控制器透过一连接垫传递一控制信号启动一多工器时,会使所述共用线路断开为一第一线段和一第二线段,并使所述第二线段连接至所述专用线路。When the delay controller transmits a control signal through a connection pad to activate a multiplexer, the common line is disconnected into a first line segment and a second line segment, and the second line segment is connected to the dedicated line.

由于所述共用线路原本串接了多个内存单元,在断开为两个线段后,在逻辑上就形成了两个子阵列。换句话说,所述第一线段对应的内存区域形成一第一子阵列,而所述第二线段对应的内存区域形成一第二子阵列。为了便于实施于管理,本实施例的断开方式可以是二等分。因此一内存阵列可等分为两个大小相等的子阵列,而两个子阵列可进一步再透过更多多工器分割为四个,依此类推。Since the shared line originally connects a plurality of memory units in series, after being disconnected into two line segments, two sub-arrays are logically formed. In other words, the memory area corresponding to the first line segment forms a first sub-array, and the memory area corresponding to the second line segment forms a second sub-array. In order to facilitate implementation and management, the disconnection mode in this embodiment may be bisected. Therefore, a memory array can be equally divided into two sub-arrays of equal size, and the two sub-arrays can be further divided into four by more multiplexers, and so on.

在一具体的实施方式中,改变内存阵列维度的方式,即子阵列的形成方式,可以是将共用数据线断开为两个较短的资料线。所述线路驱动器包含一数据驱动器。所述共用线路在此代表一或多条共用数据线,每一共用数据线连接所述数据驱动器和所述内存单元中对应的一列内存单元,用于传送所述内存单元的数据信号。在所述多工器被启动后,由于共用数据线断开,所述第二子阵列不再共用所述第一子阵列的共用数据线。而是由多工器另外提供专用线路,给所述第二子阵列中的内存单元传送数据信号。这个做法可让共用线路上的内存单元数量减少,进而使电容负载降低,加快了数据驱动的反应速度。In a specific implementation manner, the way of changing the dimension of the memory array, that is, the way of forming the sub-arrays, may be to break the common data line into two shorter data lines. The line driver includes a data driver. The common line here represents one or more common data lines, and each common data line is connected to the data driver and a corresponding row of memory units in the memory units, and is used for transmitting data signals of the memory units. After the multiplexer is activated, the second sub-array no longer shares the common data line of the first sub-array because the common data line is disconnected. Instead, the multiplexer additionally provides dedicated lines to transmit data signals to the memory units in the second sub-array. This approach can reduce the number of memory cells on the shared line, thereby reducing the capacitive load and speeding up the response speed of the data driver.

至于第二子阵列,由于数据线改由专用线路连接至数据驱动器,独立接收不同的数据信号源,也同样的享有低负载高速度的效果。更进一步地说,可将第二子阵列的地址线改为共用第一子阵列的地址线。如此,等于是将原本的内存阵列维度改变,数据线数(阵列宽度)倍增,而地址线数(阵列高度)减半。所述内存装置中原本就包含多条共用地址线,每一共用地址线连接所述地址译码器和所述内存单元中对应的一行内存单元,用于传送所述内存单元的地址信号。所述线路驱动器包含一地址译码器,透过所述共用地址线连接所述内存阵列中的每行内存单元。实作上,在所述多工器被启动后,所述地址译码器根据所述控制信号,使所述第二子阵列中的内存单元共用所述第一子阵列的共用地址线,或使用相同的地址信号同步驱动所述第一子阵列和所述第二子阵列。换句话说,使所述第二子阵列与所述第一子阵列中对应行数的共用地址线接收相同的地址信号。As for the second sub-array, because the data line is connected to the data driver by a dedicated line, it can independently receive different data signal sources, and also enjoy the effect of low load and high speed. Furthermore, the address lines of the second sub-array can be changed to share the address lines of the first sub-array. In this way, the dimension of the original memory array is changed, the number of data lines (array width) is doubled, and the number of address lines (array height) is halved. The memory device originally includes a plurality of common address lines, and each common address line is connected to the address decoder and a corresponding row of the memory units for transmitting address signals of the memory units. The line driver includes an address decoder connected to each row of memory units in the memory array through the common address line. In practice, after the multiplexer is activated, the address decoder enables the memory cells in the second sub-array to share the common address line of the first sub-array according to the control signal, or The first sub-array and the second sub-array are synchronously driven using the same address signal. In other words, make the common address lines of the second sub-array and the corresponding number of rows in the first sub-array receive the same address signal.

在另一具体的实施方式中,改变内存阵列维度的方式,即子阵列的形成方式,也可以是将共用地址线断开为两个较短的地址线。在这种情况中,上述共用线路代表的是一或多条共用地址线,用于传送所述内存单元的地址信号。在所述多工器被启动后,所述第二子阵列中的内存单元使用所述专用线路传送地址信号。同时,所述地址译码器根据所述控制信号,使所述第二子阵列中的内存单元共用所述第一子阵列的共用数据线,或使用相同的数据信号驱动所述第一子阵列和所述第二子阵列。由于第二子阵列中的内存单元使用了和第一子阵列不同的地址线,内存阵列维度等于是将数据位数(阵列宽度)砍半,并将地址线数(阵列高度)倍增。In another specific implementation manner, the way of changing the dimension of the memory array, that is, the way of forming the sub-array, may also be to break the common address line into two shorter address lines. In this case, the above-mentioned common lines represent one or more common address lines for transmitting address signals of the memory units. After the multiplexer is activated, the memory units in the second sub-array transmit address signals using the dedicated lines. At the same time, the address decoder makes the memory cells in the second sub-array share the common data line of the first sub-array according to the control signal, or uses the same data signal to drive the first sub-array and the second subarray. Since the memory cells in the second sub-array use address lines different from those in the first sub-array, the memory array dimension is equivalent to halving the number of data bits (array width) and doubling the number of address lines (array height).

在进一步的实施例中,其中所述逻辑电路层进一步包含一内存控制器,透过所述连接垫耦接所述内存阵列。一内核连接所述内存控制器和所述延迟控制器,用于执行一应用程序。所述内核可根据所述应用程序要求的一应用程序条件,透过所述延迟控制器设定所述内存阵列中的所述多工器,使所述内存阵列改变维度,即分割为二的多乘方个子阵列,再重组为符合所述应用程序条件的新阵列维度。在执行所述程序时,内核可透过所述内存控制器使用重组后的内存阵列。In a further embodiment, the logic circuit layer further includes a memory controller coupled to the memory array through the connection pad. A kernel connects the memory controller and the delay controller for executing an application program. The kernel can set the multiplexer in the memory array through the delay controller according to an application condition required by the application program, so that the memory array can be divided into two Multiply subarrays, reorganized into new array dimensions that meet the application criteria. When executing the program, the kernel can use the reorganized memory array through the memory controller.

在进一步的实施例中,所述应用程序条件包含所述应用程序需要的反应时间。在断开数据线的实施例中,所述延迟控制器启动的多工器数量越多,形成的新内存阵列反应时间越短,In a further embodiment, the application condition includes a response time required by the application. In the embodiment of disconnecting the data line, the more the number of multiplexers activated by the delay controller, the shorter the response time of the new memory array formed,

本申请另外提出一种内存控制方法,应用于前述的计算机系统和内存装置中。由一内核执行一应用程序时,所述内核根据所述应用程序要求的一应用程序条件,透过所述延迟控制器设定所述内存阵列中的所述多工器,使所述内存阵列分割为二或多个符合所述应用程序条件的子阵列,并在执行所述应用程序时透过所述内存控制器使用所述内存子阵列。The present application further proposes a memory control method, which is applied to the aforementioned computer system and memory device. When an application program is executed by a core, the core sets the multiplexer in the memory array through the delay controller according to an application condition required by the application program, so that the memory array Divide into two or more sub-arrays meeting the conditions of the application program, and use the memory sub-arrays through the memory controller when executing the application program.

综上所述,本申请基于晶圆堆迭技术,提出了一种可以弹性调整阵列维度的内存架构,使得区块链服务器产品有能力可以适应未来算法的需求。To sum up, based on the wafer stacking technology, this application proposes a memory architecture that can flexibly adjust the array dimension, so that blockchain server products have the ability to adapt to the needs of future algorithms.

附图说明Description of drawings

此处所说明的附图用来提供对本申请的进一步理解,构成本申请的一部分,本申请的示意性实施例及其说明用于解释本申请,并不构成对本申请的不当限定。在附图中:The drawings described here are used to provide a further understanding of the application and constitute a part of the application. The schematic embodiments and descriptions of the application are used to explain the application and do not constitute an improper limitation to the application. In the attached picture:

图1是本申请的计算机系统的实施例。FIG. 1 is an embodiment of the computer system of the present application.

图2是本申请的内存装置的实施例。FIG. 2 is an embodiment of a memory device of the present application.

图3至5是本申请的内存阵列和多工器的实施例。3 to 5 are embodiments of memory arrays and multiplexers of the present application.

图6是本申请的内存装置112的进一步实施例。FIG. 6 is a further embodiment of the memory device 112 of the present application.

图7至8是本申请的各种内存阵列的实施例。7 to 8 are embodiments of various memory arrays of the present application.

图9是本申请的计算机系统700中的内存层600的实施例。FIG. 9 is an embodiment of a memory layer 600 in a computer system 700 of the present application.

图10是本申请的计算机系统的进一步实施例。Figure 10 is a further embodiment of the computer system of the present application.

图11是本申请的内存阵列和多工器的进一步实施例。Figure 11 is a further embodiment of the memory array and multiplexer of the present application.

图12是本申请的内存阵列和多工器的另一实施例。FIG. 12 is another embodiment of the memory array and multiplexer of the present application.

图13是本申请内存控制方法的流程图。FIG. 13 is a flow chart of the memory control method of the present application.

具体实施方式Detailed ways

下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The following will clearly and completely describe the technical solutions in the embodiments of the present application with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of this application.

图1是本申请的立体晶圆产品100的实施例。立体晶圆产品100由至少一内存晶体层110,一逻辑电路层120,及一基底130层层堆迭。基底130除了提供基本的支撑,也提供额外的布线空间。每一层之间配置有多个连接垫102或104以提供信号信道。本实施例的立体晶圆产品100是计算机系统700的半成品,经过切割后可产生多个独立运作的计算机系统700。如图1所示,每个计算机系统700可各包含若干个内存装置112和若干个逻辑电路122,具备相同的立体晶圆结构。换句话说,每个计算机系统700中包含的内存装置112和逻辑电路122,是事先各别布局于内存晶体层110和逻辑电路层120中,再以晶圆堆迭的形式制成的立体结构。在立体结构中,芯片组之间的电路导线不需要占用多余的面积,可直接以成千上万个连接垫102和104做为信号传递的路径,使数据传递的效能问题有效被解决,借此实现本申请的计算机系统700。FIG. 1 is an embodiment of a 3D wafer product 100 of the present application. The three-dimensional wafer product 100 is composed of at least one memory crystal layer 110 , one logic circuit layer 120 , and one substrate 130 stacked layer by layer. In addition to providing basic support, the base 130 also provides additional wiring space. A plurality of connection pads 102 or 104 are disposed between each layer to provide signal channels. The 3D wafer product 100 of this embodiment is a semi-finished product of the computer system 700 , which can produce multiple independently operating computer systems 700 after cutting. As shown in FIG. 1 , each computer system 700 may include several memory devices 112 and several logic circuits 122 and have the same three-dimensional wafer structure. In other words, the memory device 112 and the logic circuit 122 included in each computer system 700 are respectively laid out in the memory crystal layer 110 and the logic circuit layer 120 in advance, and then formed into a three-dimensional structure in the form of wafer stacking . In the three-dimensional structure, the circuit wires between the chipsets do not need to occupy an extra area, and thousands of connection pads 102 and 104 can be directly used as signal transmission paths, so that the performance problem of data transmission can be effectively solved. This implements the computer system 700 of the present application.

图2是本申请的内存装置的实施例。本申请的内存装置112,布局在内存专用的内存晶体层110上。其制造可以是模块化的形式,每一内存装置112中可包含多个内存阵列200,或称为内存矩阵(BANK)。每一内存矩阵的运作可受到一阵列选择信号#SL的控制。每一内存阵列200主要由多个内存单元202组成。内存单元202排列成多行和多列,每一行共用一条地址线,接收编号为R0至Rn的地址信号。而每一列共用一条数据线,传送编号为B0至Bn的数据信号。换言之,每条共用线路各对应地连接所述内存单元202的其中一行或一列。内存单元202指的是储存位信息的基本单位,通常受到地址信号的控制而开启,并受到数据信号的控制而读出或写入数据。所述地址线连接一地址译码器210,用于传递所述地址译码器产生的地址信号214,使被选定的一或多行内存单元202被开启。所述数据线连接一数据驱动器220,用于传送内存单元202写入或读出的数据。图2所揭示的架构仅为示例,在实际制作中,内存阵列200,地址译码器210,和数据驱动器220的数量不限定为一,其间的链接关系也不限定为一对一,或多对多。综上所述,所述内存装置中的地址译码器210和数据驱动器220是一种线路驱动器。而连接的数据线和地址线是一种共用线路,以网状交织的方式驱动多个内存单元202。FIG. 2 is an embodiment of a memory device of the present application. The memory device 112 of the present application is arranged on the memory crystal layer 110 dedicated to memory. It can be manufactured in a modular form, and each memory device 112 can include multiple memory arrays 200 , or called memory matrix (BANK). The operation of each memory matrix can be controlled by an array selection signal #SL. Each memory array 200 is mainly composed of a plurality of memory units 202 . The memory units 202 are arranged in multiple rows and columns, each row shares one address line, and receives address signals numbered R0 to Rn. Each column shares one data line, and transmits data signals numbered B0 to Bn. In other words, each common line is correspondingly connected to one row or one column of the memory units 202 . The memory unit 202 refers to a basic unit for storing bit information, and is usually turned on under the control of an address signal, and read or written into data under the control of a data signal. The address line is connected to an address decoder 210 for transmitting an address signal 214 generated by the address decoder to enable one or more rows of memory cells 202 selected to be turned on. The data line is connected to a data driver 220 for transmitting data written or read by the memory unit 202 . The architecture disclosed in FIG. 2 is only an example. In actual production, the number of memory array 200, address decoder 210, and data driver 220 is not limited to one, and the link relationship between them is not limited to one-to-one, or multiple To many. To sum up, the address decoder 210 and the data driver 220 in the memory device are a kind of line driver. The connected data lines and address lines are shared lines, which drive multiple memory units 202 in a mesh-like interleaving manner.

图3至5是本申请的内存阵列200和多工器302的实施例。为了达成动态调整延迟特性的功效,本实施例在每一内存阵列200中配置了多个多工器302。每个多工器302之间相隔特定数量的行数或列数。这些多工器302可将一内存阵列200定义为多个内存区域310,每一内存区域310各包含特定行数或特定列数的内存单元202。以图3为例,每两个内存区域310相邻之处配置有一个多工器302。3 to 5 are embodiments of the memory array 200 and the multiplexer 302 of the present application. In order to achieve the effect of dynamically adjusting delay characteristics, multiple multiplexers 302 are configured in each memory array 200 in this embodiment. Each multiplexer 302 is separated by a specific number of rows or columns. These multiplexers 302 can define a memory array 200 as a plurality of memory regions 310 , and each memory region 310 includes a specific number of rows or columns of memory cells 202 . Taking FIG. 3 as an example, a multiplexer 302 is disposed adjacent to each two memory areas 310 .

图4显示多工器302启动时的运作情形。所述多工器302透过一专用线路224连接所述数据驱动器220。当一控制信号#S从图1中所示的逻辑电路层120透过所述连接垫102其中之一传递至所述多工器时,会使共用的数据线222在多工器302所在处断开,使上下两个内存区域310不再共用相同的数据线222。数据线222被分割为上半部内存区域310中的第一线段,和下半部内存区域310中的第二线段。在本实施例中,上半部的内存区域310可继续接收原本的数据信号B0至B7,但是由于第一线段上共用的内存单元数减少了,电容负载显着地降低,因此内存区域310的延迟时间可以有效缩短,也就是加快了反应的速度。多工器302将所述下半部的内存区域310的第二线段改接至专用线路224,使所述下方内存区域310继续受到数据驱动器220的控制。举例来说,所述多工器302透过所述专用线路224继续从所述数据驱动器220接收编号为B0至B7的数据信号。由于第二线段上共用的内存单元202比原来少,所以也达成了减少延迟的效果。FIG. 4 shows the operation of the multiplexer 302 when it starts up. The multiplexer 302 is connected to the data driver 220 through a dedicated line 224 . When a control signal #S is transmitted to the multiplexer from the logic circuit layer 120 shown in FIG. disconnected so that the upper and lower memory areas 310 no longer share the same data line 222 . The data line 222 is divided into a first line segment in the upper half memory area 310 and a second line segment in the lower half memory area 310 . In this embodiment, the memory area 310 in the upper half can continue to receive the original data signals B0 to B7, but because the number of shared memory cells on the first line segment is reduced, the capacitive load is significantly reduced, so the memory area 310 The delay time can be effectively shortened, that is, the speed of reaction is accelerated. The multiplexer 302 reconnects the second line segment of the lower memory area 310 to the dedicated line 224 , so that the lower memory area 310 continues to be controlled by the data driver 220 . For example, the multiplexer 302 continues to receive data signals numbered B0 to B7 from the data driver 220 through the dedicated line 224 . Since the shared memory units 202 on the second line segment are less than before, the effect of reducing delay is also achieved.

图5显示多工器302启动时的另一具体的实施方式。本实施例除了改变内存阵列200的延迟特性,也可以改变内存阵列200维度。子阵列的形成方式,同样是将共用数据线222断开为较短的上下两部份。换句话说,由于数据线222原本串接了多个内存单元202,在断开为两个线段后,在逻辑上就形成了两个子阵列。所述第一线段对应的内存区域310形成一第一子阵列,而所述第二线段对应的内存区域310形成一第二子阵列。在所述多工器302被启动后,由于共用数据线222断开,所述第二子阵列不再共用所述第一子阵列的共用数据线222。而是由多工器302另外提供专用线路224,给所述第二子阵列中的内存单元202传送数据信号。在本实施例中,可透过数据驱动器220的改良,使透过专用线路224传送的数据信号,不再是编号B0至B7,而是新增的B8至B15。更进一步地,本实施例可透过地址译码器的改良,使第二子阵列共用第一子阵列的地址线,或是接收到与第一子阵列相同的地址信号R0至R3。FIG. 5 shows another specific implementation manner when the multiplexer 302 starts. In this embodiment, in addition to changing the delay characteristic of the memory array 200 , the dimensions of the memory array 200 can also be changed. The formation method of the sub-array is also to divide the common data line 222 into two shorter parts, the upper part and the lower part. In other words, since the data line 222 originally connects a plurality of memory units 202 in series, two sub-arrays are logically formed after being disconnected into two line segments. The memory area 310 corresponding to the first line segment forms a first sub-array, and the memory area 310 corresponding to the second line segment forms a second sub-array. After the multiplexer 302 is activated, the second sub-array no longer shares the common data line 222 of the first sub-array because the common data line 222 is disconnected. Instead, the multiplexer 302 additionally provides a dedicated line 224 to transmit data signals to the memory units 202 in the second sub-array. In this embodiment, through the improvement of the data driver 220, the data signals transmitted through the dedicated line 224 are no longer numbered B0 to B7, but newly added B8 to B15. Furthermore, in this embodiment, through the improvement of the address decoder, the second sub-array can share the address lines of the first sub-array, or receive the same address signals R0 to R3 as the first sub-array.

也就是说,所述地址译码器230可根据所述控制信号#S,使所述第二子阵列中的内存单元共用所述第一子阵列的共用地址线,或使用相同的地址信号同步驱动所述第一子阵列和所述第二子阵列。如此,等于是将原本的内存阵列200维度改变。数据线数(阵列宽度)从原本的8条倍增为16条,而地址线数(阵列高度)从原本的8条减半为4条。本实施例虽然以8x8为例说明内存阵列200的分割重组方式,但可以理解的是,在实际制造中,每个内存阵列200的维度可以是容量百兆位的大型阵列。That is to say, the address decoder 230 can make the memory cells in the second sub-array share the common address line of the first sub-array according to the control signal #S, or use the same address signal to synchronize The first sub-array and the second sub-array are driven. In this way, it is equivalent to changing the original 200 dimensions of the memory array. The number of data lines (array width) is doubled from the original 8 to 16, and the number of address lines (array height) is halved from the original 8 to 4. Although this embodiment uses 8x8 as an example to illustrate the division and reorganization method of the memory array 200, it can be understood that in actual manufacturing, the dimension of each memory array 200 may be a large array with a capacity of 100 megabits.

图6是本申请的内存装置112的进一步实施例。一内存阵列200中可配置有n个多工器402#1至402#n,将内存阵列200区分为n个内存区域410#1至410#n。在多工器不启动时,所述内存阵列200维持习知的运作方式。数据驱动器220除了透过传统的共用数据线222传送数据信号之外,也提供多条专用线路224连接至所述多工器402#1至402#n。所述内存装置112中进一步包含一地址译码器230,透过地址线232传送地址信号#A至每一内存区域410#1至410#n。数据线222和地址线232虽以单线条表示,但可以理解的是实作上可包含多条线路,各别连接所述内存阵列中的每行或列。与传统设计相似的是,内存阵列200中的每一内存单元共同连接至一基准电压,或是地线#Gnd。FIG. 6 is a further embodiment of the memory device 112 of the present application. A memory array 200 can be configured with n multiplexers 402 #1 to 402 #n to divide the memory array 200 into n memory areas 410 #1 to 410 #n. When the multiplexer is not activated, the memory array 200 maintains the conventional operation mode. In addition to transmitting data signals through the traditional shared data lines 222, the data driver 220 also provides a plurality of dedicated lines 224 connected to the multiplexers 402 #1 to 402 #n. The memory device 112 further includes an address decoder 230 for transmitting address signal #A to each memory area 410 #1 to 410 #n through an address line 232 . Although the data line 222 and the address line 232 are shown as a single line, it can be understood that they may include multiple lines in practice, respectively connected to each row or column in the memory array. Similar to the conventional design, each memory cell in the memory array 200 is commonly connected to a reference voltage, or ground #Gnd.

实作上,每一多工器可以是收到一控制信号#S而决定是否启动。举例来说,控制信号#S可以是一个二乘方的数值,即2,4,8,或16等,用以指示所述些多工器402#1至402#n将所述内存阵列200分为对应数量个子阵列。在控制信号#S为2时,表示需要一个多工器将所述内存阵列200等分为两个子阵列。这时,位于所述内存阵列200中编号为n/2的多工器可响应所述控制信号而启动,以达成此目的。同理,当控制信号#S的数值为4时,表示需要三个多工器将所述内存阵列200等分为四个子阵列。这时,位于所述内存阵列200中,编号为n/4,2n/4,3n/4的多工器可响应所述控制信号#S而启动,以达成分割四块的效果。在这种设计方式中,n的数值可预先设定为一个二的乘方数,以方便实现上述分割法。In practice, each multiplexer may receive a control signal #S to determine whether to start. For example, the control signal #S can be a square value, that is, 2, 4, 8, or 16, etc., to instruct the multiplexers 402 #1 to 402 #n to use the memory array 200 Divide into a corresponding number of sub-arrays. When the control signal #S is 2, it means that a multiplexer is needed to divide the memory array 200 into two sub-arrays. At this time, the multiplexer numbered n/2 in the memory array 200 can be activated in response to the control signal to achieve this purpose. Similarly, when the value of the control signal #S is 4, it means that three multiplexers are needed to equally divide the memory array 200 into four sub-arrays. At this time, the multiplexers numbered n/4, 2n/4, 3n/4 located in the memory array 200 can be activated in response to the control signal #S to achieve the effect of dividing into four blocks. In this design method, the value of n can be preset as a power of two, so as to facilitate the realization of the above division method.

在另一种实作方式中,也可以用所述控制信号#S来决定每隔几个内存区域需要切开。举例来说,当控制信号#S的值为1时,表示每一个内存区域都需要独立出来,也就是全部的多工器402#1至402#n都被启动,使所述内存阵列200成为n个子阵列,每个子阵列包含一个内存区域。当控制信号#S的值为2时,表示需要将所述内存阵列200以每两个内存区域为一组而分割。因此,编号2,4,6,8等可被2整除的多工器会响应所述控制信号而启动,使所述内存阵列成为n/2个子阵列,每个子阵列包含2个内存区域。In another implementation manner, the control signal #S may also be used to determine that every few memory areas need to be cut. For example, when the value of the control signal #S is 1, it means that each memory area needs to be separated, that is, all the multiplexers 402#1 to 402#n are activated, so that the memory array 200 becomes a n subarrays, each subarray contains a memory region. When the value of the control signal #S is 2, it indicates that the memory array 200 needs to be divided into groups of two memory areas. Therefore, multiplexers numbered 2, 4, 6, 8, etc. that are divisible by 2 will be activated in response to the control signal, so that the memory array becomes n/2 sub-arrays, and each sub-array includes 2 memory areas.

在更进一步的实作方式中,内存阵列200的分割方式可以更加灵活。例如每一多工器各别接收不同的控制信号而决定是否启动。因此实际上可产生的分割可能性不限定于上述实施例。In a further implementation manner, the partitioning manner of the memory array 200 can be more flexible. For example, each multiplexer receives different control signals to determine whether to start. The actually possible division possibilities are therefore not limited to the above-described exemplary embodiments.

在图6的实施例中,数据驱动器220和地址译码器230也可进一步的改良,根据控制信号#S分割的情况,而改变提供给每一内存区域的数据信号,或改变提供给每一内存区域的地址信号。这个做法如同图5的实施例所述,能使内存阵列200在逻辑上动态改变长宽维度。In the embodiment of FIG. 6, the data driver 220 and the address decoder 230 can also be further improved, and the data signal provided to each memory area can be changed according to the division of the control signal #S, or the data signal provided to each memory area can be changed. The address signal of the memory area. This approach, as described in the embodiment of FIG. 5 , enables the memory array 200 to dynamically change the length and width dimensions logically.

图7至8是本申请的各种内存阵列的实施例。图7中显示图6的内存阵列200被多工器分割重组后产生的内存阵列500a。原本每一内存区域410#1至410#n各具有W列数的内存单元(宽度)和H行数的地址线(高度)。经过维度重组后,形成了包含多个子阵列502a的内存阵列500a。其中所有的子阵列502a共用所述H行地址线,而位宽度则扩展为nW列。这表示多工器需提供为nW列内存单元提供专用线路连接至数据驱动器,即nW条。在晶圆堆迭技术的支持下,可以轻易克服实作的技术难度。在图7的实施例中,原本的内存阵列维度可能是nH*W,在这里被重组为H*nW。因此每一列内存单元的数据线被驱动时,需要克服的电容负载变小n倍,使内存单元的反应速度变快。7 to 8 are embodiments of various memory arrays of the present application. FIG. 7 shows a memory array 500a generated after the memory array 200 in FIG. 6 is divided and reorganized by a multiplexer. Originally, each of the memory areas 410 # 1 to 410 #n has W columns of memory cells (width) and H rows of address lines (height). After dimension reorganization, a memory array 500a including a plurality of sub-arrays 502a is formed. All the sub-arrays 502a share the H row address lines, and the bit width is extended to nW columns. This means that the multiplexer needs to provide dedicated lines for nW columns of memory cells to connect to the data drivers, that is, nW lines. With the support of wafer stacking technology, the technical difficulty of implementation can be easily overcome. In the embodiment of FIG. 7, the original memory array dimension may be nH*W, which is reorganized as H*nW here. Therefore, when the data lines of each row of memory cells are driven, the capacitive load to be overcome becomes n times smaller, so that the response speed of the memory cells becomes faster.

图8中显示图6的内存阵列200被多工器分割重组后产生的内存阵列500b的情况。原本每一内存区域410#1至410#n各具有W列数的内存单元(宽度)和H行数的地址线(高度)。在此以每隔两个内存区块启动一个多工器的方式重组维度,形成了包含多个子阵列502b的内存阵列500b。其中每个子阵列502b包含两个内存区域,高2H行,宽W列。所述内存阵列500b中的子阵列502a共用2H行地址线。更确切地说,通过地址译码器230的改良,可弹性地根据内存阵列500b的分割情况,使所有的内存子阵列502b共用相同的地址线,或使地址译码器230传送相同的地址信号至这些内存子阵列。所述内存阵列500b的位宽度扩展为nW/2列。这表示多工器需为nW/2列内存单元提供对应数量的专用线路连接至数据驱动器。在图8的实施例和图7相比,由于子阵列502b的高度(地址行数)较多,延迟时间不如图7的架构,但需要的专用线路数较少。这说明了本实施例的架构可根据不同的需求权衡而弹性调整。FIG. 8 shows the situation of the memory array 500b generated after the memory array 200 in FIG. 6 is divided and reorganized by a multiplexer. Originally, each of the memory areas 410 # 1 to 410 #n has W columns of memory cells (width) and H rows of address lines (height). Here, the dimensions are reorganized by starting a multiplexer every two memory blocks, forming a memory array 500b including a plurality of sub-arrays 502b. Each sub-array 502b includes two memory areas, 2H rows high and W columns wide. The sub-arrays 502a in the memory array 500b share 2H row address lines. More precisely, through the improvement of the address decoder 230, all the memory sub-arrays 502b can share the same address line, or the address decoder 230 can transmit the same address signal flexibly according to the division of the memory array 500b. to these memory subarrays. The bit width of the memory array 500b is extended to nW/2 columns. This means that the multiplexer needs to provide a corresponding number of dedicated lines for nW/2 columns of memory cells to connect to the data driver. In the embodiment of FIG. 8 compared with that of FIG. 7 , since the height of the sub-array 502b (the number of address rows) is larger, the delay time is not as good as that of the architecture of FIG. 7 , but the required number of dedicated lines is less. This shows that the architecture of this embodiment can be flexibly adjusted according to different demands.

图9显示计算机系统700中的内存层600的进一步实施例。基于前述实施例介绍的概念,内存层600可以是从图1的内存晶体层110中切割出来的计算机系统的其中一个区域,包含多个内存装置510a至510d。每个内存装置510a至510d可各别应用多种不同的控制信号,配置不同的延迟特性。举例来说,一计算机系统700可在固件中预先设定每一内存装置510a至510d的配置,在开机后透过控制信号#S1至#S4各别配置所述内存装置510a至510d,然后再开机加载操作系统。在更进一步的情况下,本申请的计算机系统700也可设计为允许在运作中动态无缝的改变内存延迟特性。例如在加载一应用程序时,判断所述应用程序对内存延迟的需求,而动态发出控制信号改变所述些内存装置的维度,使延迟特性改变。FIG. 9 shows a further embodiment of a memory layer 600 in a computer system 700 . Based on the concepts introduced in the foregoing embodiments, the memory layer 600 may be one of the areas of the computer system cut out from the memory crystal layer 110 in FIG. 1 , including a plurality of memory devices 510a to 510d. Each of the memory devices 510a to 510d can apply a variety of different control signals to configure different delay characteristics. For example, a computer system 700 can pre-set the configuration of each memory device 510a to 510d in firmware, configure the memory devices 510a to 510d respectively through control signals #S1 to #S4 after booting, and then Boot up and load the operating system. In a further situation, the computer system 700 of the present application can also be designed to allow dynamic and seamless change of memory latency characteristics during operation. For example, when an application program is loaded, it is judged that the application program requires memory delay, and a control signal is dynamically issued to change the dimensions of the memory devices to change the delay characteristics.

图10是本申请的计算机系统700的进一步实施例。图1所示的立体晶圆产品100,完成晶圆堆迭程序之后,进一步经过晶圆切割的程序,形成多个计算机系统700。内存层600中显示的是根据图9实施例所设定的内存装置510a至510c。一系统层620与所述内存层600堆迭。系统层620是从图1的逻辑电路层120切割而成,包含各种计算器架构必备的逻辑电路,例如内核616和内存控制器614a至614c。每一内存控制器614a至614c各透过一接口模块612a至612c连接所述内存层600中的内存装置510a至510c。接口模块是专为确保数据传输而设计的界面,俗称物理层界面(PHY)。与图1相同,内存层600和系统层620的堆迭之间透过多个连接垫(未图标)传递信号。系统层620也透过多个连接垫104固定在基底130上。基底130除了提供基本的支撑作用,也可提供额外的布线空间。内存控制器614a至614c可透过这些接口模块612a至612c和连接垫提供地址信号#A至所述内存装置510a至510c,以存取数据信号#D。此计算机系统700的架构仅为示例。内存控制器、接口模块、内存装置的配置数量不限定为三组。内核616可以是多内核的架构。FIG. 10 is a further embodiment of a computer system 700 of the present application. The three-dimensional wafer product 100 shown in FIG. 1 , after completing the wafer stacking process, further undergoes the wafer cutting process to form a plurality of computer systems 700 . Shown in the memory layer 600 are the memory devices 510 a to 510 c configured according to the embodiment of FIG. 9 . A system layer 620 is stacked with the memory layer 600 . The system layer 620 is cut from the logic circuit layer 120 in FIG. 1 , and includes logic circuits necessary for various computer architectures, such as the core 616 and memory controllers 614a to 614c. Each memory controller 614a-614c is connected to the memory devices 510a-510c in the memory layer 600 through an interface module 612a-612c. The interface module is an interface specially designed to ensure data transmission, commonly known as the physical layer interface (PHY). Similar to FIG. 1 , the stacks of the memory layer 600 and the system layer 620 transmit signals through a plurality of connection pads (not shown). The system layer 620 is also fixed on the substrate 130 through a plurality of connection pads 104 . In addition to providing basic support, the base 130 can also provide additional wiring space. The memory controllers 614a to 614c can provide the address signal #A to the memory devices 510a to 510c through the interface modules 612a to 612c and the connection pads to access the data signal #D. The architecture of this computer system 700 is merely an example. The configuration quantity of memory controllers, interface modules, and memory devices is not limited to three groups. The core 616 may be a multi-core architecture.

在图10的计算机系统700中,配置了一个延迟控制器602,透过一或多个连接垫连接至所述内存层600,将控制信号#S传送至所述内存装置510a至510c中。这些内存装置510a和510c,如前述实施例所说明的,可根据控制信号#S弹性地调整每一内存阵列中的共用线路上连接的内存单元个数,以动态改变内存阵列的延迟特性。所述延迟控制器602可受到内核616的控制。当内核616在执行一应用程序时,可实时判断所述应用程序的延迟需求,指示所述延迟控制器602调整所述内存装置510a至510c。举例来说,使所述内存阵列改变维度机,可将每一内存阵列分割为二的乘方个子阵列,再重组为符合所述应用程序条件的新阵列维度。在执行所述程序时,内核616可透过所述内存控制器614a至614c适应性的使用符合应用程序要求的内存装置510a至510c。In the computer system 700 of FIG. 10, a delay controller 602 is configured, connected to the memory layer 600 through one or more pads, and transmits the control signal #S to the memory devices 510a to 510c. These memory devices 510a and 510c, as described in the foregoing embodiments, can flexibly adjust the number of memory units connected to the common line in each memory array according to the control signal #S, so as to dynamically change the delay characteristics of the memory array. The delay controller 602 may be controlled by a core 616 . When the kernel 616 is executing an application program, it can determine the delay requirement of the application program in real time, and instruct the delay controller 602 to adjust the memory devices 510a to 510c. For example, the memory array re-dimension machine can divide each memory array into sub-arrays that are the power of two, and then reorganize into a new array dimension that meets the conditions of the application program. When executing the program, the kernel 616 can adaptively use the memory devices 510a to 510c that meet the requirements of the application program through the memory controllers 614a to 614c.

图11是本申请的内存阵列和多工器的进一步实施例。在此以内存单元202a和202b来说明多工器402如何减少数据在线的电容负载。一个内存单元中的基本逻辑是由一开关来控制一电容,使电容充电或放电,来代表一位的数据。传送地址信号R0和R1的地址线串接内存单元202a和202b的闸极。用于传递数据信号B0的数据线串接内存单元202a和202b的开关一端。在多工器402不启动时,数据线的第一线段222a和第二线段是相连成同一线段的,使内存单元202a和202b共用同一数据线接收数据信号B0而正常运作。当多工器402受到控制信号#S的指示而开启时,所述多工器402中的开关使第二线段222b与第一线段222a断开,并使第二线段222b改接至一专用线路224。所述专用线路连接至数据驱动器220,使内存单元202b仍然可接收到数据信号。在这种架构下,由于第一线段222a和第二线段222b各别驱动一半数量的内存单元,因负载电容产生的延迟效应可以降低,因此内存的反应速度可以提高。本实施例虽然仅以一列两行内存单元做说明,但可以理解的是多工器402实际上可以是多个数量,安插在一内存阵列中的多行中间,同时控制多条数据线的共用与断开。所述专用线路224上传送的数据信号,不限定是和第一线段222a的数据信号相同。数据驱动器220也可以改良,使内存单元202a和202b在断开后接收不同数据信号。相对地,地址译码器也可以改良,使原本各别传递地址信号R0和R1的地址线在多工器402启动后,共用相同地址信号,例如R0,使内存单元202a和202b同时开启。在这种情况下,内存单元202a和202b在逻辑上可视为是同一行上的不同位。也就是内存阵列的宽高维度从原本的1*2改变为2*1了。这个架构对于弹性适应不同延迟需求具有显着的功效。上述内存单元202a和202b的实际电路结构已存在成熟的现有技术,因此本实施例仅为示意,并不限定详细的实施方式。Figure 11 is a further embodiment of the memory array and multiplexer of the present application. Here, the memory units 202a and 202b are used to illustrate how the multiplexer 402 reduces the capacitive load on the data lines. The basic logic in a memory cell is to control a capacitor with a switch to charge or discharge the capacitor to represent one bit of data. The address lines transmitting the address signals R0 and R1 are connected in series with the gates of the memory units 202a and 202b. The data line for transmitting the data signal B0 is connected in series with the switch ends of the memory units 202a and 202b. When the multiplexer 402 is not activated, the first line segment 222a and the second line segment of the data line are connected to form the same line segment, so that the memory units 202a and 202b share the same data line to receive the data signal B0 and operate normally. When the multiplexer 402 is turned on by the instruction of the control signal #S, the switch in the multiplexer 402 disconnects the second line segment 222b from the first line segment 222a, and reconnects the second line segment 222b to a dedicated Line 224. The dedicated line is connected to the data driver 220 so that the memory unit 202b can still receive data signals. Under this structure, since the first line segment 222 a and the second line segment 222 b respectively drive half the number of memory cells, the delay effect caused by the load capacitance can be reduced, so the response speed of the memory can be improved. Although this embodiment is only illustrated with one column and two rows of memory cells, it can be understood that the number of multiplexers 402 can actually be multiple, inserted in the middle of multiple rows in a memory array, and simultaneously control the sharing of multiple data lines. with disconnect. The data signal transmitted on the dedicated line 224 is not limited to be the same as the data signal on the first line segment 222a. The data driver 220 can also be modified so that the memory units 202a and 202b receive different data signals after they are disconnected. Correspondingly, the address decoder can also be improved, so that the address lines originally transmitting the address signals R0 and R1 respectively share the same address signal, such as R0, after the multiplexer 402 is activated, so that the memory units 202a and 202b are turned on simultaneously. In this case, memory cells 202a and 202b can logically be considered as different bits on the same row. That is, the width and height dimensions of the memory array have changed from the original 1*2 to 2*1. This architecture has a significant effect on elastic adaptation to different latency requirements. The actual circuit structure of the above memory units 202a and 202b has a mature prior art, so this embodiment is only for illustration and does not limit the detailed implementation manner.

图12显示的是在地址在线设置多工器的实施例。前述实施例主要说明如何将共用数据线变短。然而本申请的实施方式也可以从地址线着手。改变内存阵列维度的方式,即第一子阵列810和第二子阵列820的形成方式,也可以是将共用地址线断开为两个较短的地址线。在这种情况中,多条共用地址线传送所述内存单元202的地址信号R0至R7。在多工器802接收到控制信号#S而启动后,将第二子阵列820中被断开的多条地址线转接至专用线路,与地址译码器连接。进一步地,数据驱动器220和地址译码器230也可加以改良,根据所述控制信号#S,使所述第二子阵列820共用所述第一子阵列810的共用数据线,或使用相同的数据信号驱动所述第一子阵列810和所述第二子阵列820。同时,使第二子阵列820使用和第一子阵列不同的地址信号源,例如R8至R15(未图示)。如此,在逻辑上等于是产生了新的内存阵列,维度是原数据位数(阵列宽度)的一半,且地址线数(阵列高度)倍增。图11的做法,由于地址线变短,所以地址线的驱动负载可以减少,同样具有改变内存阵列延迟特性的效果。Fig. 12 shows an embodiment in which a multiplexer is set on the address line. The foregoing embodiments mainly describe how to shorten the common data line. However, embodiments of the present application can also start from address lines. The way to change the dimension of the memory array, that is, the way to form the first sub-array 810 and the second sub-array 820, can also be to break the common address line into two shorter address lines. In this case, a plurality of common address lines transmit address signals R0 to R7 of the memory cells 202 . After the multiplexer 802 receives the control signal #S and starts up, the plurality of disconnected address lines in the second sub-array 820 are switched to dedicated lines and connected to the address decoder. Further, the data driver 220 and the address decoder 230 can also be improved, according to the control signal #S, the second sub-array 820 can share the common data line of the first sub-array 810, or use the same Data signals drive the first sub-array 810 and the second sub-array 820 . At the same time, the second sub-array 820 uses a different address signal source from the first sub-array, such as R8 to R15 (not shown). In this way, logically, a new memory array is generated, the dimension is half of the original data number (array width), and the number of address lines (array height) is doubled. The approach in Fig. 11, because the address lines are shortened, the driving load of the address lines can be reduced, which also has the effect of changing the delay characteristics of the memory array.

图13是本申请内存控制方法的流程图。本申请另外提出一种内存控制方法,应用于前述的计算机系统和内存装置中。在步骤901中,由一内核执行一应用程序时,所述内核根据所述应用程序要求的一应用程序条件,指示延迟控制器发出一控制信号。在步骤903中,所述内存阵列中的所述多工器,根据所述控制信号,使所述内存阵列改变维度。例如将所述记忆阵列分割为二或多个符合所述应用程序条件的子阵列。在步骤905中,所述内核在执行所述应用程序时透过所述内存控制器使用所述内存子阵列。FIG. 13 is a flow chart of the memory control method of the present application. The present application further proposes a memory control method, which is applied to the aforementioned computer system and memory device. In step 901, when an application is executed by a kernel, the kernel instructs the delay controller to send a control signal according to an application condition required by the application. In step 903, the multiplexer in the memory array causes the memory array to change dimensions according to the control signal. For example, the memory array is divided into two or more sub-arrays meeting the conditions of the application program. In step 905, the kernel uses the memory sub-array through the memory controller when executing the application.

综上所述,本申请基于晶圆堆迭技术,提出了一种可以弹性调整阵列维度的内存架构,使得区块链服务器产品有能力可以适应未来算法的需求。To sum up, based on the wafer stacking technology, this application proposes a memory architecture that can flexibly adjust the array dimension, so that blockchain server products have the ability to adapt to the needs of future algorithms.

需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者装置中还存在另外的相同要素。It should be noted that, in this document, the term "comprising", "comprising" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not exclude the presence of additional same elements in the process, method, article or apparatus comprising said element.

上面结合附图对本申请的实施例进行了描述,但是本申请并不局限于上述的具体实施方式,上述的具体实施方式仅仅是示意性的,而不是限制性的,本领域的普通技术人员在本申请的启示下,在不脱离本申请宗旨和权利要求所保护的范围情况下,还可做出很多形式,均属于本申请的保护之内。The embodiments of the present application have been described above in conjunction with the accompanying drawings, but the present application is not limited to the above-mentioned specific implementations. The above-mentioned specific implementations are only illustrative and not restrictive. Those of ordinary skill in the art will Under the inspiration of this application, without departing from the purpose of this application and the scope of protection of the claims, many forms can also be made, all of which belong to the protection of this application.

Claims (13)

1. A computer system based on a wafer stacking architecture, comprising:
a memory crystal layer comprising a plurality of memory devices, wherein each memory device comprises:
a memory array including a common line and a plurality of memory cells, the common line connecting the memory cells; and
a line driver connected to the common line for driving the memory unit;
a logic circuit layer, coupled to the memory crystal layer as a wafer stack, comprising:
a plurality of connection pads for transmitting signals; and
and the delay controller is connected with the memory array through the connecting pad and is used for adjusting the number of the memory units connected on the common line so as to dynamically change the delay characteristic of the memory array.
2. The wafer stack-based architecture of claim 1 wherein the memory array comprises:
a plurality of memory areas, each memory area comprising memory cells of a specific column number or a specific row number;
a plurality of multiplexers each disposed adjacent to each other in the memory area, each connected to the line driver by a dedicated line; wherein:
when the delay controller transmits a control signal through a connecting pad to start the multiplexer, the common line is disconnected into a first line segment and a second line segment, and the second line segment is connected to the special line;
the memory area corresponding to the first line segment forms a first subarray; and
and the memory area corresponding to the second line segment forms a second subarray.
3. The wafer stack-based computer system of claim 2, wherein:
the line driver includes a data driver;
the common line comprises a plurality of common data lines, and each common data line is connected with the data driver and a corresponding column of memory units in the memory units and is used for transmitting data signals; and
after the multiplexer is activated, the memory cells in the second sub-array transmit data signals using the dedicated lines.
4. The wafer stack-based computer system of claim 3 wherein:
the line driver includes an address decoder;
the memory device further comprises a plurality of common address lines, wherein each common address line is connected with the address decoder and a corresponding row of memory units in the memory units and is used for transmitting address signals; and
after the multiplexer is started, the address decoder enables the second subarray and the shared address line of the corresponding line number in the first subarray to receive the same address signal according to the control signal.
5. The wafer stack-based computer system of claim 2 wherein the logic circuit layer further comprises:
a memory controller coupled to the memory array through the connection pad; and
the kernel is connected with the memory controller and the delay controller and is used for executing an application program; wherein:
the kernel sets the multiplexer in the memory array through the delay controller according to the application program condition required by the application program, so that the memory array is divided into two or more subarrays conforming to the application program condition, and the memory array is used through the memory controller when the application program is executed.
6. The wafer stack-based computer system of claim 5 wherein:
the application conditions include a reaction time required by the application; and
the shorter the reaction time required, the greater the number of multiplexers that the delay controller activates.
7. A memory control method applied to the computer system based on the wafer stacking architecture according to claim 5, comprising:
executing an application program;
the kernel sets the multiplexer in the memory array through the delay controller according to the application condition required by the application program, so that the memory array is divided into two or more subarrays conforming to the application program condition, and the memory array is used through the memory controller when the application program is executed, wherein:
the application conditions include a reaction time required by the application.
8. A memory device configured in a memory crystal layer and combined with a logic circuit layer to form a wafer stack-based computer system, comprising:
a memory array including a common line and a plurality of memory cells, the common line connecting the memory cells; and
a line driver connected to the common line for driving the memory unit; wherein:
the memory crystal layer receives a control signal transmitted by the logic circuit layer, and adjusts the number of memory units connected on the common line so as to dynamically change the delay characteristic of the memory array.
9. The memory device of claim 8, wherein the memory array comprises:
a plurality of memory areas, each memory area comprising memory cells of a specific column number or a specific row number;
a plurality of multiplexers each disposed adjacent to each other in the memory area, each connected to the line driver by a dedicated line; wherein:
when the multiplexer is started by the control signal, the common line is disconnected into a first line segment and a second line segment, and the second line segment is connected to the special line;
the memory area corresponding to the first line segment forms a first subarray; and
and the memory area corresponding to the second line segment forms a second subarray.
10. The memory device of claim 9, wherein:
the line driver includes a data driver;
the common line comprises a plurality of common data lines, and each common data line is connected with the data driver and a corresponding column of memory units in the memory units and is used for transmitting data signals; and
after the multiplexer is activated, the memory cells in the second sub-array transmit data signals using the dedicated lines.
11. The memory device of claim 9, wherein:
the line driver includes an address decoder;
the memory device further comprises a plurality of common address lines, wherein each common address line is connected with the address decoder and a corresponding row of memory units in the memory units and is used for transmitting address signals; and
after the multiplexer is started, the address decoder enables the second subarray and the shared address line of the corresponding line number in the first subarray to receive the same address signal according to the control signal.
12. The memory device of claim 9, wherein: when the computer system executes an application program, the control signal is generated according to an application program condition, so that the multiplexer divides the memory array into two or more subarrays which accord with the application program condition for the application program to access.
13. The memory device of claim 12, wherein: the application conditions include a reaction time required by the application.
CN202111219027.6A 2021-10-20 2021-10-20 Computer system, memory device and control method based on wafer stacking architecture Pending CN116009968A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111219027.6A CN116009968A (en) 2021-10-20 2021-10-20 Computer system, memory device and control method based on wafer stacking architecture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111219027.6A CN116009968A (en) 2021-10-20 2021-10-20 Computer system, memory device and control method based on wafer stacking architecture

Publications (1)

Publication Number Publication Date
CN116009968A true CN116009968A (en) 2023-04-25

Family

ID=86023475

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111219027.6A Pending CN116009968A (en) 2021-10-20 2021-10-20 Computer system, memory device and control method based on wafer stacking architecture

Country Status (1)

Country Link
CN (1) CN116009968A (en)

Similar Documents

Publication Publication Date Title
US11810640B2 (en) Memory interface with configurable high-speed serial data lanes for high bandwidth memory
JP4828251B2 (en) Stacked semiconductor memory device and control method thereof
US8869005B2 (en) Error correction in a stacked memory
US7215561B2 (en) Semiconductor memory system having multiple system data buses
CN102405498B (en) Configurable bandwidth memory devices and methods
KR101854579B1 (en) Apparatuses enabling concurrent communication between an interface die and a plurality of dice stacks, interleaved conductive paths in stacked devices, and methods for forming and operating the same
US10180808B2 (en) Software stack and programming for DPU operations
JP5289569B2 (en) Independently controllable and reconfigurable virtual memory device in a memory module that is pin compatible with a standard memory module
KR101746325B1 (en) Method for improving bandwidth in stacked memory devices
US10346325B2 (en) Ring bus architecture for use in a memory module
US11699468B2 (en) Memory device, semiconductor system, and data processing system
CN113704137A (en) In-memory computing module and method, in-memory computing network and construction method
TWI789956B (en) Computer system, memory device and memory control method based on wafer-on-wafer architecture
CN215932612U (en) Computer system and memory device based on wafer stacking architecture
TWM623357U (en) Computer system and memory device based on wafer-on-wafer architecture
CN113793632A (en) Non-volatile programmable chip
CN116009968A (en) Computer system, memory device and control method based on wafer stacking architecture
CN113656346B (en) Three-dimensional chip and computing system
KR101545952B1 (en) Stack memory device and method for operating the same
TW201342375A (en) Memory system and method using stacked memory device dice
CN114281240A (en) Storage system and control method thereof
US11403022B2 (en) Growing and splitting a disk array by moving RAID group members
US12008271B1 (en) Adaptive raid width and distribution for flexible storage
US20240242071A1 (en) Accelerator circuit, semiconductor device, and method for accelerating convolution calculation in convolutional neural network
WO2024192020A1 (en) Compos able memory ranks for full duplex memory

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination