The present disclosure relates to a binarized neural network circuitry using silicon-gated diodes, and more particularly, to a technology for implementing a self-activating binarized neural network circuitry that performs a self-activating function using a silicon-gated diode that performs memory and switching functions in a single element based on a positive feedback loop.
In a conventional von Neumann-based computer system, a processor and a memory are separated and data signals are transmitted through a bus line.
As computing performance improves, a bottleneck phenomenon occurs due to a difference in data processing speed between a processor and a memory, and there are limits to processing large amounts of data.
That is, the von Neumann-based system, which is a product of revolutionary development in the semiconductor industry, has improved the integration density and performance of modern computers, but has the disadvantage of consuming much energy and having long data transmission and standby times due to the physical separation between a processor and a memory hierarchy.
Considering the increase in data-intensive applications such as 5G communication standards, Internet of Things (IoT), and Artificial Intelligence (AI) after the Fourth Industrial Revolution, a new computing paradigm is an essential requirement for large-scale data processing.
To solve the above-mentioned problems, research is being actively conducted on logic-in-memory (LIM) technology that combines calculation and memory functions.
According to the logic-in-memory technology, since the calculating function of a processor and the memory function of a memory are performed in the same space, the time and power required during data transmission may be reduced, and the integration of the system may be greatly improved.
The conventional logic-in-memory technology has been actively researched based on static random-access memory (SRAM) and dynamic RAM (DRAM), which are volatile memory elements, and resistive RAM (ReRAM), magnetoresistive RAM (MRAM), and phase-change RAM (PCRAM), which are non-volatile memory elements.
To overcome limitations in processing large amounts of data, Package-on-Package (POP) and through-silicon via (TSV) technologies, which integrate logic memory into a single chip, are being researched. However, since logic and memory functions are not performed simultaneously in a transistor, problems related to the bottleneck phenomenon, power consumption, computational efficiency, and integration still remain.
In addition, in the case of logic-in-memory technology based on a non-volatile memory element, a complex manufacturing process is required because non-silicon materials are used. In addition, this technology has low element uniformity and stability, making it difficult to put into practical use.
In addition, previously researched logic-in-memory technologies cannot implement all basic complementary metal-oxide semiconductor (CMOS) logic operations in one cell and have low integration because individual circuitry and wiring are required for each logic operation.
Accordingly, the development of a binarized neural network technology using silicon-gated diodes that can be applied to the CMOS process and perform both switching and memory functions simultaneously and a self-activated neural network technology that performs the activation function of a neuron circuitry in a neural network is required.
Therefore, the present disclosure has been made in view of the above problems, and it is an object of the present disclosure to implement a self-activating binarized neural network circuitry that performs a self-activating function using a silicon-gated diode that performs memory and switching functions in a single element based on a positive feedback loop.
It is another object of the present disclosure to implement a self-activating binarized neural network circuitry having unidirectional switching characteristics by controlling a potential barrier; and being capable of increasing the area and computational efficiency of an artificial neural network by having excellent memory characteristics of a silicon-gated diode that also has memory characteristics as holes or electrons accumulate in a potential well due to a positive feedback loop and a self-activating function through the linearity of an output signal according to an input signal.
It is still another object of the present disclosure to increase the computational accuracy of an artificial neural network by enabling the use of the CMOS process, enabling the fabrication of large-scale arrays, and almost eliminating the deviation between the characteristics of a silicon-gated diode and the characteristics of a synaptic device.
It is still another object of the present disclosure to implement a binarized neural network circuitry including a memory array that is composed of silicon-gated diodes based on a positive feedback loop and performs the activation function of a neuron circuitry; and having excellent uniformity and stability.
It is yet another object of the present disclosure to implement a binarized neural network circuitry capable of reducing standby power by using the excellent memory characteristics of silicon-gated diodes; and capable of being used in the next-generation artificial intelligence computing technology by increasing computational efficiency with low power consumption through excellent switching characteristics.
In accordance with one aspect of the present disclosure, provided is a binarized neural network circuitry including a plurality of silicon-gated diodes in which a diode structure as a channel area is located between an anode terminal and a cathode terminal, a gate terminal is located on the diode structure to implement unidirectional switching through potential barrier control in the channel area based on different voltages applied to each of the anode terminal and the gate terminal, and memory characteristics are realized as holes or electrons accumulate in the potential well due to a positive feedback loop, wherein the silicon-gated diodes operate as synaptic elements in a memory array connected in parallel, and results of a multiply-accumulate (MAC) operation are output based on an input signal applied from an input line processor connected to the memory array and a weight update signal applied from a synapse line processor connected to the memory array.
The silicon-gated diode may receive an anode voltage of the anode terminal as the input signal, generate a latch-up phenomenon due to the positive feedback loop as the applied input signal increases in a positive direction, have any one of two memory states for the channel area, receive a gate voltage of the gate terminal as the weight update signal to control the input signal causing the latch-up phenomenon, update a synaptic state associated with the memory state according to application of the input signal and the weight update signal, and perform the MAC calculating function according to the synaptic weight.
The silicon-gated diode may output a current signal corresponding to any one of “0” to “1” as an operation result by performing a multiplication operation between a continuous input from “0” to “1” as the applied input signal increases in a positive direction and a synaptic weight of any one of a “0” state and a “1” state corresponding to the two memory states.
In the silicon-gated diode, in a case of a potentiation operation in which the synaptic weight is updated to the “1” state, an operation result may be output as a current signal in proportion to the applied input signal, and in a case of a depression operation in which the synaptic weight is updated to the “0” state, a current signal may be output at 0 mA regardless of the applied input signal.
When the synaptic weight is in a “1” state, the silicon-gated diode may output a cathode current of the cathode terminal in a form similar to a graph shape of a rectified linear unit (ReLU) function, regardless of the gate voltage.
The memory array may connect the anode terminal, the gate terminal, and the cathode terminal in parallel in the silicon-gated diodes to form an input line, a weight line, and an output line, respectively. The input line may be arranged perpendicular to the weight line and the output line, and the weight line and the output line may be arranged in parallel.
The input line may receive the input signal, the weight line may receive the weight update signal, and the output line may output the MAC operation result based on the input signal and the weight update signal to a next artificial neural network.
In the memory array, the silicon-gated diodes may be connected in a form of N×M, and through a multiplication operation of a synaptic weight that is updated based on an input signal applied to the input line and a weight update signal applied to the weight line, each current may be added through the output line and a sum operation may be performed to output the MAC operation results as a synaptic weight matrix.
In the memory array, when the N and the M are “2”, the input signal may consist of a first input signal and a second input signal, the synaptic weight may consist of a first synaptic weight to a fourth synaptic weight, a first current and a second current may be output to the output line, and the synaptic weight matrix consisting of the first current and the second current may be calculated as a result of a vector matrix multiplication operation of the synaptic weight and the input signal.
In the memory array, when both the first and second input signals are applied while the weight update signal is fixed at 1 V, a value of the first current may be twice a value of the second current, and afterwards, when only the first input signal is applied, a value of the first current may be equal to a value of the second current.
In the memory array, when only the second input signal is applied, only the first current may have a value proportional to the second input signal, and when the first input signal and the second input signal are not applied, the first current and the second current may be measured close to 0 mA.
The silicon-gated diode may include any one of a single silicon-gated diode, a double silicon-gated diode, and a triple silicon-gated diode.
The present disclosure can implement a self-activating binarized neural network circuitry that performs a self-activating function using a silicon-gated diode that performs memory and switching functions in a single element based on a positive feedback loop.
The present disclosure can implement a self-activating binarized neural network circuitry having unidirectional switching characteristics by controlling a potential barrier; and being capable of increasing the area and computational efficiency of an artificial neural network by having excellent memory characteristics of a silicon-gated diode that also has memory characteristics as holes or electrons accumulate in a potential well due to a positive feedback loop and a self-activating function through the linearity of an output signal according to an input signal.
The present disclosure can increase the computational accuracy of an artificial neural network by enabling the use of the CMOS process, enabling the fabrication of large-scale arrays, and almost eliminating the deviation between the characteristics of a silicon-gated diode and the characteristics of a synaptic device.
The present disclosure can implement a binarized neural network circuitry including a memory array that is composed of silicon-gated diodes based on a positive feedback loop and performs the activation function of a neuron circuitry; and having excellent uniformity and stability.
The present disclosure can implement a binarized neural network circuitry capable of reducing standby power by using the excellent memory characteristics of silicon-gated diodes; and capable of being used in the next-generation artificial intelligence computing technology by increasing computational efficiency with low power consumption through excellent switching characteristics.
Hereinafter, the embodiments of the present disclosure will be described in detail with reference to the drawings.
However, it should be understood that the present disclosure is not limited to the embodiments according to the concept of the present disclosure, but includes changes, equivalents, or alternatives falling within the spirit and scope of the present disclosure.
In the following description of the present disclosure, detailed description of known functions and configurations incorporated herein will be omitted when it may make the subject matter of the present disclosure unclear.
In addition, the terms used in the specification are defined in consideration of functions used in the present disclosure, and can be changed according to the intent or conventionally used methods of clients, operators, and users. Accordingly, definitions of the terms should be understood on the basis of the entire description of the present specification.
In description of the drawings, like reference numerals may be used for similar elements.
The singular expressions in the present specification may encompass plural expressions unless clearly specified otherwise in context.
In this specification, expressions such as “A or B” and “at least one of A and/or B” may include all possible combinations of the items listed together.
Expressions such as “first” and “second” may be used to qualify the elements irrespective of order or importance, and are used to distinguish one element from another and do not limit the elements.
It will be understood that when an element (e.g., first) is referred to as being “connected to” or “coupled to” another element (e.g., second), the first element may be directly connected to the second element or may be connected to the second element via an intervening element (e.g., third).
As used herein, “configured to” may be used interchangeably with, for example, “suitable for”, “ability to”, “changed to”, “made to”, “capable of”, or “designed to” in terms of hardware or software.
In some situations, the expression “device configured to” may mean that the device “may do ˜” with other devices or components.
For example, in the sentence “processor configured to perform A, B, and C”, the processor may refer to a general purpose processor (e.g., CPU or application processor) capable of performing corresponding operation by running a dedicated processor (e.g., embedded processor) for performing the corresponding operation, or one or more software programs stored in a memory device.
In addition, the expression “or” means “inclusive or” rather than “exclusive or”.
That is, unless mentioned otherwise or clearly inferred from context, the expression “x uses a or b” means any one of natural inclusive permutations.
Terms, such as “unit” or “module”, etc., should be understood as a unit that processes at least one function or operation and that may be embodied in a hardware manner, a software manner, or a combination of the hardware manner and the software manner.
Referring to
A gate insulating film 105 is located on the second channel area 103, and a gate terminal 106 is located on the gate insulating film 105.
In a circuitry symbol 110 of the silicon-gated diode that constitutes the binarized neural network circuitry according to one embodiment of the present disclosure, a channel area 112 is located between an anode terminal 111 and a cathode terminal 113, and a gate terminal 114 is connected thereto.
Referring to
A gate insulating film 125 is located on the first channel area 122, and a gate terminal 126 is located on the gate insulating film 125.
In a circuitry symbol 130 of the silicon-gated diode that constitutes the binarized neural network circuitry according to one embodiment of the present disclosure, a channel area 132 is located between an anode terminal 131 and a cathode terminal 133, and a gate terminal 134 is connected thereto.
According to one embodiment of the present disclosure, an input line is connected to the anode terminal 111 or the anode terminal 131, an output line is connected to the cathode terminal 113 or the cathode terminal 133, and the memory state of the channel area is determined according to the update of a synaptic weight.
Referring to
The intrinsic area 202 operates as a channel area based on a gate voltage input from any one of the first gate terminal 205 and the second gate terminal 206.
In a circuitry symbol 210 of the silicon-gated diode that constitutes the binarized neural network circuitry according to one embodiment of the present disclosure, an intrinsic area 212 is located between an anode terminal 211 and a cathode terminal 213, and a first gate terminal 214 and a second gate terminal 215 are connected thereto.
Referring to
The intrinsic area 222 operates as a channel area based on a gate voltage input from the programming gate terminal 225.
More specifically, when the level of a program voltage (VPG) input from the intrinsic area 222 to the programming gate terminal 225 is high, the intrinsic area 222 operates as an n channel corresponding to a first channel operation. When the level of the program voltage (VPG) is low, the intrinsic area 222 operates as a p channel corresponding to a second channel operation.
When the first channel operation is performed, when the level of a control voltage (VCG) input through the control gate terminal 226 is high, the silicon-gated diode 220 is determined to be in the on state. When the level of a control voltage (VCG) input through the control gate terminal 226 is low, the silicon-gated diode 220 is determined to be in the off state.
In a circuitry symbol 230 of the silicon-gated diode that constitutes the binarized neural network circuitry according to one embodiment of the present disclosure, an intrinsic area 232 is located between an anode terminal 231 and a cathode terminal 233, and a programming gate terminal 234 and a control gate terminal 235 are connected thereto.
For example, the anode terminal, the gate terminal, and the cathode terminal may be referred to as an anode electrode, a gate electrode, and a cathode electrode, respectively.
That is, the silicon-gated diode that constitutes the binarized neural network circuitry according to one embodiment of the present disclosure may have a structure in which, on a silicon channel, a single gated electrode terminal is deposited on an n- or p-type doped channel area or multiple gated electrode terminals are deposited on an intrinsically doped area.
In addition, the silicon-gated diode operates through the induction of a positive feedback loop by a potential barrier and has switching and memory characteristics, so the silicon-gated diode may be configured as a synaptic element of the binarized neural network.
That is, the silicon-gated diode has switching and memory characteristics, making it possible to implement a self-activating binarized neural network circuitry in which the activation function of a neuron circuitry is performed by a neural network.
Accordingly, the present disclosure may implement a self-activating binarized neural network circuitry that performs a self-activating function using a silicon-gated diode that performs memory and switching functions in a single element based on a positive feedback loop.
Referring to
The graph 300 shows that, when the silicon-gated diode that constitutes the binarized neural network circuitry increases an anode voltage (VIN) in the positive direction, a latch-up phenomenon occurs due to a positive feedback loop and two states are exhibited.
In addition, the anode voltage at which latch-up occurs varies depending on the gate voltage (VW) of the silicon-gated diode.
The binarized neural network circuitry may perform synaptic weight update and multiply-accumulate (MAC) calculating functions by applying an anode voltage and a gate voltage based on the characteristics of the single element.
Regardless of the gate voltage, the cathode current (IDiode) of an element in the “1” state may be similar to the graph shape of the rectified linear unit (ReLU) function.
The ReLU function may be the activation function of a neuron circuitry used in most artificial neural network architectures.
That is, in the silicon-gated diode that constitutes the binarized neural network of the present disclosure, the output current has the shape of the ReLU function in the synaptic weight state of “1” regardless of the gate voltage, so a self-activating binarized neural network circuitry that self-activates a neuron circuitry may be implemented.
Through the unidirectional switching or self-rectification characteristics and high linearity of a single silicon-gated diode, a self-activating binarized neural network that performs a self-activating function may be implemented.
Referring to
Referring to
When the anode voltage (VIN) is 0 V, it may be in a standby state.
A graph 320 shows the Bayesian neural network (BNN) operating conditions, in which the cathode current (IDiode) increases rapidly when the gate voltage (VW) is 1 V at about 2.5 V as the anode voltage (VIN) increases.
The graph 320 shows that a positive feedback loop is eliminated at around 1 V as the anode voltage (VIN) decreases.
The bistable nature is expressed as cathode current (IDiode) to anode voltage (VIN) and the high ratio of current magnitudes of states 1 and 0 may be about 108.
When a gate voltage (VW) is 1 V for an anode voltage (VIN) of 2 V, the electrical properties of a p-n diode are inherited with unipolar switching characteristics.
Regardless of the gate voltage (VW), the shape of the curve for cathode current (IDiode) versus anode voltage (VIN) in state 1 is similar to the shape of the rectified linear unit function used in the activation function of a neural network.
The difference in cathode current (IDiode) depending on the change between state 1 and state 0 may be used in MAC operation.
Referring to
For example, the binarized neural network circuitry 400 using silicon-gated diodes may be an artificial neural network that reduces calculation time and power consumption by binarizing synaptic weights.
Accordingly, the present disclosure may implement a self-activating binarized neural network circuitry having unidirectional switching characteristics by controlling a potential barrier; and being capable of increasing the area and computational efficiency of an artificial neural network by having excellent memory characteristics of a silicon-gated diode that also has memory characteristics as holes or electrons accumulate in a potential well due to a positive feedback loop and a self-activating function through the linearity of an output signal according to an input signal.
A plurality of silicon-gated diodes may operate as synaptic elements 401 in a memory array connected in parallel and may output multiply-accumulate (MAC) operation results based on an input signal applied from an input line processor 402 connected to the memory array and a weight update signal applied from a synapse line processor 403 connected to the memory array.
In the memory array as the binarized neural network circuitry 400, in a plurality of silicon-gated diodes, an anode terminal, a gate terminal, and a cathode terminal may be connected in parallel to form an input line (IL), a weight line (WL), and an output line (OL), respectively. The input line (IL) may be arranged perpendicular to the weight line (WL) and output line (OL), and the weight line (WL) and the output line (OL) may be arranged parallel to each other.
The anode terminal is connected to the input line (IL), the gate terminal is connected to the weight line (WL), and the cathode terminal is connected to the output line (OL).
The gate terminal and the cathode terminal may form the weight line (WL) and the output line (OL), respectively.
For example, in the memory array, depending on M which is the number of rows formed by a plurality of silicon-gated diodes, the first weight line (WL0) to the M-th weight line (WLm) may be configured, and the first output line (OL0) to the M-th output line (OLm) may be configured.
The number of input lines (IL) corresponding to columns and the number of output lines (OL) corresponding to rows may be the same or different.
However, the weight line (WL) and the output line (OL) must have the same number (M).
For example, since the number of elements constituting rows and columns may be as many as the number of inputs and outputs, N and M may be the same number or different numbers.
For example, in the memory array, depending on N which is the number of columns formed by a plurality of silicon-gated diodes, the first input line (IL0) to the N-th input line (ILn) may be configured.
The silicon-gated diode receives the anode voltage of the anode terminal as an input signal, and generates a latch-up phenomenon due to a positive feedback loop as the applied input signal increases in the positive direction.
The silicon-gated diode has one of two memory states for the channel area. The silicon-gated diode receives the gate voltage of the gate terminal as a weight update signal, so the input signal at which the latch-up phenomenon occurs may be controlled.
In addition, the silicon-gated diode may update a synaptic state related to any one memory state according to the application of an input signal and a weight update signal, and may perform a MAC calculating function according to a synaptic weight.
For example, the silicon-gated diode may output, as an operation result, a current signal corresponding to any one of “0” to “1” by performing a multiplication operation between a continuous input from “0” to “1” as the applied input signal increases in the positive direction and a synaptic weight of any one of two memory states of “0” and “1”.
Referring to
Accordingly, the binarized neural network circuitry may implement artificial intelligence functions such as image recognition through a continuous multi-layer neural network.
The synaptic weight may be updated through the input voltage of an input line and a weight line.
The input signal (VIN) applied through the input line may have continuous values, which may mean a continuous input A from 0 to 1 in the binarized neural network.
The current signal (IOUT) corresponding to an output may be read through OL, and at this time, the input A may have the form of a ReLU function for the input signal (VIN) due to the self-activation characteristics of the silicon-gated diode.
The operations corresponding to the first case 500 and the second case 501 are summarized in Table 1 below.
The silicon-gated diode has two memory states, a “1” state and a “0” state, which may, respectively, mean a binarized synaptic weight W (0) and a binarized synaptic weight W (1).
In the first case 500, when the synaptic weight W is 0, since the case means that the silicon-gated diode is turned off (“0” state), regardless of the applied input signal (VIN), the output current (IOUT) flows at a very low level.
Regarding the anode voltage (A) of 0˜1, a voltage below the reference voltage corresponds to 0, and a voltage greater than the reference voltage corresponds to 1.
For example, when the reference voltage is 1 V, a voltage greater than 1 V may correspond to 1.
In addition, in the output, a current greater than the reference current of 0 to 1 is output as 1, and a current smaller than the reference current is output as 0.
This may mean that the multiplication operation result of an anode voltage and a synaptic weight in the MAC operation of the binarized neural network is always 0 regardless of an input signal.
In the second case 501, when the synaptic weight W is 1, since the case means that the silicon-gated diode is turned on (“1” state), the output current (IOUT) may change linearly in proportion to the applied input signal (VIN).
This means that, in the MAC operation of the binarized neural network, the multiplication operation result of an anode voltage and a synaptic weight may be proportional to an input.
According to one embodiment of the present disclosure, the silicon-gated diode may output a current signal corresponding to any one of “0” to “1” as an operation result by performing a multiplication operation between a continuous input from “0” to “1” as an applied input signal increases in the positive direction and the synaptic weight of one of two memory states of “0” and “1”.
Here, in relation to inputs from “0” to “1” as the input signals increase or decrease in the positive direction, a voltage below the reference voltage corresponds to “0”, and a voltage greater than the reference voltage corresponds to “1”.
For example, when the reference voltage is 1 V, a voltage greater than 1 V may correspond to “1”.
Referring to
In state “0” according to the first energy band diagram 510, excess charge carriers are absent in the potential wells of the n- and p-doping regions of the energy band diagram, preventing diode current (IDiode) from flowing into a diode.
In contrast, in state “1” according to the second energy band diagram 511, excess charge carriers accumulate in the potential wells of the n- and p-doping regions of the energy band diagram, causing diode current to flow into a diode.
Modulating a potential barrier allows the diode to exhibit bistable properties.
A weight voltage of 0 V is applied when performing potentiation and depression operations, and a weight voltage of 1 V is applied when performing standby and multiplication operations.
The conductance of state “0” and state “1” may be the conductance of an element that varies depending on the presence or absence of charge in a potential well depending on a rising or depression operation.
Referring to
According to one embodiment of the present disclosure, in the silicon-gated diode, in the case of a potentiation operation in which a synaptic weight is updated to the state of “1”, an operation result is output as a current signal in proportion to an applied input signal. In the case of a depression operation in which a synaptic weight is updated to the state of “0”, a current signal may be output as 0 mA regardless of an applied input signal.
When a synaptic weight is in the state of “1”, regardless of a gate voltage, the silicon-gated diode may output the cathode current of a cathode terminal in a form similar to the graph shape of a rectified linear unit (ReLU) function.
To update the synaptic weight W to 1 according to the operation method in the timing diagram 600, the voltage pulses of an input voltage (VIN) of 2 V applied to an anode terminal and a synaptic weight voltage (VW) of 0 V applied to a gate terminal are applied.
Thereafter, continuous IOUT from 0 mA to 7.4 mA may be measured proportional to continuous VIN from 1.1 to 2.0 V.
To update the synaptic weight (W) to 0, the voltage pulses of VIN of 2 V and VW of 0 V may be applied.
At this time, regardless of the same input voltage (VA), a very low level output current close to 0 mA may be measured.
This shows that a single synaptic element performs a multiplication operation based on excellent linearity.
Referring to
The graph 610 shows an experiment result 611 and straight fitting 612.
Accordingly, the present disclosure may increase the computational accuracy of an artificial neural network by enabling the use of the CMOS process, enabling the fabrication of large-scale arrays, and almost eliminating the deviation between the characteristics of a silicon-gated diode and the characteristics of a synaptic device.
Technologies using conventional next-generation memories, such as ReRAM and MRAM, have the disadvantages of not being able to apply the CMOS process, poor element uniformity and stability, and being difficult to put into practical use due to complex processing processes.
In conventional artificial neural network technology, multiple neural network layers are connected through an activation function, but to implement this, a complex additional neuron circuitry is required, which may deteriorate the area and energy efficiency of an artificial neural network.
However, the present disclosure may implement a binarized neural network circuitry capable of reducing standby power by using the excellent memory characteristics of silicon-gated diodes; and capable of being used in the next-generation artificial intelligence computing technology by increasing computational efficiency with low power consumption through excellent switching characteristics.
Referring to
Referring to
The timing diagram 710 shows the results of a sum operation in which an output current from each diode is added to the output current (IOUT) through the common output line (OL).
The MAC operation of the binarized neural network uses a silicon-gated diode array connected in a 2×2 form as the binarized neural network circuitry 800.
In the memory array related to the binarized neural network circuitry 800, a plurality of silicon-gated diodes are connected in an N×N form. Through the multiplication operation of a synaptic weight that is updated based on an input signal applied to an input line and a weight update signal applied to a weight line, each current is added through the output line, a sum operation is performed, and the MAC operation result may be output to a synaptic weight matrix.
In addition, in the memory array, when N is “2”, the input signal consists of a first input signal and a second input signal, the synaptic weight consists of a first synaptic weight to a fourth synaptic weight and a first current and a second current output to the output line, and the synaptic weight matrix consisting of the first current and the second current may be calculated as the result of a vector matrix multiplication operation of the synaptic weight and the input signal.
The silicon-gated diode may perform multiplication operations of each input signal (VIN1, VIN2) and synaptic weights (W11, W12, W21, W22), and may perform a sum operation in which each output current (IOUT1, IOUT2) is added through the common first output line (OL1) and the second output line (OL2).
That is, mathematically, the 2×2 array may perform a vector-matrix multiplication operation
and the synaptic weight matrix 801 is exemplified as
In the binarized neural network circuitry 800, after updating the synaptic weight of each element, when both VIN1 and VIN2 are applied while VW is fixed at 1 V, IOUT1 has exactly twice the value of IOUT2. Afterwards, when only VIN1 is applied, IOUT1 and IOUT2 have the same value.
Conversely, when only VIN2 is applied, only IOUT1 may have a value proportional to the input signal.
Lastly, when all input signals are not applied, both IOUT1 and IOUT2 may be measured at low levels.
These results are consistent with the multiplication operation between the input vector and the synaptic weight matrix.
A silicon-gated diode array connected in a 2×2 form through the binarized neural network circuitry 800 may implement the MAC operation of a self-activating binarized neural network.
The binarized neural network circuitry 800 is described in a 2×2 form, but may be expanded to an N×M form.
The timing diagram 810 shows that, in a first area 811, when both VIN1 and VIN2 are applied with VW fixed at 1 V, IOUT1 has exactly twice the value of IOUT2.
A second area 812 shows that IOUT1 and IOUT2 may have the same value when only VIN1 is applied.
A third area 813 shows that, when only VIN2 is applied, only IOUT1 may have a value proportional to the input signal.
A fourth area 814 shows that both IOUT1 and IOUT2 may be measured at low levels when no input signal is applied.
That is, in the memory array related to the binarized neural network circuitry 800, when both the first and second input signals are applied while the weight update signal is fixed at 1 V, the value of the first current is twice that of the second current. Afterwards, when only the first input signal is applied, the first current and the second current may have the same value.
In addition, in the memory array, when only the second input signal is applied, only the first current has a value proportional to the second input signal. When the first input signal and the second input signal are not applied, the first current and the second current may be measured to be approximate 0 mA.
Accordingly, the present disclosure may implement a binarized neural network circuitry including a memory array that is composed of silicon-gated diodes based on a positive feedback loop and performs the activation function of a neuron circuitry; and having excellent uniformity and stability.
The matrix operation of the timing diagrams shown in
During all matrix operations, the weight voltage is fixed at 1 V and the input voltage may have 0 V or a value between 1 V and 2V.
All matrix equations commonly include
which indicates that the VIN pulse in
That is, as the weights of the 2×2 array elements change while both inputs “1” and “2” are fixed to “1”, the operation result meets the matrix equation.
The matrix operation according to the timing diagram 900 may be the same as the matrix operation on the timing diagram 900.
During matrix operation, the weight voltage is 1 V and the input voltage may have 0 V or a value between 1 V and 2 V.
The matrix operation according to the timing diagram 910 may be the same as the matrix operation on the timing diagram 910.
The matrix operation according to the timing diagram 920 may be the same as the matrix operation on the timing diagram 920.
The matrix operation according to the timing diagram 930 may be the same as the matrix operation on the timing diagram 930.
The matrix operation according to the timing diagram 940 may be the same as the matrix operation on the timing diagram 940.
The matrix operation according to the timing diagram 950 may be the same as the matrix operation on the timing diagram 950.
The above-described matrix operations may be summarized as shown in Equation 1 below.
In Equation 1, W may represent a binarized weight (conductance), IN may represent an input signal (voltage), and OUT may represent an output signal (current) according to a matrix operation.
According to the timing diagrams 900 to 950, when the output current according to the first output current (IOUT1) and the second output current (IOUT2) is 7.5 mA, it is “1”. When the output current is 15 mA, it is “2”. When the output current is less than 7.5 mA, it is “0”.
However, according to the above-mentioned example, 7.5 mA is not limited to meaning output “1”. Regardless of the input voltage, when the output current is 0 mA, it means output “0”. In contrast, when the output current is measured at a certain level according to the input voltage, it may mean “1”.
This proportional relationship may be confirmed through
The binarized neural network circuitry using silicon-gated diodes may perform a matrix operation for each weight, in this case, the output of the matrix MAC operation has high consistency with a simplified vector-matrix multiplication (VMM) equation due to the high uniformity of the silicon-gated diode.
In
For example, the image 1000 and the image 1010 may be optical images of a silicon-gated diode for performing a matrix MAC operation between the binarized weight of the BNN and an analog input.
Referring to
Referring to
The silicon-gated diode maintains a memory state with a very low output current (IDiode).
Referring to
IOUT, the product of an input voltage and a binarized weight, may be obtained from the sum of the output current (IDiode).
In the above-described specific embodiments, elements included in the disclosure are expressed in singular or plural in accordance with the specific embodiments shown.
It should be understood, however, that the singular or plural representations are to be chosen as appropriate to the situation presented for the purpose of description and that the above-described embodiments are not limited to the singular or plural constituent elements. The constituent elements expressed in plural may be composed of a single number, and constituent elements expressed in singular form may be composed of a plurality of elements.
In addition, the present disclosure has been described with reference to exemplary embodiments, but it should be understood that various modifications may be made without departing from the scope of the present disclosure.
Therefore, the scope of the present disclosure should not be limited by the embodiments, but should be determined by the following claims and equivalents to the following claims.
| Number | Date | Country | Kind |
|---|---|---|---|
| 10-2023-0114457 | Aug 2023 | KR | national |