This application is based on and claims priority under 35 U.S.C. § 119 to Korean Patent Application No. 10-2022-0021736, filed on Feb. 18, 2022, in the Korean Intellectual Property Office, the disclosure of which is incorporated by reference herein in its entirety.
The disclosure relates to a ferroelectric field effect transistor having electrode structures with various curvature, a neural network apparatus having electrode structures with various curvature, and an electronic apparatus including the neural network apparatus.
There has been an increased interest in neuromorphic processors that perform neural network operations. A neuromorphic processor may be used as a neural network apparatus for driving various neural networks such as a convolutional neural network (CNN), a recurrent neural network (RNN), and a feedforward neural network (FNN) and may be utilized in fields including data classification, image recognition, autonomous control, speak-to-text, etc.
The neuromorphic processor may include (or be connected to) a plurality of memory cells for storing weights. A memory cell may be implemented by various devices, and recently, a non-volatile memory having a simple structure has been proposed as a memory cell for a neuromorphic processor in order to reduce an area of the memory cell and power consumption.
Provided is a ferroelectric field effect transistor having a linear response characteristic to an applied voltage.
Provided is a neural network apparatus including a synaptic element having a linear response characteristic to an applied voltage.
In addition, provided is an electronic apparatus including a neural network apparatus.
Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented embodiments of the disclosure.
A ferroelectric field effect transistor according to an embodiment includes: a substrate; a source protruding from an upper surface of the substrate in a first direction; a drain protruding from an upper surface of the substrate in the first direction; a channel spaced apart from the upper surface of the substrate and extending between and connected to the source and the drain in a second direction different from the first direction; a ferroelectric film surrounding an outer circumferential surface of the channel; and a gate electrode surrounding the ferroelectric film, wherein the channel may have curved cross-sections having a plurality of different curvatures.
The channel may have an elliptical pillar shape with a continuously changing radius in an azimuth direction different from the first direction and the second direction.
The channel may have a tapered shape.
The channel may include a first channel having a first radius and a second channel having a second radius different from the first radius.
The ferroelectric film may include a first ferroelectric film surrounding an outer circumferential surface of the first channel and a second ferroelectric film surrounding an outer circumferential surface of the second channel, and the gate electrode may be arranged to surround both the first ferroelectric film and the second ferroelectric film.
In addition, the channel further includes a third channel having a third radius different from the first radius and the second radius, the ferroelectric film further comprising a third ferroelectric film surrounding an outer circumferential surface of the third channel, and the gate electrode may surround all of the first ferroelectric film, the second ferroelectric film, and the third ferroelectric film.
For example, a ratio of a thickness of the ferroelectric film to a radius of the channel may be greater than 0 and equal to or less than 2.
For example, the gate electrode may include at least one of TiN, TaN, WN, NbN, W, Mo, or Pt.
For example, the ferroelectric film may include an oxide of at least one of Si, Al, Hf, or Zr and a dopant including at least one of Si, Al, Y, La, Gd, Mg, Ca, Sr Ba, Ti, Zr, Hf, and N, and at least one material from among MgZnO, AlScN, BaTiO3, Pb(Zr, Ti)O3, SrBiTaO7, or polyvinylidene fluoride (PVDF).
A neural network apparatus according to an embodiment includes: a plurality of word lines; a plurality of bit lines crossing the plurality of word lines; and a plurality of synaptic elements at intersections between the plurality of word lines and the plurality of bit lines, each of the plurality of synaptic elements electrically connected to a corresponding word line of the plurality of word lines and a corresponding bit line of the plurality of bit lines, wherein each of the plurality of synaptic elements includes comprises an inner pillar, a ferroelectric film surrounding an outer circumferential surface of the inner pillar, and an outer electrode surrounding an outer circumferential surface of the ferroelectric film, and the inner pillar has a plurality of different curvatures.
The inner pillar may have an elliptical pillar shape.
The inner pillar may have a tapered shape.
The inner pillar may include a first element having a first radius and a second element having a second radius different from the first radius, and the first element and the second element may be connected in parallel to each other.
In addition, the inner pillar may further include a third element in which having a third radius different from the first radius and the second radius, and the first element, the second element, and the third element may be connected in parallel to one another.
For example, a ratio of a thickness of the ferroelectric film to a radius of the inner electrode may be greater than 0 and equal to or less than 2.
For example, the inner pillar and the outer electrode may each include at least one of TiN, TaN, WN, NbN, W, Mo, or Pt.
Each of the plurality of synaptic elements further comprises an access transistor, and one of the inner pillar or the outer electrode is electrically connected to a drain of the access transistor, and a gate of the access transistor is electrically connected to the corresponding word line, and a source of the access transistor is electrically connected to the corresponding bit line.
The neural network apparatus may further include an output circuit configured to output a signal from the plurality of bit lines.
Each of the plurality of synaptic elements may further include a field effect transistor, and one of the inner pillar or the outer electrode may be electrically connected to a gate of the field effect transistor.
Each of the plurality of synaptic elements further may include an access transistor and a field effect transistor, wherein a gate of the access transistor is electrically connected to a corresponding one of the plurality of word lines, a source of the access transistor is electrically connected to a corresponding one of the plurality of bit lines, a gate of the field effect transistor is electrically connected to a drain of the access transistor, and one of the inner electrode or the outer electrode of the capacitor may be electrically connected to a drain of the field effect transistor.
A neural network apparatus according to another embodiment includes: a plurality of word lines; a plurality of bit lines crossing the plurality of word lines; and a plurality of synaptic elements at intersections between the plurality of word lines and the plurality of bit lines, each of the plurality of synaptic elements electrically connected to a corresponding word line of the plurality of word lines and a corresponding bit line of the plurality of bit lines, wherein each of the plurality of synaptic elements includes an access transistor and a ferroelectric field effect transistor, and the ferroelectric field effect transistor comprising a source, a drain, a channel extending between and electrically connecting the source and the drain, a ferroelectric film surrounding an outer circumferential surface of the channel, and a gate electrode surrounding the ferroelectric film, wherein the channel has curved cross-sections having a plurality of different curvatures.
A gate of the access transistor may be electrically connected to a corresponding one of the plurality of word lines, a source of the access transistor may be electrically connected to a corresponding one of the plurality of bit lines, and a drain of the access transistor may be electrically connected to a gate of the ferroelectric field effect transistor.
The neural network apparatus may further include a plurality of input lines and a plurality of output lines, a source of the ferroelectric field effect transistor may be electrically connected to a corresponding one of the plurality of input lines, and a drain of the ferroelectric field effect transistor may be electrically connected to a corresponding one of the plurality of output lines.
An electronic apparatus according to an embodiment may include: a neural network apparatus having at least one of the above-described structure; a memory including computer-executable instructions; and a processor configured to control the neural network apparatus by executing the computer-executable instructions stored in the memory such that the neural network apparatus performs a neural network operation based on input data received from the processor, and generate an information signal corresponding to the input data based on the result of the neural network operation.
The above and other aspects, features, and advantages of certain embodiments of the disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout. In this regard, the present embodiments may have different forms and should not be construed as being limited to the descriptions set forth herein. Accordingly, the embodiments are described below, by referring to the figures, to merely explain aspects. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
Hereinafter, a ferroelectric field effect transistor, a neural network apparatus, and an electronic apparatus will be described in detail with reference to the accompanying drawings. In the following drawings, the same reference numerals refer to the same components, and the size of each component in the drawings may be exaggerated for clarity and convenience of description. In addition, the embodiments described below are merely examples and various modifications are possible from these embodiments.
Hereinafter, the term “upper portion” or “on” may also include “to be present on a non-contact basis” as well as “to be in directly contact with”. For example, it will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is otherwise oriented (e.g., rotated 90 degrees or at other orientations), the spatially relative descriptors used herein are to be interpreted accordingly. The singular expression includes multiple expressions unless the context clearly implies otherwise. In addition, when a part “includes” a component, this means that it may further include other components, not excluding other components unless otherwise opposed.
The use of the term “the” and similar indicative terms may correspond to both singular and plural. If there is no explicit description of an order for steps that make up a method or vice versa, these steps can be done in an appropriate order and are not necessarily limited to the order described.
When the terms “about” or “substantially” are used in this specification in connection with a numerical value, it is intended that the associated numerical value includes a manufacturing tolerance (e.g., ±10%) around the stated numerical value. Further, regardless of whether numerical values are modified as “about” or “substantially,” it will be understood that these values should be construed as including a manufacturing or operational tolerance (e.g., ±10%) around the stated numerical values.
Further, the terms “unit”, “module” or the like mean a unit that processes at least one function or operation, which may be implemented in processing circuitry such as hardware, software and/or a combination of hardware and software. For example, the processing circuitry more specifically may include (and/or be included in), but is not limited to, a processor, Central Processing Unit (CPU), a controller, an arithmetic logic unit (ALU), an artificial intelligence accelerator, a digital signal processor, a microcomputer, a field programmable gate array (FPGA), a System-on-Chip (SoC), a programmable logic unit, a microprocessor, application-specific integrated circuit (ASIC), semiconductor elements in an integrated circuit, circuits enrolled as an Intellectual Property (IP), electrical components (such as at least one of transistors, resistors, capacitors, electrical components such as logic gates including at least one of AND gates, OR gates, NOR gates, NAND gates, NOT gates, XOR gates, and/or the like), etc.
The connection or connection members of lines between the components shown in the drawings example represent functional connection and/or physical or circuit connections and may be replaceable or represented as various additional functional connections, physical connections, and/or circuit connections in an actual device.
The use of all examples and/or exemplary terms is simply to describe technical ideas in detail, and the scope is not limited by the use of these examples and/or exemplary terms unless the scope is expressly limited by the claims.
In the convolution layer, a first feature map FM1 may correspond to an input feature map, and a second feature map FM2 may correspond to an output feature map. The feature map may mean a data set in which various characteristics of input data are expressed. The feature maps FM1 and FM2 may be high-dimensional matrices of two-dimensional or more and have activation parameters, respectively. When the feature maps FM1 and FM2 correspond to, for example, three-dimensional feature maps, the feature maps FM1 and FM2 have a width W (or referred to as a column), a height H (or a row), and a depth C. In these cases, the depth C may correspond to, e.g., the number of channels.
In the convolution layer, a convolution operation on the first feature map FM1 and a weight map WM may be performed, and as a result, the second feature map FM2 may be generated. The weight map WM may filter the first feature map FM1 and may be referred to as a weight filter or a weight kernel. In at least one example, the depth of the weight map WM (e.g., the number of channels) is the same as the depth C of the first feature map FM1. In at least one example, the weight map WM is shifted to transverse the first feature map FM1 as a sliding window. During each shift, each of the weights included in the weight map WM may be multiplied by and added to all feature values in a region overlapping the first feature map FM1. As the first feature map FM1 and the weight map WM are convoluted, one channel of the second feature map FM2 may be generated.
Although one weight map WM is illustrated in
As described above, the neural network 20 may include a DNN or n-layers neural network including two (or more) hidden layers. For example, as shown in
Each of the layers included in the neural network 20 may include a plurality of channels. The channel may correspond to a plurality of artificial nodes, known as a neuron, a processing element (PE), a unit, or similar terms. For example, as shown in
Channels included in each of the layers of the neural network 20 may be connected to each other to process data. For example, one channel may receive data from other channels to perform an arithmetic operation and/or may output the arithmetic operation result to other channels.
The input and output of the channel may be referred to as an input activation and an output activation, respectively. For example, the activation may be a parameter corresponding to an output of one channel and/or simultaneously inputs of channels included in the next layer. Meanwhile, in some example embodiments, each of the channels may determine its own activation based on activations and/or weights received from channels included in the previous layer. The weight is a parameter used to calculate output activation in each channel and may be a value allocated to a connection relationship between channels.
Each of the channels may be processed by a computational unit and/or processing element (such as a component and/or components included in processing circuitry) that receives an input and outputs an output activation, and the input-output of each of the channels may be mapped. For example, σ is an activation function, wjki is the weight from the k-th channel included in the (i−1)-th layer to the j-th channel included in the i-th layer, bji is the a bias of the j-th channel included in the i-th layer, and when aji is the activation of the j-th channel included in the i-th layer, the activation aji may be calculated by using the following Equation 1.
As shown in
As described above, in the neural network 20, numerous datasets may be exchanged between multiple interconnected channels and undergo an arithmetic operation process passing through the layer. In such an arithmetic operation process, a number of multiply-accumulate (MAC) operations are performed, and a number of memory access operations to load activations and weights, which are operands of MAC operations at an appropriate point in time, should be performed together.
Meanwhile, a typical digital computer uses a Von Neumann architecture that separates a computational unit and a memory and includes a common data bus for data transfer between two separated blocks. Therefore, in the process of performing the neural network 20 in which the data movement and arithmetic operation are continuously repeated, a large amount of time may be required for data transmission, and excessive power may be consumed. To overcome this problem, an in-memory computing neural network apparatus has been proposed as an architecture that integrates memory and operation units for performing MAC operations into one.
The plurality of word lines WL and the plurality of bit lines BL may be arranged to cross each other. A plurality of synaptic elements 550 may be arranged at the intersections of the plurality of word lines WL and the plurality of bit lines BL. Accordingly, the plurality of synaptic elements 550 may be arranged in the form of a two-dimensional array.
During the learning operation of the neural network apparatus 500, one row of synaptic elements 550 may be individually selected through an individual word line WL, and a program pulse or pulses may be applied to each row of the synaptic elements 550 through the plurality of bit lines BL in the selected row. In addition, learning data signals may be applied to the selected synaptic element 550 through the plurality of input lines IL. Through this process, weights may be stored in each selected synaptic element 550. During the learning operation of the neural network apparatus 500, the word line driver 510 may be configured to sequentially provide turn-on signals to the plurality of word lines WL, the bit line driver 520 may be configured to provide weight signals to the plurality of bit lines BL, and the input circuit 530 may be configured to provide the learning data signals to the plurality of input lines IL.
During the inference operation of the neural network apparatus 500, all of the synaptic elements 550 may be selected through the word lines WL, and the read voltage may be applied to the synaptic elements 550 through the bit lines BL. In addition, in some example embodiments, input data signals to be used for inference may be applied to the synaptic elements 550 through the input lines IL. Then, currents from the synaptic elements 550 connected in parallel to the output line OL are summed and the summed current flows through each output line OL. The current flowing through each output line OL may vary according to the weights stored in the plurality of synaptic elements 550. The output circuit 540 may convert the current flowing through each output line OL into a digital signal. During the inference operation of the neural network apparatus 500, the word line driver 510 may be configured to provide turn-on signals to the plurality of word lines WL, the bit line driver 520 may be configured to apply read voltages to the plurality of bit lines BL, and the input circuit 530 may be configured to provide input data signals to the plurality of input lines IL.
Each of the synaptic elements 550 may be implemented as a nonvolatile memory. For example, the synaptic elements 550 may be implemented using a ferroelectric memory that can be operated with low voltage and low power, has a relatively long operating repetition life, and has a relatively fast operating speed. Meanwhile, in order to ensure accuracy and precision of the neural network operation performed by the neural network apparatus 500, it is advantageous for the synaptic elements 550 to have linear state change characteristics along with multi-level characteristics. In general, since ferroelectrics may undergo rapid polarization switching near the coercive voltage, thereby causing non-linear state changes, a curved pillar shape ferroelectric capacitor having electrode structures of various curvatures is proposed to induce linear state change characteristics of a ferroelectric memory.
For example,
First, referring to
The inner electrode 101 and the outer electrode 103 may include at least one metallic material selected from, for example, TiN, TiAlN, TaN, Co, WN, NbN, W, Mo, Pt, and/or the like. The ferroelectric film 102 may include at least one ferroelectric material selected from among an oxide of Si, an oxide of Al, an oxide of Hf, and/or an oxide of Zr, and may be doped with at least one of Si, Al, Y, La, Gd, Mg, Ca, Sr Ba, Ti, Zr, Hf, N, MgZnO, AlScN, BaTiO3, Pb(Zr, Ti)O3, SrBiTaO7, polyvinylidene fluoride (PVDF), and/or the like. The ferroelectric material is a material that maintains a spontaneous polarization by aligning an internal electric dipole moment even if an electric field is not applied thereto from the outside. Characteristics, such as conductance, of the ferroelectric capacitor 100, may change according to the polarization direction of the ferroelectric material in the ferroelectric film 102. The ferroelectric film 102 may have an inner radius and an outer radius. The difference between the inner radius and the outer radius may be referred to as a thickness t of the ferroelectric film 102. The ratio t/r of the thickness t of the ferroelectric film 102 to the radius r of the inner electrode 101 may be greater than 0 and less than or equal 2.
Although
In
Referring to
Referring to
Referring to
When a voltage of V is applied to a general flat capacitor with a dielectric film thickness t, an electric field between two flat electrodes is represented by E=V/t. In contrast, when a voltage is applied between the inner electrode and the outer electrode of the capacitor with a cylindrical shape, the electric field increases from the outer electrode to the inner electrode. For example, when a voltage of V is applied to a cylindrical capacitor having a dielectric film of thickness t and an internal electrode of radius r, the electric field inside the dielectric film may be expressed as Equation 2 below.
Here, x is a radial distance of a point inside the dielectric film, and is 0≤x≤t. For example, assuming a cylindrical capacitor having a radius of the inner electrode of 5 nm and a thickness of the dielectric film of 10 nm, the electric field strength on the inner electrode surface is about twice the electric field strength on the outer electrode surface, and an electric field on the surface of the internal electrode is generated about twice as large as an electric field on a flat capacitor. In addition, as the radius of the inner electrode decreases, the strength of the electric field on the surface of the inner electrode may increase further.
Meanwhile, when the voltage applied to the ferroelectric capacitor is gradually increased, the electric field strength inside the ferroelectric film continues to increase until the electric field reaches the coercive electric field, and when the strength of the electric field reaches the coercive electric field, polarization switching occurs in the ferroelectric film. In the case of ferroelectric capacitors having a cylindrical shape, even if the thickness of the ferroelectric film is the same, the electric field strength inside the ferroelectric film varies according to the radius of the inner electrode. For example, the smaller the radius of the inner electrode, the higher the strength of the electric field on the surface of the inner electrode, so even a smaller applied voltage to the ferroelectric capacitor may induce polarization switching.
When two or more cylindrical ferroelectric capacitors having different radii of inner electrodes are connected, e.g., in parallel, two or more polarization switching may occur in two or more different voltage regions. For example,
Therefore, when the ferroelectric capacitors have curved pillar shapes having various curvature electrode structures, as shown in
In addition, the principle of the ferroelectric capacitor described above may also be applied to the ferroelectric field effect transistor. For example,
In some embodiments, the source 210 and the drain 220 may include a semiconductor material doped with a first conductive type, and the channel 230 may be doped with a second conductive type electrically opposite to the first conductive type. For example, the channel 230 may include a p-type semiconductor, and the source 210 and the drain 220 may include an n-type semiconductor. Otherwise, the channel 230 may include an n-type semiconductor, and the source 210 and the drain 220 may include a p-type semiconductor. In some embodiments, the channel 230 may be doped at a relatively low concentration of about 1014 to 1018/cm3, and the source 210 and the drain 220 may be doped at a relatively high concentration of about 1019 to 1021/cm3 for low resistance. The source 210, the drain 220, and the channel 230 may include at least one semiconductor material from among, for example, a Group IV semiconductor such as silicon (Si), germanium (Ge), SiGe, etc., a Group III-V compound semiconductor such as GaN, SiC, GaAs, InGaAs, GaP, etc., an oxide semiconductor such as ZnO, SnO, GaO, InO, InGaZnO, ZnSnO, etc., and/or a two-dimensional semiconductor such as MoS2, SnS2, WTe2, etc.
The gate structure of the ferroelectric field effect transistor 200 shown in
For example, the ferroelectric film 241 may include at least one ferroelectric material selected from among an oxide of Si, an oxide of Al, an oxide of Hf, or an oxide of Zr, and may be doped with at least one of Si, Al, Y, La, Gd, Mg, Ca, Sr Ba, Ti, Zr, Hf, N, MgZnO, AlScN, BaTiO3, Pb(Zr, Ti)O3, SrBiTaO7, and/or polyvinylidene fluoride (PVDF). In addition, the gate electrode 240 may include at least one metallic material, for example, TiN, TiAlN, TaN, Co, WN, NbN, W, Mo, Pt, and/or the like.
In the ferroelectric field effect transistor 200 according to at least one embodiment, conductance between the source 210 and the drain 220 may change according to the polarization direction of the ferroelectric film 241. In addition, a threshold voltage of the ferroelectric field effect transistor 200 may change according to the polarization direction of the ferroelectric film 241. Meanwhile, the channel 230 may have a curvature that continuously changes in the azimuth direction on the XZ plane. Therefore, when a gate voltage is applied to the gate electrode 240, the strength of the electric field on the surface of the channel 230 may vary according to the azimuth direction on the XZ plane. As a result, the ferroelectric field effect transistor 200 may have the same effect as a parallel connection of a myriad of channels in which the threshold voltage continuously changes.
The gate structure of the ferroelectric field effect transistor shown in
The ferroelectric field effect transistor having the gate structure illustrated in
The first to fourth channels 330a, 330b, 330c, and 330d may be sequentially arranged between the source 310 and the drain 320 in the Y-direction. In addition, the first to fourth channels 330a, 330b, 330c, and 330d may be arranged on a common central axis. The gate structure of the ferroelectric field effect transistor shown in
The channel 430 may have a tapered structure having radii which gradually change in the Y-direction. For example, the channel 430 may have a conical shape or a truncated cone shape. The gate structure of the ferroelectric field effect transistor shown in
For example, in the case of the first channel having the first radius, polarization switching of the ferroelectric film occurs in the gate voltage range corresponding to the section a, in the case of the second channel having the second radius different from the first radius, polarization switching of the ferroelectric film occurs in the gate voltage range corresponding to the section b, and in the case of the third channel having the third radius different from the first radius and the second radius, polarization switching of the ferroelectric film occurs in the gate voltage range corresponding to the section c. Then, conductance of the first channel may change significantly in the section a, conductance of the second channel may change significantly in the section b, and conductance of the third channel may change significantly in the section c.
The change in conductance of the first channel corresponding to the section a of the voltage-current characteristic curve 50 may be represented by a first potentialization and depression (PD) curve 51, the change in conductance of the second channel corresponding to the section b of the voltage-current characteristic curve 50 may be represented by a second PD curve 52, and the change in conductance of the third channel corresponding to the section c of the voltage-current characteristic curve 50 may be represented by a third PD curve 53. Each of the first PD curve 51, the second PD curve 52, and the third PD curve 53 has nonlinear characteristics with respect to the gate voltage.
Meanwhile, when comparing a composite PD curve 54 as a result of normalizing and combining the first PD curve 52 and the third PD curve 53 with the second PD curve 52, it may be seen that the synthetic PD curve 54 is more linear with respect to a gate voltage change. The composite PD curve 54 may correspond to the composite conductance of the first channel and the third channel. Therefore, the first channel and the third channel having different gate voltage sections for causing polarization switching are connected in parallel, and thus more linear conductance may be obtained with respect to the change in the gate voltage, and the composite conductance of the first to third channels connected in parallel may have more linear characteristics. By applying this principle to a synaptic element using ferroelectric memory, a relatively linear change in synapse connection strength may be obtained.
During the learning operation of the neural network apparatus 500, the access transistor 551 may be individually turned on through an individual word line WL, and a program pulse may be applied to the gate of the ferroelectric field effect transistor 552 through the bit line BL. Signals of the learning data may be applied through the input lines IL. Through this process, weights may be stored in the ferroelectric field effect transistor 552, respectively. To this end, the word line driver 510 may be configured to sequentially apply turn-on signals to the plurality of word lines WL during a learning operation of the neural network apparatus 500. The bit line driver 520 may be configured to apply weight signals to the plurality of bit lines BL during a learning operation of the neural network apparatus 500.
In the inference operation of the neural network apparatus 500, all of the access transistors 551 may be turned on through the word lines WL, and the read voltage may be applied through the bit lines BL. Then, currents from the synaptic elements 550 connected in parallel to the output line OL are summed and the summed current flows through each output line OL. The output circuit 540 may convert the current flowing through each output line OL into a digital signal. To this end, the word line driver 510 may be configured to apply turn-on signals to the entire plurality of word lines WL during an inference operation of the neural network apparatus 500. The bit line driver 520 may be configured to apply the read voltage to the plurality of bit lines BL during am inference operation of the neural network apparatus 500.
According to at least one embodiment, since the ferroelectric field effect transistor 552 has a linear state change characteristic, the weight may be linearly updated in the synaptic element 550 of the neural network apparatus 500. For example, the weight (or synapse connection strength) may linearly change in proportion to a program pulse provided through the bit lines BL.
Alternatively, in the synaptic element 550a, the ferroelectric capacitor 553 may be arranged as the gate of the field effect transistor 554. In this case, the ferroelectric capacitor 553 and the field effect transistor 554 may form one metal-ferroelectric-metal-insulator (MFMIS) field effect transistor together. In this case, the ferroelectric capacitor 553 and the field effect transistor 554 may serve as a memory together.
The plurality of word lines WL and the plurality of bit lines BL may be arranged to cross each other. The plurality of synaptic elements 640 may be arranged at intersections of the plurality of word lines WL and the plurality of bit lines BL. Accordingly, the plurality of synaptic elements 640 may be arranged in the form of a two-dimensional array. The plurality of synapse elements 640 include any one of the ferroelectric capacitors according to the embodiments described in
During the learning operation of the neural network apparatus 600, the synaptic elements 640 in one row may be individually selected through the individual word line WL, and a signal of the learning data may be applied to the synaptic elements 640 arranged in the selected row through the word line WL. In addition, in the selected row, a weight signal may be applied to each synaptic element 550 for each column through a bit line BL. Through this process, weights may be stored in the plurality of synaptic element 640. During the learning operation of the neural network apparatus 600, the word line driver 610 may be configured to sequentially apply learning data signals to the plurality of word lines WL, and the bit line driver 620 may be configured to apply weight signals to the plurality of bit lines BL.
During the inference operation of the neural network apparatus 600, input data signals to be used for inference may be applied to the synaptic elements 640 through the entire word lines WL. Then, currents from the synaptic elements 640 connected in parallel to the bit line BL are summed and the summed current flows through each bit line BL. The current flowing through each bit line BL may vary according to the weights stored in the plurality of synaptic elements 640. The output circuit 630 may convert the current flowing through each bit line BL into a digital signal.
The electronic apparatus 700 may include a processor 710, a random access memory (RAM) 720, a neural network apparatus 730, a memory 740, a sensor module 750, and a communication module 760. The electronic apparatus 700 may further include an input/output module, a security module, a power control device, and the like. Some of the hardware components of the electronic apparatus 700 may be mounted on at least one semiconductor chip.
The processor 710 controls the overall operation of the electronic apparatus 700. The processor 710 may include one processor core and/or a plurality of processor cores (e.g., Multi-Core). The processor 710 may process and/or execute programs and/or data stored in the memory 740. In some embodiments, the processor 710 may control the function of the neural network apparatus 730 by executing programs stored in the memory 740. The processor 710 may be implemented as, e.g., a central processing unit (CPU), a graphics processing unit (GPU), an application processor (AP), and/or the like.
The RAM 720 may temporarily store programs, data, or instructions. For example, programs and/or data stored in the memory 740 may be temporarily stored in the RAM 720 according to the control or boot code of the processor 710. The RAM 720 may be implemented as a memory, e.g., such as dynamic RAM (DRAM), static RAM (SRAM), and/or the like.
The neural network apparatus 730 may perform an operation of the neural network based on the received input data and generate an information signal based on the execution result. The neural network may include, but is not limited to, CNN, RNN, FNN, long short-term memory (LSTM), stacked neural network (SNN), state-space dynamic neural network (SSDNN), deep belief networks (DBN), restricted Boltzmann machine (RBM), and/or the like. The neural network apparatus 730 may be a hardware accelerator itself dedicated to a neural network or an apparatus including the same. The neural network apparatus 730 may perform a read and/or write operation as well as an operation of the neural network.
The neural network apparatus 730 may correspond to the neural network apparatuses 500 and 600 according to the embodiments shown in
The information signal may include one of various types of recognition signals such as a voice recognition signal, an object recognition signal, an image recognition signal, a biometric information recognition signal, and the like. For example, in at least one embodiment, the neural network apparatus 730 may receive frame data included in the video stream as input data and generate, from frame data, a recognition signal for an object included in an image represented by the frame data. However, the neural network apparatus is not limited thereto, and the neural network apparatus 730 may receive various types of input data and generate a recognition signal according to the input data according to the type or function of the device on which the electronic apparatus 700 is mounted.
The neural network apparatus 730 may perform, for example, machine learning model such as linear regression, logistic regression, statistical clustering, Bayesian classification, decision trees, principal component analysis, and/or expert system, and/or machine learning model of ensemble techniques, etc., such as random forest. The machine learning model may be used to provide various services such as, for example, image classification service, user authentication service based on biometric information or biometric data, advanced driver assistance system (ADAS), voice assistant service, automatic speech recognition (ASR) service, and/or the like.
The memory 740 is a storage place for storing data and may store an operating system (OS), various programs, and various pieces of data. In at least one embodiment, the memory 740 may store intermediate results generated during the operation of the neural network apparatus 730.
The memory 740 may be a DRAM, but is not limited thereto. The memory 740 may include at least one of a volatile memory and/or a nonvolatile memory. The nonvolatile memory may include at least one of read-only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable and programmable ROM (EEPROM), flash memory, phase-change ROM (PROM), magnetic ROM (MROM), resistive ROM (RROM), ferroelectric ROM (FROM), and/or the like. The volatile memory may include at least one of dynamic RAM (DRAM), static RAM (SRAM), synchronous DRAM (SDRAM), phase-change RAM (PRAM), magnetic RAM (MRAM), resistive RAM (RRAM), ferroelectric RAM (FeRAM), and/or the like. In an at least one embodiment, the memory 740 may include at least one of a hard disk drive (HDD), a solid state drive (SSD), a compact flash (CF), a secure digital (SD), a micro secure digital (Micro-SD), a mini secure digital (Mini-SD), and/or a memory stick.
The sensor module 750 may collect information around a device on which the electronic apparatus 700 is mounted. The sensor module 750 may sense and/or receive a signal (e.g., an image signal, a voice signal, a magnetic signal, a bio signal, a touch signal, etc.) from the outside of the electronic apparatus 700 and convert the sensed and/or received signal into data. For example, the sensor module 750 may include at least one of various types of sensing devices such as a sensing device, for example, a microphone, an imaging device, an image sensor, a light detection and ranging (LIDAR) sensor, an ultrasonic sensor, an infrared sensor, a biosensor, and/or a touch sensor.
In some example embodiments, the sensor module 750 may provide the converted data to the neural network device 730 as input data. For example, the sensor module 750 may include an image sensor, generate a video stream by photographing an external environment of the electronic device 700, and sequentially provide the continuous data frame of the video stream to the neural network device 730 as input data. However, the example embodiments are not limited thereto, and the sensor module 750 may provide various types of data to the neural network device 730.
The communication module 760 may include various wired or wireless interfaces capable of communicating with an external device. For example, the communication module 760 may be configured to transmit Tx and receive Rx signals and may include a wired local area network (LAN), a wireless local area network (WLAN) such as a wireless fidelity (Wi-Fi), a wireless personal area network (WPAN) such as Bluetooth, a wireless universal serial bus (USB), Zigbee, near field communication (NFC), radio-frequency identification (RFID), power line communication (PLC), and a communication interface capable of connecting to a mobile cellular network, such as 3rd generation (3G), 4th generation (4G), 5th generation (5G), long term evolution (LTE), and/or the like.
The ferroelectric field effect transistors, neural network devices, and electronic devices described above are described with reference to the embodiments shown in the drawings, but it should be understood that embodiments described herein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments. While one or more embodiments have been described with reference to the figures, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope as defined by the following claims.
Number | Date | Country | Kind |
---|---|---|---|
10-2022-0021736 | Feb 2022 | KR | national |