The present disclosure relates to computers, and more particularly, to neural computers including an image sensor capable of controlling a photocurrent.
An artificial neural network algorithm that imitates the structure of a human optic nerve may be used in deep learning. This neural network algorithm may be used in neural network computing. So far, various neural network algorithms have been introduced, and representative products are, for example, a convolution neural network (CNN), a recurrent neural network (RNN), etc. In the case of the CNN, a feature map is extracted from multiple convolutional layers, and only an important part of the feature map is taken by reducing dimensions through subsampling. CNN is an essential technology used in most computer vision fields, such as image classification, semantic segmentation, optical flow, etc., and compared to multi-layer perceptron (MLP) which is an early algorithm, it has less computational load and higher performance.
Provided are neural computers capable of simplifying an operation of processing an input image.
Provided are neural computers capable of reducing a processing capacity of input image and time.
Provided are neural computers capable of improving the degree of integration and constructing a large area device.
Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented embodiments of the disclosure.
According to an embodiment, a neural computer includes a preprocessor configured to receive an image and generate a feature map of the received image, a flattening unit configured to transform the feature map generated by the preprocessor into tabular data to provide data output, and an image classifier configured to classify images received through the preprocessor by using the data output by the flattening unit as an input value. The preprocessor includes an optical signal processor configured to receive the image and generate the feature map.
In one example, the optical signal processor may include an image sensor configured to receive the image and generate the feature map, and a controller connected to the image sensor and configured to control a voltage for generating the feature map.
In one example, the image sensor may include a plurality of pixels, wherein each of the plurality of pixels may include a substrate, an insulating layer on the substrate, first and second gate electrodes separated from each other on the insulating layer, and source and drain electrodes separated from each other on the substrate with the insulating layer therebetween. A voltage of the first and second gate electrodes may be controlled by the controller.
In one example, the substrate may include Si, Ge, or a compound including Si and Ge.
In one example, the substrate may include graphene, a transition metal dichalcogenide (TMD), or black phosphorus (BP).
In one example, the TMD may include any one of Mo, W, Ti, Ni, Ta, Hf and Zr and the TMD may include one of S, Se and Te.
In one example, the insulating layer may include an oxide of any one of Si, Al, Hf, Zr, Ti, Ta, and W.
In an example, the insulating layer may include a perovskite material having the structure of ABO3. “A,” in ABO3, may be any one of Li, Na, K, Ca, Rb, Sr, Y, Ag, Cs, Ba, La, Ce, Pr, Nd, Sm, Gd, Dy, Ho, Yb and Bi. “B,” in ABO3, may be any one of Mg, Al, Sc, Ti, V, Cr, Mn, Fe, Co, Ni, Cu, Zn, Ga, Ge, Y, Zr, Nb, Mo, Ru, Pd, Cd, In, Sn, Sb, La, Ce, Pr, Nd, Sm, Gd, Ta, W, Ir, Pb, and Bi.
The above and other aspects, features, and advantages of certain embodiments of the disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout. In this regard, the present embodiments may have different forms and should not be construed as being limited to the descriptions set forth herein. Accordingly, the embodiments are merely described below, by referring to the figures, to explain aspects. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
Hereinafter, a neural computer including an image sensor capable of controlling a photocurrent according to an embodiment will be described in detail with reference to the accompanying drawings. In the drawings, thicknesses of layers and regions may be exaggerated for clarification of the specification. The embodiments of the inventive concept are capable of various modifications and may be embodied in many different forms. When an element or layer is referred to as being “on” or “above” another element or layer, the element or layer may be directly on another element or layer or intervening elements or layers. In the drawings, like reference numerals refer to like elements throughout.
Referring to
The process of generating a feature map may include a process of generating a first feature map with respect to the received image and a process of generating a second feature map with respect to the first feature map. The second feature map may be a compression of the first feature map. In one example, the process of generating the feature map may include only the process of generating the first feature map. In one example, the process of generating the feature map may further include processes of generating a third feature map, a fourth feature map and so on in addition to the process of generating the second feature map. In an example, the process of generating the first feature map may include a process of applying a first convolution filter to the received image. The first convolution filter may be generated in consideration of the characteristics (type) of the received image. The first feature map may be obtained by applying the first convolution filter to the received image data. In an example, the process of generating the second feature map may include the process of applying a second convolution filter to the first feature map, and what is obtained by applying the second convolution filter to the first feature map may be the second feature map. In one example, the process of generating the second feature map may include a process of first pooling the first feature map, and what is obtained as a result of the first pooling may be the second feature map. A feature map finally generated by the preprocessor 110 is formed through the process of applying the convolutional filter and/or pooling.
The process of generating the convolutional filter and the pooling process may be performed by using the optical signal processor. In one example, the optical signal processor may include an image sensor.
Referring to
The image sensor 410 may include a plurality of pixels 410A aligned to accommodate a given image. The controller 420 may be electrically connected to each of the plurality of pixels 410A. In an example, the connection between the controller 420 and the pixel 410A may denote the connection between the controller 420 and a device, for example, an optical device, that is disposed in the pixel 410A to perform a pixel role. The connection between the controller 420 and the image sensor 410 is designed so that the controller 420 selects an arbitrary pixel from among the plurality of pixels 410A of the image sensor 410 and controls a voltage applied to the selected pixel. In one example, an independent controller may be connected to each of the plurality of pixels 410A, and the independent controller may be controlled by a main controller.
Referring to
In the case of a conventional neural computer that captures an image, stores the captured image and generates a feature map by applying a convolutional filter to the stored image, the processes of storing the captured image and generating the feature map with respect to the stored image are performed in another unit outside the image sensor. On the other hand, in the case of a neural computer according to an embodiment, as described above, the process from capturing an image to generating a feature map is performed in an image sensor, and thus, the operation of the neural computer may be simplified, and a separate unit for storing a captured image and generating a feature map with respect to the stored image is not required, and thus, the volume of the computer may be reduced and the degree of integration may be increased.
The optical signal processor 400 including the image sensor 410 in which the optical device 500 is provided in each pixel 410A may be applied to other devices, for example, a CMOS image sensor, machine vision, neuromorphic, analog computing, etc.
Referring to
Referring to
From the results of
The disclosed neural computer includes an image sensor capable of controlling a voltage applied to a gate electrode of an optical device provided in a pixel. In such an image sensor, a process from receiving an image (e.g., capturing an image) to generating a feature map may be performed. Therefore, in the case of using the neural computer according to an embodiment, the process from receiving an image to classifying (recognizing) an image may be simplified, the computer volume may be reduced, and the degree of integration may be increased.
One or more of the elements disclosed above may include or be implemented in processing circuitry such as hardware including logic circuits; a hardware/software combination such as a processor executing software; or a combination thereof. For example, the processing circuitry more specifically may include, but is not limited to, a central processing unit (CPU), an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable gate array (FPGA), a System-on-Chip (SoC), a programmable logic unit, a microprocessor, application-specific integrated circuit (ASIC), etc.
It should be understood that embodiments described herein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments. While one or more embodiments have been described with reference to the figures, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope as defined by the following claims.
Number | Date | Country | Kind |
---|---|---|---|
10-2021-0042228 | Mar 2021 | KR | national |
This application claims the benefit of U.S. Provisional Application No. 63/112,720, filed on Nov. 12, 2020, in the US Patent Office and Korean Patent Application No. 10-2021-0042228, filed on Mar. 31, 2021, in the Korean Intellectual Property Office, the disclosures of which are incorporated herein in their entireties by reference.
Number | Name | Date | Kind |
---|---|---|---|
20060150149 | Chandhoke et al. | Jul 2006 | A1 |
20070139541 | Fein et al. | Jun 2007 | A1 |
20160358038 | Jaderberg | Dec 2016 | A1 |
20180039882 | Ikeda | Feb 2018 | A1 |
20180315870 | Snaith et al. | Nov 2018 | A1 |
20190034748 | Matsumoto et al. | Jan 2019 | A1 |
20190277957 | Chandrasekhar | Sep 2019 | A1 |
20200090028 | Huang | Mar 2020 | A1 |
20200320677 | Ion et al. | Oct 2020 | A1 |
20210036024 | Kim | Feb 2021 | A1 |
20210151678 | Lee et al. | May 2021 | A1 |
20210152732 | Eki | May 2021 | A1 |
20210224643 | Li | Jul 2021 | A1 |
20210256311 | Olarig | Aug 2021 | A1 |
20220284552 | Yang | Sep 2022 | A1 |
20220292332 | Koumura | Sep 2022 | A1 |
20230109524 | Hirose | Apr 2023 | A1 |
20230298145 | Miscuglio | Sep 2023 | A1 |
20230377094 | Horii | Nov 2023 | A1 |
Number | Date | Country |
---|---|---|
6725733 | Jul 2020 | JP |
20150056851 | May 2015 | KR |
WO-2013110803 | Aug 2013 | WO |
Entry |
---|
Mirko Hansen et al., ‘Double-Barrier Memristive Devices for Unsupervised Learning and Pattern Recognition’ Frontiers in Neuroscience, vol. 11, Article 91, Feb. 2017. |
Seunghwan Seo et al., ‘Artificial optic-neural synapse for colored and color-mixed pattern recognition’ Nature Communications, 9:5106, 2018. |
Houk Jang et al., ‘An Atomically Thin Optoelectronic Machine Vision Processor’ Advanced Materials, vol. 32, 2020. |
‘ML Practicum: Image Classification’ Machine Learning, Google Developers, Jul. 2022 https://developers.google.com/machine-learning/practica/image-classification/convolutional-neural-networks. |
Number | Date | Country | |
---|---|---|---|
20220147799 A1 | May 2022 | US |
Number | Date | Country | |
---|---|---|---|
63112720 | Nov 2020 | US |