The specification relates generally to machine vision systems, and more particularly to a polarimetric sensor array for machine vision systems.
Learning from human vision have informed visible light techniques. Learnings from the vision of animals, such as bees, may inform polarimetry techniques. Combining these techniques may result in what may be termed an interspecies and/or chimeric system.
According to an aspect of the present specification an example device includes: a polarimetric sensor array comprising an optically controlled polarimetry memtransistor, the polarimetric sensor array configured to: detect incoming light representing a field of view of the polarimetric sensor array; and generate a polarization signal representing a polarization of the incoming light; and an artificial intelligence subsystem interconnected with the polarimetric sensor array, the artificial intelligence subsystem configured to: process the polarization signal for a machine vision function on the field of view.
According to another aspect of the present specification, an example method includes: detecting, at a polarimetric sensor array comprising an optically controlled polarimetry memtransistor, incoming light representing a field of view of the polarimetric sensor array; generating a polarization signal representing a polarization of the incoming light; and processing the polarization signal for a machine vision function on the field of view.
Implementations are described with reference to the following figures, in which:
The patent or application file contains at least one drawing executed in color.
Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
Current humanoid machine vision techniques mimic human systems and lacks polarimetric functionalities that convey the information of navigation and authentic images. Interspecies-chimera vision reserving multiple hosts' capacities will lead to improved machine vision, but implementation of visual functions of multiple species (human and non-human) in one optoelectronic device is still elusive.
Traditional visual systems are bulky, energy-intensive, and lengthy, especially for cognitive tasks owing to discrete computation hierarchy. The sensory, memory, and computing units are separated from each other. The location and format (analog/digital) of image data need to be changed frequently, giving rise to the penalty of energy consumption and time delay. To solve this issue, mimicking human visual systems is a promising strategy. In human visions, photoreceptors and neurons in the human retina detect and pre-process images that are later sent to the visual cortex for cognitive signal processing. For current machine visions, intelligent photodetectors can simultaneously sense and pre-process light stimuli like a human retina. Received images are then directly transferred into artificial neural networks (ANNs) for complex visual processing. ANNs fundamentally imitate the fundamental principles of human brains relying on the activities of synapses and neurons, which can realize huge parallel computing and high-energy efficiency. However, human eyes only provide very limited photodetection abilities regarding the light wavelength (380 nm to 700 nm), intensity (comfortable intensity 200 lux to 750 lux), and vector (not sensitive to linear and circular polarity). This restrains the application of humanoid machine vision. On the other hand, interspecies-chimera machine vision integrates the cognitive function of humans and the special visual function of other species. This can provide functions beyond traditional artificial intelligence (AI) by leveraging machines to solve problems like a human while processing super-human capabilities enabled by other appealing functions.
Humanoid machine vision has been reported to reduce background noise, scotopic/photopic adaptation, broadband sensing, convolution processing, etc. In the human vision system, the image information is received and pre-processed by the retina and transferred into the visual cortex region for cognitive processing. Human vision, although, is capable of cognitive tasks. It is not sensitive to the polarization of light which includes critical information. It is rudimentary to integrate polarimetric functions into humanoid machine vision. Honeybees are known as excellent navigators with the help of ommateum measuring linearly polarized lights. The skylight pattern is determined by the position of the sun, described by the Rayleigh sky model. For instance, real-time navigation can be achieved via monitoring the sun-position-related celestial polarization cues. Indeed, the polarimetric capability makes honeybees a master for navigating from the honeycomb to a flower patch. But honeybees are not intelligent species like humans. Therefore, a rational design philosophy is to integrate the polarimetry from honeybees and the intelligence from humans, creating an interspecies-chimera machine vision. Particularly, the reflected light from shiny surfaces, including cars' windows, buildings, water on the road, and the road itself, forms glare spots. This compromises the processing accuracy in traditional machine visions due to glare-induced distortion. The formation of glare spots is because the reflected light is polarized light parallel to the surface. The polarimetric function can suppress the light intensity with a certain polarization (glare spots) and keep high fidelity in authentic images. This interspecies-chimera machine vision, on the one hand, can detect the polarization pattern in the sky for real-time navigation and on the other hand, the system can realize anti-glare pattern recognition.
As presently described, an optically-controlled polarimetry memtransistor (OCPM) based on a van der Waals heterostructure (ReS2/GeSe2). The device provides polarization sensitivity, nonvolatility, and positive/negative photoconductance simultaneously. The polarimetric measurement can identify celestial polarizations for real-time navigation like a honeybee. Meanwhile, cognitive tasks can be completed like a human by sensing, memory, and synaptic functions. Particularly, the anti-glare recognition with polarimetry saves an order of magnitude energy compared to the traditional humanoid counterpart. This technique promotes the concept of interspecies-chimera visual systems that may be leveraged in machine vision functions of autonomous vehicles, medical diagnoses, intelligent robotics, etc.
The system 100 may include a single device integrating the sensor array 104 and the artificial intelligence subsystem 108, or the sensor array 104 and the artificial intelligence subsystem 108 may be implemented in distinct devices.
Generally, the artificial intelligence subsystem is configured for navigation using the natural light polarization pattern of the sky. The polarimetric sensor array 104 senses this pattern and provides corresponding polarization signals (e.g., in the form of a polarization-dependent current measurement) to the artificial intelligence subsystem 108. The artificial intelligence subsystem 108 may be configured to process the signals to output a direction, heading, location, speed, or similar navigational information.
The artificial intelligence subsystem 108 may additionally be configured for glare reduction. In particular, the artificial intelligence subsystem 108 may process a visible light image to remove or reduce glare as informed by the polarimetric sensor array 104. This may be done prior to or simultaneously with feature recognition or other visible light techniques performed on signals captured by the image sensor.
In addition to navigation and glare reduction, other examples of using light polarization to provide useful data or to augment data captured by other techniques are contemplated.
The artificial intelligence subsystem 108 may be implemented on a computing device, such as an example computing device 120 depicted in
The computing device 120 may include a includes a processor 124 and a memory 128.
The processor 124 may include a central processing unit (CPU), a microcontroller, a microprocessor, a processing core, a field-programmable gate array (FPGA), or similar. The processor 124 may include multiple cooperating processors. The processor 124 may cooperate with the memory 128 to realize the functionality described herein.
The memory 128 may include a combination of volatile (e.g., Random Access Memory or RAM) and non-volatile memory (e.g., read-only memory or ROM, Electrically Erasable Programmable Read Only Memory or EEPROM, flash memory). All or some of the memory 128 may be integrated with the processor 124. The memory stores applications, each including a plurality of computer-readable instructions executable by the processor 124. The execution of the instructions by the processor 124 configures the device 120 to perform the actions discussed herein. In particular, the applications stored in the memory 128 include an application 132. When executed by the processor 124, the application 132 configures the processor 124 to perform various functions discussed below in greater detail and related to the machine vision operation of the device 120. For example, the application 132 may implement some or all of the artificial intelligence subsystem 108. The application 132 may also be implemented as a suite of distinct applications. Further, some or all of the functionality of the application 136 may be implemented as dedicated hardware components, such as one or more FPGAs or application-specific integrated circuits (ASICs).
In some examples, the device 120 may further include a communications interface (not shown) including suitable components (e.g., transmitters, receivers, antennae, ports, etc.) allowing the device 120 to communicate over wired or wireless links. The device 120 may further include one or more input and/or output devices (not shown), such as displays, buttons, microphones, speakers, and the like.
In various examples, the polarimetric sensor array 104, artificial intelligence subsystem 108, and image sensor 110 (if used) may be provided to the same electronic device or on the same integrated circuit (IC), for example as part of the device 120. In other examples, two or more of these components 104, 108, and 110 may be distant from each other and may be connect by wired or wireless data pathways.
The polarimetric sensor array 104 includes individual polarimetric sensor elements 106 that may be arranged in a grid, similar to a visible-light image sensor, such as a charge-coupled device (CCD) or complementary metal-oxide-semiconductor (CMOS) device. Each polarimetric sensor element may be considered a pixel that corresponds to one or more image sensor elements of an image sensor, if used, for example, a polarimetric sensor element may correspond to one image sensor elements, four image sensor elements, sixteen image sensor elements, or similar group of image elements or image pixels.
The polarimetric sensor array 104 or the individual pixel elements 106 thereof provide an optically controlled polarimetry memtransistor (OCPM). In particular, the OCPM provides polarization sensitivity, nonvolatility, and positive/negative photoconductance. That is, the polarimetric sensor array 104 is photosensitive, and therefore configured to detect incoming light from a field of view 112 of the polarimetric sensor array 104. More particularly, the polarimetric sensor array 104 is sensitive to the polarity or polarization of the incoming light. The polarimetric sensor array 104 may be configured to generate a polarization signal representing the polarization of the incoming light. For example, the polarimetric sensor array 104 may generate different currents according to the polarization of the incoming light. Other sensitivities are also contemplated.
The polarimetric sensor array 104 may further be supported on a rotatable platform 116, configured to rotate the array 104 for detecting the light and to generate the polarization signals according to the reception of the light at different angles relative to the atomic structure of each of the polarimetric sensor elements 106, as further described herein. In other examples, different detection angles of the polarimetric sensor array 104 may be achieved via orientation of the polarimetric sensor elements 106 within the array 104. That is, some sensor elements 106, or subsets thereof may be oriented in different directions in order to detect light at the different angles relative to the atomic structures.
The OCPM can further realize optically-programmed non-volatile states. In particular, the OCPM provides both positive photoconductivity (PPC) and negative photoconductivity, which mimics antagonistic shunting and memory of bipolar cells, demonstrating the sensing function of the retina and the computation function of the visual cortex. The conductance of the device can be gradually increased and decreased by light stimuli, corresponding to long-term potentiation and depression. This can be used to construct artificial neural networks (ANNs) for neuromorphic computing.
Accordingly, the array 104 may encode a set of node weights of at least one layer of an ANN. That is, the array 104 may include an embedded artificial intelligence subsystem, given by the one or more layers defined by the node weights, as stored in the optically-programmed memory of each of the elements 106. The embedded artificial intelligence subsystem of the array 104 may cooperate with the artificial intelligence subsystem 108 to realize the machine vision functionality described herein. In particular, with respect to the glare reduction function of the system 100, the array 104 may receive an input image (e.g., via the image sensor 110 or via sensing capabilities of the array 104 itself), and process the input image via the embedded artificial intelligence subsystem to generate a modified or filtered output for each pixel of the input image. That is, the embedded artificial intelligence subsystem of the array 104 filters the input image. Specifically, the configured node weights of the embedded ANN may be configured for glare reduction filtering, for example via factoring the polarization of the light in the received input. The filtered output may then be provided to the artificial intelligence subsystem 108 to generate an output image for machine vision functionality. In particular, the output image may represent the input image with reduced glare.
In particular, Van der Waals (vdW) heterostructures are promising for optical sensing, memory, and computing. Two-dimensional (2D) materials have excellent optoelectronic properties, atomically thin thickness, high-carrier mobility, and tuneable electrical transports. In particular the OCPM as described herein may include 2D materials with in-plane anisotropic structures which can realize polarimetric functions to lights. The presently described 2D polarimetric devices utilizing atomic-level anisotropy are compatible with complementary metal-oxide semiconductor (CMOS) fabrication techniques. 2D materials devices have the potential to realize high-density integration and excellent scalability. Furthermore, the presently described 2D materials exhibit outstanding memory, and computing capabilities, which are easier to fuse with polarimetry functionalities. Multi-terminal memtransistors based on vdW heterostructures can achieve programmable optoelectronics and complex functions of hetero-synaptic plasticity that can be used for brain-inspired neuromorphic computing. Accordingly, the presently described system 100 seamlessly integrates polarization sensitivity, light sensing, memory, and neuromorphic computing, which will leverage novel AI applications.
Thus, for example, polarimetric sensor array 104 may include a Van der Waals heterostructure formed by stacked layers of rhenium disulfide (ReS2) and germanium diselenide (GeSe2) with anisotropic structures. Such a structure may provide the OCPM properties of the polarimetric sensor array 104. For example, the ReS2/GeSe2 structure provided both positive photoconductivity (PPC) and negative photoconductivity (NPC) under the light of 808 nm and 405 nm, respectively.
For example, referring to
The atomic structures of ReS2 and GeSe2 are shown in greater detail in
ReS2 is a transition metal dichalcogenide (TMDC), demonstrating very stable properties in ambient conditions. It is an n-type semiconductor with a direct bandgap. ReS2 has a distorted 1T structure deriving from the hexagonal structure. A 2D ReS2 layer is consisting of three atomic layers of ‘S-Re-S’. Re-S forms covalent bonds. Each Re is covalently bonded with 6 S atoms in an octahedral geometry. Each S atom is covalently bonded to three Re atoms. Four Re atoms are constructed into a parallelogram that demonstrates in-planar anisotropy. The anisotropic structure determines bi-axial optical and electrical properties, indicating polarization-sensitive properties. Layers are stacked together due to van der Waals (vdW) forces, similar to other 2D layered materials. For the OCPM in this work, the b-axis direction was defined as the reference direction of the device. The channel current (source-drain current (Ids)) flowed along the b-axis of ReS2.
GeSe2 is a two-dimensional crystal with a monoclinic crystal structure. GeSe2 is a layered transition metal dichalcogenide (TMDC). It has a direct wide bandgap that shows excellent optoelectronic properties. The GeSe2 is a two-dimensional crystal with a tetragonal structure. As depicted in
ReS2 and GeSe2 have in-plane anisotropic structures and direct bandgaps, indicating excellent bi-axial optical and electrical properties. Polarization-sensitive absorption spectra were measured to further investigate the bi-axial optical properties.
In an experimental setup, a linear polarizer was utilized to polarize the irradiated light. For the initial state (rotation angle=0°), the polarization direction of irradiated light was parallel with the b-axis direction of measure materials. To acquire the polarization-sensitive absorption spectra mapping, the sample was rotated clockwise by an angle of 10° for each step. Polarization-sensitive absorption spectra mapping of the ReS2 layer was recorded and is depicted in
A higher absorption coefficient was observed when the polarized light was parallel with the b-axis of ReS2 ([010] direction of the ReS2 crystal) compared to the perpendicular counterpart. Besides, the blueshift feature of absorption spectra was found when the rotation angle increased from 0° to 90°. This implied the anisotropic electronic properties. The polarization-sensitive absorption spectra mapping of the GeSe2 layer was also measured and is depicted in
Furthermore, to confirm the anisotropic characteristic of ReS2 and GeSe2, the Raman spectra of the materials were measured. The typical Raman spectrum of the ReS2/GeSe2 double layer, in particular at the heterojunction, was shown, as can be seen in
The angle-dependent Raman intensity of the A1g mode was studied. The experimental and fitting plots of ReS2 and GeSe2 are presented in
The thickness of ReS2 and GeSe2 layers were identified by the Atomic Force Microscopy (AFM).
The HAADF atomic scale image of ReS2 indicated a high crystallinity of the exfoliated ReS2, as can be seen in
Further, the fast Fourier transform (FFT) patterns were obtained to get more insight into crystalline structures. In the FFT pattern of ReS2 crystalline as depicted in
Multilayer GeSe2 was exfoliated from bulk crystal using Nitto tape and directly transferred onto a highly p-doped silicon substrate that was covered by silicon dioxide (90 nm). ReS2 was mechanically exfoliated and transferred onto the GeSe2 flake, which was assisted by an aligned transfer system equipped with an optical microscope. The GeSe2 holder can be rotated to change the alignment direction. This can fabricate different alignment angles of ReS2/GeSe2.
Further, the optoelectronic properties and working mechanism of the OCPM were systematically investigated.
The light-wavelength-dependent transfer curves were measured to identify the operation gate voltage, as depicted in
The l-t plot under an unpolarised pulsed light stimulation (808 nm wavelength) indicates non-volatile positive photoconductivity (PPC) properties, with (Vgs=−10V), as can be seen in
In comparison, after the device was programmed to the high conductivity by 808 nm, the conductance decreased when the light with the wavelength of 405 nm was shined on the device, and the lowered conductance was remained after the illumination, which meant the non-volatile negative photoconductivity (NPC), with (Vgs=−10 V), as depicted in
The retention time of the device programmed by pulsed gate voltages was measured and is illustrated in
The programmable conductance can represent synaptic weights in ANNs, implementing the matrix-vector multiplication (MVM) for deep learning algorithms. The architecture used lights to program synaptic weights. The photonic device increases processing speed owing to high bandwidth, lowered parasitic crosstalk, and achieves ultralow power consumption comparing with electrical counterpart.51 Minimizing the asymmetric nonlinearity (ANL) of weight updating (potentiation/depression) is used for computing accuracy. A small ANL of 0.19 was obtained in the OCPM as can be seen in
In particular, the value of ANL can be calculated according to the equation (6):
The feature was comparable with the electronic-controlled artificial synapses. Further, the device was operated for over 30 cycles without obvious variation of conductance. In
Moreover, conductance modulations under multiple optical pulses with different pulse numbers (PN), pulses widths (PW), and pulse power (PP) were characterized, and are illustrated in
The endurance of the OCPM was tested by programming the device with electronic stimuli as illustrated in
The working mechanism of the OCPM has been investigated. The density functional theory (DFT) was employed to calculate the energy band structure of the ReS2/GeSe2 heterojunction, which is depicted in
First-principles calculations were performed using the projector-augmented wave (PAW) method as implemented in the Vienna ab initio simulation package (VASP). The Perdew-Burke-Ernzerhofer (PBE) formula within the generalized gradient approximation (GGA) was used to describe the exchange correlation. The DFT-D2 method of Grimme was used to describe the vdW heterostructure interaction. The vacuum region of 20 Å was set along the x direction to avoid virtual interaction between adjacent images for 2D vdW heterostructure. The cutoff energy was 450 eV. All the structures were relaxed until the forces on all unconstrained atoms were smaller than 0.01 eV/Å, and the total energy convergence criterion was 10−4 eV.
The working mechanisms of PPC and NPC are schematically depicted in
As for the NPC, light with a wavelength of 405 nm was used to decrease the conductance of the ReS2 channel. When the light was shined on the device, hole-electrons pairs were mainly generated in the GeSe2 layer due the larger area. Hole charges were attracted to the SiO2 layer due to the negative gate voltage and trapped at the interface GeSe2/SiO2. Some holes moved from GeSe2 to the ReS2 side owing to the small valance band offset and combined with electrons in ReS2, which decreased the carrier density in the ReS2 layer. Meanwhile, the electrons generated in the GeSe2 also migrated to the ReS2 side, but they were mainly trapped at the ReS2/GeSe2 interface, which was due to the quantum well formed by energy band bending. Those trapped electrons did not contribute to conductivity. The ReS2/GeSe2 heterostructure has bi-axial optoelectronic properties due to in-plane anisotropic structures. This was the basis for the PPC and NPC with polarimetric functions.
Moreover, the influence of the twist angles (the twist angle between the b-axis of ReS2 and the b-axis of GeSe2) on optoelectronic responses has been investigated. The twist angle in the twistronic optoelectronic devices was adjusted from 0° to 90° by in-situ rotating the ReS2 layer as can be seen in
Accordingly, as described herein, the system 100 with an OCPM-based polarimetric sensor array 104 is configured for multiple machine vision functions, including image sensing, polarimetric measurement, and optical-controlled synaptic weights. The sensory functions realize front-end image sensing, which is the input for high-level back-end cognitive tasks. On the other hand, the fully optical-controlled conductance can represent the weight updating in ANNs. We demonstrate that OCPM arrays deploy deep learning algorithms for cognitive tasks. Moreover, the ability of polarimetric measurement can be used for navigation and anti-glare pattern recognition, which goes beyond traditional humanoid machine vision and revolutionizes interspecies-chimera machine vision.
In particular, many animals (for example, honeybees) rely on the sun compass for spatial orientation. They can effectively identify directions even in poor weather (e.g. invisible sun, cloudy) by measuring the celestial polarization patterns, which cannot be done by human eyes. Honeybees are masters at real-time navigating to attractive flower patches with the help of polarised-light patterns. A scout honeybee uses waggle dance (waggles back and forth as moving forward in a straight line) on the honeycomb to transmit the location information to its nestmates, for example as depicted in
In the sky, the polarized sunlight is symmetrically distributed about the solar meridian (SM). The polarization direction (PD) is perpendicular to the direction of SM. Therefore, direction identification can be achieved by measuring the direction of SM or PD. As a proof-of-concept, the navigation accuracy was experimentally evaluated. A typical cloudy day was selected. In practical application, the degree of linear polarization is smaller than 60%, which is calculated by the equation (I∥−I⊥)/(I∥+I⊥). I∥ and I⊥ are light intensities along two perpendicular directions. It is challenging to effectively detect weakly polarized skylights with monolithic polarization-sensitive devices. The perceptive architecture of OCPM array coupling with algebraic analysis can realize navigation applications (
In particular, referring to
The celestial polarization pattern is symmetrical about the solar meridian (SM) based on the single-scattering Rayleigh model. The polarization direction (PD) (or e-vector E) is perpendicular to the SM. Monitoring the direction of PD to deduce the direction of SM is the basis of honeybee-biomimetic navigation. Particularly, the OCPM is sensitive to polarized natural light, which simplifies the system's complexity due to the filter-free and polarizer-free features. To demonstrate the practical applications, the OCPM was tested outdoors to measure the SM direction. The crystalline b-axis of ReS2 and GeSe2 was initially aligned with the north, which was defined as 0°.
A schematic diagram of an experimental test setup for measuring celestial polarization patterns is shown in
Polarization-dependent Ids-Vgs mapping was measured and is depicted in
The reconfigurable electronic feature makes it more compatible with other ancillary circuits. Because the operation currents have different margins in two HRS (0-0.2 nA) and LRS (20-94 nA) working models, which can have a better compatibility with a wider range of peripheral circuits.
The fitting curve matched well with experimental results. The values of φ were obtained for HRS and LRS was 29.3° and 29.6° respectively, which can be utilized to deduce the direction of SM.
Typical polar coordinate plots of angle-dependent Ids at LRS (
The evolution of solar azimuth angle over a day can be calculated as shown in
Thus, based on the experimental results, the polarization sensitivity of the system 100 can be used for real-time navigation inspired by honeybees. Specifically, referring to
At block 3405, the system 100, and more particularly, the polarimetric sensor array 104 detects incoming light representing one or more objects and/or the environment within the field of view 112 of the array 104.
At block 3410, the system 100 selects a detection angle at which to arrange the array 104. The detection may be relative to a reference direction, which may preferably be a known direction, such as North, but in other examples may be an arbitrarily selected reference direction, for example based on the application of the navigation method 3400. For example, the system 100 may rotate the rotatable platform 116 to the selected detection angle. In other examples, when multiple sensor elements 106 are disposed within the array 104 at different angles, the system 100 may select the subset of the sensor elements 106 which are angled at the selected detection angle.
At block 3415, the system 100 is configured to measure a current detected by the array 104 (or subset of sensor elements 106 thereof). In particular, the OCPM nature of the sensor elements 106 and the array 104 causes the current to be polarization dependent, and accordingly the current acts as a polarization signal representative of the polarization of the detected light.
At block 3420, the system 100 determines whether a sufficient amount of polarization signals or points (i.e., current values at different detection angles) have been collected. For example, the system 100 may have a predefined threshold number of points (e.g., 10, 16, etc.), e.g., based on a predefined number of detection angles through which the array 104 is to rotate, or the like.
If the determination at block 3420 is negative, then the system 100 returns to block 3410 to select another detection angle to obtain another point.
If the determination at block 3420 is affirmative, then the system 100 proceeds to block 3425.
At block 3425, the system 100 is configured to fit a curve to the combination of polarization signals and detection angles obtained. For example, the system 100 may apply the curve given by equation (7) to the points.
At block 3430, according to the curve fitting at block 3425, the system 100 may determine the value θ as the angle between the PD and the reference direction.
At block 3435, the system 100 may determine a location of the system 100 based on the value θ. In particular, the value θ may represent the solar azimuth angle (i.e., relative to the North direction as a the reference direction) and hence the system 100 may apply the solar azimuth angle to determine the location given by a latitude and longitude of the system 100. In other examples, such as if the reference direction is not objectively known, the system 100 may apply the value θ to determine a relative location, for example for navigation to and from a given starting point, or the like. In still further examples, other known constants may be applied to enable objective location determination and/or navigational functions.
In some examples, in addition to navigational capabilities, human-bee chimera machine vision integrates polarimetry and cognitive recognition functions, which overmatch traditional neuromorphic vision systems. Reflected light from a surface (usually non-metallic) is linearly polarized. The reflected light is usually very strong, inducing glare spots that will distort visualized images and decrease cognitive accuracy in machine vision. Specifically, automatic vehicles and drones use real-time sensing and processing traffic information. Therefore, three major challenges for image processing in automatic machines are sensing and processing accuracy, computing speed, and energy consumption. However, reflective materials used in buildings, advertising boards with smooth surfaces, and a body of water produce lots of glare spots. This will exacerbate the processing accuracy and increase the risk. Note, the formation of glare spots is because the reflected light is polarized light parallel to the surface. The polarimetric capabilities of OCPM can filter the polarized light and reduce the influence of glare spots. Besides, fully optic-controlled synapse can be used to build ANNs for cognitive pattern recognition. An anti-glare machine vision can be developed by integrating polarimetric and synaptic functions.
The comparison between human vision systems and OCPM-based honeybee-chimera machine vision is illustrated in
As an example, the anti-glare pattern recognition with OCPM has been presented. A dataset with three letters of “O”, “F”, and “U” with 6×7 pixels was adopted for the training and testing. The scale of each pixel was 0-1. Ideal patterns (no background noise and glare-spot-like noise), no anti-glare patterns, and anti-glare patterns are shown in
In particular, with reference to
Thus, in operation with respect to the system 100, the image may be captured by the image sensor 110 or directly by the polarimetric sensor elements 106 of the polarimetric sensor array 104 and may correspond to at least a portion of the field of view 112. The image may be exposed to the OCPM array 104, and in particular, to the embedded artificial intelligence subsystem, which may be, for example, an embedded artificial neural network (e.g., given by at least one layer) encoded by node weights stored by the sensor elements 106. More particularly, the node weights may be stored by leveraging the atomic structure, large bandgap, and optically programmed storage structure of the ReS2/GeSe2 structure. Specifically, as can be seen in
The glare spot from reflected light was assumed as linear polarized light parallel with the objective surface. The OCPM was perpendicular to the polarized reflected light. The brightness of glare spots was reduced to 1/6.4 based on the dichroic ratio of the OCPM. The neuromorphic computing for pattern recognition was based on the code written with MATLAB. The illustration of dataset images for pattern recognition training is shown in
Three example patterns (letter “F”) with and without anti-glare are shown in
The pattern recognition accuracies with and without OCPM-based anti-glare processing were calculated and depicted in
The recognition accuracy plots for 9 running cycles were calculated to implement statistics analysis, as depicted in
To quantitatively evaluate the improvement regarding converge computing speed and energy consumption, the epochs of all simulation running cycles were extracted. Every training epoch should spend the same time and energy for the ANNs with and without anti-glare because they shared the same architecture. Thus, the computing speed and energy consumption are roughly proportional to training epochs, which is the foundation of quantitatively identifying the computing efficiency of the two candidates. The summary of training epochs obtaining different recognition accuracies is given in Table 1.
The data was extracted from 9 stimulation running cycles. The epochs were the average values of 9 running cycles. In particular, the average values were summarized in Table 1. To get the recognition accuracy of 70%, 57 and 99 training epochs were required for with and without anti-glare functions respectively. 2410 epochs were needed to get 99% accuracy for the humanoid neuromorphic machine vision without anti-glare functions. Remarkably, the anti-glare function can significantly reduce the training epochs to merely 245 using interspecies-chimera machine vision, almost one magnitude of difference. The ratio of improvement (RoI) was as high as nearly 90% which was much higher than previous reported results with image pre-processing abilities.
The rate of improvement (RoI) is defined by the following equation (8):
The number of epochs was the average of 9 running simulations. The RoI was 89.8% to realize an accuracy of 99% with the anti-glare function. The training efficiency improvement was higher than the previous reported results, in which optoelectronic resistive random access memory (ORRAM) suppressed the background noise. Furthermore, the energy budget was estimated for the training with and without anti-glare. The ratio of energy consumption (RoEC) is described by the following equation (9):
Considering the energy consumption for the training process, by utilizing the anti-glare function, the RoEC was merely 11.4% to get a recognition accuracy of 99% compared to that without the anti-glare function. A remarkable improvement in computing speed and energy efficiency was achieved by equipping ANNs with anti-glare abilities. OCPM demonstrates great prospects for advanced neuromorphic vision systems.
The same conclusion can be obtained that the anti-glare interspecies-chimera machine vision showed superior performances despite minor differences among different running cycles. A statistical comparison of training epochs to get the same recognition accuracies with and without anti-glare abilities is shown in
Besides, the training processing is the most energy-intensive step which is determined by neural network size, dataset size, training epochs, etc. As we fixed other variables (network size, dataset size, etc.) during the training process. The training epochs qualitatively reflect the significantly improved energy efficiency. The training energy consumption was reduced to 11.4% aided by the anti-glare abilities, as depicted in the inset of
As described herein, an interspecies-chimera machine vision by integrating the polarimetric function of honeybees and the intelligence of human beings. In particular, the device based on a van der Waals heterostructure (ReS2/GeSe2) provides polarization sensitivity, nonvolatility, and positive/negative photoconductance simultaneously. The polarimetric measurement can identify celestial polarizations for real-time navigation. Meanwhile, the anti-glare recognition with polarimetry improved one magnitude of energy efficiency compared to the traditional neuromorphic machine vision counterpart. This technique may have applications in autonomous vehicles, medical diagnoses, intelligent robotics, etc.
The scope of the claims should not be limited by the embodiments set forth in the above examples but should be given the broadest interpretation consistent with the description as a whole.
Number | Date | Country | |
---|---|---|---|
63545001 | Oct 2023 | US |