The present disclosure relates generally to image sensors and, more specifically, to pixels with integrated processing capabilities.
Computer vision often relies on light intensity-based pixel data collected through state-of-the-art CMOS image sensors. However, in almost all cases, appropriate context for the signals transmitted by the pixels is missing (or are extremely vague) with respect to the ‘real-world events’ being captured by the sensor. Thus, the onus of processing is rather put on algorithms, such as intelligent machine learning algorithms, to pre-process, extract appropriate context, and make intelligent decisions based on light intensity-based pixel data. Such a vision pipeline (e.g., processing procedure from pixel to processor) may lead to 1) complex machine learning algorithms designed to cater to image/video data without appropriate context 2) increased the time to decision, associated with the processing time of the machine learning algorithms 3) energy-hungry and slow access to pixel data being captured and generated by the CMOS image sensor. None of which is to suggest that any technique suffering to some degree from these issues is disclaimed or that any other subject matter is disclaimed.
The following is a non-exhaustive listing of some aspects of the present techniques. These and other aspects are described in the following disclosure.
Some aspects include sensors with circuitry which may generate specific responses to types of movement and/or shapes of objects.
Some aspects include circuit structures for implementation of retinal-analogous computations, including using advanced 3D integration of semiconductor chips.
Some aspects include a structure and/or method of forming a structure for Active Pixel Sensor (APS) based bipolar spike generator circuitry.
Some aspects include a structure and/or method of forming a structure for transistor-based circuitry for object motion sensitivity.
Some aspects include a structure and/or method of forming a structure for mapping overlapping center-surround receptive field in two-dimensional arrays, including for cameras.
Some aspects include implementation of some computations in inner layers of a camera's image array (e.g., in pixels analogous to retina components).
Some aspects include a structure and/or method of forming a structure capable of extracting object motion and/or shape features, including in faster and/or less energy consuming ways that in CMOS and DVS image sensors.
The above-mentioned aspects and other aspects of the present techniques will be better understood when the present application is read in view of the following figures in which like numbers indicate similar or identical elements:
While the present techniques are susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. The drawings may not be to scale. It should be understood, however, that the drawings and detailed description thereto are not intended to limit the present techniques to the particular form disclosed, but to the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the present techniques as defined by the appended claims.
To mitigate the problems described herein, the inventors had to both invent solutions and, in some cases just as importantly, recognize problems overlooked (or not yet foreseen) by others in the field of image sensing. Indeed, the inventors wish to emphasize the difficulty of recognizing those problems that are nascent and will become much more apparent in the future should trends in industry continue as the inventors expect. Further, because multiple problems are addressed, it should be understood that some embodiments are problem-specific, and not all embodiments address every problem with traditional systems described herein or provide every benefit described herein. That said, improvements that solve various permutations of these problems are described below.
Animal eyes (e.g., eye structure) are extremely diverse and may be specialized for the environment and behavioral niche of each species. Specialization may be particularly robust in the retina, a part of the central nervous system containing parallel circuits for representing or capturing different visual features. In contrast, the engineered ‘eyes,’ e.g., image sensor technology, used in machine vision, may be highly stereotyped. Even though cameras or other imaging devices may have different optics on the front end, the image sensor chip, which may represent the electronic analogue of the biological retina, may essentially be a two-dimensional array of pixels with each transmitting a luminance signal at a fixed frame rate. In some embodiments, the efficiency and performance of machine vision may be improved by using specialized image sensors that replicate some of the feature-selective computations performed in biological retinas.
Efforts to bring biologically-inspired functionality to electronic image sensors date back to at least the 1980s with the advent of neuromorphic sensors. Two related aspects of visual computation, which were recognized in retinal neurobiology by the late 1980s, have dominated the field of neuromorphic vision sensors. The first idea was to mimic luminance adaptation, the computation used by the retina to adjust the dynamic range of its biological components to that of the visual scene. Humans may use vision over 10 orders of magnitude of luminance and even single natural images may vary in brightness by more than a factor of 105. These high dynamic ranges may be poorly represented by linear photodiodes and digitization to 8 or even 12 bits. High dynamic range (HDR) cameras use multiple exposures to reconstruct an image, trading bit depth for frame rate, while logarithmic detectors use range compression to avoid saturation. The second aspect of retinal computation to take hold in neuromorphic image sensors is change detection—the propensity of retinal neurons to adapt to the mean luminance over time and only transmit information about its change. Event based cameras, or Dynamic Vision Sensors (DVS), may implement temporal differentiation at each pixel and asynchronously transmit binary ‘spike’ events when the luminance change exceeds a threshold. The asynchronous transmission of DVS cameras has critical advantages for high-speed operation, since it is not limited by frame rate, and for efficiency (e.g., energy efficiency, data transmission efficiency, etc.), since pixels that do not change do not transmit data.
In some embodiments, a new class of neuromorphic sensors referred to herein as Integrated Retinal Functionality in Image Sensors (IRIS) is presented. By leveraging understanding of inner retinal circuits, IRIS technology may go beyond luminance adaptation and change detection—features mostly confined to phototransduction and the first retinal synapse—to implement computations that occur in the circuits of the inner retina (or analog), mimicking the feature-selective spike trains of RGCs. In some embodiments, IRIS circuits implementing either (or both) of two retinal motion computations are presented: Object Motion Sensitivity (OMS) and Looming Detection (LD). In some embodiments, the effect of the present disclosure may not be to implement the detailed electro-chemical dynamics of retinal cell types, but rather to functionally mimic the computational behavior of retinal circuits on image sensing platforms.
In some embodiments, OMS may include a computation that enables the visual system to discriminate motion of objects in the world (object motion) from motion due to one's own eye, head, and body movements (self-motion). A subset of RGCs respond to either local motion in the receptive field ‘center’ or differential motion of the receptive field ‘center’ and ‘surround’ regions but remains silent for global motion. OMS RGCs may be important in detecting movements of predators and prey amidst a background of self-motion. In some embodiments, for example for machine vision applications, a fast sensor with built-in OMS may detect moving objects even if the camera itself was moving, for example, on an autonomous vehicle.
In some embodiments, LD may be include a computation analogous to that which may have evolved to warn animals of approaching threats, especially those from overhead. Loom-sensitive RGCs may respond selectively to expanding (e.g., approaching) dark objects—with much weaker responses to other movement, such as translational motion across the visual field. Experiments in flies, zebrafish, frogs, and mice may have established a causal role for LD RGCs signal transmission in eliciting stereotyped escape responses. In machine vision, an LD-equipped sensor may be used on an autonomous vehicle to avoid collisions by enabling fast detection of approaching objects in large area receptive fields.
In some embodiments, OMS and LD circuits may be built on standard Complementary Metal Oxide Semiconductor (CMOS) pixels as well as on DVS pixels. In some embodiments, advances in semiconductor chip stacking technology and highly-scaled, dense CMOS transistors are exploited, such as to embed retina-inspired circuits in a hierarchical manner analogous to the processing layers of the biological retina (shown in
Feature selective circuits in the vertebrate retina, like OMS and LD, may be built from some 5 classes of neurons. Photoreceptors may form the input layer (like the pixels in a camera) and retinal ganglion cells (RGCs) may represent the output. The computations that transform the pixel-like representation of the photoreceptors to the feature selective representation of RGCs may be carried out by the some 3 interneuron classes, including: horizontal cells, bipolar cells, and amacrine cells. Horizontal cells may mainly be involved in lateral inhibition and color processing, but may not play a major role in OMS and LD circuits, which is not to say that circuity performing functions analogous to horizontal cells may not be included in one or more embodiment of the present disclosure. In some embodiments, the components of IRIS circuits may be designed to match the functionality of bipolar and amacrine cells in these computations.
Amacrine cells may be the most diverse class of neurons in the retina, comprising more than 60 cell types. While some of the cellular and synaptic details of amacrine cells may remain imperfectly understood, their algorithmic role in the OMS and LD circuits may be well characterized. In the OMS circuit of
As described above, the OMS computation in the retina may start by detection of change in temporal contrast of input light by the bipolar cells. In other words, for OMS behavior, functionally, the bipolar cells generate a spike for a change in light intensity above a certain threshold.
In some embodiments, for APS-based implementation the focal plane array may be formed by a two-dimensional array of APS pixels with additional circuitry to enable light contrast-change detection. In some embodiments, the array of such contrast-change sensitive APS pixels may sample the input light intensity for each frame, such as in parallel, and compare it to the light intensity of the next frame. In some embodiments, if the light intensity sensed by each APS pixel increases (decreases), the contrast sensitive APS pixels may generate an ON (OFF) spike.
In some embodiments, the APS-based contrast-change detection circuit may be implemented such as as-shown in
In some embodiments, the DVS-based contrast-sensitive pixel circuit may be implemented such as as-shown in (
In some embodiments, when the pixels in the center region 400 generate bipolar-spikes, while at the same time, pixels in the surround region 410 also generate bipolar-spikes, it may indicate that the receptive field comprising of the center region 400 and the surround region 410 is experiencing global or background motion without any object motion. In such a case, the voltage accumulated on the capacitor Cint 420 from center pixels may be offset by the discharge effect of surround pixels and the CMOS buffer 422 output may remain low. In some embodiments, if the center pixels receive bipolar-spikes, without significant corresponding bipolar-spikes on the surround pixels, the voltage accumulated on the capacitor Cint 420 may not experience a significant discharging path through the surround pixels, which may result in higher voltage that pulls the output of the CMOS buffer 422 high. The generated spike from the CMOS buffer 422, thus, represents the output OMS-feature-spike, which may indicate an object motion detected in the center region with respect to the surround region.
In some embodiments, the design aspects of the OMS circuit, such as the circuit proposed in
In some embodiments, the method mimicking the center-surround receptive field (such as shown in
As seen in
An object moving left (e.g., in time frame 660) may result in a low signal from the ON bipolar spike transistor 632 and a high signal from the OFF bipolar spike transistor 634 from the left boundary pair 630, and a high signal from the ON bipolar spike transistor 642 and a low signal from the OFF bipolar spike transistor 644 from the right boundary pair 640. This may correspond to a transient spike in the OUT 680 current of a circuit containing both the left boundary pair 630 and the right boundary pair 640. The transient spike may resolve to a low OUT 680 current, once the object has been detected.
A looming object (e.g., in time frame 670) may result in may result in a low signal from the ON bipolar spike transistor 632 and a high signal from the OFF bipolar spike transistor 634 from the left boundary pair 630, and a low signal from the ON bipolar spike transistor 642 and a high signal from the OFF bipolar spike transistor 644 from the right boundary pair 640. This may correspond to a transient dip in the OUT 680 current of a circuit containing both the left boundary pair 630 and the right boundary pair 640. The transient dip may resolve to a high OUT 680 current, once the object has been detected.
In a further example, consider the dark object within the receptive field which may be approaching (or looming). In some embodiments, in such a case, the pair of transistors on the left and the right boundary of the object may simultaneously experience decrease in light intensity (e.g., as the boundary of the dark object expands), thereby generating OFF bipolar-spikes. The OFF spike transistors at the left and the right boundary may be activated by the OFF bipolar spikes, while all the other transistors may remain OFF. In some embodiments, therefore, the boundary OFF spike transistors may pull the voltage across Cint low. In some embodiments, in response to a low voltage on Cint 610 the logic circuit may generate a high output voltage (or an LD feature-spike) indicating an approaching or looming object in the receptive field. In some embodiments, instead of a dark object the LD circuit may also generate an LD feature-spike if a bright object is approaching in the receptive field. For example, in this case, the ON spike transistors at the left and right boundary of the object may be active (as depicted in
In some embodiments, IRIS sensors may embed retinal feature extraction behavior using retina-inspired circuits within image sensors. In some embodiments, similar circuit-technology design techniques (e.g., similar to OMS and LD) may be used to embed a rich class of retinal functionality including color, object orientation, object shape, etc. in image sensors. In some embodiments, IRIS sensors can be implemented based on underlying APS or DVS pixels. In some embodiments, for APS pixels to achieve high dynamic range a coarse grained (at pixel-array-level) or fine-grained (at individual pixel level) exposure timing control may be implemented. In some embodiments, the photodiodes associated with IRIS sensors may span wide range of wavelengths including visible light, infra-red, or near infra-red light.
In some embodiments, advances in 3D integration of semiconductor chips may enable IRIS sensors. In some embodiments, 3D integration may allow integration of routing metal layers and transistor-based circuits required from implementing spatio-temporal computations similar to retinal circuit directly above (or under) the pixel array. Such 3D integrated IRIS sensors may use various 3D packaging technologies like metal-to-metal fusion bonding, through silicon vias (TSVs), etc. Further, heterogeneous sensors operating at different wavelengths may be co-integrated to extract retina-like feature vectors over different spectrums of light.
In some embodiments, IRIS sensors may have a significant impact on computer vision. Today's computer vision may rely exclusively on light intensity-based (APS) or light change-detection-based (DVS) pixels data collected through state-of-the-art CMOS image sensors. However, in almost all cases, appropriate context for the pixels may be missing (or may be extremely vague) with respect to the ‘real-world events’ being captured by the sensor. Thus, the onus of processing may be put on intelligent machine learning algorithms to pre-process, extract appropriate context, and make intelligent decisions based on pixel data. Unfortunately, such a vision pipeline may lead to 1) complex machine learning algorithms designed to cater to image/video data without appropriate context 2) increases the time to decision, associated with the processing time of the machine learning algorithms 3) energy-hungry and slow access to pixel data being captured and generated by the CMOS image sensor. In some embodiments, IRIS sensors may usher in new frontiers in vision-based decision making by generating highly specific motion and shape-based features, providing valuable context to pixels captured by the camera. In some embodiments, the underlying algorithms processing data generated from IRIS sensors may be based on traditional deep learning models or on emerging set of spiking neural networks that could process feature-spikes generated from IRIS sensors. In some embodiments, since IRIS cameras may use APS pixels, in general, they may generate both feature-spikes and light intensity map as and when required by the computer vision algorithms.
Computing system 1100 may include one or more processors (e.g., processors 1120a-1120n) coupled to system memory 1130, and a user interface 1140 via an input/output (I/O) interface 1150. A processor may include a single processor or a plurality of processors (e.g., distributed processors). A processor may be any suitable processor capable of executing or otherwise performing instructions. A processor may include a central processing unit (CPU) that carries out program instructions to perform the arithmetical, logical, and input/output operations of computing system 1100. A processor may execute code (e.g., processor firmware, a protocol stack, a database management system, an operating system, or a combination thereof) that creates an execution environment for program instructions. A processor may include a programmable processor. A processor may include general or special purpose microprocessors. A processor may receive instructions and data from a memory (e.g., system memory 1130). Computing system 1100 may be a uni-processor system including one processor (e.g., processor 1120a-1120n), or a multi-processor system including any number of suitable processors (e.g., 1120a-1120n). Multiple processors may be employed to provide for parallel or sequential execution of one or more portions of the techniques described herein. Processes, such as logic flows, described herein may be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating corresponding output. Processes described herein may be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit). Computing system 1300 may include a plurality of computing devices (e.g., distributed computing systems) to implement various processing functions.
Computing system 1100 may include one or more photodetectors (e.g., photodetectors 1152) or other sensors in an integrated sensor 1160. The integrated sensor 1160 may be coupled to system memory 1130, and a user interface 1140 via an input/output (I/O) interface 1150. Photodetectors 1152 may be pixels, sub-pixels, photodiodes, resistive photodetectors, etc. Photodetectors 1152 may be coupled to spike circuits 1102a-1102n, where each of the spike circuits 1102a-1102n may be coupled to one or more of the photodetectors 1152. The spike circuits 1102a-1102n may be bipolar spike circuits, such as previously described. The spike circuits 1102a-1102n may operate on outputs of the photodetectors 1152. The spike circuits 1102a-1102n may generate bipolar spikes based on some conditions (e.g., signals) or one or more of the photodetectors 1152. The spike circuits 1102a-1102n may be coupled to an integration circuit 1104, which may integrate (or accumulate) output of one or more of the spike circuits 1102a-1102n.
The spike circuits 1102a-1102n may operate on outputs of the photodetectors 1152, which may allow weighting, transmission, pass though, collection, amplification, etc. of the outputs of the photodetectors 1152. The spike circuits 1102a-1102n may select for outputs of the photodetectors 1152 corresponding specific vision sensing schemes, such as OMS or LD. The spike circuits 1102a-1102n or another computation element may contain additional memory elements, such as ROM, eDRAM, accumulation elements, etc. which may be readable or readable and writable memory. The photodetectors 1152 and spike circuits 1102a-1102n may be controlled by one or more reset element, such as a reset element (not depicted) in communication with the I/O interface 1150 or controlled by one or more of the processors 1120a-1120n. The photodetectors 1152 may be exposed to input, such as light (e.g., in the case of a photosensor) or other input, an analyte (such as temperature), or other sensing material. The photodetectors 1152 may comprise transistors, diodes, etc.
The user interface 1140 may comprise one or more I/O device interface, for example to provide an interface for connection of one or more I/O devices to computing system 1100. The user interface 1140 may include devices that receive input (e.g., from a user) or output information (e.g., to a user). The user interface 1140 may include, for example, graphical user interface presented on displays (e.g., a cathode ray tube (CRT) or liquid crystal display (LCD) monitor), pointing devices (e.g., a computer mouse or trackball), keyboards, keypads, touchpads, scanning devices, voice recognition devices, gesture recognition devices, printers, audio speakers, microphones, cameras, or the like. The user interface 1140 may be connected to computing system 1100 through a wired or wireless connection. The user interface 1140 may be connected to computing system 1100 from a remote location. The user interface 1140 may be in communication with one or more other computing systems. Other computing units, such as located on remote computer system, for example, may be connected to computing system 1100 via a network.
System memory 1130 may be configured to store program instructions 1132 or data 1134. Program instructions 1132 may be executable by a processor (e.g., one or more of processors 1120a-1120n) to implement one or more embodiments of the present techniques. Program instructions 1132 may include modules of computer program instructions for implementing one or more techniques described herein with regard to various processing modules. Program instructions may include a computer program (which in certain forms is known as a program, software, software application, script, or code). A computer program may be written in a programming language, including compiled or interpreted languages, or declarative or procedural languages. A computer program may include a unit suitable for use in a computing environment, including as a stand-alone program, a module, a component, or a subroutine. A computer program may or may not correspond to a file in a file system. A program may be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program may be deployed to be executed on one or more computer processors located locally at one site or distributed across multiple remote sites and interconnected by a communication network.
System memory 1130 may include a tangible program carrier having program instructions stored thereon. A tangible program carrier may include a non-transitory computer readable storage medium. A non-transitory computer readable storage medium may include a machine-readable storage device, a machine-readable storage substrate, a memory device, or any combination thereof. Non-transitory computer readable storage medium may include non-volatile memory (e.g., flash memory, ROM, PROM, EPROM, EEPROM memory), volatile memory (e.g., random access memory (RAM), static random-access memory (SRAM), synchronous dynamic RAM (SDRAM)), bulk storage memory (e.g., CD-ROM and/or DVD-ROM, hard-drives), or the like. System memory 1130 may include a non-transitory computer readable storage medium that may have program instructions stored thereon that are executable by a computer processor (e.g., one or more of processors 1120a-1120n) to cause the subject matter and the functional operations described herein. A memory (e.g., system memory 1130) may include a single memory device and/or a plurality of memory devices (e.g., distributed memory devices). Instructions or other program code to provide the functionality described herein may be stored on a tangible, non-transitory computer readable media. In some cases, the entire set of instructions may be stored concurrently on the media, or in some cases, different parts of the instructions may be stored on the same media at different times.
I/O interface 1150 may be configured to coordinate I/O traffic between processors 1120a-1120n, spike circuits 1102a-1102n, integration circuit 1104, photodetectors 1152, system memory 1130, user interface 1140, etc. I/O interface 1150 may perform protocol, timing, or other data transformations to convert data signals from one component (e.g., system memory 1130) into a format suitable for use by another component (e.g., processors 1120a-1120n). I/O interface 1150 may include support for devices attached through various types of peripheral buses, such as a variant of the Peripheral Component Interconnect (PCI) bus standard or the Universal Serial Bus (USB) standard.
Embodiments of the techniques described herein may be implemented using a single instance of computing system 1100 or multiple computing systems 1100 configured to host different portions or instances of embodiments. Multiple computing systems 1100 may provide for parallel or sequential processing/execution of one or more portions of the techniques described herein.
Those skilled in the art will appreciate that computing system 1100 is merely illustrative and is not intended to limit the scope of the techniques described herein. Computing system 1100 may include any combination of devices or software that may perform or otherwise provide for the performance of the techniques described herein. For example, computing system 1100 may include or be a combination of a cloud-computing system, a data center, a server rack, a server, a virtual server, a desktop computer, a laptop computer, a tablet computer, a server device, a client device, a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a vehicle-mounted computer, or a Global Positioning System (GPS), or the like. Computing system 1100 may also be connected to other devices that are not illustrated, or may operate as a stand-alone system. In addition, the functionality provided by the illustrated components may in some embodiments be combined in fewer components or distributed in additional components. Similarly, in some embodiments, the functionality of some of the illustrated components may not be provided or other additional functionality may be available.
Those skilled in the art will also appreciate that while various items are illustrated as being stored in memory or on storage while being used, these items or portions of them may be transferred between memory and other storage devices for purposes of memory management and data integrity. Alternatively, in other embodiments some or all of the software components may execute in memory on another device and communicate with the illustrated computer system via inter-computer communication. Some or all of the system components or data structures may also be stored (e.g., as instructions or structured data) on a computer-accessible medium or a portable article to be read by an appropriate drive, various examples of which are described above. In some embodiments, instructions stored on a computer-accessible medium separate from computing system 1100 may be transmitted to computing system 1100 via transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as a network or a wireless link. Various embodiments may further include receiving, sending, or storing instructions or data implemented in accordance with the foregoing description upon a computer-accessible medium. Accordingly, the present techniques may be practiced with other computer system configurations.
While various items are illustrated as being stored in memory or on storage while being used, these items or portions of them may be transferred between memory and other storage devices for purposes of memory management and data integrity. Alternatively, in other embodiments some or all of the software components may execute in memory on another device and communicate with the illustrated computer system via inter-computer communication. Some or all of the system components or data structures may also be stored (e.g., as instructions or structured data) on a computer-accessible medium or a portable article to be read by an appropriate drive, various examples of which are described above. In some embodiments, instructions stored on a computer-accessible medium separate from computer system may be transmitted to computer system via transmission media or signals such as electrical, electromagnetic, or digital signals, conveyed via a communication medium such as a network or a wireless link. Various embodiments may further include receiving, sending, or storing instructions or data implemented in accordance with the foregoing description upon a computer-accessible medium. Accordingly, the present techniques may be practiced with other computer system configurations.
In block diagrams, illustrated components are depicted as discrete functional blocks, but embodiments are not limited to systems in which the functionality described herein is organized as illustrated. The functionality provided by each of the components may be provided by software or hardware modules that are differently organized than is presently depicted, for example such software or hardware may be intermingled, conjoined, replicated, broken up, distributed (e.g., within a data center or geographically), or otherwise differently organized. The functionality described herein may be provided by one or more processors of one or more computers executing code stored on a tangible, non-transitory, machine readable medium. In some cases, notwithstanding use of the singular term “medium,” the instructions may be distributed on different storage devices associated with different computing devices, for instance, with each computing device having a different subset of the instructions, an implementation consistent with usage of the singular term “medium” herein. In some cases, third party content delivery networks may host some or all of the information conveyed over networks, in which case, to the extent information (e.g., content) is said to be supplied or otherwise provided, the information may be provided by sending instructions to retrieve that information from a content delivery network.
The reader should appreciate that the present application describes several independently useful techniques. Rather than separating those techniques into multiple isolated patent applications, applicants have grouped these techniques into a single document because their related subject matter lends itself to economies in the application process. But the distinct advantages and aspects of such techniques should not be conflated. In some cases, embodiments address all of the deficiencies noted herein, but it should be understood that the techniques are independently useful, and some embodiments address only a subset of such problems or offer other, unmentioned benefits that will be apparent to those of skill in the art of reviewing the present disclosure. Due to costs constraints, some techniques disclosed herein may not be presently claimed and may be claimed in later filings, such as continuation applications or by amending the present claims. Similarly, due to space constraints, neither the Abstract nor the Summary of the Invention sections of the present document should be taken as containing a comprehensive listing of all such techniques or all aspects of such techniques.
It should be understood that the description and the drawings are not intended to limit the present techniques to the particular form disclosed, but to the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the present techniques as defined by the appended claims. Further modifications and alternative embodiments of various aspects of the techniques will be apparent to those skilled in the art in view of this description. Accordingly, this description and the drawings are to be construed as illustrative only and are for the purpose of teaching those skilled in the art the general manner of carrying out the present techniques. It is to be understood that the forms of the present techniques shown and described herein are to be taken as examples of embodiments. Elements and materials may be substituted for those illustrated and described herein, parts and processes may be reversed or omitted, and certain features of the present techniques may be utilized independently, all as would be apparent to one skilled in the art after having the benefit of this description of the present techniques. Changes may be made in the elements described herein without departing from the spirit and scope of the present techniques as described in the following claims. Headings used herein are for organizational purposes only and are not meant to be used to limit the scope of the description.
As used throughout this application, the word “may” is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must). The words “include”, “including”, and “includes” and the like mean including, but not limited to. As used throughout this application, the singular forms “a,” “an,” and “the” include plural referents unless the content explicitly indicates otherwise. Thus, for example, reference to “an element” or “a element” includes a combination of two or more elements, notwithstanding use of other terms and phrases for one or more elements, such as “one or more.” The term “or” is, unless indicated otherwise, non-exclusive, i.e., encompassing both “and” and “or.” Terms describing conditional relationships, e.g., “in response to X, Y,” “upon X, Y,”, “if X, Y,” “when X, Y,” and the like, encompass causal relationships in which the antecedent is a necessary causal condition, the antecedent is a sufficient causal condition, or the antecedent is a contributory causal condition of the consequent, e.g., “state X occurs upon condition Y obtaining” is generic to “X occurs solely upon Y” and “X occurs upon Y and Z.” Such conditional relationships are not limited to consequences that instantly follow the antecedent obtaining, as some consequences may be delayed, and in conditional statements, antecedents are connected to their consequents, e.g., the antecedent is relevant to the likelihood of the consequent occurring. Statements in which a plurality of attributes or functions are mapped to a plurality of objects (e.g., one or more processors performing steps A, B, C, and D) encompasses both all such attributes or functions being mapped to all such objects and subsets of the attributes or functions being mapped to subsets of the attributes or functions (e.g., both all processors each performing steps A-D, and a case in which processor 1 performs step A, processor 2 performs step B and part of step C, and processor 3 performs part of step C and step D), unless otherwise indicated. Similarly, reference to “a computer system” performing step A and “the computer system” performing step B can include the same computing device within the computer system performing both steps or different computing devices within the computer system performing steps A and B. Further, unless otherwise indicated, statements that one value or action is “based on” another condition or value encompass both instances in which the condition or value is the sole factor and instances in which the condition or value is one factor among a plurality of factors. Unless otherwise indicated, statements that “each” instance of some collection have some property should not be read to exclude cases where some otherwise identical or similar members of a larger collection do not have the property, i.e., each does not necessarily mean each and every. Limitations as to sequence of recited steps should not be read into the claims unless explicitly specified, e.g., with explicit language like “after performing X, performing Y,” in contrast to statements that might be improperly argued to imply sequence limitations, like “performing X on items, performing Y on the X'ed items,” used for purposes of making claims more readable rather than specifying sequence. Statements referring to “at least Z of A, B, and C,” and the like (e.g., “at least Z of A, B, or C”), refer to at least Z of the listed categories (A, B, and C) and do not require at least Z units in each category. Unless specifically stated otherwise, as apparent from the discussion, it is appreciated that throughout this specification discussions utilizing terms such as “processing,” “computing,” “calculating,” “determining” or the like refer to actions or processes of a specific apparatus, such as a special purpose computer or a similar special purpose electronic processing/computing device. Features described with reference to geometric constructs, like “parallel,” “perpendicular/orthogonal,” “square”, “cylindrical,” and the like, should be construed as encompassing items that substantially embody the properties of the geometric construct, e.g., reference to “parallel” surfaces encompasses substantially parallel surfaces. The permitted range of deviation from Platonic ideals of these geometric constructs is to be determined with reference to ranges in the specification, and where such ranges are not stated, with reference to industry norms in the field of use, and where such ranges are not defined, with reference to industry norms in the field of manufacturing of the designated feature, and where such ranges are not defined, features substantially embodying a geometric construct should be construed to include those features within 15% of the defining attributes of that geometric construct. The terms “first”, “second”, “third,” “given” and so on, if used in the claims, are used to distinguish or otherwise identify, and not to show a sequential or numerical limitation. As is the case in ordinary usage in the field, data structures and formats described with reference to uses salient to a human need not be presented in a human-intelligible format to constitute the described data structure or format, e.g., text need not be rendered or even encoded in Unicode or ASCII to constitute text; images, maps, and data-visualizations need not be displayed or decoded to constitute images, maps, and data-visualizations, respectively; speech, music, and other audio need not be emitted through a speaker or decoded to constitute speech, music, or other audio, respectively. Computer implemented instructions, commands, and the like are not limited to executable code and can be implemented in the form of data that causes functionality to be invoked, e.g., in the form of arguments of a function or API call. To the extent bespoke noun phrases (and other coined terms) are used in the claims and lack a self-evident construction, the definition of such phrases may be recited in the claim itself, in which case, the use of such bespoke noun phrases should not be taken as invitation to impart additional limitations by looking to the specification or extrinsic evidence.
In this patent, to the extent any U.S. patents, U.S. patent applications, or other materials (e.g., articles) have been incorporated by reference, the text of such materials is only incorporated by reference to the extent that no conflict exists between such material and the statements and drawings set forth herein. In the event of such conflict, the text of the present document governs, and terms in this document should not be given a narrower reading in virtue of the way in which those terms are used in other materials incorporated by reference.
Grouped, numerated embodiments are listed below by way of example. Reference to prior characterizations of embodiments within are within each group.
1. An object motion sensitive circuit comprising: a photodetector, wherein the photodetector detects light intensity; and a thresholding circuit, wherein the thresholding circuit determines if a magnitude of a difference in the light intensity detected by the photodetector between a first occurrence and a second occurrence exceeds a threshold; and wherein the thresholding circuit outputs a signal corresponding to the determination to a processor or storage.
2. The circuit of embodiment 1, wherein the photodetector is a photodiode.
3. The circuit of embodiment 1, further comprising a sensor, wherein the sensor comprises the photodetector.
4. The circuit of embodiment 3, wherein the sensor is an active pixel sensor.
5. The circuit of embodiment 4, wherein the active pixel sensor is a three-transistor sensor.
6. The circuit of embodiment 4, further comprising a sampling circuit, wherein the sampling circuit determines the difference in the light intensity detected by the photodetector between a first time and a second time.
7. The circuit of embodiment 3, wherein the sensor is a dynamic vision sensor.
8. The circuit of embodiment 7, wherein the dynamic vision sensor comprises a logarithmic photoreceptor.
9. The circuit of embodiment 7, further comprising a difference amplifier, wherein the difference amplifier determines the difference in the light intensity detected by the photodetector between a first time and a second time.
10. The circuit of embodiment 1, further comprising a buffer, wherein the buffer isolates the photodetector from feedback of the thresholding circuit.
11. The circuit of embodiment 1, wherein the signal output by the circuit is a bipolar spike.
12. The circuit of embodiment 1, wherein the signal output by the circuit is determined at times controlled by a timing control signal.
13. The circuit of embodiment 1, wherein the signal output by the circuit is determined asynchronously.
14. The circuit of embodiment 1, further comprising signal conditioning circuitry.
15. The circuit of embodiment 1, further comprising 2T or 3T NVM circuity, wherein the thresholding circuit comprises 2T or 3T NVM circuitry or is in series with 2T or 3T NVM circuitry, and wherein the thresholding circuit determines if a magnitude of a difference in the light intensity detected by the photodetector between a first occurrence and a second occurrence exceeds a threshold based on the 2T or 3T NVM circuitry.
16. The circuit of embodiment 1, wherein the first occurrence is at a first time and wherein the second occurrence is at a second time, different from the first time.
17. The circuit of embodiment 1, wherein the photodetector and the thresholding circuit are homogeneously integrated.
18. The circuit of embodiment 1, wherein the photodetector and the thresholding circuit are heterogeneously integrated.
19. The circuit of any one of embodiments 1 to 18, wherein the circuit is distributed over multiple die.
20. The circuit of any one of embodiments 1 to 18, wherein the circuit comprises multiple dies integrated by three-dimensional integration.
21. A sensor array comprising:
multiple of any one of the circuits of embodiments 1 to 19.
22. The sensor array of embodiment 21, further comprising:
wherein a set of the photodetectors of the multiple circuits correspond to a center region and wherein a set of photodetectors of the multiple circuits correspond to a surround region;
a integration circuit,
multiple photodetectors,
multiple transistors,
a capacitor,
a comparator circuit,
This application claims the benefit of U.S. Provisional Patent Application 63/395,725, titled IRIS: INTEGRATED RETINAL FUNCTIONALITY IN IMAGE SENSORS, filed 5 Aug. 2022. The entire contents of each afore-mentioned patent filing is hereby incorporated by reference.
| Number | Date | Country | |
|---|---|---|---|
| 63395725 | Aug 2022 | US |
| Number | Date | Country | |
|---|---|---|---|
| Parent | PCT/US2023/071788 | Aug 2023 | WO |
| Child | 19046353 | US |