The present disclosure generally relates to implementation of sensor devices, and specifically relates a sensor system comprising a plurality of stacked sensor layers that can be part of an artificial reality system.
Artificial reality systems such as head-mounted display (HMD) systems employ complex sensor devices (cameras) for capturing features of objects in a surrounding area in order to provide satisfactory user experience. A limited number of conventional sensor devices can be implemented in an HMD system and utilized for, e.g., eye tracking, hand tracking, body tracking, scanning of a surrounding area with a wide field-of-view, etc. Most of the time, the conventional sensor devices capture a large amount of information from the surrounding area. Due to processing a large amount of data, the conventional sensor devices can be easily saturated negatively affecting processing speed. Furthermore, the conventional sensor devices employed in artificial reality systems dissipate a large amount of power while having a prohibitively large latency due to performing computationally intensive operations.
A sensor assembly for determining one or more features of a local area surrounding some or all of the sensor assembly is presented herein. The sensor assembly includes a plurality of stacked sensor layers, i.e., sensor layers stacked on top of each other. A first sensor layer of the plurality of stacked sensor layers located on top of the sensor assembly can be implemented as a photodetector layer and includes an array of pixels. The top sensor layer can be configured to capture one or more images of light reflected from one or more objects in the local area. The sensor assembly further includes one or more sensor layers located beneath the photodetector layer. The one or more sensor layers can be configured to process data related to the captured one or more images for determining the one or more features of the local area, e.g., depth information for the one or more objects, an image classifier, etc.
A head-mounted display (HMD) can further integrate a plurality of sensor assemblies. The HMD displays content to a user wearing the HMD. The HMD may be part of an artificial reality system. The HMD further incudes an electronic display, at least one illumination source and an optical assembly. The electronic display is configured to emit image light. The at least one illumination source is configured to illuminate the local area with light captured by at least one sensor assembly of the plurality of sensor assemblies. The optical assembly is configured to direct the image light to an eye box of the HMD corresponding to a location of a user's eye. The image light may comprise depth information for the local area determined by the at least one sensor assembly based in part on the processed data related to the captured one or more images.
In some embodiments, a sensor apparatus includes a first sensor layer and one or more semiconductor layers that, together with the first sensor layer, form a stack. The first sensor layer includes an array of pixels. The one or more semiconductor layers are located beneath the first sensor layer and include a machine learning (ML) model accelerator configured to implement a convolutional neural network (CNN) model that processes pixel data output by the array of pixels, the pixel data corresponding to one or more frames. The one or more semiconductor layers further include a first memory, a second memory, and a controller. The first memory is configured to store coefficients of the CNN model and instruction codes. The second memory is configured to store the pixel data. The controller is configured to execute the instruction codes to control operations of the ML model accelerator, the first memory, and the second memory.
The figures depict examples of the present disclosure for purposes of illustration only. One skilled in the art will readily recognize from the following description that alternative examples of the structures and methods illustrated herein may be employed without departing from the principles, or benefits touted, of the disclosure described herein.
The disclosed techniques may include or be implemented in conjunction with an artificial reality system. Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, haptic feedback, or some combination thereof, any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Additionally, in some examples, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, e.g., create content in an artificial reality and/or are otherwise used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head-mounted display (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
A stacked sensor system for determining various features of an environment is presented herein, which may be integrated into an artificial reality system. The stacked sensor system includes a plurality of stacked sensor layers. Each sensor layer of the plurality of stacked sensor layers may represent a signal processing layer for performing a specific signal processing function. Analog sensor data related to intensities of light reflected from the environment can be captured by a photodetector layer located on top of the stacked sensor system. The captured analog sensor data can be converted from analog domain to digital domain, e.g., via an analog-to-digital conversion (ADC) layer located beneath the photodetector layer. The digital sensor data can be then provided to at least one signal processing layer of the stacked sensor system located beneath the ADC layer. The at least one signal processing layer would process the digital sensor data to determine one or more features of the environment.
In some examples, a plurality of stacked sensor systems is integrated into an HMD. The stacked sensor systems (e.g., sensor devices) may capture data describing various features of an environment, including depth information of a local area surrounding some or all of the HMD. The HMD displays content to a user wearing the HMD. The HMD may be part of an artificial reality system. The HMD further includes an electronic display and an optical assembly. The electronic display is configured to emit image light. The optical assembly is configured to direct the image light to an eye box of the HMD corresponding to a location of a user's eye. The image light may comprise the depth information for the local area determined by at least one of the plurality of stacked sensor systems.
In some other examples, a plurality of stacked sensor systems can be integrated into an eyeglass-type platform representing a NED. The NED may be part of an artificial reality system. The NED presents media to a user. Examples of media presented by the NED include one or more images, video, audio, or some combination thereof. The NED further includes an electronic display and an optical assembly. The electronic display is configured to emit image light. The optical assembly is configured to direct the image light to an eye box of the NED corresponding to a location of a user's eye. The image light may comprise the depth information for the local area determined by at least one of the plurality of stacked sensor systems.
The front rigid body 105 includes one or more electronic display elements (not shown in
The HMD 100 includes a distributed network of sensor devices (cameras) 130, which may be embedded into the front rigid body 105. Note that, although not shown in
Note that it would be impractical for each sensor device 130 in the distributed network of sensor devices 130 to have its own direct link (bus) to a central processing unit (CPU) or a controller 135 embedded into the HMD 100. Instead, each individual sensor device 130 may be coupled to the controller 135 via a shared bus (not shown in
Note that it is not required to always keep active (i.e., turned on) all the sensor devices 130 embedded into the HMD 100. In some examples, the controller 135 is configured to dynamically activate a first subset of the sensor devices 130 and deactivate a second subset of the sensor devices 130, e.g., based on a specific situation. In one or more examples, depending on a particular simulation running on the HMD 100, the controller 135 may deactivate a certain portion of the sensor devices 130. For example, after locating a preferred part of an environment for scanning, specific sensor devices 130 can remain active, whereas other sensor devices 130 can be deactivated in order to save power dissipated by the distributed network of sensor devices 130.
A sensor device 130 or a group of sensor devices 130 can, e.g., track, during a time period, one or more moving objects and specific features related to the one or more moving objects. The features related to the moving objects obtained during the time period may be then passed to another sensor device 130 or another group of sensor devices 130 for continuous tracking during a following time period, e.g., based on instructions from the controller 135. For example, the HMD 100 may use the extracted features in the scene as a “land marker” for user localization and head pose tracking in a three-dimensional world. A feature associated with a user's head may be extracted by, e.g., one sensor device 130 at a time instant. In a next time instant, the user's head may move and another sensor device 130 may be activated to locate the same feature for performing head tracking. The controller 135 may be configured to predict which new sensor device 130 could potentially capture the same feature of a moving object (e.g., the user's head). In one or more examples, the controller 135 may utilize the IMU data obtained by the IMU 120 to perform coarse prediction. In this scenario, information about the tracked feature may be passed from one sensor device 130 to another sensor device 130, e.g., based on the coarse prediction. A number of active sensor devices 130 may be dynamically adjusted (e.g., based on instructions from the controller 135) in accordance with a specific task performed at a particular time instant. Furthermore, one sensor device 130 can perform an extraction of a particular feature of an environment and provide extracted feature data to the controller 135 for further processing and passing to another sensor device 130. Thus, each sensor device 130 in the distributed network of sensor devices 130 may process a limited amount of data. In contrast, conventional sensor devices integrated into an HMD system typically perform continuous processing of large amounts of data, which consumes much more power.
In some examples, each sensor device 130 integrated into the HMD 100 can be configured for a specific type of processing. For example, at least one sensor device 130 can be customized for tracking various features of an environment, e.g., determining sharp corners, hand tracking, etc. Furthermore, each sensor device 130 can be customized to detect one or more particular landmark features, while ignoring other features. In some examples, each sensor device 130 can perform early processing that provides information about a particular feature, e.g., coordinates of a feature and feature description. To support the early processing, certain processing circuitry may be incorporated into the sensor device 130, as discussed in more detail in conjunction with
In an embodiment, a sensor device 130 can include an array of 100×100 pixels or an array of 200×200 pixels coupled to processing circuitry customized for extracting of, e.g., up to 10 features of an environment surrounding some or all of the HMD 100. In another embodiment, processing circuitry of a sensor device 130 can be customized to operate as a neural network trained to track, e.g., up to 20 joint locations of a user's hand, which may be required for performing accurate hand tracking. In yet other embodiment, at least one sensor device 130 can be employed for face tracking where, e.g., a user's mouth and facial movements can be captured. In this case, the at least one sensor device 130 can be oriented downward to facilitate tracking of user's facial features.
Note that each sensor device 130 integrated into the HMD 100 may provide a level of signal-to-noise ratio (SNR) above a threshold level defined for that sensor device 130. Because a sensor device 130 is customized for a particular task, sensitivity of the customized sensor device 130 can be improved in comparison with conventional cameras. Also note that the distributed network of sensor devices 130 is a redundant system and it is possible to select (e.g., by the controller 135) a sensor device 130 of the distributed network that produces a preferred level of SNR. In this manner, tracking accuracy and robustness of the distributed network of sensor devices 130 can be greatly improved. Each sensor device 130 may be also configured to operate in an extended wavelength range, e.g., in the infrared and/or visible spectrum.
In some examples, a sensor device 130 includes a photodetector layer with an array of silicon-based photodiodes. In alternate examples, a photodetector layer of a sensor device 130 can be implemented using a material and technology that is not silicon based, which may provide improved sensitivity and wavelength range. In one embodiment, a photodetector layer of a sensor device 130 is based on an organic photonic film (OPF) photodetector material suitable for capturing light having wavelengths larger than 1000 nm. In another embodiment, a photodetector layer of a sensor device 130 is based on Quantum Dot (QD) photodetector material. A QD-based sensor device 130 can be suitable for, e.g., integration into AR systems and applications related to outdoor environments at low visibility (e.g., at night). Available ambient light is then mostly located in the long wavelength non-visible range between, e.g., approximately 1 μm and 2.5 μm, i.e., in the short wave infrared range. The photodetector layer of the sensor device 130 implemented based on an optimized QD film can detect both visible and short wave infrared light, whereas the silicon based film may be sensitive only to wavelengths of light around approximately 1.1 μm.
In some examples, the controller 135 embedded into the front rigid body 105 and coupled to the sensor devices 130 of the distributed sensor network is configured to combine captured information from the sensor devices 130. The controller 135 may be configured to properly integrate data associated with different features collected by different sensor devices 130. In some examples, the controller 135 determines depth information for one or more objects in a local area surrounding some or all of the HMD 100, based on the data captured by one or more of the sensor devices 130.
The electronic display 155 emits image light toward the optical assembly 160. In various examples, the electronic display 155 may comprise a single electronic display or multiple electronic displays (e.g., a display for each eye of a user). Examples of the electronic display 155 include: a liquid crystal display (LCD), an organic light emitting diode (OLED) display, an inorganic light emitting diode (ILED) display, an active-matrix organic light-emitting diode (AMOLED) display, a transparent organic light emitting diode (TOLED) display, some other display, a projector, or some combination thereof. The electronic display 155 may also include an aperture, a Fresnel lens, a convex lens, a concave lens, a diffractive element, a waveguide, a filter, a polarizer, a diffuser, a fiber taper, a reflective surface, a polarizing reflective surface, or any other suitable optical element that affects the image light emitted from the electronic display 155. In some examples, the electronic display 155 may have one or more coatings, such as anti-reflective coatings.
The optical assembly 160 receives image light emitted from the electronic display 155 and directs the image light to the eye box 165 of the user's eye 170. The optical assembly 160 also magnifies the received image light, corrects optical aberrations associated with the image light, and the corrected image light is presented to a user of the HMD 100. In some examples, the optical assembly 160 includes a collimation element (lens) for collimating beams of image light emitted from the electronic display 155. At least one optical element of the optical assembly 160 may be an aperture, a Fresnel lens, a refractive lens, a reflective surface, a diffractive element, a waveguide, a filter, or any other suitable optical element that affects image light emitted from the electronic display 155. Moreover, the optical assembly 160 may include combinations of different optical elements. In some examples, one or more of the optical elements in the optical assembly 160 may have one or more coatings, such as anti-reflective coatings, dichroic coatings, etc. Magnification of the image light by the optical assembly 160 allows elements of the electronic display 155 to be physically smaller, weigh less, and consume less power than larger displays. Additionally, magnification may increase a field-of-view of the displayed media. For example, the field-of-view of the displayed media is such that the displayed media is presented using almost all (e.g., 110 degrees diagonal), and in some cases all, of the user's field-of-view. In some examples, the optical assembly 160 is designed so its effective focal length is larger than the spacing to the electronic display 155, which magnifies the image light projected by the electronic display 155. Additionally, in some examples, the amount of magnification may be adjusted by adding or removing optical elements.
In some examples, the front rigid body 105 further comprises an eye tracking system (not shown in
In some examples, the front rigid body 105 further comprises a varifocal module (not shown in
Stacking of multiple sensor layers (wafers) as shown in
In some examples, by employing wafer scaling, the sensor assembly 200 of a small size can be efficiently implemented. For example, a wafer of the photodetector layer 205 can be implemented using, e.g., 45 nm process technology, whereas a wafer of the processing layer 220 can be implemented using more advanced process technology, e.g., 28 nm or smaller process technology. Since a transistor in the 28 nm process technology occupies a very small area, a large number of transistors can be fit into a small area of the processing layer 220. In the illustrative embodiment, the sensor assembly 200 can be implemented as a cube of 1 mm×1 mm×1 mm having a power dissipation of approximately 10 mW. In comparison, conventional sensors (cameras) comprise a photodetector pixel array and processing circuitry implemented on a single silicon layer, and a total sensor area is determined as a sum of areas of all functional blocks. Without the benefit of vertical stacking as in the embodiment shown in
In some examples, a feature extraction layer 320 with processing circuitry customized for feature extraction may be placed immediately beneath the ADC layer 315. The feature extraction layer 320 may also include a memory for storing, e.g., digital sensor data generated by the ADC layer 315. The feature extraction layer 320 may be configured to extract one or more features from the digital sensor data obtained from the ADC layer 315. As the feature extraction layer 320 is customized for extracting specific features, the feature extraction layer 320 may be efficiently designed to occupy a small area size and dissipate a limited amount of power. More details about the feature extraction layer 320 are provided in conjunction with
In some examples, a convolutional neural network (CNN) layer 325 may be placed immediately beneath the feature extraction layer 320, or in the same layer as the processing circuitry of feature extraction. A neural network logic of the CNN layer 325 may be trained and optimized for particular input data, e.g., data with information about a specific feature or a set of features obtained by the feature extraction layer 320. As the input data are fully expected, the neural network logic of the CNN layer 325 may be efficiently implemented and customized for a specific type of feature extraction data, resulting into a reduced processing latency and lower power dissipation.
In some examples, the CNN layer 325 is designed to perform image classification and recognition applications. Training of the neural network logic of the CNN layer 325 may be performed offline, and network weights in the neural network logic of the CNN layer 325 may be trained prior to utilizing the CNN layer 325 for image classification and recognition. In one or more examples, the CNN layer 325 is implemented to perform inference, i.e., to apply the trained network weights to an input image to determine an output, e.g., an image classifier. In contrast to designing a generic CNN architecture, the CNN layer 325 may be implemented as a custom and dedicated neural network, and can be designed for a preferred level of power dissipation, area size and efficiency (computational speed).
An image to be classified, such as input image 327a, may be represented by a matrix of pixel values. Input image 327a may include multiple channels, each channel representing a certain component of the image. For example, an image from a digital camera may have a red channel, a green channel, and a blue channel. Each channel may be represented by a 2-D matrix of pixels having pixel values in the range of 0 to 255 (i.e., 8-bit). A gray-scale image may have only one channel. In the following description, the processing of a single image channel using CNN 326 is described. Other channels may be processed similarly.
As shown in
Matrix 327c may be processed by a second convolution layer 328c using a second weights array (labelled [W1 in
Matrix 327e can then be passed through a fully-connected layer 328e, which can include a multi-layer perceptron (MLP). Fully-connected layer 328e can perform a classification operation based on matrix 327e (e.g., to classify whether the object in image 327a represents a hand). Fully-connected layer 328e can also multiply matrix 327e with a third weights array (labelled [W2] in
CNN 326 can be implemented in CNN layer 325 using various techniques. For example, as to be described below, CNN 326 can be implemented in machine learning hardware accelerator supported with in-memory compute. The in-memory compute can include, for example, performing read/write operations at a memory to perform matrix transpose, reshaping, etc. In some examples, the in-memory compute can include matrix multiplication which can be performed by an array of memristors as to be described in
In some examples, each sensor layer 305, 315, 320, 325 in the sensor assembly 300 customized for a particular processing task can be implemented using silicon-based technology. Alternatively, at least one of the sensor layers 305, 315, 320, 325 may be implemented based on a non-silicon photo-detection material, e.g., OPF photo-detection material and/or QD photo-detection material. In some examples, instead of the silicon-based photodetector layer 305 that includes the array of pixels 310 based on photodiodes, a non-silicon photodetector layer 330 can be placed on top of the sensor assembly 300. In one embodiment, the non-silicon photodetector layer 330 is implemented as a photodetector layer of QD photo-detection material, and can be referred to as a QD photodetector layer. In some examples, the non-silicon photodetector layer 330 is implemented as a photodetector layer of OPF photo-detection material, and can be referred to as an OPF photodetector layer. In some examples, more than one photodetector layer can be used for photo detection in the sensor assembly 300, e.g., at least one silicon-based photodetector layer 305 and at least one non-silicon based photodetector layer 330.
In some examples, a direct copper bonding can be used for inter-layer coupling between the photodetector layer 305 and the ADC layer 315. As shown in
In some examples, as discussed, interconnection between sensor layers located in the sensor assembly 300 beneath the photodetector layer 305 can be achieved using, e.g., TSV technology. Referring back to
In some examples, an optical assembly 350 may be positioned on top of the silicon-based photodetector layer 305 (or the non-silicon based photodetector layer 330). The optical assembly 350 may be configured to direct at least a portion of light reflected from one or more objects in a local area surrounding the sensor assembly 300 to the pixels 310 of the silicon-based photodetector layer 305 (or sensor elements of the non-silicon based photodetector layer 330). In some examples, the optical assembly 350 can be implemented by stacking one or more layers of wafers (not shown in
In some examples, all glass wafers of the optical assembly 350 and all silicon wafers of the sensor layers 305, 315, 320, 325 can be manufactured and stacked together before each individual sensor-lens unit is diced from a wafer stack to obtain one instantiation of the sensor assembly 300. Once the manufacturing is finished, each cube obtained from the wafer stack becomes a complete, fully functional camera, e.g., the sensor assembly 300 of
In some examples, when the non-silicon based photodetector layer 330 (e.g., QD photodetector layer or OPF photodetector layer) is part of the sensor assembly 300, the non-silicon based photodetector layer 330 may be directly coupled to the ADC layer 315. Electrical connections between sensor elements (pixels) in the non-silicon based photodetector layer 330 and the ADC layer 315 may be made as copper pads. In this case, the non-silicon based photodetector layer 330 can be deposited on the ADC layer 315 after all the other sensor layers 315, 320, 325 are stacked. After the non-silicon based photodetector layer 330 is deposited on the ADC layer 315, the optical assembly 350 is applied on top of the non-silicon based photodetector layer 330.
The sensor circuitry 405 may acquire and pre-process sensor data, before providing the acquired sensor data to the feature extraction circuitry 410, e.g., via a TSV interface. The sensor data may correspond to an image captured by a two-dimensional array of pixels 415, e.g., M×N array of digital pixels, where M and N are integers of same or different values. Note that the two-dimensional array of pixels 415 may be part of the photodetector layer 305 of the sensor assembly 300 of
The feature extraction circuitry 410 may determine one or more features from the captured image represented by the pixel data 420. In the illustrative embodiment of
It should be understood that the sensor architecture 400 shown in
In some examples, the neural network 500 may be optimized for neuromorphic computing having a memristor crossbar suitable for performing vector-matrix multiplication. Learning in the neural network 500 is represented in accordance with a set of parameters that include values of conductance G=Gn,m (n=1, 2, . . . , N; m=1, 2, . . . , M) and resistance RS (e.g., vector of M resistance values rS) at cross-bar points of the neural network 500. An op-amp 502 and its associated resistor rS serves as an output driver and a column-wise weighting coefficient of each column of memristor elements, respectively.
In some examples, instead of fetching parameters from, e.g., a dynamic random-access memory (DRAM), the parameters in the form of conductance and resistance values are directly available at the cross-bar points of the neural network 500 and can be directly used during computation, e.g., during the vector-matrix multiplication. The neural network 500 based on the memristor crossbar shown in
Initial weights of the neural network 500, Gn,m, can be written via an input 505 with values organized in, e.g., N rows and M columns, which may represent a matrix input. In one or more examples, the matrix input 505 may correspond to a kernel for a convolution operation. In some examples, an input 510 may correspond to digital pixel values of an image, e.g., captured by the photodetector layer 305 and processed by ADC layer 315 and the feature extraction layer 320 of the sensor assembly 300 of
In some examples, the neural network 500 can be efficiently interfaced with the photodetector layer 305, the ADC layer 315 and the feature extraction layer 320 of the sensor assembly 300 of
After receiving the key-point map 620, the sensor system 605 may activate a portion of pixels, e.g., that correspond to a vicinity of the predicted feature(s). The sensor system 605 would then capture and process only those intensities of light related to the activated portion of pixels. By activating only the portion of pixels and processing only a portion of intensity values captured by the activated portion of pixels, power dissipated by the sensor system 605 can be reduced. The sensor system 605 may derive one or more updated locations of the one or more key features. The sensor system 605 may then send the one or more updated locations of the one or more key features to the host system 610 as an updated key-point map 625 at an increased rate of, e.g., 100 frames per second since the updated key-point map 625 includes less data than the full resolution key-frame 615. The host system 610 may then process the updated key-point map 625 having a reduced amount of data in comparison with the full resolution key-frame 615, which provides saving in power dissipated at the host system 610 while a computational latency at the host system 610 is also decreased. In this manner, the sensor system 605 and the host system 610 form the host-sensor closed loop system 600 with predictive sparse capture. The host-sensor closed loop system 600 provides power savings at both the sensor system 605 and the host system 610 with an increased communication rate between the sensor system 605 and the host system 610.
In the following, example techniques are provided which can 1) reduce the computing and memory power of sensor system 605; 2) improve privacy and security of image data generated by sensor system 605; and 3) customize the machine learning system (e.g., CNN 360) at sensor system 605.
In the following, example techniques are provided which can 1) reduce the computing and memory power of sensor system 605; 2) improve privacy and security of image data generated by sensor system 605; and 3) customize sensor system 605 for different users.
In
To reduce power consumption, MCU 704 can control various components of semiconductor layer 702 to perform sequence of operations 700 of
In addition, based on their different power consumption behaviors, different types of memories can be used to store different types of data to reduce the power consumption by the memory system. Specifically, during compute, a read from a non-volatile memory (NVM) can use a much higher power than static random access memory (SRAM), but during sleep state, the retention power of SRAM is much higher than those of NVM. Examples of NVM include magnetoresistive random access memory (MRAM), resistive random-access memory (RRAM) which can include memristors such as those illustrated in
Various techniques can be used to reduce the memory and computation power involved in the storage of sparse weight matrix 740. As shown in the top right of
In addition to using different bit lengths/voltages to represent zero and non-zero entries, other techniques can be used to further reduce power involved in transmission of sparse weight matrix 740 between memory 742 and DSP and ML accelerator 706. For example, as shown in
In addition, DSP and ML accelerator 706 may implement techniques to reduce computation power in performing computations with sparse weight matrix 740. For example, as shown in
Besides providing a ML accelerator to perform feature extraction, in-memory compute can also provide other image processing capabilities to facilitate control of sensor system 605, such as embedded matching, layer pre-processing, and depth-wise convolution layers.
Specifically, in-memory compute can provide embedded matching functionalities, such as computing distances between an input vector with a reference vector in a vector database provided by the in-memory compute, and to look up the closest match. The matching can be used to perform a similarity search for an input vector to augment, or to replace, the feature extraction capabilities provided by CNN, to support various applications such as simultaneous localization and mapping (SLAM), sentence/image infrastructure service, etc. The distance being computed can be of L0 distance, L1 distance, L2 distance, etc.
Example structures of MSB bit cells 772 (e.g., MSB bit cell 772a) and LSB bit cells 774 (e.g., LSB bit cell 774a) are illustrated in
The distance computation/similarity search operation for an input vector can be performed in two phases by controller 788 together with computing/matching logic 782, row peripheral 784, and column peripheral 786. In the first phase, a search can be performed to identify reference vectors having MSBs matching the input. To perform the search, column peripheral 786 can drive search data lines 776 based on the MSBs of the input vector, and the sl and sl_bar signals of each MSB cell, which stores a MSB of a reference vector, can be driven by the MSB of the input vector. The state of the ml signal can reflect whether the MSB of the reference vector (stored in the MSB bit cell) and the MSB of the input vector matches. Controller 788 can detect the state of the ml signals of the MSB bit cells via output data line 778 and identify reference vectors having the same MSBs as the input vector.
In a second phase, based on identifying which of the reference vectors have the same MSBs as the input vector, controller 788 can turn on assert control line 780 of LSB cells that belong to the matching reference vectors to perform distance/similarity compute. Row peripheral 784 can include bit cells having similar structure as the LSB cells shown in
In addition, in-memory compute can support layer pre-processing and depth-wise convolution layers. For example, in-memory compute can support pre-processing operation on images, such as image filtering, low-level vision, etc., with programmable and small set of kernels. In addition, in-memory compute can support depth-wise convolution layers, in which image data of each input channel (e.g., R, G, and B) convolve with a kernel of the corresponding input channel to generate immediate data for each input channel, followed by a pointwise convolution to combine the intermediate data into convolution output data for one output channel.
In some examples, the gating model can include a user-specific model 806 and a base model 808. User-specific model 806 can be different between different sensor systems 605 (e.g., on the same HMD platform used to capture different scenes, on different HMD platforms operated by different users, etc.), whereas base model 808 can be common between different sensor systems 605. For example, base model 808 can reflect a general distribution of pixels of interest in a scene under a particular operating condition, whereas user-specific model 806 can reflect the actual distribution of pixels in a scene captured by a specific sensor system 605. User-specific model 806 can be applied to pixel values in a frame 802 to compute an importance matrix 810 for each pixel in the frame. Importance matrix 810 can indicate, for example, regions of interests in frame 802. Base model 808 can then be applied to the regions of interests in frame 802 indicated by importance matrix 810 to select the pixel values input to DSP and ML accelerator 706. Base model 808 can include different gating functions to select different subsets of pixels for different channels. Both user-specific model 806 and base model 808 can change between frames, so that different subsets of pixels can be selected in different frames (e.g., to account for movement of an object).
Both user-specific model 806 and base model 808 can be generated from various sources, such as via statistical analysis, training, etc. For example, through a statistical analysis of pixel values of frames captured in different operating conditions, the probability of each pixel carrying useful information for a certain application can be determined, and the models can be determined based on the probabilities. As another example, both user-specific model 806 and base model 808 can be trained, using training data, to learn about which subset of pixels likely to include useful information for the application, and to provide those pixels to DSP and ML accelerator 706.
The gating scheme in
In addition, to reduce the updating of weights array in first memory 710 of
To improve privacy and data security, sensor system 605 can implement an encryption mechanism to encrypt the pixel data stored in the frame buffer, as well as other outputs (e.g., key frames). The encryption can be based on random numbers, which can be generated using NVMs of sensor system 605 (e.g., first memory 710 of
As described above, CNN 360 is trained to perform feature extractions to facilitate control of sensor system 605. The training operation can be used to customize the CNN for different sensor systems 605. The customization can be user-specific, application-specific, scene-specific, etc.
At each mobile platform, an in-situ training operation (e.g., in-situ training operations 1010a, 1010b, etc.) can be performed to further customize CNN 360. The customization can be user-specific (e.g., to detect a hand of a particular user), application-specific, scene-specific (e.g., to detect a particular set of objects in a particular scene), etc. The in-situ training operation can generate, from base model parameters 1004, customized model parameters 1014, which can then be stored in first memory 710 to support feature extraction operations at mobile platforms 1006 and 1008.
In-situ training operation 1010 can include different types of learning operations, such as a supervised learning operation, an unsupervised learning operation, and a reinforcement learning operation. Specifically, in a supervised learning operation, the user can provide labelled image data captured locally by sensor system 605 to train CNN 360 at sensor system 605. In an unsupervised learning operation, sensor system 605 can train CNN 360 to classify the pixel data into different groups by determining the similarity (e.g., based on cosine distance, Euclidean distance, etc.) between pixel values of the image data, which are not labelled. In a reinforcement learning operation, sensor system 605 can learn and adjust the weights of CNN 360 based on interaction with the environment at different times to maximize a reward. The reward can be based on a goal of detection. For example, in a case where the goal is to find a region of pixels corresponding to a hand, the reward can be measured by a number of pixels having target features of a hand within the region in a frame. The weights of CNN 360 can then be updated in a subsequent reinforcement learning operation on a different frame to increase the number of hand pixels within the region. The rules for updating the weights in a reinforcement learning operation can be stochastic. For example, the outputs of CNN 360 can be compared with thresholds generated from random numbers to compute the reward. Both the unsupervised learning operation and reinforcement learning operation can run in the background without requiring user input.
In-situ training operation 1010 can be customized for different use cases. In one example, a transferring learning operation can be performed, in which the weights of the lower layers (obtained from ex-situ training) are frozen, and only the weights of the upper layers are adjusted. For example, referring to
In some examples, ex-situ training operation 1000 and in-situ training operation operations 1010 can be performed in a federated/collaborative learning scheme, in which CNN 360 is trained across multiple decentralized worker machines/platforms holding local pixel data samples without exchanging their data samples. For example, in
In some examples, the reinforcement learning and unsupervised learning operations can be performed using an array of memristors, such as the one shown in
In addition, unsupervised learning operations can also be performed using an array of memristors, such as the one shown in
In some examples, the neural network can be trained by exploiting the Spike-timing-dependent plasticity (STDP), which is an example of bio-inspired algorithm that enables unsupervised learning. The assumption underlying STDP is that when the presynaptic neuron spikes just before the postsynaptic neuron spikes, the synapse/weight between the two becomes stronger, and vice-versa. Therefore, if the presynaptic neuron spikes again, the synapse will allow the postsynaptic neuron to spike faster or with a higher occurrence probability.
In sensor system 605, the input and output spike can correspond to an event at a pixel. The event can correspond to, for example, the intensity of light received by a photodiode within the frame exposure period exceeding one or more thresholds, which can be indicated by one or more flag bits in the digital pixel cell. In some examples, a pattern of the flag bits can indicate, for example, sensor system 605 operating in a certain environment (e.g., an environment having sufficient ambient light, a target environment for an application, etc.), which can lead to sensor system 605 being woken up to process the image data captured in the environment.
An array of memristors, such as the one shown in
In some examples, array of memristors 1070 can implement other types of multiplications, such as vector-vector and vector-matrix multiplications. For example, column lines (C0, C1, C2, etc.) can carry values representing a one-dimensional vector (e.g., a 1×128 vector), whereas row lines (R0, R1, R2, etc.) can carry values representing another one dimensional vector (e.g., another 1×128 vector), and array of memristors 1070 can implement a vector-vector multiplication between the two vectors. The vector-vector multiplication can represent the computations of, for example, a fully-connected neural network layer in
The HMD 1105 is a head-mounted display that presents content to a user comprising virtual and/or augmented views of a physical, real-world environment with computer-generated elements (e.g., two-dimensional (2D) or three-dimensional (3D) images, 2D or 3D video, sound, etc.). In some examples, the presented content includes audio that is presented via an external device (e.g., speakers and/or headphones) that receives audio information from the HMD 1105, the console 1110, or both, and presents audio data based on the audio information. The HMD 1105 may comprise one or more rigid bodies, which may be rigidly or non-rigidly coupled together. A rigid coupling between rigid bodies causes the coupled rigid bodies to act as a single rigid entity. In contrast, a non-rigid coupling between rigid bodies allows the rigid bodies to move relative to each other. An example of the HMD 1105 may be the HMD 100 described above in conjunction with
The HMD 1105 includes one or more sensor assemblies 1120, an electronic display 1125, an optical assembly 1130, one or more position sensors 1135, an IMU 1140, an optional eye tracking system 1145, and an optional varifocal module 1150. Some examples of the HMD 1105 have different components than those described in conjunction with
Each sensor assembly 1120 may comprise a plurality of stacked sensor layers. A first sensor layer located on top of the plurality of stacked sensor layers may include an array of pixels configured to capture one or more images of at least a portion of light reflected from one or more objects in a local area surrounding some or all of the HMD 1105. At least one other sensor layer of the plurality of stacked sensor layers located beneath the first (top) sensor layer may be configured to process data related to the captured one or more images. The HMD 1105 or the console 1110 may dynamically activate a first subset of the sensor assemblies 1120 and deactivate a second subset of the sensor assemblies 1120 based on, e.g., an application running on the HMD 1105. Thus, at each time instant, only a portion of the sensor assemblies 1120 would be activated. In some examples, information about one or more tracked features of one or more moving objects may be passed from one sensor assembly 1120 to another sensor assembly 1120, so the other sensor assembly 1120 may continue to track the one or more features of the one or more moving objects.
In some examples, each sensor assembly 1120 may be coupled to a host, i.e., a processor (controller) of the HMD 1105 or the console 1110. The sensor assembly 1120 may be configured to send first data of a first resolution to the host using a first frame rate, the first data being associated with an image captured by the sensor assembly 1120 at a first time instant. The host may be configured to send, using the first frame rate, information about one or more features obtained based on the first data received from the sensor assembly 1120. The sensor assembly 1120 may be further configured to send second data of a second resolution lower than the first resolution to the host using a second frame rate higher than the first frame rate, the second data being associated with another image captured by the sensor assembly at a second time instant.
Each sensor assembly 1120 may include an interface connection between each pixel in the array of the top sensor layer and logic of at least one sensor layer of the one or more sensor layers located beneath the top sensor layer. At least one of the one or more sensor layers located beneath the top sensor layer of the sensor assembly 1120 may include logic configured to extract one or more features from the captured one or more images. At least one of the one or more sensor layers located beneath the top sensor layer of the sensor assembly 1120 may further include a CNN based on an array of memristors for storage of trained network weights.
At least one sensor assembly 1120 may capture data describing depth information of the local area. The at least one sensor assembly 1120 can compute the depth information using the data (e.g., based on a captured portion of a structured light pattern). Alternatively, the at least one sensor assembly 1120 can send this information to another device such as the console 1110 that can determine the depth information using the data from the sensor assembly 1120. Each of the sensor assemblies 1120 may be an embodiment of the sensor device 130 in
The electronic display 1125 displays two-dimensional or three-dimensional images to the user in accordance with data received from the console 1110. In various examples, the electronic display 1125 comprises a single electronic display or multiple electronic displays (e.g., a display for each eye of a user). Examples of the electronic display 1125 include: a LCD, an OLED display, an ILED display, an AMOLED display, a TOLED display, some other display, or some combination thereof. The electronic display 1125 may be an embodiment of the electronic display 155 in
The optical assembly 1130 magnifies image light received from the electronic display 1125, corrects optical errors associated with the image light, and presents the corrected image light to a user of the HMD 1105. The optical assembly 1130 includes a plurality of optical elements. Example optical elements included in the optical assembly 1130 include: an aperture, a Fresnel lens, a convex lens, a concave lens, a filter, a reflecting surface, or any other suitable optical element that affects image light. Moreover, the optical assembly 1130 may include combinations of different optical elements. In some examples, one or more of the optical elements in the optical assembly 1130 may have one or more coatings, such as partially reflective or anti-reflective coatings.
Magnification and focusing of the image light by the optical assembly 1130 allows the electronic display 1125 to be physically smaller, weigh less and consume less power than larger displays. Additionally, magnification may increase the field-of-view of the content presented by the electronic display 1125. For example, the field-of-view of the displayed content is such that the displayed content is presented using almost all (e.g., approximately 110 degrees diagonal), and in some cases all, of the field-of-view. Additionally in some examples, the amount of magnification may be adjusted by adding or removing optical elements.
In some examples, the optical assembly 1130 may be designed to correct one or more types of optical error. Examples of optical error include barrel or pincushion distortions, longitudinal chromatic aberrations, or transverse chromatic aberrations. Other types of optical errors may further include spherical aberrations, chromatic aberrations or errors due to the lens field curvature, astigmatisms, or any other type of optical error. In some examples, content provided to the electronic display 1125 for display is pre-distorted, and the optical assembly 1130 corrects the distortion when it receives image light from the electronic display 1125 generated based on the content. In some examples, the optical assembly 1130 is configured to direct image light emitted from the electronic display 1125 to an eye box of the HMD 1105 corresponding to a location of a user's eye. The image light may include depth information for the local area determined by at least one of the plurality of sensor assemblies 1120 based in part on the processed data. The optical assembly 1130 may be an embodiment of the optical assembly 160 in
The IMU 1140 is an electronic device that generates data indicating a position of the HMD 1105 based on measurement signals received from one or more of the position sensors 1135 and from depth information received from the at least one sensor assembly 1120. A position sensor 1135 generates one or more measurement signals in response to motion of the HMD 1105. Examples of position sensors 1135 include: one or more accelerometers, one or more gyroscopes, one or more magnetometers, another suitable type of sensor that detects motion, a type of sensor used for error correction of the IMU 1140, or some combination thereof. The position sensors 1135 may be located external to the IMU 1140, internal to the IMU 1140, or some combination thereof.
Based on the one or more measurement signals from one or more position sensors 1135, the IMU 1140 generates data indicating an estimated current position of the HMD 1105 relative to an initial position of the HMD 1105. For example, the position sensors 1135 include multiple accelerometers to measure translational motion (forward/back, up/down, left/right) and multiple gyroscopes to measure rotational motion (e.g., pitch, yaw, roll). In some examples, the position sensors 1135 may represent the position sensors 125 of
The IMU 1140 receives one or more parameters from the console 1110. The one or more parameters are used to maintain tracking of the HMD 1105. Based on a received parameter, the IMU 1140 may adjust one or more IMU parameters (e.g., sample rate). In some examples, certain parameters cause the IMU 1140 to update an initial position of the reference point so it corresponds to a next position of the reference point. Updating the initial position of the reference point as the next calibrated position of the reference point helps reduce accumulated error associated with the current position estimated the IMU 1140. The accumulated error, also referred to as drift error, causes the estimated position of the reference point to “drift” away from the actual position of the reference point over time. In some examples of the HMD 1105, the IMU 1140 may be a dedicated hardware component. In other examples, the IMU 1140 may be a software component implemented in one or more processors. In some examples, the IMU 1140 may represent a sensor device 130 of
In some examples, the eye tracking system 1145 is integrated into the HMD 1105. The eye tracking system 1145 determines eye tracking information associated with an eye of a user wearing the HMD 1105. The eye tracking information determined by the eye tracking system 1145 may comprise information about an orientation of the user's eye, i.e., information about an angle of an eye-gaze. In some examples, the eye tracking system 1145 is integrated into the optical assembly 1130. An embodiment of the eye-tracking system 1145 may comprise an illumination source and an imaging device (camera).
In some examples, the varifocal module 1150 is further integrated into the HMD 1105. The varifocal module 1150 may be coupled to the eye tracking system 1145 to obtain eye tracking information determined by the eye tracking system 1145. The varifocal module 1150 may be configured to adjust focus of one or more images displayed on the electronic display 1125, based on the determined eye tracking information obtained from the eye tracking system 1145. In this way, the varifocal module 1150 can mitigate vergence-accommodation conflict in relation to image light. The varifocal module 1150 can be interfaced (e.g., either mechanically or electrically) with at least one of the electronic display 1125, and at least one optical element of the optical assembly 1130. Then, the varifocal module 1150 may be configured to adjust focus of the one or more images displayed on the electronic display 1125 by adjusting position of at least one of the electronic display 1125 and the at least one optical element of the optical assembly 1130, based on the determined eye tracking information obtained from the eye tracking system 1145. By adjusting the position, the varifocal module 1150 varies focus of image light output from the electronic display 1125 towards the user's eye. The varifocal module 1150 may be also configured to adjust resolution of the images displayed on the electronic display 1125 by performing foveated rendering of the displayed images, based at least in part on the determined eye tracking information obtained from the eye tracking system 1145. In this case, the varifocal module 1150 provides appropriate image signals to the electronic display 1125. The varifocal module 1150 provides image signals with a maximum pixel density for the electronic display 1125 only in a foveal region of the user's eye-gaze, while providing image signals with lower pixel densities in other regions of the electronic display 1125. In one embodiment, the varifocal module 1150 may utilize the depth information obtained by the at least one sensor assembly 1120 to, e.g., generate content for presentation on the electronic display 1125.
The I/O interface 1115 is a device that allows a user to send action requests and receive responses from the console 1110. An action request is a request to perform a particular action. For example, an action request may be an instruction to start or end capture of image or video data or an instruction to perform a particular action within an application. The I/O interface 1115 may include one or more input devices. Example input devices include: a keyboard, a mouse, a game controller, or any other suitable device for receiving action requests and communicating the action requests to the console 1110. An action request received by the I/O interface 1115 is communicated to the console 1110, which performs an action corresponding to the action request. In some examples, the I/O interface 1115 includes an IMU 1140 that captures IMU data indicating an estimated position of the I/O interface 1115 relative to an initial position of the I/O interface 1115. In some examples, the I/O interface 1115 may provide haptic feedback to the user in accordance with instructions received from the console 1110. For example, haptic feedback is provided when an action request is received, or the console 1110 communicates instructions to the I/O interface 1115 causing the I/O interface 1115 to generate haptic feedback when the console 1110 performs an action.
The console 1110 provides content to the HMD 1105 for processing in accordance with information received from one or more of: the at least one sensor assembly 1120, the HMD 1105, and the I/O interface 1115. In the example shown in
The application store 1155 stores one or more applications for execution by the console 1110. An application is a group of instructions, that when executed by a processor, generates content for presentation to the user. Content generated by an application may be in response to inputs received from the user via movement of the HMD 1105 or the I/O interface 1115. Examples of applications include: gaming applications, conferencing applications, video playback applications, or other suitable applications.
The tracking module 1160 calibrates the HMD system 1100 using one or more calibration parameters and may adjust one or more calibration parameters to reduce error in determination of the position of the HMD 1105 or of the I/O interface 1115. For example, the tracking module 1160 communicates a calibration parameter to the at least one sensor assembly 1120 to adjust the focus of the at least one sensor assembly 1120 to more accurately determine positions of structured light elements captured by the at least one sensor assembly 1120. Calibration performed by the tracking module 1160 also accounts for information received from the IMU 1140 in the HMD 1105 and/or an IMU 1140 included in the I/O interface 1115. Additionally, if tracking of the HMD 1105 is lost (e.g., the at least one sensor assembly 1120 loses line of sight of at least a threshold number of structured light elements), the tracking module 1160 may re-calibrate some or all of the HMD system 1100.
The tracking module 1160 tracks movements of the HMD 1105 or of the I/O interface 1115 using information from the at least one sensor assembly 1120, the one or more position sensors 1135, the IMU 1140 or some combination thereof. For example, the tracking module 1160 determines a position of a reference point of the HMD 1105 in a mapping of a local area based on information from the HMD 1105. The tracking module 1160 may also determine positions of the reference point of the HMD 1105 or a reference point of the I/O interface 1115 using data indicating a position of the HMD 1105 from the IMU 1140 or using data indicating a position of the I/O interface 1115 from an IMU 1140 included in the I/O interface 1115, respectively. Additionally, in some examples, the tracking module 1160 may use portions of data indicating a position or the HMD 1105 from the IMU 1140 as well as representations of the local area from the at least one sensor assembly 1120 to predict a future location of the HMD 1105. The tracking module 1160 provides the estimated or predicted future position of the HMD 1105 or the I/O interface 1115 to the engine 1165.
The engine 1165 generates a 3D mapping of the local area surrounding some or all of the HMD 1105 based on information received from the HMD 1105. In some examples, the engine 1165 determines depth information for the 3D mapping of the local area based on information received from the at least one sensor assembly 1120 that is relevant for techniques used in computing depth. The engine 1165 may calculate depth information using one or more techniques in computing depth from structured light. In various examples, the engine 1165 uses the depth information to, e.g., update a model of the local area, and generate content based in part on the updated model.
The engine 1165 also executes applications within the HMD system 1100 and receives position information, acceleration information, velocity information, predicted future positions, or some combination thereof, of the HMD 1105 from the tracking module 1160. Based on the received information, the engine 1165 determines content to provide to the HMD 1105 for presentation to the user. For example, if the received information indicates that the user has looked to the left, the engine 1165 generates content for the HMD 1105 that mirrors the user's movement in a virtual environment or in an environment augmenting the local area with additional content. Additionally, the engine 1165 performs an action within an application executing on the console 1110 in response to an action request received from the I/O interface 1115 and provides feedback to the user that the action was performed. The provided feedback may be visual or audible feedback via the HMD 1105 or haptic feedback via the I/O interface 1115.
In some examples, based on the eye tracking information (e.g., orientation of the user's eye) received from the eye tracking system 1145, the engine 1165 determines resolution of the content provided to the HMD 1105 for presentation to the user on the electronic display 1125. The engine 1165 provides the content to the HMD 1105 having a maximum pixel resolution on the electronic display 1125 in a foveal region of the user's gaze, whereas the engine 1165 provides a lower pixel resolution in other regions of the electronic display 1125, thus achieving less power consumption at the HMD 1105 and saving computing cycles of the console 1110 without compromising a visual experience of the user. In some examples, the engine 1165 can further use the eye tracking information to adjust where objects are displayed on the electronic display 1125 to prevent vergence-accommodation conflict.
The foregoing description of the examples of the disclosure has been presented for the purpose of illustration; it is not intended to be exhaustive or to limit the disclosure to the precise forms disclosed. Persons skilled in the relevant art can appreciate that many modifications and variations are possible in light of the above disclosure.
Some portions of this description describe the examples of the disclosure in terms of algorithms and symbolic representations of operations on information. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, hardware, or any combinations thereof.
Any of the steps, operations, or processes described herein may be performed or implemented with one or more hardware or software modules, alone or in combination with other devices. In one embodiment, a software module is implemented with a computer program product comprising a computer-readable medium containing computer program code, which can be executed by a computer processor for performing any or all of the steps, operations, or processes described.
Examples of the disclosure may also relate to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, and/or it may comprise a general-purpose computing device selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a non-transitory, tangible computer readable storage medium, or any type of media suitable for storing electronic instructions, which may be coupled to a computer system bus. Furthermore, any computing systems referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
Examples of the disclosure may also relate to a product that is produced by a computing process described herein. Such a product may comprise information resulting from a computing process, where the information is stored on a non-transitory, tangible computer readable storage medium and may include any embodiment of a computer program product or other data combination described herein.
Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the disclosure be limited not by this detailed description, but rather by any claims that issue on an application based hereon. Accordingly, the disclosure of the examples is intended to be illustrative, but not limiting, of the scope of the disclosure, which is set forth in the following claims.
The present application is a continuation-in-part of U.S. patent application Ser. No. 16/910,844, filed on Jun. 24, 2020, entitled “SENSOR SYSTEM BASED ON STACKED SENSOR LAYERS” which is a continuation of U.S. patent application Ser. No. 15/909,162, filed on Mar. 1, 2018, now bearing U.S. Pat. No. 10,726,627, issued on Jul. 28, 2020, entitled “SENSOR SYSTEM BASED ON STACKED SENSOR LAYERS,” and which claims the benefit of and priority to U.S. Provisional Application No. 62/536,605, filed on Jul. 25, 2017, and entitled “STACKED SENSOR SYSTEM USING MEMRISTORS.” The present application also claims the benefit of and priority to U.S. Provisional Application No. 63/021,476, filed on May 7, 2020, and entitled “SMART SENSOR,” and U.S. Provisional Application No. 63/038,636, filed on Jun. 12, 2020, and entitled “SMART SENSOR.” The contents of each of the above-identified applications are hereby incorporated by reference in their entirety for all purposes.
Number | Date | Country | |
---|---|---|---|
62536605 | Jul 2017 | US | |
63021476 | May 2020 | US | |
63038636 | Jun 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15909162 | Mar 2018 | US |
Child | 16910844 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16910844 | Jun 2020 | US |
Child | 17313884 | US |