Seamless readout mode transitions in image sensors

Information

  • Patent Grant
  • 11659298
  • Patent Number
    11,659,298
  • Date Filed
    Monday, May 24, 2021
    2 years ago
  • Date Issued
    Tuesday, May 23, 2023
    11 months ago
  • Inventors
  • Original Assignees
  • Examiners
    • Prabhakher; Pritham D
    Agents
    • Brownstein Hyatt Farber Schreck, LLP
  • CPC
    • H04N25/75
  • Field of Search
    • US
    • 348 294000
    • CPC
    • H04N5/378
    • H04N5/3765
    • H04N5/3535
    • H04N5/376
    • H04N5/374
    • H04N5/3742
    • H04N5/345
    • H04N5/353
  • International Classifications
    • H04N5/378
    • H04N25/75
    • Disclaimer
      This patent is subject to a terminal disclaimer.
Abstract
An image sensor for an electronic device can adjust a scan rate of a reset pointer of a frame relative to a scan rate of a readout pointer of a previous frame. As a result, the image sensor can transition between various readout modes without interrupting image output from the image sensor.
Description
FIELD

Embodiments described herein relate to image sensors and, in particular, to image sensors configured to seamlessly transition between readout modes.


BACKGROUND

An image sensor for an electronic device typically includes an array of photosensitive electronic components conventionally referred to as “photosites.” When exposed to light, each photosite accumulates an electric charge proportional to the brightness of light illuminating that photosite for the duration of the exposure. After exposure, the image sensor reads out the accumulated electric charge of all or some of the photosites to generate an image. The photosites of a conventional image sensor are generally read out in sequence (e.g., by rows or by columns) for each image output from the image sensor.


Some image sensors support multiple photosite readout modes. Example readout modes include full-resolution readout modes, binning readout modes, oversampling readout modes, and so on. However, conventional image sensors undesirably interrupt—and/or otherwise disable—image output while transitioning between readout modes.


SUMMARY

Many embodiments described herein reference a method of operating an image sensor. The method includes the operations of: applying a readout signal to a first group of photosites defined by a readout pointer; applying a reset signal to a second group of photosites defined by a reset pointer; incrementing the readout pointer at a first scan rate; and incrementing the reset pointer at a second scan rate different from the first scan rate. In these examples, the reset signal and the readout signal are applied to the first and second group of photosites, respectively, in phase.


Some embodiments described reference an image sensor including a photosite array. The photosite array includes a first group of photo sites and a second group of photosites. The image sensor also includes a controller configured to: apply a readout signal (associated with a readout pointer) to the first group of photosites; apply a reset signal (associated with a reset pointer) in phase with the readout signal, to the second group of photosites; increment the readout pointer at a first scan rate; and increment the reset pointer at a second scan rate different from the first scan rate.





BRIEF DESCRIPTION OF THE DRAWINGS

Reference will now be made to representative embodiments illustrated in the accompanying figures. It should be understood that the following descriptions are not intended to limit this disclosure to one included embodiment. To the contrary, the disclosure provided herein is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the described embodiments, and as defined by the appended claims.



FIG. 1 depicts an example electronic device incorporating an image sensor, such as described herein.



FIG. 2 is a simplified system diagram of an image sensor configured for seamless transition between readout modes, such as described herein.



FIG. 3A is a timing diagram illustrating operations of an image sensor operating in a first readout mode, such as described herein.



FIG. 3B is a timing diagram illustrating operations of an image sensor operating in a second readout mode, such as described herein.



FIG. 4 is a timing diagram depicting a seamless transition between readout modes of an image sensor, such as described herein.



FIG. 5 is a timing diagram depicting a seamless transition between readout modes of an image sensor, such as described herein.



FIG. 6A is a detail view of the enclosed circle A-A of the timing diagram of FIG. 3A, showing synchronization between read and reset signals applied when an image sensor, such as described herein, is continuously operated in a selected readout mode.



FIG. 6B is a detail view of the enclosed circle B-B of the timing diagram of FIG. 4, showing synchronization between read and reset signals applied when an image sensor, such as described herein, is seamlessly transitioned between readout modes.



FIG. 6C is a detail view of the enclosed circle C-C of the timing diagram of FIG. 5, showing synchronization between read and reset signals applied when an image sensor, such as described herein, is seamlessly transitioned between readout modes.



FIG. 7 is a flowchart depicting example operations of a method of synchronizing a readout signal and a reset signal by an image sensor, such as described herein.



FIG. 8 is a flowchart depicting example operations of a method of correcting gain offset in a frame by an image sensor, such as described herein.





The use of the same or similar reference numerals in different figures indicates similar, related, or identical items.


Additionally, it should be understood that the proportions and dimensions (either relative or absolute) of the various features and elements (and collections and groupings thereof) and the boundaries, separations, and positional relationships presented therebetween, are provided in the accompanying figures merely to facilitate an understanding of the various embodiments described herein and, accordingly, may not necessarily be presented or illustrated to scale, and are not intended to indicate any preference or requirement for an illustrated embodiment to the exclusion of embodiments described with reference thereto.


DETAILED DESCRIPTION

Embodiments described herein reference an image sensor for an electronic device that is configured to seamlessly transition between different readout modes.


For simplicity of description, the embodiments that follow reference a complementary metal oxide semiconductor (“CMOS”) image sensor that is sensitive to light in a traditionally visible spectral range—and methods for operating the same—but it may be appreciated that this is merely one example construction and that other image sensor topologies (e.g., charge coupled devices, non-visible light image sensors, front side or back side illuminated image sensors, flat panel detectors, and so on) may be suitable in other implementations.


A CMOS image sensor, such as described herein, includes—among other components, structures, and/or elements—an array of “photosites,” referred to as a “photosite array.” Typically, the photosite array includes a set of columns and a set of rows arranged in a grid. In some implementations, the term “pixel” can be used as a synonym for the term “photosites.”


Each photosite of a photosite array includes at least one “photosensitive area.” When a photosensitive area of a photosite is exposed to light for a particular exposure duration, the photosensitive area accumulates an electric charge. This is referred to as an “integration” operation because the accumulated electric charge corresponds to the total quantity of photons received at the photosensitive area over the exposure duration (herein, the “integration time”). In other words, the electric charge accumulated by a particular photosensitive area of a particular photosite of a photosite array is proportional to the brightness of light illuminating that photosensitive area for the integration time.


After integration, the image sensor “reads out” the accumulated electric charge (as a current, as a voltage, or as a digital value) of all or some of the photosensitive areas of all or some of the photosites as an initial step in a process to generate a digital image referred to herein as a “frame.” The photosensitive areas of the photosites of a photosite array, such as described herein, are typically read out in groups. For example, in some embodiments, the photosensitive areas of the photosites of a photosite array can be grouped by rows or columns, read out in a row-wise or a column-wise sequence (e.g., read out sequentially by columns or read out sequentially by rows). In addition, after a readout operation is performed, each photosensitive area of each photosite of the photosite array is “reset” in order to discharge any residual electric charge prior to integration of the next frame.


For embodiments described herein, a CMOS image sensor (hereinafter, simply, “image sensor”) is understood to be configured to output frames at a constant or, optionally, variable “frame rate.” For each frame output from an image sensor, the image sensor typically performs at least three operations. First, the image sensor sends a “reset signal” to all or some of the photosites to reset all or some of the photosensitive areas of those photosites. Second, the image sensor waits for an integration time period to elapse. Third, the image sensor reads out the accumulated electric charge (e.g., as a voltage, as a current, or as a digital value) of all or some of the photosensitive areas of all or some of the photosites of the photosite array. The accumulated electric charge information, regardless of form or format (e.g., charge, voltage, digital value, and so on), is referred to herein for simplicity of description as “data” read out from the photosite array.


The image sensor assembles a frame based on the data read out from the photosite array. In some cases, the image sensor can apply one or more filters to, and/or perform one or more operations with, the data read out from the photosite array prior to (and/or subsequent to) assembling the frame. Examples include, but are not limited to: operations to apply one or more demosaicing/debayering filters; binning operations (e.g., combining charges accumulated by adjacent or nearby photosites); averaging operations; gain correction operations; and so on.


For simplicity of description, the embodiments that follow reference an image sensor that resets, integrates, and reads out each photosensitive area of each photosite of a photosite array in a row-wise sequence (e.g., the image sensor resets, integrates, and reads out photosensitive areas in a row then sequentially advances to the next row). It may be appreciated, however, that this is merely one example and that other embodiments can be configured to reset, integrate, and/or read out one or more photosensitive areas of one or more photosites of a photosite array in any other suitable manner.


More specifically, the embodiments that follow reference an image sensor configured to perform reset and readout operations on entire rows of photosites at a time. For example, to perform a reset operation, the image sensor can send a reset signal to each photosite in a selected row of photosites at substantially the same time in order to efficiently discharge any residual electric charge accumulated by any photosensitive area of any photosite in the selected row. The selected row can be addressed by a pointer referred to herein as the “reset pointer.” Thereafter, the next row is selected by incrementing the reset pointer and the process is repeated until all rows of the photosite array have been reset.


Similarly, to perform a readout operation after a selected integration period has elapsed, each photosensitive area of each photosite in a selected row of photosites can be read out (e.g., into temporary storage capacitors, buffers, registers, analog-to-digital converters, amplifiers, and so on) substantially simultaneously. As with the reset operation described above, the selected row can be addressed by a pointer referred to herein as the “readout pointer.” Thereafter, the readout pointer is incremented and the process is repeated until all rows have been read out.


In typical embodiments, the readout pointer of a particular frame leads the reset pointer of the next frame by a certain amount referred to herein as the “inter-frame delay.” In other words, an image sensor—such as described herein—is generally configured to perform a readout operation simultaneously with a reset operation. More specifically, the image sensor is typically configured to perform a readout operation for a first frame (herein, “frame FN”) while simultaneously performing a reset operation for a second frame (herein, “frame FN+1”).


In many embodiments, the reset pointer of frame FN+1 and the readout pointer of frame FN are synchronously incremented at a particular “scan rate.” As may be appreciated, the pointers are incremented at the same time in order to reduce and/or eliminate switching noise in the image sensor. The scan rate (e.g., the rate at which the reset pointer and the readout pointer are incremented together) can be based, at least in part, on the total time required to reset and read out a particular row of photosites. Collectively, the total time required to reset and read out a particular row of photosites is referred to herein as the “line time.”


As noted above, precise synchronization while incrementing the reset pointer of frame FN+1 and the readout pointer of frame FN may be required to reduce and/or eliminate switching noise in the image sensor. To facilitate this synchronization, embodiments described herein include a timing controller that can include a vertical synchronization controller (herein, a “VSYNC” controller). The VSYNC controller can be any suitable digital or analog circuit, including one or more discrete or analog components, application-specific integrated circuits, general-purpose processors, and so on. The VSYNC controller can, in some examples, be implemented in whole or in part in software.


The VSYNC controller is configured to set an appropriate scan rate for the reset pointer of frame FN+1 and the readout pointer of frame FN and to maintain synchronization while incrementing the two pointers. In some implementations, the VSYNC controller includes at least one asynchronous first-in first-out controller (also referred to as an “ASYNC FIFO”) that can be used to buffer or otherwise receive and store addresses of the readout pointer of frame FN and the reset pointer of frame FN+1. The rate at which addresses in the ASYNC FIFO are flushed may depend upon the scan rate maintained by the VSYNC controller. It may be appreciated, however, that this is merely one example; other circuit topologies can be used in other embodiments.


Further, as noted above, an image sensor such as described herein is typically configured to operate in multiple readout modes. Example readout modes include, but are not limited to: full-resolution readout modes (e.g., a readout mode in which a frame is generated by reading out each photosite of the image sensor); binning readout modes (e.g., a readout mode in which a frame is generated by combining charges accumulated by adjacent or nearby photosites with each other to reduce noise); photosite-skipping readout modes (e.g., a readout mode in which select photosites are skipped); oversampling readout modes (e.g., a readout mode in which more photosites are sampled than an output frame requires to increase noise tolerance); high frame-rate readout modes (e.g., a readout mode in which frame integration times are reduced to increase frame count per second); low frame-rate readout modes (e.g., a readout mode in which frame integration times are increased to decrease frame count per second); and so on.


It may be appreciated that different readout modes may be associated with different integration times, different readout times, different scan rates, different line times, and so on. For example, a binning readout mode may be associated with a different line time and scan rate than a full-resolution readout mode. As noted above, conventional image sensors undesirably interrupt frame output (e.g., inserting a vertical blanking frame) in order to transition between readout modes.


For embodiments described herein, however, an image sensor is configured to adjust the incrementation of the reset pointer of frame FN+1 relative to the incrementation of the readout pointer of frame FN to smoothly transition between readout modes in real-time (e.g., transitioning from an “original readout mode” to a “target readout mode” without interrupting frame output).


For example, in some cases, the reset pointer of frame FN+1 may selectively not increment when the readout pointer of frame FN increments. By selectively not incrementing, the reset pointer of frame FN+1 gradually lags farther behind the readout pointer of frame FN. Phrased in another non-limiting manner, the reset pointer of frame FN+1 increments through the rows in the photosite array at a slower rate (e.g., a slower scan rate) than the readout pointer of frame FN. In this manner, by controlling the number of times that the reset pointer of frame FN+1 is not incremented when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter (and after an integration period has elapsed) the scan rate of the readout pointer of frame FN+1 can be set to match the slowed scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the target readout mode. To illustrate this example, Table 1 and Table 2 are provided below.


Table 1 corresponds to an example embodiment in which a photosite array includes ten rows. In this example, the readout pointer of frame FN is incremented once—or skipped—for every time unit t. Similarly, the reset pointer of frame FN+1 is incremented once for every time unit t. The two frames FN, FN+1 are separated by an inter-frame delay of one time unit t and the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 are the same at one row per time unit t.











TABLE 1





Time
Readout pointer
Reset Pointer


(t)
FN
FN+1

















0
Row 0



1
Row 1
Row 0


2
Row 2
Row 1


3
Row 3
Row 2


4
Row 4
Row 3


5
Row 5
Row 4


6
Row 6
Row 5


7
Row 7
Row 6


8
Row 8
Row 7


9
Row 9
Row 8


10

Row 9









Table 2 corresponds to an example embodiment such as described above in which the image sensor transitions from an original readout mode to a target readout mode associated with a slower scan rate than the original readout mode.


More specifically, the reset pointer of frame FN+1 in this example increments through the ten rows of the example photosite array at a slower rate than the readout pointer of frame FN. In this example, the readout pointer of frame FN is incremented once, or skipped for every time unit t according to the scan rate of the original readout mode. In contrast, the reset pointer of frame FN+1 is incremented three times for every four time units t, according to the scan rate of the target readout mode. In this example, the two frames FN, FN+1 remain separated by an inter-frame delay of one time unit t, but the effective scan rates are different. More specifically, the readout pointer of frame FN proceeds with a scan rate of one row per time unit t (e.g., the scan rate of the original readout mode) while the reset pointer of frame FN+1 proceeds with a scan rate approximately equal to 0.75 rows per time unit t (e.g., the scan rate of the target readout mode).











TABLE 2





Time
Readout pointer
Reset Pointer


(t)
FN
FN+1

















0
Row 0



1
Row 1
Row 0


2
Row 2



3
Row 3
Row 1


4
Row 4
Row 2


5
Row 5



6
Row 6
Row 3


7
Row 7
Row 4


8
Row 8



9
Row 9
Row 5


10

Row 6


11




12

Row 7


13

Row 8


14




15

Row 9









As a result of the approach described in the preceding example, the scan rate of the reset pointer of frame FN+1 can be changed to match the scan rate of the target readout mode, while still maintaining time synchronization with the readout pointer for frame FN to prevent and/or mitigate effects of switching noise. In another non-limiting phrasing, this example configuration smoothly transitions—in a single frame—from a higher scan rate readout mode for frame FN to a lower scan rate readout mode for frame FN+1.


In another example configuration of an image sensor described herein, more than one row of photosites can be reset simultaneously. As a result of this approach, the effective scan rate of the reset pointer(s) of frame FN+1 conforms, generally, to the scan rate of the target readout mode while still maintaining time synchronization with the scan rate of the current readout mode. In another non-limiting phrasing, this example configuration transitions from a lower scan rate readout mode to a higher scan rate readout mode by increasing the pace at which the reset pointers of frame FN+1 increment relative to the pace at which the readout pointer of frame FN increments.


Table 3 corresponds to an example embodiment, such as described above, in which the image sensor transitions to a target readout mode associated with a faster scan rate than the original readout mode.


More specifically, the reset pointer of frame FN+1 in this example, increments through the ten rows of the example photosite array at a faster rate than the readout pointer of frame FN. In this example, the readout pointer of frame FN is incremented once for every time unit t according to the scan rate of the original readout mode. In contrast, the reset pointer of frame FN+1 is delayed by a certain amount (e.g., 6 time units t) and, thereafter, two rows on average are reset simultaneously for every one time unit t, according to the scan rate of the target readout mode. More specifically, the readout pointer of frame FN proceeds with a scan rate of one row per time unit t (e.g., the scan rate of the original readout mode) while the reset pointer of frame FN+1 proceeds with a scan rate approximately equal to two rows per time unit t (e.g., the scan rate of the target readout mode).











TABLE 3





Time
Readout pointer
Reset Pointer


(t)
FN
FN+1

















0
Row 0



1
Row 1



2
Row 2



3
Row 3



4
Row 4



5
Row 5



6
Row 6
Row 0, Row 1


7
Row 7
Row 2, Row 3


8
Row 8
Row 4, Row 5


9
Row 9
Row 6, Row 7


10

Row 8, Row 9









As a result of the approach described in reference to the preceding example, the scan rate of the reset pointer of frame FN+1 can be changed to match the scan rate of the target readout mode, while still maintaining time synchronization with the readout pointer for frame FN to prevent and/or mitigate effects of switching noise. In another non-limiting phrasing, this example configuration smoothly transitions—in a single frame—from a lower scan rate readout mode for frame FN to a higher scan rate readout mode for frame FN+1.


In some cases, after a transition from a first scan rate to a second scan rate is performed, line timing may also be changed (e.g., increased or decreased).


It may be appreciated that the preceding examples described in reference to Table 2 and Table 3 are not exhaustive. To the contrary, it may be appreciated that the systems, methods, and techniques described herein can be applied to effect real-time and/or smooth transitions between any number of suitable readout modes.


These foregoing and other embodiments are discussed below with reference to FIGS. 1-7. However, those skilled in the art will readily appreciate that the detailed description given herein with respect to these figures is for explanatory purposes only and should not be construed as limiting.



FIG. 1 depicts an example electronic device 100. The electronic device 100 includes a housing 102, a display 104, and an image sensor 106.


The image sensor 106 is disposed within the housing 102, and is positioned to align with—and/or extend at least partially through—an aperture defined by the housing 102. As illustrated, the aperture is circular, but this may not be required. In other cases, the aperture may not be required at all; the image sensor 106 can be positioned behind a transparent portion of the housing.


As with other embodiments described herein, the image sensor 106 includes a photosite array of individual photosites also referred to as “pixels.” Typically, a photosite array is rectangular and the photosites of the array are arranged in a regular grid pattern of columns and rows. Further, a photosite array is typically planar, although in some embodiments, a photosite array can be concave or convex or may take another shape.


Each photosite of the photosite array can include one or more of, without limitation: a photosensitive area (e.g., photodiode, charge-coupled device, and so on); a reset switch or circuit; a microlens; a hot mirror (e.g., or other band stop optical filter); a color filter (e.g., a portion of a Bayer pattern); an output amplifier (e.g., charge amplifier, voltage amplifier, trans-impedance amplifier, and so on); an analog-to-digital converter; a charge storage capacitor; digital memory; and so on.


Some image sensor embodiments can include photosites with more than one photosensitive area. In these examples, different photosensitive areas may be formed from different materials, or material combinations, and/or may have different photosensitivity or electrical characteristics (e.g., rise time, fall time, reverse bias, dark current, and so on). In another example, a first photosensitive area is positioned adjacent to a second photosensitive area of the same photosite to facilitate autofocus adjustments or parallax-based depth sensing.


For simplicity of description, the image sensor 106 is understood to include a photosite array in which each photosite includes a single photosensitive area defined by a semiconductor material (e.g., silicon, germanium, indium gallium arsenide, gallium phosphide, and so on). It is appreciated, however, that this is merely one example and, that in other implementations, a photosite can include more than one photosensitive area and/or an image sensor can include a photosite array in which some photosites include more than one photosensitive area and some photosites include only a single photosensitive area.


It may be appreciated that the foregoing description of FIG. 1, and the various alternatives thereof and variations thereto, are presented, generally, for purposes of explanation, and to facilitate a thorough understanding of various possible configurations of an electronic device incorporating an image sensor such as described herein. However, it will be apparent to one skilled in the art that some of the specific details presented herein may not be required in order to practice a particular described embodiment, or an equivalent thereof.


For example, although the electronic device 100 is depicted as a cellular phone, it may be appreciated that other electronic devices can incorporate an image sensor such as described herein including, but not limited to: tablet devices; laptop devices; wearable devices; desktop computers; computing accessories; peripheral input devices; vehicle control devices; mobile entertainment devices; augmented reality devices; virtual reality devices; industrial control devices; digital wallet devices; home security devices; business security devices; and so on.


Further, it is appreciated that an electronic device, such as the electronic device 100, can include any number of image sensors. In some examples, an electronic device can include image sensors oriented outwardly from a front surface of a housing (such as shown in FIG. 1), oriented outwardly from a rear or back surface of a housing, or oriented outwardly from a side surface of a housing. In some embodiments, an electronic device can include multiple image sensors, some of which may be conventionally configured image sensors and some of which may be image sensors such as described herein.


Further, it is appreciated that the electronic device 100—or, more generally, any portable or stationary electronic device incorporating one or more image sensors such as described herein—can additionally include, without limitation, a processor, memory, power supply and/or battery, network connections, sensors, input/output ports, acoustic elements, haptic elements, digital and/or analog circuits for performing and/or coordinating tasks, and so on. For simplicity of illustration, the electronic device 100 is depicted in FIG. 1 without many of these elements, each of which may be included, partially and/or entirely, within the housing 102 and may be operationally or functionally associated with, or coupled to, the display 104 of the electronic device 100 or the image sensor 106 of the electronic device 100.


For example, a processor of the electronic device 100 can be configured to access and execute instructions stored in a memory in order to instantiate any number of suitable classes, objects, virtual machines, threads, pipelines, and/or routines to perform, monitor, and/or coordinate one or more operations of the electronic device 100 and/or the image sensor 106. In some cases, the processor can be communicably coupled—either directly (e.g., via general-purpose input/output pins) or indirectly (e.g., via an intermediary circuit or integrated circuit)—to one or more of the hardware components of the electronic device 100, including the display 104 and the image sensor 106. In this manner, frames output by the image sensor 106 can be shown and/or rendered on the display 104 in substantially real time. In other embodiments, the processor may not be a required intermediary between the display 104 and the image sensor 106; frames output by the image sensor 106 can be received directly by the display 104.


In one embodiment, the processor can cause the display 104 to render a graphical user interface. In many cases, the graphical user interface can include a live view of frames output from the image sensor 106. As a result of this construction, a user of the electronic device 100 can observe a scene captured by the image sensor 106 in substantially real time.


In further examples, the processor of the electronic device 100 can render the graphical user interface while monitoring a user input sensor (e.g., capacitive touch sensor, capacitive force sensor, and so on) to detect one or more characteristics of a user's physical interaction with the display 104 and/or the user's physical interaction with the live view of frames output from the image sensor 106. Example detectable characteristics can include, but are not limited to, characteristics of: touch input events (e.g., location and duration); force input events (e.g., location, duration, and magnitude); touch gesture events (e.g., swipe, zoom in, zoom out, rotate, pinch); force gesture events (e.g., heavy press, light press); and so on.


As a result of this construction, a user of the electronic device 100 is encouraged to interact with content shown on the display 104, including the live view of frames output from the image sensor 106. In response to detecting specific user interactions and/or specific events (herein, generally, “user input”), the processor and/or another digital or analog circuit in the electronic device 100 can perform a specific task or operation related to one or more of the display 104, the graphical user interface, and/or the image sensor 106.


For example, in one embodiment, the graphical user interface shown on the display 104 includes a live view of frames output from the image sensor 106 operating in a particular readout mode suitable for capturing high-resolution images of a well-illuminated scene (such as a full-resolution readout mode). In some circumstances, a user of the electronic device 100 may prefer to change the readout mode of the image sensor 106 to a readout mode more suitable for capturing low-noise images of a poorly-illuminated scene (such as a binned readout mode). In this example, the user can interact with the graphical user interface (e.g., by selecting a button, by swiping in a particular direction, by applying a force beyond a threshold, and so on) to instruct the electronic device 100 to change the readout mode of the image sensor 106. In response to receiving the user input, the processor—and/or another digital or analog circuit of the electronic device 100 or the image sensor 106—can cause the image sensor 106 to change readout modes.


Accordingly, generally and broadly, it is understood that in some embodiments a graphical user interface of an electronic device can be used to trigger changes in readout modes of an image sensor, such as described herein. In other cases, however, an input sensor can change readout modes in response to other triggers, states, or events such as, but not limited to: a setting of the electronic device; output received from a remote or local application; output from an image processing algorithm; and so on.


Thus, it is understood that the foregoing descriptions of specific embodiments are presented for the limited purposes of illustration and description. These descriptions are not exhaustive nor intended to limit the disclosure to the precise forms recited herein. To the contrary, it will be apparent to one of ordinary skill in the art that many modifications and variations are possible in view of the above teachings.



FIG. 2 is a simplified system diagram of an image sensor 200, such as described herein. The image sensor 200 includes a photosite array 202 that, in turn, includes m rows and n columns of photosites.


As with other embodiments described herein, the photosites of the photosite array 202 can each include a number of discrete circuits and/or components that may vary from embodiment to embodiment. For simplicity of description and illustration, the embodiments that follow reference a single photosite positioned within the photosite array 202 at position (0, n); it may be appreciated that other photosites of the photosite array 202 may be similarly configured. In other examples, however, the photosites of the photosite array 202 may be configured in a different manner than shown and described relative to the photosite (0, n).


In the illustrated example, the photosite (0, n) includes a photosensitive element 204 (e.g., a photodiode) defining a photosensitive area that is sensitive to light in a particular spectral range. In typical examples, the photosensitive area is sensitive to light in a traditionally visible spectral range, but this may not be required of all embodiments. For example, in some cases, the photosensitive area of the photosensitive element 204 may be sensitive to infrared light.


In many examples, the photosite (0, n) also includes an amplifier 206. Output from the amplifier 206 is gated by a switch 208.


As noted with respect to other embodiments described herein, when the photosite array 202 of the image sensor 200 is exposed to light during an integration operation, electric charge is accumulated on photosensitive area(s) of the photosensitive element 204. During a readout operation, the switch 208 can be closed, connecting an output of the amplifier 206 to a readout circuit of the image sensor 200. In further examples, the photosite (0, n) can include a second switch associated with a reset operation; when closed (or in other circuit topologies, opened) the second switch can discharge any latent charge held by the photosensitive element 204.


As noted with respect to other embodiments described herein, the image sensor 200 is configured to perform readout operations and reset operations in a row-wise fashion. To facilitate this mode of operation, the image sensor 200 includes a vertical synchronization controller 210 configured to apply one or more signals (e.g., readout signals and/or reset signals) to one or more rows simultaneously, typically in phase with a clock signal (not shown). In one example, frame FN can be read out while frame FN+1 is being reset; the readout signal and the reset signals (of the next frame) are applied to rows in phase with one another. In this example, the vertical synchronization controller 210 can be configured to release and/or otherwise send a reset signal (for frame FN+1) to a first row of the set of rows at the same time a readout signal (for frame FN) is sent to a second row of the set of rows. By maintaining precise timing between the readout signals associated with frame FN and the reset signals associated with frame FN+1, switching noise can be mitigated.


The specific row (or rows) driven by the vertical synchronization controller 210 can be selected by a row decoder 212. The row decoder 212 enables a specific row trace from a set of row traces (e.g., a bus) coupling the vertical synchronization controller 210 and the row decoder 212 in response to receiving an address and/or a pointer—such as a readout pointer or a reset pointer, such as described above—from a timing controller 214.


In one specific example, the timing controller 214 can initiate a reset operation by setting the reset pointer to reference row 0. Next, the timing controller 214 sends the reset pointer to the row decoder 212. The row decoder 212 decodes the pointer and determines a row trace associated with the reset pointer. Next, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the readout operation on row zero. In one example, the vertical synchronization controller 210 buffers the received signal until an edge of a clock signal (not shown) is generated or received by the vertical synchronization controller 210.


Thereafter, and after a delay based on a scan rate (such as described above), the timing controller 214 increments the reset pointer to reference row one, which is decoded by the row decoder 212 into a specific row trace. Next, as with the previous example, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the reset operation on row one.


The image sensor 200 includes a readout circuit. The readout circuit can include, in some examples, an (optional) column decoder/encoder 216 and an (optional) analog to digital converter 218. Much like the row decoder 212, the column decoder/encoder 216 can be configured to translate physical trace locations into addresses, and vice versa. The analog to digital converter 218 is configured to convert voltage, current, or charge information received from one or more photosites into a digital representation of an appropriate or selected bit depth (e.g., 8-bit, 10-bit, 12-bit, 16-bit, and so on). Thereafter, optionally, the output from the analog to digital converter 218 can be received by a gain correction controller (not shown) that can provide for gain correction.


In one specific example, the timing controller 214 can initiate a readout operation by setting the readout pointer to reference row 0. Next, the timing controller 214 sends the readout pointer to the row decoder 212. The row decoder 212 decodes the pointer and determines a row trace associated with the readout pointer. Next, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the readout operation on row zero.


Once the vertical synchronization controller 210 applies the readout signal to row zero, the accumulated electrical charge of each photosensitive element of each photosite in row zero is converted into a digital value (either serially or in parallel) by the analog to digital converter 218. Thereafter, output from the column decoder/encoder 216 can be used to associate the digital value(s) output from the analog to digital converter 218 with specific row and column addresses.


Thereafter, and after a delay based on a scan rate such as described above, the timing controller 214 increments the readout pointer to reference row 1, and the readout operation continues.


As with other embodiments described herein, the image sensor 200 can be configured to adjust the incrementation of reset pointers relative to the incrementation of readout pointers in order to smoothly transition between readout modes in real-time. To facilitate transitions such as these, the image sensor 200 includes a transition timing controller 220. The transition timing controller 220 is configured—as noted above—to either reduce or increase the effective scan rate of a reset pointer relative the scan rate of a readout pointer in order to transition to a slower or faster scan rate for the subsequent frame. In other cases, a transition timing controller 220 may not be required; in these examples, the vertical synchronization controller 210 may be configured to reduce or increase the effective scan rate of a reset pointer relative to the scan rate of a readout pointer in order to transition to a faster or slower scan rate for the subsequent frame. Operations that can be performed by a transition timing controller 220—and/or a vertical synchronization controller 210—are described in greater detail below in reference to FIGS. 3A-5.


For example, FIG. 3A depicts a timing diagram 300—and a detailed view of the same, identified as the detailed timing diagram 302—illustrating operations of an image sensor, such as the image sensor 200 depicted in FIG. 2, operating in a selected readout mode. As noted with respect to other embodiments described herein, the selected readout mode can specify a line time (e.g., a specific reset duration, integration duration, and/or readout duration) and/or a scan rate.


In this figure, two frames are shown and identified as the frames FN and FN+1. The reset pointer of frame FN is incremented at a scan rate represented by the slope of the dotted line 304. In other words the dotted line 304 depicts a sequence of reset pointer increments or, more generally, a sequence of reset operations performed on sequential rows of the photosite array of the image sensor.


Similarly, the readout pointer of frame FN is incremented at a scan rate represented by the slope of the solid line 306. In other words the solid line 306 depicts a sequence of readout pointer increments or, more generally, a sequence of readout operations performed on sequential rows of the photosite array of the image sensor.


The reset pointer of frame FN+1 is incremented at a scan rate represented by the slope of the dotted line 308. As with other embodiments described herein, the readout pointer of frame FN leads the reset pointer of frame FN+1 by an inter-frame delay 310.


In this example, the readout mode of both frames FN and FN+1 are the same. In other words, the read out and reset pointers of the frames FN and FN+1 are incremented at precisely the same scan rate, as shown by the parallel relationship between the dotted lines 304, 308 and the solid line 306. As noted above, incrementing a readout pointer and a reset pointer at the same rate—or, in another non-limiting phrasing, performing readout and reset operations in a timed/synchronized sequence—reduces and/or eliminates switching noise in an image sensor.



FIG. 3B depicts the timing diagram 300 operating in a different readout mode associated with a slower scan rate than the readout mode illustrated in FIG. 3A. In this example, the slower scan rate can be influenced, at least in part, by a shorter readout duration than the readout mode shown in FIG. 3A.



FIG. 4 depicts a timing diagram 400—and a detailed view of the same, identified as the detailed timing diagram 402—illustrating operations of an image sensor, such as the image sensor 200 depicted in FIG. 2. In this example, the image sensor transitions from a high scan rate readout mode to a low scan rate readout mode.


In this figure, as with the embodiment depicted in FIGS. 3A-3B, two frames are shown and identified as the frames FN and FN+1. The reset pointer of frame FN is incremented at a first scan rate represented by the slope of the dotted line 404a. Similarly, the readout pointer of frame FN is incremented at the first scan rate represented by the slope of the solid line 406a.


In this embodiment, the reset pointer of frame FN+1 is incremented at a second scan rate represented by the slope of the dotted line 404b. Similarly, the readout pointer of frame FN+1 is incremented at the second scan rate represented by the slope of the solid line 406b. In order to effect the second scan rate, and as noted with respect to other embodiments described herein, the reset pointer of frame FN+1 may selectively not increment when the readout pointer of frame FN increments. By selectively not incrementing, the reset pointer of frame FN+1 gradually lags farther behind the readout pointer of frame FN. This is shown in the figure by the progressively larger delay D1-Dn.


As described above, by controlling the number of times that the reset pointer of frame FN+1 is not incremented when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter and after the integration period has elapsed, the scan rate of the readout pointer of frame FN+1 can be set to match the slowed scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the lower readout mode.



FIG. 5 depicts a timing diagram 500—and a detailed view of the same, identified as the detailed timing diagram 502—illustrating operations of an image sensor, such as the image sensor 200 depicted in FIG. 2. In this example, the image sensor transitions from a low scan rate readout mode to a high scan rate readout mode.


In this figure, as with the embodiment depicted in FIGS. 3A-4, two frames are shown and identified as the frames FN and FN+1. The reset pointer of frame FN is incremented at a first scan rate represented by the slope of the dotted line 504a. Similarly, the readout pointer of frame FN is incremented at the first scan rate represented by the slope of the solid line 506a.


In this embodiment, the reset pointer of frame FN+1 is incremented at a second scan rate represented by the slope of the dotted line 504b. Similarly, the readout pointer of frame FN+1 is incremented at the second scan rate represented by the slope of the solid line 506b. In order to effect the second scan rate, and as noted with respect to other embodiments described herein, the reset pointer of frame FN+1 may selectively reset multiple rows at a time when the readout pointer of frame FN increments.


In addition, the image sensor is configured to delay resetting the first row of frame FN+1. The delay, referred to as the staggered start delay 508, is set such that the reset pointer of frame FN+1 does not overtake the readout pointer of frame FN.


In many cases, the staggered start delay 508 has the effect of decreasing integration time in the frame FN+1. In these examples, one or more gain correction operations can be optionally performed on the frame FN+1.


In this manner, by selectively incrementing multiple rows simultaneously, the reset pointer of frame FN+1 gradually advances toward the readout pointer of frame FN. This is shown in the figure by the progressively smaller delays D1-Dn.


As described above, by controlling the number of times and the manner in which multiple rows of frame FN+1 are simultaneously reset when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter, and after the integration period has elapsed, the scan rate of the readout pointer of frame FN+1 can be set to match the faster scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the higher readout mode.



FIG. 6A is a detail view of the enclosed circle A-A of the timing diagram of FIG. 3A. This figure shows two step-wise functions corresponding to the operations of synchronously incrementing the readout pointer of frame FN and the reset pointer of frame FN+1. More specifically, the operation of incrementing the readout pointer of frame FN is shown by the step-wise function 602 and the operation of incrementing the reset pointer of frame FN+1 is shown by the step-wise function 604. As shown by the reference line 606, the step-wise function 602 and the step-wise function 604 are in phase, thereby reducing switching noise in the image sensor. Two arrows are shown generally through the centers of the step-wise functions 602, 604 illustrating that the average or effective scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 are identical.



FIG. 6B is a detail view of the enclosed circle B-B of the timing diagram of FIG. 4. As with other embodiments described herein, the figure shows two step-wise functions corresponding to the operations of synchronously incrementing the readout pointer of frame FN and the reset pointer of frame FN+1, while gradually changing the average or effective scan rate of the reset pointer of frame FN+1.


More specifically, the operation of incrementing the readout pointer of frame FN is shown by the step-wise function 602 and the operation of incrementing the reset pointer of frame FN+1 is shown by the step-wise function 604. As shown by the reference line 606, the rising edges of the step-wise function 602 and the step-wise function 604 are in phase, thereby reducing switching noise in the image sensor. Phrased in another way, the readout pointer incrementation illustrated by the step-wise function 602 is in phase with the reset pointer incrementation illustrated by the step-wise function 604.


In this embodiment, however, certain selected steps of the step-wise function 604 are skipped and/or repeated. For example, incrementing of the step 608 is skipped once (e.g., the respective row is reset at least once), effectively doubling the step's duration. Similarly, the step 610 is skipped once (e.g., the respective row is reset at least once), effectively doubling the step's duration. Other steps, such as the step 612 may not be skipped. In one phrasing, the step-wise function 604 is selectively “jittered” in order to affect a scan rate different from the step-wise function 602 while still maintaining time/phase synchronization with the step-wise function 602.


As with the embodiment described in reference to FIG. 6A, two arrows are shown generally through the centers of the step-wise functions 602, 604 illustrating that the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 are different. In particular, the figure illustrates that the scan rate of the reset pointer of frame FN+1 is less than that of the readout pointer of frame FN.



FIG. 6C is a detail view of the enclosed circle C-C of the timing diagram of FIG. 5. As with other embodiments described herein, the figure shows two step-wise functions corresponding to the operations of synchronously incrementing the readout pointer of frame FN and the reset pointer of frame FN+1, while gradually changing the effective scan rate of the reset pointer of frame FN+1.


More specifically, as with the embodiment described in reference to FIG. 6B, the operation of incrementing the readout pointer of frame FN is shown by the step-wise function 602 and the operation of incrementing the reset pointer of frame FN+1 is shown by the step-wise function 604. As shown by the reference line 606, the rising edges of the step-wise function 602 and the step-wise function 604 are in phase, thereby reducing switching noise in the image sensor.


In this embodiment, however, certain selected steps of the step-wise function 604 are performed in groups. In other words, certain rows can be reset at the same time. For example, each step in the group of three steps 614 is reset at the same time. In one phrasing, the step-wise function 604 is selectively “jittered” in order to affect a scan rate different from the step-wise function 602 while still maintaining time/phase synchronization with the step-wise function 602.


As with the embodiment described in reference to FIGS. 6A-6B, two arrows are shown generally through the centers of the step-wise functions 602, 604 illustrating that the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 are different. In particular, the figure illustrates that the scan rate of the reset pointer of frame FN+1 is greater than that of the readout pointer of frame FN.


It may be appreciated that the foregoing description of FIGS. 2-5, and the various alternatives thereof and variations thereto are presented, generally, for purposes of explanation, and to facilitate a thorough understanding of various possible configurations of an electronic device incorporating an image sensor such as described herein. However, it will be apparent to one skilled in the art that some of the specific details presented herein may not be required in order to practice a particular described embodiment, or an equivalent thereof.


For example, it may be appreciated that a linear or substantially linear relationship between the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 may not be required. In some cases, the scan rate of the reset pointer of frame FN+1 may vary according to a curve or function.


Further, although only two readout modes are described in reference to the foregoing embodiments, it may be appreciated that an image sensor such as described herein can transition to and between any number of readout modes. In some examples, multiple transitions can occur in sequence. For example, the image sensor may transition from a first readout mode at frame FN, to a second readout mode at frame FN+1, to a third readout mode at frame FN+2.


Thus, it is understood that the foregoing descriptions of specific embodiments are presented for the limited purposes of illustration and description. These descriptions are not exhaustive nor intended to limit the disclosure to the precise forms recited herein. To the contrary, it will be apparent to one of ordinary skill in the art that many modifications and variations are possible in view of the above teachings.



FIG. 7 is a flowchart depicting example operations of a method of synchronizing a readout signal and a reset signal by an image sensor, such as described herein. In some embodiments, the method 700 can be performed by a processor or controller of an image sensor, such as the processor and/or VSYNC controller described in reference to FIG. 2.


The method 700 includes operation 702 in which a target readout mode is selected, which may specify a new line timing and/or a new scan rate, such as described above. Next, at operation 704, the scan rate of the current line timing of the current frame can be used to modify the line timing of the next frame (see, e.g., FIGS. 6B-6C). Next, at operation 706, the line timing of a reset operation of the next frame is modified based on the modification determined at operation 704. Finally, at operation 708, the line timing of the readout operation of the next frame is modified.



FIG. 8 is a flowchart depicting example operations of a method of correcting gain offset in a frame by an image sensor, such as described herein. As noted above, gain correction may be preferred or required if an image sensor transitions from a low scan rate readout mode to a high scan rate readout mode. The method 800 includes operation 802 in which a line timing is transitioned between frames using one or more methods or techniques such as described herein. Next, at operation 804, a gain adjustment operation can be applied to the frame associated with the transition.


One may appreciate that although many embodiments are disclosed above, that the operations and steps presented with respect to methods and techniques described herein are meant as exemplary and accordingly are not exhaustive. One may further appreciate that alternate step order or fewer or additional operations may be required or desired for particular embodiments.


Although the disclosure above is described in terms of various exemplary embodiments and implementations, it should be understood that the various features, aspects and functionality described in one or more of the individual embodiments are not limited in their applicability to the particular embodiment with which they are described, but instead can be applied, alone or in various combinations, to one or more of the some embodiments of the invention, whether or not such embodiments are described and whether or not such features are presented as being a part of a described embodiment. Thus, the breadth and scope of the present invention should not be limited by any of the above-described exemplary embodiments but are instead defined by the claims herein presented.

Claims
  • 1. A method of operating an image sensor comprising an array of photosites, the method comprising: reading the array of photosites at a first scan rate during a readout time for a binned mode of operation of the array of photosites; andresetting the array of photosites at a second scan rate for a full-resolution readout mode of the array of photosites;wherein,the resetting commences during, and after a commencement of, the readout time; andthe second scan rate is different than the first scan rate.
  • 2. The method of claim 1, wherein the reading and the resetting are time-synchronized.
  • 3. The method of claim 1, wherein: the reading is associated with a first frame; andthe resetting is associated with a second frame.
  • 4. The method of claim 1, wherein: the binned mode of operation is associated with a first line time; andthe full-resolution mode is associated with a second line time different from the first line time.
  • 5. The method of claim 1, further comprising: after an integration period following the resetting, reading the array of photosites at the second scan rate and at a second readout time.
  • 6. The method of claim 5, further comprising: resetting the array of photosites at the first scan rate; wherein,the resetting of the array of photosites at the first scan rate commences during the second readout time.
  • 7. The method of claim 5, further comprising: resetting the array of photosites at the second scan rate; wherein,the resetting of the array of photosites at the first scan rate commences during the second readout time.
  • 8. An image sensor, comprising: a photosite array; anda controller configured to, read the photosite array at a first scan rate during a readout time for a binned mode of operation of the photosite array; andreset the photosite array at a second scan rate for a full-resolution readout mode of the photosite array;wherein,the reset commences during, and after a commencement of, the readout time; andthe second scan rate is different than the first scan rate.
  • 9. The image sensor of claim 8, wherein the read and the reset are time-synchronized.
  • 10. The image sensor of claim 8, wherein: the read is associated with a first frame; andthe reset is associated with a second frame.
  • 11. The image sensor of claim 8, wherein: the binned mode of operation is associated with a first line time; andthe full-resolution readout mode is associated with a second line time different from the first line time.
  • 12. The image sensor of claim 8, wherein: the controller is configured to, read the photosite array at the second scan rate and at a second readout time, after an integration period following the reset.
  • 13. The image sensor of claim 12, wherein: the controller is configured to, reset the photosite array at the first scan rate; wherein,the reset of the photosite array at the first scan rate commences during the second readout time.
  • 14. The image sensor of claim 12, wherein: the controller is configured to, reset the photosite array at the second scan rate; wherein,the reset of the photosite array at the first scan rate commences during the second readout time.
  • 15. A method of operating an image sensor comprising an array of photosites, the method comprising: reading the array of photosites at a first scan rate during a readout time for a full-resolution readout mode of the array of photosites; andresetting the array of photosites at a second scan rate for a binned mode of operation of the array of photosites;wherein,the resetting commences during, and after a commencement of, the readout time; andthe first scan rate is different than the second scan rate.
  • 16. The method of claim 15, wherein: the reading is associated with a first frame; andthe resetting is associated with a second frame.
  • 17. The method of claim 15, wherein: the full-resolution readout mode is associated with a first line time; andthe binned mode of operation is associated with a second line time different from the first line time.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 16/226,491, filed Dec. 19, 2018, which claims the benefit under 35 U.S.C. § 119(e) of U.S. Provisional Patent Application No. 62/700,113, filed on Jul. 18, 2018, the contents of which are incorporated herein by reference as if fully disclosed herein.

US Referenced Citations (282)
Number Name Date Kind
4686572 Takatsu Aug 1987 A
4686648 Fossum Aug 1987 A
5105264 Erhardt et al. Apr 1992 A
5329313 Keith Jul 1994 A
5396893 Oberg et al. Mar 1995 A
5471515 Fossum et al. Nov 1995 A
5541402 Ackland Jul 1996 A
5550677 Schofield et al. Aug 1996 A
5781312 Noda Jul 1998 A
5841126 Fossum et al. Nov 1998 A
5880459 Pryor et al. Mar 1999 A
5949483 Fossum et al. Sep 1999 A
6008486 Stam et al. Dec 1999 A
6040568 Caulfield et al. Mar 2000 A
6233013 Hosier et al. May 2001 B1
6348929 Acharya et al. Feb 2002 B1
6448550 Nishimura Sep 2002 B1
6528833 Lee et al. Mar 2003 B2
6541751 Bidermann Apr 2003 B1
6670904 Yakovlev Dec 2003 B1
6713796 Fox Mar 2004 B1
6714239 Guidash Mar 2004 B2
6798453 Kaifu Sep 2004 B1
6816676 Bianchi et al. Nov 2004 B2
6905470 Lee et al. Jun 2005 B2
6931269 Terry Aug 2005 B2
6956605 Hashimoto Oct 2005 B1
6982759 Goto Jan 2006 B2
7075049 Rhodes et al. Jul 2006 B2
7084914 Van Blerkom Aug 2006 B2
7091466 Bock Aug 2006 B2
7119322 Hong Oct 2006 B2
7133073 Neter Nov 2006 B1
7259413 Rhodes Aug 2007 B2
7262401 Hopper et al. Aug 2007 B2
7271835 Iizuka Sep 2007 B2
7282028 Kim et al. Oct 2007 B2
7319218 Krymski Jan 2008 B2
7332786 Altice Feb 2008 B2
7342270 Kuwazawa Mar 2008 B2
7390687 Boettiger Jun 2008 B2
7415096 Sherman Aug 2008 B2
7437013 Anderson Oct 2008 B2
7443421 Stavely et al. Oct 2008 B2
7446812 Ando et al. Nov 2008 B2
7471315 Silsby et al. Dec 2008 B2
7502054 Kalapathy Mar 2009 B2
7525168 Hsieh Apr 2009 B2
7554067 Zarnoski et al. Jun 2009 B2
7555158 Park et al. Jun 2009 B2
7622699 Sakakibara et al. Nov 2009 B2
7626626 Panicacci Dec 2009 B2
7636109 Nakajima et al. Dec 2009 B2
7671435 Ahn Mar 2010 B2
7728351 Shim Jun 2010 B2
7733402 Egawa et al. Jun 2010 B2
7737475 Hynecek Jun 2010 B2
7742090 Street Jun 2010 B2
7764312 Ono et al. Jul 2010 B2
7773138 Lahav et al. Aug 2010 B2
7786543 Hsieh Aug 2010 B2
7796171 Gardner Sep 2010 B2
7817198 Kang et al. Oct 2010 B2
7821547 Suzuki Oct 2010 B2
7873236 Li et al. Jan 2011 B2
7880785 Gallagher Feb 2011 B2
7884402 Ki Feb 2011 B2
7906826 Martin et al. Mar 2011 B2
7952121 Arimoto May 2011 B2
7952635 Lauxtermann May 2011 B2
7982789 Watanabe et al. Jul 2011 B2
8026966 Altice Sep 2011 B2
8032206 Farazi et al. Oct 2011 B1
8089036 Manabe et al. Jan 2012 B2
8089524 Urisaka Jan 2012 B2
8094232 Kusaka Jan 2012 B2
8116540 Dean Feb 2012 B2
8140143 Picard et al. Mar 2012 B2
8153947 Barbier et al. Apr 2012 B2
8159570 Negishi Apr 2012 B2
8159588 Boemler Apr 2012 B2
8164669 Compton et al. Apr 2012 B2
8174595 Honda et al. May 2012 B2
8184188 Yaghmai May 2012 B2
8194148 Doida Jun 2012 B2
8194165 Border et al. Jun 2012 B2
8222586 Lee Jul 2012 B2
8227844 Adkisson Jul 2012 B2
8233071 Takeda Jul 2012 B2
8259228 Wei et al. Sep 2012 B2
8310577 Neter Nov 2012 B1
8324553 Lee Dec 2012 B2
8330839 Compton et al. Dec 2012 B2
8338856 Tai et al. Dec 2012 B2
8340407 Kalman Dec 2012 B2
8350940 Smith et al. Jan 2013 B2
8400546 Itano et al. Mar 2013 B2
8441545 Hoda et al. May 2013 B2
8456540 Egawa Jun 2013 B2
8456559 Yamashita Jun 2013 B2
8462247 Kim Jun 2013 B2
8508637 Han et al. Aug 2013 B2
8514308 Itonaga et al. Aug 2013 B2
8520913 Dean Aug 2013 B2
8546737 Tian et al. Oct 2013 B2
8547388 Cheng Oct 2013 B2
8575531 Hynecek et al. Nov 2013 B2
8581992 Hamada Nov 2013 B2
8594170 Mombers et al. Nov 2013 B2
8619163 Ogua Dec 2013 B2
8619170 Mabuchi Dec 2013 B2
8629484 Ohri et al. Jan 2014 B2
8634002 Kita Jan 2014 B2
8648947 Sato et al. Feb 2014 B2
8723975 Solhusvik May 2014 B2
8730345 Watanabe May 2014 B2
8754983 Sutton Jun 2014 B2
8755854 Addison et al. Jun 2014 B2
8759736 Yoo Jun 2014 B2
8767104 Makino et al. Jul 2014 B2
8803990 Smith Aug 2014 B2
8810703 Mabuchi Aug 2014 B2
8817154 Manabe et al. Aug 2014 B2
8860871 Aoki Oct 2014 B2
8879686 Okada et al. Nov 2014 B2
8902330 Theuwissen Dec 2014 B2
8902341 Machida Dec 2014 B2
8908062 Ito Dec 2014 B2
8908073 Minagawa Dec 2014 B2
8923994 Laikari et al. Dec 2014 B2
8934030 Kim et al. Jan 2015 B2
8936552 Kateraas et al. Jan 2015 B2
8946610 Iwabuchi et al. Feb 2015 B2
8981517 Oshiyama et al. Mar 2015 B2
8982237 Chen Mar 2015 B2
8982260 Eshraghian et al. Mar 2015 B2
8988598 Ovsiannikov Mar 2015 B2
9001251 Smith et al. Apr 2015 B2
9041837 Li May 2015 B2
9017748 Theuwissen Jun 2015 B2
9054009 Oike et al. Jun 2015 B2
9066017 Geiss Jun 2015 B2
9066660 Watson et al. Jun 2015 B2
9088727 Trumbo Jul 2015 B2
9094623 Kawaguchi Jul 2015 B2
9099604 Roy Aug 2015 B2
9100597 Hu Aug 2015 B2
9106859 Kizuna et al. Aug 2015 B2
9131171 Aoki et al. Sep 2015 B2
9154750 Pang Oct 2015 B2
9225948 Hasegawa Dec 2015 B2
9232150 Kleekajai et al. Jan 2016 B2
9232161 Suh Jan 2016 B2
9270906 Peng et al. Feb 2016 B2
9276031 Wan Mar 2016 B2
9277144 Kleekajai et al. Mar 2016 B2
9287304 Park et al. Mar 2016 B2
9287423 Mori et al. Mar 2016 B2
9288380 Nomura Mar 2016 B2
9288404 Papiashvili Mar 2016 B2
9293500 Sharma et al. Mar 2016 B2
9307161 Lee Apr 2016 B2
9319611 Fan Apr 2016 B2
9344649 Bock May 2016 B2
9380245 Guidash Jun 2016 B1
9392237 Toyoda Jul 2016 B2
9438258 Yoo Sep 2016 B1
9445018 Fettig et al. Sep 2016 B2
9451887 Watson et al. Sep 2016 B2
9467553 Heo et al. Oct 2016 B2
9473706 Malone et al. Oct 2016 B2
9479688 Ishii Oct 2016 B2
9490285 Itonaga Nov 2016 B2
9497397 Kleekajai et al. Nov 2016 B1
9503616 Taniguchi et al. Nov 2016 B2
9521337 Shen Dec 2016 B1
9538067 Hamada Jan 2017 B2
9538106 McMahon et al. Jan 2017 B2
9549099 Fan Jan 2017 B2
9571758 Hashimoto et al. Feb 2017 B2
9584743 Lin et al. Feb 2017 B1
9584744 Lenchenkov et al. Feb 2017 B2
9596420 Fan et al. Mar 2017 B2
9596423 Molgaard Mar 2017 B1
9609250 Lee et al. Mar 2017 B2
9654689 Gleason May 2017 B2
9661210 Haneda May 2017 B2
9666618 Meynants May 2017 B2
9686485 Agranov et al. Jun 2017 B2
9700240 Letchner et al. Jul 2017 B2
9741754 Li et al. Aug 2017 B2
9749556 Fettig et al. Aug 2017 B2
9754994 Koo et al. Sep 2017 B2
9774318 Song Sep 2017 B2
9781368 Song Oct 2017 B2
9819890 Wang et al. Nov 2017 B2
9857469 Oggier et al. Jan 2018 B2
9888198 Mauritzson et al. Feb 2018 B2
9894304 Smith Feb 2018 B1
9912883 Lee Mar 2018 B1
9936105 Furuya Apr 2018 B2
9936151 Wang et al. Apr 2018 B2
9952323 Deane Apr 2018 B2
9973678 Mandelli et al. May 2018 B2
10019294 van der Lugt et al. Jul 2018 B2
10044954 Ikeda et al. Aug 2018 B2
10051217 Kondo Aug 2018 B2
10104318 Smith Oct 2018 B2
10120446 Pance et al. Nov 2018 B2
10136090 Vogelsang et al. Nov 2018 B2
10178329 Vogelsang Jan 2019 B2
10212378 Negishi Feb 2019 B2
10217889 Dhulla et al. Feb 2019 B2
10249660 Guidash et al. Apr 2019 B2
10263032 Wan Apr 2019 B2
10269857 Wu et al. Apr 2019 B2
10271037 Oh Apr 2019 B2
10285626 Kestelli et al. May 2019 B1
10306167 Shimasaki May 2019 B2
10334181 Guenter et al. Jun 2019 B2
10379317 Shimokawa et al. Aug 2019 B2
10205904 Kobayashi Sep 2019 B2
10431608 Ebihara Oct 2019 B2
10440301 Li et al. Oct 2019 B2
10447950 Wang Oct 2019 B2
10484627 Zhou Nov 2019 B2
10609348 Agranov et al. Mar 2020 B2
10630920 Matsunaga Apr 2020 B2
10630929 Koizumi et al. Apr 2020 B2
10748955 Oh et al. Aug 2020 B2
10775605 Ollila Sep 2020 B2
10848693 Agranov et al. Nov 2020 B2
10943935 Li et al. Mar 2021 B2
10943940 Wu et al. Mar 2021 B2
10951838 Eyama Mar 2021 B2
11252381 Pang et al. Feb 2022 B2
11258993 Hoshino Feb 2022 B2
11546532 Bahukhandi et al. Jan 2023 B1
20030036685 Goodman et al. Feb 2003 A1
20040207836 Chhibber et al. Oct 2004 A1
20050026332 Fratti et al. Feb 2005 A1
20060274161 Ing et al. Dec 2006 A1
20070263099 Motta et al. Nov 2007 A1
20080049132 Suzuki Feb 2008 A1
20080177162 Bae et al. Jul 2008 A1
20080315198 Jung Dec 2008 A1
20090096901 Bae et al. Apr 2009 A1
20090101914 Hirotsu et al. Apr 2009 A1
20090146234 Luo et al. Jun 2009 A1
20090201400 Zhang et al. Aug 2009 A1
20090219266 Lim et al. Sep 2009 A1
20100134631 Voth Jun 2010 A1
20110080500 Wang et al. Apr 2011 A1
20110109776 Kawai May 2011 A1
20110156197 Tivarus et al. Jun 2011 A1
20110164162 Kato Jul 2011 A1
20120092541 Tuulos et al. Apr 2012 A1
20120098964 Oggier Apr 2012 A1
20130147981 Wu Jun 2013 A1
20140071321 Seyama Mar 2014 A1
20140078356 Vaartstra Mar 2014 A1
20140078381 Ovsiannikov Mar 2014 A1
20150062391 Murata Mar 2015 A1
20150350583 Mauritzson Dec 2015 A1
20160050379 Jiang et al. Feb 2016 A1
20160191826 Furuya Jun 2016 A1
20160205311 Mandelli Jul 2016 A1
20160219232 Murata Jul 2016 A1
20160255262 Lee Sep 2016 A1
20160286108 Fettig Sep 2016 A1
20160309140 Wang Oct 2016 A1
20160323524 Smith Nov 2016 A1
20160344920 Iwahara Nov 2016 A1
20160344958 Kondo Nov 2016 A1
20170047363 Choi et al. Feb 2017 A1
20170111600 Wang Apr 2017 A1
20170195596 Vogelsang Jul 2017 A1
20170295326 Eyama Oct 2017 A1
20210021778 Bairo Jan 2021 A1
20210281789 McMahon Sep 2021 A1
20220046196 Li Feb 2022 A1
20220375986 Rosenblum et al. Nov 2022 A1
Foreign Referenced Citations (93)
Number Date Country
1630350 Jun 2005 CN
1774032 May 2006 CN
1833429 Sep 2006 CN
1842138 Oct 2006 CN
1947414 Apr 2007 CN
101189885 May 2008 CN
101221965 Jul 2008 CN
101233763 Jul 2008 CN
101472059 Jul 2009 CN
101567977 Oct 2009 CN
101622859 Jan 2010 CN
101739955 Jun 2010 CN
101754029 Jun 2010 CN
101803925 Aug 2010 CN
102036020 Apr 2011 CN
102067584 May 2011 CN
102208423 Oct 2011 CN
102451160 May 2012 CN
102668542 Sep 2012 CN
102820309 Dec 2012 CN
102821255 Dec 2012 CN
103024297 Apr 2013 CN
103051843 Apr 2013 CN
103329513 Sep 2013 CN
103546702 Jan 2014 CN
104041009 Sep 2014 CN
104052919 Sep 2014 CN
204761615 Nov 2015 CN
205211754 May 2016 CN
814606 Feb 2002 EP
1763228 Mar 2007 EP
2023611 Feb 2009 EP
2107610 Oct 2009 EP
2230690 Sep 2010 EP
2512126 Oct 2012 EP
2601833 Jun 2022 GB
S61123287 Jun 1986 JP
2000059697 Feb 2000 JP
2001211455 Aug 2001 JP
2001358994 Dec 2001 JP
2004111590 Apr 2004 JP
2005318504 Nov 2005 JP
2006287361 Oct 2006 JP
2007504670 Mar 2007 JP
2007516654 Jun 2007 JP
2008507908 Mar 2008 JP
2008271280 Nov 2008 JP
2008543061 Nov 2008 JP
2009021809 Jan 2009 JP
4255223 Apr 2009 JP
2009159186 Jul 2009 JP
2009212909 Sep 2009 JP
2009296465 Dec 2009 JP
2010080604 Apr 2010 JP
2010114834 May 2010 JP
2011040926 Feb 2011 JP
2011049697 Mar 2011 JP
2011091775 May 2011 JP
2011216970 Oct 2011 JP
2011217315 Oct 2011 JP
2011097646 Dec 2011 JP
2012010306 Jan 2012 JP
2012019516 Jan 2012 JP
2012513160 Jun 2012 JP
2013005397 Jan 2013 JP
2013051523 Mar 2013 JP
2013070240 Apr 2013 JP
2013529035 Jul 2013 JP
20030034424 May 2003 KR
20030061157 Jul 2003 KR
20050103732 Nov 2005 KR
20080069851 Jul 2008 KR
20100008239 Jan 2010 KR
20100065084 Jun 2010 KR
20130074459 Jul 2013 KR
200520551 Jun 2005 TW
200803481 Jan 2008 TW
201110689 Mar 2011 TW
201301881 Jan 2013 TW
WO 05041304 May 2005 WO
WO 06014641 Feb 2006 WO
WO 06130443 Dec 2006 WO
WO 07049900 May 2007 WO
WO 10120945 Oct 2010 WO
WO 12053363 Apr 2012 WO
WO 12088338 Jun 2012 WO
WO 12122572 Sep 2012 WO
WO 12138687 Oct 2012 WO
WO 13008425 Jan 2013 WO
WO 13179018 Dec 2013 WO
WO 13179020 Dec 2013 WO
WO 19102887 May 2019 WO
WO 20228392 Nov 2020 WO
Non-Patent Literature Citations (17)
Entry
U.S. Appl. No. 16/805,558, filed Feb. 28, 2020, Lee et al.
U.S. Appl. No. 17/075,547, filed Oct. 20, 2020, Agranov et al.
U.S. Appl. No. 17/380,852, filed Jul. 20, 2021, Li et al.
Aoki, et al., “Rolling-Shutter Distortion-Free 3D Stacked Image Sensor with -160dB Parasitic Light Sensitivity In-Pixel Storage Node,” ISSCC 2013, Session 27, Image Sensors, 27.3 27.3 A, Feb. 20, 2013, retrieved on Apr. 11, 2014 from URL:http://ieeexplore.ieee.org/stamp/stamp.jsp?tp-&arnumber=6487824.
Elgendi, “On the Analysis of Fingertip Photoplethysmogram Signals,” Current Cardiology Reviews, 2012, vol. 8, pp. 14-25.
Feng, et al., “On the Stoney Formula for a Thin Film/Substrate System with Nonuniform Substrate Thickness,” Journal of Applied Mechanics, Transactions of the ASME, vol. 74, Nov. 2007, pp. 1276-1281.
Fu, et al., “Heart Rate Extraction from Photoplethysmogram Waveform Using Wavelet Multui-resolution Analysis,” Journal of Medical and Biological Engineering, 2008, vol. 28, No. 4, pp. 229-232.
Han, et al., “Artifacts in wearable photoplethysmographs during daily life motions and their reduction with least mean square based active noise cancellation method,” Computers in Biology and Medicine, 2012, vol. 42, pp. 387-393.
Lopez-Silva, et al., “Heuristic Algorithm for Photoplethysmographic Heart Rate Tracking During Maximal Exercise Test,” Journal of Medical and Biological Engineering, 2011, vol. 12, No. 3, pp. 181-188.
Santos, et al., “Accelerometer-assisted PPG Measurement During Physical Exercise Using the LAVIMO Sensor System,” Acta Polytechnica, 2012, vol. 52, No. 5, pp. 80-85.
Sarkar, et al., “Fingertip Pulse Wave (PPG signal) Analysis and Heart Rate Detection,” International Journal of Emerging Technology and Advanced Engineering, 2012, vol. 2, No. 9, pp. 404-407.
Schwarzer, et al., On the determination of film stress from substrate bending: STONEY'S formula and its limits, Jan. 2006, 19 pages.
Shen et al., “Stresses, Curvatures, and Shape Changes Arising from Patterned Lines on Silicon Wafers,” Journal of Applied Physics, vol. 80, No. 3, Aug. 1996, pp. 1388-1398.
Yan, et al., “Reduction of motion artifact in pulse oximetry by smoothed pseudo Wigner-Ville distribution,” Journal of NeuroEngineering and Rehabilitation, 2005, vol. 2, No. 3, pp. 1-9.
Yousefi, et al., “Adaptive Cancellation of Motion Artifact in Wearable Biosensors,” 34th Annual International Conference of the IEEE EMBS, San Diego, California, Aug./Sep. 2012, pp. 2004-2008.
Fontaine, “The State-of-the-Art of Smartphone Imagers,” 2019 International Image Sensor Workshop, Snowbird, Utah, Jun. 23-27, 2019, 3 pages.
Jang et al., “A new PDAF correction method of CMOS image sensor with Nonacell and Super PD to improve image quality in binning mode,” IS&T International Symposium on Electronic Imaging 2021, Image Quality and System Performance XVIII, 5 pages.
Related Publications (1)
Number Date Country
20210281789 A1 Sep 2021 US
Provisional Applications (1)
Number Date Country
62700113 Jul 2018 US
Continuations (1)
Number Date Country
Parent 16226491 Dec 2018 US
Child 17329010 US