Embodiments described herein relate to image sensors and, in particular, to image sensors configured to seamlessly transition between readout modes.
An image sensor for an electronic device typically includes an array of photosensitive electronic components conventionally referred to as “photosites.” When exposed to light, each photosite accumulates an electric charge proportional to the brightness of light illuminating that photosite for the duration of the exposure. After exposure, the image sensor reads out the accumulated electric charge of all or some of the photosites to generate an image. The photosites of a conventional image sensor are generally read out in sequence (e.g., by rows or by columns) for each image output from the image sensor.
Some image sensors support multiple photosite readout modes. Example readout modes include full-resolution readout modes, binning readout modes, oversampling readout modes, and so on. However, conventional image sensors undesirably interrupt—and/or otherwise disable—image output while transitioning between readout modes.
Many embodiments described herein reference a method of operating an image sensor. The method includes the operations of: applying a readout signal to a first group of photosites defined by a readout pointer; applying a reset signal to a second group of photosites defined by a reset pointer; incrementing the readout pointer at a first scan rate; and incrementing the reset pointer at a second scan rate different from the first scan rate. In these examples, the reset signal and the readout signal are applied to the first and second group of photosites, respectively, in phase.
Some embodiments described reference an image sensor including a photosite array. The photosite array includes a first group of photo sites and a second group of photosites. The image sensor also includes a controller configured to: apply a readout signal (associated with a readout pointer) to the first group of photosites; apply a reset signal (associated with a reset pointer) in phase with the readout signal, to the second group of photosites; increment the readout pointer at a first scan rate; and increment the reset pointer at a second scan rate different from the first scan rate.
Reference will now be made to representative embodiments illustrated in the accompanying figures. It should be understood that the following descriptions are not intended to limit this disclosure to one included embodiment. To the contrary, the disclosure provided herein is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the described embodiments, and as defined by the appended claims.
The use of the same or similar reference numerals in different figures indicates similar, related, or identical items.
Additionally, it should be understood that the proportions and dimensions (either relative or absolute) of the various features and elements (and collections and groupings thereof) and the boundaries, separations, and positional relationships presented therebetween, are provided in the accompanying figures merely to facilitate an understanding of the various embodiments described herein and, accordingly, may not necessarily be presented or illustrated to scale, and are not intended to indicate any preference or requirement for an illustrated embodiment to the exclusion of embodiments described with reference thereto.
Embodiments described herein reference an image sensor for an electronic device that is configured to seamlessly transition between different readout modes.
For simplicity of description, the embodiments that follow reference a complementary metal oxide semiconductor (“CMOS”) image sensor that is sensitive to light in a traditionally visible spectral range—and methods for operating the same—but it may be appreciated that this is merely one example construction and that other image sensor topologies (e.g., charge coupled devices, non-visible light image sensors, front side or back side illuminated image sensors, flat panel detectors, and so on) may be suitable in other implementations.
A CMOS image sensor, such as described herein, includes—among other components, structures, and/or elements—an array of “photosites,” referred to as a “photosite array.” Typically, the photosite array includes a set of columns and a set of rows arranged in a grid. In some implementations, the term “pixel” can be used as a synonym for the term “photosites.”
Each photosite of a photosite array includes at least one “photosensitive area.” When a photosensitive area of a photosite is exposed to light for a particular exposure duration, the photosensitive area accumulates an electric charge. This is referred to as an “integration” operation because the accumulated electric charge corresponds to the total quantity of photons received at the photosensitive area over the exposure duration (herein, the “integration time”). In other words, the electric charge accumulated by a particular photosensitive area of a particular photosite of a photosite array is proportional to the brightness of light illuminating that photosensitive area for the integration time.
After integration, the image sensor “reads out” the accumulated electric charge (as a current, as a voltage, or as a digital value) of all or some of the photosensitive areas of all or some of the photosites as an initial step in a process to generate a digital image referred to herein as a “frame.” The photosensitive areas of the photosites of a photosite array, such as described herein, are typically read out in groups. For example, in some embodiments, the photosensitive areas of the photosites of a photosite array can be grouped by rows or columns, read out in a row-wise or a column-wise sequence (e.g., read out sequentially by columns or read out sequentially by rows). In addition, after a readout operation is performed, each photosensitive area of each photosite of the photosite array is “reset” in order to discharge any residual electric charge prior to integration of the next frame.
For embodiments described herein, a CMOS image sensor (hereinafter, simply, “image sensor”) is understood to be configured to output frames at a constant or, optionally, variable “frame rate.” For each frame output from an image sensor, the image sensor typically performs at least three operations. First, the image sensor sends a “reset signal” to all or some of the photosites to reset all or some of the photosensitive areas of those photosites. Second, the image sensor waits for an integration time period to elapse. Third, the image sensor reads out the accumulated electric charge (e.g., as a voltage, as a current, or as a digital value) of all or some of the photosensitive areas of all or some of the photosites of the photosite array. The accumulated electric charge information, regardless of form or format (e.g., charge, voltage, digital value, and so on), is referred to herein for simplicity of description as “data” read out from the photosite array.
The image sensor assembles a frame based on the data read out from the photosite array. In some cases, the image sensor can apply one or more filters to, and/or perform one or more operations with, the data read out from the photosite array prior to (and/or subsequent to) assembling the frame. Examples include, but are not limited to: operations to apply one or more demosaicing/debayering filters; binning operations (e.g., combining charges accumulated by adjacent or nearby photosites); averaging operations; gain correction operations; and so on.
For simplicity of description, the embodiments that follow reference an image sensor that resets, integrates, and reads out each photosensitive area of each photosite of a photosite array in a row-wise sequence (e.g., the image sensor resets, integrates, and reads out photosensitive areas in a row then sequentially advances to the next row). It may be appreciated, however, that this is merely one example and that other embodiments can be configured to reset, integrate, and/or read out one or more photosensitive areas of one or more photosites of a photosite array in any other suitable manner.
More specifically, the embodiments that follow reference an image sensor configured to perform reset and readout operations on entire rows of photosites at a time. For example, to perform a reset operation, the image sensor can send a reset signal to each photosite in a selected row of photosites at substantially the same time in order to efficiently discharge any residual electric charge accumulated by any photosensitive area of any photosite in the selected row. The selected row can be addressed by a pointer referred to herein as the “reset pointer.” Thereafter, the next row is selected by incrementing the reset pointer and the process is repeated until all rows of the photosite array have been reset.
Similarly, to perform a readout operation after a selected integration period has elapsed, each photosensitive area of each photosite in a selected row of photosites can be read out (e.g., into temporary storage capacitors, buffers, registers, analog-to-digital converters, amplifiers, and so on) substantially simultaneously. As with the reset operation described above, the selected row can be addressed by a pointer referred to herein as the “readout pointer.” Thereafter, the readout pointer is incremented and the process is repeated until all rows have been read out.
In typical embodiments, the readout pointer of a particular frame leads the reset pointer of the next frame by a certain amount referred to herein as the “inter-frame delay.” In other words, an image sensor—such as described herein—is generally configured to perform a readout operation simultaneously with a reset operation. More specifically, the image sensor is typically configured to perform a readout operation for a first frame (herein, “frame FN”) while simultaneously performing a reset operation for a second frame (herein, “frame FN+1”).
In many embodiments, the reset pointer of frame FN+1 and the readout pointer of frame FN are synchronously incremented at a particular “scan rate.” As may be appreciated, the pointers are incremented at the same time in order to reduce and/or eliminate switching noise in the image sensor. The scan rate (e.g., the rate at which the reset pointer and the readout pointer are incremented together) can be based, at least in part, on the total time required to reset and read out a particular row of photosites. Collectively, the total time required to reset and read out a particular row of photosites is referred to herein as the “line time.”
As noted above, precise synchronization while incrementing the reset pointer of frame FN+1 and the readout pointer of frame FN may be required to reduce and/or eliminate switching noise in the image sensor. To facilitate this synchronization, embodiments described herein include a timing controller that can include a vertical synchronization controller (herein, a “VSYNC” controller). The VSYNC controller can be any suitable digital or analog circuit, including one or more discrete or analog components, application-specific integrated circuits, general-purpose processors, and so on. The VSYNC controller can, in some examples, be implemented in whole or in part in software.
The VSYNC controller is configured to set an appropriate scan rate for the reset pointer of frame FN+1 and the readout pointer of frame FN and to maintain synchronization while incrementing the two pointers. In some implementations, the VSYNC controller includes at least one asynchronous first-in first-out controller (also referred to as an “ASYNC FIFO”) that can be used to buffer or otherwise receive and store addresses of the readout pointer of frame FN and the reset pointer of frame FN+1. The rate at which addresses in the ASYNC FIFO are flushed may depend upon the scan rate maintained by the VSYNC controller. It may be appreciated, however, that this is merely one example; other circuit topologies can be used in other embodiments.
Further, as noted above, an image sensor such as described herein is typically configured to operate in multiple readout modes. Example readout modes include, but are not limited to: full-resolution readout modes (e.g., a readout mode in which a frame is generated by reading out each photosite of the image sensor); binning readout modes (e.g., a readout mode in which a frame is generated by combining charges accumulated by adjacent or nearby photosites with each other to reduce noise); photosite-skipping readout modes (e.g., a readout mode in which select photosites are skipped); oversampling readout modes (e.g., a readout mode in which more photosites are sampled than an output frame requires to increase noise tolerance); high frame-rate readout modes (e.g., a readout mode in which frame integration times are reduced to increase frame count per second); low frame-rate readout modes (e.g., a readout mode in which frame integration times are increased to decrease frame count per second); and so on.
It may be appreciated that different readout modes may be associated with different integration times, different readout times, different scan rates, different line times, and so on. For example, a binning readout mode may be associated with a different line time and scan rate than a full-resolution readout mode. As noted above, conventional image sensors undesirably interrupt frame output (e.g., inserting a vertical blanking frame) in order to transition between readout modes.
For embodiments described herein, however, an image sensor is configured to adjust the incrementation of the reset pointer of frame FN+1 relative to the incrementation of the readout pointer of frame FN to smoothly transition between readout modes in real-time (e.g., transitioning from an “original readout mode” to a “target readout mode” without interrupting frame output).
For example, in some cases, the reset pointer of frame FN+1 may selectively not increment when the readout pointer of frame FN increments. By selectively not incrementing, the reset pointer of frame FN+1 gradually lags farther behind the readout pointer of frame FN. Phrased in another non-limiting manner, the reset pointer of frame FN+1 increments through the rows in the photosite array at a slower rate (e.g., a slower scan rate) than the readout pointer of frame FN. In this manner, by controlling the number of times that the reset pointer of frame FN+1 is not incremented when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter (and after an integration period has elapsed) the scan rate of the readout pointer of frame FN+1 can be set to match the slowed scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the target readout mode. To illustrate this example, Table 1 and Table 2 are provided below.
Table 1 corresponds to an example embodiment in which a photosite array includes ten rows. In this example, the readout pointer of frame FN is incremented once—or skipped—for every time unit t. Similarly, the reset pointer of frame FN+1 is incremented once for every time unit t. The two frames FN, FN+1 are separated by an inter-frame delay of one time unit t and the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 are the same at one row per time unit t.
Table 2 corresponds to an example embodiment such as described above in which the image sensor transitions from an original readout mode to a target readout mode associated with a slower scan rate than the original readout mode.
More specifically, the reset pointer of frame FN+1 in this example increments through the ten rows of the example photosite array at a slower rate than the readout pointer of frame FN. In this example, the readout pointer of frame FN is incremented once, or skipped for every time unit t according to the scan rate of the original readout mode. In contrast, the reset pointer of frame FN+1 is incremented three times for every four time units t, according to the scan rate of the target readout mode. In this example, the two frames FN, FN+1 remain separated by an inter-frame delay of one time unit t, but the effective scan rates are different. More specifically, the readout pointer of frame FN proceeds with a scan rate of one row per time unit t (e.g., the scan rate of the original readout mode) while the reset pointer of frame FN+1 proceeds with a scan rate approximately equal to 0.75 rows per time unit t (e.g., the scan rate of the target readout mode).
As a result of the approach described in the preceding example, the scan rate of the reset pointer of frame FN+1 can be changed to match the scan rate of the target readout mode, while still maintaining time synchronization with the readout pointer for frame FN to prevent and/or mitigate effects of switching noise. In another non-limiting phrasing, this example configuration smoothly transitions—in a single frame—from a higher scan rate readout mode for frame FN to a lower scan rate readout mode for frame FN+1.
In another example configuration of an image sensor described herein, more than one row of photosites can be reset simultaneously. As a result of this approach, the effective scan rate of the reset pointer(s) of frame FN+1 conforms, generally, to the scan rate of the target readout mode while still maintaining time synchronization with the scan rate of the current readout mode. In another non-limiting phrasing, this example configuration transitions from a lower scan rate readout mode to a higher scan rate readout mode by increasing the pace at which the reset pointers of frame FN+1 increment relative to the pace at which the readout pointer of frame FN increments.
Table 3 corresponds to an example embodiment, such as described above, in which the image sensor transitions to a target readout mode associated with a faster scan rate than the original readout mode.
More specifically, the reset pointer of frame FN+1 in this example, increments through the ten rows of the example photosite array at a faster rate than the readout pointer of frame FN. In this example, the readout pointer of frame FN is incremented once for every time unit t according to the scan rate of the original readout mode. In contrast, the reset pointer of frame FN+1 is delayed by a certain amount (e.g., 6 time units t) and, thereafter, two rows on average are reset simultaneously for every one time unit t, according to the scan rate of the target readout mode. More specifically, the readout pointer of frame FN proceeds with a scan rate of one row per time unit t (e.g., the scan rate of the original readout mode) while the reset pointer of frame FN+1 proceeds with a scan rate approximately equal to two rows per time unit t (e.g., the scan rate of the target readout mode).
As a result of the approach described in reference to the preceding example, the scan rate of the reset pointer of frame FN+1 can be changed to match the scan rate of the target readout mode, while still maintaining time synchronization with the readout pointer for frame FN to prevent and/or mitigate effects of switching noise. In another non-limiting phrasing, this example configuration smoothly transitions—in a single frame—from a lower scan rate readout mode for frame FN to a higher scan rate readout mode for frame FN+1.
In some cases, after a transition from a first scan rate to a second scan rate is performed, line timing may also be changed (e.g., increased or decreased).
It may be appreciated that the preceding examples described in reference to Table 2 and Table 3 are not exhaustive. To the contrary, it may be appreciated that the systems, methods, and techniques described herein can be applied to effect real-time and/or smooth transitions between any number of suitable readout modes.
These foregoing and other embodiments are discussed below with reference to
The image sensor 106 is disposed within the housing 102, and is positioned to align with—and/or extend at least partially through—an aperture defined by the housing 102. As illustrated, the aperture is circular, but this may not be required. In other cases, the aperture may not be required at all; the image sensor 106 can be positioned behind a transparent portion of the housing.
As with other embodiments described herein, the image sensor 106 includes a photosite array of individual photosites also referred to as “pixels.” Typically, a photosite array is rectangular and the photosites of the array are arranged in a regular grid pattern of columns and rows. Further, a photosite array is typically planar, although in some embodiments, a photosite array can be concave or convex or may take another shape.
Each photosite of the photosite array can include one or more of, without limitation: a photosensitive area (e.g., photodiode, charge-coupled device, and so on); a reset switch or circuit; a microlens; a hot mirror (e.g., or other band stop optical filter); a color filter (e.g., a portion of a Bayer pattern); an output amplifier (e.g., charge amplifier, voltage amplifier, trans-impedance amplifier, and so on); an analog-to-digital converter; a charge storage capacitor; digital memory; and so on.
Some image sensor embodiments can include photosites with more than one photosensitive area. In these examples, different photosensitive areas may be formed from different materials, or material combinations, and/or may have different photosensitivity or electrical characteristics (e.g., rise time, fall time, reverse bias, dark current, and so on). In another example, a first photosensitive area is positioned adjacent to a second photosensitive area of the same photosite to facilitate autofocus adjustments or parallax-based depth sensing.
For simplicity of description, the image sensor 106 is understood to include a photosite array in which each photosite includes a single photosensitive area defined by a semiconductor material (e.g., silicon, germanium, indium gallium arsenide, gallium phosphide, and so on). It is appreciated, however, that this is merely one example and, that in other implementations, a photosite can include more than one photosensitive area and/or an image sensor can include a photosite array in which some photosites include more than one photosensitive area and some photosites include only a single photosensitive area.
It may be appreciated that the foregoing description of
For example, although the electronic device 100 is depicted as a cellular phone, it may be appreciated that other electronic devices can incorporate an image sensor such as described herein including, but not limited to: tablet devices; laptop devices; wearable devices; desktop computers; computing accessories; peripheral input devices; vehicle control devices; mobile entertainment devices; augmented reality devices; virtual reality devices; industrial control devices; digital wallet devices; home security devices; business security devices; and so on.
Further, it is appreciated that an electronic device, such as the electronic device 100, can include any number of image sensors. In some examples, an electronic device can include image sensors oriented outwardly from a front surface of a housing (such as shown in
Further, it is appreciated that the electronic device 100—or, more generally, any portable or stationary electronic device incorporating one or more image sensors such as described herein—can additionally include, without limitation, a processor, memory, power supply and/or battery, network connections, sensors, input/output ports, acoustic elements, haptic elements, digital and/or analog circuits for performing and/or coordinating tasks, and so on. For simplicity of illustration, the electronic device 100 is depicted in
For example, a processor of the electronic device 100 can be configured to access and execute instructions stored in a memory in order to instantiate any number of suitable classes, objects, virtual machines, threads, pipelines, and/or routines to perform, monitor, and/or coordinate one or more operations of the electronic device 100 and/or the image sensor 106. In some cases, the processor can be communicably coupled—either directly (e.g., via general-purpose input/output pins) or indirectly (e.g., via an intermediary circuit or integrated circuit)—to one or more of the hardware components of the electronic device 100, including the display 104 and the image sensor 106. In this manner, frames output by the image sensor 106 can be shown and/or rendered on the display 104 in substantially real time. In other embodiments, the processor may not be a required intermediary between the display 104 and the image sensor 106; frames output by the image sensor 106 can be received directly by the display 104.
In one embodiment, the processor can cause the display 104 to render a graphical user interface. In many cases, the graphical user interface can include a live view of frames output from the image sensor 106. As a result of this construction, a user of the electronic device 100 can observe a scene captured by the image sensor 106 in substantially real time.
In further examples, the processor of the electronic device 100 can render the graphical user interface while monitoring a user input sensor (e.g., capacitive touch sensor, capacitive force sensor, and so on) to detect one or more characteristics of a user's physical interaction with the display 104 and/or the user's physical interaction with the live view of frames output from the image sensor 106. Example detectable characteristics can include, but are not limited to, characteristics of: touch input events (e.g., location and duration); force input events (e.g., location, duration, and magnitude); touch gesture events (e.g., swipe, zoom in, zoom out, rotate, pinch); force gesture events (e.g., heavy press, light press); and so on.
As a result of this construction, a user of the electronic device 100 is encouraged to interact with content shown on the display 104, including the live view of frames output from the image sensor 106. In response to detecting specific user interactions and/or specific events (herein, generally, “user input”), the processor and/or another digital or analog circuit in the electronic device 100 can perform a specific task or operation related to one or more of the display 104, the graphical user interface, and/or the image sensor 106.
For example, in one embodiment, the graphical user interface shown on the display 104 includes a live view of frames output from the image sensor 106 operating in a particular readout mode suitable for capturing high-resolution images of a well-illuminated scene (such as a full-resolution readout mode). In some circumstances, a user of the electronic device 100 may prefer to change the readout mode of the image sensor 106 to a readout mode more suitable for capturing low-noise images of a poorly-illuminated scene (such as a binned readout mode). In this example, the user can interact with the graphical user interface (e.g., by selecting a button, by swiping in a particular direction, by applying a force beyond a threshold, and so on) to instruct the electronic device 100 to change the readout mode of the image sensor 106. In response to receiving the user input, the processor—and/or another digital or analog circuit of the electronic device 100 or the image sensor 106—can cause the image sensor 106 to change readout modes.
Accordingly, generally and broadly, it is understood that in some embodiments a graphical user interface of an electronic device can be used to trigger changes in readout modes of an image sensor, such as described herein. In other cases, however, an input sensor can change readout modes in response to other triggers, states, or events such as, but not limited to: a setting of the electronic device; output received from a remote or local application; output from an image processing algorithm; and so on.
Thus, it is understood that the foregoing descriptions of specific embodiments are presented for the limited purposes of illustration and description. These descriptions are not exhaustive nor intended to limit the disclosure to the precise forms recited herein. To the contrary, it will be apparent to one of ordinary skill in the art that many modifications and variations are possible in view of the above teachings.
As with other embodiments described herein, the photosites of the photosite array 202 can each include a number of discrete circuits and/or components that may vary from embodiment to embodiment. For simplicity of description and illustration, the embodiments that follow reference a single photosite positioned within the photosite array 202 at position (0, n); it may be appreciated that other photosites of the photosite array 202 may be similarly configured. In other examples, however, the photosites of the photosite array 202 may be configured in a different manner than shown and described relative to the photosite (0, n).
In the illustrated example, the photosite (0, n) includes a photosensitive element 204 (e.g., a photodiode) defining a photosensitive area that is sensitive to light in a particular spectral range. In typical examples, the photosensitive area is sensitive to light in a traditionally visible spectral range, but this may not be required of all embodiments. For example, in some cases, the photosensitive area of the photosensitive element 204 may be sensitive to infrared light.
In many examples, the photosite (0, n) also includes an amplifier 206. Output from the amplifier 206 is gated by a switch 208.
As noted with respect to other embodiments described herein, when the photosite array 202 of the image sensor 200 is exposed to light during an integration operation, electric charge is accumulated on photosensitive area(s) of the photosensitive element 204. During a readout operation, the switch 208 can be closed, connecting an output of the amplifier 206 to a readout circuit of the image sensor 200. In further examples, the photosite (0, n) can include a second switch associated with a reset operation; when closed (or in other circuit topologies, opened) the second switch can discharge any latent charge held by the photosensitive element 204.
As noted with respect to other embodiments described herein, the image sensor 200 is configured to perform readout operations and reset operations in a row-wise fashion. To facilitate this mode of operation, the image sensor 200 includes a vertical synchronization controller 210 configured to apply one or more signals (e.g., readout signals and/or reset signals) to one or more rows simultaneously, typically in phase with a clock signal (not shown). In one example, frame FN can be read out while frame FN+1 is being reset; the readout signal and the reset signals (of the next frame) are applied to rows in phase with one another. In this example, the vertical synchronization controller 210 can be configured to release and/or otherwise send a reset signal (for frame FN+1) to a first row of the set of rows at the same time a readout signal (for frame FN) is sent to a second row of the set of rows. By maintaining precise timing between the readout signals associated with frame FN and the reset signals associated with frame FN+1, switching noise can be mitigated.
The specific row (or rows) driven by the vertical synchronization controller 210 can be selected by a row decoder 212. The row decoder 212 enables a specific row trace from a set of row traces (e.g., a bus) coupling the vertical synchronization controller 210 and the row decoder 212 in response to receiving an address and/or a pointer—such as a readout pointer or a reset pointer, such as described above—from a timing controller 214.
In one specific example, the timing controller 214 can initiate a reset operation by setting the reset pointer to reference row 0. Next, the timing controller 214 sends the reset pointer to the row decoder 212. The row decoder 212 decodes the pointer and determines a row trace associated with the reset pointer. Next, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the readout operation on row zero. In one example, the vertical synchronization controller 210 buffers the received signal until an edge of a clock signal (not shown) is generated or received by the vertical synchronization controller 210.
Thereafter, and after a delay based on a scan rate (such as described above), the timing controller 214 increments the reset pointer to reference row one, which is decoded by the row decoder 212 into a specific row trace. Next, as with the previous example, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the reset operation on row one.
The image sensor 200 includes a readout circuit. The readout circuit can include, in some examples, an (optional) column decoder/encoder 216 and an (optional) analog to digital converter 218. Much like the row decoder 212, the column decoder/encoder 216 can be configured to translate physical trace locations into addresses, and vice versa. The analog to digital converter 218 is configured to convert voltage, current, or charge information received from one or more photosites into a digital representation of an appropriate or selected bit depth (e.g., 8-bit, 10-bit, 12-bit, 16-bit, and so on). Thereafter, optionally, the output from the analog to digital converter 218 can be received by a gain correction controller (not shown) that can provide for gain correction.
In one specific example, the timing controller 214 can initiate a readout operation by setting the readout pointer to reference row 0. Next, the timing controller 214 sends the readout pointer to the row decoder 212. The row decoder 212 decodes the pointer and determines a row trace associated with the readout pointer. Next, the row decoder 212 sends a signal via the row trace to the vertical synchronization controller 210. The vertical synchronization controller 210 buffers the received signal until a time at which it is appropriate to perform the readout operation on row zero.
Once the vertical synchronization controller 210 applies the readout signal to row zero, the accumulated electrical charge of each photosensitive element of each photosite in row zero is converted into a digital value (either serially or in parallel) by the analog to digital converter 218. Thereafter, output from the column decoder/encoder 216 can be used to associate the digital value(s) output from the analog to digital converter 218 with specific row and column addresses.
Thereafter, and after a delay based on a scan rate such as described above, the timing controller 214 increments the readout pointer to reference row 1, and the readout operation continues.
As with other embodiments described herein, the image sensor 200 can be configured to adjust the incrementation of reset pointers relative to the incrementation of readout pointers in order to smoothly transition between readout modes in real-time. To facilitate transitions such as these, the image sensor 200 includes a transition timing controller 220. The transition timing controller 220 is configured—as noted above—to either reduce or increase the effective scan rate of a reset pointer relative the scan rate of a readout pointer in order to transition to a slower or faster scan rate for the subsequent frame. In other cases, a transition timing controller 220 may not be required; in these examples, the vertical synchronization controller 210 may be configured to reduce or increase the effective scan rate of a reset pointer relative to the scan rate of a readout pointer in order to transition to a faster or slower scan rate for the subsequent frame. Operations that can be performed by a transition timing controller 220—and/or a vertical synchronization controller 210—are described in greater detail below in reference to
For example,
In this figure, two frames are shown and identified as the frames FN and FN+1. The reset pointer of frame FN is incremented at a scan rate represented by the slope of the dotted line 304. In other words the dotted line 304 depicts a sequence of reset pointer increments or, more generally, a sequence of reset operations performed on sequential rows of the photosite array of the image sensor.
Similarly, the readout pointer of frame FN is incremented at a scan rate represented by the slope of the solid line 306. In other words the solid line 306 depicts a sequence of readout pointer increments or, more generally, a sequence of readout operations performed on sequential rows of the photosite array of the image sensor.
The reset pointer of frame FN+1 is incremented at a scan rate represented by the slope of the dotted line 308. As with other embodiments described herein, the readout pointer of frame FN leads the reset pointer of frame FN+1 by an inter-frame delay 310.
In this example, the readout mode of both frames FN and FN+1 are the same. In other words, the read out and reset pointers of the frames FN and FN+1 are incremented at precisely the same scan rate, as shown by the parallel relationship between the dotted lines 304, 308 and the solid line 306. As noted above, incrementing a readout pointer and a reset pointer at the same rate—or, in another non-limiting phrasing, performing readout and reset operations in a timed/synchronized sequence—reduces and/or eliminates switching noise in an image sensor.
In this figure, as with the embodiment depicted in
In this embodiment, the reset pointer of frame FN+1 is incremented at a second scan rate represented by the slope of the dotted line 404b. Similarly, the readout pointer of frame FN+1 is incremented at the second scan rate represented by the slope of the solid line 406b. In order to effect the second scan rate, and as noted with respect to other embodiments described herein, the reset pointer of frame FN+1 may selectively not increment when the readout pointer of frame FN increments. By selectively not incrementing, the reset pointer of frame FN+1 gradually lags farther behind the readout pointer of frame FN. This is shown in the figure by the progressively larger delay D1-Dn.
As described above, by controlling the number of times that the reset pointer of frame FN+1 is not incremented when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter and after the integration period has elapsed, the scan rate of the readout pointer of frame FN+1 can be set to match the slowed scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the lower readout mode.
In this figure, as with the embodiment depicted in
In this embodiment, the reset pointer of frame FN+1 is incremented at a second scan rate represented by the slope of the dotted line 504b. Similarly, the readout pointer of frame FN+1 is incremented at the second scan rate represented by the slope of the solid line 506b. In order to effect the second scan rate, and as noted with respect to other embodiments described herein, the reset pointer of frame FN+1 may selectively reset multiple rows at a time when the readout pointer of frame FN increments.
In addition, the image sensor is configured to delay resetting the first row of frame FN+1. The delay, referred to as the staggered start delay 508, is set such that the reset pointer of frame FN+1 does not overtake the readout pointer of frame FN.
In many cases, the staggered start delay 508 has the effect of decreasing integration time in the frame FN+1. In these examples, one or more gain correction operations can be optionally performed on the frame FN+1.
In this manner, by selectively incrementing multiple rows simultaneously, the reset pointer of frame FN+1 gradually advances toward the readout pointer of frame FN. This is shown in the figure by the progressively smaller delays D1-Dn.
As described above, by controlling the number of times and the manner in which multiple rows of frame FN+1 are simultaneously reset when the readout pointer of frame FN is incremented, the image sensor can effectively control the scan rate of the reset pointer of frame FN+1. Thereafter, and after the integration period has elapsed, the scan rate of the readout pointer of frame FN+1 can be set to match the faster scan rate of the reset pointer of frame FN+1, thereby transitioning the image sensor to the higher readout mode.
More specifically, the operation of incrementing the readout pointer of frame FN is shown by the step-wise function 602 and the operation of incrementing the reset pointer of frame FN+1 is shown by the step-wise function 604. As shown by the reference line 606, the rising edges of the step-wise function 602 and the step-wise function 604 are in phase, thereby reducing switching noise in the image sensor. Phrased in another way, the readout pointer incrementation illustrated by the step-wise function 602 is in phase with the reset pointer incrementation illustrated by the step-wise function 604.
In this embodiment, however, certain selected steps of the step-wise function 604 are skipped and/or repeated. For example, incrementing of the step 608 is skipped once (e.g., the respective row is reset at least once), effectively doubling the step's duration. Similarly, the step 610 is skipped once (e.g., the respective row is reset at least once), effectively doubling the step's duration. Other steps, such as the step 612 may not be skipped. In one phrasing, the step-wise function 604 is selectively “jittered” in order to affect a scan rate different from the step-wise function 602 while still maintaining time/phase synchronization with the step-wise function 602.
As with the embodiment described in reference to
More specifically, as with the embodiment described in reference to
In this embodiment, however, certain selected steps of the step-wise function 604 are performed in groups. In other words, certain rows can be reset at the same time. For example, each step in the group of three steps 614 is reset at the same time. In one phrasing, the step-wise function 604 is selectively “jittered” in order to affect a scan rate different from the step-wise function 602 while still maintaining time/phase synchronization with the step-wise function 602.
As with the embodiment described in reference to
It may be appreciated that the foregoing description of
For example, it may be appreciated that a linear or substantially linear relationship between the scan rates of the readout pointer of frame FN and the reset pointer of frame FN+1 may not be required. In some cases, the scan rate of the reset pointer of frame FN+1 may vary according to a curve or function.
Further, although only two readout modes are described in reference to the foregoing embodiments, it may be appreciated that an image sensor such as described herein can transition to and between any number of readout modes. In some examples, multiple transitions can occur in sequence. For example, the image sensor may transition from a first readout mode at frame FN, to a second readout mode at frame FN+1, to a third readout mode at frame FN+2.
Thus, it is understood that the foregoing descriptions of specific embodiments are presented for the limited purposes of illustration and description. These descriptions are not exhaustive nor intended to limit the disclosure to the precise forms recited herein. To the contrary, it will be apparent to one of ordinary skill in the art that many modifications and variations are possible in view of the above teachings.
The method 700 includes operation 702 in which a target readout mode is selected, which may specify a new line timing and/or a new scan rate, such as described above. Next, at operation 704, the scan rate of the current line timing of the current frame can be used to modify the line timing of the next frame (see, e.g.,
One may appreciate that although many embodiments are disclosed above, that the operations and steps presented with respect to methods and techniques described herein are meant as exemplary and accordingly are not exhaustive. One may further appreciate that alternate step order or fewer or additional operations may be required or desired for particular embodiments.
Although the disclosure above is described in terms of various exemplary embodiments and implementations, it should be understood that the various features, aspects and functionality described in one or more of the individual embodiments are not limited in their applicability to the particular embodiment with which they are described, but instead can be applied, alone or in various combinations, to one or more of the some embodiments of the invention, whether or not such embodiments are described and whether or not such features are presented as being a part of a described embodiment. Thus, the breadth and scope of the present invention should not be limited by any of the above-described exemplary embodiments but are instead defined by the claims herein presented.
This application is a continuation of U.S. patent application Ser. No. 16/226,491, filed Dec. 19, 2018, which claims the benefit under 35 U.S.C. § 119(e) of U.S. Provisional Patent Application No. 62/700,113, filed on Jul. 18, 2018, the contents of which are incorporated herein by reference as if fully disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
4686572 | Takatsu | Aug 1987 | A |
4686648 | Fossum | Aug 1987 | A |
5105264 | Erhardt et al. | Apr 1992 | A |
5329313 | Keith | Jul 1994 | A |
5396893 | Oberg et al. | Mar 1995 | A |
5471515 | Fossum et al. | Nov 1995 | A |
5541402 | Ackland | Jul 1996 | A |
5550677 | Schofield et al. | Aug 1996 | A |
5781312 | Noda | Jul 1998 | A |
5841126 | Fossum et al. | Nov 1998 | A |
5880459 | Pryor et al. | Mar 1999 | A |
5949483 | Fossum et al. | Sep 1999 | A |
6008486 | Stam et al. | Dec 1999 | A |
6040568 | Caulfield et al. | Mar 2000 | A |
6233013 | Hosier et al. | May 2001 | B1 |
6348929 | Acharya et al. | Feb 2002 | B1 |
6448550 | Nishimura | Sep 2002 | B1 |
6528833 | Lee et al. | Mar 2003 | B2 |
6541751 | Bidermann | Apr 2003 | B1 |
6670904 | Yakovlev | Dec 2003 | B1 |
6713796 | Fox | Mar 2004 | B1 |
6714239 | Guidash | Mar 2004 | B2 |
6798453 | Kaifu | Sep 2004 | B1 |
6816676 | Bianchi et al. | Nov 2004 | B2 |
6905470 | Lee et al. | Jun 2005 | B2 |
6931269 | Terry | Aug 2005 | B2 |
6956605 | Hashimoto | Oct 2005 | B1 |
6982759 | Goto | Jan 2006 | B2 |
7075049 | Rhodes et al. | Jul 2006 | B2 |
7084914 | Van Blerkom | Aug 2006 | B2 |
7091466 | Bock | Aug 2006 | B2 |
7119322 | Hong | Oct 2006 | B2 |
7133073 | Neter | Nov 2006 | B1 |
7259413 | Rhodes | Aug 2007 | B2 |
7262401 | Hopper et al. | Aug 2007 | B2 |
7271835 | Iizuka | Sep 2007 | B2 |
7282028 | Kim et al. | Oct 2007 | B2 |
7319218 | Krymski | Jan 2008 | B2 |
7332786 | Altice | Feb 2008 | B2 |
7342270 | Kuwazawa | Mar 2008 | B2 |
7390687 | Boettiger | Jun 2008 | B2 |
7415096 | Sherman | Aug 2008 | B2 |
7437013 | Anderson | Oct 2008 | B2 |
7443421 | Stavely et al. | Oct 2008 | B2 |
7446812 | Ando et al. | Nov 2008 | B2 |
7471315 | Silsby et al. | Dec 2008 | B2 |
7502054 | Kalapathy | Mar 2009 | B2 |
7525168 | Hsieh | Apr 2009 | B2 |
7554067 | Zarnoski et al. | Jun 2009 | B2 |
7555158 | Park et al. | Jun 2009 | B2 |
7622699 | Sakakibara et al. | Nov 2009 | B2 |
7626626 | Panicacci | Dec 2009 | B2 |
7636109 | Nakajima et al. | Dec 2009 | B2 |
7671435 | Ahn | Mar 2010 | B2 |
7728351 | Shim | Jun 2010 | B2 |
7733402 | Egawa et al. | Jun 2010 | B2 |
7737475 | Hynecek | Jun 2010 | B2 |
7742090 | Street | Jun 2010 | B2 |
7764312 | Ono et al. | Jul 2010 | B2 |
7773138 | Lahav et al. | Aug 2010 | B2 |
7786543 | Hsieh | Aug 2010 | B2 |
7796171 | Gardner | Sep 2010 | B2 |
7817198 | Kang et al. | Oct 2010 | B2 |
7821547 | Suzuki | Oct 2010 | B2 |
7873236 | Li et al. | Jan 2011 | B2 |
7880785 | Gallagher | Feb 2011 | B2 |
7884402 | Ki | Feb 2011 | B2 |
7906826 | Martin et al. | Mar 2011 | B2 |
7952121 | Arimoto | May 2011 | B2 |
7952635 | Lauxtermann | May 2011 | B2 |
7982789 | Watanabe et al. | Jul 2011 | B2 |
8026966 | Altice | Sep 2011 | B2 |
8032206 | Farazi et al. | Oct 2011 | B1 |
8089036 | Manabe et al. | Jan 2012 | B2 |
8089524 | Urisaka | Jan 2012 | B2 |
8094232 | Kusaka | Jan 2012 | B2 |
8116540 | Dean | Feb 2012 | B2 |
8140143 | Picard et al. | Mar 2012 | B2 |
8153947 | Barbier et al. | Apr 2012 | B2 |
8159570 | Negishi | Apr 2012 | B2 |
8159588 | Boemler | Apr 2012 | B2 |
8164669 | Compton et al. | Apr 2012 | B2 |
8174595 | Honda et al. | May 2012 | B2 |
8184188 | Yaghmai | May 2012 | B2 |
8194148 | Doida | Jun 2012 | B2 |
8194165 | Border et al. | Jun 2012 | B2 |
8222586 | Lee | Jul 2012 | B2 |
8227844 | Adkisson | Jul 2012 | B2 |
8233071 | Takeda | Jul 2012 | B2 |
8259228 | Wei et al. | Sep 2012 | B2 |
8310577 | Neter | Nov 2012 | B1 |
8324553 | Lee | Dec 2012 | B2 |
8330839 | Compton et al. | Dec 2012 | B2 |
8338856 | Tai et al. | Dec 2012 | B2 |
8340407 | Kalman | Dec 2012 | B2 |
8350940 | Smith et al. | Jan 2013 | B2 |
8400546 | Itano et al. | Mar 2013 | B2 |
8441545 | Hoda et al. | May 2013 | B2 |
8456540 | Egawa | Jun 2013 | B2 |
8456559 | Yamashita | Jun 2013 | B2 |
8462247 | Kim | Jun 2013 | B2 |
8508637 | Han et al. | Aug 2013 | B2 |
8514308 | Itonaga et al. | Aug 2013 | B2 |
8520913 | Dean | Aug 2013 | B2 |
8546737 | Tian et al. | Oct 2013 | B2 |
8547388 | Cheng | Oct 2013 | B2 |
8575531 | Hynecek et al. | Nov 2013 | B2 |
8581992 | Hamada | Nov 2013 | B2 |
8594170 | Mombers et al. | Nov 2013 | B2 |
8619163 | Ogua | Dec 2013 | B2 |
8619170 | Mabuchi | Dec 2013 | B2 |
8629484 | Ohri et al. | Jan 2014 | B2 |
8634002 | Kita | Jan 2014 | B2 |
8648947 | Sato et al. | Feb 2014 | B2 |
8723975 | Solhusvik | May 2014 | B2 |
8730345 | Watanabe | May 2014 | B2 |
8754983 | Sutton | Jun 2014 | B2 |
8755854 | Addison et al. | Jun 2014 | B2 |
8759736 | Yoo | Jun 2014 | B2 |
8767104 | Makino et al. | Jul 2014 | B2 |
8803990 | Smith | Aug 2014 | B2 |
8810703 | Mabuchi | Aug 2014 | B2 |
8817154 | Manabe et al. | Aug 2014 | B2 |
8860871 | Aoki | Oct 2014 | B2 |
8879686 | Okada et al. | Nov 2014 | B2 |
8902330 | Theuwissen | Dec 2014 | B2 |
8902341 | Machida | Dec 2014 | B2 |
8908062 | Ito | Dec 2014 | B2 |
8908073 | Minagawa | Dec 2014 | B2 |
8923994 | Laikari et al. | Dec 2014 | B2 |
8934030 | Kim et al. | Jan 2015 | B2 |
8936552 | Kateraas et al. | Jan 2015 | B2 |
8946610 | Iwabuchi et al. | Feb 2015 | B2 |
8981517 | Oshiyama et al. | Mar 2015 | B2 |
8982237 | Chen | Mar 2015 | B2 |
8982260 | Eshraghian et al. | Mar 2015 | B2 |
8988598 | Ovsiannikov | Mar 2015 | B2 |
9001251 | Smith et al. | Apr 2015 | B2 |
9041837 | Li | May 2015 | B2 |
9017748 | Theuwissen | Jun 2015 | B2 |
9054009 | Oike et al. | Jun 2015 | B2 |
9066017 | Geiss | Jun 2015 | B2 |
9066660 | Watson et al. | Jun 2015 | B2 |
9088727 | Trumbo | Jul 2015 | B2 |
9094623 | Kawaguchi | Jul 2015 | B2 |
9099604 | Roy | Aug 2015 | B2 |
9100597 | Hu | Aug 2015 | B2 |
9106859 | Kizuna et al. | Aug 2015 | B2 |
9131171 | Aoki et al. | Sep 2015 | B2 |
9154750 | Pang | Oct 2015 | B2 |
9225948 | Hasegawa | Dec 2015 | B2 |
9232150 | Kleekajai et al. | Jan 2016 | B2 |
9232161 | Suh | Jan 2016 | B2 |
9270906 | Peng et al. | Feb 2016 | B2 |
9276031 | Wan | Mar 2016 | B2 |
9277144 | Kleekajai et al. | Mar 2016 | B2 |
9287304 | Park et al. | Mar 2016 | B2 |
9287423 | Mori et al. | Mar 2016 | B2 |
9288380 | Nomura | Mar 2016 | B2 |
9288404 | Papiashvili | Mar 2016 | B2 |
9293500 | Sharma et al. | Mar 2016 | B2 |
9307161 | Lee | Apr 2016 | B2 |
9319611 | Fan | Apr 2016 | B2 |
9344649 | Bock | May 2016 | B2 |
9380245 | Guidash | Jun 2016 | B1 |
9392237 | Toyoda | Jul 2016 | B2 |
9438258 | Yoo | Sep 2016 | B1 |
9445018 | Fettig et al. | Sep 2016 | B2 |
9451887 | Watson et al. | Sep 2016 | B2 |
9467553 | Heo et al. | Oct 2016 | B2 |
9473706 | Malone et al. | Oct 2016 | B2 |
9479688 | Ishii | Oct 2016 | B2 |
9490285 | Itonaga | Nov 2016 | B2 |
9497397 | Kleekajai et al. | Nov 2016 | B1 |
9503616 | Taniguchi et al. | Nov 2016 | B2 |
9521337 | Shen | Dec 2016 | B1 |
9538067 | Hamada | Jan 2017 | B2 |
9538106 | McMahon et al. | Jan 2017 | B2 |
9549099 | Fan | Jan 2017 | B2 |
9571758 | Hashimoto et al. | Feb 2017 | B2 |
9584743 | Lin et al. | Feb 2017 | B1 |
9584744 | Lenchenkov et al. | Feb 2017 | B2 |
9596420 | Fan et al. | Mar 2017 | B2 |
9596423 | Molgaard | Mar 2017 | B1 |
9609250 | Lee et al. | Mar 2017 | B2 |
9654689 | Gleason | May 2017 | B2 |
9661210 | Haneda | May 2017 | B2 |
9666618 | Meynants | May 2017 | B2 |
9686485 | Agranov et al. | Jun 2017 | B2 |
9700240 | Letchner et al. | Jul 2017 | B2 |
9741754 | Li et al. | Aug 2017 | B2 |
9749556 | Fettig et al. | Aug 2017 | B2 |
9754994 | Koo et al. | Sep 2017 | B2 |
9774318 | Song | Sep 2017 | B2 |
9781368 | Song | Oct 2017 | B2 |
9819890 | Wang et al. | Nov 2017 | B2 |
9857469 | Oggier et al. | Jan 2018 | B2 |
9888198 | Mauritzson et al. | Feb 2018 | B2 |
9894304 | Smith | Feb 2018 | B1 |
9912883 | Lee | Mar 2018 | B1 |
9936105 | Furuya | Apr 2018 | B2 |
9936151 | Wang et al. | Apr 2018 | B2 |
9952323 | Deane | Apr 2018 | B2 |
9973678 | Mandelli et al. | May 2018 | B2 |
10019294 | van der Lugt et al. | Jul 2018 | B2 |
10044954 | Ikeda et al. | Aug 2018 | B2 |
10051217 | Kondo | Aug 2018 | B2 |
10104318 | Smith | Oct 2018 | B2 |
10120446 | Pance et al. | Nov 2018 | B2 |
10136090 | Vogelsang et al. | Nov 2018 | B2 |
10178329 | Vogelsang | Jan 2019 | B2 |
10212378 | Negishi | Feb 2019 | B2 |
10217889 | Dhulla et al. | Feb 2019 | B2 |
10249660 | Guidash et al. | Apr 2019 | B2 |
10263032 | Wan | Apr 2019 | B2 |
10269857 | Wu et al. | Apr 2019 | B2 |
10271037 | Oh | Apr 2019 | B2 |
10285626 | Kestelli et al. | May 2019 | B1 |
10306167 | Shimasaki | May 2019 | B2 |
10334181 | Guenter et al. | Jun 2019 | B2 |
10379317 | Shimokawa et al. | Aug 2019 | B2 |
10205904 | Kobayashi | Sep 2019 | B2 |
10431608 | Ebihara | Oct 2019 | B2 |
10440301 | Li et al. | Oct 2019 | B2 |
10447950 | Wang | Oct 2019 | B2 |
10484627 | Zhou | Nov 2019 | B2 |
10609348 | Agranov et al. | Mar 2020 | B2 |
10630920 | Matsunaga | Apr 2020 | B2 |
10630929 | Koizumi et al. | Apr 2020 | B2 |
10748955 | Oh et al. | Aug 2020 | B2 |
10775605 | Ollila | Sep 2020 | B2 |
10848693 | Agranov et al. | Nov 2020 | B2 |
10943935 | Li et al. | Mar 2021 | B2 |
10943940 | Wu et al. | Mar 2021 | B2 |
10951838 | Eyama | Mar 2021 | B2 |
11252381 | Pang et al. | Feb 2022 | B2 |
11258993 | Hoshino | Feb 2022 | B2 |
11546532 | Bahukhandi et al. | Jan 2023 | B1 |
20030036685 | Goodman et al. | Feb 2003 | A1 |
20040207836 | Chhibber et al. | Oct 2004 | A1 |
20050026332 | Fratti et al. | Feb 2005 | A1 |
20060274161 | Ing et al. | Dec 2006 | A1 |
20070263099 | Motta et al. | Nov 2007 | A1 |
20080049132 | Suzuki | Feb 2008 | A1 |
20080177162 | Bae et al. | Jul 2008 | A1 |
20080315198 | Jung | Dec 2008 | A1 |
20090096901 | Bae et al. | Apr 2009 | A1 |
20090101914 | Hirotsu et al. | Apr 2009 | A1 |
20090146234 | Luo et al. | Jun 2009 | A1 |
20090201400 | Zhang et al. | Aug 2009 | A1 |
20090219266 | Lim et al. | Sep 2009 | A1 |
20100134631 | Voth | Jun 2010 | A1 |
20110080500 | Wang et al. | Apr 2011 | A1 |
20110109776 | Kawai | May 2011 | A1 |
20110156197 | Tivarus et al. | Jun 2011 | A1 |
20110164162 | Kato | Jul 2011 | A1 |
20120092541 | Tuulos et al. | Apr 2012 | A1 |
20120098964 | Oggier | Apr 2012 | A1 |
20130147981 | Wu | Jun 2013 | A1 |
20140071321 | Seyama | Mar 2014 | A1 |
20140078356 | Vaartstra | Mar 2014 | A1 |
20140078381 | Ovsiannikov | Mar 2014 | A1 |
20150062391 | Murata | Mar 2015 | A1 |
20150350583 | Mauritzson | Dec 2015 | A1 |
20160050379 | Jiang et al. | Feb 2016 | A1 |
20160191826 | Furuya | Jun 2016 | A1 |
20160205311 | Mandelli | Jul 2016 | A1 |
20160219232 | Murata | Jul 2016 | A1 |
20160255262 | Lee | Sep 2016 | A1 |
20160286108 | Fettig | Sep 2016 | A1 |
20160309140 | Wang | Oct 2016 | A1 |
20160323524 | Smith | Nov 2016 | A1 |
20160344920 | Iwahara | Nov 2016 | A1 |
20160344958 | Kondo | Nov 2016 | A1 |
20170047363 | Choi et al. | Feb 2017 | A1 |
20170111600 | Wang | Apr 2017 | A1 |
20170195596 | Vogelsang | Jul 2017 | A1 |
20170295326 | Eyama | Oct 2017 | A1 |
20210021778 | Bairo | Jan 2021 | A1 |
20210281789 | McMahon | Sep 2021 | A1 |
20220046196 | Li | Feb 2022 | A1 |
20220375986 | Rosenblum et al. | Nov 2022 | A1 |
Number | Date | Country |
---|---|---|
1630350 | Jun 2005 | CN |
1774032 | May 2006 | CN |
1833429 | Sep 2006 | CN |
1842138 | Oct 2006 | CN |
1947414 | Apr 2007 | CN |
101189885 | May 2008 | CN |
101221965 | Jul 2008 | CN |
101233763 | Jul 2008 | CN |
101472059 | Jul 2009 | CN |
101567977 | Oct 2009 | CN |
101622859 | Jan 2010 | CN |
101739955 | Jun 2010 | CN |
101754029 | Jun 2010 | CN |
101803925 | Aug 2010 | CN |
102036020 | Apr 2011 | CN |
102067584 | May 2011 | CN |
102208423 | Oct 2011 | CN |
102451160 | May 2012 | CN |
102668542 | Sep 2012 | CN |
102820309 | Dec 2012 | CN |
102821255 | Dec 2012 | CN |
103024297 | Apr 2013 | CN |
103051843 | Apr 2013 | CN |
103329513 | Sep 2013 | CN |
103546702 | Jan 2014 | CN |
104041009 | Sep 2014 | CN |
104052919 | Sep 2014 | CN |
204761615 | Nov 2015 | CN |
205211754 | May 2016 | CN |
814606 | Feb 2002 | EP |
1763228 | Mar 2007 | EP |
2023611 | Feb 2009 | EP |
2107610 | Oct 2009 | EP |
2230690 | Sep 2010 | EP |
2512126 | Oct 2012 | EP |
2601833 | Jun 2022 | GB |
S61123287 | Jun 1986 | JP |
2000059697 | Feb 2000 | JP |
2001211455 | Aug 2001 | JP |
2001358994 | Dec 2001 | JP |
2004111590 | Apr 2004 | JP |
2005318504 | Nov 2005 | JP |
2006287361 | Oct 2006 | JP |
2007504670 | Mar 2007 | JP |
2007516654 | Jun 2007 | JP |
2008507908 | Mar 2008 | JP |
2008271280 | Nov 2008 | JP |
2008543061 | Nov 2008 | JP |
2009021809 | Jan 2009 | JP |
4255223 | Apr 2009 | JP |
2009159186 | Jul 2009 | JP |
2009212909 | Sep 2009 | JP |
2009296465 | Dec 2009 | JP |
2010080604 | Apr 2010 | JP |
2010114834 | May 2010 | JP |
2011040926 | Feb 2011 | JP |
2011049697 | Mar 2011 | JP |
2011091775 | May 2011 | JP |
2011216970 | Oct 2011 | JP |
2011217315 | Oct 2011 | JP |
2011097646 | Dec 2011 | JP |
2012010306 | Jan 2012 | JP |
2012019516 | Jan 2012 | JP |
2012513160 | Jun 2012 | JP |
2013005397 | Jan 2013 | JP |
2013051523 | Mar 2013 | JP |
2013070240 | Apr 2013 | JP |
2013529035 | Jul 2013 | JP |
20030034424 | May 2003 | KR |
20030061157 | Jul 2003 | KR |
20050103732 | Nov 2005 | KR |
20080069851 | Jul 2008 | KR |
20100008239 | Jan 2010 | KR |
20100065084 | Jun 2010 | KR |
20130074459 | Jul 2013 | KR |
200520551 | Jun 2005 | TW |
200803481 | Jan 2008 | TW |
201110689 | Mar 2011 | TW |
201301881 | Jan 2013 | TW |
WO 05041304 | May 2005 | WO |
WO 06014641 | Feb 2006 | WO |
WO 06130443 | Dec 2006 | WO |
WO 07049900 | May 2007 | WO |
WO 10120945 | Oct 2010 | WO |
WO 12053363 | Apr 2012 | WO |
WO 12088338 | Jun 2012 | WO |
WO 12122572 | Sep 2012 | WO |
WO 12138687 | Oct 2012 | WO |
WO 13008425 | Jan 2013 | WO |
WO 13179018 | Dec 2013 | WO |
WO 13179020 | Dec 2013 | WO |
WO 19102887 | May 2019 | WO |
WO 20228392 | Nov 2020 | WO |
Entry |
---|
U.S. Appl. No. 16/805,558, filed Feb. 28, 2020, Lee et al. |
U.S. Appl. No. 17/075,547, filed Oct. 20, 2020, Agranov et al. |
U.S. Appl. No. 17/380,852, filed Jul. 20, 2021, Li et al. |
Aoki, et al., “Rolling-Shutter Distortion-Free 3D Stacked Image Sensor with -160dB Parasitic Light Sensitivity In-Pixel Storage Node,” ISSCC 2013, Session 27, Image Sensors, 27.3 27.3 A, Feb. 20, 2013, retrieved on Apr. 11, 2014 from URL:http://ieeexplore.ieee.org/stamp/stamp.jsp?tp-&arnumber=6487824. |
Elgendi, “On the Analysis of Fingertip Photoplethysmogram Signals,” Current Cardiology Reviews, 2012, vol. 8, pp. 14-25. |
Feng, et al., “On the Stoney Formula for a Thin Film/Substrate System with Nonuniform Substrate Thickness,” Journal of Applied Mechanics, Transactions of the ASME, vol. 74, Nov. 2007, pp. 1276-1281. |
Fu, et al., “Heart Rate Extraction from Photoplethysmogram Waveform Using Wavelet Multui-resolution Analysis,” Journal of Medical and Biological Engineering, 2008, vol. 28, No. 4, pp. 229-232. |
Han, et al., “Artifacts in wearable photoplethysmographs during daily life motions and their reduction with least mean square based active noise cancellation method,” Computers in Biology and Medicine, 2012, vol. 42, pp. 387-393. |
Lopez-Silva, et al., “Heuristic Algorithm for Photoplethysmographic Heart Rate Tracking During Maximal Exercise Test,” Journal of Medical and Biological Engineering, 2011, vol. 12, No. 3, pp. 181-188. |
Santos, et al., “Accelerometer-assisted PPG Measurement During Physical Exercise Using the LAVIMO Sensor System,” Acta Polytechnica, 2012, vol. 52, No. 5, pp. 80-85. |
Sarkar, et al., “Fingertip Pulse Wave (PPG signal) Analysis and Heart Rate Detection,” International Journal of Emerging Technology and Advanced Engineering, 2012, vol. 2, No. 9, pp. 404-407. |
Schwarzer, et al., On the determination of film stress from substrate bending: STONEY'S formula and its limits, Jan. 2006, 19 pages. |
Shen et al., “Stresses, Curvatures, and Shape Changes Arising from Patterned Lines on Silicon Wafers,” Journal of Applied Physics, vol. 80, No. 3, Aug. 1996, pp. 1388-1398. |
Yan, et al., “Reduction of motion artifact in pulse oximetry by smoothed pseudo Wigner-Ville distribution,” Journal of NeuroEngineering and Rehabilitation, 2005, vol. 2, No. 3, pp. 1-9. |
Yousefi, et al., “Adaptive Cancellation of Motion Artifact in Wearable Biosensors,” 34th Annual International Conference of the IEEE EMBS, San Diego, California, Aug./Sep. 2012, pp. 2004-2008. |
Fontaine, “The State-of-the-Art of Smartphone Imagers,” 2019 International Image Sensor Workshop, Snowbird, Utah, Jun. 23-27, 2019, 3 pages. |
Jang et al., “A new PDAF correction method of CMOS image sensor with Nonacell and Super PD to improve image quality in binning mode,” IS&T International Symposium on Electronic Imaging 2021, Image Quality and System Performance XVIII, 5 pages. |
Number | Date | Country | |
---|---|---|---|
20210281789 A1 | Sep 2021 | US |
Number | Date | Country | |
---|---|---|---|
62700113 | Jul 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16226491 | Dec 2018 | US |
Child | 17329010 | US |