The present invention generally relates to high dynamic range (HDR), and more particularly to a high-dynamic-range (HDR) detecting system and method.
Dynamic range (DR) is the ratio between the largest value and the smallest value, for example, used in the context of signals such as light.
High Dynamic Range (HDR) is a video and image technology that improves the way light is represented in displays. HDR offers the possibility to represent substantially brighter highlights, darker shadows, more details and more colorful colors, in order to make better use of displays that have high brightness, contrast and colors capabilities.
In conventional HDR detecting systems, camera capture control is commonly based on image intensity distribution, which may be insufficient for user performance due to artificial intelligence (AI) systems may accept different intensity distribution and object details. Moreover, HDR image information of wide dynamic range intensity distribution in real scene may ordinarily incur higher cost in image capture and computation.
A need has thus arisen to propose a novel scheme to overcome drawbacks of the conventional HDR detecting systems.
In view of the foregoing, it is an object of the embodiment of the present invention to provide a high-dynamic-range (HDR) detecting system and method performed with reduced cost and computation. In one embodiment, the HDR detecting system and method may provide metadata such as objects with associated locations and objects count for an application.
According to one embodiment, a high-dynamic-range (HDR) detecting system includes a camera module and a vision module. The camera module includes an image sensor. The vision module receives data generated by the camera module and accordingly transmits feedback to the camera module. The camera module comprises a multiple-exposure unit that provides different exposure settings, according to which a plurality of images are captured by the image sensor. The vision module comprises an object detection device that detects objects in the plurality of images, thereby resulting in object detection results as metadata.
The HDR detecting system 100 of the embodiment may include a camera module 11 and a vision module 12. Specifically, the vision module 12 receives (original) data generated by the camera module 11, and accordingly transmits feedback (information) to the camera module 11. The HDR detecting system 100 of the embodiment may also include a display device 13 operatively coupled to the vision module 12 for human viewing.
In the embodiment, the camera module 11 may include an image sensor 111 configured to capture an image. The camera module 11 may include a capture control device 112, which may include an auto-exposure (or automatic-exposure or AE) unit 1121, such as an exposure meter, configured to automatically calculate and adjust exposure settings for the image sensor 111.
The camera module 11 of the embodiment may include an HDR device 113, such as an image signal processor (ISP), which may include an HDR fusion device 1131 configured to blend multiple images of the same subject captured with different exposure settings (i.e., bracketing) into a single image; and may include a tone mapping device 1132 configured to map one set of colors to another in order to approximate appearance of a high-dynamic-range image.
In the embodiment, the vision module 12, such as an image signal processor (ISP), may include a motion detection device 121 configured to detect a change in a position of an object relative to its surroundings. Generally speaking, the image signal processor of the vision module 12 may perform image signal processing such as distortion correction, sample-rate conversion (e.g., upscaling and downscaling), de-noise filtering, de-blur filtering, gamma correction and contrast limited adaptive histogram equalization.
Referring to the HDR detecting method 200 as illustrated in
Next, in step 22, a captured image (by the image sensor 111), as the data, is sent to the vision module 12, and is then subject to motion detection by the motion detection device 121, thereby resulting in a motion detection result as the feedback.
In step 23, HDR mode is determined according to the AE statistics and the motion detection result (i.e., the feedback). In the embodiment, HDR mode is decided if one or more of the following criteria have been met: change in the motion detection is less than a predetermined threshold; non-optimized (intensity) distribution after AE converges; single exposure cannot obtain enough texture details for all regions (of the image). On the other hand, non-HDR mode is decided if one or more of the following criteria have been met: change in the motion detection is greater than the predetermined threshold; AE does not converge or scene change happens; single exposure can substantially achieve intensity range of real scene, and obtain enough texture details for all regions; power or frame rate (expressed in frame per second or fps) is limited.
If HDR mode is decided in step 23, multiple images are captured with different exposure settings and blended into a single image by the HDR fusion device 1131 (step 24), followed by performing tone mapping (step 25) by the tone mapping device 1132. It is noted that HDR fusion and tone mapping may be performed by conventional techniques, details of which are thus omitted for brevity. The flow then goes to step 26, in which the image is displayed (by the display device 13) for human viewing or is used for computer vision. If non-HDR mode is decided in step 23, the flow directly goes to step 26.
The HDR detecting system 300 of the embodiment is similar to the HDR detecting system 100 of
The vision module 12 of the embodiment may further included a neural network device 122 (for computer vision) configured to perform object detection or recognition, such as pedestrian detection by a pedestrian detection unit 1221, face detection by a face detection unit 1222, or face recognition by a face recognition unit 1223. The neural network device 122 may obtain object features, such as locations, sizes and colors, which may be stored in a buffer (not shown).
Referring to the HDR detecting method 400 as illustrated in
In step 43, a mode is determined according to the AE statistics and the feedback. In the embodiment, one of linear mode, enhance mode and HDR mode (in order of image quality) is determined. Linear mode (adapted to normal light) is decided if single exposure can substantially obtain enough texture details and contrast for all regions (of the image). Enhance mode (adapted to low light or less detail) is decided if a local region has narrow intensity distribution. HDR mode (adapted to wide range illumination environment) is decided if single exposure cannot obtain enough texture details for all regions, or change in the motion detection is less than a predetermined threshold.
If linear mode is decided, the flow goes directly to step 44, in which the image is displayed (by the display device 13) for human viewing or is used for computer vision. If enhance mode is decided, object details such as contrast or edge may be enhanced in step 45, and the flow then goes to step 44. If HDR mode is decided, HDR fusion is performed by the HDR fusion device 1131 in step 46, tone mapping is performed by the tone mapping device 1132 in step 47, and the flow then goes to step 44.
Subsequently, in step 48, it determines whether requirement of image quality is met, for example, by a user or according to detection/recognition result (of the neural network device 122). If the requirement is not met, the mode is switched to next mode with higher image quality. For example, if linear mode fails, the mode is then switched to enhance mode; if enhance mode fails, the mode is then switched to HDR mode.
If the requirement is met, the flow goes to step 49, in which a region of interest (ROI) is determined, for example, according to the motion detection result (from the motion detection device 121) or detection/recognition result (from the neural network device 122).
Next, in step 50, it determines whether requirement of image quality is met, for example, by the user or according to detection/recognition result (of the neural network device 122) in a manner similar to step 48.
The HDR detecting system 500 of the embodiment is similar to the HDR detecting system 100 of
The vision module 12 may further include an object detection device 123 configured to detect objects in the images, thereby resulting in object detection results as metadata. The object detection device 123 may, for example, include a neural network device 122 of
Referring to the HDR detecting method 600 as illustrated in
In step 62, the object detection device 123 detects objects in the images, thereby resulting in object detection results as (plural) metadata. Specifically, the object detection results (or metadata) may include object features, such as locations, sizes and colors.
In one exemplary embodiment, a group of (e.g., three) low-resolution images are first captured with different exposure settings. If an object is detected in step 62, a group of (e.g., three) high-resolution images may be captured with different exposure settings next time, therefore improving recognition rate but with increased power consumption. Alternatively, a group of low-resolution images may be captured next time, therefore maintaining low power consumption.
If no object is detected in step 62, the flow goes back to step 61. Otherwise, the flow goes to step 63, in which the metadata are analyzed (to determine objects count and position distribution, for example).
In step 64, one or more applications 14 may be activated according to the metadata. In one example, the application 14 is related to system interaction, in which a signal is sent to trigger another equipment with different scene condition and detection result. In another example, the application 14 is related to providing auxiliary suggestion with current scene condition, people or object information. In a further example, the application 14 is related to human postures and tracking, object classification and analysis.
Although specific embodiments have been illustrated and described, it will be appreciated by those skilled in the art that various modifications may be made without departing from the scope of the present invention, which is intended to be limited solely by the appended claims.