The present disclosure relates to an apparatus and method for providing an image. More particularly, the present disclosure relates to an apparatus and method for providing an image having increased-resolution.
Mobile terminals were first developed to provide wireless communication between users. As technology has advanced, mobile terminals now provide many additional features beyond the simple telephone conversation. For example, mobile terminals are now able to provide advanced functions such as an alarm, a Short Messaging Service (SMS), a Multimedia Messaging Service (MMS), E-mail, games, short range communication, an image capturing function using a mounted digital camera, a multimedia function for providing audio and video content, a scheduling function, and many more. With the plurality of features now provided, a mobile terminal has effectively become a necessity of daily life for most people.
As is known in the art, an image may be captured by a digital camera mounted on the mobile terminal. For example, when a user selects an image capturing function, a Graphical User Interface (GUI) may be displayed, allowing the user to select a capturing button of the GUI to ultimately capture a desired image. When capturing an image, an image sensor of the digital camera is controlled to receive information on a plurality of photosites. However, the number of photosites may be limited and thus provide an image having a low resolution. Accordingly, there is a need for an improved apparatus and method for providing an image having super-resolution using an existing image sensor.
The above information is presented as background information only to assist with an understanding of the present disclosure. No determination has been made, and no assertion is made, as to whether any of the above might be applicable as prior art with regard to the present disclosure
Aspects of the present disclosure are to address at least the above-mentioned problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of the present disclosure is to provide an apparatus and method for providing an image having super-resolution.
In accordance with an aspect of the present disclosure, a method for increasing the resolution of an image is provided. The method includes capturing a plurality of frames of an image, determining a reference frame from among the plurality of frames, iteratively determining an offset of each of the plurality of frames to the reference frame until unity scaling is reached, and determining a pixel value for insertion between pixels of the reference frame.
In accordance with another aspect of the present disclosure, an apparatus for increasing the resolution of an image is provided. The apparatus includes a camera unit configured to capture a plurality of frames of an image, and a control unit configured to determine a reference frame from among the plurality of frames, to iteratively determine an offset of each of the plurality of frames to the reference frame until unity scaling is reached, and to determine a pixel value for insertion between pixels of the reference frame.
Other aspects, advantages, and salient features of the disclosure will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses various embodiments of the disclosure.
The above and other aspects, features, and advantages of various embodiments of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
Throughout the drawings, it should be noted that like reference numbers are used to depict the same or similar elements, features, and structures.
Detailed descriptions of various aspects of the present disclosure will be discussed below with reference to the attached drawings. The descriptions are set forth as examples only, and shall not limit the scope of the present disclosure.
The detailed description with reference to the accompanying drawings is provided to assist in a comprehensive understanding of various embodiments of the disclosure as defined by the claims and their equivalents. It includes various specific details to assist in that understanding but these are to be regarded as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the disclosure. In addition, descriptions of well-known functions and constructions are omitted for clarity and conciseness.
The terms and words used in the following description and claims are not limited to the bibliographical meanings, but, are merely used by the inventor to enable a clear and consistent understanding of the disclosure. Accordingly, it should be apparent to those skilled in the art that the following description of various embodiments of the present disclosure are provided for illustration purpose only and not for the purpose of limiting the disclosure as defined by the appended claims and their equivalents.
It is to be understood that the singular forms “a,” “an,” and “the” include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to “a component surface” includes reference to one or more of such surfaces.
By the term “substantially” it is meant that the recited characteristic, parameter, or value need not be achieved exactly, but that deviations or variations, including for example, tolerances, measurement error, measurement accuracy limitations and other factors known to those of skill in the art, may occur in amounts that do not preclude the effect the characteristic was intended to provide.
Unless defined differently, all terms used in the present disclosure, including technical or scientific terms, have meanings that are understood generally by a person having ordinary skill in the art. Ordinary terms that may be defined in a dictionary should be understood to have the meaning consistent with their context, and unless clearly defined in the present disclosure, should not be interpreted to be excessively idealistic or formalistic.
According to various embodiments of the present disclosure, an electronic device may include communication functionality. For example, an electronic device may be a smart phone, a tablet Personal Computer (PC), a mobile phone, a video phone, an e-book reader, a desktop PC, a laptop PC, a netbook PC, a Personal Digital Assistant (PDA), a Portable Multimedia Player (PMP), an MP3 player, a mobile medical device, a camera, a wearable device (e.g., a Head-Mounted Device (HMD), electronic clothes, electronic braces, an electronic necklace, an electronic appcessory, an electronic tattoo, or a smart watch), and/or the like.
According to various embodiments of the present disclosure, an electronic device may be a smart home appliance with communication functionality. A smart home appliance may be, for example, a television, a Digital Video Disk (DVD) player, an audio, a refrigerator, an air conditioner, a vacuum cleaner, an oven, a microwave oven, a washer, a dryer, an air purifier, a set-top box, a TV box (e.g., Samsung HomeSync™, Apple TV™, or Google TV™), a gaming console, an electronic dictionary, an electronic key, a camcorder, an electronic picture frame, and/or the like.
According to various embodiments of the present disclosure, an electronic device may be a medical device (e.g., Magnetic Resonance Angiography (MRA) device, a Magnetic Resonance Imaging (MRI) device, Computed Tomography (CT) device, an imaging device, or an ultrasonic device), a navigation device, a Global Positioning System (GPS) receiver, an Event Data Recorder (EDR), a Flight Data Recorder (FDR), an automotive infotainment device, a naval electronic device (e.g., naval navigation device, gyroscope, or compass), an avionic electronic device, a security device, an industrial or consumer robot, and/or the like.
According to various embodiments of the present disclosure, an electronic device may be furniture, part of a building/structure, an electronic board, electronic signature receiving device, a projector, various measuring devices (e.g., water, electricity, gas or electro-magnetic wave measuring devices), and/or the like that include communication functionality.
According to various embodiments of the present disclosure, an electronic device may be any combination of the foregoing devices. In addition, it will be apparent to one having ordinary skill in the art that an electronic device according to various embodiments of the present disclosure is not limited to the foregoing devices.
The term “super resolution” refers to a process of the related art that is performed to enhance an image using a sequence of captured frames of the image abstracted from a video. In the related art, motion within the video sequence, either through panning, or motion of objects within the sequence, is used to introduce non-redundant additional samples of the image. This sequence of images, or samples, is composited into a single static image wherein the non-redundant samples are used to fill in a higher order mosaic of potential sample point locations. From this, a higher resolution image (i.e., an image having super-resolution) may be generated.
As part of the present invention, it has been discovered that when a human being attempts to hold a camera still for a photograph, the user naturally “shakes” or moves the mobile terminal in a random pattern around the desired framing position. An aspect of this invention takes advantage of this natural movement.
Referring to
Referring to
Referring to
As evidenced by
Referring to
In
In
The example illustrated in
Referring to
Referring to
Referring to
While only two image captures are illustrated in the examples of
A first aspect of the present disclosure is to provide an apparatus and method for aligning (i.e., registering) one image relative to another image. In that regard, a progressive approximation method is provided. In more detail, an apparatus and method are provided that use a multi-resolution successive approximation of the alignment. This method has the advantage of being fast as well as robust in the presence of noise and repeating structures in images that may “fool” other alignment methods. In yet another embodiment of the present disclosure, the alignment method includes a morphing routine to correct for changes due to camera angle. This is important for handheld camera “snapshots” that are likely to be taken at fairly close range.
Referring to
In the example of
Referring to
The nine possible shift offsets 510 in Image B include an upper left offset (ul), an up offset (up), an upper right offset (ur), a left offset (lf), a center or no offset (cn), a right offset (rt), a lower left offset (ll), a down offset (dn), and a lower right offset (lr). The first iteration determines which of the nine possible shift offsets 510 will minimize the differences between pixel values of Image A and Image B. In an embodiment of the present disclosure, the difference in pixel values at each offset is determined according to Equation (1).
In Equation (1), it is assumed that the offset determination is performed with the full Red, Green, and Blue (RGB) data set when comparing the pixel values of Image A and Image B. However, as an alternative, the algorithms may be run on the green channels only as a method of saving power or processing resources. In that case, the offsets from the green channels can be used to composite all channels of the pixels. The results of each determination of Equation (1) are compared using Equation (2) to determine a value of Xoffset and Yoffset. offset.
Based on the results of Equation (2), a first displacement (i.e., an Xoffset and a Yoffset) of Image B with respect to Image A can be determined. That is, although all pixels of Image B are displaced according to the nine shift offsets, only one of the shifted images is chosen based on the result of Equation (2). Again, the displacement value is based on a scaled down image (e.g., a 1/64th image) such that the displacement value is only a rough approximation. To further refine the determination of the displacement, a second and subsequent iterations are needed. In the example of
Referring to
X
offset 1/32
=X
offset 1/64*2
Y
offset 1/32
=Y
offset 1/64*2 Equation (3)
Using Equation (3) to determine the starting values for the example of
In the example of
X
offset
=X
offset 1/64*2+new Xoffset
Y
offset
=Y
offset 1/64*2+new Yoffset Equation (4)
In the example of
Although only two iterations are illustrated in
Referring to
To achieve a resultant image having super-resolution, fractional pixel offsets are needed for insertion as samples between pixels of the reference image. As can be seen in Table 1, by the seventh iteration, the algorithm calculates a displacement to the nearest pixel. To achieve the necessary fractional pixel offsets, additional iterations are necessary. In one embodiment, the algorithm is further designed to iterate two more times beyond unity scaling. However, rather than using downscaled image data, both Image A and Image B are ‘up-sampled’ by a factor of two for each iteration beyond unity scaling. This up-sampling is reflected in Table 1.
In another embodiment, an optimization is performed in which Image B is re-sampled by a ½ pixel in all directions and compared to Image A at unity scale for the eighth iteration. For the final iteration, Image B is re-sampled to the nearest ¼ pixel in all directions and again compared to Image A at unity scale.
Referring to
Referring to
Although two images are illustrated in the above examples beginning with
To better determine a number of image captures that should be used to obtain a satisfactory image without imparting a significant computation burden, A combinatorial model was constructed to compute the cumulative probability of landing on a certain number of unique fractional states (referred to as the number of “Hits”) within a certain number of captured frames. Every new frame until the 4th frame reliably offers new resolution information. Thus, four Hits (4=2× resolution) is highly likely with only four frames, and seven captured frames reliably produces six hits (6=2.4× resolution). To achieve greater than 50% probability of 8 hits (8=2.8× resolution) at every locality of the super-resolution image, one would need to invest at least 10 frames of memory and processing.
Based on experimental results, it has been discovered that it is generally efficient to populate half (or less than half) of the available resolution states. Beyond this half way point, an increasing number of redundancies occur resulting in progressively less efficient use of available memory and processing resources. Hence, for super-resolution system implementations with tight constraints on frame memory and processing resources, it has been found that 10 or fewer frame captures is sufficient.
It has further been discovered that the number of image captures had an impact on noise reduction when performing the super resolution algorithm according to embodiments of the present disclosure. More specifically, it has been discovered that super-imposing low-resolution frames has a significant noise reduction benefit that increases proportionally by the square root of the number of frames. Based on experimental results, it has been discovered that when four frames are used for the super-resolution algorithm, noise is reduced to 50% its original value. When 10 frames are used, noise is reduced to 31% its original value. And when 25 frames are used, noise is reduced to 20% its original value. In terms of Signal to Noise Ratio (SNR), SNR is increased 2× with four frames, 3.2× with 10 frames, and 5× with 25 frames. Finally, in terms of SNR gain in (dB), four frames increased SNR gain by 6 dB, 10 frames increased SNR gain by 10 dB and 25 frames increased SNR gain by 14 dB.
Also for ease of discussion, the above examples beginning in
As part of the present disclosure, it was discovered that sequential image captures typically include perspective, scale, rotation, and translation differences caused by random camera tilt along various axes. Thus, an embodiment of the present disclosure provides an image registration algorithm that not only aligns (i.e., registers) sequential images but also transforms them to the same camera perspective.
Referring to
Referring to
Referring to
Referring to
When it is determined in operation 1107 that all image pairs have been considered, the process proceeds to operation 1115 at which the counter is again set to 1. In operation 1117, it is determined if the counter is less than or equal to the number of image captures N. The purpose of the counter in this instance is to ensure that operations 1119 to 1123 are performed for all image captures. In operation 1119, morphing is performed of image capture X according to the four corners determined in operation 1111. In operation 1121, the image capture X is scaled up by a factor of four to achieve a super-resolution for that image capture. In operation 1123, the pixel sums for image capture X are read into a sum buffer for later use in determining a final value for each pixel. In operation 1125, the counter is increased by 1 and the process returns to operation 1117.
When it is determined in operation 1117 that all image captures have been considered, the process proceeds to operation 1127 at which the counter is set to the number of pixels. In operation 1129, it is determined if the counter is equal to zero. The purpose of the counter in this instance is to ensure that all pixels of the image are considered. In operation 1131, an average pixel value is determined based on the sum of pixel values determined in operation 1123. In operation 1133, the counter is increased by 1 and the process returns to operation 1129.
When it is determined in operation 1129 that all pixels have been considered, the process proceeds to operation 1135 in which image sharpening may be performed by an appropriate filter. Notably, operation 1135 is optional.
Referring to
According to various embodiments of the present disclosure, the electronic device 1200 comprises at least one control unit 1210. The at least one control unit 1210 may be configured to operatively control the electronic device 1200. For example, the at least one control unit 1210 may control operation of the various components or units included in the electronic device 1200. The at least one control unit 1210 may transmit a signal to the various components included in the electronic device 1200 and control a signal flow between internal blocks of the electronic device 1200. The at least one control unit 1210 may be or otherwise include at least one processor. For example, the at least one control unit 1210 may include an Application Processor (AP), and/or the like.
The storage unit 1220 may be configured to store user data, and the like, as well a program which performs operating functions according to various embodiments of the present disclosure. The storage unit 1220 may include a non-transitory computer-readable storage medium. As an example, the storage unit 1220 may store a program for controlling general operation of an electronic device 1200, an Operating System (OS) which boots the electronic device 1200, and application program for performing other optional functions such as a camera function, a sound replay function, an image or video replay function, a signal strength measurement function, a route generation function, image processing, and the like. Further, the storage unit 1220 may store user data generated according to a user of the terminal 1220, such as, for example, a text message, a game file, a music file, a movie file, and the like. According to various embodiments of the present disclosure, the storage unit 1220 may store an application or a plurality of applications that individually or in combination operate a camera unit 1270 to capture (e.g., contemporaneously) one or more images of substantially the same viewpoint, and/or the like. According to various embodiments of the present disclosure, the storage unit 1220 may store an application or a plurality of applications that individually or in combination operate the image processing unit 1230 or the control unit 1210 to perform any of the functions, operations or steps as described above. The storage unit 1220 may store an application or a plurality of applications that individually or in combination operate the control unit 1210 and the communication unit 1260 to communicate with a counterpart electronic device to receive one or more images from the counterpart electronic device, and/or the like. The storage unit 1220 may store an application or a plurality of applications that individually or in combination operate display unit 1240 to display a graphical user interface, an image, a video, and/or the like.
The display unit 1240 displays information inputted by a user or information to be provided to the user as well as various menus of the electronic device 1200. For example, the display unit 1240 may provide various screens according to the user such as an idle screen, a message writing screen, a calling screen, a route planning screen, and the like. According to various embodiments of the present disclosure, the display unit 1240 may display an interface which the user may manipulate or otherwise enter inputs via a touch screen to enter selection of the function relating to the signal strength of the electronic device 1200. The display unit 1240 can be formed as a Liquid Crystal Display (LCD), an Organic Light Emitting Diode (OLED), an Active Matrix Organic Light Emitting Diode (AMOLED), and the like. However, various embodiments of the present disclosure are not limited to these examples. Further, the display unit 1240 can perform the function of the input unit 1250 if the display unit 1240 is formed as a touch screen.
The input unit 1250 may include input keys and function keys for receiving user input. For example, the input unit 1250 may include input keys and function keys for receiving an input of numbers or various sets of letter information, setting various functions, and controlling functions of the electronic device 1200. For example, the input unit 1250 may include a calling key for requesting a voice call, a video call request key for requesting a video call, a termination key for requesting termination of a voice call or a video call, a volume key for adjusting output volume of an audio signal, a direction key, and the like. In particular, according to various embodiments of the present disclosure, the input unit 1250 may transmit to the at least one control unit 1210 signals related to the operation of a camera unit (not shown), to selection of an image, to selection of a viewpoint, and/or the like. Such an input unit 1250 may be formed by one or a combination of input means such as a touch pad, a touchscreen, a button-type key pad, a joystick, a wheel key, and the like.
The communication unit 1260 may be configured for communicating with other electronic devices and/or networks. According to various embodiments of the present disclosure, the communication unit 1260 may be configured to communicate using various communication protocols and various communication transceivers. For example, the communication unit 1260 may be configured to communicate via Bluetooth technology, NFC technology, WiFi technology, 2G technology, 3G technology, LTE technology, or another wireless technology, and/or the like.
The camera unit 1270 may be configured to capture one or a plurality of images and provide the data of the captured one or more images to the control unit 1210 for processing.
Referring to
The pre-processor 1310 may be configured to receive an input 1301, such as data from an image sensor or digital camera. After processing the received data (e.g., lens shading, addressing flicker, etc.), the pre-processor provides the pre-processed data to the ISP 1340. The ISP 1340 performs additional functions on the data such as conversion from RGB format to Ycbcr format, white balancing, color saturation enhancement, and the like. Moreover, the ISP 1340 provides the received data to the super-resolution multi-frame processor 1350 for performing any or all of the functions as described above. Based on the additional memory needs of the super-resolution multi-frame processor 1350, a frame memory 1360 may also be provided. Upon completion of the super-resolution processing, the super-resolution multi-frame processor 1350 provides an output 1305 including an image having an enhanced resolution. The output may be provided to an external storage, a display unit, and the like. In a normal mode, the ISP 1340 provides an output for further processing to the FIIMD 1320 and the CREO 1330 which may ultimately provide an output signal to a display unit, such as display unit 1240 for
In an alternative embodiment, the pre-processor 1310 may output raw data directly to the super-resolution multi-frame processor 1350. In that case, the super-resolution multi-frame processor 1350 may output super raw data to the ISP 1340. However, this option requires the ISP 1340 to operate on a much larger data set.
It will be appreciated that various embodiments of the present disclosure according to the claims and description in the specification can be realized in the form of hardware, software or a combination of hardware and software.
Any such software may be stored in a non-transitory computer readable storage medium. The non-transitory computer readable storage medium stores one or more programs (software modules), the one or more programs comprising instructions, which when executed by one or more processors in an electronic device, cause the electronic device to perform a method of the present disclosure.
Any such software may be stored in the form of volatile or non-volatile storage such as, for example, a storage device like a Read Only Memory (ROM), whether erasable or rewritable or not, or in the form of memory such as, for example, Random Access Memory (RAM), memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a Compact Disk (CD), Digital Versatile Disc (DVD), magnetic disk or magnetic tape or the like. It will be appreciated that the storage devices and storage media are various embodiments of non-transitory machine-readable storage that are suitable for storing a program or programs comprising instructions that, when executed, implement various embodiments of the present disclosure. Accordingly, various embodiments provide a program comprising code for implementing apparatus or a method as claimed in any one of the claims of this specification and a non-transitory machine-readable storage storing such a program.
While the disclosure has been shown and described with reference to various embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the disclosure as defined by the appended claims and their equivalents. Various embodiments of the present disclosure are described as examples only and are noted intended to limit the scope of the present disclosure. Accordingly, the scope of the present disclosure should be understood as to include any and all modifications that may be made without departing from the technical spirit of the present disclosure.
This application claims the benefit under 35 USC §119(e) of U.S. Provisional Application No. 62/011,311, filed Jun. 12, 2014, the entire disclosure of which is hereby incorporated by reference.
Number | Date | Country | |
---|---|---|---|
62011311 | Jun 2014 | US |