The present disclosure relates to an image output device and a control method for the same, and more particularly, to an image output device that communicates with another image output device, and a control method for the same.
In recent years, various types of display devices have been developed and supplied.
In various image display environments, there are increasing cases of using a plurality of display devices, especially a plurality of projector devices, rather than just one display device.
Using the plurality of projector devices may have various effects and advantages, such as increasing the size or maximum output luminance of an image, and may also cause a problem such as excessive increase in the luminance of the image or decrease in a contrast ratio or dynamic range.
In case that the respective images output by the plurality of projector devices are provided in an overlay manner, the image may have advantages such as increased luminance to the maximum output luminance and enhanced three-dimensionality. Therefore, there has been a demand for a method to maintain or increase the contrast ratio and dynamic range of the image while maintaining these advantages.
The disclosure relates to an image output device that provides a user with one image by communicating with a plurality of image output devices, and a control method for the same.
According to an aspect of the disclosure, there is provided an image output device including: a projector; a communication interface including a circuit; memory storing instructions; and at least one processor, wherein the instructions, when executed by the at least one processor, cause the image output device to: acquire brightness information of an input image received via the communication interface, identify whether another image output device outputs the input image by adjusting its luminance, by communicating with the other image output device that receives a same image as the input image and outputs the same image onto a projection surface, adjust the luminance of at least one area included in the input image based on the brightness information according to a result of the identifying, and control the projector to output the adjusted input image onto the projection surface.
The instructions, when executed by the at least one processor, may cause the image output device to acquire the adjusted input image, where the luminance of a relatively bright area is increased and the luminance of a relatively dark area is decreased, among the at least one area included in the input image, based on the brightness information and a tone map curve for expanding a dynamic range.
The instructions, when executed by the at least one processor, may cause the image output device to: acquire one of a depth map, an object map, or a saliency map, corresponding to the input image, based on the brightness information of the input image, identify distance information of a plurality of objects included in the input image based on one of the depth map or the object map, increase the luminance of an area corresponding to a first object disposed at a relatively short distance among the plurality of objects, and decrease the luminance of an area corresponding to a second object disposed at a relatively long distance among the plurality of objects.
The instructions, when executed by the at least one processor, may cause the image output device to transmit, to the other image output device, at least one of a tone map curve used by the image output device or a depth map or an object map, corresponding to the input image, based on the other image output device being identified as outputting the input image by adjusting its luminance.
The instructions, when executed by the at least one processor, may cause the image output device to control the projector to output the adjusted input image to a corresponding position on the projection surface based on received position information if position information on the projection surface onto which the other image output device outputs the input image is received.
The image output device may include a sensor, wherein the instructions, when executed by the at least one processor, may cause the image output device to: acquire, through the sensor, position information on the projection surface onto which the other image output device outputs the input image, and control the projector to output the adjusted input image to a corresponding position on the projection surface based on the acquired position information.
The instructions, when executed by the at least one processor, may cause the image output device to: set one of the other image output device and the image output device as a master and set the other as a slave, transmit a first control signal to the other image output device for the other image output device to output the input image by adjusting its luminance, and control the projector to output the input image without adjusting the luminance of the input image based on the other image output device being set as the slave, and transmit a second control signal to the other image output device for the other image output device to output the input image without adjusting its luminance, and control the projector to output the adjusted input image after adjusting the luminance of the input image based on the other image output device being set as the master.
The instructions, when executed by the at least one processor, may cause the image output device to: identify at least one object corresponding to a user selection among a plurality of objects included in the input image, and acquire the adjusted input image by increasing the luminance of an area corresponding to the identified at least one object and decreasing the luminance of a remaining area.
A projection screen output by the image output device and a projection screen output by the other image output device may be overlaid at the same position on the projection surface.
According to an aspect of the disclosure, there is provided a control method for an image output device, the control method including: acquiring brightness information of an input image; identifying whether another image output device outputs the image by adjusting its luminance by communicating with the other image output device that receives the same image as the input image and outputs the same image onto a projection surface; adjusting the luminance of at least one area included in the input image based on the brightness information according to a result of the identifying; and outputting the adjusted input image onto the projection surface.
In the adjusting, the adjusted input image may be acquired where the luminance of a relatively bright area is increased and the luminance of a relatively dark area is decreased, among the at least one area included in the input image, based on the brightness information and a tone map curve for expanding a dynamic range.
The adjusting may include acquiring one of a depth map, an object map, or a saliency map, corresponding to the input image, based on the brightness information of the input image, identifying distance information of a plurality of objects included in the input image based on one of the depth map or the object map, increasing the luminance of an area corresponding to a first object disposed at a relatively short distance among the plurality of objects, and decreasing the luminance of an area corresponding to a second object disposed at a relatively long distance among the plurality of objects.
The control method may further include transmitting, to the other image output device, at least one of a tone map curve used by the image output device or a depth map or an object map, corresponding to the input image, based on the other image output device being identified as outputting the image by adjusting its luminance.
In the outputting, the adjusted input image may be output to a corresponding position on the projection surface based on received position information if position information on the projection surface onto which the other image output device outputs the image is received.
The control method may further include acquiring, through the sensor, position information on the projection surface onto which the other image output device outputs the image, wherein in the outputting, the adjusted input image is output to a corresponding position on the projection surface based on the acquired position information.
According to the various embodiments of the present disclosure, it is possible to provide the image by using the plurality of image output devices, and provide the image having the improved visibility by increasing the maximum output luminance of the image.
According to the various embodiments of the present disclosure, it is possible to increase the maximum output luminance of the image, thereby increasing the contrast ratio and dynamic range of the image simultaneously while increasing the visibility, immersion, and three-dimensionality of the image.
The above and other aspects and/or features of one or more embodiments of the disclosure will be more apparent from the following detailed description, taken in conjunction with the accompanying drawings, in which:
Terms used in the specification are briefly described, and the present disclosure is then described in detail.
General terms currently widely used are selected as terms used in embodiments of the present disclosure in consideration of their functions in the present disclosure, and may be changed based on the intentions of those skilled in the art or a judicial precedent, the emergence of a new technique, or the like. In addition, in a specific case, terms arbitrarily chosen by an applicant may exist. In this case, the meanings of such terms are mentioned in detail in corresponding description portions of the present disclosure. Therefore, the terms used in the present disclosure need to be defined on the basis of the meanings of the terms and the contents throughout the present disclosure rather than simple names of the terms.
The present disclosure may be variously modified and have diverse embodiments, and specific embodiments of the present disclosure are thus shown in the drawings and described in detail in the detailed description. However, it is to be understood that the present disclosure is not limited to the specific embodiments, and includes all modifications, equivalents, and substitutions without departing from the scope and spirit of the present disclosure. Omitted is a detailed description of a case where it is decided that the detailed description of the known art related to the present disclosure may obscure the gist of the present disclosure.
Terms “first”, “second”, and the like, may be used to describe various components. However, the components are not to be construed as being limited to these terms. The terms are used only to distinguish one component and another component from each other.
A term of a singular number may include its plural number unless explicitly indicated otherwise in the context. It is to be understood that a term “include,” “formed of,” or the like used in this application specifies the presence of features, numerals, steps, operations, components, parts, or combinations thereof, mentioned in the specification, and does not preclude the presence or addition of one or more other features, numerals, steps, operations, components, parts, or combinations thereof.
In the present disclosure, a “module” or a “˜er/˜or” may perform at least one function or operation, and be implemented by hardware, software, or a combination of hardware and software. In addition, a plurality of “modules” or a plurality of “˜ers/˜ors” may be integrated in at least one module and implemented by at least one processor (not shown) except for a “module” or a “˜er/or” that needs to be implemented by specific hardware.
Hereinafter, the embodiments of the present disclosure are described in detail with reference to the accompanying drawings so that those skilled in the art to which the present disclosure pertains may easily practice the present disclosure. However, the present disclosure may be modified in various different forms, and is not limited to the embodiments provided herein. However, the present disclosure may be modified in various different forms, and is not limited to the embodiments provided herein.
Referring to
In particular, each of the image output device 100 and another image output device 200 may be implemented as a projector device that enlarges and projects an image onto a wall or screen, and the projector device may be a liquid crystal display (LCD) projector or a digital light processing (DLP) projector using a digital micromirror device (DMD).
According to various embodiments of the present disclosure, the image output device 100 and another image output device 200 may be the same device or may be different devices.
For example, the details or specifications of image output device 100 and another image output device 200 may be the same as each other. For example, the image output device 100 and another image output device 200 may have the same brightness unit (e.g., American national standards institute (ANSI)-Lumens unit, light emitting diode (LED)-lumens unit), contrast ratio, resolution, or the like. However, this configuration is only an example, and the details or specifications of the image output device 100 and another image output device 200 may be different from each other.
Each of the image output device 100 and another image output device 200 according to an embodiment of the present disclosure may be implemented as the projector device (or a projector display) that projects the image, is not limited thereto, and may be implemented as any of various forms of display devices, such as a television (TV), a video wall, a large format display (LFD), a digital signage, or a digital information display (DID).
Here, the display device may have various types of display panels such as a liquid crystal display (LCD) panel, an organic light-emitting diode (OLED) panel, a liquid crystal on silicon (LCoS) panel, a digital light processing (DLP) panel, a quantum dot (QD) panel, a quantum dot light-emitting diode (QLED) panel, a micro light-emitting diode (μLED) panel, or a mini LED panel.
In addition, each of the image output device 100 and another image output device 200 may be a display device for home or industrial use, may be a lighting device used in daily life, may be an audio device including an audio module, and may be implemented as a portable communication device (e.g., smartphone), a computer device, a portable multimedia device, a wearable device, or a home appliance device.
Meanwhile, each of the image output device 100 and another image output device 200 according to an embodiment of the present disclosure is not limited to the above-described device, and the image output device 100 may be implemented as the image output device 100 having two or more functions of the above-described devices. For example, the image output device 100 may be used as the display device, the lighting device, or the audio device by turning off its projector function and turning on its lighting function or speaker function based on a manipulation of a processor, and may be used as an artificial intelligence (AI) speaker by including a microphone or a communication device.
According to an embodiment of the present disclosure, the image output device 100 and another image output device 200 may be combined with each other to provide one image. For example, an image (hereinafter, a first image) output by the image output device 100 and an image (hereinafter, a second image) output by another image output device 200 may be overlaid with each other (hereinafter, an overlay manner) to form one image (hereinafter, a third image).
As another example, the image (i.e., first image) output by the image output device 100 and the image (i.e., second image) output by another image output device 200 may be arrayed (hereinafter, an array manner) to form one image (for example, one of the images output by the image output device 100 and the images output by another image output device 200 is disposed on a left side and the other is disposed on a right side, or one of the images output by the image output device 100 and the images output by another image output device 200 is disposed on an upper side and the other is disposed on a lower side).
According to an embodiment, in the overlay manner, the images respectively output by the image output device 100 and another image output device 200 (i.e., first image and second image) may be overlaid with each other to form the third image, thus increasing the brightness (or, luminance) of the image.
According to another embodiment, in the array manner, the images respectively output by the image output device 100 and another image output device 200 may be arrayed to form one image, thus increasing the size or resolution of the image.
In case of providing the image in the overlay manner, the image output device 100 according to the various embodiments of the present disclosure may output an original image by adjusting its luminance in such a way that a dark portion (e.g., low grayscale or low luminance area) of the third image formed by overlaying the images (i.e., first image and second image) output by the image output device 100 and another image output device 200 with each other is prevented from being excessively brightened compared to a dark portion of the original image (or the image input into the image output device 100) (by limiting increase in the luminance of the dark portion), and a light portion (e.g., high grayscale or high luminance area) of the third image is more brightened compared to a light portion of the original image (by increasing the luminance of the light portion).
Referring to
The projection unit 110 according to an embodiment may output the image that the image output device 100 is to output onto the projection surface. The projection unit 110 according to an embodiment may include a projection lens (not shown).
Here, the projection lens may be disposed on one surface of a main body of the image output device 100, and project light passed through a lens array to the outside of the main body. The projection lens in the various embodiments may be an optical lens which is low-dispersion coated to reduce chromatic aberration. The projection lens may be a convex lens or a condensing lens, and the projection lens according to the various embodiments may adjust a focus by adjusting positions of a plurality of sub-lenses.
The projection unit 110 may perform a function of outputting the image onto the projection surface. Here, the projection surface may be a portion of a physical space onto which the image is output, or may be a separate screen. Meanwhile, a configuration of the projection unit 110 is not limited to the above-described example, and the image output device 100 may output the image onto the projection surface by using various methods.
The communication interface 120 according to an embodiment of the present disclosure may communicate with various types of external devices (e.g., image providing device or display device), external servers, or the like, and receive various types of data and information.
For example, the communication interface 120 may receive the various types of data and information from the image providing device (e.g., source device), an external storage medium (e.g., universal serial bus (USB) memory), the external server (e.g., cloud server or web hard), or the like by using a communication method/communication standard such as an access point (AP)-based wireless fidelity (Wi-Fi), i.e., wireless local area network (LAN)), a Bluetooth, a Zigbee, a wired/wireless LAN, a wide area network (WAN), Ethernet, an IEEE 1394, a high definition multimedia interface (HDMI), a USB, Thunderbolt™, a mobile high-definition link (MHL), an audio engineering society/European broadcasting union (AES/EBU) communication, an optical communication, or a coaxial communication.
In particular, the communication interface 120 according to an embodiment of the present disclosure may communicate with another image output device 200 (e.g., projector device) adjacent to the image output device 100. For example, the communication interface 120 may automatically detect another image output device 200 adjacent to the image output device 100 under control of the processor 130, and communicate with the detected another image output device 200.
As another example, the communication interface 120 may acquire a list including at least one another image output device 200 adjacent to the image output device 100 under the control of the processor 130, and communicate with a selected another image output device 200 in case that another image output device 200 is selected from the list.
At least one processor 130 may be electrically connected to a memory (not shown), and control overall operations of the image output device 100.
According to an embodiment of the present disclosure, the processor 130 may be implemented as a digital signal processor (DSP), a microprocessor, or a Time controller (TCON) that processes a digital signal. However, the processor 130 is not limited thereto, may include at least one of a central processing unit (CPU), a micro controller unit (MCU), a micro processing unit (MPU), a controller, an application processor (AP), a communication processor (CP), an advanced RISC machine (ARM) processor, or an artificial intelligence (AI) processor, or may be defined by these terms. In addition, the processor 130 may be implemented in a system-on-chip (SoC) or a large scale integration (LSI), in which a processing algorithm is embedded, or may be implemented in the form of a field programmable gate array (FPGA). The processor 130 may perform various functions by executing computer executable instructions stored in the memory.
First, the processor 130 according to an embodiment of the present disclosure may acquire the brightness information of the input image in case that the input image is received via the communication interface 120.
Here, the brightness information may include each grayscale value of a plurality of pixels included in the input image, and the output luminance corresponding to each grayscale value of the plurality of pixels.
At least one processor 130 may identify whether another image output device 200 outputs the image by adjusting its luminance by communicating with another image output device 200 that receives the same image as the input image and outputs the same image onto the projection surface.
Next, the processor 130 may adjust the luminance of at least one area included in the input image on the basis of the brightness information according to an identification result. Next, the processor 130 may control the projection unit 110 to output the luminance-adjusted image onto the projection surface.
A detailed description thereof is provided with reference to
According to the overlay manner, the image output device 100 may output a first screen 10 onto the projection surface, and another image output device 200 may also output a second screen 20 onto the projection surface. Here, a position where the image output device 100 outputs the screen (hereinafter, a first projection surface) and a position where another image output device 200 outputs the screen (hereinafter, a second projection surface) may be the same as each other.
The first projection surface and the second projection surface may be the same as each other, the first screen 10 output by the image output device 100 and the second screen 20 output by another image output device 200 may thus be overlaid with each other to form a third image 30.
Meanwhile, if the image output device 100 and another image output device 200 are each implemented as the projector device, the image output device 100 and another image output device 200 may each have their own ANSI-Lumens.
A recommended luminance range (or recommended brightness range) may exist based on a usage environment, as shown in Table 1 below.
According to an embodiment, even though the ANSI lumens of the image output device 100 do not satisfy the recommended luminance range, as shown in
Referring to
Here, the depth map 2 may indicate an image or an image channel that includes information related to a distance per area or pixel from an observation viewpoint in the input image 1. The object map may indicate an image or an image channel that shows the shape, appearance, or the like of a plurality of objects, a predefined object, a main object having a priority of a threshold value or more, or the like, in the input image 1.
According to an embodiment, at least one processor 130 may identify distance information of the plurality of objects included in the input image 1 on the basis of one of the depth map 2 or the object map.
Next, the processor 130 may increase the luminance of an area corresponding to a first object disposed at a relatively short distance among the plurality of objects, and decrease the luminance of an area corresponding to a second object disposed at a relatively long distance among the plurality of objects, thereby adjusting the luminance of at least one area included in the input image 1.
According to an embodiment, the processor 130 may increase the luminance of a near object (i.e., object disposed at a relatively short distance from the observation viewpoint based on the distance information), and decrease the luminance of a far object (i.e., object disposed at a relatively long distance from the observation viewpoint based on the distance information).
Next, the processor 130 may control the projection unit 110 to output the luminance-adjusted image onto the projection surface.
According to an embodiment, the first screen 10 output by the image output device 100 may indicate the luminance-adjusted image. According to an embodiment, the second screen 20 output by another image output device 200 may indicate the luminance-unadjusted image.
The third image 30, which is formed by overlaying the first screen 10 output by the image output device 100 and the second screen 20 output by another image output device 200 with each other, may be expressed in Equation 1 below.
Here, (x, y) indicates pixel coordinates, Ioverlay indicates the third image 30, and Ioriginal indicates the input image (or the original image). 2D Region Map may include the depth map, the object map, a saliency mask, or the like. In the description above, the image output device 100 may output the luminance-adjusted image, and another image output device 200 may output the luminance-unadjusted image. Therefore, Ioriginal (x,y)*2D Region Map (x,y) may indicate the first screen 10, and Ioriginal (x,y) may indicate the second screen 20.
Meanwhile, the opposite of the description above is also possible. It is apparent that the image output device 100 may output the luminance-unadjusted image, and another image output device 200 may output the luminance-adjusted image, as shown in
Referring to
Meanwhile, according to an embodiment of the present disclosure, the image output device 100 and another image output device 200 may communicate with each other. Accordingly, one of the image output device 100 and another image output device 200 may output the luminance-adjusted image and the other may output the luminance-unadjusted image.
For example, at least one processor 130 included in the image output device 100 may communicate with another image output device 200 via the communication interface 120 to thus set one of the image output device 100 and another image output device 200 as a master and the other as a slave.
For example, the image output device 100 may receive identification information from another image output device 200, and the processor 130 may compare the identification information of the image output device 100 with the identification information of another image output device 200 to thus set one of the devices as the master and the other as the slave on the basis of a comparison result. For example, the processor 130 may compare the identification information of the image output device 100 with the identification information of another image output device 200, and set the image output device 100 as the master and another image output device 200 as the slave if the image output device 100 is capable of outputting an image having relatively higher brightness than another image output device 200. Here, the identification information may indicate the details or specifications (e.g., brightness (or ANSI lumens), contrast ratio, or resolution) of a component included in the device. However, this configuration is only an example, and may not be limited thereto.
For example, it is apparent that at least one processor 130 may communicate with another image output device 200 via the communication interface 120 to thus randomly set one of the image output device 100 and another image output device 200 as the master and the other as the slave.
Next, at least one processor 130 may transmit a control signal to another image output device 200 for another image output device 200 to output the image by adjusting its luminance, and control the projection unit 110 to output the input image 1 without adjusting the luminance of the input image 1 if another image output device 200 is set as the slave.
As another example, at least one processor 130 may transmit a control signal to another image output device 200 for another image output device 200 to output the image without adjusting its luminance, and control the projection unit 110 to output the adjusted input image after adjusting the luminance of the input image 1 if another image output device 200 is set as the master.
Referring to
For example, at least one processor 130 included in the image output device 100 may increase the luminance of an area corresponding to the first object disposed at the relatively short distance among the plurality of objects included in the input image 1 on the basis of one of the depth map 2 or the object map, and decrease the luminance of an area corresponding to the second object disposed at the relatively long distance, thereby acquiring the luminance-adjusted image, that is, the first image 10. Next, at least one processor 130 may control the projection unit 110 to output the luminance-adjusted image onto the projection surface.
In addition, another image output device 200 may increase the luminance of the area corresponding to the first object disposed at the relatively short distance among the plurality of objects included in the image 1 on the basis of one of the depth map 2 or the object map, and decrease the luminance of the area corresponding to the second object disposed at the relatively long distance, thereby acquiring the luminance-adjusted image, that is, the second image 20. Next, another image output device 200 may output the luminance-adjusted image onto the projection surface.
Here, the depth map 2 (or the object map) used by the image output device 100 may be the same as or different from the depth map 2 (or the object map) used by another image output device 200.
For example, if another image output device 200 is identified as outputting the image by adjusting the luminance, the image output device 100 may transmit the depth map 2 or the object map, corresponding to the input image 1, to another image output device 200, or receive the depth map 2 or the object map from another image output device 200.
Next, in the third image 30 formed by overlaying the first screen 10 and the second screen 20 with each other, the luminance of the light portion may be increased compared to the luminance of the input image 1 (or the original image), and the luminance of the dark portion may not be increased excessively (or the increase in the luminance may be limited) as shown in
The third image 30 formed by overlaying the first screen 10 output by the image output device 100 and the second screen 20 output by another image output device 200 with each other, as shown in
Here, (x, y) indicates the pixel coordinates, Ioverlay indicates the third image 30, and Ioriginal indicates the input image (or the original image). 2D Region Map may include the depth map, the object map, the saliency mask, or the like. In the description above, the image output device 100 may apply a weight of Alpha to the luminance-adjusted image, and another image output device 200 may apply a weight of Beta to the luminance-adjusted image. Therefore, Alpha*Ioriginal (x,y)*2D Region Map (x,y) may indicate the first screen 10, and Beta*Ioriginal (x,y)*2D Region Map (x,y) may indicate the second screen 20.
Here, each weight of Alpha and Beta may have a value between 0 and 1, and the weights may be the same or different values.
According to an embodiment, each of the image output device 100 and another image output device 200 may acquire the weights of Alpha and Beta on the basis of a difference in light sources (or brightness (i.e., ANSI lumens)) and a difference in 2D Region Maps.
At least one processor 130 according to an embodiment of the present disclosure may identify at least one object corresponding to a user selection among the plurality of objects included in the input image 1.
Next, the processor 130 may increase the luminance of an area corresponding to at least one identified object and decrease the luminance of the remaining area, thereby acquiring the luminance-adjusted image.
In detail, at least one processor 130 may acquire the saliency map (or the saliency mask) from the input image 1.
Next, at least one processor 130 may identify at least one object corresponding to the user selection among the plurality of objects included in the input image 1. Next, at least one processor 130 may increase the luminance of an area (hereinafter, an area of interest) corresponding to at least one object identified and decrease the luminance of the remaining area, on the basis of the saliency map.
As another example, at least one processor 130 may output only the area of interest in the input image without outputting the remaining area.
Referring to
The third image 30, which is formed by overlaying the first screen 10 output by the image output device 100 and the second screen 20 output by another image output device 200 with each other, may be expressed in Equation 3 below.
Here, (x, y) indicates the pixel coordinates, Ioverlay indicates the third image 30, and Ioriginal indicates the input image (or the original image). 2D Region Map may include the depth map, the object map, the saliency mask, or the like. In the description above, the image output device 100 may output an image where the area of interest has the increased luminance, and another image output device 200 may output the image as it is. Therefore, Ioriginal (x,y)*2D Region Map (x,y) may indicate the first screen 10, and Ioriginal (x,y) may indicate the second screen 20.
Meanwhile, the opposite of the description above is also possible. It is apparent that the image output device 100 may output the input image as it is, and another image output device 200 may output the image where the area of interest has the increased luminance, as shown in
At least one processor 130 according to an embodiment of the present disclosure may acquire the adjusted image, where the luminance of a relatively bright area (e.g., high grayscale or high luminance) is increased and the luminance of a relatively dark area (e.g., low grayscale or low luminance) is decreased, among the plurality of areas included in the input image 1, on the basis of the brightness information of the input image 1 and the tone map curve for expanding its dynamic range DR.
At least one processor 130 according to an embodiment of the present disclosure may perform tone mapping on each pixel value of the plurality of pixels included in the input image 1 on the basis of the tone map curve.
Next, at least one processor 130 may acquire the luminance-adjusted image based on the tone mapping. Here, the luminance of the dark portion in the luminance-adjusted image may be maintained as the luminance of the dark portion (e.g., low grayscale or black area) in the input image 1, and the luminance of the light portion in the luminance-adjusted image may be increased compared to the luminance of the light portion (e.g., medium grayscale or high grayscale area) in the input image 1. Therefore, the dynamic range DR of the luminance-adjusted image may be increased compared to the dynamic range of the input image 1, and its contrast ratio and depth may also be increased.
As shown in
Here, the third image may be expressed in Equation 4 below.
Here, (x, y) indicates the pixel coordinates, Ioverlay indicates the third image 30, and Ioriginal indicates the input image (or the original image). Tone Map Curve 1 may indicate the tone map curve used by the image output device 100 for the tone mapping, and Tone Map Curve 2 may indicate the tone map curve used by another image output device 200 for the tone mapping. In the description above, the image output device 100 may apply the weight of Alpha to the luminance-adjusted image, and another image output device 200 may apply the weight of Beta to the luminance-adjusted image. Therefore, Alpha*Tone Map Curve 1 (Ioriginal (x,y)) may indicate the first screen 10, and Beta*Tone Map Curve 2 (Ioriginal (x,y)) may indicate the second screen 20.
Here, each weight of Alpha and Beta may have a value between 0 and 1, and the weights may be the same or different values.
According to an embodiment, each of the image output device 100 and another image output device 200 may acquire the weights of Alpha and Beta on the basis of the difference in the light sources (or the brightness (i.e., ANSI lumens)) and the difference in the tone map curves.
Meanwhile, Tone Map Curve 1 and Tone Map Curve 2 may be the same or different. For example, the image output device 100 and another image output device 200 may share the tone map curve and may use the same tone map curve for the tone mapping. As another example, it is apparent that the image output device 100 and another image output device 200 may each use the different tone map curves pre-stored in the tone mapping.
Referring to an upper side of
For example, if the maximum brightness (e.g., ANSI lumens) of the projector device is 400 lumens, and a surrounding environment of the projector device is bright daytime and outdoors, the luminance of the image provided by the projector device may be below the recommended luminance range, and the image may appear slightly dark or have poor visibility from the user's viewpoint.
The third image 30, which is formed by overlaying the first image 10 output by the image output device 100 and the second image 20 output by another image output device 200 with each other in the overlay manner, may be provided while having the same light intensity as the sum of the maximum light intensity that the image output device 100 may output and the maximum light intensity that another image output device 200 may output, thus maintaining its visibility even in case that the surrounding environment is bright daytime and outdoors.
However, the luminance of an area in the third screen 30 that corresponds to a relatively dark area in the input image 1 may be increased, thus causing the problem of decreasing the dynamic range, contrast ratio, or the like of the third screen 30.
According to the various embodiments of the present disclosure, at least one of the image output device 100 or another image output device 200 may adjust the luminance of at least one area included in the input image 1, and output the luminance-adjusted image onto the projection surface, thereby increasing the dynamic range, contrast ratio, depth, three-dimensionality, or the like of the third image 30.
As another example, referring to a lower side of
Referring to an upper side of
In addition, the image output device 100 may communicate with another image output device 200 via a second communication interface.
For example, at least one processor 130 included in the image output device 100 may control the projection unit 110 to output the luminance-adjusted image to a corresponding position on the projection surface on the basis of received position information if position information on the projection surface onto which another image output device 200 outputs the image is received via the second communication interface.
The image output device 100 according to an embodiment of the present disclosure may communicate with another image output device 200 via the second communication interface, thereby synchronizing the output of the first image 10 with the output of the second image 20.
The image output device 100 according to an embodiment of the present disclosure may include a sensor. Here, the sensor may include a camera.
At least one processor 130 may acquire, through the sensor, the position information on the projection surface onto which another image output device 20 outputs the second image 20, and may control the projection unit 110 to output the luminance-adjusted image to the corresponding position on the projection surface on the basis of the acquired position information.
At least one processor 130 according to an embodiment may perform a keystone correction function, a leveling correction function, or the like for controlling the projection unit 110 to output the image to the corresponding position on the projection surface.
Here, the keystone correction function may be a function for solving a problem that a trapezoidal image is output onto the projection surface due to the tilt of the electronic device 100.
Here, the keystone correction function may be a function for correcting the image for the trapezoidal image output onto the projection surface to be output as a rectangular or square image. The keystone correction function may be classified into horizontal keystone correction or vertical keystone correction based on a direction of keystone correction.
Here, the leveling correction function may indicate a function of rotating the image. In detail, the processor 130 may control the projection unit 110 to output the image by rotating the image by a certain angle using the leveling correction function.
Referring to the upper side of
The image output device 100 and the additional device may be connected to each other through an external signal processing device. Here, the external signal processing device may also be referred to as a connector.
The external signal processing device may connect the image processing device 100 and the additional device to each other in a wired manner or a wireless manner to thus transmit and receive an electrical signal from the additional device. The connector may be physically connected to the additional device. Here, it is apparent that the connector may include various types of wired communication modules such as an HDMI connection terminal and a USB connection terminal, and may also include various types of wireless communication modules such as Bluetooth and Wi-Fi.
The control method of an image output device according to an embodiment for achieving the above-described purpose of the present disclosure may first include acquiring brightness information of an input image (S1010).
Next, the method may include identifying whether another image output device outputs the image by adjusting its luminance by communicating with another image output device that receives the same image as the input image and outputs the same image onto a projection surface (S1020).
Next, the method may include adjusting the luminance of at least one area included in the input image on the basis of the brightness information according to an identification result (S1030).
Next, the method may include outputting the luminance-adjusted image onto the projection surface (S1040).
Here, in the adjusting (S1020), the adjusted image, where the luminance of a relatively bright area is increased and the luminance of a relatively dark area is decreased, among the plurality of areas included in the input image, may be acquired on the basis of the brightness information and a tone map curve for expanding a dynamic range DR.
The adjusting (S1020) according to an embodiment of the present disclosure may include acquiring one of a depth map, an object map, or a saliency map, corresponding to the input image, on the basis of the brightness information of the input image, identifying distance information of a plurality of objects included in the input image on the basis of one of the depth map, the object map, or the saliency map, increasing the luminance of an area corresponding to a first object disposed at a relatively short distance among the plurality of objects, and decreasing the luminance of an area corresponding to a second object disposed at a relatively long distance among the plurality of objects.
The control method according to an embodiment of the present disclosure may further include transmitting, to another image output device, at least one of a tone map curve used by the image output device or a depth map or an object map, corresponding to the input image, if another image output device is identified as outputting the image by adjusting its luminance.
In the outputting (S1040) according to an embodiment of the present disclosure, the adjusted image may be output to a corresponding position on the projection surface on the basis of received position information if position information on the projection surface onto which another image output device outputs the image is received.
The image output device according to an embodiment of the present disclosure may include a sensor, and the control method may further include acquiring, through the sensor, position information on the projection surface onto which another image output device outputs the image, wherein in the outputting (S1040), the adjusted image may be output to the corresponding position on the projection surface on the basis of the acquired position information.
The control method according to an embodiment of the present disclosure may further include: setting one of another image output device and the image output device as a master and the other as a slave, transmitting a control signal to another image output device for another image output device to output the image by adjusting its luminance, and outputting the input image without adjusting the luminance of the input image if another image output device is set as the slave, and transmitting a control signal to another image output device for another image output device to output the image without adjusting its luminance, and outputting the adjusted input image after adjusting the luminance of the input image if another image output device is set as the master.
The control method according to an embodiment of the present disclosure may further include identifying at least one object corresponding to a user selection among the plurality of objects included in the input image, wherein in the adjusting (S1020), the adjusted image may be acquired by increasing the luminance of an area corresponding to the identified at least one object and decreasing the luminance of the remaining area.
According to an embodiment of the present disclosure, a projection screen output by the image output device and a projection screen output by another image output device may be overlaid with each other at the same position on the projection surface.
However, it is apparent that the various embodiments of the present disclosure may be applied not only to the electronic device but also to every type of electronic device including the display.
Meanwhile, the various embodiments of the present disclosure described above may be implemented in a computer or a computer-readable recording medium using software, hardware, or a combination of software and hardware. In some cases, the embodiments described in the present disclosure may be implemented by the processor itself. According to software implementation, the embodiments such as the procedures and functions described in the specification may be implemented by separate software modules. Each of the software modules may perform one or more functions and operations described in the specification.
Meanwhile, a non-transitory computer-readable medium may store computer instructions for performing processing operations of the electronic device according to the various embodiments of the present disclosure described above. The computer instructions stored in the non-transitory computer-readable medium may allow a specific device to perform the processing operations of the image output device 100 according to the various embodiments described above in case of being executed by a processor of the specific device.
The non-transitory computer-readable medium is not a medium that stores data therein for a while, such as a register, a cache, or a memory, and indicates a medium that semi-permanently stores data therein and is readable by a machine. A specific example of the non-transitory computer-readable medium may include a compact disk (CD), a digital versatile disk (DVD), a hard disk, a Blu-ray disk, a universal serial bus (USB), a memory card, a read-only memory (ROM), or the like.
Although the embodiments are shown and described in the present disclosure hereinabove, the present disclosure is not limited to the above-mentioned specific embodiments, and may be variously modified by those skilled in the art to which the present disclosure pertains without departing from the gist of the present disclosure as claimed in the accompanying claims. These modifications should also be understood to fall within the scope and spirit of the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
10-2022-0087184 | Jul 2022 | KR | national |
This application is a bypass continuation application of International Application No. PCT/KR2023/007263 designating the United States, filed on May 26, 2023, in the Korean Intellectual Property Receiving Office and claiming priority to Korean Patent Application No. 10-2022-0087184, filed on Jul. 14, 2022, in the Korean Intellectual Property Office, the disclosures of which are incorporated by reference herein in their entireties.