One or more embodiments described herein relate to a display device and a method of operating the display device.
As display devices operate over time, the pixels of the display device may become degraded. The degradation experienced by the pixels may severely degrade display quality, especially for pixels used to display a logo for an extended period of time. These effects are exacerbated when the logo includes a high gray level image. If prolonged, an afterimage may be displayed in the pixel region where the logo is displayed.
One or more embodiments described herein provide a display device which may reduce degradation and afterimage effects, including but not exclusively in a logo region.
One or more embodiments described herein may reduce or prevent grayscale banding in a logo region and a peripheral region.
One or more embodiments described herein provide a method of operating a display device which may achieve the aforementioned effects.
In accordance with some embodiments, a display device includes a display panel including a plurality of pixels; a controller configured to detect a logo region including a logo in input image data, determine a correction gain based on a first average gray level of the logo region and a second average gray level of a peripheral region adjacent to the logo region, and generate corrected image data by correcting the input image data based on the correction gain; and a data driver configured to provide data signals to the plurality of pixels based on the corrected image data.
In accordance with some embodiments, a method of operating a display device includes detecting a logo region including a logo in input image data; determining a correction gain based on a first average gray level of the logo region and a second average gray level of a peripheral region adjacent to the logo region; generating corrected image data by correcting the input image data based on the correction gain; and driving a display panel based on the corrected image data.
Illustrative, non-limiting embodiments will be more clearly understood from the following detailed description in conjunction with the accompanying drawings.
Hereinafter, embodiments of the present inventive concept will be explained in detail with reference to the accompanying drawings.
Referring to
The display panel 110 may include a plurality of data lines, a plurality of scan lines, and the plurality of pixels PX coupled to the plurality of data lines and the plurality of scan lines. Each pixel PX may include, for example, a self-luminous light emitter, e.g., an organic light emitting diode (OLED). In this case, the display panel 110 may be an OLED panel. In other embodiments, the display panel 110 may be an inorganic light emitting diode display panel, a quantum dot light emitting diode display panel, a liquid crystal display (LCD) panel, or any other suitable display panel.
The scan driver 120 may provide the scan signals SS to the plurality of pixels PX based on a scan control signal SCTRL from the controller 140. In some embodiments, the scan control signal SCTRL may include, but is not limited to, a scan start signal and a scan clock signal. In some embodiments, the scan driver 120 may be integrated or formed in a peripheral portion of the display panel 110. In some embodiments, the scan driver 120 may be implemented with one or more integrated circuits.
The data driver 130 may provide data signals DS to the plurality of pixels PX, through the plurality of data lines, based on a data control signal DCTRL and corrected image data CDAT from the controller 140. In some embodiments, the data control signal DCTRL may include, but is not limited to, an output data enable signal, a horizontal start signal and a load signal. In some embodiments, the data driver 130 and the controller 140 may be implemented with a single integrated circuit, which, for example, may be referred to as a timing controller embedded data driver (TED). In other embodiments, the data driver 130 and the controller 140 may be implemented with separate integrated circuits.
The controller 140 (e.g., a timing controller (TCON)) may receive input image data IDAT and a control signal CTRL from an external host processor (e.g., a graphic processing unit (GPU), an application processor (AP) or a graphic card). In some embodiments, the input image data IDAT may be RGB image data including red image data, green image data and blue image data. In other embodiments, the input image data IDAT may be image data of a different combination of colors.
In some embodiments, the control signal CTRL may include, but is not limited to, a vertical synchronization signal, a horizontal synchronization signal, an input data enable signal, a master clock signal, and/or one or more other types of signals. The controller 140 may generate the scan control signal SCTRL, data control signal DCTRL and corrected image data CDAT based on the control signal CTRL and the input image data IDAT. The controller 140 may control an operation of the scan driver 120 by providing the scan control signal SCTRL to the scan driver 120, and may control operation of the data driver 130 by providing the data control signal DCTRL and corrected image data CDAT to data driver 130.
In the display device 100 according to embodiments, the controller 140 may receive the input image data IDAT, detect a logo region LR including a logo by analyzing the input image data IDAT, and determine a correction gain based on a first average gray level of the logo region LR and a second average gray level of a peripheral region PR adjacent to the logo region LR. In addition the controller 140 may generate the corrected image data CDAT by correcting the input image data IDAT based on the correction gain.
In some embodiments, the controller 140 may calculate the correction gain by dividing the second average gray level of the peripheral region PR by the first average gray level of the logo region LR, and may generate the corrected image data CDAT by multiplying (e.g., calculating the product of) the input image data IDAT for the logo region LR and/or the peripheral region PR by (and) the correction gain, In some embodiments, a gray level of the corrected image data CDAT for the logo region LR and/or the peripheral region PR may be linearly proportional to a gray level of the input image data IDAT for the logo region LR and/or the peripheral region PR. In some embodiments, the gray level of the corrected image data CDAT for the logo region LR and/or the peripheral region PR may be adjusted (e.g., decreased otherwise adjusted), for example, relative to the gray level of the input image data IDAT for the logo region LR and/or the peripheral region PR.
The data driver 130 may provide the plurality of pixels PX in the logo region LR and/or the peripheral region PR with the data signals DS corresponding to the corrected image data CDAT, which corrected image data CDAT corresponds to gray levels that have been adjusted (e.g., decreased) relative to the input image data IDAT. If not adjusted (e.g., if the data signals DS corresponding to the input image data IDAT are provided to the plurality of pixels PX in the logo region LR), degradation of the pixels PX in the logo region LR and/or the peripheral region PR may occur, along with an afterimage effect in that/those regions(s). However, according to one or more embodiments, the adjusted data signals DS corresponding to the corrected image data CDAT may reduce degradation of the pixels in the logo region LR and/or peripheral region PR and may prevent an afterimage effect from occurring in the logo region LR and/or the peripheral region PR.
To perform these operations, as illustrated, for example, in
The logo region detecting block (e.g., detector) 150 may detect the logo region LR including the logo (e.g., “LOGO” in
The peripheral region setting block (e.g., setting logic) 160 may set the peripheral region PR adjacent to the logo region LR. For example, the peripheral region setting block 160 may set a region surrounding the logo region LR. The surrounding peripheral region PR may have a predetermined shape, e.g., a substantially rectangular shape, elliptical shape or another shape. In some embodiments, the peripheral region setting block 160 may store one or more parameters corresponding to a size and/or a shape of the peripheral region PR, and may set the peripheral region PR based on the one or more parameters. In some embodiments, the one or more parameters corresponding to the size and/or shape of the peripheral region PR may be selected, set or changed by the host processor or by a user.
The correction gain determining block (e.g., gain logic) 170 may determine the correction gain CGAIN based on the first average gray level of the logo region LR and the second average gray level of the peripheral region PR. In some embodiments, the correction gain determining block 170 may calculate the first average gray level of the logo region LR by calculating an average of gray levels of the input image data IDAT for the logo region LR, may calculate the second average gray level of the peripheral region PR by calculating an average of gray levels of the input image data IDAT for the peripheral region PR, and may calculate a luminance ratio (of a luminance of the peripheral region PR to a luminance of the logo region LR) by dividing the second average gray level by the first average gray level.
In addition, the correction gain determining block 170 may determine the correction gain CGAIN based on the luminance ratio and a predetermined or preset (e.g., minimum) correction gain. In this case, the correction gain CGAIN may be determined to be greater than or equal to the predetermined or present (e.g., minimum) correction gain and less than or equal to 1. In one embodiment, the predetermined or preset correction gain may be different from a minimum gain.
In some embodiments, because a region having a luminance higher than that of the peripheral region PR may be detected as the logo region LR, the luminance ratio of the luminance of the peripheral region PR to the luminance of the logo region LR may be less than or equal to 1. In some cases, even if the luminance ratio is greater than 1, the correction gain determining block 170 may determine the correction gain CGAIN as 1.
For example, the correction gain determining block 170 may calculate the luminance ratio of the luminance of the peripheral region PR to the luminance of the logo region LR based on equation (1):
LUM_RATIO=AVG_PERI/AVG_LOGO (1)
where LUM_RATIO may represent the luminance ratio, AVG_PERI may represent the second average gray level of the peripheral region PR, and AVG_LOGO may represent the first average gray level of the logo region LR.
In addition, the correction gain determining block 170 may calculate correction gain CGAIN based on equation (2):
CGAIN=LUM_RATIO*(1−GAIN_LIMIT)+GAIN_LIMIT″ (2)
where CGAIN represents the correction gain CGAIN and GAIN_LIMIT may represent the predetermined or preset (e.g., minimum) correction gain. Hereinafter, the predetermined or preset correction gain will be assumed to be a minimum correction gain for the sake of discussion. The predetermined correction gain may be a value different from a minimum correction gain in another embodiment.
In view of equations (1) and (2), it is seen that as the second average gray level of the peripheral region PR decreases or as the first average gray level of the logo region LR increases, the correction gain CGAIN may be reduced from 1. The corrected image data CDAT may be generated based on the correction gain CGAIN and thus may be different from (e.g., reduced in gray level compared with) the input image data IDAT. Accordingly, degradation of the pixels PX in the logo region LR may be reduced, which, in turn, may reduce an afterimage effect that may be prone to develop in the logo region LR.
In some embodiments, the correction gain determining block 170 may calculate the first average gray level of the logo region LR and divide the peripheral region PR into a plurality of peripheral sub-regions (e.g., having a predetermined (e.g., ring) shape surrounding the logo region LR). In addition, the correction gain determining block 170 calculate the second average gray level of the peripheral region PR to correspond to a weighted-average gray level of one or more peripheral sub-regions. One or more weights used to calculate weighted-average gray level may change (e.g., decrease) as distances of the one or more peripheral sub-regions to the logo region LR increase.
In addition, the correction gain determining block 170 may calculate a luminance ratio, which may correspond to a weighted-luminance of the peripheral region PR to a luminance of the logo region LR. This calculation may include, for example, dividing (calculating the quotient of) the weighted-average gray level by (and) the first average gray level. The correction gain determining block 170 may determine the correction gain CGAIN, based on the luminance ratio and the minimum correction gain, to be a value greater than or equal to the minimum correction gain and less than or equal to 1. In some embodiments, a relatively high weight may be applied to one or more of the peripheral sub-regions closer to the logo region LR, a relatively low weight may be applied to one or more of the peripheral sub-regions farther away from the logo region LR. Accordingly, in one or more embodiments, the correction gain CGAIN may have a more pronounced effect on a peripheral image closer to the logo.
The data correcting block (e.g., data corrector) 180 may generate the corrected image data CDAT, for example, by correcting the input image data IDAT for the logo region LR and the peripheral region PR based on the correction gain CGAIN. In some embodiments, the data correcting block 180 may generate the corrected image data CDAT by multiplying the input image data IDAT for the logo region LR and the peripheral region PR by the correction gain CGAIN. Since the correction gain CGAIN is less than or equal to 1, the corrected image data CDAT for the logo region LR and the peripheral region PR may be decreased compared with the input image data IDAT for the logo region LR and the peripheral region PR. Accordingly, degradation of pixels PX in the logo region LR may be reduced, which, in turn, may reduce the likelihood of an afterimage effect occurring in the logo region LR.
In other embodiments, the data correcting block 180 may generate the corrected image data CDAT for the logo region LR by multiplying the input image data IDAT for the logo region LR by the correction gain. Further, to generate the corrected image data CDAT for the peripheral region PR, the data correcting block 180 may divide the peripheral region PR into a plurality of peripheral sub-regions (e.g., having a predetermined (e.g., ring) shape surrounding the logo region LR) and may determine a plurality of sub-region correction gains for the plurality of peripheral sub-regions. The plurality of sub-region correction gains may be, for example, greater than the correction gain CGAIN and less than 1.
The data correcting block 180 may multiply the input image data IDAT for the plurality of peripheral sub-regions by the plurality of sub-region correction gains, respectively. For example, the data correcting block 180 may determine the plurality of sub-region correction gains for peripheral sub-regions, so that the sub-region correction gains for the peripheral sub-regions are linearly proportional to distances of the peripheral sub-regions relative to the logo region LR. Thus, for example, the sub-region correction gain for one or more peripheral sub-regions closer to the logo region LR may be closer to the correction gain CGAIN, and the sub-region correction gain(s) for one or more peripheral sub-region(s) farther away from the logo region LR may be close to 1.
In this case, the amount of decrease of the corrected image data CDAT relative to the input image data IDAT corresponding to the peripheral sub-region(s) farther away from the logo region LR may be less than the amount of decrease of the corrected image data CDAT relative to the input image data IDAT corresponding to peripheral sub-region(s) closer to the logo region LR. Thus, the luminance difference between the peripheral region PR and a region outside (or surrounding) the peripheral region PR may be reduced.
In some embodiments, the correction gain CGAIN may be determined based on the input image data IDAT in a previous frame period. The corrected image data CDAT in a current frame period may then be generated by correcting the input image data IDAT in the current frame period based on the correction gain CGAIN in the previous frame period. For example, the input image data IDAT in the previous frame period may be stored in the frame memory 190, and the correction gain CGAIN may be determined based on the input image data IDAT stored in the frame memory 190. In some embodiments, the correction gain CGAIN may be determined based on the input image data IDAT in the current frame period, and the corrected image data CDAT in the current frame period may be generated by correcting the input image data IDAT in the current frame period based on the correction gain CGAIN in the current frame period.
In some display devices which have been proposed (e.g., organic light emitting diode (OLED) display devices), performance of pixels PX may degrade over time. For example, pixels PX displaying a logo that includes a high gray level image may be degraded more severely than pixels PX in other areas of the display device. Accordingly, an afterimage effect may be displayed in the logo region LR where the logo is displayed.
In an attempt to reduce degradation and the occurrence of an afterimage in the logo region, the proposed display devices perform a clamping operation that limits gray levels for the logo and peripheral regions to a predetermined reference gray level. For example, as illustrated by the curve 210 in
However, in the display device 100 according to embodiments, no such conversion is performed. As illustrated as a graph 230 in
Referring to
At S320, peripheral region setting block 160 may set a peripheral region PR adjacent to the logo region LR. In an example, as illustrated in
At S330, correction gain determining block 170 may calculate a first average gray level of the logo region LR. The correction gain determining block 170 may calculate the first average gray level of the logo region LR, for example, by calculating an average of gray levels of the input image data IDAT for the logo region LR.
At 340, correction gain determining block 170 may calculate a second average gray level of the peripheral region PR. For example, the correction gain determining block 170 may calculate the second average gray level of the peripheral region PR by calculating an average of gray levels of the input image data IDAT for the peripheral region PR.
At S350, correction gain determining block 170 may calculate a luminance ratio (of a luminance of the peripheral region PR to a luminance of the logo region LR) by dividing the second average gray level of the peripheral region PR by the first average gray level of the logo region LR. For example, the correction gain determining block 170 may calculate the luminance ratio of the luminance of the peripheral region PR to the luminance of the logo region LR using equation: LUM_RATIO=AVG_PERI/AVG_LOGO, where LUM_RATIO may represent the luminance ratio, AVG_PERI may represent the second average gray level of the peripheral region PR, and AVG_LOGO may represent the first average gray level of the logo region LR.
Because a region having a luminance higher than that of the peripheral region PR may be detected as the logo region LR, the luminance ratio (of the luminance of the peripheral region PR to the luminance of the logo region LR) may be less than or equal to 1. Further, even if the second average gray level of the peripheral region PR is higher than the first average gray level of the logo region LR, the correction gain determining block 170 may determine a correction gain CGAIN as 1.
Accordingly, as illustrated in
As S360, the correction gain determining block 170 may determine the correction gain CGAIN based on the luminance ratio LUM_RATIO and the minimum correction gain. In this case, the correction gain CGAIN may be determined to be greater than or equal to the minimum correction gain and less than or equal to 1. In some embodiments, the correction gain determining block 170 may calculate the correction gain CGAIN using the following equation: CGAIN=LUM_RATIO*(1−GAIN_LIMIT)+GAIN_LIMIT, where CGAIN may represent the correction gain CGAIN, LUM_RATIO may represent the luminance ratio, and GAIN_LIMIT may represent the minimum correction gain. Accordingly, as illustrated in
At S370, data correcting block 180 may generate the corrected image data CDAT by multiplying the input image data IDAT for the logo region LR and the peripheral region PR by the correction gain CGAIN. In the example of
At S380, data driver 130 may drive a display panel 110 based on the corrected image data CDAT. With respect to the logo region LR and the peripheral region PR, since the gray level of the corrected image data CDAT is reduced compared with the gray level of the input image data IDAT, degradation of pixels PX in the logo region LR may be reduced, which, in turn, may reduce an afterimage effect in the logo region LR. Further, since the gray level of the corrected image data CDAT is proportional (e.g., linearly proportional) to the gray level of the input image data IDAT, a grayscale banding phenomenon may be prevented.
Referring to
At S420, peripheral region setting block 160 may set a peripheral region PR adjacent to the logo region LR.
At S430, correction gain determining block 170 may calculate a first average gray level of the logo region LR.
At S442, correction gain determining block 170 may divide the peripheral region PR into a plurality of peripheral sub-regions. For example, as illustrated in
At S444, correction gain determining block 170 may calculate a weighted-average gray level of the plurality of peripheral sub-regions PSR1, PSR2, PSR3 and PSR4, based on one or more weights that decrease with increasing distance away from the logo region LR. For example, as illustrated in
At S450, correction gain determining block 170 may calculate a luminance ratio (of a weighted-luminance of the peripheral region PR to a luminance of the logo region LR) by dividing the weighted-average gray level of the plurality of peripheral sub-regions PSR1, PSR2, PSR3 and PSR4, or the weighted-average gray level of the peripheral region PR by the first average gray level of the logo region LR.
At S460, correction gain determining block 170 may determine a correction gain CGAIN to be greater than or equal to the minimum correction gain and less than or equal to 1. The correction gain determining block 170 may determine the correction gain CGAIN in this range based on the luminance ratio and a minimum correction gain. Since a relatively high weight SR1_W may be applied to the first peripheral sub-region PSR1 (which is close to the logo region LR) and a relatively low weight SR4_W may be applied to the fourth peripheral sub-region PSR4 (which is farther away from the logo region LR), the correction gain CGAIN may have a more profound effect on a peripheral image at areas closer to the logo.
At S470, data correcting block 180 may generate corrected image data CDAT by multiplying the input image data IDAT for the logo region LR and the peripheral region PR by the correction gain CGAIN.
At 480, data driver 130 may drive a display panel 110 based on the corrected image data CDAT.
Accordingly, in a method of operating a display device 100 according to embodiments, degradation and afterimage effects in the logo region LR may be reduced. Also, a grayscale banding phenomenon may be prevented from occurring in the logo region LR and peripheral region PR.
The method of
Referring to
At S520, peripheral region setting block 160 may set a peripheral region PR adjacent to the logo region LR.
At S530, correction gain determining block 170 may calculate a first average gray level of the logo region LR.
At S540, correction gain determining block 170 may calculate a second average gray level of the peripheral region PR.
At S550, correction gain determining block 170 may calculate a luminance ratio (of a luminance of the peripheral region PR to a luminance of the logo region LR) by dividing the second average gray level of the peripheral region PR by the first average gray level of the logo region LR.
At S560, correction gain determining block 170 and may determine a correction gain CGAIN to be greater than or equal to the minimum correction gain and is less than or equal to 1. The correction gain determining block 170 may determine the correction gain CGAIN to be within this range based on the luminance ratio and a minimum correction gain
At S572, data correcting block 180 may generate corrected image data CDAT for the logo region LR by multiplying the input image data IDAT for the logo region LR by the correction gain CGAIN.
At 574, to generate the corrected image data CDAT for the peripheral region PR, data correcting block 180 may divide the peripheral region PR into a plurality of peripheral sub-regions (S574).
At S576, data correcting block 180 may determine a plurality of sub-region correction gains for respective ones of the plurality of peripheral sub-regions, so that the sub-region correction gains are greater than the correction gain CGAIN and less than 1.
At S578, data correcting block 180 may multiply the input image data IDAT for the plurality of peripheral sub-regions by the plurality of sub-region correction gains, respectively.
For example, as illustrated in
In some embodiments, as illustrated in
In this case, the following conversions may be performed: the input image data IDAT representing a 0-gray level to a 255-gray level may be converted to the corrected image data CDAT representing the 0-gray level to a 255-gray level 255*CGAIN multiplied by the correction gain CGAIN of about 0.5 with respect to the logo region LR as illustrated by curve 610 in
At S580, data driver 130 may drive a display panel 110 based on the corrected image data CDAT. Since the first sub-region correction gain SR1_CGAIN for the first peripheral sub-region PSR1 close to the logo region LR is close to the correction gain CGAIN, and the fourth sub-region correction gain SR4_CGAIN for the fourth peripheral sub-region PSR4 distant from the logo region LR is close to 1, the decreasing amount of the corrected image data CDAT from the input image data IDAT for the fourth peripheral sub-region PSR4 (which is distant from the logo region LR) may be less than the decreasing amount of the corrected image data CDAT from the input image data IDAT for the first peripheral sub-region PSR1 (which is close to the logo region LR). Thus, the luminance difference between the peripheral region PR and a region outside (or surrounding) the peripheral region PR may be reduced. Also, degradation and an afterimage in the logo region LR may be reduced, and a grayscale banding phenomenon in the logo region LR and peripheral region PR may be prevented.
Referring to
At S720, peripheral region setting block 160 may set a peripheral region PR adjacent to the logo region.
At S730, correction gain determining block 170 may calculate a first average gray level of the logo region LR.
At S742, correction gain determining block 170 may divide the peripheral region PR into a plurality of peripheral sub-regions.
At S744, correction gain determining block 170 may calculate a weighted-average gray level of the plurality of peripheral sub-regions with weights that decrease with increasing distance of the plurality of peripheral sub-regions to the logo region LR (S744).
At S750, correction gain determining block 170 may calculate a luminance ratio (of a weighted-luminance of the peripheral region PR to a luminance of the logo region LR) by dividing the weighted-average gray level of the plurality of peripheral sub-regions by the first average gray level of the logo region LR.
At S760, correction gain determining block 170 may determine a correction gain CGAIN to be greater than or equal to the minimum correction gain and less than or equal to 1. The correction gain CGAIN may be determined based on the luminance ratio and a minimum correction gain Since relatively high weights are applied to the peripheral sub-region(s) closer to the logo region LR and relatively low weights are applied to peripheral sub-region(s) more distant from the logo region LR, the correction gain CGAIN may produce a more pronounced effect for a peripheral image close to the logo.
At S772, data correcting block 180 may generate corrected image data CDAT for the logo region LR by multiplying the input image data IDAT for the logo region LR by the correction gain CGAIN.
At S776, to generate the corrected image data CDAT for the peripheral region PR, data correcting block 180 may divide the peripheral region PR into a plurality of peripheral sub-regions (e.g., substantially the same as the plurality of peripheral sub-regions determined by correction gain determining block 170) and may determine a plurality of sub-region correction gains for the plurality of peripheral sub-regions, with the plurality of sub-region correction gains being greater than the correction gain CGAIN and less than 1.
At S778, data correcting block 180 may multiply the input image data IDAT for the plurality of peripheral sub-regions by the plurality of sub-region correction gains, respectively.
At S780, data driver 130 may drive a display panel 110 based on the corrected image data CDAT. Since the sub-region correction gain for the peripheral sub-region close to the logo region LR is close to the correction gain CGAIN and the sub-region correction gain for the peripheral sub-region distant from the logo region LR is close to 1, the decreasing amount of the corrected image data CDAT from the input image data IDAT for the peripheral sub-region distant from the logo region LR may be less than the decreasing amount of the corrected image data CDAT from the input image data IDAT for the peripheral sub-region close to the logo region LR. Thus, the luminance difference between the peripheral region PR and a region outside (or surrounding) the peripheral region PR may be reduced. Further, degradation and afterimage effects in the logo region LR may be reduced, and a grayscale banding phenomenon in the logo region LR and the peripheral region PR may be prevented.
The processor 1110 may perform various computing functions or tasks. The processor 1110 may be, for example, an application processor (AP), a microprocessor, or a central processing unit (CPU). The processor 1110 may be coupled to one or more other components, for example, via an address bus, a control bus, a data bus, etc. In some embodiments, the processor 1110 may be coupled to an extended bus, e.g., a peripheral component interconnection (PCI) bus.
The memory device 1120 may store data for operations of the electronic device 1100 and may include at least one non-volatile memory device, such as an erasable programmable read-only memory (EPROM) device, an electrically erasable programmable read-only memory (EEPROM) device, a flash memory device, a phase change random access memory (PRAM) device, a resistance random access memory (RRAM) device, a nano floating gate memory (NFGM) device, a polymer random access memory (PoRAM) device, a magnetic random access memory (MRAM) device, a ferroelectric random access memory (FRAM) device, etc, and/or at least one volatile memory device such as a dynamic random access memory (DRAM) device, a static random access memory (SRAM) device, and a mobile dynamic random access memory (mobile DRAM) device.
The storage device 1130 may be a solid state drive (SSD) device, a hard disk drive (HDD) device, a CD-ROM device, or another type of storage device. The I/O device 1140 may be an input device such as a keyboard, a keypad, a mouse, a touch screen, etc, and an output device such as a printer, a speaker, etc. The power supply 1150 may supply power for operations of the electronic device 1100. The display device 1160 may be coupled to other components through the buses or other communication links.
In the display device 1160, a logo region may be detected, a correction gain may be determined based on a first average gray level of the logo region and a second average gray level of a peripheral region adjacent to the logo region, corrected image data may be generated by correcting input image data based on the correction gain, and a display panel may be driven based on the corrected image data. Accordingly, degradation and an afterimage effect in the logo region may be reduced. Also, grayscale banding in the logo region and the peripheral region may be prevented.
The inventive concepts according to one or more embodiments may be applied to any type of electronic device 1100 including display device 1160. Examples include a television (TV), a digital TV, a 3D TV, a smart phone, a wearable electronic device, a tablet computer, a mobile phone, a personal computer (PC), a home appliance, a laptop computer, a personal digital assistant (PDA), a portable multimedia player (PMP), a digital camera, a music player, a portable game console, a navigation device, etc.
The methods, processes, and/or operations described herein may be performed by code or instructions to be executed by a computer, processor, controller, or other signal processing device. The computer, processor, controller, or other signal processing device may be those described herein or one in addition to the elements described herein. Because the algorithms that form the basis of the methods (or operations of the computer, processor, controller, or other signal processing device) are described in detail, the code or instructions for implementing the operations of the method embodiments may transform the computer, processor, controller, or other signal processing device into a special-purpose processor for performing the methods herein.
Also, another embodiment may include a computer-readable medium, e.g., a non-transitory computer-readable medium, for storing the code or instructions described above. The computer-readable medium may be a volatile or non-volatile memory or other storage device, which may be removably or fixedly coupled to the computer, processor, controller, or other signal processing device which is to execute the code or instructions for performing the method embodiments or operations of the apparatus embodiments herein.
The controllers, processors, devices, blocks, modules, units, multiplexers, logic, interfaces, decoders, drivers, generators and other signal generating and signal processing features of the embodiments disclosed herein may be implemented, for example, in non-transitory logic that may include hardware, software, or both. When implemented at least partially in hardware, the controllers, processors, devices, blocks, modules, units, multiplexers, logic, interfaces, decoders, drivers, generators and other signal generating and signal processing features may be, for example, any one of a variety of integrated circuits including but not limited to an application-specific integrated circuit, a field-programmable gate array, a combination of logic gates, a system-on-chip, a microprocessor, or another type of processing or control circuit.
When implemented in at least partially in software, the controllers, processors, devices, blocks, modules, units, multiplexers, logic, interfaces, decoders, drivers, generators and other signal generating and signal processing features may include, for example, a memory or other storage device for storing code or instructions to be executed, for example, by a computer, processor, microprocessor, controller, or other signal processing device. The computer, processor, microprocessor, controller, or other signal processing device may be those described herein or one in addition to the elements described herein. Because the algorithms that form the basis of the methods (or operations of the computer, processor, microprocessor, controller, or other signal processing device) are described in detail, the code or instructions for implementing the operations of the method embodiments may transform the computer, processor, controller, or other signal processing device into a special-purpose processor for performing the methods described herein.
The foregoing is illustrative of embodiments and is not to be construed as limiting thereof. Although a few embodiments have been described, those skilled in the art will readily appreciate that many modifications are possible in the embodiments without materially departing from the novel teachings and advantages of the present inventive concept. Accordingly, all such modifications are intended to be included within the scope of the present inventive concept as defined in the claims. Therefore, it is to be understood that the foregoing is illustrative of various embodiments and is not to be construed as limited to the specific embodiments disclosed, and that modifications to the disclosed embodiments, as well as other embodiments, are intended to be included within the scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
10-2020-0113365 | Sep 2020 | KR | national |
This application is a continuation of U.S. application Ser. No. 17/326,684 filed on May 21, 2021, which claims the benefit of Korean Patent Application No. 10-2020-0113365, filed on Sep. 4, 2020, in the Korean Intellectual Property Office, the disclosures of which are incorporated by reference herein in their entirety.
Number | Date | Country | |
---|---|---|---|
Parent | 17326684 | May 2021 | US |
Child | 18313837 | US |