The application claims priority to Korean Patent Application No. 10-2020-0075230, filed Jun. 19, 2020, and all the benefits accruing therefrom under 35 U.S.C. § 119, the content of which in its entirety is herein incorporated by reference.
Embodiments of the invention relate to a display device and a method of driving the display device.
With the development of information technology, the importance of display devices, which are a connection medium between users and information, has been emphasized. In response to this, the use of display devices such as a liquid crystal display device, an organic light emitting display device, a plasma display device, and the like has been increasing.
A display device may include a plurality of pixels and display an image (frame) through a combination of light emitted from the pixels. When a plurality of different images are continuously displayed, a user may recognize the images as a moving image. In addition, when a plurality of identical images are continuously displayed, the user may recognize the images as a still image.
In a display device, when a still image is displayed for a long time, or when a part of a moving image such as a logo is displayed for a long time with a same luminance, pixel deterioration and afterimages may occur. In such a display device, grayscales of the logo can be corrected to prevent the afterimages.
Embodiments of the invention are directed to a display device in which a white logo and a color logo displayed in a logo area are accurately extracted and grayscales of the extracted logo are effectively corrected.
An embodiment of a display device according to the invention includes: pixels; an image converter which generates a second image by correcting grayscales of a first logo in a first image for the pixels; and a data driver which provides data signals corresponding to the second image to the pixels. In such an embodiment, the image converter detects the first logo based on value and saturation of the first image, generates first map data corresponding to the first logo, and specifies pixels corresponding to the first logo based on the first map data.
In an embodiment, the image converter may detect a second logo in the first image, generate second map data corresponding to the second logo, specify pixels corresponding to the second logo based on the second map data, and generate the second image by further correcting grayscales of the second logo.
In an embodiment, the image converter may include: a first logo detector which generates first sub-map data based on the value of the first image, generating second sub-map data based on the saturation of the first image, and generates the first map data by combining the first sub-map data and the second sub-map data; a second logo detector which generates the second map data based on a white mark of the first image; a logo determiner which generates third map data using the first map data and the second map data; and a grayscale converter which specifies the pixels corresponding to the first logo and the pixels corresponding to the second logo based on the third map data, and generates the second image by converting grayscales of the pixels corresponding to the first logo and the pixels corresponding to the second logo in the first image.
In an embodiment, the first logo detector may include a coordinate converter which converts the first image of RGB color space coordinates to a third image of HSV color space coordinates.
In an embodiment, the first logo detector may further include: a first map data extractor which generates the first sub-map data corresponding to an area having a value equal to or greater than a threshold value among the third image; and a second map data extractor which generates the second sub-map data corresponding to an area having a saturation equal to or greater than a threshold saturation among the third image.
In an embodiment, the first map data may be generated based on an intersection of the first sub-map data and the second sub-map data.
In an embodiment, the second logo detector may generate the second map data corresponding to an area having a white mark equal to or greater than a threshold white mark in the first image.
In an embodiment, the white mark may be a grayscale value of the first image.
In an embodiment, the second logo detector may generate the second map data based on the value of the first image.
In an embodiment, the third map data may be generated based on a combination of the first map data and the second map data.
In an embodiment, the first logo may include a color mark, and the second logo may include a white mark.
In an embodiment, the first logo detector and the second logo detector may generate the first map data and the second map data based on an Otsu binarization method.
An embodiment of a method of driving a display device according to the invention includes: detecting a first logo in a first image based on value and saturation of the first image; generating first map data corresponding to the first logo; detecting a second logo in the first image based on a white mark of the first image; generating second map data corresponding to the second logo; generating third map data using the first map data and the second map data; specifying pixels corresponding to the first logo and pixels corresponding to the second logo based on the third map data; and generating a second image by correcting grayscales of the pixels corresponding to the first logo and the pixels corresponding to the second logo in the first image.
In an embodiment, the generating the first map data may include: converting the first image of RGB color space coordinates to a third image of HSV color space coordinates; generating first sub-map data corresponding to an area having a value equal to or greater than a threshold value among the third image; generating second sub-map data corresponding to an area having a saturation equal to or greater than a threshold saturation among the third image; and generating the first map data by combining the first sub-map data and the second sub-map data.
In an embodiment, the first map data may be generated based on an intersection of the first sub-map data and the second sub-map data.
In an embodiment, the second map data may be generated corresponding to an area having a white mark equal to or greater than a threshold white mark in the first image.
In an embodiment, the white mark may be a grayscale value of the first image.
In an embodiment, the second map data may be generated based on the white mark and the value of the first image.
In an embodiment, the third map data may be generated based on a combination of the first map data and the second map data.
The above and other features of the invention will become more apparent by describing in further detail embodiments thereof with reference to the accompanying drawings, in which:
The invention now will be described more fully hereinafter with reference to the accompanying drawings, in which various embodiments are shown. This invention may, however, be embodied in many different forms, and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like reference numerals refer to like elements throughout.
It will be understood that, although the terms “first,” “second,” “third” etc. may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, regions, layers and/or sections should not be limited by these terms. These terms are only used to distinguish one element, component, region, layer or section from another element, component, region, layer or section. Thus, “a first element,” “component,” “region,” “layer” or “section” discussed below could be termed a second element, component, region, layer or section without departing from the teachings herein.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting. As used herein, “a”, “an,” “the,” and “at least one” do not denote a limitation of quantity, and are intended to include both the singular and plural, unless the context clearly indicates otherwise. For example, “an element” has the same meaning as “at least one element,” unless the context clearly indicates otherwise. “At least one” is not to be construed as limiting “a” or “an.” “Or” means “and/or.” As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” or “includes” and/or “including” when used in this specification, specify the presence of stated features, regions, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, regions, integers, steps, operations, elements, components, and/or groups thereof.
In addition, when an element is “coupled to” or “connected to” another element, this includes not only the case where the element is directly coupled to the other element, but also the case where another element is coupled therebetween. In contrast, when an element is referred to as being “coupled directly to” or “connected directly to” another element, there are no intervening elements present.
Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and the disclosure, and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Embodiments described herein should not be construed as limited to the particular shapes of regions as illustrated herein but are to include deviations in shapes that result, for example, from manufacturing. For example, a region illustrated or described as flat may, typically, have rough and/or nonlinear features. Moreover, sharp angles that are illustrated may be rounded. Thus, the regions illustrated in the figures are schematic in nature and their shapes are not intended to illustrate the precise shape of a region and are not intended to limit the scope of the claims.
Hereinafter, embodiments of the invention will be described in detail with reference to the accompanying drawings.
Referring to
The timing controller 100 may receive grayscales and control signals for each first image (frame) from an external processor. In one embodiment, for example, in the case of displaying a still image, the grayscales of consecutive first images may be substantially the same as each other. In one embodiment, for example, in the case of displaying a moving image, the grayscales of consecutive first images may be substantially different from each other. In such an embodiment, a part of the moving image may be a still area such as a logo.
The image converter 500 may generate a second image by correcting the grayscales of the logo in the first image.
In an embodiment, the image converter 500 may generate (or extract) map data corresponding to a logo area larger than the logo in the first image, and correct the grayscales of the logo using the generated map data.
In one embodiment, for example, the image converter 500 may generate first map data corresponding to a first logo including a color mark in the first image. In such an embodiment, the image converter 500 may generate second map data corresponding to a second logo including a white mark in the first image. In such an embodiment, the image converter 500 may generate third map data using the first map data and the second map data. The image converter 500 may specify (determine or select) pixels corresponding to the logo (for example, the first logo and/or the second logo) based on the third map data. In such an embodiment, the image converter 500 may generate the second image by correcting the grayscales of the pixels specified as corresponding to the logo.
The timing controller 100 may provide the grayscales of the second image to the data driver 200. In an embodiment, the timing controller 100 may provide control signals suitable for each specification to the data driver 200, the scan driver 300, or the like to display the second image.
In an embodiment, as shown in
The data driver 200 may provide data signals corresponding to the second image to pixels. In one embodiment, for example, the data driver 200 may generate the data signals to be provided to data lines DL1, DL2, DL3, . . . , and DLn using the grayscales of the second image and the control signals. In one embodiment, for example, the data driver 200 may sample the grayscales using a clock signal and apply the data signals corresponding to the grayscales to the data lines DL1 to DLn in units of pixel rows. A pixel row may mean pixels connected to a same scan line, where n may be an integer greater than 0.
The scan driver 300 may receive a clock signal, a scan start signal, or the like from the timing controller 100 and generate scan signals to be provided to scan lines SL1, SL2, SL3, . . . , and SLm, where m may be an integer greater than 0.
The scan driver 300 may sequentially supply the scan signals having a turn-on level pulse to the scan lines SL1 to SLm. In one embodiment, for example, the scan driver 300 may include scan stages configured in the form of a shift register. The scan driver 300 may generate the scan signals by sequentially transmitting the scan start signal in the form of a turn-on level pulse to a next scan stage based on the clock signal.
The pixel unit 400 may include the pixels. Each pixel PXij may be connected to a corresponding data line and a corresponding scan line, where i and j may be integers greater than 0. The pixel PXij may mean a pixel whose scan transistor is connected to an i-th scan line and a j-th data line. In an embodiment, each pixel PXij may receive voltages of a first power source VDD and a second power source VSS from outside. Here, the first power source VDD and the second power source VSS may be voltages used for the operation of the pixels. In one embodiment, for example, the first power source VDD may have a voltage level higher than a voltage level of the second power source VSS.
Referring to
A first electrode (for example, an anode electrode) of the light emitting element LD may be connected to the first power source VDD via the driving circuit DC, and a second electrode (for example, a cathode electrode) of the light emitting element LD may be connected to the second power source VSS. The light emitting element LD may emit light at a luminance corresponding to the amount of driving current controlled by the driving circuit DC.
The light emitting element LD may include or be composed of an organic light emitting diode. Alternatively, the light emitting element LD may include or be composed of an inorganic light emitting diode such as a micro light emitting diode (“LED”) or a quantum dot light emitting diode. Alternatively, the light emitting element LD may be an element including or composed of an organic material and an inorganic material. In an embodiment, as shown in
The first power source VDD and the second power source VSS may have different potentials from each other. In one embodiment, for example, a voltage applied through the first power source VDD may be greater than a voltage applied through the second power source VSS.
The driving circuit DC may include a first transistor T1, a second transistor T2, and a storage capacitor Cst.
A first electrode of the first transistor T1 (a driving transistor) may be connected to the first power source VDD, and a second electrode of the first transistor T1 may be electrically connected to the first electrode (for example, the anode electrode) of the light emitting element LD. A gate electrode of the first transistor T1 may be connected to a first node N1. The first transistor T1 may control the amount of driving current supplied to the light emitting element LD in response to a data signal supplied to the first node N1 through a data line DLj.
A first electrode of the second transistor T2 (a switching transistor) may be connected to the data line DLj, and a second electrode of the second transistor T2 may be connected to the first node N1. A gate electrode of the second transistor T2 may be connected to a scan line SLi.
The second transistor T2 may be turned on when a scan signal of a voltage (for example, a gate-on voltage) in a turn-on level, at which the second transistor T2 is turned on, is supplied from the scan line SLi, and thus the data line DLj and the first node N1 may be electrically connected. When the second transistor is turned on, the data signal of a corresponding frame may be supplied to the data line DLj, and accordingly, the data signal may be transmitted to the first node N1. A voltage corresponding to the data signal transmitted to the first node N1 may be stored in the storage capacitor Cst.
One electrode of the storage capacitor Cst may be connected to the first node N1, and another electrode of the storage capacitor Cst may be connected to the first electrode of the light emitting element LD. The storage capacitor Cst may be charged with the voltage corresponding to the data signal supplied to the first node N1, and may maintain the charged voltage until the data signal of the next frame is supplied.
In an embodiment, as shown in
Referring to
The logo area (or an area including the first logo LG1 and/or the second logo LG2) may be a still image area in which the position and grayscale are maintained in consecutive first images IMG1. In one embodiment, for example, the first logo LG1 may be a logo including the color mark, and the second logo LG2 may be a logo including the white mark. In such an embodiment, the first logo LG1 may be displayed in a form surrounding a part of the second logo LG2 (e.g., the letter “S” shown in
A logo area LGA may include the first and second logos LG1 and LG2 and may be an area larger than the first and second logos LG1 and LG2. In one embodiment, for example, the logo area LGA may be a rectangular area, such that the logo area LGA may be easily defined with coordinate values based on the x and y axes. In an alternative embodiment, the logo area LGA may be defined as other shapes such as a circle or an oval. An area other than the first and second logos LG1 and LG2 among the logo area LGA may be defined as a background.
Referring to
In an embodiment, the image converter 500 may generate (or extract) map data (first to third map data LMR1, LMR2, and LMF) corresponding to the logo area LGA in the first image IMG1, and correct the grayscales of the first logo LG1 and/or the second logo LG2 using the generated map data LMR1, LMR2, and LMF.
In one embodiment, for example, the image converter 500 may generate the first map data LMR1 corresponding to the first logo LG1 including the color mark in the first image IMG1. In such an embodiment, the image converter 500 may generate the second map data LMR2 corresponding to the second logo LG2 including the white mark in the first image IMG1. In such an embodiment, the image converter 500 may generate the third map data LMF using the first map data LMR1 and the second map data LMR2. The image converter 500 may specify the pixels corresponding to the first logo LG1 and/or the second logo LG2 based on the third map data LMF. In an embodiment, the image converter 500 may generate second image IMG2 by correcting the grayscales of the pixels specified as corresponding to the first logo LG1 and/or the second logo LG2.
The first logo detector 510 may detect the first logo LG1 in the first image IMG1 and generate the first map data LMR1 corresponding to the first logo LG1.
In an embodiment, the first logo detector 510 may convert the first image IMG1 from RGB color space coordinates to HSV color space coordinates to detect the first logo LG1 including the color mark, and detect the first logo LG1 based on value (or brightness) and saturation in the logo area LGA among the converted first image IMG1 (hereinafter, referred to as a third image).
Referring to
The coordinate converter 511 may convert the first image IMG1 of the RGB color space coordinates to a third image IMG1_1 of the HSV color space coordinates. In an embodiment, each pixel (for example, the pixel Pxij shown in
The first map data extractor 512 may generate (or extract) first sub-map data LMD1 based on the third image IMG1_1 of the HSV color space coordinates.
In an embodiment, the first map data extractor 512 may generate the first sub-map data LMD1 based on an area having the value equal to or greater than a predetermined threshold value in the logo area LGA.
In one embodiment, for example, as shown in
In an embodiment, the first logo LG1 including the color mark as well as the second logo LG2 including the white mark may have a high value. In such an embodiment, when a relatively bright image is displayed in the area (or background) excluding the first and second logos LG1 and LG2 among the logo area LGA according to the image displayed by the first image IMG1, the value in the corresponding area may be high. In this case, on the first sub-map data LMD1, the pixels corresponding to the first logo LG1 as well as the pixels corresponding to the second logo LG2 and/or the area in which the bright image is displayed (or a noise area NS) may be extracted as pixels having the threshold value Vth or higher.
The second map data extractor 513 may generate (or extract) second sub-map data LMD2 based on the third image IMG1_1 of the HSV color space coordinates.
In an embodiment, the second map data extractor 513 may generate the second sub-map data LMD2 based on an area having the saturation equal to or greater than a predetermined threshold saturation in the logo area LGA.
In one embodiment, for example, as shown in
In an embodiment, in the image displayed by the first image IMG1 as well as the first logo LG1 including the color mark, a high saturation image may be displayed in the area excluding the first and second logos LG1 and LG2 (or the background) among the logo area LGA. In this case, on the second sub-map data LMD2, the pixels corresponding to the first logo LG1 as well as the pixels corresponding to the area in which the high saturation image is displayed (or a noise area NS) may be extracted as pixels having the threshold saturation Sth or higher.
The map data generator 514 may generate the first map data LMR1 corresponding to the first logo LG1 by detecting the first logo LG1 including the color mark.
In an embodiment, the map data generator 514 may generate the first map data LMR1 using the first sub-map data LMD1 and the second sub-map data LMD2. In one embodiment, for example, since the first logo LG1 displayed in the logo area LG includes the color mark, the value and saturation of the first logo LG1 may be relatively high. The map data generator 514 may generate the first map data LMR1 of
Referring back to
In an embodiment, the second logo detector 520 may generate the second map data LMR2 based on an area having the white mark equal to or greater than a predetermined threshold white mark to detect the second logo LG2 including the white mark.
In one embodiment, for example, as shown in
In an embodiment, the white mark may be a grayscale value of the first image IMG1.
In an embodiment, the second logo detector 520 may generate the second map data LMR2 using the value as well as the white mark. In one embodiment, for example, the second logo detector 520 may generate the second map data LMR2 by extracting pixels having the white mark of 714 or more, which is the threshold white mark Wth, and the value of 714 or more, which is the threshold value Vth among the logo area LGA. Since the second logo LG2 including the white mark is displayed as a relatively bright image, when the second logo detector 520 generates the second map data LMR2 using the value as well as the white mark, accuracy may be further improved in extracting the second logo LG2.
In an embodiment, the first and second logo detectors 510 and 520 may use a conventional logo detection algorithm to extract the first and second logos LG1 and LG2. In one embodiment, for example, a logo detection algorithm using Otsu binarization method may be performed. Otsu binarization method is an adaptive thresholding way for binarization in image processing, which is well known in the art.
The logo determiner 530 may generate the third map data LMF using the first map data LMR1 and the second map data LMR2. In one embodiment, for example, the logo determiner 530 may generate the third map data LMF by extracting pixels extracted corresponding to the first logo LG1 on the first map data LMR1 and pixels extracted corresponding to the second logo LG2 on the second map data LMR2 as the pixels corresponding to the logo. In one embodiment, for example, the third map data LMF may be generated in the form of a union (or based on a combination) of the first map data LMR1 and the second map data LMR2 as shown in
The grayscale converter 540 may specify the pixels corresponding to the first and second logos LG1 and LG2 based on the third map data LMF, and generate the second image IMG2 by converting the grayscales of the specified pixels in the first image IMG1.
The grayscale converter 540 may generate the second image IMG2 by reducing the grayscales of the pixels corresponding to the first and second logos LG1 and LG2 in the first image IMG1. Accordingly, luminance of light emitted from the pixels corresponding to the first and second logos LG1 and LG2 among consecutive frame periods may be reduced to prevent afterimages.
In embodiments of the invention, as described above with reference to
Embodiments of the display device according to the invention may accurately extract a color logo as well as a white logo displayed in the logo area and correct the grayscales of the extracted logo. Accordingly, the pixel deterioration and afterimages in the logo area LGA may be removed (or reduced).
The invention should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete and will fully convey the concept of the invention to those skilled in the art.
While the invention has been particularly shown and described with reference to embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit or scope of the invention as defined by the following claims.
Number | Date | Country | Kind |
---|---|---|---|
10-2020-0075230 | Jun 2020 | KR | national |
Number | Name | Date | Kind |
---|---|---|---|
10027411 | Hong et al. | Jul 2018 | B2 |
20100008426 | Madden | Jan 2010 | A1 |
20150030247 | Shu | Jan 2015 | A1 |
20170069244 | Lim | Mar 2017 | A1 |
20180174529 | de Greef | Jun 2018 | A1 |
Number | Date | Country |
---|---|---|
101645136 | Aug 2016 | KR |
Number | Date | Country | |
---|---|---|---|
20210398493 A1 | Dec 2021 | US |