This application is based upon and claims the benefit of priority under 35 U.S.C. §119 from Japanese Patent Applications No. 2012-089037, filed on Apr. 10, 2012, and No. 2012-143772, filed on Jun. 27, 2012, the entire contents of both of which are incorporated herein by reference.
The embodiments relate to a stereoscopic image display processing device, a stereoscopic image display processing method, and a stereoscopic image display processing program.
In recent years, 3D televisions (TVs), which are able to display three-dimensional (3D) contents, have been flourishing in the market. The 3DTVs are classified into a glasses method in which 3DTVs are viewed through glasses, and a glasses-free method in which 3DTVs are viewed without the need to wear glasses.
The glasses method includes two types of methods that are a frame sequential method and a line alternative method. In the frame sequential method, left and right (L and R) images are output alternately per frame on a TV side, and viewed using liquid crystal shutter glasses. In the line alternative method, L and R images are output alternately for each line in the same frame on a TV side, and viewed through polarized glasses.
The glasses-free method includes a parallax barrier method, a lenticular method, and so on.
3D contents include broadcast, and package software such as movies and video games. Currently, formats for 3D contents are defined by the high-definition multimedia interface (HDMI) 1.4a specification.
It is specified that a frame-packing format is supported at “1080p@23.98/24 Hz” and “720p@50 or 59.94/60 Hz”, a side-by-side format is supported at “1080i@50 or 59.94/60 Hz”, and a top-and-bottom is supported at “720p@50 or 59.94/60 Hz” and “1080p@23.98/24 Hz”.
The frame-packing format at “1080p” is mainly used for package software of movies, the frame-packing format at “720p” is mainly used for package software of video games, and the side-by-side format is mainly used for broadcast.
When viewing 3D contents on a 3DTV in a 3D viewing mode, images are decoded by the 3DTV in accordance with a 3D format. Images decoded and output are viewed through glasses in the glasses method, and viewed without glasses in the glasses-free method, thereby enabling the images to be viewed stereoscopically.
In viewing 3D broadcast aired by a TV station in a 3D viewing mode, some scenes are aired where a superimposed text showing emergency breaking news and so on is superimposed on the broadcast. When the TV station does not superimpose the same characters at the same positions on L and R images in the side-by-side format, but instead superimposes the superimposed text across the L and R images using an existing text superimposer similarly to 2D broadcast, a problem arises that the superimposed text is not readable in a 3D viewing state.
Japanese Patent Application Publication No. 2010-288234 (Patent Document 1) proposed a method for switching display between 3D and 2D on a receiving device side in accordance with display information contained in broadcast signals, when showing emergency information as a superimposed text while transmitting 3D contents from a transmission side.
In order to put the method of Patent Document 1 into practical use, it is required to decide specifications of display information to be contained in broadcast signals, and support the specifications on both transmission side and receiving side. In such a case, it is difficult in reality to decide and support details of specifications on both the transmission side, that is broadcasting stations, and the receiving side, that is receiving devices such as TVs produced by each company.
An object of the embodiments is to provide a stereoscopic image display processing device, a stereoscopic image display processing method, and a stereoscopic image display processing program by which a superimposed text is readable even when viewing 3D broadcast.
A first aspect of the embodiments provides a stereoscopic image display processing device, comprising: a decoding section configured to output a display image signal based on an input processed image signal after executing 3D decoding processing of the processed image signal based on a 3D format of the processed image signal, or without executing the 3D decoding processing; a determination section configured to determine a state of superimposed text information in the input processed image signal; and a control section configured to control the decoding section to execute or not to execute the 3D decoding processing of the processed image signal based on a determination result by the determination section.
A second aspect of the embodiments provides a stereoscopic image display processing method, comprising: determining a state of superimposed text information in an input processed image signal; and controlling execution or non execution of 3D decoding processing of the processed image signal based on a determination result regarding the state of the superimposed text information.
A third aspect of the embodiments provides a stereoscopic image display processing program for causing a computer to execute the instructions, comprising: an instruction for determining a state of superimposed text information in an input processed image, and an instruction for controlling execution or non execution of 3D decoding processing of the processed image signal based on a determination result regarding the state of the superimposed text information.
The embodiments will be explained below with reference to the accompanying drawings. In the following statements regarding the drawings, same or similar parts are denoted by the same or similar reference numerals. However, it should be noted that the drawings are schematic, and a device, a system configuration, and so on illustrated therein are different from those in reality. Therefore, specific configurations shall be determined in consideration of the following explanation.
The embodiments described below represent examples of a device and a method for embodying technical thoughts of the present invention, and the technical thoughts of the present invention do not limit materials, shapes, configurations, arrangements, and so on of constituents to those described below. Various changes may be added to the technical thoughts of the present invention without departing from the technical scope set forth in the claims.
As shown in
The image processing unit 1 is configured by various processors and the like for controlling the stereoscopic image display processing device 100, and processing various signals input into the input section 10.
An input image signal Fi, and an identification signal Fm that defines a format of the input image signal Fi are input to the input section 10 of the image processing unit 1, and the input section 10 outputs a processed image signal Fp based on the identification signal Fm.
For example, an image signal output from an external device such as a Blue-ray disc (BD) recorder, or an image signal of broadcast received by a TV tuner is input as the input image signal Fi. Simultaneously, when an image signal is input through an HDMI cable that is connected to an external device such as a BD recorder, the identification signal Fm is input from the external device in order to identify the 3D format of the input image signal Fi based on the HDMI 1.4a specification.
When an external device is connected via a cable other than the HDMI cable, and broadcast received by a TV tuner is input as an image signal, a set value corresponding to a 3D decoding method that is arbitrarily set by a viewer via remote control is input as the format identification signal Fm.
When the identification signal Fm that is input in the input section 10 represents an interlaced signal for either 2D or 3D, the input section 10 executes interlace-progressive (IP) conversion processing of the input image signal Fi. The input section 10 outputs a progressive image after the IP conversion processing as the processed image signal Fp. Meanwhile, when a format other than above is input in the input section 10, the input section 10 outputs the input image signal Fi as it is as the processed image signal Fp.
The determination section 12 determines a state of superimposed text information in the processed image signal Fp. The “state of superimposed text information” means whether or not the processed image signal Fp contains superimposed text information. For example, the determination section 12 determines the state of superimposed text information based on whether or not a previously registered string is present in the processed image signal Fp. When a determination result indicates that the string is present, the determination section 12 outputs a value 1, and, when the string is not present, the determination section 12 outputs a value 0, as a determination signal Tp of superimposed text information.
For the determination section 12, a method described in, for example, Japanese Patent Application Publication No. 2009-217303 may be used.
The determination section 12 may be configured by dedicated hardware, or may use software to have a substantially equivalent function by using a CPU of a normal computer system. The determination section 12 includes an auxiliary storage device (not shown). The auxiliary storage device stores a program that determines superimposed text information. Also, the auxiliary storage temporarily stores data being processed and data of processing results in determination of superimposed text information.
Determination processing of superimposed text information executed at the determination section 12 according to the embodiment will be explained using a flowchart shown in
In step S102, the rectangular region image is converted into two values where pixels containing a character are converted into 1, and pixels containing no character are converted into 0 with respect to the rectangular image in the pixel region, and a character image is thus generated. In step S103, characters in the generated character image are recognized by using an existing optical character recognition (OCR) engine or the like. It is determined whether previously registered characters such as “breaking news”, “earthquake”, “tsunami”, “news”, and “seismic intensity” are present in the character image, and, a value 1 is output when such characters are present, and a value 0 is output when such characters are not present, as the determination signal Tp of superimposed text information.
When it is determined at the determination section 12 that superimposed text information is not contained in the processed image signal Fp, and a 3D image signal is contained in the processed image signal Fp, the decoding section 14 executes 3D decoding processing in accordance with a format of the identification signal Fm and generates a display image signal Fd. As shown in
When a 3D image signal is contained in the processed image signal Fp, the control section 16 controls the decoding section 14 based on the sate of superimposed text information determined by the determination section 12.
When it is determined that superimposed text information is contained in the processed image signal Fp, the control section 16 controls the decoding section 14 so that the decoding section 14 does not execute 3D decoding processing of the processed image signal Fp, but outputs the processed image signal Fp and the superimposed text information as they are as the display image signal Fd.
When it is determined by the determination section 12 that superimposed text information is not contained in the processed image signal Fp, the control section 16 controls the decoding section 14 so that the decoding section 14 executes 3D decoding processing and outputs the display image signal Fd generated by the 3D decoding processing.
As shown in
As shown in
The control section 16 outputs a control signal Cd as 1 only when the mode selection signal Sd is 1 and the determination signal Tp is 0. This means that the control signal Cd is limited to 0 when the determination signal Tp is 1, even if the mode selection signal Sd is 1.
The frame memory 20 of the decoding section 14 stores the processed image signal Fp input from the input section 10. The frame memory 20 separates a pair of 3D images within the processed image signal Fp based on the control signal Cd from the control section 16 and the identification signal Fm from the input section 10, and reads the pair of 3D images as an L image signal Fl and an R image signal Fr. The upscale section 22 enlarges the L and R image signals Fl and Fr based on the control signal Cd and the identification signal Fm, and outputs an L enlarged image signal FL and an R enlarged image signal FR to the display unit 3 as the display image signals Fd, respectively.
Specific operations of the frame memory 20 and the upscale section 22 of the decoding section 14 will be explained. When the determination signal Fm represents a 2D signal, the processed image signal Fp is read as it is for both the L image signal Fl and the R image signal Fr regardless of the value of the control signal Cd. The L and R image signals Fl and Fr are output as the L enlarged image signal Fl and the R enlarged image signal FR without being enlarged at the upscale section 22. In short, the processed image signal Fp is output as it is as the L and R enlarged image signals FL and FR.
Next, as shown in
In this case, a superimposed text in the L image 30a and a superimposed text in the R image 30b do not have the same string. Therefore, conventionally, a superimposed text on an L enlarged image 30L and a superimposed text on an R enlarged image 30R have different strings, as shown in
In the embodiment, when the control signal Cd is 1, in other words, when it is determined that no superimposed text information is present, a left half of the processed image 30 of the processed image signal Fp is read as the L image 30a, and the right half thereof is read as the R image 30b by the frame memory 20. The sizes of the L and R images 30a and 30b are doubled horizontally at the upscale section 22, respectively, and the L and R enlarged images 30L and 30R are output.
On the other hand, when the control signal Cd is 0, in other words, when superimposed text information is present, or the 3D viewing mode is off, the processed image 30 is output as the L and R enlarged images 30A and 30B, while remaining in the side-by-side state.
As shown in
As described above, in the embodiment, the side-by-side format is used to explain a 3D format. In the embodiment, the 3D decoding processing is not executed when the control signal Cd is 0 in a case of other formats. Processing of the decoding section 14 in a case of other formats will be explained below.
For example, when the identification signal Fm represents the top-and-bottom format at “1080p@23.98/24 Hz”, and when the control signal Cd is 1, an upper half of the processed image signal Fp is read as an L image signal Fl, and a lower half thereof is read as an R image signal Fr. The sizes of the L and R image signals Fl and Fr are doubled vertically at the upscale section 22, and output as an L enlarged image signal FL and an R enlarged image signal FR.
Meanwhile, when the control signal Cd is 0, the processed image signal Fp is output as the L and R image signals Fl and Fr while remaining in the top-and-bottom state. The L and R image signals Fl and Fr are output as L and R enlarged image signals FL and FR, respectively, without being enlarged at the upscale section 22. In short, the processed image signal Fp is output as it is as the L and R enlarged image signals FL and FR.
When the identification signal Fm represents the top-and-bottom format at “720p@50 or 59.94/60 Hz”, and when the control signal Cd is 1, an upper half of the processed image signal Fp is read as an L image signal Fl, and a lower half thereof is read as an R image signal Fr. The L and R image signals Fl and Fr are enlarged 1.5 times horizontally and 3 times vertically at the upscale section 22 so as to be enlarged to an image size equivalent to 1080p, and output as an L enlarged image signal FL and an R enlarged image signal FR, respectively.
Meanwhile, when the control signal Cd is 0, the processed image signal Fp is output as the L and R image signals Fl and Fr while remaining in the top-and-bottom state. The L and R image signals Fl and Fr are enlarged 1.5 times horizontally and 1.5 times vertically at the upscale section 22, and output as L and R enlarged image signals FL and FR, respectively. In short, the processed image signal Fp is enlarged to an image size equivalent to 1080p in the top-and-bottom state, and output as the L and R enlarged image signals FL and FR.
When the identification signal Fm represents the frame-packing format at “1080p@23.98/24 Hz”, an L image and an R image in an LR packing image that is input as the processed image signal Fp are read as an L image signal Fl, and an R image signal Fr, respectively, regardless of the value of the control signal Cd. The L and R image signals Fl and Fr are output as L and R enlarged image signals FL and FR without being enlarged at the upscale section 2.
When the identification signal Fm represents the frame-packing type at “720p@50 or 59.94/60 Hz”, an L image and an R image in a LR packing image that is input as the processed image signal Fp are read as an L image signal Fl and an R image signal Fr, respectively, regardless of the value of the control signal Cd. The L and R image signals Fl and Fr are enlarged 1.5 times in both horizontally and vertically to be enlarged to an image size equivalent to 1080p, and output as L and R enlarged signals FL and FR, respectively.
The display unit 3 is, for example, a display with Full-HD resolution using a frame sequential method. The display unit 3 stores the input L and R enlarged image signals FL and FR in a frame memory in the display unit 3, and reads and displays the L and R enlarged image signals FL and FR alternately per frame.
Ina case of the frame-packing format at “1080p@23.98/24 Hz”, and in a case of the top-and-bottom format at “1080p@23.98/24 Hz”, a frame frequency is different from other formats. Therefore, the display unit 3 reads the L and R enlarged image signals FL and FR stored in the frame memory while executing pull-down conversion of the L an R enlarged image signals FL and FR, and displays the same in the frame sequential method.
In the embodiment, a frame sequential display was used as the display unit 3, but the display unit 3 is not limited thereto. A line-alternate or glasses-free display may also be used, and input L and R enlarged image signals FL and FR may be displayed as L and R images in accordance with the display method.
The storage unit 5 is configured of a ROM (a read only memory), a RAM (a random access memory), or the like, and stores a stereoscopic image display processing program executed in the image processing unit 1 (a computer).
As explained above, in the stereoscopic image display processing device according to the embodiment, when it is determined at the determination section 12 that a previously registered characters such as “news” and “breaking news” is contained in the processed image signal that has been input, the processed image signal is output as it is without executing 3D decoding.
In the conventional technology, while 3D broadcast aired by a TV station is viewed in a 3D viewing mode, when a scene is broadcasted with a superimposed text such as breaking news that is superimposed across L and R images of the broadcast as shown in
On the other hand, in the stereoscopic image display processing device according to the embodiment, images as shown in
With the configuration of the foregoing stereoscopic image display processing device, subtitles of movies or the like that are present at a lower part of a 3D screen could be detected as a superimposed text, and a problem happens that 3D decoding is stopped even though broadcast without any superimposed text superimposed thereon is actually viewed. Generally, a superimposed text showing breaking news and the like is often superimposed in a region in an upper half of a screen.
For example, as shown in
The determination section 12 determines superimposed text information in the determination region 34 that is set in the processed image 30, based on the region information Ta output from the setting section 11. Although the superimposed text information is determined in the determination region 34 that is set in an upper half of the processed image 30 as shown in
As explained so far, since it is possible to arbitrarily set a region where superimposed text information is determined, false detection of strings such as subtitles of movies is prevented.
A recording and reproduction apparatus such as a BD recorder is used as an external device in order to play back package software or record and reproduce 3D broadcast. For example, in a case where a 3DTV is used for 3D viewing through a BD recorder, and when reproducing BD 3D software, both the recorder and the TV are automatically switched to operation modes suitable for a 3D format based on the HDMI 1.4a specification.
Meanwhile, when 3DTV broadcast recorded in a recorder (the side-by-side format is used in the 3DTV broadcast) is viewed in 3D, it is not possible to identify the 3D format and automatically switch operations. In such a case, the recorder side reproduces images in a 2D operation mode, and the TV side is manually switched to the side-by-side decoding mode and operated.
As the recorder is controlled during 3D viewing to play back or pause, or to display a menu screen, an OSD character image of the recorder is displayed. For example, as shown in
In this case, since the recorder operates in a 2D mode, the same OSD characters are not displayed on an L image 30a and an R image 30b of a side-by-side image, but the OSD characters are superimposed on one frame of a processed image 30 of the side-by-side image as one frame of a 2D image. Therefore, normal stereoscopic view of the OSD characters is not possible.
Namely, when different OSD characters are present in L and R images 30a and 30b, the OSD characters are not readable. Also, when OSD characters are present only in either one of the L and R images 30a and 30b, flicker is felt. This leads to a problem that accumulation of eye fatigue is caused.
In the modification example of the embodiment, the determination section 12 determines not only whether or not superimposed text information is contained, but also whether or not the L and R images 30a and 30b have the same superimposed text information. In other words, the determination section 12 distinguishes between the L image 30a and the R image 30b in accordance with the identification signal Fm with respect to the input image signal Fi that has been input to the determination section 12, and detects strings from the respective L and R images 30a and 30b. The determination section 12 determines whether or not there is a string contained only in either one of the L and R images 30a and 30b.
The determination section 12 outputs a superimposed text information determination signal Tp which takes a value 1 when there is a string that is present only in either one of the L and R images 30a and 30b, and a value 0 when there is not such a string. This determination is carried out for each pair of the L and R images 30a and 30b.
The modification example of the embodiment is different from the embodiment in that it is determined not only whether or not superimposed text information is contained, but also whether or not the L and R images 30a and 30b have the same superimposed text information. The rest of the configurations are the same as those of the embodiment, and duplicated statements will be omitted.
Determination processing of superimposed text information executed at the determination section 12 according to the modification example of the embodiment will be explained by using a flowchart shown in
In step S200, superimposed text information is detected in the L image 30a. For example, the processing is executed in accordance with the flowchart shown in
In step S201, superimposed text information is detected in the R image 30b. Character code data is generated based on a character recognition result in the R image 30b, and stored in the auxiliary storage device of the determination section 12.
In step S202, the character code data is compared between the L and R images 30a and 30b, and it is determined whether or not there is a string that is present only in either one of the L and R image 30a and 30b. Based on a result of the character code data comparison, the determination signal Tp of the superimposed text information is generated.
Specifically, first, the determination section 12 determines whether or not there is a string that is present only in the L image 30a, and generates a 1-bit determination result signal J1. The determination section 12 determines whether or not each piece of the character code data detected in the L image 30a is detected in the R image 30b.
As a result of the comparison, when there is at least one piece of character code data that is not detected in the R image 30b, the determination section 12 outputs the determination result signal J1 as 1, and, when there is no such character code data, the determination section 12 outputs the determination result signal J1 as 0. The determination section 12 also outputs the determination result signal J1 as 0 when there is no superimposed text information detected in the L image 30a at all.
Similarly, the determination section 12 determines whether or not there is a string that is present only in the R image 30b, and generates a 1-bit determination result signal J2. The determination section 12 determines whether or not each piece of the character code data detected in the R image 30b is detected in the L image 30a.
As a result of the comparison, when there is at least one piece of character code data that is not detected in the L image 30a, the determination section 12 outputs the determination result signal J2 as 1, and, when there is no such character code data, the determination section 12 outputs the determination result signal J2 as 0. The determination section 12 also outputs the determination result signal J2 as 0 when there is no superimposed text information detected in the R image 30b at all.
The determination section 12 generates the determination signal Tp of the superimposed text information by using the determination result signals J1 and J2. For example, the determination section 12 carries out logical operation in which the value 0 is taken only when both the determination judgment signals J1 and J2 are 0, and the value 1 is taken otherwise. In other words, when there is superimposed text information that is present only in the L image 30a, or there is superimposed text information that is present only in the R image 30b, the determination signal Tp becomes the value 1. The determination signal Tp is a 1-bit signal, and is output from the determination section 12 as a final determination signal.
In the modification example of the embodiment, when different OSD characters are superimposed in the L and R images 30a and 30b or when OSD characters are superimposed on either one of the L and R images 30a and 30b in a 3D viewing operation mode, 3D display is automatically switched off. As a result, the OSD characters become viewable. In addition, since a viewer no longer sees different OSD characters with the left and right eyes or sees OSD characters with only one eye, eye fatigue is not induced.
<Stereoscopic Image Display Processing Program>
The stereoscopic image display processing program causes a computer to execute each processing in the foregoing image processing unit 1 as an instruction. It is also possible to store the stereoscopic image display processing program in a computer readable storage medium, and provide the stereoscopic image display processing program as a stereoscopic image display processing program stored in a computer readable storage medium.
Although the embodiment of the present invention was explained as above, it should not be understood that the present invention is limited to the statements and drawings included in this disclosure. Various alternative embodiments, examples, operation techniques will be obvious to those skilled in the art from this disclosure. Therefore, the technical scope of the present invention is defined only by a matter specifying the invention according to the reasonable scope of patent claims based on the foregoing explanation.
Number | Date | Country | Kind |
---|---|---|---|
2012-089037 | Apr 2012 | JP | national |
2012-143772 | Jun 2012 | JP | national |