This application is based on and claims priority under 35 USC 119 from Japanese Patent Application No. 2019-164658 filed Sep. 10, 2019.
The present disclosure relates to an information processing apparatus and a non-transitory computer readable medium.
In Japanese Unexamined Patent Application Publication No. 2009-164969, a content playback apparatus that plays back video content with a structure in which additional information and a video are associated with each other through a playback time to display both the additional information and the video on a display device is disclosed. The content playback apparatus includes playback time calculating means for calculating, as a playback time of additional information in video content, at least part of the time from a playback time of the additional information to a playback time of the next additional information, playback speed control means for setting, as a playback speed for displaying the additional information and the video associated with the additional information, the ratio of the playback time to the time required for a viewer to view the additional information, and display control means for causing the additional information and the video associated with the additional information to be displayed on the display device at the playback speed.
In Japanese Unexamined Patent Application Publication No. 2009-16910, a video playback apparatus that plays back at least video data and subtitle data from a recording medium including the video data, audio data, and the subtitle data is disclosed. In the video playback apparatus, the subtitle data is extracted, the language of the extracted subtitle data is translated into another language, and subtitle data in the translated language is played back along with the video data.
In translating a subtitle on a video and displaying the translated subtitle, if a subtitle display time associated with the original subtitle before translation is applied to the translated subtitle, the subtitle display time may be shorter than a recognition time required to recognize the translated subtitle. For example, the number of letters in the translated subtitle may be larger than the number of letters in the original subtitle. In this case, a user is not able to recognize the translated subtitle correctly. To prevent this, for example, in the case where a video with a subtitle is translated and there is a part in which a display time for displaying the subtitle is shorter than a recognition time required to recognize the translated subtitle, the part is specified in advance and adjustment such as editing the video is performed for the part. Thus, before the video with the translated subtitle is played back, a user needs to understand the part in which the display time is shorter than the recognition time.
Aspects of non-limiting embodiments of the present disclosure relate to providing an information processing apparatus and a non-transitory computer readable medium that are capable of allowing, in the case where a subtitle on a video is translated and displayed, a user to understand, before the video with the translated subtitle is played back, that there is a part in which a display time associated with the subtitle is shorter than a recognition time required to recognize the translated subtitle.
Aspects of certain non-limiting embodiments of the present disclosure address the above advantages and/or other advantages not described above. However, aspects of the non-limiting embodiments are not required to address the advantages described above, and aspects of the non-limiting embodiments of the present disclosure may not address advantages described above.
According to an aspect of the present disclosure, there is provided an information processing apparatus comprising a processor programmed to acquire from a video a first subtitle in a first language, translate the first subtitle in the first language into a second subtitle in a second language, and display a notification for the second subtitle in a case where a display time for the first subtitle in the first language is shorter than a recognition time for the second subtitle in the second language.
Exemplary embodiments of the present disclosure will be described in detail based on the following figures, wherein:
Exemplary embodiments of the present disclosure will be described in detail with reference to drawings.
A multimedia content generation system according to an exemplary embodiment of the present disclosure includes, as illustrated in
The multimedia content generation system according to this exemplary embodiment generates multimedia content, which is a combination of various types of content such as videos, still images, sounds, characters, automatic translation, and the like. The multimedia content generation system according to this exemplary embodiment is able to generate, for example, multimedia content in which a subtitle is added to a video, the added subtitle is translated into a different language, and the translated subtitle is added to the video.
A subtitle mentioned herein represents information such as a commentary, conversation, or translation displayed as text on a screen of a video such as a movie or television. Such a subtitle may be transferred as subtitle information between the terminal apparatus 20 and the editing processing server 10.
The editing processing server 10 is an information processing apparatus into which editing software for editing various types of content to generate multimedia content is installed. The terminal apparatus 20 captures a video and generates multimedia content using the editing software running on the editing processing server 10.
Such editing software may not be installed in the editing processing server 10 but may be directly installed in the terminal apparatus 20 such as a PC and used.
The editing processing server 10 includes, as illustrated in
The CPU 11 performs predetermined processing based on a control program stored in the memory 12 or the storage device 13 to control an operation of the editing processing server 10. In this exemplary embodiment, explanation will be provided based on the assumption that the CPU 11 reads and executes the control program stored in the memory 12 or the storage device 13. However, the program may be stored in a recording medium such as a compact disc-read only memory (CD-ROM) and provided to the CPU 11.
The editing processing server 10 according to this exemplary embodiment includes, as illustrated in
The data communication unit 31 performs data communication with the terminal apparatus 20 via the network 30.
The controller 32 controls an operation of the editing processing server 10. The controller 32 includes a subtitle acquisition unit 41, a translation unit 42, a recognition time acquisition unit 43, a display time acquisition unit 44, a display control unit 45, and a user operation reception unit 46.
The data storing unit 33 stores various content data including video data on which editing processing is to be performed. The data storing unit 33 also stores a table indicating, for each language, the number of letters or the number of words in a subtitle that are able to be recognized per unit time.
The display control unit 45 controls screens displayed on the terminal apparatus 20.
The subtitle acquisition unit 41 acquires a subtitle in a first language from a video to which the subtitle is added.
The translation unit 42 translates the subtitle in the first language into a second language.
The display time acquisition unit 44 acquires a subtitle display time, which is a time in which the subtitle in the first language is displayed. Specifically, the display time acquisition unit 44 acquires, as the subtitle display time, a time from the display start time at which displaying the subtitle starts to the display end time at which displaying the subtitle ends.
Furthermore, in a case where a plurality of subtitles in the first language whose display times overlap at least partially are displayed in a section (also called a scene), which is a section of a video, the display time acquisition unit 44 acquires, as the subtitle display time, a time from the display start time at which displaying the first subtitle displayed out of the plurality of subtitles starts to the display end time at which displaying the last subtitle displayed out of the plurality of subtitles ends.
The recognition time acquisition unit 43 acquires a subtitle recognition time, which is a time required to recognize a subtitle in the second language obtained by translation by the translation unit 42.
The subtitle recognition time represents a time required to recognize a subtitle in the second language obtained by translation by the translation unit 42. In this example, the subtitle recognition time is a time required to read a subtitle and is calculated based on the number of letters or the number of words for each language. That is, the recognition time acquisition unit 43 acquires the subtitle recognition time based on the number of letters or the number of words in a translated subtitle in the second language. The subtitle recognition time may be set to be different according to the language.
Furthermore, in the case where a plurality of subtitles in the first language whose display times overlap at least partially are displayed in a section of a video, the recognition time acquisition unit 43 sums a plurality of subtitle recognition times for the second language after translation in the section of the video to obtain the subtitle recognition time.
In the case where the subtitle display time for the first language is shorter than the subtitle recognition time for the second language translated by the translation unit 42, the display control unit 45 functions as notifying means for displaying and notifying of a corresponding part.
Furthermore, the display control unit 45 performs control such that a still image at the display start time in a section in which the subtitle display time for the first language acquired by the subtitle acquisition unit 41 is shorter than the subtitle recognition time for the second language translated by the translation unit 42 is displayed, and notifies of a corresponding part.
Furthermore, the display control unit 45 performs control such that a section in which the subtitle display time for the first language acquired by the subtitle acquisition unit 41 is shorter than the subtitle recognition time for the second language translated by the translation unit 42 is repeatedly played back and displayed, and notifies of a corresponding part.
Furthermore, the display control unit 45 performs control such that a playback section in which the subtitle display time for the first language acquired by the subtitle acquisition unit 41 is shorter than the subtitle recognition time for the second language translated by the translation unit 42 is displayed in a display form different from other sections, and notifies of a corresponding part.
Furthermore, the display control unit 45 performs control such that a playback section in which the subtitle display time for the first language acquired by the subtitle acquisition unit 41 is shorter than the subtitle recognition time for the second language translated by the translation unit 42 is displayed in different forms between the case where the ratio of the subtitle display time for the first language to the subtitle recognition time for the second language is lower than a preset value and the case where the ratio of the subtitle display time for the first language to the subtitle recognition time for the second language is equal to or higher than the preset value, and notifies of a corresponding portion.
In the case where the subtitle display time for the subtitle in the first language is shorter than the subtitle recognition time for the second language, the translation unit 42 translates part of the subtitle in the first language into the second language. For example, in the case where the subtitle display time for the first language is shorter than the subtitle recognition time for the second language, the translation unit 42 translates part of words in the subtitle in the first language into the second language.
Furthermore, in the case where a subtitle in the first language is displayed in a section of a video, the translation unit 42 translates part of words in the subtitle.
Furthermore, in the case where a plurality of subtitles in the first language are displayed in a section of a video and the subtitle display time for the first language is shorter than the subtitle recognition time for the second language, the translation unit 42 translates any one of the plurality of subtitles in the first language into the second language.
Furthermore, the translation unit 42 translates part of a subtitle in the first language into the second language, based on a predetermined priority order, such that the subtitle recognition time for the translated second language is shorter than the subtitle display time for the original first language before translation.
Specifically, the translation unit 42 translates part of a subtitle in the first language into the second language, based on a priority order corresponding to an arrangement position of the subtitle in the first language in a video. Furthermore, the translation unit 42 preferentially translates a subtitle whose display form is different from other subtitles in a video, such as a subtitle in which the size of letters is different from those in other subtitles or a subtitle in which the color of letters is different from those in other subtitles.
The user operation reception unit 46 receives a priority order for a subtitle to be translated into the second language, out of subtitles in the first language acquired by the subtitle acquisition unit 41. Then, the translation unit 42 translates part of a subtitle in the first language into the second language, based on the priority order received by the user operation reception unit 46.
Next, an operation of the editing processing server 10 in the multimedia content generation system according to this exemplary embodiment will be described in detail with reference to drawings.
First, an operation of the editing processing server 10 will be schematically explained with reference to a flowchart of
First, in step S10, the subtitle acquisition unit 41 acquires a subtitle added to a video. Specifically, a Japanese subtitle illustrated in
In step S11, the display time acquisition unit 44 acquires a subtitle display time t for the subtitle acquired in step S10. The subtitle display time t represents a time from a subtitle display start time at which displaying a subtitle starts to a subtitle display end time at which displaying the subtitle ends. Furthermore, the subtitle display time t is set based on the table stored in the data storing unit 33. Specifically, in the case where the number of letters in a Japanese subtitle is 10, as illustrated in
In step S12, the translation unit 42 translates the Japanese subtitle acquired in step S10 into English. Specifically, the translation unit 42 translates the Japanese subtitle into an English subtitle “I like dogs.”.
In step S13, the recognition time acquisition unit 43 counts the number of translated English letters or the number of translated English words, and calculates the subtitle recognition time based on the number of letters or the number of words for English in the table stored in the data storing unit 33. Specifically, the recognition time acquisition unit 43 counts the number of translated English words as 3, and thus calculates the subtitle recognition time T as 1.5 seconds.
In step S14, the controller 32 determines whether the subtitle display time t is longer than the subtitle recognition time T.
In the case where, as illustrated in
In contrast, in the case where, as illustrated in
Specifically, for example, in a section of a video illustrated in
As illustrated in
The standard settings represents settings corresponding to the position in which a subtitle is displayed. In the case where the “standard settings” is selected on the priority setting screen illustrated in
Furthermore, in the case where “size of letters” is selected on the priority setting screen illustrated in
Furthermore, in the case where “color of letters” is selected on the priority setting screen illustrated in
Furthermore, in the case where “customized settings” is selected on the priority setting screen illustrated in FIG. 10, the translation unit 42 performs translation by giving priority to a part in a specified translation range of a Japanese subtitle in the video.
In step S16, for example, any one of the buttons on the priority setting screen illustrated in
Specifically, for example, when “size of letters” is selected and the “set” button is pressed on the priority setting screen illustrated in
Next, a second exemplary embodiment of the present disclosure will be described. In the second exemplary embodiment, a case where display times for a plurality of Japanese subtitles overlap in a section of a video, as illustrated in
As illustrated in
Furthermore, the subtitle recognition time T is calculated by summing subtitle recognition times for the plurality of translated subtitles in the second language in the section. That is, the recognition time acquisition unit 43 calculates the subtitle recognition time by summing the plurality of subtitle recognition times for the translated second language in a section of a video.
Specifically, for example, a section of a video illustrated in
An English subtitle “Hello”, which is obtained by translating the first Japanese subtitle displayed, is one word. Thus, a subtitle recognition time T1 is calculated as 0.5 seconds. In a similar manner, for an English subtitle obtained by translating the second Japanese subtitle displayed, a subtitle recognition time T2 is calculated as 0.5 seconds. Furthermore, an English subtitle “Nice to meet you!!”, which is obtained by translating the last Japanese subtitle displayed, is four words. Thus, a subtitle recognition time T3 is calculated as 2 seconds. Therefore, the subtitle recognition time T is calculated as 3 seconds by adding the subtitle recognition times T1=0.5 seconds, T2=0.5 seconds, and T3=2 seconds.
As illustrated in
In contrast, as illustrated in
Next, a modification of the priority setting screen in
As illustrated in
Furthermore, in the case where “customized settings” is selected on the display screen illustrated in
A user is able to select a part to be preferentially translated, while confirming an image of a part in which the subtitle display time t is shorter than the subtitle recognition time T as described above.
Next, the modification of the playback switch bar 50 described above will be described with reference to
In
Specifically, a part of a video in which the subtitle display time t is shorter than the subtitle recognition time T is displayed on the playback switch bar 50 in different display forms between the case where the ratio of the subtitle display time t for the first language to the subtitle recognition time T for the second language is lower than a preset value and the case where the ratio of the subtitle display time t for the first language to the subtitle recognition time T for the second language is equal to or higher than the preset value. For example, as illustrated in
For selection of priority for translation, by selecting a pointer for a part with a high priority for translation from among a plurality of pointers on the playback switch bar 50 and confirming a still image in a video, the playback position of the video is able to be switched.
In this exemplary embodiment, an example in which translation is performed from Japanese as the first language into English as the second language has been explained. However, the present disclosure is not limited to this. The present disclosure is also applicable to other languages, for example, in a case where translation is performed from English as the first language into Japanese as the second language.
Furthermore, in this exemplary embodiment, an example in which a translated subtitle in the second language is displayed on the subtitle in the first language is described. However, the present disclosure is not limited to this example. A translated subtitle in the second language may be displayed in place of a subtitle in the first language.
Furthermore, in this exemplary embodiment, an example in which in the case where the subtitle display time t is shorter than the subtitle recognition time T, a still image at the display start time of a part in which the subtitle display time t is shorter than the subtitle recognition time T is displayed has been explained. However, the present disclosure is not limited to this example. Moving images in a playback section for several seconds before and after a part in which the subtitle display time t is shorter than the subtitle recognition time T may be played backed repeatedly.
The foregoing description of the exemplary embodiments of the present disclosure has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the disclosure to the precise forms disclosed. Obviously, many modifications and variations will be apparent to practitioners skilled in the art. The embodiments were chosen and described in order to best explain the principles of the disclosure and its practical applications, thereby enabling others skilled in the art to understand the disclosure for various embodiments and with the various modifications as are suited to the particular use contemplated. It is intended that the scope of the disclosure be defined by the following claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
2019-164658 | Sep 2019 | JP | national |