This application is based upon and claims the benefit of priority from Japanese Patent Application No. 2009-49522, filed on Mar. 3, 2009; the entire contents of which are incorporated herein by reference.
The present invention relates to an apparatus and a method for presenting a content and another content related to the content.
As to a PC (Personal Computer), a retrieval function is a general technique widely used in daily business. For example, Web retrieval to retrieve information related with a specified keyword from a Web site on Internet is used by many persons. In the retrieval function with the PC, by inputting a keyword related with things (retrieval objects) from a user, a retrieval result is acquired. Briefly, a retrieval method which the user need actively input the keyword is ordinary. Hereinafter, this retrieval method is called “active retrieval method”. Furthermore, IT (Information Technology) of an AV equipment (such as a TV, a DVD recorder, a HDD recorder) is progressing, and the AV equipment is connectable with an Internet. Accompanying with this, the AV equipment prepares the active retrieval function same as the PC.
However, an ability to store a content in the AV equipment, a service of Video-On-Demand (VOD), and a seamlessness of the content between equipments (outside of the AV equipment) or the Internet, are enlarged, and the number of contents accessible by the user increases quickly. Furthermore, the user's style to acquire and view the content by spread of the HDD recorder is changing. Briefly, a style to record the user's desired content is changing as a style to previously record the user's interest contents and select the user's favorite one from the contents, or a passive style (a recommendation service, an automatic recording service). Accordingly, the contents accessible by the user often include one which the user cannot understand.
On the other hand, it is an important problem to effectively select the user's desired content from a group of contents accessible by the AV equipment. As mentioned-above, the user does not always understand all contents because the style to acquire and view the content is changing. Accordingly, when the user retrieves the content using the active retrieval function, the user suitably select a retrieval keyword, and it is difficult for the user to effectively retrieve the user's desired content.
Furthermore, as to the AV equipment, a user is not always strong in the PC. In other words, the user is not always familiar with the active retrieval of keyword input type, which is a general function in the PC. As a result, the active retrieval method in the AV equipment is a function only used by a partial active user. Briefly, this method is not convenient for many users.
In order to solve above-mentioned problem, instead of the active retrieval method, a passive retrieval method is used. As to the passive retrieval method, for example, contents related with some content are presented, and the user selects his/her favorite one from the presented contents (called “relational contents”). In this method, the user's desired content is not actively (positively) searched using the keyword. By setting some content as a retrieval source, relevancy between this content and other contents is presented to the user, and the user implicitly retrieves his/her desired content from the other contents.
Briefly, the user retrieves his/her favorite content without an explicit keyword. As a result, as to a plurality of contents accessible by the AV equipment, while watching a visualized result of relevance among the plurality of contents, the user can passively select the desired one. In this way, in the passive retrieval method, it is very important for the user to visualize the relevancy between contents.
On the other hand, in a field of photograph, various techniques to visualize relationship among a large number of images (taken by a digital camera) are proposed. For example, in WO 00/33572 (Kokai) ( . . . Patent Reference 1), in order to represent a relationship among a plurality of images in time series, the plurality of images is enlarged in order of time series, and spirally displayed from an inner circle to an outer circle in order of time series.
However, as to a dynamic image such as the TV or the video, which is an object of the AV equipment, the image of each frame is changing in order while playing, and the relevancy between images is also changing dynamically. In the technique disclosed in the Patent Reference 1, the dynamic image cannot be processed because a static image is a processing object. Accordingly, as to the content (the dynamic image) changeable with passage of time, it is desired that the relevancy among contents is visualized with the change.
The present invention is directed to an apparatus and a method for visually presenting contents related to the processing source content changing with passage of time.
According to an aspect of the present invention, there is provided an apparatus for presenting contents, comprising: a selection unit configured to select a first content as a processing object; a timer configured to generate a time information; a detection unit configured to detect a change of at least a part of the first content based on the time information; a layout unit configured to acquire a second content as a next processing object after the change is detected, a third content related to the first content and a fourth content related to the second content, and symbolically lay out the second content, the third content and the fourth content in a virtual space; and a presentation unit configured to visually present the virtual space.
Hereinafter, embodiments of the present invention will be explained by referring to the drawings. The present invention is not limited to the following embodiments.
(1) Component:
A content presentation apparatus includes a controller comprising a CPU (Central Processing Unit) and a GPU (Graphics Processing Unit), a storage unit comprising a ROM (Read Only Memory), a RAM (Random Access Memory) or a HDD (Hard Disk Drive) each storing various program and various data (image, video), a communication unit communicating with an external apparatus, and a bus connecting thereof. The content presentation apparatus has a hardware component using a regular computer. Furthermore, as to the content presentation apparatus, a display unit displaying information, and an operation input unit (such as a keyboard or a mouse) accepting an input of the user's indication, are (wirelessly or wired-) connected. The display unit is a display device such as a CRT (Cathode Ray Tube) or a liquid crystal monitor. The operation input unit is an input device preparing an operation key, an operation button or a mouse to input the user's operation.
Next, in this hardware component, each function realized by executing each program stored in the storage unit with the CPU in the content presentation apparatus is explained.
The content information database 102 correspondingly stores a content and metadata (content information) of the content. Concretely, in case of indicating a specified content, the content information database 102 supplies metadata corresponding to the content. Furthermore, in case of indicating a predetermined item of the specified content, the content information database 102 supplies information corresponding to the predetermined item of the metadata. At the same time, in case of presenting a predetermined keyword, the content information database 102 retrieves metadata (or item of the metadata) including the keyword, and presents a retrieval result representing which metadata (or which item of metadata) matches the keyword and which content correspond the metadata. This function is called “relational database function”.
In this case, “content” is defined as a block of information viewed by a person, i.e., digital data of a video, an image, music, text, or combination thereof. As an example of the video, a TV broadcasting program, a commodity such as a cinema or a video clip delivered/sold/distributed by a recording medium (such as a DVD) or a VOD service, a dynamic image delivered on Web, a dynamic image taken by a camera or a cellular-phone with a camera function, and a program recorded by a video recorder, an HDD recorder, a DVD recorder or TV/PC having recording function, are listed. As an example of the image, digital photograph data taken by a digital camera are listed. As an example of the music, a music album sold/distributed as a memory medium (such as a CD), and tune data (for example, MP3 data) delivered/sold/distributed via an on-line store of the Web, are listed.
The metadata (content information) of the content is a group of information belonging to the content, and mainly information for the purpose to explain the content. For example, if the content is a TV program, the metadata is a group of a title, a sub-title, a broadcasting station, a broadcast type, a broadcast start date (time), a broadcast end date (time), a broadcast time length, contents of the program, a performer information, a program genre information, a producer information, and a caption information.
For example, in case of a TV program by a Japanese digital terrestrial broadcasting, ARIB (Association of Radio Industries and Businesses) prescribes a standard format to set the group of above-mentioned information. In the digital terrestrial broadcasting, a broadcasting wave on which metadata is superimposed as the prescribed format is delivered. In the same way, as to another kind of content, a standard format is often prescribed. For example, if the content is a photograph taken by a digital camera, as to metadata of the content, a format advocating as Exif (Exchangeable Image File Format) is widely used (de facto standard).
The metadata is not always sold/delivered/distributed by a content creator or a distributor, after assigning the metadata to the content thereby. The metadata is often assigned to the content by a user afterward. For example, if the content is a photograph taken by a digital camera, the user often adds a comment text to the content afterward, in order to discriminate the photograph. Furthermore, a person except for the creator and the distributor often assigns the metadata using any method. For example, in a recent video recorder (including the HDD recorder), a scene change or a CM (commercial) part is automatically detected from a TV broadcast program recorded, and this change information is automatically added as chapter information to the TV broadcast program. In such equipment, a function to detect a scene change is previously installed. The chapter information is generated using this function, and added as metadata. Above explanation is only one example. Metadata has various formats, i.e., the metadata is one unit of additional information related with some content, and information arranged by each item according to a predetermined rule.
The metadata includes metadata of all of the content, and metadata of a part of the content. In case of TV program, metadata includes a title, a sub-title, a broadcasting station, a broadcast type, a broadcast start date (time), a broadcast end date (time), a broadcast time length, contents of the program, a performer information, a program genre information, a producer information, and a caption information. In these data, the title, the sub-title, the broadcasting station, the broadcast type, the broadcast start date (time), the broadcast end date (time), the broadcast time length, the contents of the program, the performer information, the program genre information, and the producer information, are metadata of all of the content. On the other hand, in a closed caption information (one datum stored as the caption information), contents information corresponding to a position of broadcasting time of the TV program is supplied. Concretely, contents “OOO” is broadcasted in a period of “1˜5” minutes passed from the broadcast start time, and contents “XXX” is broadcasted in a period of “6˜10” minutes passed from the broadcast start time, i.e., a pair of text representing a time schedule and the broadcast contents is supplied. In this case, (a part of) the closed caption information is metadata of a part of content based on time. In case of the chapter information to the content, program contents and a chapter name in some chapter segment are stored as the metadata. This is also one example of the metadata of the at least of the content.
For example, correspondence between metadata and the content is explained. A content identifier (ID) to identify the content is assigned to the content, and the content identifier is corresponded with the metadata. The content identifier is a content name or a content number.
Hereinafter, in order to simplify the explanation, the content presentation apparatus 100 is AV (audio visual) equipment or a recording equipment of a TV program, the content is the TV program, and metadata is assigned to the TV program. In the case of the AV equipment of the TV program, the content presentation apparatus 100 is a TV set, a video recorder or a personal computer (having AV view function), each includes a function to display the TV program (broadcasted from a broadcasting station) on a screen. In the case of the recording equipment of TV program, the content presentation apparatus 100 is a TV set (having a recording function), a video recorder or a personal computer (having AV recording/view function), each includes a function to record a TV program and play the recorded TV program. However, forms of the content and the content presentation apparatus are not limited to above examples.
In
(a1) Selection from TV Program List Supplied as EPG (Electric Program Guide):
In AV equipments of TV program widely spread, a GUI (Graphical User Interface) screen to select a program by referring to a program list (EPG) similar to TV program list of a news paper is provided. In this screen (EPG screen), typically, programs of one week broadcasted on digital terrestrial broadcasting, BS digital broadcasting and CS digital broadcasting, are listed as a TV program list by classifying into each channel with date (time). From this TV program list, a TV program selected by a user via an operation input unit is set to the processing source content.
(a2) Selection from a List of TV Programs Recorded:
In recording equipments of TV program widely spread, the GUI screen to select a program by referring to a list of TV programs recorded is provided. From the list of TV programs on the GUI screen, a TV program selected by a user via the operation input unit is set to the processing source content.
(a3) Selection from a List of TV Programs Presently Broadcasting (Live Broadcasting):
In AV equipments of TV program widely spread, by changing a channel or a broadcasting type via the operation input unit, the user can select and view a program presently broadcasting. Furthermore, some AV equipment presents a list of TV programs presently broadcasting on the GUI screen. If the content presentation apparatus 100 is such AV equipment, from the list of TV programs on the GUI screen, a TV program selected by the user via the operation input unit is set to the processing source content.
(a4) Selection from a List of TV Programs Listed Based on Another Predetermined Reference:
For example, as to general AV/recording equipment of TV program, a function to search a TV program (a TV program having a specified keyword is searched from EPG screen, and the search result is listed, or a TV program belonging to specified genre is searched, and the search result is listed) is prepared. From the list of TV programs (search result) on the GUI screen, a TV program selected by the user via the operation input unit is set to the processing source content. The list of TV programs by this search function means TV programs listed based on another predetermined reference. However, another predetermined reference is not limited to this search function, but applied to all functions to list TV programs.
(a5) Selection of a Regular TV Program:
The regular TV program means a TV program which is a source content according to a predetermined rule. For example, a TV program presently viewed (a TV program broadcasting on a channel presently selected, or a TV program (recorded) presently replaying) is selected as the processing source content. If any content is not viewed presently, a content most previously viewed may be selected as the processing source content. Furthermore, by preparing a profile information (indicating a specified keyword or user's liking), the processing source content may be automatically selected based on the profile information. For example, in case of “succor program” as the profile information, a content having program information “succor program” is selected as the processing source content. Furthermore, by storing a user's operation history (operation history information) to the content presentation apparatus 100, a TV program which the user often viewed may be selected as the processing source content based on the operation history information. For example, the case that the user often viewed means, in the operation history information, the number of contents corresponding to metadata of the same program information is above a threshold.
The content information acquisition unit 103 acquires metadata corresponding to the content is acquired. Concretely, in response to a content identifier of a predetermined content, the content information acquisition unit 103 retrieves all or a part of items of metadata corresponding to the content identifier from the content information database 102. In case of TV program, as the part of items of metadata, for example, program contents are retrieved. As to the part of items of metadata, the content information acquisition unit 103 acquires based on a predetermined reference. The reference is, regularly, a time, chapter information, or a specified keyword. In case of the closed caption information, as to the reference “Is information matched after OO minutes from start time of broadcast ?”, a program as a segment including this time is acquired. In this case, the content information acquisition unit 103 may acquire a segment corresponding a period including the time (after OO minutes from start time of broadcast. Furthermore, the reference is not limited to above-mentioned example. For example, if the content presentation apparatus 100 includes a function to calculate a surge degree of a TV broadcast by analyzing acoustic information included in all the TV broadcast, “a part having the surge degree above a predetermined threshold” may be set to the reference. Furthermore, if the content presentation apparatus 100 includes a function to estimate a person appeared in a TV broadcast by analyzing image of the TV broadcast, “a segment in which a specified person appears” may be set to the reference. The reference is not limited to above-mentioned examples.
The time control unit 104 controls time information. Concretely, the time control unit 104 times the date, the present time and day of the week, and outputs as a time information. Furthermore, the time control unit 104 calculates a difference between the present data and a specified data, a difference between the present time and a specified time, and a difference between two specified days, and outputs as the time information.
Based on the time information (controlled by the time control unit 104), the source content update unit 105 detects a change of the processing source content (presently selected by the source content selection unit 101), and updates the processing source content. Concretely, the source content update unit 105 decides whether the processing source content (presently selected by the source content selection unit 101) is valid. In case of invalid, by retrieving a valid content, the source content update unit 105 selects the valid content as a new processing source content, and updates the processing source content to the new processing source content.
As mentioned-above, the source content selection unit 101 selects the processing source content using various methods. For example, the source content selection unit 101 selects a TV program presently broadcasted via a channel (presently selected) as the processing source content. In this case, after the present time is over a broadcast end time of the TV program, a TV program presently broadcasted via this channel is different from the processing source content selected by the source content selection unit 101. For example, a program A is broadcasted in a period “13:00˜15:00” via a channel A, and a program B is broadcasted from “15:00” via the channel A. When the source content selection unit 101 selects the processing source content at “14:50”, the program A is set to the processing source content. However, at “15:00” when ten minutes has passed from 14:50, broadcast of the program A is completed, and the program B starts to be broadcasted. In this case, the source content update unit 105 automatically updates the processing source content.
A condition to decide whether the content is valid is as follows.
(b1) Assume that a processing source content is selected from “TV programs presently broadcasted” (irrespective of selection method). When the present time (generated from the time control unit 104) is not over a broadcast end time of the processing source content, the processing source content is valid. When the present time is equal to or over the broadcast end time, the processing source content is invalid.
(b2) Assume that the processing source content is selected from a TV program except for “TV programs presently broadcasted”. The processing source content is always valid.
The condition to decide whether the content is valid is not limited to above example, and may be suitably changed as follows.
(c1) Assume that the processing source content is a video content as VOD delivery and selected by the source content selection unit 101 while delivering the video content. When the present time is equal to or over a delivery end time of the video content, the processing source content is invalid.
(c2) Assume that the processing source content is selected from “TV programs presently broadcasted” (irrespective of selection method). When the present time is equal to or over a broadcast end time of the processing source content, if the processing source content is already recorded, the processing source content is valid. Furthermore, if the processing source content is selected from “TV programs to be broadcasted in future”, the same decision condition may be used.
When the source content update unit 105 updates the processing source content, the source content history control unit 106 sets the processing source content (before updating) to a content as a previous source content, and updates a history of the past processing source content. Concretely, at timing when the source content update unit 105 updates the processing source content, the source content history control unit 106 stores the previous processing source content and the update time as processing source history information. The number of the previous processing source content is not limited to one. The previous processing source contents as arbitrary units and update order of the contents are stored as the processing source history information.
The relational content calculation unit 107 specifies a content related to the present processing source content selected by the source content selection unit or updated by the source content update unit 105, and a content related to the past processing source content stored in the source content history control unit 106. Then, the relational content calculation unit 107 acquires metadata corresponding to each of specified contents (candidate contents), the present processing source content and the past processing source content from the content information database 102 via the content information acquisition unit 103, and calculates a relevance between each of the specified contents and each of the present processing source content and the past processing source content, using the metadata. The relational content calculation unit 107 lists a plurality of contents each having the relevance in higher order from candidate contents. The plurality of contents listed is called “relational content”. The number of the contents listed (relational contents) may be limited. Alternatively, all candidate contents may be listed without limiting the number.
The candidate content (specified candidate for relational content) is, concretely, a content to be selected as the processing source content by the source content selection unit 101. At least one candidate content exists. This content (a selection source content) is changed by realization form of the content presentation apparatus 100. In case that the content presentation apparatus 100 is AV equipment or recording equipment of TV program, typically, contents listed on the GUI screen (presenting EPG information) and contents (corresponding to the number of channels) presently broadcasted are regarded as selection source contents. Furthermore, in case that the content presentation apparatus 100 is the recording equipment of TV program, a content recorded by the apparatus 100 are regarded as the selection source content.
Recently, the AV equipment or the recording equipment of TV program mutually connectable with another equipment using DLNA (Digital Living Network Alliance) function exists. In this case, a content stored in the another equipment is regarded as the selection source content. Furthermore, in case of an IPTV view equipment or the equipment connectable with a VOD service or a Network, a content provided by IPTV, a content provided by a VOD service and a content existing on the Network are regarded as the selection source content. These selection source contents are candidate contents for the relational content, and objects to calculate the relevance. From the candidate contents (a set of source contents to calculate the relevance), a plurality of contents is listed in higher order of the relevance, as the relational contents.
As to the relevance, the calculation method used by a relevance calculation unit 17 disclosed in JP-A 2009-80580 (Kokai) ( . . . Patent reference 2) is applied. In this case, a relevance r between a processing source content X and some content Y is calculated by a function F, as following equation (1).
r=F(X,Y) (1)
This calculation example is described in the patent reference 2. As to actual value of X and Y, at least one item included in metadata is used. For example, actual value of X is a broadcast start date (time) of the content X, and actual value of Y is a broadcast start date (time) of the content Y. The function F(X, Y) becomes larger if the broadcast start date of X is nearer to the broadcast start date of Y. The function F(X, Y) becomes smaller if the broadcast start date of X is farther from the broadcast start date of Y.
In the first embodiment, a source content used for calculation of the relevance different from that in the patent reference 2. In this reference, the source content is the processing source content and a content of a calculation object of the relevance. However, in the first embodiment, a next processing source content (updated by the source content update unit 105) and a past processing source content (controlled by the source content history control unit 106) are further included.
An example that above-mentioned relevance is applied to the first embodiment is explained. In the first embodiment, the relational content calculation unit 107 calculates a relevance R using a following equation (2).
R=G(Δt)×F(X,Y)+(1−G(t))×F(Z,Y) (2)
In the equation (2), X represents a processing source content, Y represents a candidate content of a relational content, and Z represents a previous processing source content in past processing source contents (controlled by the source content history control unit 106). As actual values of X, Y and Z, at least one item included in metadata corresponding to the content is respectively used. “Δt” represents a passing time after selecting (or updating to) the present processing source content. Briefly, if the present processing source content is selected by the source content selection unit 101, “Δt” represents a passing time from the selection time to the present time. If the present processing source content is newly updated by the source content update unit 105, “Δt” represents a passing time from the update time to the present time. “G(Δt)” is a function of an interval “Δt”, and outputs a value within “0“˜”1” based on “Δt”. For example, “G(Δt)” is realized as a function shown in
A concrete example to calculate the relevance is explained. Assume that the processing source content transits as shown in
R=G(t−t1)×F(A,Y)+(1−G(t−t1))×F(B,Y) (3)
With regard to the relevance R, as the passing time is nearer to “T (arbitrary constant)”, a contribution ratio of relevance between X and Y (calculated by F(X,Y)) becomes larger. On the other hand, as the passing time is nearer to “0”, a contribution ratio of relevance between Z and Y (calculated by F(Z,Y)) becomes larger. Briefly, if a passing time from selection of or update to the present processing source content is not so long, a relevance ratio with a previous processing source content Z is larger while a relevance ratio with the present processing source content X is smaller. If the passing time becomes longer, the relevance ratio with the present processing source content X is larger while the relevance ratio with the previous present processing source content Z is smaller.
In the equation (3), as the past processing source content used for calculation of the relevance, the previous processing source content (immediately before the present processing source content) is used. However, the past processing source content before a plurality of contents from the present processing source content may be used. In this case, the relevance is calculated by following equation (4).
R=G(Δt)×F(X,Y)+Σ(1−Gn(Δt))×F(Zn,Y) (4)
In the equation (4), “Zn” is a processing source content before n-units from the present processing object, and “Gn(Δt)” is a function to determine a weight of the processing source content “Zn”, which satisfies “ΣGn(Δt)=1”.
Based on the relevance calculated by the relational content calculation unit 107, the layout unit 108 symbolically lays out at least one content included in relational contents (specified by the relational content calculation unit 107) and the present processing source content into a virtual space. Then, the layout unit 108 visualizes the virtual space, and presents the visualized virtual space via the presentation unit 109.
Concretely, from relational contents specified by the relational content calculation unit 107, the layout unit 108 selects a relational content to be laid in the virtual space, according to a predetermined condition. The predetermined condition is, for example, the number of relational contents, the relevance degree (calculated by the relational content calculation unit 107) above a threshold, a broadcast date (or recording date) within a predetermined range, or a specified genre, but not limited to them. A condition originated from metadata corresponding to the relational content may be set to the predetermined condition. As to a method for selecting the relational content, the method used by the display source content selector in the patent reference 2 is applied.
The layout unit 108 symbolizes the present processing content and the relational content (selected as mentioned-above) respectively, and lays out them into a first-dimensional space, a second-dimensional space or a third-dimensional space (virtual space) based on their relevance. Then, the layout unit 108 visualizes the virtual space, and presents the visualized virtual space via the presentation unit 109. As to a method for symbolizing the content, the method used by the content visualization unit in the patent reference 2 is applied. As to a method for laying out the content, the method used by the visualized content layout unit in the patent reference 2 is applied. As to a method for visualizing the virtual space, the method used by the space visualization unit in the patent reference 2 is applied.
For example, as to symbolization and layout of the content, by using all or a part of items included in metadata corresponding to the content, the layout unit 108 generates an image (representing the content as a picture or a text), and lays out the image in the virtual space. In this case, the present processing source content is laid at the origin, and each relational content is laid around the origin so that the relevance of the relational content is larger as the relational content is nearer to the origin. Furthermore, as to visualization of the virtual space, the virtual space (each content is symbolically laid out) is rendered, and an image representing the virtual space is generated.
The presentation unit 109 displays the image representing the virtual space (visualized by the layout unit 108) via a display. Briefly, by presenting the virtual space, the present processing source content and the relational content are presented.
(2) Operation:
Next, processing of the content presentation apparatus 100 of the first embodiment is explained by referring to
In
In
At S63, it is decided whether all of candidate contents are already processed. If at least one candidate content is not processed yet (No at S63), at S64, metadata corresponding to the at least one candidate content is acquired from the content information database 102. At S65, by using metadata of the present processing source content (acquired at S61), metadata of the past processing source content (acquired at S62) and metadata of the candidate content (acquired at S64), the relevance is calculated by the equation (2).
After S65, processing is returned to S63. If all candidate contents are already processed (Yes at S63), at S66, from the all candidate contents, the candidate content with the relevance is listed in order of higher relevance. Each of contents listed is specified as the relational content.
In
As mentioned-above, as to a TV program of the present processing source content, another TV program related to the TV program is presented as the relational content. Accordingly, the user can easily view which content exists in relation to the present specified content. Concretely, for example, in case that one TV program is selected as the processing source content using EPG, the user can view which TV programs related to the one TV program are included in EPG of one week. As a result, the user can reserve recording of all TV programs related. As to the prior art, in order to operate the same reservation via a EPG screen, the user need search the related TV program by watching TV programs listed on the EPG screen. Furthermore, the user need search his/her desired TV program using above-mentioned program search function. In former case, it takes a long time for the user to search TV programs in all channels of one week, and this is not actual search method. In latter case, the user must consider a suitable keyword. If the user considers unsuitable keyword, the user's desired TV program cannot be searched. On the other hand, in the first embodiment, by presenting TV programs related to the (present and past) processing source content, the user can reserve all TV programs related, at once.
In case that a recorded TV program is the present processing source content, the same operation is possible for the user. In this case, by presenting another recorded TV program (as a relational content) related to the recorded TV program, the user can utilize the relational content as a candidate to view the other recorded TV program. Furthermore, by presenting TV programs scheduled to be broadcasted hereafter (as a relational content) related to the recorded TV program, the user can utilize the relational content as a candidate to reserve recording of TV programs scheduled to be broadcasted.
Furthermore, irrespective of a type of content, the relational content can be specified. For example, as to a TV program on some digital terrestrial broadcasting, a dynamic image (delivered on Network) related to the TV program is specified as the relational content. In this case, the user can search his/her desired content from relational contents over the type of content.
Recently, the number of contents accessible by the user are rapidly increasing. Accordingly, as shown in the first embodiment, it is important to support the user to search his/her desired content from many contents, and it is important to present significant content for the user to urge to view or record. In the patent reference 2, based on relevance between contents, a content highly related to some content is listed, and dynamically visualized. Briefly, by presenting the relational content, the user can be urged to select the significant content. However, in the patent reference 2, if the processing source content (as a source to present the relational content) changes with passage of time, the user's utility is damaged.
For example, even if a TV program (selected as the processing source content) is presently broadcasted, time is passing while the user is operating the content presentation apparatus. As a result, the TV program is completely broadcasted, and a next TV program starts to be broadcasted, i.e., the processing source content changes. In this case, in the patent reference 2, the next TV program is set to a next processing source content, and relational contents of the next processing object are specified. In general, before and after change of the processing source content, the relational content largely changes. Because, if the processing source content is not correlative with the next processing source content, the relational content of the processing source content is not also correlative with the relational content of the next processing source content.
If a user changes the processing source content intentionally, a problem does not occur. However, irrespective of the user's intention, when the processing source content changes with passage of time, the relational content also changes before and after change of the processing source content, and the user is confused. For example, assume that relational contents of the processing source content (before change) are presented, and a user is searching a content to be viewed or recorded from the relational contents. If the relational contents are changed by change of the processing source content, the relational contents to be searched by the user are not presented. Briefly, the relational contents are discontinuously changed, and the user cannot continue to search from the relational contents on condition before change of the processing source content.
However, in the first embodiment, when the processing source content is changed, a new processing source content and previous processing source content are taken into consideration, and contents related to each of the new processing source content and the previous processing source content are presented. Accordingly, immediately after the (previous) processing source content is changed to the new processing source content, contents related to the previous processing source content are mainly presented. As time has passed, these relational contents are not presented gradually. On the other hand, contents related to the new processing source content are presented gradually as time has passed. When a predetermined time (T in
In this way, in the first embodiment, even if the processing source content is changed with passage of time, relational contents to be presented are continually changed. Accordingly, it is avoided that the user suddenly misses a content to be searched. At the same time, new contents related to the new processing source content are also presented. Accordingly, the relational contents can be continually presented without damaging the user's utility.
Next, the content presentation apparatus of the second embodiment is explained. As to the same unit as the first embodiment, same sign is assigned, and its explanation is omitted.
(1) Component:
The relational content calculation unit 107 specifies a content related to the present processing source content selected by the source content selection unit 101 or updated by the source content update unit 105. Next, the relational content calculation unit 107 acquires metadata corresponding to each of specified contents (candidate contents), the present processing source content from the content information database 102 via the content information acquisition unit 103, and calculates the relevance between each of the candidate contents and the present processing source content, using the metadata. The relational content calculation unit 107 lists a plurality of contents (relational contents) in order of higher relevance from candidate contents. In the second embodiment, the relational content calculation unit 107 does not process the past processing source content (controlled by the source content history control unit 106). The relational content calculation unit 107 has the same component as that in the patent reference 2.
In the second embodiment, the relational content calculation unit 107 calculates the relevance R by following equation (5).
R=F(X,Y) (5)
In the equation (5), “X” represents the processing source content, and “Y” represents a content as an object to calculate the relevance. As actual value of “X” and “Y”, in the same way as the first embodiment, metadata corresponding to the content is used.
When the source content update unit 105 updates the processing source content, the relational content calculation-history control unit 110 sets the processing source content (before update) to the past processing source content (previous processing source content), and controls processing source history information of the update time and the past processing source content. Concretely, the relational content calculation-history control unit 110 stores the processing source history information. The processing source history information represents the update time, the past processing source content, the relational content (specified by the relational content calculation unit 107) of the past processing object and the relevance.
As to the relational content (present relational content) specified by the relational content calculation unit 107, the relational content (past relational content) represented by the processing source history information (controlled by the relational content calculation-history control unit 110) and the present processing source content, the layout unit 108 symbolically lays out them in a virtual space. In the same way as the first embodiment, the layout unit 108 visualizes the virtual space, and presents the visualized virtual space via the presentation unit 109.
Concretely, based on the relevance of the present relational content and the relevance of the past relational content (represented by the processing source history information), the layout unit 108 symbolizes the present processing source content, the present relational content and the past relational content respectively, and lays out each symbolized content in the virtual space. As a method for laying out the content, for example, the method used by the visualized content layout unit disclosed in the patent reference 1 is applied. Furthermore, in the second embodiment, in order to separately lay out the present relational content and the past relational content, the layout unit 108 divides a region of the virtual space.
A method for dividing the region is explained. In this example, the virtual space is two-dimensional space (xy-plane), the present processing source content is laid at the origin, and relational contents are laid around the origin based on the relevance. The region of the virtual space is divided by following equations (6) and (7).
S1:X2+y2≦N×G(Δt) (6)
S2:S−S1 (7)
In equations (6) and (7), S represents all region of the virtual space, S1 represents a region to lay out the present relational content, S2 represents a region to lay out the past relational content, N represents an arbitrary constant, and G(Δt) represents a function of passing time Δt. For example, G(Δt) is realized as a function shown in
Based on a relevance of the present relational content (specified by the relational content calculation unit 107) and a relevance of the past relational content represented by the processing source history information (controlled by the relational content calculation-history control unit 110), the layout unit 108 symbolically lays out the present relational content in the region S1 and the past relational content in the region S2. A method for laying out the symbolized present relational content is same as that of the first embodiment. In the region S2, for example, by setting a layout start point as a point having a distance “N×G(Δt)” from the origin, the symbolized past relational content is laid out.
In above explanation, as a method for dividing the region, an example that the region is increasing as a circle region is explained. However, the method is not limited to this example. For example, a fan-shaped region having an angle θ° from a positive direction of x-axis may be the region S1, and the region S excluding the region S1 may be S2. Furthermore, a shape of the divided region may not be a circle, or a continuous region. For example, the region S1 may be set by following equation (8).
S1:M×G(Δt)≦x2+y2N×G(Δt) (8)
In the equation (8), a region excluding the region S1 can be set as the region S2 (M<N). In this case, a doughnut-shaped region is S1, and two discontinuous regions (inner side region and outer side region of S1) are S2.
In above explanation, an example that the divided region is changing with passage of time is shown. However, this is only one example, and the divided region is not limited to this example. A factor to divide the region may not be passage of time. Irrespective of the passage of time, the region S may be divided into a plurality of fixed regions.
Furthermore, in above-explanation, an example that the region is represented in two-dimensional space is shown. However, this is only one example, and the region is not limited to this example. For example, the region may be represented in three-dimensional space
After that, the layout unit 108 visualizes the virtual space in which each symbolized content is laid out, and presents the visualized virtual space via the presentation unit 109. This processing is same as the first embodiment.
(2) Operation:
Next, processing of the content presentation apparatus 100′ of the second embodiment is explained. Steps of content presentation processing are same as steps in
At S6, based on a relevance of the present relational content (specified at S5) and a relevance of the past relational content represented by the processing source history information (stored at S4), the layout unit 108 symbolically lays out the present relational content, the present relational content and the past relational content in the virtual space. As mentioned-above, the virtual space is divided into a plurality of regions. The symbolized present processing source content, the symbolized present relational content and the symbolized past relational content, are laid out in each region. After that, the virtual space in which the symbolized content is laid out in each region is visualized. Step S7 is the same as that of the first embodiment.
As mentioned-above, in the second embodiment, in case of calculating the relevance, the past processing source content is not taken into consideration. However, in case of laying out in the virtual space, the past processing source content is taken into consideration. Briefly, in addition to the relational content of the present processing source content, the relational content of the past processing source content is symbolically laid out. Accordingly, even if the processing source content changes with passage of time, relational contents to be laid out in the virtual space is changing continuously (not suddenly) before and after the change of the processing source content.
Furthermore, in the second embodiment, a region to lay out the present relational content is different from a region to lay out the past relational content. Accordingly, the present relational content and the past relational content are not presented in confusion. The user can correctly understand that the relational content relates to the present processing source content or the past processing source content.
Next, the content presentation apparatus of the third embodiment is explained. As to the same unit as the first embodiment or the second embodiment, same sign is assigned, and its explanation is omitted.
(1) Component:
Based on time information generated from the time control unit 104, the inter-content reference position control unit 111 detects a reference position as a position of a part of the present processing source content (selected by the source content selection unit 101), and detects update of the reference position. This reference position can be detected using metadata. As explained in the content information acquisition unit 103 of the first embodiment, as to the content, metadata of all of the content and metadata of a part of the content are corresponded. For example, assume that the content presentation apparatus 100″ is the recording equipment and the present processing source content is a TV program recorded. In the content presentation apparatus 100″, while the present processing source content is played, a play position is gradually changing with passage of time. Based on change of the play position, metadata corresponding to the play position (or adjacency of the play position) are also changing. The inter-content reference position control unit 111 acquires the metadata via the content information acquisition unit 103, and detects change of the reference position when the metadata corresponding to the play position changes. Above-explanation is only one example, and a method for detecting change of the reference position is not limited to this example. For example, the reference position is not limited to one part of the content, and may be some range of the content. Furthermore, the inter-content reference position control unit 111 may simultaneously control a plurality of reference positions.
When the inter-content reference position control unit 111 detects change of the reference position in the processing source content, the source content history control unit 106 controls a reference position history information to represent metadata of the reference position (before update) as metadata of a past reference position. Concretely, when the inter-content reference position control unit 111 detects change of the reference position, the source content history control unit 106 acquires the detection time (update time) from the time control unit 104, and stores the reference position history information correspondingly representing the update time and metadata of the reference position (before update). The number of the reference position history information is not limited to one. The reference position history information of arbitrary units may be stored with update order of the reference position.
When the inter-content reference position control unit 111 detects change of the reference position, the relational content calculation unit 107 specifies a relational content of the present processing source content (selected by the object) and a relational content of the past reference position represented by the reference position history information (controlled by the source content history control unit 106). Concretely, as to the present processing source content and each of candidate contents (of relational content), the relational content calculation unit 107 acquires metadata of all of the content from the content information database 102 via the content information acquisition unit 103. Furthermore, the relational content calculation unit 107 acquires metadata corresponding to the past reference position in the present processing source content from the content information database 102. By using metadata of each content and metadata of the reference position, the relational content calculation unit 107 calculates a relevance between the present processing source content and each of the candidate contents, and calculates a relevance between the past reference position and each of the candidate contents. From the candidate contents, the relational content calculation unit 107 lists a candidate content in order of higher relevance. The candidate content listed is a relational content related to the present processing source content and the past reference position. A method for calculating the relevance is same as that of the first embodiment, for example, the equation (2) is used.
(2) Operation:
Next, processing of the content presentation apparatus 100″ of the third embodiment is explained by referring to
As mentioned-above, even if metadata corresponding to all of the processing source content does not change, when update of the reference position is detected based on change of metadata corresponding to a part of the processing source content, by calculating the relevance with each of candidate contents again, the relational content to be presented can be changed. For example, assume that a TV program (song program) is selected as the processing source content, a singer A appears at a time t1, and a singer B appears at a time t2. In a period when the singer A appears, metadata of the singer A is acquired. In a period when the singer B appears, metadata of the singer B is acquired. In this case, at the time t1, in addition to metadata of all of song program, metadata of the singer A is acquired. Accordingly, by calculating the relevance using both metadata, the relational content related to the song program and the singer A is presented. On the other hand, at the time t2, in addition to metadata of all of song program, metadata of the singer B is acquired. Accordingly, by calculating the relevance using both metadata, the relational content related to the song program and the singer B is presented.
As mentioned-above, in the third embodiment, even if the processing source content does not change on the whole, when a part of the processing source content changes with passage of time, a relational content can be dynamically presented based on change of the part. For example, while a relational content of a scene in a TV program (presently broadcasted) is presented, when the scene changes to another scene in the TV program (subject of the TV program largely changes), a relational content of another scene is presented, i.e., the relational content to be presented largely changes. As explained in the first embodiment, if a relational content to be presented based on change of the processing source content changes discontinuously, it often happens that the user suddenly misses the relational content to be searched. However, in the third embodiment, even if the subject of the TV program changes by changing a scene broadcasted, in addition to a relational content related to all the TV program, a relational content related to a scene (after changed) is presented. Briefly, the relational content more suitable for change of the scene in the processing source content (TV program) can be presented, and the relational content to be presented (before and after change of the part) does not change discontinuously.
Next, the content presentation apparatus of the fourth embodiment is explained. As to the same unit as the first-third embodiments, same sign is assigned, and its explanation is omitted.
(1) Component:
The relational content calculation unit 107 specifies a relational content of the present processing source content (selected by the source content selection unit 101). Furthermore, when the inter-content reference position control unit 111 detects update of a reference position, the relational content calculation unit 107 specifies a relational content of the reference position. Concretely, the relational content calculation unit 107 acquires metadata used when update of the reference position is detected, and metadata corresponding to candidate contents (of the relational content) from the content information database 102 via the content information acquisition unit 103. By using the metadata, the relational content calculation unit 107 calculates a relevance between the reference position and each of the candidate contents, and lists a candidate content in order of higher relevance from the candidate contents. The candidate content listed is a reference content specified in relation to the reference position. In the same way as the second embodiment, calculation of the relevance is executed by the equation (5).
When the inter-content reference position control unit 111 detects update of a reference position, the relational content calculation-history control unit 110 sets the reference position (before update) to a past reference position (previous reference position), and controls processing source history information of the update time and the past reference position. Concretely, the relational content calculation-history control unit 110 stores the processing source history information. The processing source history information represents the update time, metadata of the past reference position, a relational content (specified by the relational content calculation unit 107) of the past reference position, and the relevance between the relational content and the past reference position.
Based on the relevance of the relational content (present relational content) specified by the relational content calculation unit 107, and the relevance of the relational content (past relational content) represented by the processing source history information (controlled by the relational content calculation-history control unit 110), the layout unit 108 symbolically lays out the present processing source content, the present relational content and the past relational content in a virtual space. Then the layout unit 108 visualizes the virtual space, and presents the visualized virtual space via the presentation unit 109. A method for laying out the visualized content in the virtual space is same as that of the second embodiment.
(2) Operation:
Next, processing of the content presentation apparatus 100′″ of the fourth embodiment is explained. Steps of the content presentation processing are same as steps in
As mentioned-above, in the fourth embodiment, even if the processing source content does not change on the whole, when a part of the processing source content changes with passage of time, a relational content is dynamically presented based on change of the part. In this case, the relational content related to all of the present processing source content, and the relational content related to the part of the processing source content, are separately presented in different regions. Under this component, even if the part of the processing source content changes with passage of time, the relational content to be symbolically laid in the virtual space (before and after change of the part) can change not suddenly but continuously. Furthermore, relational contents are not presented in confusion, and the user can correctly understand that the relational content relates to all of the present processing source content or the part of the present processing source content.
In above-mentioned embodiments, the content presentation apparatus 100 is the view equipment or the recording equipment of the TV program. However, the content presentation apparatus 100 is not limited to these equipments. For example, the content presentation apparatus 100 may be a computer comprising one apparatus such as a personal computer or a micro computer. Furthermore, the content presentation apparatus 100 may be realized by a system which a plurality of computers is connected via a network.
In the disclosed embodiments, the processing can be performed by a computer program stored in a computer-readable medium.
In the embodiments, the computer readable medium may be, for example, a magnetic disk, a flexible disk, a hard disk, an optical disk (e.g., CD-ROM, CD-R, DVD), an optical magnetic disk (e.g., MD). However, any computer readable medium, which is configured to store a computer program for causing a computer to perform the processing described above, may be used.
Furthermore, based on an indication of the program installed from the memory device to the computer, OS (operation system) operating on the computer, or MW (middle ware software), such as database management software or network, may execute one part of each processing to realize the embodiments.
Furthermore, the memory device is not limited to a device independent from the computer. By downloading a program transmitted through a LAN or the Internet, a memory device in which the program is stored is included. Furthermore, the memory device is not limited to one. In the case that the processing of the embodiments is executed by a plurality of memory devices, a plurality of memory devices may be included in the memory device.
A computer may execute each processing stage of the embodiments according to the program stored in the memory device. The computer may be one apparatus such as a personal computer or a system in which a plurality of processing apparatuses are connected through a network. Furthermore, the computer is not limited to a personal computer. Those skilled in the art will appreciate that a computer includes a processing unit in an information processor, a microcomputer, and so on. In short, the equipment and the apparatus that can execute the functions in embodiments using the program are generally called the computer.
Other embodiments of the invention will be apparent to those skilled in the art from consideration of the specification and embodiments of the invention disclosed herein. It is intended that the specification and embodiments be considered as exemplary only, with the scope and spirit of the invention being indicated by the claims.
Number | Date | Country | Kind |
---|---|---|---|
P2009-049522 | Mar 2009 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
6263501 | Schein et al. | Jul 2001 | B1 |
6389424 | Kim et al. | May 2002 | B1 |
6466193 | Anai | Oct 2002 | B1 |
6597380 | Wang et al. | Jul 2003 | B1 |
6628313 | Minakuchi et al. | Sep 2003 | B1 |
6646980 | Yamamoto et al. | Nov 2003 | B1 |
6661426 | Jetha et al. | Dec 2003 | B1 |
6665668 | Sugaya et al. | Dec 2003 | B1 |
6674484 | Boland et al. | Jan 2004 | B1 |
6675174 | Bolle et al. | Jan 2004 | B1 |
6734877 | Honda et al. | May 2004 | B1 |
6745205 | Choi et al. | Jun 2004 | B2 |
6756972 | Komatsu et al. | Jun 2004 | B2 |
6816551 | Kasutani | Nov 2004 | B1 |
6853374 | Iwasaki | Feb 2005 | B2 |
6853389 | Ikeda | Feb 2005 | B1 |
6956812 | Okada et al. | Oct 2005 | B2 |
7171344 | Lind | Jan 2007 | B2 |
7213206 | Fogg | May 2007 | B2 |
7245677 | Pare, Jr. | Jul 2007 | B1 |
7281054 | Ishioka et al. | Oct 2007 | B2 |
7298384 | Anabuki et al. | Nov 2007 | B2 |
7333096 | Washio et al. | Feb 2008 | B2 |
7369142 | Kondo et al. | May 2008 | B2 |
7421455 | Hua et al. | Sep 2008 | B2 |
7511710 | Barrett | Mar 2009 | B2 |
7519121 | Matsushita et al. | Apr 2009 | B2 |
7519685 | Nonaka et al. | Apr 2009 | B2 |
7557774 | Baudisch et al. | Jul 2009 | B2 |
7557865 | Minami et al. | Jul 2009 | B2 |
7590948 | Narita et al. | Sep 2009 | B2 |
7607150 | Kobayashi et al. | Oct 2009 | B1 |
7805022 | Tanaka | Sep 2010 | B2 |
7840892 | Pyhalammi et al. | Nov 2010 | B2 |
7930634 | Nakamura | Apr 2011 | B2 |
7940285 | Would et al. | May 2011 | B2 |
8002633 | Shimizu | Aug 2011 | B2 |
8041155 | Mihara et al. | Oct 2011 | B2 |
8174523 | Higuchi et al. | May 2012 | B2 |
8244738 | Higuchi et al. | Aug 2012 | B2 |
20010007980 | Ishibashi et al. | Jul 2001 | A1 |
20010031129 | Tajima | Oct 2001 | A1 |
20010035855 | Komatsu et al. | Nov 2001 | A1 |
20020055924 | Liming | May 2002 | A1 |
20020057294 | Ejima et al. | May 2002 | A1 |
20020145603 | Iwasaki | Oct 2002 | A1 |
20030046694 | Istvan et al. | Mar 2003 | A1 |
20030103675 | Endo et al. | Jun 2003 | A1 |
20030122773 | Washio et al. | Jul 2003 | A1 |
20030142115 | Endo et al. | Jul 2003 | A1 |
20040172410 | Shimojima et al. | Sep 2004 | A1 |
20050010599 | Kake et al. | Jan 2005 | A1 |
20050055639 | Fogg | Mar 2005 | A1 |
20050076361 | Choi et al. | Apr 2005 | A1 |
20050097606 | Scott et al. | May 2005 | A1 |
20050138564 | Fogg | Jun 2005 | A1 |
20050193408 | Sull et al. | Sep 2005 | A1 |
20050210410 | Ohwa et al. | Sep 2005 | A1 |
20050212726 | Namba et al. | Sep 2005 | A1 |
20060064716 | Sull et al. | Mar 2006 | A1 |
20060086022 | Would et al. | Apr 2006 | A1 |
20060150215 | Wroblewski | Jul 2006 | A1 |
20070027855 | Kagawa | Feb 2007 | A1 |
20070106661 | Narita et al. | May 2007 | A1 |
20070106949 | Narita et al. | May 2007 | A1 |
20070107015 | Kazama et al. | May 2007 | A1 |
20070112940 | Morisawa et al. | May 2007 | A1 |
20070130318 | Roast | Jun 2007 | A1 |
20070136755 | Sakai | Jun 2007 | A1 |
20070252822 | Kim et al. | Nov 2007 | A1 |
20080074985 | Mukaiyama et al. | Mar 2008 | A1 |
20080136837 | Iguchi | Jun 2008 | A1 |
20080215548 | Ohashi et al. | Sep 2008 | A1 |
20080235729 | Doi et al. | Sep 2008 | A1 |
20080267582 | Yamauchi et al. | Oct 2008 | A1 |
20080316229 | Terayoko | Dec 2008 | A1 |
20080316518 | Watari et al. | Dec 2008 | A1 |
20090019031 | Krovitz et al. | Jan 2009 | A1 |
20090022480 | Yabe | Jan 2009 | A1 |
20090025030 | Kasutani | Jan 2009 | A1 |
20090080698 | Mihara et al. | Mar 2009 | A1 |
20090083814 | Sekine et al. | Mar 2009 | A1 |
20090138914 | Fukuda et al. | May 2009 | A1 |
20090228800 | Yasuda | Sep 2009 | A1 |
20090262139 | Tanaka et al. | Oct 2009 | A1 |
20100050098 | Turner | Feb 2010 | A1 |
20100054703 | Tanaka et al. | Mar 2010 | A1 |
20100057696 | Miyazawa et al. | Mar 2010 | A1 |
20100057722 | Nakamura et al. | Mar 2010 | A1 |
20100058173 | Kizuka et al. | Mar 2010 | A1 |
20100058213 | Higuchi et al. | Mar 2010 | A1 |
20100058388 | Baba et al. | Mar 2010 | A1 |
20100156893 | Mihara et al. | Jun 2010 | A1 |
20100229126 | Mihara et al. | Sep 2010 | A1 |
20100302595 | Yamada et al. | Dec 2010 | A1 |
20100333140 | Onodera et al. | Dec 2010 | A1 |
20110083552 | Vicars | Apr 2011 | A1 |
20110252447 | Suzuki et al. | Oct 2011 | A1 |
20120266192 | Crew et al. | Oct 2012 | A1 |
Number | Date | Country |
---|---|---|
1 051 034 | Nov 2000 | EP |
1 513 341 | Sep 2005 | EP |
1 785 814 | May 2007 | EP |
2 159 713 | Mar 2010 | EP |
2 159 722 | Mar 2010 | EP |
2 160 032 | Mar 2010 | EP |
2 166 752 | Mar 2010 | EP |
09-219835 | Aug 1997 | JP |
09-259130 | Oct 1997 | JP |
11-003346 | Jan 1999 | JP |
11-175534 | Jul 1999 | JP |
11-249774 | Sep 1999 | JP |
2000-076267 | Mar 2000 | JP |
2000-242648 | Sep 2000 | JP |
2000-305946 | Nov 2000 | JP |
2001-175380 | Jun 2001 | JP |
2001-186497 | Jul 2001 | JP |
2001-275048 | Oct 2001 | JP |
2001-285787 | Oct 2001 | JP |
2001-292383 | Oct 2001 | JP |
2001-326881 | Nov 2001 | JP |
2001-359006 | Dec 2001 | JP |
2002-044543 | Feb 2002 | JP |
2002-175269 | Jun 2002 | JP |
2002-175321 | Jun 2002 | JP |
2002-209178 | Jul 2002 | JP |
2002-325214 | Nov 2002 | JP |
2003-085884 | Mar 2003 | JP |
2003-167914 | Jun 2003 | JP |
2003-203035 | Jul 2003 | JP |
2003-330969 | Nov 2003 | JP |
2004-178085 | Jun 2004 | JP |
2004-178384 | Jun 2004 | JP |
2004-258390 | Sep 2004 | JP |
2005-010854 | Jan 2005 | JP |
2005-072826 | Mar 2005 | JP |
2005-080247 | Mar 2005 | JP |
2005-267390 | Sep 2005 | JP |
2005-303741 | Oct 2005 | JP |
2006-094520 | Apr 2006 | JP |
2006-514451 | Apr 2006 | JP |
2006-129122 | May 2006 | JP |
2006-260178 | Sep 2006 | JP |
2006-311163 | Nov 2006 | JP |
2006-352378 | Dec 2006 | JP |
2007-034663 | Feb 2007 | JP |
2007-049502 | Feb 2007 | JP |
2007-066285 | Mar 2007 | JP |
2007-122496 | May 2007 | JP |
2007-148669 | Jun 2007 | JP |
2007-149313 | Jun 2007 | JP |
2007-249641 | Sep 2007 | JP |
2007-272651 | Oct 2007 | JP |
2007-281945 | Oct 2007 | JP |
2007-311892 | Nov 2007 | JP |
2008-004985 | Jan 2008 | JP |
2008-077774 | Apr 2008 | JP |
2008-078785 | Apr 2008 | JP |
2008-129942 | Jun 2008 | JP |
2008-130032 | Jun 2008 | JP |
2008-134725 | Jun 2008 | JP |
2008-204193 | Sep 2008 | JP |
2008-242504 | Oct 2008 | JP |
2008-270872 | Nov 2008 | JP |
2008-312178 | Dec 2008 | JP |
2009-080580 | Apr 2009 | JP |
0033572 | Jun 2000 | WO |
2004061711 | Jul 2004 | WO |
2004091216 | Oct 2004 | WO |
Entry |
---|
Yamauchi et al., U.S. Appl. No. 12/049,455, filed Mar. 17, 2008. |
Mihara et al., U.S. Appl. No. 12/076,439, filed Mar. 18, 2008. |
Sekine et al., U.S. Appl. No. 12/232,186, filed Sep. 11, 2008. |
Tanaka et al., U.S. Appl. No. 12/416,786, filed Apr. 1, 2009. |
Kizuka et al., U.S. Appl. No. 12/418,474, filed Apr. 3, 2009. |
Baba et al., U.S. Appl. No. 12/426,756, filed Apr. 20, 2009. |
Miyazawa et al., U.S. Appl. No. 12/431,663, filed Apr. 28, 2009. |
Nakamura et al., U.S. Appl. No. 12/461,761, filed Aug. 24, 2009. |
Higuchi et al., U.S. Appl. No. 12/461,924, filed Aug. 27, 2009. |
Higuchi et al., U.S. Appl. No. 12/585,269, filed Sep. 9, 2009. |
Mihara et al., U.S. Appl. No. 12/585,458, filed Sep. 15, 2009. |
Shimizu et al., U.S. Appl. No. 13/110,279, filed May 18, 2011. |
Yoshiyuki Kokojima et al., “Resolution Independent Rendering of Deformable Vector Objects Using Graphics Hardware”, Sketches, Siggraph, (2006). |
Masaru Suzuki et al., “Sasatto Search” Human Interface Technology for Information Retrieval, Toshiba Review vol. 62, No. 12, 2007, pp. 54-57. |
Loviscach, “Motion Blur for Textures by Means of Anisotropic Filtering,” Eurographics Symposium on Rendering (2005), pp. 105-110. |
Takeuchi et al., “Motion sharpening in moving natural images,” Journal of Vision (2002), 2:377 (Abstract). |
Brostow et al., “Image-Based Motion Blur for Stop Motion Animation”, Proc. of Siggraph '01, pp. 561-566, (2001). |
Ramos, et al., “Fluid Interaction Techniques for the Control and Annotation of Digital Video”, Proceedings of the 16th annual ACM Symposium on User Interface Software and Technology, vol. 5, Issue 2, pp. 105-114, (2003). |
Irani et al., “Efficient Representations of Video Sequence and Their Applications”, Signal processing Image Communication (Journal), pp. 1-39, (1996). |
Teodosio et al., “Salient Stills”, ACM Transaction on Multimedia Computing, Communications and Appllications, pp. 16-36, (2005). |
Agarwala et al., “Interactive Digital Photomontage”, ACM Siggraph, pp. 1-9, (2004). |
Huan-Bo Luan et al., Interactive Spatio-Temporal Visual Map Model Web Video Retrieval, Multimedia and Expo, 2007, IEEE International Conference on, IEEE, P1, Jul. 1, 2007, pp. 560-563, ISBN: 978-1-4244-1016-3. (Abstract). |
Office Action dated Apr. 2, 2013 in Japanese Appl No. 2009-049522 with English-language translation. |
Shibata, M. et al., Proceedings of Information Science and Technology Forum, 7(3), Aug. 20, 2008, pp. 77-78, URL, http://ci.nii.ac.jp/naid/110007641466. |
Office Action dated Jun. 25, 2013 in JP Application No. 2009-049522. |
U.S. Appl. No. 12/585,269, filed Sep. 10, 2009. |
U.S. Appl. No. 12/426,756. |
U.S. Appl. No. 12/416,780. |
U.S. Appl. No. 12/418,474. |
U.S. Appl. No. 12/461,761, filed Aug. 24, 2009. |
U.S. Appl. No. 12/461,924, filed Aug. 27, 2009. |
Number | Date | Country | |
---|---|---|---|
20100229126 A1 | Sep 2010 | US |