1. Field of the Invention
The present invention relates to an information search apparatus, an information search method, and an information recording medium on which an information search program is computer-readably recorded, and more particularly to a song search apparatus and a song search method in which one song or a plurality of songs comprised of lyrics and performances (prelude, accompaniment, interlude, and postlude; hereinafter, ‘prelude, accompaniment, interlude, and postlude’ referred to as ‘performance’) are searched, and an information recording medium on which a song search program is computer-readably recorded.
2. Related Art
Recently, digital data for reproducing a plurality of songs are stored by an onboard navigation apparatus or a home server apparatus or the like, and favorite songs among them are selected and reproduced.
At this time, in prior arts, as a first song search method, it is common that part of constituent words (phrases) constituting lyrics included in a song to be reproduced is inputted as it is and the song having lyrics including the constituent words is searched and reproduced.
Further, US 2003 078919 A1 (FIGS. 1 to 3) discloses a second song search method reflecting subjective feeling of a user. There is a search method for preparing a sensitivity table which can be updated with a correlation value between a search keyword (for example, cheerful music or refreshing music) and a feature word (for example, cheerfulness or energetic), preparing each feature word list representing ‘1’ and ‘0’ depending on whether or not there is a feature related to the feature word, and searching a plurality of matched songs on the basis of the sensitivity table and the feature word list when the user inputs a desired search keyword.
However, according to the above-mentioned conventional song search methods, there are the following prevalent problems. First, in the first song search method, there is a problem in that different search routines are performed for search words which are closely related to each other such as ‘coast’, ‘waterside’, and ‘beach’, respectively, and that an entirely different songs are searched for each of the words constituting lyrics in some cases. In addition, the search process is not efficient.
On the other hand, according to the second song search method, there is a problem in that appropriate songs are not searched since the sensitivity table is used with respect to only a search keyword inputted by a user and thus the correlation to the song to be searched is not taken into consideration.
Accordingly, the present invention is made in consideration of the above-mentioned problems, and an object of the present invention is to provide a song search apparatus and a song search method in which songs matched with the subjective feeling of a user can be quickly searched, a program for searching song, and an information recording medium in which the program for searching song is recorded.
The present invention will be described below. Although reference numerals in the accompanying drawings will be accessorily written as parenthetic numerals for descriptive convenience, the present invention is not limited to the illustrated features.
The above object of the present invention can be achieved by an information search apparatus (S) which searches one or plural songs among a plurality of songs comprised of lyrics and performances. The information search apparatus (S) is provided with: a song feature information storing device (3) which distinguishably stores song feature information comprised of constituent word information which indicates contents of a plurality of constituent words included in the song, and first sound feature information which indicates the sound feature of the performance included in the song, for every song; a search word input device (9) which is used to input a search word representing the song to be searched; a search word feature information storing device (4) which distinguishably stores, as search word feature information characterizing an inputted search word, the search word feature information comprised of first lyric feature information, which is a collection of weighted constituent words obtained by applying a weight value to each constituent word constituting the lyrics included in any one of the songs to be searched by using the search word, and second sound feature information which indicates sound feature of the performance included in any one of the songs to be searched by using the search word, for every search word; a comparison device (8) which compares input search word feature information as the search word feature information corresponding to the inputted search word with each of the stored first sound feature information; and an extracting device (8) which extracts songs to have the best similarity to the input search word feature information, as the songs corresponding to the inputted search word, based on the comparison result of the comparison device (8).
The above object of the present invention can be achieved by an information search method which searches one or plural songs among a plurality of songs comprised of lyrics and performances. The information search method is provided with: a search word input process in which a search word indicating a song to be searched is inputted; a comparison process in which, as input search word feature information characterizing an inputted search word, the input search word feature information comprised of first lyric feature information, which is a collection of weighted constituent words obtained by applying a weight value to each of the constituent words constituting the lyrics included in any one of the songs to be searched by using the inputted search word, and second sound feature information, which indicates sound feature of the performance included in any one of the songs to be searched by using the inputted search word, is compared with a plurality of entities of the first sound feature information which indicates the sound feature of the song included in each of the songs; and an extracting process in which songs to have the best similarity to the input search word feature information are extracted, as the songs which are appropriate for the inputted search word, based on the comparison result of the comparison process.
The above object of the present invention can be achieved by an information recording medium on which an information search program is recorded so as to be readable through a computer which is included in an information search apparatus (S) which searches one or plural songs among a plurality of songs comprised of lyrics and performances with a pre-installed re-writable recording medium. The information search program causes the computer to function as: a song feature information storing device (3) which distinguishably stores song feature information comprised of constituent word information that represents contents of a plurality of constituent words included in the song and first sound feature information which indicates the sound feature of the performance included in the song, for every song; a search word input device (9) which is used to input a search word representing the song to be searched; a search word feature information storing device (4) which distinguishably stores, as search word feature information characterizing an inputted search word, the search word feature information comprised of first lyric feature information which is a collection of weighted constituent words obtained by applying a weight value to each constituent word constituting the lyrics included in any one of the songs to be searched by using a corresponding search word and second sound feature information indicating the sound feature of the performance included in any one of the songs to be searched by using the corresponding search word, for every search word; a comparison device (8) which compares input search word feature information as the search word feature information corresponding to the inputted search word with each entity of the stored first sound feature information; and an extracting device (8) which extracts the songs to have the best similarity to the input search word feature information, as the songs appropriate for the inputted search word, based on the comparison result of the comparison device (8).
Hereinafter, best embodiments of the present invention will now be described with reference to the drawings. Also, the below-mentioned embodiment is a case in which the present invention is applied to a song search apparatus for storing a plurality of songs and searching and outputting (reproducing) any one of songs according to a request of a user.
(I) Overall Configuration and Whole Operation
First, the overall configuration and the whole operation of the song search apparatus according to the present embodiment will be explained with reference to
As shown in
At this time, in the song database 2, a plurality of songs are accumulated and stored as an object to be searched by the below-mentioned song search process. Further, each of the songs is comprised of at least lyrics and a performance including prelude, accompaniment, interlude, and postlude.
Here, a song which is desired to be accumulated is inputted into the song database 2, when song information Ssg corresponding to the song is inputted into the song input unit 1 from the outside, by performing the format conversion process etc. for storing the song information Ssg in the song database 2 at the song input unit 1 to input the processed song information Ssg into the song database 2.
Next, the song feature information database 3 is accumulated with song feature information corresponding to all the songs accumulated in the song database 2 such that the song feature information can be identified with respect to each of the songs.
Here, the song feature information is accumulated in the song feature information database 3 such that it corresponds to each of all the songs accumulated in the song database 2, and is the information characterizing the respective lyrics and performance of each of the songs.
Next, the song feature information will be explained with reference to
When a new song is inputted into the song database 2 as song information Ssg, the song feature information is newly generated in accordance with the song and the generated song feature information is newly registered and accumulated in the song feature information database 3.
Here, when a new song is accumulated in the song database 2, the song information Ssg corresponding to the song is read from the song database 2 and is outputted to the sound feature information extracting unit 5 and the constituent word information extracting unit 6, as shown in
Further, a plurality of parameters representing the sound feature of the song are extracted from the song information Ssg and are outputted to the song feature information generating unit 7 as sound feature information Sav by the sound feature information extracting unit 5.
At this time, as the plurality of parameters included in the sound feature information Sav, there are, as shown in the right side of
In parallel to the above, the constituent word information extracting unit 6 extracts lyrics (words of the song) included in the song from the song information Ssg, searches whether or not the constituent words (phrases, hereinafter referred to simply as constituent words) set previously in the extracted lyrics are included, generates the constituent word feature information Swd for each of the constituent words indicating the search result (regarding whether or not the constituent words are included in the lyrics), and outputs them to the song feature information generating unit 7.
At this time, the constituent word feature information Swd shows whether or not the constituent words set previously such as ‘love’, ‘sea’, ‘thought’, or ‘hope’ are included in lyrics constituting a song. When the constituent word is included in the lyrics, the value of the constituent word feature information Swd for the constituent word is defined to be ‘1’, and, when the constituent word is not included in the lyrics, the value of the constituent word feature information Swd for the constituent word is defined to be ‘0’. More specifically, for example, as shown in the left side of
Thereby, the song feature information generating unit 7 combines the sound feature information Sav and the corresponding constituent word feature information Swd for every song, outputs the song feature information Ssp composed of a plurality of entities of song feature information 20 corresponding to each of the songs to the song feature information database 3, as shown in
In addition, the search word feature information corresponding to the whole search words previously set as search keywords (that is, the search keywords subjectively characterizing the song which a user wants to listen to at that time, hereinafter, referred to simply as the search words) inputted by the user in the below-mentioned song search process is distinguishably accumulated in the search word feature information database 4 for each of the search words.
Here, the search word feature information must be selected and inputted when a user searches songs from the songs accumulated in the song database 2, and it is the information characterizing each of the search words inputted by the user.
In the above embodiment, the constituent word information extracting unit 6 searches whether or not the constituent words set previously in the lyrics extracted from the song information Ssg, generates the constituent word feature information Swd for each of the constituent words, and outputs them to the song feature information generating unit 7. However, the constituent word feature information Swd can be generated in other modified ways as follows: More specifically, in the case that images of a constitution word can be changeable by its adjacent modifier, the meanings of the constitution word can be analyzed in consideration of the changed images, and the constituent word feature information Swd can be generated. In addition, on the basis of the normalized appearance numbers obtained by dividing the appearance numbers of the constitution word in one song with the length of the whole lyrics, the constituent word feature information Swd can be generated. Further, on the basis of a constitution word used in a title of a song, the constituent word feature information Swd can be generated. In this case, the constitution words of the song title can be assigned different weights to from ones to constitution words of lyrics in a song. Furthermore, by getting constitution words grouped by use of a thesaurus database (not shown), the constituent word feature information Swd can be generated on the basis of appearance numbers of each group to which a constitution word belongs. When lyrics in a song contain constitution words in different languages (such as English, French, German, Spanish, and Japanese), the constituent word feature information Swd can be generated on the basis of appearance numbers of each of the constitution words in the different language. Further, in the above-mentioned explanation, depending on whether or not the constituent words set previously are included in lyrics constituting a song, the value of the constituent word feature information Swd for the constituent word is defined to be ‘1’ or ‘0’, or binarized. However, the appearance numbers of the constituent words in a song are counted. When the counted appearance numbers are equal to or over a predetermined threshold, the value of the constituent ward feature information Swd for the constituent word can be defined to be ‘1’, and when the counted appearance numbers are less than the threshold, the value of the constituent word feature information Swd can be defined to be ‘0’. These modifications of generating the constituent word feature information can be performed independently each other and can be simultaneously performed together with all modifications. When all modifications are simultaneously performed, these modifications are performed in an order shown in
Next, the search word feature information will be explained with reference to
As shown in
Here, the sound feature information constituting the search word feature information 30 specifically includes sound parameters similar to those parameters included in the sound feature information Sav.
Also, similarly, the lyric feature information constituting the search word feature information 30 is a collection composed by applying a weight value to each of the plurality of subjective pieces of the lyric feature information characterizing the lyrics included in the song to be searched and extracted by using the search word feature information 30, according to the specific content of the respective lyrics included in the song to be searched and extracted by using the search word corresponding to the search word feature information 30.
More specifically, as shown in
Moreover, as mentioned below, in order to obtain the search result of favorite songs of a user, the same subjective concept as the subjective concept represented by each of the lyric feature information is represented by any one of the search words. Also, each of the lyric feature information itself represents the feature of the songs (more specifically, the lyrics included in the song) accumulated in the song database 2 but is independent from the search word itself. The detailed description thereof will be mentioned below.
Also, in case a user subjectively searches a desired song by using the information accumulated in the song feature information database 3 and the search word feature information database 4, after any search word is inputted at the input unit 9 by the user, input information Sin representing the inputted search word is outputted to the search process unit 8.
Thereby, the search process unit 8 extracts one entity of the search word feature information 30 corresponding to the inputted search word from the search word feature information database 4 as search word feature information Swds on the basis of the input information Sin, and simultaneously extracts a plurality of entities of the song feature information 20 corresponding to the entire songs accumulated in the song database 2 from the song feature information database 3 as song feature information Ssp. Then, the extracted entity of the search word feature information 30 is compared with those of the song feature information 20, and song identifying information Srz representing the songs corresponding to the entities of the song feature information 20 which have the best similarity to the entity of the search word feature information 30 are generated and outputted to the song database 2.
Thereby, the song database 2 outputs the song represented by the song identifying information Srz to the song output unit 10 as song information Ssg.
Then, the song output unit 10 performs an output interface process etc. needed for the output song information Ssg and outputs the processed song information Ssg to an external amplifier unit or a broadcasting unit or the like (not shown).
Also, after the song information Ssg representing one song is outputted from the song output unit 10, evaluation information regarding whether or not the song corresponding to the outputted song information Ssg is appropriate for a song required by a user who initially inputs a search word, is again inputted into the input unit 9, and the corresponding input information Sin is outputted to the search process unit 8.
Thereby, the search process unit 8 generates history information representing the result of the past song search process on the basis of the evaluation information inputted as the input information Sin, and temporarily stores history information Sm in the history storage unit 11 and reads out it if necessary to perform the below-mentioned history managing process.
(II) Song Search Process
Next, the song search process related to the embodiment performed by using the song search apparatus S comprising the above-mentioned configuration will be explained with reference to FIGS. 3 to 8 in detail. In addition,
As shown in
In addition, in parallel with the process of the steps S1 and S2, for every song, the constituent words constituting the lyrics included in the entire songs are read from the song feature information database 3 and then outputted to the search process unit 8 (step S3). Then, for every song, the search process unit 8 performs the process for converting the constituent words into the lyric feature information corresponding to the lyric included in each song by using the conversion table stored in the memory (not shown) in the search process unit 8 (step S4).
Here, the conversion table will be explained with reference to
The lyric feature information generated by the process of the step S4 is identical to the lyric feature information with the weight values included in the search word feature information 30, and is a collection composed by applying the weight value to each of the constituent words constituting the lyrics included in the song according to the specific content of each lyric feature information, for every song corresponding to the lyric feature information.
More specifically, in the example of the conversion table T shown in
In addition, in the process of the step S4, the corresponding lyric feature information 40 is generated from each constituent word of the song by using the conversion table T shown in
In parallel with the processes of the steps S1, S2, S3, and S4, only the sound feature information in each of the song feature information 20 corresponding to the entire songs is read from the song feature information database 3 and is outputted to the search process unit 8 (step S5).
Based on these, in the search process unit 8, for every song, the lyric feature information included in one entity of the search word feature information 30 (including the weight values of the lyric feature information 40) extracted at the step S2 is compared with the entity of the lyric feature information 40 corresponding to each song converted at the step S4, and the sound feature information included in the search word feature information 30 is compared with the sound feature information corresponding to each song extracted at the step S5, and then the similarity between the lyric feature information 40 and sound feature information corresponding to each song and those corresponding to the inputted search word is calculated for every song (step S6).
Further, based on the similarity calculated for each song, a reproduction list in which the songs to be outputted are arranged in higher order of the similarity is prepared (step S7), and the songs arranged in the order equal to that of the reproduction list are extracted from the song database 2 and are outputted to the song output unit 10 (step S8).
When one song is outputted, the user listening to the outputted song evaluates whether or not the outputted song is appropriate for the inputted search word at the step S1 and inputs the evaluation result by using the input unit 9 (step S9).
Further, when the outputted song is appropriate for the search word (step S9; matching), the below-mentioned matching history information is updated (step S11), and then the process proceeds to the step S12. On the other hand, when the outputted song is inappropriate for the search word at the evaluating step S9 (step S9: non-matching), the below-mentioned non-matching history information is updated (step S10), and then the process proceeds to the step S12.
Here, the non-matching history information updated at the above-mentioned step S10 and the matching history information updated at the above-mentioned step S11 will be explained with reference to
First, as the matching history information G, as shown in
On the other hand, as the non-matching history information NG, as shown in
Moreover, when the updating of the history information is finished with respect to the predetermined number of songs, based on the result, the content of the above-mentioned conversion table T and the content of the search word feature information 30 are updated (step S12).
Next, whether the output of the final song in the reproduction list prepared at the above-mentioned step S7 is finished is determined (step S13). When the output of the final song is not finished (step S13: NO), the process returns to the above-mentioned step S8. At the step S8, the next song in the reproduction list is outputted and then the steps S9 to S12 are repeated with respect to the next song. On the other hand, in the determination of the step 13, when the output of the final song is finished (step S13: YES), the song search process is completed.
Next, the update process of the conversion table T at the above-mentioned step S12 will be explained with reference to FIGS. 6 to 8.
In the update process of the conversion table T, as shown in
On the other hand, if the subjective feeling is matched (step S20: YES), then the process proceeds to the update process of the actual conversion table T.
In addition, in the below-mentioned update process, with respect to the predetermined number of the songs (forty songs (twenty appropriate songs and twenty inappropriate songs) in
In the actual update process, first, for every constituent word of the song which is evaluated to be appropriate for the matching history information (one storage address corresponding to one song, in
Next, similarly to the above, with respect to the non-matching history information, for every constituent word of the songs which are evaluated to be inappropriate, all ‘0’s and ‘1’s in the vertical direction of
Here, the larger the difference between the average value AA and the average value DA for every constituent word, the higher the probability of representing the lyric feature information 40 corresponding to the current search word by the constituent word.
However, note that the average values AA and DA calculated by the above-mentioned process are not the result obtained by evaluating for the entire songs stored in the song database 2. Accordingly, the confidence interval of each calculated average value (statistically, referred to as a sample ratio) must be obtained and the difference between the average value AA and the average value DA corresponding to each constituent word, (the absolute value of average value AA−average value DA), must be determined whether it can be used as the weight value of the history information of the constituent word. More specifically, assuming the confidence level of the confidence interval to be 90%, the difference is determined whether it can be used as the weight value of the history information of the constituent word by using the confidence interval calculated by the below-mentioned equation (1). That is, the confidence interval is calculated by the following equation for each constituent word:
Confidence interval=2×1.65×[(AA×(1−AA))/N]1/2 (1),
where N is a number of songs.
Next, when the confidence interval is calculated, the absolute value of the value obtained by subtracting the average value DA from the average value AA is determined whether it is larger than the calculated confidence interval (step S23).
Further, if the absolute value of the value obtained by subtracting the average value DA from the average value AA is larger than the calculated confidence interval (step S23: YES), the difference becomes a reliable value and is employed as the weight value of the corresponding lyric feature information (the lyric feature information of ‘heart-warming’ in
In addition, in the initial state of the song search apparatus S, since the history information includes the initial values set in advance and the number of songs which becomes the target of the history information is limited, as a result, new history information is stored by overwriting the old history information. Thereby, as the evaluation of the outputted songs is made, the subjective feeling of the user is reflected on the weight values of the conversion table T. This shows that the conversion table T can be updated accordingly in case the subjective feeling changes.
Next, when the evaluation for twenty songs (ten songs having the matching history information+(plus) ten songs having the non-matching history information) from the state in
For example, for the constituent word ‘sea’ shown in
Further, similarly, since the constituent word of ‘love’ is stored in the matching history information and the non-matching history information without any bias, the difference between the average value AA and the average value DA for the constituent word ‘love’ becomes smaller than the above-mentioned confidence interval, and thus the constituent word ‘love’ becomes the constituent word which is not related to the lyric feature information ‘heart-warming’.
As explained above, according to the operation of the song search apparatus S of the present embodiment, since songs are searched by comparing the entity of the search word feature information 30 corresponding to an inputted search word and the entity of the song feature information 20 corresponding to the stored songs, appropriate songs can be certainly searched by the inputted search word, and songs matched to the subjective feeling of a user can be much easily searched, compared with the case of searching songs using only an inputted search word.
Also, since at least the lyric feature information 40 corresponding to the subjective feeling identical to the subjective feeling represented by an inputted search word is generated, songs are searched by using the lyric feature information 40 reflecting the subjective feeling of a user, and thus songs suitable for the subjective feeling of the user can be easily searched.
Further, since the lyric feature information 40 is generated by using the conversion table T, the suitable lyric feature information 40 can be generated by an arrangement of a simple apparatus.
Moreover, since songs corresponding to an inputted search word are extracted by comparing the lyric feature information 40 which constitutes the search word feature information 30 corresponding to the inputted search word with the lyric feature information 40 corresponding to the song feature information 20 and simultaneously by comparing the sound feature information which constitutes the search word feature information 30 corresponding to the inputted search word with the sound feature information which constitutes the song feature information 20, songs which are appropriate for the search word can be easily searched.
In addition, since the conversion table T is updated based on the evaluation of a user, the certainty of the conversion table T based on the evaluation of the user becomes higher, and thus songs according to the subjective feeling of the user can be easily searched.
Further, since each entity of the search word feature information 30 is updated based on the evaluation of a user, the certainty of the search word feature information 30 based on the evaluation of the user becomes higher, and thus songs according to the subjective feeling of the user can be easily searched.
Further, since the conversion table T or the search word feature information 30 is updated based on an inputted evaluation data by using the history information of each song, the conversion table T or the search word feature information 30 is updated according to the subjective feeling of a user by reflecting the past history thereof, and thus songs according to the subjective feeling of the user can be easily searched.
Also, since, for every constituent word, the weight value of the constituent word is updated by using the difference between the average value AA and the average value DA and also since the conversion table T is updated, the certainty of the conversion table T becomes higher, and thus songs according to the subjective feeling of a user can be easily searched.
In addition, although the case of applying the present invention to the song search apparatus S for accumulating and searching a plurality of songs in the above-mentioned embodiment was explained, the present invention may be applied to the image search apparatus which searches still images or moving images accumulated in accordance with the subjective feeling of a user.
Also, by recording a program corresponding to the flow chart shown in
The invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein.
The entire disclosure of Japanese Patent Application No. 2003-385714 filed on Nov. 14, 2003 and Japanese Patent Application No. 2004-327334 filed on Nov. 11, 2004 including the specification, claims, drawings and summary is incorporated herein by reference in its entirety.
Number | Date | Country | Kind |
---|---|---|---|
P2003-385714 | Nov 2003 | JP | national |
P2004-327334 | Nov 2004 | JP | national |