The present invention relates to a technology for analyzing audio signals.
A score alignment technique for estimating a position in a music piece at which sound is actually being generated (hereinafter referred to as “sound generation position”) by means of analyzing an audio signal that represents the sound that is generated by the performance of the music piece has been proposed in the prior art. For example, Japanese Laid-Open Patent Application No. 2015-79183 discloses a configuration for calculating the likelihood (observation likelihood) that each time point in a music piece corresponds to the actual sound generation position by means of analyzing an audio signal, to thereby calculate the posterior probability of the sound generation position by means of updating the likelihood using a hidden semi-Markov model (HSMM).
It should be noted in passing that, in practice, it is difficult to completely eliminate the possibility of the occurrence of an erroneous estimation of the sound generation position. Thus, in order, for example, to predict the occurrence of an erroneous estimation and carry out appropriate countermeasures in advance, it is important to quantitatively evaluate the validity of a probability distribution of the posterior probability.
In consideration of such circumstances, an object of a preferred aspect of the present disclosure is to appropriately evaluate the validity of the probability distribution relating to the sound generation position.
In order to solve the problem described above, in an audio analysis method according to a preferred aspect of this disclosure, a sound generation probability distribution which is a distribution of probabilities that sound representing an audio signal is generated at each position in a music piece, is calculated from the audio signal, a sound generation position of the sound in the music piece is estimated from the sound generation probability distribution, and an index of validity of the sound generation probability distribution is calculated from the sound generation probability distribution.
An audio analysis device according to a preferred aspect of this disclosure comprises a distribution calculation module that calculates a sound generation probability distribution which is a distribution of probabilities that sound representing an audio signal is generated at each position in a music piece, from the audio signal; a position estimation module that estimates a sound generation position of the sound in the music piece from the sound generation probability distribution; and an index calculation module that calculates an index of validity of the sound generation probability distribution from the sound generation probability distribution.
Selected embodiments will now be explained with reference to the drawings. It will be apparent to those skilled in the field of musical performances from this disclosure that the following descriptions of the embodiments are provided for illustration only and not for the purpose of limiting the invention as defined by the appended claims and their equivalents.
As shown in
The performance device 12 executes an automatic performance of a target music piece under the control of the audio analysis device 10. From among the plurality of parts that constitute the target music piece, the performance device 12 according to the first embodiment executes an automatic performance of parts other than the parts performed by the performer P. For example, a main melody part of the target music piece is performed by the performer P, and the automatic performance of an accompaniment part of the target music piece is executed by the performance device 12.
As shown in
The sound collection device 14 generates an audio signal A by collecting, sounds generated by the performance by the performer P (for example, instrument sounds or singing sounds). The audio signal A represents the waveform of the sound. Moreover, an audio signal A that is output from an electric musical instrument, such as an electric string instrument, can also be used. Therefore, the sound collection device 14 can be omitted. The audio signal A can also be generated by adding signals that are generated by a plurality of the sound collection devices 14. The display device 16 (for example, a liquid-crystal display panel) displays various images under the control of the audio analysis device 10.
As shown in
The storage device 24 of the first embodiment stores music data M. The music data M is in the form of ten SMF (Standard MIDI File) file conforming to the MIDI (Musical Instrument Digital Interface) standard, which designates the performance content of the target music piece. As shown in
The reference data MA designates performance content of part of the target music piece to be performed by the performer P (for example, a sequence of notes that constitute the main melody part of the target music piece). The performance data MB designates performance content of part of the target music piece that is automatically performed by the performance device 12 (for example, a sequence of notes that constitute the accompaniment part of the target music piece). Each of the reference data MA and the performance data MB is time-series data, in which are arranged, in a time series, instruction data designating performance content (sound generation/mute) and time data designating the generation time point of said instruction data. The instruction data assigns pitch (note number) and intensity (velocity), and provide instruction for various events, such as sound generation and muting. The time data, on the other hand, designates, for example, an interval for successive instruction data.
The electronic controller 22 has a plurality of functions for realizing the automatic performance of the target music piece (audio analysis module 32; performance control module 34; and evaluation processing module 36) by the execution of a program that is stored in the storage device 24. Moreover, a configuration in which the functions of the electronic controller 22 are realized by a group of a plurality of devices (that is, a system) or a configuration in which some or all of the functions of the electronic controller 22 are realized by a dedicated electronic circuit can also be employed. In addition, a server device, which is located away from the space in which the performance device 12 and the sound collection device 14 are installed, such as a music hall, can realize some or all of the functions of the electronic controller 22.
As shown in
Specifically, the distribution calculation module 42 of the first embodiment crosschecks the audio signal A of each unit segment and the reference data MA of the target, music piece to thereby calculate the likelihood (observation likelihood) that the sound generation position of the unit segment corresponds to each position t in the target music piece. Then, under the condition that the unit segment of the audio signal A has been observed, the distribution calculation module 42 calculates, as the sound generation probability distribution D, the probability distribution of the posterior probability (posterior distribution) that the time point of the sound generation of said unit segment was the position t in the target music piece, from the likelihood for each position t. Known statistical processing, such as Bayesian estimation using a hidden semi-Markov model (HSMM) can be suitably used for calculating the sound generation probability distribution D that uses the observation likelihood, as disclosed in, for example, Patent Document 1.
The position estimation module 44 estimates the sound generation position Y of the sound represented by the unit segment of the audio signal A in the target music piece from the sound generation probability distribution D calculated by the distribution calculation module 42. Known statistical processing estimation methods, such as MAP (Maximum A Posteriori) estimation, can be freely used to estimate the sound generation position Y using the sound generation probability distribution D. The estimation of the sound generation position Y by the position estimation module 44 is repeated for each unit segment of the audio signal A. That is, for each of a plurality of unit segments of the audio signal A, one of a plurality of positions t of the target music piece is specified as the sound generation position Y.
The performance control module 34 of
The performance device 12 executes the automatic performance of the target music piece in accordance with the instructions from the performance control module 34. Since the sound generation position Y moves with time toward the end of the target music piece as the performance of the performer P progresses, the automatic performance of the target music piece by the performance device 12 will also progress with the movement of the sound generation position Y. That is, the automatic performance of the target music piece by the performance device 12 is executed at the same tempo as that of the performance of the performer P. As can be understood from the foregoing explanation, in order to synchronize the automatic performance with the performance of the performer P, the performance control module 34 provides instruction to the performance device 12 for carrying out the automatic performance in accordance with the content specified by the performance data MB while maintaining the intensity of each note and the musical expressions, such as phrase expressions, of the target music piece. Thus, for example, if performance data MB that represents the performance of a specific performer, such as a performer of the past who is no longer alive, are used, it is possible to create an atmosphere as if the performer were cooperatively and synchronously playing together with a plurality of actual performers P, while accurately reproducing musical expressions that are unique to said performer by means of the automatic performance.
Moreover, in practice, time on the order of several hundred milliseconds is required for the performance device 12 to actually generate a sound (for example, for the hammer of the sound generation mechanism 124 to strike a string), after the performance control module 34 provides instruction to the performance device 12 to carry out the automatic performance by means of an output of instruction data in the performance data MB. That is, the actual generation of sound by the performance device 12 can be delayed with respect to the instruction from the performance control module 34. Therefore, the performance control module 34 can also provide instruction to the performance device 12 regarding the performance at a (future) point in time that is subsequent to the sound generation position Y in the target music piece estimated by the audio analysis module 32.
The evaluation processing module 36 of
Based on the tendency described above, the index calculation module 52 calculates the index Q in accordance with the shape of the sound generation probability distribution D. The index calculation module 52 of the first embodiment calculates the index Q in accordance with the degree of dispersion d at the peak of the sound generation probability distribution D. Specifically, the index calculation module 52 calculates as the index Q the variance of one peak that is present in the sound generation probability distribution D (hereinafter referred to as “selected peak”). Thus, the validity of the sound generation probability distribution D can be evaluated as increasing as the index Q becomes smaller (that is, the selected peak becomes sharper). If, as shown in
The validity determination module 54 of
The operation control module 56 controls the operation of the automatic performance system 100 in accordance with the determination result of the validity determination module 54 (presence/absence of validity of the sound generation probability distribution D). When the validity determination module 54 determines that the sound generation probability distribution D is not valid, the operation control module 56 of the first embodiment notifies the user to that effect. Specifically, the operation control module 56 causes the display device 16 to display a message indicating that the sound generation probability distribution D is not valid. The message can be a character string, such as “the estimation accuracy of the performance position has decreased,” or the message can report the decline in the estimation accuracy by means of a color change. By visually checking the display of the display device 16, the user can ascertain that the automatic performance system 100 is not able to estimate the sound generation position Y with sufficient accuracy. In the foregoing description, the determination result by the validity determination module 54 is visually reported to the user by means of an image display, but it is also possible to audibly notify the user of the determination result by means of sound, for example. For instance, the operation control module 56 reproduces sound from a sound-emitting device, such as a loudspeaker or an earphone. The sound can be an announcement, such as “the estimation accuracy of the performance position has decreased,” or can be an alarm.
The index calculation module 52 calculates the index Q of the validity of the sound generation probability distribution D calculated by the distribution calculation module 42 (S4). Specifically, the degree of dispersion d of the selected peak of the sound generation probability distribution D is calculated as the index Q. The validity determination module 54 determines the presence/absence of validity of the sound generation probability distribution D based on the index Q (S5). Specifically, the validity determination module 54 determines whether the index Q is lower than the threshold value QTH.
If the index Q exceeds the threshold value QTH (Q>QTH), that the sound generation probability distribution D is not valid can be tested. If the validity determination module 54 determines that the sound generation probability distribution D is not valid (S5: NO), the operation control module 56 notifies the user that the sound generation probability distribution D is not valid (S6). On the other hand, if the index Q is below the threshold value QTH (Q<QTH), it can be determined whether the sound generation probability distribution D is valid. If the validity determination module 54 determines that the sound generation probability distribution D is valid (S5: YES), the operation (S6) to report the sound generation probability distribution D as not valid is not executed. However, if the validity determination module 54 determines that the sound generation probability distribution D is valid, the operation control module 56 can notify the user to that effect.
As described above, in the first embodiment, the index Q of the validity of the sound generation probability distribution D is calculated from the sound generation probability distribution D. Thus, it is possible to quantitatively evaluate the validity of the sound generation probability distribution D (and, thus, the validity of the sound generation position Y that can be estimated front the sound generation probability distribution D). In the first embodiment, the index Q is calculated in accordance with the degree of dispersion d (for example, variance) at the peak of the sound generation probability distribution D. Accordingly, it is possible to calculate the index Q, which can highly accurately test the validity of the sound generation probability distribution D, based on the tendency that the validity (statistical reliability) of the sound generation probability distribution D increases as the degree of dispersion d of the peak of the sound generation probability distribution D becomes smaller.
In addition, in the first embodiment, the user is notified of the determination result that the sound generation probability distribution is not valid. The user might therefore respond by changing the automatic control that utilizes the estimation result of the sound generation position Y to manual control.
A second embodiment now be described. In each of the embodiments illustrated below, elements that have the same actions or functions as in the first embodiment have been assigned the same reference symbols as those used to describe the first embodiment, and detailed descriptions thereof have been appropriately omitted.
In the automatic performance system 100 according to the second embodiment, the method with which the index calculation module 52 calculates the index Q of the validity of the sound generation probability distribution D differs from the first embodiment. The operations and configurations other than those of the index calculation module 52 are the same as in the first embodiment.
Specifically, the index calculation module 52 calculates as the index Q the difference δ between the highest peak (that is, the maximum peak) and the second highest peak from an array of local maximum values of a plurality of peaks of the sound generation probability distribution D sorted in descending order. However, the method for calculating the index Q in the second embodiment is not limited to the example described above. For example, the differences δ between the local maximum values of the maximum peak and each of the remaining plurality of peaks in the sound generation probability distribution D can be calculated, and a representative value (for example, the mean) of the plurality of differences δ can be calculated as the index Q.
As described above, in the second embodiment, it is assumed that the validity of the sound generation probability distribution tends to increase as the index Q becomes larger. In light of this tendency, the validity determination module 54 of the second embodiment determines the presence/absence of validity of the sound generation probability distribution D in accordance with the result of comparing the index Q with the threshold value QTH. The validity determination module 54 can compare the index Q with the threshold value QTH. Specifically, the validity determination module 54 determines that the sound generation probability distribution D is valid when the index Q exceeds the threshold value QTH (S5: YES) and determines that the sound generation probability distribution D is not valid when the index Q is below the threshold value QTH (S5: NO). The validity determination module 54 can determine that the sound generation probability distribution is valid when the index Q is equal to or higher than the threshold value QTH and determine that the sound generation probability distribution D is not valid when the index Q is lower than the threshold value QTH. The validity determination module 54 can determine that the sound generation probability distribution D is valid when the index Q is higher than the threshold value QTH and determine that the sound generation probability distribution D is not valid when the index Q is equal to or lower than the threshold value QTH. The other operations are the same as in the first embodiment.
In the second embodiment as well, since the index Q of the validity of the sound generation probability distribution D is calculated from the sound generation probability distribution D, there is the advantage that it is possible to quantitatively evaluate the validity of the sound generation probability distribution D (and, thus, the validity of the sound generation position Y that can be estimated from the sound generation probability distribution D) in the same manner as in the first embodiment. In addition, in the second embodiment the index Q is calculated in accordance with the differences δ between the local maximum values of the peaks of the sound generation probability distribution D. Accordingly, based on the tendency for the validity of the sound generation probability distribution D to increase as the local maximum value of a specific peak of the sound generation probability distribution D becomes greater than the local maximum values of the other peaks (that is, the difference δ is larger), it is possible to calculate the index Q that can evaluate the validity of the sound generation probability distribution D with great accuracy.
If the validity determination module 54 determines that the sound generation probability distribution D is not valid (S5: NO), the operation control module 56 cancels the control in which the performance control module 34 synchronizes the automatic performance of the performance device 12 with the progression of the sound generation position Y (S10). For example, the performance control module 34 can set the tempo of the automatic performance of the performance device 12 to a tempo that is unrelated to the progression of the sound generation position Y in accordance with an instruction from the operation control module 56. For example, the performance control module 34 can control the performance device 12 so that the automatic performance is executed at the tempo immediately before it was determined by the validity determination module 54 that the sound generation probability distribution D is not valid, or at a standard tempo designated by the music data M (S3). If, on the other hand, the validity determination module 54 determines that the sound generation probability distribution D is valid (S5: YES), the operation control module 56 causes the performance control module 34 to continue the control to synchronize the automatic performance with the progression of the sound generation position Y (S11). Accordingly, the performance control module 34 controls the performance device 12 such that the automatic performance is synchronized with the progression of the sound generation position Y (S3).
The same effects as those of the first embodiment or the second embodiment are also achieved in the third embodiment. In the third embodiment as well, if the validity determination module 54 determines that the sound generation probability distribution D is not valid, the control to synchronize the automatic performance with the progression of the sound generation position Y is canceled. Thus, the possibility that the sound generation position Y estimated from the sound generation probability distribution D with a low validity (for example, an erroneously estimated sound generation position Y) will be reflected in the automatic performance can be reduced.
The embodiment illustrated above can be variously modified. Specific modified embodiments are illustrated below. Two or more embodiments arbitrarily selected from the following examples can be appropriately combined as long as they are not mutually contradictory.
(1) In the first embodiment, the degree of dispersion d (for example, variance) of the peak of the sound generation probability distribution D is calculated as the index Q, but the method for calculating the index Q based on the degree of dispersion d is not limited to the this particular example. For instance, the index Q can also be found by means of a prescribed calculation that uses the degree of dispersion d. As can be understood from the foregoing example, calculating the index Q in accordance with the degree of dispersion d at the peak of the sound generation probability distribution D includes, in addition to the configuration in which the degree of dispersion d is calculated as the index Q (Q=d), configuration in which the index Q that differs from the degree of dispersion d (Q≠d) can be calculated in accordance with said degree of dispersion d.
(2) in the second embodiment, the differences δ between the local maximum values of the peaks of the sound generation probability distribution D is calculated as the index Q, but the method for calculating the index Q in accordance with the difference δ is not limited to the foregoing example. For example, it is also possible to calculate the index Q by means of a prescribed calculation that uses the difference δ. As can be understood from the foregoing example, calculating the index Q in accordance with the differences δ between the local maximum values of the peaks of the sound generation probability distribution D includes, in addition to the configuration in which the difference δ is calculated as the index Q (Q=δ), a configuration in which the index Q that is different from the difference δ (Q≠δ) is calculated in accordance with said difference δ.
(3) In the embodiments described above, the presence/absence of the validity of the sound generation probability distribution D is determined based on the index Q, but the determination of the presence/absence of the validity of the sound generation probability distribution D can be omitted. For example, the determination of the presence/absence of validity of the sound generation probability distribution D is not necessary in a configuration in which the index Q calculated by the index calculation module 52 is reported to the user by means of an image display or by outputting sound, or in a configuration in which the time series of the index Q is stored in the storage device 24 as a history. As can be understood from the foregoing example, the validity determination module 54 exemplified in each of the above-described embodiments and the operation control module 56 can be omitted from the audio analysis device 10.
(4) In the embodiments described above, the distribution calculation module 42 calculates the sound generation probability distribution D over the entire segment of the target music piece, but the distribution calculation module 42 can also calculate the sound generation probability distribution D over a partial segment of the target music piece. For example, the distribution calculation module 42 calculates the sound generation probability distribution D with respect to a partial segment of the target music piece located in the vicinity of the sound generation position Y estimated for the immediately preceding unit segment (that is, the probability distribution at each position tin said segment).
(5) In the embodiments described above, the sound generation position Y estimated by the position estimation module 44 is used by the performance control module 34 to control the automatic performance, but the use of the sound generation position Y is not limited in this way. For example, it is possible to play the target music piece by supplying music data representing sounds of the performance of the target music piece to a sound-emitting device (for example, a loudspeaker or an earphone) so as to be synchronized with the progression of the sound generation position Y. In addition, it is possible to calculate the tempo of the performance of the performer P from the temporal change of the sound generation position Y, and to evaluate the performance from the calculation result (for example, to determine the presence/absence of a change in tempo). As can be understood from the foregoing example, the performance control module 34 can be omitted from the audio analysis device 10.
(6) As exemplified in the above-described embodiments, the audio analysis device 10 is realized by cooperation between the electronic controller 22 and the program. The program according to a preferred aspect of causes a computer to function as the distribution calculation module 42 for calculating the sound generation probability distribution D, which is a distribution of probabilities that sound representing the audio signal A is generated at each position t in the target music piece from the audio signal A; as the position estimation module 44 for estimating the sound generation position Y of the sound in the target music piece from the sound generation probability distribution D; and as the index calculation module 52 for calculating the index Q of the validity of the sound generation probability distribution D from the sound generation probability distribution D. The program exemplified above can be stored on a computer-readable storage medium and installed in the computer.
The storage medium is, for example, a non-transitory (non-transitory) storage medium, a good example of which is an optical storage medium, such as a CD-ROM, but can include other known arbitrary storage medium formats, such as semiconductor storage media and magnetic storage media. “Non-transitory storage media” include any computer-readable storage medium that excludes transitory propagation signals (transitory propagation signal) and does not exclude volatile storage media. Furthermore, it is also possible to deliver the program to a computer in the form of distribution via a communication network.
(7) For example, the following configurations can be understood from the embodiments exemplified above.
In an audio analysis method according to a preferred aspect (first aspect), a computer system calculates from an audio signal a sound generation probability distribution, which is a distribution of probabilities that sound representing the audio signal is generated at each position in a music piece, estimates the sound generation position of the sound in the music piece from the sound generation probability distribution, and calculates an index of the validity of the sound generation probability distribution from the sound generation probability distribution. In the first aspect, the index of the validity of the sound generation probability distribution is calculated from the sound generation probability distribution. Thus, it is possible to quantitatively evaluate the validity of the sound generation probability distribution (and, hence, the validity of the result of estimating the sound generation position from the sound generation probability distribution).
In a preferred example (second aspect) of the first aspect, when calculating the index, the index is calculated in accordance with the degree of dispersion at a peak of the sound generation probability distribution. It is assumed that the validity (statistical reliability) of the sound generation probability distribution tends to increase as the degree of dispersion (for example, variance) of the peak of the sound generation probability distribution decreases. If this tendency is assumed, by means of the second aspect in which the index is calculated in accordance with the degree of dispersion of the peak of the sound generation probability distribution, it is possible to calculate the index that can evaluate the validity of the sound generation probability distribution with high accuracy. For example, in a configuration in which the degree of dispersion of the peaks of the sound generation probability distribution is calculated as the index, the sound generation probability distribution can be evaluated as being valid when the index is below the threshold value (for example, when the variance is small), and the sound generation probability distribution can be evaluated as not being valid when the index is higher than the threshold value (for example, when the variance is large).
In a preferred example (third aspect) of the first aspect, when calculating the index, the index is calculated in accordance with the difference between a local maximum value of maximum peaks of the sound generation probability distribution and the local maximum value of another peak. It is assumed that the validity (statistical reliability) of the sound generation probability distribution tends to increase as the local maximum value of a specific peak of the sound generation probability distribution increases with respect to the local maximum value of the other peak. If the tendency described above is assumed, by means of the third aspect, in which the index is calculated in accordance with the difference between the local maximum value at the maximum peak and the local maximum value at the other peak, it is possible to calculate the index that can test the validity of the sound generation probability distribution with high accuracy. For example, in a configuration in which the difference between the local maximum value at the maximum peak and the local maximum value at another peak is calculated as the index, it is possible to determine that the sound generation probability distribution is valid when the index is greater than the threshold value and that the sound generation probability distribution is not valid when the index is below the threshold value.
In a preferred example (fourth aspect) of any one of the first aspect to the third aspect, the computer system further determines the presence/absence of the validity of the sound generation probability distribution based cats the index. By means of the fourth aspect, it is possible to objectively determine the presence/absence of the validity of the sound generation probability distribution.
In a preferred example (fifth aspect) of the fourth aspect, the computer system further notifies a user when it is determined that the sound generation probability distribution is not valid. In the fifth aspect, the user is notified when it is determined that the sound generation probability distribution is not valid. The user might therefore respond by changing the automatic control that utilizes the estimation result of the sound generation position to manual control
hi a preferred example (sixth aspect) of the fourth aspect, the computer system further executes the automatic performance of the music piece so that the automatic performance is synchronized with the progression of the estimated sound generation position, and when it is determined that the sound generation probability distribution is not valid, the computer system cancels the control to synchronize the automatic performance with the progression of the sound generation position. In the sixth aspect, when it is determined that the sound generation probability distribution is not valid, the control to synchronize the automatic performance with the progression of the sound generation position is canceled. Accordingly, it is possible prevent a sound generation position estimated from a sound generation probability distribution of low validity (for example, an erroneously estimated sound generation position) from being reflected in the automatic performance.
An audio analysis device according to a preferred aspect (seventh aspect) comprises a distribution calculation module that calculates from an audio signal a sound generation probability distribution, which is a distribution of probabilities that sound representing the audio signal is generated at each position in a music piece; a position estimation module that estimates the sound generation position of the sound in the music piece from the sound generation probability distribution; and an index calculation module that calculates an index of the validity of the sound generation probability distribution from the sound generation probability distribution. In the seventh aspect, the index of the validity of the sound generation probability distribution, is calculated from the sound generation probability distribution. Accordingly, it is possible to quantitatively evaluate the validity of the sound generation probability distribution (and, thus, the validity of the result of estimating the sound generation position from the sound generation probability distribution).
The present embodiments are useful because it is possible to appropriately evaluate the validity of the probability distribution relating to the sound generation position.
Number | Date | Country | Kind |
---|---|---|---|
2016-216886 | Nov 2016 | JP | national |
This application is a continuation application of International Application No. PCT/JP2017/040143, filed on Nov. 7, 2017, which claims priority to Japanese Patent Application No. 2016-216886 filed in Japan on Nov. 7, 2016. The entire disclosures of International Application No. PCT/JP2017/040143 and Japanese Patent Application No. 2016-216886 are hereby incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5913259 | Grubb et al. | Jun 1999 | A |
9069065 | Coley | Jun 2015 | B1 |
20100170382 | Kobayashi | Jul 2010 | A1 |
20110214554 | Nakadai et al. | Sep 2011 | A1 |
20130231761 | Eronen | Sep 2013 | A1 |
Number | Date | Country |
---|---|---|
2001-117580 | Apr 2001 | JP |
2007-241181 | Sep 2007 | JP |
2011-180590 | Sep 2011 | JP |
2012-168538 | Sep 2012 | JP |
2015-079183 | Apr 2015 | JP |
Entry |
---|
International Search Report in PCT/JP2017/040143 dated Jan. 30, 2018. |
Translation of Office Action in the corresponding Japanese Patent Application No. 2016-216886, dated Sep. 2, 2020. |
Number | Date | Country | |
---|---|---|---|
20190251940 A1 | Aug 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2017/040143 | Nov 2017 | US |
Child | 16393592 | US |