The present invention relates to a non-transitory computer-readable recording medium storing an information presentation program and the like.
It is desired to find useful knowledge from data by machine learning. In a conventional technique, since it is difficult to perform perfect machine learning, a plurality of training models is generated and presented to a user.
In
As illustrated in
Therefore, the conventional technique takes measures by listing the top K training models in descending order of the objective function from the collection of all training models.
Examples of the related art include: [Non-Patent Document 1] Satoshi Hara, Takanori Maehara “Enumerate Lasso Solutions for Feature Selection” AAAI-17; and [Non-Patent Document 2] Satoshi Hara, Masakazu Ishihata “Approximate and Exact Enumeration of Rule Models” AAAI-18.
According to an aspect of the embodiments, there is provided a non-transitory computer-readable recording medium storing an information presentation program for causing a computer to perform processing including: performing a training processing that generates a plurality of training models by executing machine learning that uses training data; and performing a generation processing that generates hierarchical information that represents, in a hierarchical structure, a relationship between hypotheses shared as common and the hypotheses regarded as differences for a plurality of the hypotheses extracted from each of the plurality of training models and each designated by a combination of one or more explanatory variables.
The object and advantages of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the claims.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are not restrictive of the invention.
However, when there is a bias in knowledge included in the top K training models, the completeness of knowledge is lowered if the top K training models are listed as in the conventional technique.
As described with reference to
In addition, as described with reference to
That is, it is desired to easily compare complicated training models with each other.
In one aspect, an object of the present invention is to provide an information presentation program, an information presentation method, and an information presentation device capable of easily comparing complicated training models with each other.
Hereinafter, embodiments of an information presentation program, an information presentation method, and an information presentation device disclosed in the present application will be described in detail with reference to the drawings. Note that these embodiments do not limit the present invention.
An example of processing of the information presentation device according to the present embodiment will be described.
The information presentation device acquires a hypothesis set from the training model M. The hypothesis set of the training model M will serve as information that explains an output result of the training model M. In the example illustrated in
A hypothesis set H4 of the training model M4 includes hypotheses hy1, hy2, hy8, and hy9. A hypothesis set Hn of the training model Mn includes hypotheses hy1, hy2, hy8, hy10, hy11, and hy12. Description of hypothesis sets of other training models M will be omitted.
The information presentation device executes similarity determination based on the hypothesis sets of the training models M and classifies the training models M into families of similar training models M. In the example illustrated in
The information presentation device compares the hypothesis sets H1 to H3 of the training models M1 to M3 belonging to the first group and specifies the hypotheses hy1, hy2, hy3, and hy4 shared as common. The information presentation device compares the hypothesis sets H4, Hn, and others of the training models M4, Mn, and others belonging to the second group and specifies the hypotheses hy1, hy2, and hy8 shared as common.
The information presentation device compares the “hypotheses hy1, hy2, hy3, and hy4” shared as common to the first group with the “hypotheses hy1, hy2, and hy8” shared as common to the second group to specify the “hypotheses hy1 and hy2” shared as common to the first and second groups.
The information presentation device generates hierarchical information in which common hypothesis sets Hc1, Hc2-1, and Hc2-2 and unique hypothesis sets Hc3-1, Hc3-2, Hc3-3, Hc3-4, and Hc3-n are coupled, based on the execution result of the above.
The common hypothesis set Hc1 includes “hypotheses hy1 and hy2” shared as common to all training models M. The common hypothesis set Hc2-1 is a hypothesis set shared as common to the training models M1 to M3 belonging to the first group and includes “hypotheses hy3 and hy4” obtained by excluding the hypotheses of the common hypothesis set Hc1. The common hypothesis set Hc2-2 is a hypothesis set shared as common to the training models M4, Mn, and others belonging to the second group and includes the “hypothesis hy8” obtained by excluding the hypotheses of the common hypothesis set Hc1.
The unique hypothesis set Hc3-1 includes the “hypothesis hy5” unique to the training model M1 obtained by excluding the hypotheses of the common hypothesis sets Hc1 and Hc2-1 from the hypothesis set H1 included in the training model M1. The unique hypothesis set Hc3-2 includes the “hypothesis hy6” unique to the training model M2 obtained by excluding the hypotheses of the common hypothesis sets Hc1 and Hc2-1 from the hypothesis set H2 included in the training model M2. The unique hypothesis set Hc3-3 includes the “hypothesis hy7” unique to the training model M3 obtained by excluding the hypotheses of the common hypothesis sets Hc1 and Hc2-1 from the hypothesis set H3 included in the training model M3.
The unique hypothesis set Hc3-4 includes the “hypothesis hy9” unique to the training model M4 obtained by excluding the hypotheses of the common hypothesis sets Hc1 and Hc2-2 from the hypothesis set H4 included in the training model M4. The unique hypothesis set Hc3-n includes the “hypotheses hy10, hy11, and hy12” unique to the training model Mn obtained by excluding the hypotheses of the common hypothesis sets Hc1 and Hc2-2 from the hypothesis set Hn included in the training model Mn.
As described with reference to
Subsequently, processing in which the information presentation device according to the present embodiment determines similarity based on the hypothesis sets of the training models M will be described.
In
The hypothesis hy1-1 is a hypothesis constituted by a combination of the attributes “winning the election once”, “having a relative as a politician”, “policy_ABC bill”, and “ranking rate_less than 0.8” and has a weight of “−0.95”. The hypothesis hy1-2 is a hypothesis constituted by a combination of the attributes “, rookie (denial of rookie)”, “having a relative as a politician”, “policy_ABC bill”, and “ranking rate_less than 0.8” and has a weight of “−0.96”. The hypothesis hy1-3 is a hypothesis constituted by a combination of the attributes “incumbent”, “having a relative as a politician”, “policy_ABC bill”, and “ranking rate_less than 0.8” and has a weight of “−0.85”. The attribute is an example of the explanatory variable.
Comparing each attribute of the hypothesis hy1-3 with each attribute of the hypothesis hy1-1, the attribute “incumbent” of the hypothesis hy1-3 includes the attribute “winning the election once” of the hypothesis hy1-1. Since the other attributes coincide with each other between the hypotheses hy1-1 and hy1-3, the hypothesis hy1-3 is a hypothesis including the hypothesis hy1-1.
Comparing each attribute of the hypothesis hy1-3 with each attribute of the hypothesis hy1-2, the attribute “incumbent” of the hypothesis hy1-3 includes the attribute “, rookie” of the hypothesis hy1-2. Since the other attributes coincide with each other between the hypotheses hy1-2 and hy1-3, the hypothesis hy1-3 is a hypothesis including the hypothesis hy1-2.
The hypothesis hy2-1 is a hypothesis constituted by the attribute “, rookie” and has a weight of “0.69”. The hypothesis hy2-2 is a hypothesis constituted by the attribute “policy_ABC bill” and has a weight of “0.81”. The hypothesis hy2-3 is a hypothesis constituted by the attribute “winning the election once” and has a weight of “0.82”. The hypothesis hy2-4 is a hypothesis constituted by the attribute “ranking rate_less than 0.8” and has a weight of “−0.94”.
The hypothesis sets H1-1 and H2-1 illustrated in
The hypotheses hy2-1′, hy2-2′, hy2-3′, and hy2-4′ are included in the hypothesis hy1-1. The hypotheses hy2-1′, hy2-2′, hy2-3′, and hy2-4′ are also included in the hypothesis hy1-2. In addition, it is assumed that the hypotheses hy1-1 and hy1-2 have an inclusion relationship with each other.
The information presentation device adds the weights of the hypotheses hy2-1′, hy2-2′, hy2-3′, and hy2-4′ (the weights are zero) to the weight of the hypothesis hy1-1 as a destination of inclusion. In addition, since the hypotheses hy1-1 and hy1-2 have an inclusion relationship with each other, the information presentation device updates the weight of the hypothesis hy1-1 to “−1.93” by adding the weight of the hypothesis hy1-2 to the weight of the hypothesis hy1-1.
The information presentation device adds the weights of the hypotheses hy2-1′, hy2-2′, hy2-3′, and hy2-4′ (the weights are zero) to the weight of the hypothesis hy1-2 as a destination of inclusion. In addition, since the hypotheses hy1-1 and hy1-2 have an inclusion relationship with each other, the information presentation device updates the weight of the hypothesis hy1-2 to “−1.93” by adding the weight of the hypothesis hy1-1 to the weight of the hypothesis hy1-2.
Since the hypotheses hy1-1 and hy1-2 are in an inclusion relationship with each other, the information presentation device updates the weight of the hypothesis hy1-3 to “−2.78” by adding the weight of the hypothesis hy1-1 or the weight of the hypothesis hy1-2 to the weight of the hypothesis hy1-3 as a destination of inclusion.
The information presentation device executes the processing in
The hypotheses hy2-1, hy2-2, hy2-3, and hy2-4 are included in the hypothesis hy1-1. The hypotheses hy2-1, hy2-2, hy2-3, and hy2-4 are also included in the hypothesis hy1-2. In addition, the hypotheses hy1-1 and hy1-2 have an inclusion relationship with each other.
Since the hypotheses hy1-1′ and hy1-2′ have an inclusion relationship with each other, the information presentation device adds the weight (initial value 0) of the hypothesis hy1-2′ to the hypothesis hy1-1′. In addition, the information presentation device updates the weight of the hypothesis hy1-1′ to “1.39” by adding the weights (initial values) of the hypotheses hy2-1, hy2-2, hy2-3, and hy2-4 to the weight of the hypothesis hy1-1′ as a destination of inclusion.
Since the hypotheses hy1-1′ and hy1-2′ have an inclusion relationship with each other, the information presentation device adds the weight (initial value 0) of the hypothesis hy1-1′ to the hypothesis hy1-2′. In addition, the information presentation device updates the weight of the hypothesis hy1-2′ to “1.39” by adding the weights (initial values) of the hypotheses hy2-1, hy2-2, hy2-3, and hy2-4 to the weight of the hypothesis hy1-2′ as a destination of inclusion.
Since the hypotheses hy1-1′ and hy1-2′ are in an inclusion relationship with each other, the information presentation device updates the weight of the hypothesis hy1-3′ to “1.39” by adding the weight of the hypothesis hy1-1′ or the weight of the hypothesis hy1-2′ to the weight of the hypothesis hy1-3′ as a destination of inclusion.
The information presentation device executes the processing in
Next, an example of a configuration of the information presentation device according to the present embodiment will be described.
The communication unit 110 is coupled to an external device or the like in a wired or wireless manner and transmits and receives information to and from the external device or the like. For example, the communication unit 110 is implemented by a network interface card (NIC) or the like. The communication unit 110 may be coupled to a network (not illustrated).
The input unit 120 is an input device that inputs various types of information to the information presentation device 100. The input unit 120 corresponds to a keyboard, a mouse, a touch panel, or the like.
The display unit 130 is a display device that displays information output from the control unit 150. The display unit 130 corresponds to a liquid crystal display, an organic electro luminescence (EL) display, a touch panel, or the like.
The storage unit 140 includes training data 141, a training model table 142, a hypothesis database 143, a common hypothesis set table 144, and hierarchical information 145. The storage unit 140 corresponds to a semiconductor memory element such as a random access memory (RAM), a read only memory (ROM), or a flash memory, or a storage device such as a hard disk drive (HDD).
The training data 141 is data in which a hypothesis and a label corresponding to this hypothesis are associated with each other.
The training model table 142 is a table that holds the plurality of training models M. The training model is generated by a training unit 151. Description of the data structure of the training model table 142 will be omitted.
The hypothesis database 143 is a table that holds the hypothesis sets extracted from the training models M.
The common hypothesis set table 144 is a table that holds the hypothesis sets shared as common, among the hypothesis sets of the respective training models.
The hierarchical information 145 indicates information obtained by hierarchically coupling the common hypothesis set indicating hypotheses shared as common and the unique hypothesis set indicating hypotheses regarded as differences in the hypothesis sets of the training models M. For example, the hierarchical information 145 corresponds to the common hypothesis sets Hc1, Hc2-1, and Hc2-2 and the unique hypothesis sets Hc3-1 to Hc3-n described with reference to
The control unit 150 includes the training unit 151, a classification unit 152, and a generation unit 153. The control unit 150 can be implemented by a central processing unit (CPU), a micro processing unit (MPU), or the like. In addition, the control unit 150 can also be implemented by a hard wired logic such as an application specific integrated circuit (ASIC) or a field programmable gate array (FPGA).
The training unit 151 generates the training model M by executing machine learning based on the training data 141. When executing machine learning, the training unit 151 generates a plurality of training models M by altering parameters, random seeds, preprocessing, and the like of the training models M. The training unit 151 registers the plurality of generated training models M in the training model table 142.
For example, the training unit 151 may execute machine learning based on a technique described in Patent Document (Japanese Laid-open Patent Publication No. 2020-46888) or the like, or may execute machine learning using another conventional technique. The training model M generated by machine learning includes a hypothesis set for explaining an output result of this training model M, and weights are set individually in each hypothesis. Note that the training unit 151 may generate different training models M by further using a plurality of pieces of training data (not illustrated).
The classification unit 152 classifies the plurality of training models M into a plurality of groups according to the similarity. It is assumed that training models belonging to the same group are similar to each other. The classification unit 152 outputs the classification result for the training models M to the generation unit 153. Hereinafter, an example of processing of the classification unit 152 will be described. For example, the classification unit 152 executes processing of generating the hypothesis database, processing of specifying the similarity between the training models, and processing of classifying the training models.
Processing in which the classification unit 152 generates the hypothesis database 143 will be described. The classification unit 152 extracts the hypothesis set of the training model M and a weight included in this hypothesis set from the training model M registered in the training model table 142 and registers the extracted hypothesis set and weight in the hypothesis database 143. When registering the hypothesis set and the weight in the hypothesis database 143, the classification unit 152 associates the hypothesis set and the weight with the identification information on the training model M. The classification unit 152 repeatedly executes the above processing for each training model M.
Processing in which the classification unit 152 specifies the similarity between the training models will be described. The processing in which the classification unit 152 specifies the similarity corresponds to the processing described above with reference to
The classification unit 152 compares the hypothesis set of the training model M1 with the hypothesis set of the training model M2, based on the hypothesis database 143. For convenience, the hypothesis set of the training model M1 will be referred to as a first hypothesis set, and the hypothesis set of the training model M2 will be referred to as a second hypothesis set.
The classification unit 152 adds, to the first hypothesis set, a hypothesis that exists in the second hypothesis set but does not exist in the first hypothesis set. The classification unit 152 adds, to the second hypothesis set, a hypothesis that exists in the first hypothesis set but does not exist in the second hypothesis set. By executing such processing, the classification unit 152 aligns the granularity of the hypotheses of the first hypothesis set with the granularity of the hypotheses of the second hypothesis set.
The classification unit 152 determines the inclusion relationship between the hypotheses for the first hypothesis set and the second hypothesis set after aligning the granularity of the hypotheses. The classification unit 152 may determine the inclusion relationship in any manner and, for example, determines the inclusion relationship of each hypothesis based on a table defining the inclusion relationships regarding each attribute. In such a table, information such as “winning the election once” and “, rookie” being included in “incumbent” is defined.
The classification unit 152 allocates a weight to each hypothesis for the first hypothesis set and the second hypothesis set, by calculating the cumulative value of weights set in hypotheses, based on the inclusion relationships between the hypotheses. The processing in which the classification unit 152 calculates the cumulative values to calculates the weights and allocates the weights to each hypothesis corresponds to the processing described with reference to
The classification unit 152 specifies a first vector in which each hypothesis of the first hypothesis set is assigned as one dimension and the value of each dimension is assigned by the cumulative value of one of the hypotheses. The classification unit 152 specifies a second vector in which each hypothesis of the second hypothesis set is assigned as one dimension and the value of each dimension is assigned by the cumulative value of one of the hypotheses. The classification unit 152 specifies the distance between the first vector and the second vector as the similarity.
The classification unit 152 specifies the similarity between the respective training models by repeatedly executing the above processing for all the combinations of the training models M.
Processing in which the classification unit 152 classifies the training models will be described. The training unit 151 specifies the similarity between the training models by executing the above processing and classifies training models having similarity equal to or higher than a threshold value into the same group. For example, when the similarity between the training models M1 and M2 is equal to or higher than the threshold value and the similarity between the training models M2 and M3 is equal to or higher than the threshold value, the classification unit 152 classifies the training models M1, M2, and M3 into the same group. The classification unit 152 classifies the plurality of training models into a plurality of groups by executing the above processing and outputs the classification result to the generation unit 153.
Here, it is assumed that the hypothesis added to each hypothesis set by the classification unit 152 in order to align the granularity of the hypotheses is used only when the classification unit 152 generates a vector and will not be used by the generation unit 153 to be described below.
By executing the processing described with reference to
As described with reference to
Here, an example of a processing procedure in which the generation unit 153 specifies the hypothesis set shared as common will be described.
The generation unit 153 of the information presentation device 100 acquires the hypothesis set Hn of the training model Mn from the hypothesis database 143 (step S10). The generation unit 153 acquires a list of the training models M from the hypothesis database 143 (step S11).
The generation unit 153 acquires a hypothesis set Hi of an undetermined training model M in the list of the training models M (step S12). The generation unit 153 excludes a hypothesis inconsistent between the hypothesis sets Hi and Hn (step S13). Here, the hypothesis sets Hi and Hn from which inconsistent hypotheses have been excluded will be referred to as hypothesis sets Hi′ and Hn′, respectively.
The generation unit 153 generates the hypothesis set Hcommon shared as common to the hypothesis sets Hi′ and Hn′ (step S14). The generation unit 153 registers information on the training models having the hypothesis set Hcommon in the common hypothesis set table 144 and records a relationship between the training models corresponding to the hypothesis set Hcommon (step S15).
When the processing has not been executed on all the training models M included in the list (step S16, No), the generation unit 153 proceeds to step S12. When the processing has been executed on all the training models M included in the list (step S16, Yes), the generation unit 153 ends the processing.
Here, an example of the processing of excluding a hypothesis inconsistent between the hypothesis sets described in step S13 in
In
It is assumed that the hypothesis set Hn of the training model Mn includes hypotheses {Hn,1, Hn,2, Hn,3, Hn,4, Hn,5}. Each hypothesis is assumed as indicated below. Each of A, B, C, D, E, and F in the hypotheses is an example of an attribute (explanatory variable).
Hn,1: A→True
Hn,2: B∧F→True
Hn,3: C→True
Hn,4: D→False
Hn,5: E→True
It is assumed that the hypothesis set H1 of the training model M1 includes hypotheses {H1,1, H1,2, H1,3, H1,4, H1,5}. Each hypothesis is assumed as indicated below. Each of A, B, C, D, E, and F in the hypotheses is an example of an attribute (explanatory variable).
H1,1: A→True
H1,2: B→True
H1,3: C∧D→True
H1,4: E→False
When executing the above inconsistency determination, the generation unit 153 determines that Hn,4 of the hypothesis set Hn and H1,3 of the hypothesis set H1 are inconsistent. In addition, the generation unit 153 determines that Hn,5 of the hypothesis set Hn and H1,4 of the hypothesis set H1 are inconsistent.
The generation unit 153 generates a hypothesis set Hn′ by excluding inconsistent Hn,4, and Hn,5 from the hypothesis set Hn, based on the result of the inconsistency determination. The generation unit 153 generates a hypothesis set H1′ by excluding inconsistent H1,4, and H1,5 from the hypothesis set H1, based on the result of the inconsistency determination.
Subsequently, an example of processing of generating a hypothesis shared as common between the hypothesis sets described in step S14 in
In
It is assumed that the hypothesis set Hn′ of the training model Mn includes hypotheses {Hn,1, Hn,2, Hn,3}. It is assumed that the hypothesis set H1′ of the training model M1 includes hypotheses {H1,1, H1,2, H1,3}.
Since the hypothesis Hn,1 of the hypothesis set Hn′ and the hypothesis H1,1 of the hypothesis set H1′ coincide with each other, the generation unit 153 generates a common hypothesis “Hc,1: A→True”.
Description will be made of the common hypothesis generation for the hypothesis Hn,2 of the hypothesis set Hn′ and the hypothesis H1,2 of the hypothesis set H1′ by the generation unit 153. In the generation unit 153, the condition part “B A F” of the hypothesis Hn,2 and the condition part “B” of the hypothesis H1,2 are in the inclusion relationship “B∧F⊃B∨B∧F⊂B”. Therefore, the generation unit 153 generates the common portion “Cc=(B)∧(B∧F)”=“Cc=B∧B∧F”=“Cc=B∧F” of the condition parts. The generation unit 153 generates the common portion “True” of the conclusion parts. By the above processing, the generation unit 153 generates the common hypothesis “B∧F→True” for the hypotheses Hn,2 and H1,2.
By executing the above processing, the generation unit 153 generates the hypothesis set Hcommon shared as common between the hypothesis set Hn′ of the training model Mn and the hypothesis set H1′ of the training model M1. For example, the hypothesis set Hcommon shared as common includes hypotheses {Hc,1, Hc,2}. Each of the hypotheses is assumed as indicated below.
Hc,1: A→True
Hc,2: B∧F→True
The generation unit 153 records a relationship between the training models corresponding to the hypothesis set Hcommon shared as common, based on the result of the processing performed in
Meanwhile, when a weight is set in a hypothesis included in the hypothesis set, the generation unit 153 updates the conclusion part of the hypothesis in consideration of a weight of a hypothesis in an inclusion relationship.
Hn,1: A→True (weight: 0.2)
Hn,2: B∧F→True (weight: 0.3)
Hn,3: C→True (weight: 0.4)
Hn,4: D→False (weight: −0.3)
Hn,5: E→True (weight: 0.2)
Here, the hypothesis Hn,3 is assumed to be included in the hypotheses Hn,4 and Hn,5. In these circumstances, the generation unit 153 updates the weight of the hypothesis Hn,4 to “0.1” by adding the weight “0.4” of the hypothesis Hn,3 to the weight “−0.3” of the hypothesis Hn,4 as a destination of inclusion. In addition, since the weight of the hypothesis Hn,4 has changed from a negative value to a positive value, the conclusion part of the hypothesis Hn,4 is updated to “True”.
The generation unit 153 updates the weight of the hypothesis Hn,5 to “0.6” by adding the weight “0.4” of the hypothesis Hn,3 to the weight “0.2” of the hypothesis Hn,5 as a destination of inclusion. In addition, since the weight of the hypothesis Hn,5 has not changed from a positive value, the conclusion part of the hypothesis Hn,5 is left as “True”.
By executing the above processing, the generation unit 153 repeatedly executes processing of specifying a hypothesis set shared as common to the hypothesis sets of the respective training models belonging to the same group. Similarly, the generation unit 153 specifies a hypothesis set shared as common to the hypothesis sets of the respective groups, based on the hypothesis sets of the respective groups. By executing such processing, the generation unit 153 specifies, for example, the common hypothesis sets Hc1, Hc2-1, and Hc2-2 and the unique hypothesis sets Hc3-1, Hc3-2, Hc3-3, Hc3-4, and Hc3-n described with reference to
Next, a processing procedure of the information presentation device 100 according to the present embodiment will be described.
The classification unit 152 of the information presentation device 100 extracts hypothesis sets and weights of hypotheses from the training models M in the training model table 142 and registers the extracted hypothesis sets and weights in the hypothesis database 143 (step S102). The classification unit 152 executes a similarity calculation process (step S103).
The classification unit 152 classifies the training models into a plurality of groups, based on the similarity between the respective training models M (step S104). The generation unit 153 of the information presentation device 100 executes a common hypothesis specifying process (step S105).
The generation unit 153 generates the hierarchical information 145, based on the result of the common hypothesis specifying process (step S106). The generation unit 153 outputs the hierarchical information 145 to the display unit 130 (step S107).
Next, an example of a processing procedure of the similarity calculation process indicated in step S103 in
As illustrated in
The classification unit 152 determines an inclusion relationship between the listed condition parts of the hypotheses (step S203). The classification unit 152 calculates the cumulative value of weights of each hypothesis and specifies the vector for each training model M (step S204).
The classification unit 152 calculates the similarity, based on the vectors of the respective training models M (step S205).
Note that the processing procedure of the common hypothesis specifying process illustrated in step S105 in
Next, an example of scatter diagrams regarding the cumulative values of weights calculated by the classification unit in
In
Next, effects of the information presentation device 100 according to the present embodiment will be described. The information presentation device 100 generates a plurality of training models M by executing machine learning that uses the training data 141. The information presentation device 100 generates the hierarchical information 145 that represents, in a hierarchical structure, a relationship between hypotheses shared as common and hypotheses regarded as differences for a plurality of hypotheses extracted from each of the plurality of training models and each designated by a combination of one or more explanatory variables. By referring to such hierarchical information 145, the user may be allowed to see the commonality and difference of the hypotheses of the plurality of training models M, from the plurality of training models M, and may easily compare the complicated training models with each other.
The information presentation device 100 specifies a common hypothesis shared as common and a difference hypothesis regarded as a difference between the hypothesis set of one training model to be compared and the hypothesis set of another training model to be compared, and generates the hierarchical information 145 by arranging the common hypothesis in an upper layer of the difference hypothesis. The common hypothesis corresponds to the common hypothesis set in
The information presentation device 100 specifies similarity between the training models, based on the hypothesis sets extracted from the training models M, and classifies the plurality of training models into a plurality of groups, based on the specified similarity. The information presentation device 100 specifies the common hypothesis and the difference hypothesis, based on the classification result. This may enable to specify the common hypothesis and the difference hypothesis based on the hypothesis sets of similar training models.
The information presentation device 100 aligns the granularity of the hypotheses of the hypothesis sets of the respective training models M to be compared and specifies the similarity between the respective training models M to be compared, based on the cumulative values of the hypothesis sets. This may enable to specify the similarity between the respective training models M even if the hypotheses of the training models to be compared do not completely correspond to each other.
Note that the processing procedure of the similarity calculation process executed by the classification unit 152 is not limited to the processing procedure in
The classification unit 152 calculates an overlap ratio between the listed hypotheses (step S303). In the processing in step S303, the classification unit 152 may calculate the overlap ratio by excluding a hypothesis added to make the granularity match.
The classification unit 152 determines an inclusion relationship between the listed condition parts of the hypotheses (step S304). The classification unit 152 calculates the cumulative value of weights of each hypothesis and corrects the cumulative value by multiplying the cumulative value by the overlap ratio for each training model M (step S305).
The classification unit 152 specifies the vector of each training model according to the corrected cumulative values (step S306). The classification unit 152 calculates the similarity, based on the vectors of the respective training models M (step S307).
As described with reference to
In addition, the classification unit 152 of the information presentation device 100 described above calculates the vectors of the training models by aligning the granularity of the hypothesis sets of the training models M to be compared, but is not limited to this. For example, the classification unit 152 may compare the hypothesis sets of the training models M to be compared to specify conjunction hypotheses and calculate the vectors using only the specified hypotheses to specify the similarity between the training models M. This allows the processing of aligning the granularity of the hypotheses to be skipped and thus may enable to specify the similar training models M while simplifying the processing.
Next, an example of a hardware configuration of a computer that implements functions similar to the functions of the information presentation device 100 indicated in the above embodiments will be described.
As illustrated in
The hard disk device 207 includes a training program 207a, a classification program 207b, and a generation program 207c. In addition, the CPU 201 reads each of the programs 207a to 207c and loads the read programs 207a to 207c into the RAM 206.
The training program 207a functions as a training process 206a. The classification program 207b functions as a classification process 206b. The generation program 207c functions as a generation process 206c.
Processing of the training process 206a corresponds to the processing of the training unit 151. Processing of the classification process 206b corresponds to the processing of the classification unit 152. Processing of the generation process 206c corresponds to the processing of the generation unit 153.
Note that each of the programs 207a to 207c does not necessarily have to be previously stored in the hard disk device 207. For example, each of the programs is stored in a “portable physical medium” to be inserted into the computer 200, such as a flexible disk (FD), a compact disc read only memory (CD-ROM), a digital versatile disc (DVD), a magneto-optical disk, or an integrated circuit (IC) card. Then, the computer 200 may read and execute each of the programs 207a to 207c.
All examples and conditional language provided herein are intended for the pedagogical purposes of aiding the reader in understanding the invention and the concepts contributed by the inventor to further the art, and are not to be construed as limitations to such specifically recited examples and conditions, nor does the organization of such examples in the specification relate to a showing of the superiority and inferiority of the invention. Although one or more embodiments of the present invention have been described in detail, it should be understood that the various changes, substitutions, and alterations could be made hereto without departing from the spirit and scope of the invention.
This application is a continuation application of International Application PCT/JP2021/013860 filed on Mar. 31, 2021 and designated the U.S., the entire contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2021/013860 | Mar 2021 | US |
Child | 18468565 | US |