The present invention relates to a feature-converting device and the like that convert features.
A learning algorithm is a basic method in various devices, for example, as seen in an action determination device disclosed in PTL 1.
The action determination device disclosed in PTL 1 estimates an action of a user having a moving body by assigning an error-reduced state to a trajectory of the moving body. On the basis of information in which trajectory information regarding the trajectory is associated with action information regarding the action, the action determination device estimates a relationship between the trajectory information and the action information. In this case, the action determination device selects a specific feature from among features constituting the trajectory information and estimates (predicts) a relationship between the specific feature and the action information.
In other words, on the basis of learning information in which explanatory variables (for example, the above-mentioned trajectory information) are associated with a target variable (for example, the above-mentioned action information), a learning algorithm computes a relationship between the explanatory variables and the target variable. The learning algorithm applies the computed relationship to predictive information, thereby estimating a value of the target variable regarding the predictive information. When the learning algorithm estimates the value regarding the predictive information, explanatory variables representing the predictive information are the same as the explanatory variables in the learning information.
PTL 1: Japanese Unexamined Patent Application Publication No. 2009-157770
In general, in predictive analysis, a predictive model (a relationship between explanatory variables and a target variable) having high classification accuracy cannot be constructed only by explanatory variables prepared by an analyst.
It is effective to perform feature selection while performing feature construction for converting the given explanatory variables instead of using the prepared explanatory variables as it is in order to generate a predictive model having high classification accuracy.
However, generally, feature selection and feature construction are those involving an extremely large amount of computations. For example, when processing for taking logarithms of given features or arithmetic processing for combining a plurality of features are performed, an enormous number of features are constructed and thus all the features are needed to be evaluated.
For example, when the number of input features is assumed to be N. (2×N) features are constructed after each feature is processed by squaring and taking a logarithm. Additionally, features with an order of (3×N)3 are constructed after processing for choosing any three features from a feature set that includes the above constructed features and the original input features, and then multiplying the chosen three features.
Thus, it is a main object of the present invention to provide a feature-converting device and the like that can provide good features quickly.
As an aspect of the present invention, a feature-converting device including:
first feature construction means for receiving one or more first features representing features including one or more elements composed of a numeral or a code, and constructing, by applying one or more unary functions that compute at least one of the features on the basis of at least one of the features to the received first features, one or more second features representing results of applying the unary operation functions to the first features;
first feature selection means for computing relevance between (i) the one or more second features and the one or more first features and (ii) a target variable that includes one or more elements composed of a numeral or a code associated with one or more of the elements included in the first features, and selecting one or more third features representing highly relevant features from among the one or more second features and the one or more first features;
second feature construction means for receiving the one or more third features and applying one or more kinds of multi-operand functions, which compute at least one of the features on the basis of one or more of the features, to the received third features, and constructing one or more fourth features representing the results of applying the multi-operand functions to the third features; and
second feature selection means for computing relevance between (iii) the one or more of the fourth features and the one or more third features and (iv) the target variable, and selecting at least one fifth feature that represents highly relevant feature from among the one or more of the fourth features and the one or more third features.
In addition, as another aspect of the present invention, a feature-converting method including causing an information processing device to includes:
receiving one or more first features representing features including one or more elements composed of a numeral or a code, and constructing, by applying one or more unary functions that compute at least one of the features on the basis of at least one of the features to the received first features, one or more second features representing results of applying the unary operation functions to the first features;
computing relevance between (i) the one or more second features and the one or more first features and (ii) a target variable that includes one or more elements composed of a numeral or a code associated with one or more of the elements included in the first features, and selecting one or more third features representing highly relevant features from among the one or more second features and the one or more first features;
receiving the one or more third features and applying one or more kinds of multi-operand functions, that compute at least one of the features on the basis of one or more of the features, to the received third features, and constructing one or more fourth features representing the results of applying the multi-operand functions to the third features; and
computing relevance between (iii) the one or more of the fourth features and the one or more third features and (iv) the target variable, and selecting at least one fifth feature that represents highly relevant feature from among the one or more of the fourth features and the one or more third features.
Furthermore, the object is also realized by a feature-converting program, and a computer-readable recording medium which records the program.
The feature-converting device and the like according to the present invention can provide good features quickly.
Now, terms and the like for helping understanding of the present invention will be described before describing exemplary embodiments for implementing the present invention.
In a learning algorithm, the more the explanatory variables are included in learning information, the more the computed relationship fits the learning information, while the less it fits predictive information representing information regarding a target to be predicted. The above-mentioned problem in the learning algorithm is known as overlearning problems. As a result of that, for example, when overlearning problems occur in the action determination device disclosed in PTL 1, accuracy of prediction declines.
According to an information criterion, appropriate setting of the amount of explanatory variables can alleviate overlearning problems in a learning algorithm. In learning algorithms, alleviating overlearning problems improves accuracy of prediction regarding predictive information.
For descriptive convenience, each feature is assumed to include a plurality of elements each including a numeral, a code, or the like.
Feature selections represent that an appropriate amount of features are selected from among features. Feature selections select an appropriate amount of features, for examples, on the basis of a score function for each feature. As for the score function, various methods are known, such as a correlation with respect to a target variable, an information gain, a chi-square value, and a Hilbert-Schmidt Independence Criterion.
In addition, feature constructions (feature conversions) are examples of methods for achieving high accuracy of prediction and convert a given feature to one or more appropriate feature.
Examples of feature constructions include a logarithmic function (log(x)), a square function (X×X), a binary function (to convert to the value of 0 or 1 on the basis of the value of X), a product function (Xi×Xj), and a quotient function (Xi÷Xj). Additionally, X represents one feature. Furthermore, Xi and Xj each represent one feature in a feature set representing a set of features, provided that 1≦i≦N and 1≦j≦N where N represents the number of features included in the feature set.
Additionally, in the present application, “log” means a logarithmic function. In addition, the base of the logarithmic function is, for example, a Napier's constant. However, the base of the logarithmic function is not limited to the Napier's constant.
Next, the technical problem to be solved by the present invention will be described in more detail. First, to facilitate understanding, the summary of related art of the present invention will be described.
The present applicant has filed U.S. Patent Application (provisional application) No. 61/883,660 (filed on Sep. 27, 2013) and International Patent Application No. PCT/JP2014/004520 that claims priority based on the US Patent Application, prior to filing of the present application. The invention disclosed in the patent application will be briefly described.
An information processing device disclosed in the patent application synthesizes a plurality of functions, thereby constructing a new function, and applies the constructed new function to a feature, thereby constructing a new feature. Next, the information processing device determines whether or not the constructed new feature satisfies a predetermined condition. For example, the information processing device synthesizes N (provided that N≧1) kinds of functions twice, thereby constructing (N×N) kinds of functions. Accordingly, when M (provided that M≧1) features are input, the information processing device constructs (M×N×N) features. In other words, since the information processing device can construct many features, the above-described overlearning problem can occur, depending on the situation, when learning processing is executed on the basis of the features.
Furthermore, the present applicant has filed U.S. Patent Application (provisional application) No. 61/883,672 (filed on Sep. 27, 2013) and International Patent Application No. PCT/JP2014/004706 that claims priority based on the US Patent Application, prior to filing of the present application. The invention disclosed in the patent application will be briefly described.
The information processing device disclosed in the patent application selects, for a function that takes a plurality of values as operands, a combination of features that serve as the operands from among a plurality of features and applies the function to the selected combination of the features, thereby constructing a new feature. Next, the information processing device determines whether or not the constructed new feature satisfies a predetermined condition. For example, the information processing device applies a function that takes two kinds of values as operands to M (provided that M≧1) kinds of features, thereby constructing (M×M) features. Accordingly, when there are N (provided that N≧1) kinds of functions that take two kinds of values as operands, the information processing device constructs (N×M×M) features. In other words, even the information processing device can construct many features, as in the application by the applicant of the present application. Thus, the above-described overlearning problem can occur depending on the situation in executing learning processing on the basis of the features.
Accordingly, feature selection needs to be executed for solving these overlearning problems. However, there is a problem in that as the number of features increases, a computation load for executing feature selection becomes larger.
Hereinafter, exemplary embodiments of the present invention capable of solving such problems will be described in detail with reference to the drawings.
A structure of a feature-converting device 105 according to a first exemplary embodiment of the present invention and processing executed by the feature-converting device 105 will be described with reference to
The feature-converting device 105 according to the first exemplary embodiment includes a first feature construction unit 101, a first feature selection unit 102, a second feature construction unit 103, and a second feature selection unit 104.
First, the first feature construction unit 101 applies arithmetic processing for computing one or more features on the basis of at least one or more features to the first features 501 in response to receipt of first features 501, thereby computing a second feature(s) 502 (step S101). For example, the arithmetic processing may be a unary function (single-operand function) for computing one feature on the basis of one feature. Examples of the unary function will be presented in a second exemplary embodiment that will be described later.
In addition, for descriptive convenience, the arithmetic processing to be applied in the first feature construction unit 101 is referred to as first arithmetic processing.
For example, the first features 501 are features (Xn, provided that n is an integer of from 1 to 4) included in learning information exemplified in
In the learning information exemplified in
In the example depicted in
In addition, with reference to
In the above-described example, for example, when each of the pieces of the information D1 to D8 is associated with sales in a specific day, sales in a specific day can be predicted on the basis of atmospheric temperatures before the specific day.
In addition, although it has been assumed that the learning information includes the information D1 to D8, the learning information may include much more information. Additionally, although it has been assumed that each of the information D1 to D8 is represented using the features X1 to X4, the information D1 to D8 may be represented using many more features. In addition, in the example depicted in
For example, the first feature construction unit 101 applies a predetermined function such as sin(feature X1) or “feature 2”×log(feature X3) to the above-described features X1 to X4, thereby converting to new features (step S101). Additionally, the sin represents a sine that is a trigonometric function. For example, the predetermined function may be a function for converting N (provided that N is a positive integer) features to M (provided that M is an integer satisfying 1≦M≦N) different features, as in a method for selecting components having high contribution rate in principal component analysis. The predetermined function is not limited to the above-described example.
With reference to
Next, the first feature selection unit 102 selects a third feature(s) 503 from the first features 501 and the second features 502 computed by the first feature construction unit 101 according to a feature selection procedure (step S102). Additionally, for descriptive convenience, the feature selection procedure in the first feature selection unit 102 is referred to as a first feature selection procedure.
When the feature selection procedure is, for example, a means that selects a feature(s) high in relevance (relativity) between features and a target variable, the first feature selection unit 102 selects a third feature(s) 503 by computing relevance between the features and the target variable.
The relevance can be computed, for example, on the basis of a Pearson's correlation coefficient, a cosine similarity, a Hilbert-Schmidt Independence Criterion (HSIC), or the like. Alternatively, the relevance can be computed on the basis of a Maximal Information Coefficient (MIC) or the like.
The feature selection procedure is not limited to the above-described example, and may be, for example, a method of selecting a specific feature(s) on the basis of a relevance among a plurality of (plural) features and a relevance between each feature and a target variable. Alternatively, the feature selection procedure may be a method of selecting a specific feature(s) on the basis of the relevance between the plurality of features. In this case, the feature selection procedure selects, for example, a feature(s) having a low relevance between the plurality of features. As the feature selection procedure, various methods are already known. Thus, the description thereof will be omitted.
Other than the above-described examples, various methods for relevance computation are already known. Thus, in the present exemplary embodiment, a detailed description regarding the method for relevance computation will be omitted.
With reference to an example in which values of the target variable are those depicted in
Referring to
For example, the values of the target variable Y represent sales in specific days. In other words, in this example, the information D1 is information in which sales in a specific day are associated with an atmospheric temperature before the specific day. In addition, the information D2 is information in which a second specific day different from the specific day is associated with an atmospheric temperature before the second specific day. Learning information is, for example, information in which regarding the specific days represented by the information D1 to D8, the example depicted in
For example, the first feature selection unit 102 computes correlation coefficients (
Next, the first feature selection unit 102 selects features highly relevant to the target variable Y. Referring to
In addition, the number of the third features 503 can be any number as long as it is smaller than a sum of the first features 501 and the second features 502. Thus, the number of the third features 503 is not limited to the above-described example.
In addition, in the above-described example, the feature selection has been assumed to be the means for selecting highly relevant features between features and a target variable. However, the feature selection may be a means in which indices representing relevance among features are further incorporated. In this case, the feature selection is a procedure for selecting features highly relevant to the target variable and lowly relevant to one another as the third feature(s) 503. Additionally, as the indices representing the relevance, indices such as correlation coefficients and information gain are already known. Thus, in the present exemplary embodiment, detailed descriptions of the indices and the feature selection will be omitted.
Next, the second feature construction unit 103 applies arithmetic processing that computes one or more features on the basis of at least one or more features to the third features 503 selected by the first feature selection unit 102 to compute fourth features 504 (step S103). For example, the second feature construction unit 103 applies a multi-operand function (polynomial function) that computes at least one feature on the basis of a plurality of features to the third features 503 to compute the fourth features 504. In addition, one example of the arithmetic processing is the polynomial function as shown in the second exemplary embodiment of the present invention.
For descriptive convenience, the arithmetic processing applied in the second feature construction unit 103 is referred to as second arithmetic processing.
In addition, the second feature construction unit 103 may compute the fourth features 504 on the basis of the first features 501 and the third features 503. In this case, since the first features 501 are features to be received, the second feature construction unit 103 computes the fourth features 504 on the basis of the features input by a user. When the features to be input by the user are previously known to be good features, the second feature construction unit 103 is highly likely to compute better features on the basis of the features.
With reference to an example depicted in
In the example depicted in
In addition, in the example depicted in
Next, the second feature selection unit 104 selects a fifth features 505 from the first features 501 to fourth features 504 according to a feature selection (step S104). Additionally, for descriptive convenience, the feature selection in the second feature selection unit 104 is referred to as second feature selection.
In addition, the feature selection in the second feature selection unit 104 may be the same as or different from the feature selection procedure in the first feature selection unit 102.
For example, when the feature selection is a means for selecting features highly relevant to a target variable, the second feature selection unit 104 selects a fifth features 505 by computing relevance between the features and the target variable.
For example, the second feature selection unit 104 computes relevance by computing correlation coefficients between the target variable Y and the features Z1 to Z6 (
For example, referring to
Next, the second feature selection unit 104 selects features highly relevant to the target variable Y. Referring to
In addition, the number of the fifth features 505 can be any number as long as it is smaller than a sum of the first features 501 to the fourth features 504. Thus, the number of the fifth features 505 is not limited to the above-described example.
In addition, in the example described above, the feature selection has been assumed to be the means for selects features highly relevant to a target variable, but may be a means in which indices representing relevance among a plurality of features are additionally incorporated. In this case, the feature selection is a procedure for selecting feature(s) highly relevant to the target variable and lowly relevant to one another as the fifth features 505. Additionally, as the indices representing the relevance, indices such as correlation coefficients and information gain are already known. Thus, in the present exemplary embodiment, detailed descriptions of the indices and the feature selection will be omitted.
In addition, as depicted in
Furthermore, the feature-converting device 105 may include an aspect in which the feature construction unit constructs features and the feature selection unit selects on the basis of features such as the constructed features. In this case, the feature-converting device 105 repeatedly performs feature construction and feature selection.
For example, the learning device 122 depicted in
The learning device 122 includes the feature-converting device 105 and a learning unit 121.
The feature-converting device 105 constructs the fifth features 505 on the basis of the first features 501 according to the above-described procedure. Next, the learning unit 121 computes relationships between the explanatory variables and a target variable on the basis of learning information including the fifth features 505 as explanatory variables. Alternatively, the learning unit 121 applies the relationships to predictive information including the fifth features 505 as explanatory variables to estimate values regarding the predictive information.
Next, a description will be given of advantageous effects regarding the feature-converting device 105 according to the present exemplary embodiment.
The feature-converting device 105 according to the present exemplary embodiment can provide good features quickly. The reason for this is that the feature-converting device 105 reduces the number of arithmetic operations while maintaining the quality of features as compared to typical feature-converting devices.
The reason will be described in detail with reference to
As exemplified in PTL 1, the typical feature-converting device includes one feature construction unit and one feature selection unit. The feature construction unit computes new features on the basis of the received features. The feature selection unit selects some features from the new features.
For example, the typical feature-converting device applies a certain function to the received features to compute new features. In the examples depicted in
Next, the feature selection unit of the typical feature-converting device computes relevance between the target variable (
Specifically, in this example, in the typical feature-converting device, the feature selection unit receives as input 36 (=4 (received features)+4 (log-applied features)+8×7÷2 (features as products of each two features) features.
On the other hand, the feature-converting device 105 according to the present exemplary embodiment performs feature constructing processing and processing for selecting features from the constructed features and the like a plurality of times. The first feature selection unit 102 receives as input 8 (=4 (received features)+4 (log-applied features)) features. The second feature selection unit 104 receives as input 16 (=4 (received features)+4 (log-applied features)+6 (features computed by the second feature construction unit 103)) features.
As described above, in the typical feature selection means, the amount of computations sharply increases according to the number of input features. In this case, it is assumed that the typical feature selection means is a linear-order computational algorithm respect to the number of input features in case that the amount of computations by the typical feature selection means is estimated to be small. Even in this case, while the typical feature-converting device needs to process 36 features, the feature-converting device 105 according to the present exemplary embodiment processes 24 (=8+16) features. Accordingly, since the number of features to be processed is reduced, the feature-converting device 105 according to the present exemplary embodiment can provide features more quickly than the typical feature-converting device.
Next, it will be described that the feature-converting device 105 according to the present exemplary embodiment maintains the quality of features.
Referring to
The feature-converting device 105 processes only a smaller number of features than the typical feature-converting device. Nevertheless, the comparison between features computed by the feature-converting device 105 and features computed by the typical feature-converting device indicates that the correlation coefficients between the features and the target variable are equal.
In addition, even when features highly relevant to the target variable are some features that constitute the features, they often have high relevance to the target variable. Conversely, in cases where there are low relevance between the some features and the target variable, even when features are constructed by combining the some features, relevance between the constructed features and the target variable are often low. The feature-converting device 105 according to the present exemplary embodiment constructs features highly relevant to a target variable in the step-by-step manner and therefore is unlikely to construct features lowly relevant to the target variable.
Furthermore, when the feature selection means is a relevance-based means, the higher the relevance of a feature with a target variable is, the better the quality of the feature is. Thus, the feature-converting device 105 according to the present exemplary embodiment can maintain the quality of features.
In addition, by repeating the feature constructing processing and the feature selecting processing in the feature-converting device 105, the number of features to be processed by the feature selection unit is further reduced. Thus, even in an aspect in which the feature-converting device 105 includes three or more feature construction units and three or more feature selection units, the feature-converting device 105 according to the present exemplary embodiment can provide good features more quickly.
A learning device that has the feature-converting device 105 according to the present exemplary embodiment estimates on the basis of good features provided by the feature-converting device 105. Accordingly, the learning device 122 according to the present exemplary embodiment can achieve high accuracy of prediction.
Next, a description will be given of the second exemplary embodiment of the present invention based on the above-described first exemplary embodiment.
In the following description, characteristic parts according to the present exemplary embodiment will be mainly described, and the same structural parts as those of the above-described first exemplary embodiment will be denoted by the same reference numerals, thereby omitting overlapping detailed descriptions thereof.
With reference to
The feature-converting device 202 according to the second exemplary embodiment includes a first feature construction unit 201, the first feature selection unit 102, the second feature construction unit 103, and the second feature selection unit 104.
First, the first feature construction unit 201 applies a unary function, which computes one value on the basis of one value, to the elements of each of the first features 501, thereby computing the second features 502.
Examples of the unary function include various functions such as sin functions (sine functions), cos functions (cosine functions), exponential functions, logarithmic functions, polynomial functions, functions that provide frequencies in classification into a histogram, and deviations. Additionally, the unary function may be a function that rounds up or down values after the decimal point in a real number, or the like. In addition, the unary function may be a function that provides weather in an area to the name of the area, or the like. In addition, the one value may be, for example, a feature that represents a set of a plurality of values. When an input value is a feature, the unary function executes computation on the basis of the feature, and outputs a feature obtained as a result of the computation. For example, when the input value is a feature and the unary function is a logarithmic function, the unary function represents a function that applies the logarithmic function to each element of the feature to output computed values.
Such a unary function may be a function that computes 1 when the value of a certain one element of one or more elements constituting a feature is equal to or more than a specific threshold value, and computes 0 when the value of the one element is less than the specific threshold value.
In addition, the unary function may be a function that computes a moving average for each element included in a feature. In this case, for example, the unary function computes an average of one or more elements adjacent to an i-th element for the i-th element in the feature. The adjacent elements may be defined, for example, on the basis of a percentage (from about 1 to 10%) of the number of elements included in the feature.
In addition, the unary function may be a function that computes a value of a (i+k)-th element (or a (i−k)-th element) for an i-th element in the feature. k may be defined, for example, on the basis of the percentage (from about 1 to 10%) of the number of elements included in the feature.
The unary function is not limited to the above-described examples.
As the first arithmetic processing, the first feature construction unit 201 applies a unary function that computes one feature on the basis of one first feature 501 to each of the first features 501, thereby computing the second features 502.
Next, according to the feature selection procedure, the first feature selection unit 102 selects the third features 503 on the basis of the second features 502.
In addition, the second feature selection unit 104 may apply a polynomial function that computes one value on the basis of two or more values to each of the third features 503 to compute the fourth features 504.
Described will be advantageous effects regarding the feature-converting device 202 according to the present exemplary embodiment.
The feature-converting device 202 according to the present exemplary embodiment can provide better features more quickly than the feature-converting device 105 according to the first exemplary embodiment.
The reasons for this are twofold: reason 1 and reason 2. That is,
(Reason 1): The structural parts of the feature-converting device 202 according to the second exemplary embodiment include the structural parts of the feature-converting device 105 according to the first exemplary embodiment; and
(Reason 2): By reducing the number the features that are constructed by the first feature construction unit 201, processing by the first feature selection unit 102 is reduced as compared to that by the first feature selection unit 102 in the first exemplary embodiment.
Referring again to the examples depicted in
In addition, an input of the polynomial function may be features that represent a set of a plurality of elements. In this case, the polynomial function executes an arithmetic operation on the basis of the input features and outputs a feature obtained as a result of the operation. For example, when input values are two features and the polynomial function is multiplication, the polynomial function represents a function that outputs a value computed by multiplication between corresponding elements of the two features. Examples of the polynomial function can include logical OR operation, logical AND operation, logical exclusive OR operation, multiplication (product), and division (quotient).
Referring to
Furthermore, the features X2, X3, log(X2), and log(X3) are found to respectively have higher correlation coefficients with the target variable than features X1, X4, log(X1), and log(X4).
On the other hand, the feature-converting device 202 can select the above-mentioned features having higher correlation coefficients (i.e., the third features) on the basis of the first features by applying the unary function to each of the first features 501 and, then, selecting the features on the basis of the results thereof. The feature-converting device 202 constructs fourth features on the basis of the third features and thus executes processing constructing features regarding the features that have high correlation coefficients and are in small numbers.
Accordingly, the feature-converting device 202 according to the present exemplary embodiment first applies a unary function and therefore can reduce processing for combining a plurality of features. Better features can be provided more quickly than in the feature-converting device 105 according to the first exemplary embodiment.
In addition, the second feature construction unit 103 performs processing for applying a polynomial function, whereby it can be prevented that the first feature construction unit 201 and the second feature construction unit 103 perform overlapping processing. As a result of this, the feature-converting device 202 according to the present exemplary embodiment can provide better features more quickly.
Next will be described a third exemplary embodiment of the present invention based on the above-described first exemplary embodiment.
In the following description, characteristic parts according to the present exemplary embodiment will be mainly described, and the same structural parts as those of the above-described first exemplary embodiment will be denoted by the same reference numerals, thereby omitting overlapping descriptions thereof.
With reference to
The feature-converting device 303 according to the third exemplary embodiment includes the first feature construction unit 101, the first feature selection unit 102, a second feature construction unit 301, and a second feature selection unit 302.
The second feature construction unit 301 applies a linear function to the third features 503 to construct the fourth features 504.
Next, the second feature selection unit 302 selects the fifth features 505 on the basis of the first features 501 to the fourth features 504 according to a feature selection procedure for selecting features according to indices based on the linear function.
For example, the linear function is an operation of a product, a sum, or the like.
Next will be described advantageous effects regarding the feature-converting device 303 according to the present exemplary embodiment.
The feature-converting device 303 according to the present exemplary embodiment can provide good features quickly, as well as can provide features easily understandable to a user.
The reasons for this are twofold: reason 1 and reason 2. That is,
(Reason 1) the structural parts of the feature-converting device 303 according to the third exemplary embodiment include the structural parts of the feature-converting device according to the first exemplary embodiment; and
(Reason 2) it can be prevented that a nonlinear function is additionally applied to features computed on the basis of a nonlinear function.
The reason 2 will be further described.
When the first feature construction unit 101 and the first feature selection unit 102 process on the basis of a nonlinear function, the third features 503 will be features computed by applying the nonlinear function to the first features 501. Accordingly, when the second feature construction unit 301 and the first feature selection unit 302 process on the basis of the nonlinear function, the fifth features 505 will be a features computed by applying the nonlinear function twice to the first features 501. In general, it is difficult for a user to understand values computed by applying a nonlinear function twice.
Accordingly, since the second feature construction unit 301 and the second feature selection unit 302 process on the basis of a linear function, it can be prevented that a nonlinear function is applied twice. As a result of this, the feature-converting device 303 according to the present exemplary embodiment can provide features easily understandable to a user.
Additionally, in the feature selection means, computation time is shorter in linear function-based processing than in nonlinear function-based processing. In other words, performing linear function-based processing by the second feature construction unit 301 and the second feature selection unit 302 reduces processing time in the second feature construction unit 301 and the second feature selection unit 302. Thus, the feature-converting device 303 according to the present exemplary embodiment can provide good features more quickly.
(Hardware Configuration Example)
A configuration example of hardware resources that realize a feature-converting device in the above-described exemplary embodiments of the present invention using a single calculation processing apparatus (an information processing apparatus or a computer) will be described. However, the feature-converting device may be realized using physically or functionally at least two calculation processing apparatuses. Further, the feature-converting device may be realized as a dedicated apparatus.
The non-transitory recording medium 24 is, for example, a computer-readable Compact Disc, Digital Versatile Disc, Universal Serial Bus (USB) memory, or Solid State Drive. The non-transitory recording medium 24 allows a related program to be holdable and portable without power supply. The non-transitory recording medium 24 is not limited to the above-described media. Further, a related program can be carried via a communication network by way of the communication I/F 27 instead of the non-transitory medium 24.
In other words, the CPU 21 copies, on the memory 22, a software program (a computer program: hereinafter, referred to simply as a “program”) stored by the disc 23 when executing the program and executes arithmetic processing. The CPU 21 reads data necessary for program execution from the memory 22. When display is needed, the CPU 21 displays an output result on the output apparatus 26. When a program is input from the outside, the CPU 21 reads the program from the input apparatus 25. The CPU 21 interprets and executes a feature-converting program present on the memory 22 corresponding to a function (processing) indicated by each unit illustrated in
In other words, in such a case, it is conceivable that the present invention can also be made using the feature-converting program. Further, it is conceivable that the present invention can also be made using a computer-readable, non-transitory recording medium storing the feature-converting program.
The present invention has been described using the above-described exemplary embodiments as exemplary cases. However, the present invention is not limited to the above-described exemplary embodiments. In other words, the present invention is applicable with various aspects that can be understood by those skilled in the art without departing from the scope of the present invention.
This application is based upon and claims the benefit of priority from U.S. patent application No. 61/971,585, filed on Mar. 28, 2014, the disclosure of which is incorporated herein in its entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2015/001120 | 3/3/2015 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
61971585 | Mar 2014 | US |