The embodiments described herein generally relate to generating an interpretive behavioral model, and more particularly, to generating an interpretive behavioral model that may be utilized to determine an extent to which an input within a dataset influences or causes an output.
Various machine learning models may be fitted to a dataset in order to generate outputs. These machine learning models, when fitted to a robust and comprehensive input dataset, may generate highly accurate outputs. While accurate, these outputs may not be interpretable or provide information regarding one or more factors or characteristics present in the input dataset that resulted in the generated outputs. As such, while the outputs may be accurate and likely remain accurate across a large and varied dataset, the reasons for the generation of the output may remain unclear, rendering the machine learning models opaque and uninterpretable. While multiple behavioral models may better facilitate interpretation of input datasets and identification of one or more characteristics that influence an extent to which certain outputs may result from certain inputs, these behavioral models are tested within and accurate only within limited environments. With an increase in the robustness and diversity of the dataset, the efficacy of these behavioral models is reduced.
Accordingly, a need exists for alternative methods of generating an interpretive behavioral model that enables the interpretation of an extent to which one or more characteristics of data within an input dataset influences or causes the generation of one or more outputs.
In one embodiment, a method for generating an interpretive behavioral model is provided. The method is implemented by a computing device and includes fitting a machine learning trained model to a plurality of data features included as part of a data feature-set for identifying a generative model, the fitting generates complete data feature-set outputs that are associated with a first set of accuracy values, iteratively fitting, after an iterative removal of each data feature from the data feature-set, the machine learning trained model to subsets of the plurality of data features to determine respective reduced feature-set outputs, each subset lacks a different data feature of the plurality of data features, determining one or more of the reduced feature-set outputs as corresponding to a second set of accuracy values that are lower than the first set of accuracy values, designating the iteratively removed data features that are associated with the one or more of the reduced feature-set outputs corresponding to the second set of accuracy values that are lower than the first set of accuracy values as accuracy-modifying data features included as part of an accuracy-modifying data feature-set, generating a first linear model based on the accuracy-modifying data features, generating a second linear model that is based on one of the accuracy-modifying data features having a weight that is highest relative to a remaining ones of the accuracy-modifying data features, and identifying the second linear model as the generative model responsive to determining that the linear model accuracy value of the second linear model exceeds each of the first set of accuracy values.
In another embodiment, a system for generating an interpretive behavioral model is provided. The system includes one or more processors included as part of a computing device, and non-transitory computer readable medium storing instructions that, when executed by the one or more processors, cause the computing device to fit a machine learning trained model to a plurality of data features included as part of a data feature-set for identifying a generative model, the fitting generates complete data feature-set outputs that are associated with a first set of accuracy values, iteratively fit, after an iterative removal of each data feature from the data feature-set, the machine learning trained model to subsets of the plurality of data features to determine respective reduced feature-set outputs, each subset lacks a different data feature of the plurality of data features, determine one or more of the reduced feature-set outputs as corresponding to a second set of accuracy values that are lower than the first set of accuracy values, designate the iteratively removed data features that are associated with the one or more of the reduced feature-set outputs corresponding to the second set of accuracy values that are lower than the first set of accuracy values as accuracy-modifying data features included as part of an accuracy-modifying data feature-set, generate a first linear model based on the accuracy-modifying data features, generate a second linear model that is based on one of the accuracy-modifying data features having a weight that is highest relative to a remaining ones of the accuracy-modifying data features, and identify the second linear model as the generative model responsive to determining that the linear model accuracy value of the second linear model exceeds each of the first set of accuracy values.
In another embodiment, a method for generating an interpretive behavioral model is provided. The method is implemented by a computing device and includes iteratively fitting, after an iterative removal of each data feature from a data feature-set, a machine learning trained model to subsets of a plurality of data features to determine respective reduced feature-set outputs, determining one or more of the reduced feature-set outputs as corresponding to a second set of accuracy values that are lower than a first set of accuracy values associated with a complete data feature-set outputs, designating one or more of the iteratively removed data features that are associated with the one or more of the reduced feature-set outputs corresponding to the second set of accuracy values as accuracy-modifying data features, generating a first linear model based on the accuracy-modifying data features, generating a second linear model that is based on one of the accuracy-modifying data features having a weight that is highest relative to a remaining ones of the accuracy-modifying data features, and identifying the second linear model as the generative model responsive to determining that the linear model accuracy value of the second linear model exceeds each of the first set of accuracy values.
These and additional features provided by the embodiments described herein will be more fully understood in view of the following detailed description, in conjunction with the drawings.
The embodiments set forth in the drawings are illustrative and exemplary in nature and not intended to limit the subject matter defined by the claims. The following detailed description of the illustrative embodiments can be understood when read in conjunction with the following drawings, where like structure is indicated with like reference numerals and in which:
Currently, different types of machine learning models may be fitted to a dataset, for the purpose of generating various outputs having a high level of accuracy. These machine learning models, when fitted to a robust and comprehensive input dataset, may enable the generation of outputs that are highly accurate. However, these outputs may not be interpretable. As such, while the outputs may be accurate and likely remain accurate across a large and varied dataset, the reasons for the generation of the output may remain unclear, rendering the machine learning models uninterpretable. While multiple behavioral models may better facilitate interpretation of input datasets and identification of one or more characteristics that influence an extent to which one or more particular outputs may result from the input datasets, these behavioral models are tested within and are accurate only within limited environments.
The embodiments described herein address and overcome the deficiencies of the above described conventional machine learning based models and behavioral models. In particular, the embodiments described herein provide for the generation of an interpretive behavior model that may be utilized to accurately generate a set of outputs based on a given set of inputs in addition to determining the extent to which each input, directly and indirectly, influenced or affected the generation of the set of outputs. For example, the interpretive behavioral model may determine the extent to which a consumer's decision to purchase a product was influenced by various factors, e.g., a discount sign near the product, a lighting near the product, the location of the product within a store, and so forth. The interpretative behavioral models may also determine the extent to which a combination of, e.g., the discount sign near a product and the lighting near the product, affected a consumer's decision to purchase the product.
The memory 102 may store operating instructions 103, each of which may be embodied as a computer program, firmware, and so forth. The memory 102 may comprise RAM, ROM, flash memories, hard drives, or any device capable of storing the operating instructions 103 such that the operating instructions 103 can be accessed by the processor 106. The operating instructions 103 may comprise logic or algorithm(s) written in any programming language of any generation (e.g., 1GL, 2GL, 3GL, 4GL, or 5GL) such as, for example, machine language that may be directly executed by the processor, or assembly language, object-oriented programming (OOP), scripting languages, microcode, etc., that may be compiled or assembled into machine readable and executable instructions and stored on the memory 102. Alternatively, the operating instructions 103 may be written in a hardware description language (HDL), such as logic implemented via either a field-programmable gate array (FPGA) configuration or an application-specific integrated circuit (ASIC), or their equivalents. Accordingly, the methods described herein may be implemented in any conventional computer programming language, as pre-programmed hardware elements, or as a combination of hardware and software components. The processor 106 along with the memory 102 may operate as a controller for the computing device 100.
A local interface 104 is also included in
The operating instructions 103 may include an operating system and/or other software for managing components of the computing device 100. It should be understood that while the component in
In embodiments, at block 210, the computing device 100 may fit a machine learning trained model to a plurality of data features included as part of a data feature-set for the purpose of identifying a generative model. The fitting of the machine learning trained model results in the generation of a set of complete data feature-set outputs that are associated with a first set of accuracy values.
At block 220, the computing device 100 may iteratively fit, after an iterative removal of each data feature from the data feature-set, the machine learning trained model to subsets of the plurality of data features to determine respective reduced feature-set outputs, and each subset lacks a different data feature of the plurality of data features.
At block 230, the computing device 100 may determine one or more of the reduced feature-set outputs as corresponding to a second set of accuracy values that are lower than the first set of accuracy values.
At block 240, the computing device 100 may designate the iteratively removed data features that are associated with the one or more of the reduced feature-set outputs corresponding to the second set of accuracy values that are lower than the first set of accuracy values as accuracy-modifying data features included as part of an accuracy-modifying data feature-set.
At block 250, the computing device 100 may generate a first linear model based on the accuracy-modifying data features.
At blocks 260 and 270, the computing device 100 may determine a respective weight for each of the accuracy-modifying data features and generate a second linear model that is based on one of the accuracy-modifying data features having a weight that is highest relative to a remaining ones of the accuracy-modifying data features. At block 280, the computing device 100 may identify the second linear model as the generative model responsive to determining that the linear model accuracy value of the second linear model exceeds each of the first set of accuracy values.
In embodiments, the example generative model 304 may be fitted to the example input dataset 302 such that the model may analyze at least a subset of the information included in the example input dataset 302. Based on the analysis, the generative model 304 may generate the example output dataset 306, which may indicate that one or more individuals purchased a particular good or a particular service that is located within the store. In embodiments, the example generative model 304, which may be a neural network trained model, may generate the example output dataset 306 that is highly accurate and indicative of a strong correlation between the input and the output. For example, the generative model 304 may be a neural network trained model that accurately establishes the correlation between the example input dataset 302 and the example output dataset 306.
However, the generative model 304 may not provide information relating to the degree to which a particular data feature or specific details included in the example input dataset 302 influences the example output dataset 306. As such, while the generative model 304 may be helpful in predicting a strong correlation between the example input dataset 302 and the example output dataset 306, the model may not be useful in determining, e.g., whether a lighting near a product influenced the purchasing decision of an individual to a higher extent than a discount sign that may have been positioned within a few feet of the product, and so forth. In other words, the generative model 304 may fail to enable a deeper and more comprehensive understanding of user behavior, in particular, the extent to which a particular individual's decision to, e.g., purchase a product, was influenced by a particular factor, for example. For this reason, the example generative model 304 may not be considered interpretable.
In embodiments, as stated above, an example behavioral model may be generated and fitting to a subset of the input dataset 302 and this model may provide insights into the extent and manner in which a particular feature (or set of features) in the subset of the input dataset 302 influences a purchase decision. For example, the example behavior model may enable a determination that a discount sign located near a product had an 80% influence on a user's decision to purchase the product and the lighting around the product (e.g., around a shelf on which a product was located) contributed to a 20% influence on the user's decision to purchase the product. However, such a behavioral model has certain limitations, namely that accuracy levels of the model may reduce with an increase in the size and diversity of the datasets. As such, both models, individually, fail to provide robust insights regarding the extent to which one or more factors influence user behavior. However, the processes described in
As a non-limiting example, the trained model 404 may include one or more artificial neural networks (ANNs) and may include connections between nodes that form a directed acyclic graph (DAG). ANNs may include node inputs, one or more hidden activation layers, and node outputs, and may be utilized with activation functions in the one or more hidden activation layers such as a linear function, a step function, logistic (sigmoid) function, a tanh function, a rectified linear unit (ReLu) function, or combinations thereof. ANNs are trained by applying such activation functions to training data sets to determine an optimized solution from adjustable weights and biases applied to nodes within the hidden activation layers to generate one or more outputs as the optimized solution with a minimized error.
In machine learning applications, new inputs may be provided (such as the generated one or more outputs) to the ANN model as training data to continue to improve accuracy and minimize error of the ANN model. The one or more ANN models may utilize one to one, one to many, many to one, and/or many to many (e.g., sequence to sequence) sequence modeling. The trained model 404 as described herein may utilize one or more ANN models as understood to those skilled in the art or as yet-to-be-developed to generate results as described in embodiments herein. Such ANN models may include artificial intelligence components selected from the group that may include, but not be limited to, an artificial intelligence engine, Bayesian inference engine, and a decision-making engine, and may have an adaptive learning engine further comprising a deep neural network learning engine. The one or more ANN models may employ a combination of artificial intelligence techniques, such as, but not limited to, Deep Learning, Random Forest Classifiers, Feature extraction from audio, images, clustering algorithms, or combinations thereof.
As another non-limiting example, a convolutional neural network (CNN) may be utilized. For example, a convolutional neural network (CNN) may be used as an ANN that, in a field of machine learning, for example, is a class of deep, feed-forward ANNs that may be applied for audio-visual analysis. CNNs may be shift or space invariant and utilize shared-weight architecture and translation invariance characteristics. Additionally or alternatively, a recurrent neural network (RNN) may be used as an ANN that is a feedback neural network. RNNs may use an internal memory state to process variable length sequences of inputs to generate one or more outputs. In RNNs, connections between nodes may form a DAG along a temporal sequence. One or more different types of RNNs may be used such as a standard RNN, a Long Short Term Memory (LSTM) RNN architecture, and/or a Gated Recurrent Unit RNN architecture.
Returning to
Thereafter, the reduced feature set outputs may be compared with the complete data feature-set outputs as described in block 210 of
After the second set of accuracy values are generated and compared with the first set of accuracy values, the feature referring to the lighting of a product, for example, may be placed back into the reduced dataset, thereby including all of the theta values 402 and completing the example input dataset 302. Thereafter, iteratively, another feature may be removed from the example input dataset 302 (i.e. from the theta values 402). For example, the discount sign positioned near a particular product that was purchased by the individual may be removed from the example input dataset 302 and the example exhaustive interaction model 408 may again be fitted to the newly generated reduced set of theta values 402, namely the reduced set that lacks the data features or information relating to the discount sign. Additionally, the latest iteration of the fitting of the example exhaustive interaction model 408 may result in the generation of an additional set of accuracy values (e.g., a third set of accuracy values), which will be compared with the first set of accuracy value of the completed data feature-set outputs as described in block 210.
In embodiments, if the second set of accuracy values are lower than that the first set of accuracy values, the computing device 100 may designate the theta value that was removed from the example input dataset 302—the lighting feature in one example—as an accuracy modifying data feature (block 410), as the removal of this feature resulted in a second set of accuracy values for the reduced feature-set outputs that were lower than the first set of accuracy values. In other words, as the removal of the lighting feature reduced the accuracy of the reduced feature-set outputs, in embodiments, the lighting feature may be determined to affect the purchasing decision of an individual relative to a product. For example, it may be determined that the removal of the lighting feature reduced the accuracy of the reduced feature-set outputs by 20 percent.
Similarly, if the third set of accuracy values are lower than the first set of accuracy values, the computing device 100 may designate the second theta value that was removed from the example input dataset 302—the discount sign feature in one example—as an accuracy modifying data feature (block 410), as the removal of this feature resulted in a third set of accuracy values for the reduced feature-set outputs that were lower than the first set of accuracy values. In other words, as the removal of the discount sign feature reduced the accuracy of the reduced feature-set output, in embodiments, the discount sign feature may be determined to affect the purchasing decision of an individual relative to a product. For example, it may be determined that the removal of the discount sign feature reduces the accuracy of the reduced feature-set outputs by 50 percent.
It is noted that, each one of the theta values 402 may be removed from the set of theta values 402 and the exhaustive interaction model 408 may be iteratively fitted after the removal of each theta value to the reduced set of theta values 402 in order to generate reduced feature-set outputs, which will then be iteratively compared with the first set of accuracy values. Thereafter, if the accuracy value of a particular set of reduced feature-set outputs is lower than the first set of accuracy values of the example input dataset 302, the particular theta value that was removed from the theta values 402 will be designated as an accuracy modifying data feature.
The computing device 100, upon identifying and designating at least a subset of the example input dataset 302 (i.e. one or more of the theta values 402) as accuracy modifying data features, may generate a first linear model (block 412) that is based on the determined accuracy modifying data features. In embodiments, the linear model may be a machine learning trained model that utilizes one or more of the artificial intelligence neural network based processes and algorithms described above. In embodiments, upon generating the first linear model (block 412), the computing device 100 may determine a weight value for each of the respective accuracy modifying data features and rank these respective weight values according their magnitude. For example, if the lighting feature described above caused a reduction in the accuracy of the reduced feature-set outputs by 50 percent, the weight value of the discount sign feature (e.g., designated as an accuracy modifying data feature) may be high (e.g., a weight value of 7 out of 10) as compared to the weight value of the lighting feature, which may be relatively lower (e.g., 3 out of 10). Thereafter, the first linear model (block 412) may be compared to the trained model 404 (block 416).
In embodiments, it is noted each of the first linear model (block 412) and any subsequent linear models that are generated may include main effect components and interaction components associated with the accuracy-modifying data features. In embodiments, each main effect component associated with a particular accuracy-modifying data feature is representative of a direct influence that at least one of the accuracy-modifying data features may have on the accuracy of the reduced feature-set outputs. Additionally, each interaction component is representative of an indirect influence that at least one of the accuracy-modifying data features has on the accuracy of the reduced feature-set outputs described above.
For example, a main effect of the lighting feature described above may be determined such the removal of the lighting feature may reduce the accuracy the reduced feature-set outputs by 20 percent and the main effect of the discount sign feature may be determined such that the removal of the discount sign feature reduces the accuracy of the reduced feature set outputs by 50 percent. As such, the direct influence (e.g., main effect) of each of the light feature and the discount sign feature on the reduced feature set outputs may be associated with, e.g., an importance of the feature in a purchase decision of an individual. For example, the lighting may affect an individual's product purchase decision to an extent that is lower than the extent to which the discount sign feature may affect the individual's product purchase decision. Consequently, the main effect of the discount sign feature may be higher than the main effect of the lighting feature. Additionally, the interaction component associated with the lighting feature may refer to an extent to which the accuracy of the reduced feature set outputs may be reduced if the discount feature in combination with lighting feature are removed from the example input dataset 302. In other words the interaction feature refer to the extent to which the combination of both the lighting feature and the discount sign feature affects the purchase decision of an individual.
In embodiments, the generated interpretive behavioral model may enable the determination and prediction of various behavior patterns of various consumers based on an analysis of data features describing the consumer 502. For example, the interpretive behavioral model may be utilized to determine that the example discount sign 506 influences the purchase decision of the consumer 502, namely the decision of the consumer 502 to purchase one of the candy bars 504 on the shelf 508 to a larger extent than the fact that the candy bars 504 are located on the shelf 508 that is situated near the checkout counter. The interpretive behavior model may also be utilize to analyze and discern the effects of other factors on the purchasing decisions of other individuals.
It should be understood that the embodiments of the present disclosure are directed to a method for generating an interpretive behavioral model is provided. The method is implemented by a computing device and includes fitting a machine learning trained model to a plurality of data features included as part of a data feature-set for identifying a generative model, the fitting generates complete data feature-set outputs that are associated with a first set of accuracy values, iteratively fitting, after an iterative removal of each data feature from the data feature-set, the machine learning trained model to subsets of the plurality of data features to determine respective reduced feature-set outputs, each subset lacks a different data feature of the plurality of data features, determining one or more of the reduced feature-set outputs as corresponding to a second set of accuracy values that are lower than the first set of accuracy values, designating the iteratively removed data features that are associated with the one or more of the reduced feature-set outputs corresponding to the second set of accuracy values that are lower than the first set of accuracy values as accuracy-modifying data features included as part of an accuracy-modifying data feature-set, generating a first linear model based on the accuracy-modifying data features, determining a respective weight for each of the accuracy-modifying data features, generating a second linear model that is based on one of the accuracy-modifying data features having a weight that is highest relative to a remaining ones of the accuracy-modifying data features, and identifying the second linear model as the generative model responsive to determining that the linear model accuracy value of the second linear model exceeds each of the first set of accuracy values.
The terminology used herein is for the purpose of describing particular aspects only and is not intended to be limiting. As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms, including “at least one,” unless the content clearly indicates otherwise. “Or” means “and/or.” As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” or “includes” and/or “including” when used in this specification, specify the presence of stated features, regions, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, regions, integers, steps, operations, elements, components, and/or groups thereof. The term “or a combination thereof” means a combination including at least one of the foregoing elements.
It is noted that the terms “substantially” and “about” may be utilized herein to represent the inherent degree of uncertainty that may be attributed to any quantitative comparison, value, measurement, or other representation. These terms are also utilized herein to represent the degree by which a quantitative representation may vary from a stated reference without resulting in a change in the basic function of the subject matter at issue.
While particular embodiments have been illustrated and described herein, it should be understood that various other changes and modifications may be made without departing from the spirit and scope of the claimed subject matter. Moreover, although various aspects of the claimed subject matter have been described herein, such aspects need not be utilized in combination. It is therefore intended that the appended claims cover all such changes and modifications that are within the scope of the claimed subject matter.
This application claims the benefit of U.S. Provisional Application Ser. No. 63/187,134, filed on May 11, 2021, the contents of which are incorporated by reference herein in their entirety.
Number | Date | Country | |
---|---|---|---|
63187134 | May 2021 | US |