The present invention generally relates to effective determination of entity and industry reputations, and, more particularly, to systems and methods that provide realistic and accurate determination of the ability of an entity to deliver on stakeholder expectations through use of media data.
It is vital for companies, brands, and corporations to manage their reputation. The reputation of a corporation is a measure of how society views the corporation and provides a good measure of public expectation that the corporation has the basic ability to fulfill the expectations of a current or potential consumer. Reputation of a corporation bears weight on a number of important factors. For example, a strong reputation of a corporation will increase chances of obtaining and maintaining a loyal customer base, resulting in increased sales and being able to charge a premium for items sold. Strong corporate reputation also allows for a stronger current and potential employee pool. A strong reputation also increases chances that a potential purchaser will commit to a purchase, and a potential investor will commit to an investment.
Reputation is especially important in e-commerce, where products are being purchased online and the benefit of face to face encounter and interaction is uncommon. In addition, the e-commerce world removes the personal brick and mortar experience, making interaction during purchasing less personal, thereby making reputation even more important since live interaction cannot be relied upon to entice current and potential customers.
Reputation determination currently is inaccurately measured, using unrealizable data that tends to be biased and self-serving. A more reliable system and method is required to be provided for accurately measuring and managing reputation of entities and industries so as to allow for adjustment to improve reputation, thereby benefitting the entity, as well as enhancing consumer experience.
Embodiments of the present invention provide a system and method for determining reputation of an entity from at least one media data source. Referring to the method, the method comprises the steps of: using at least one of the group consisting of a text analysis model and a text mining model to determine if a sentiment toward an entity for which reputation is sought is positive, negative, or neutral by use of at least one media data point in at least one media data source, wherein a sentiment is an emotion about the entity portrayed by the at least one media data point, and assigning a sentiment numerical value to the at least one media data point based on whether the determined sentiment toward the entity is positive, negative, or neutral; determining at least one media reputation score of the entity for which reputation is sought, wherein the media reputation score is a measure of emotion portrayed about the entity in the media; training a classification model so that the classification model will associate at least one media data point with at least one media reputation driver, where a media reputation driver is a driver of reputation that is considered when assessing media reputation of the entity for which reputation is sought; using the classification model to associate at least one media data point with the at least one media reputation driver, or determine that the at least one media data point cannot be associated with any of the at least one media reputation drivers; and determining an entity media reputation driver score for each of the at least one media reputation drivers based on the at least one media data source.
Other systems, methods and features of the present invention will be or become apparent to one having ordinary skill in the art upon examining the following drawings and detailed description. It is intended that all such additional systems, methods, and features be included in this description, be within the scope of the present invention and protected by the accompanying claims.
The accompanying drawings are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present invention. The drawings illustrate embodiments of the invention and, together with the description, serve to explain the principals of the invention.
The present system and method provides a reliable measure of reputation for an entity or industry, and allows for reputation management in a manner that is most efficient, where specific adjustments that would be most effective toward increasing reputation are highlighted so as to provide the entity or industry with guidance for improving reputation. It should be noted that an entity may be any of, but not limited to, a corporation, company, individual, or a group of individuals functioning under one name or label.
The present system and method may be provided in the network 1 illustrated by the schematic diagram of
As shown by
Functionality as performed by the present reputation assessment and management system and method is defined by modules within the reputation server 100. The modules may be provided together as a reputation engine consisting of the modules, or in multiple locations within a single or more than one machine. For example, in hardware, the functionality of the modules can be implemented with any or a combination of the following technologies, which are each well known in the art: a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), etc. The modules can also be provided as software modules of a reputation engine, where the reputation engine comprises a processor and a memory having software modules therein defining functionality to be performed by the present system and method.
Referring to an embodiment where the reputation engine comprises a memory having software modules therein defining functionality to be performed by the present system and method, as shown by
The processor 102 is a hardware device for executing software, particularly that stored in the memory 106. The processor 102 can be any custom made or commercially available single core or multi-core processor, a central processing unit (CPU), a Graphics processing unit (GPU), an auxiliary processor among several processors associated with the present reputation server 100, a semiconductor-based microprocessor (in the form of a microchip or chip set), a microprocessor, or generally any device for executing software instructions.
The memory 106 can include any one or combination of volatile memory elements (e.g., random access memory (RAM, such as DRAM, SRAM, SDRAM, etc.) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CDROM, etc.). Moreover, the memory 106 may incorporate electronic, magnetic, optical, and/or other types of storage media. Note that the memory 106 can have a distributed architecture, where various components are situated remotely from one another, but can be accessed by the processor 102.
The software 110 defines functionality performed by the reputation 100, in accordance with the present invention. The software 110 in the memory 106 may include one or more separate programs, each of which contains an ordered listing of executable instructions for implementing logical functions of the reputation server 100, as described below. The memory 106 may contain an operating system (O/S) 170. The operating system 170 essentially controls the execution of programs within the reputation server 100 and provides scheduling, input-output control, file and data management, memory management, and communication control and related services.
The I/O devices 174 may include input devices, for example but not limited to, a keyboard, mouse, scanner, microphone, etc. Furthermore, the I/O devices 174 may also include output devices, for example but not limited to, a printer, display, etc. Finally, the I/O devices 174 may further include devices that communicate via both inputs and outputs, for instance but not limited to, a modulator/demodulator (modem; for accessing another device, system, or network), a radio frequency (RF) or other transceiver, a telephonic interface, a bridge, a router, or other device.
When the reputation server 100 is in operation, the processor 102 is configured to execute the software 110 stored within the memory 106, to communicate data to and from the memory 106, and to generally control operations of the reputation server 100 pursuant to the software 110.
When the functionality of the reputation server 100 is in operation, the processor 102 is configured to execute the software 110 stored within the memory 106, to communicate data to and from the memory 106, and to generally control operations of the reputation server 100 pursuant to the software 110. The operating system 170 is read by the processor 102, perhaps buffered within the processor 102, and then executed.
When functionality of the reputation server 100 is implemented in software 110, as defined by software modules within the memory 106, as will be described herein, it should be noted that instructions for implementing the reputation server 100 can be stored on any computer-readable medium for use by or in connection with any computer-related device, system, or method. Such a computer-readable medium may, in some embodiments, correspond to either or both the memory 106 or the storage device 104. In the context of this document, a computer-readable medium is an electronic, magnetic, optical, or other physical device or means that can contain or store a computer program for use by or in connection with a computer-related device, system, or method. Instructions for implementing the system can be embodied in any computer-readable medium for use by or in connection with the processor or other such instruction execution system, apparatus, or device. Although the processor 102 has been mentioned by way of example, such instruction execution system, apparatus, or device may, in some embodiments, be any computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “computer-readable medium” can be any means that can store, communicate, propagate, or transport the program for use by or in connection with the processor or other such instruction execution system, apparatus, or device.
Such a computer-readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory) (electronic), an optical fiber (optical), and a portable compact disc read-only memory (CDROM) (optical). Note that the computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via for instance optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner if necessary, and then stored in a computer memory.
As shown by block 202, data is obtained for determining reputation of an entity and that data is cleaned.
As shown by block 206, a determination is then made to ensure that those to be surveyed have a predefined level of familiarity with the entity for which reputation determination is sought. Reputation determination is more accurate when only those with at least the predefined level of familiarity with the entity itself are surveyed. Only respondents that pass a pre-defined threshold of familiarity are considered in the rating. For example, familiarity can be assessed on a scale of one to seven, and only those familiar at level of four and up are considered. From the familiarity score, the present system and method can also provide the entity with an awareness score for their reference and reputation management. The level of familiarity of those to be surveyed may be determined by asking questions that are specific to the entity for which the reputation determination is sought.
Since a surveyed person could potentially lie about their level of familiarity with an entity, a screening for unreliability test, is performed to flush out those likely to lie or exaggerate about their familiarity with an entity associated with the survey questions (block 208). For example, a questionnaire can be sent that lists corporations that do not exist and requesting a familiarity level from the receiver of the survey. If the individual surveyed claims to be familiar with one or more corporation that does not exist, the responder may be considered to be unreliable and they would then be removed from the demographically representative sample that is sized to provide a ninety-five or greater confidence interval.
The previously mentioned data cleaning steps of
Returning to
Referring to
In accordance with the present invention, the emotional connection survey questions transmitted fall into one of four categories. A first category of survey questions is esteem questions, which are questions intended to determine a level of esteem that the party surveyed associates with the entity to which a reputation measurement is desired. As an example, the survey question may ask a surveyed individual to rate from one to seven whether a company is a company that the surveyed individual gets a good feeling about, with a seven rating representing that a very good feeling is felt by the surveyed individual, and a one ratings representing that a very poor feeling is felt by the surveyed individual.
A second category of survey questions is admiration questions, which are questions intended to determine a level of admiration that the party surveyed associates with the entity to which a reputation measurement is desired. As an example, the survey question may ask a surveyed individual to rate from one to seven whether a company is a company that the surveyed individual admires and respects, with a seven rating being a high level of admiration and respect, and a one rating being a lowest level of admiration and respect.
A third category of survey questions is trust questions, which are questions intended to determine a level of trust that the party surveyed associates with the entity to which a reputation measurement is desired. As an example, the survey question may ask a surveyed individual to rate from one to seven whether a company is a company that the surveyed individual trusts, with a seven rating being a high level of trust, and a one being the lowest level of trust.
A fourth category of survey questions is feeling questions, which are questions intended to determine a level of positive feeling that the party surveyed associates with the entity to which a reputation measurement is desired. As an example, the survey question may ask a surveyed individual to rate from one to seven whether a company is a company that the surveyed individual feels has a good overall reputation, with a seven rating signifying a belief that the company has a very good overall reputation, and a one rating signifying a belief that the company has a very poor overall reputation.
While it is preferred that rating for each category of emotional connection survey be between one and seven, one having ordinary skill in the art would appreciate that a different scale may be used that is smaller than a zero to one-hundred scale. This will be apparent in the following description since a conversion is performed to change from a smaller scale to a zero to one-hundred scale, as described herein.
The result of this step is X number of returned first, second, third, and fourth category survey questions, where X is the number of individuals within the unique group. Therefore, if there are, for example, two-thousand individuals within the unique group (the highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity), then assuming that all who received the emotional connection survey questions responded with ratings, there will be two-thousand first category question ratings, two-thousand second category question ratings, two-thousand third category question ratings, and two-thousand fourth category question ratings, also referred to herein as the resulting emotional connection survey question ratings. The resulting emotional connection survey question ratings may be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
As shown by block 224, each of the resulting emotional connection survey question ratings returned are then converted by the reputation engine 108 from a raw scale of one to seven, to a zero to one hundred scale, to provide a rescaled score for each of the resulting emotional connection survey question ratings returned. This process may be performed by the reputation perception score module 130 using an equation such as, but not limited to the following equation 1.
Rescaled Score=((Raw Score−1)/6)×100 (Eq. 1)
The result of this step is X number of esteem reputation perception scores, X number of admiration reputation perception scores, X number of trust reputation perception scores, and X number of feeling reputation perception scores, where X is the number of individuals within the unique group. Each of the reputation perception scores may then be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
It has been determined that people in different countries, or geographical regions in general, tend to rate companies higher or lower resulting in an artificial skew in the rating distribution. As a result, and to ensure high accuracy in the final determined reputation of the reputation server 100, the present system and method overcomes this “cultural bias” in the data by standardizing all scores, per respondent, against the aggregate distribution of all scores stored (block 226). A standardization formula is applied in each market to ensure that scores are comparable across different markets, where different weights are applied to specific data based on the specific market. It should be noted that a “market” may be considered a specific geographical region, for example, a country, or a bigger or smaller geographical region. As an example, in Italy, the car brand Ferrari is likely to have an artificially high skew because the car is made in Italy. This would make survey results from those in Italy skewed on the high side. Therefore, the present system and method applies a standardization formula that takes into account a global mean and a global standard deviation across markets, and a country mean and a country standard deviation to normalize that which may have been skewed merely as a product of cultural pride or bias in general. As a result, a lower weight would be applied in Italy for the brand Ferrari than in other markets. Another example where a cultural weighting may be applied is when markets with a tendency to skepticism resulting in general lower opinions about corporations are compared with markets with an optimism tendency resulting in elevated opinions about corporations. In such cases, the cultural weighting removes cultural biases in rating and assures a standardized representation of reputation perceptions across different markets. The standardized formula may include equations 2 and 3 herein.
In equation 2 the country mean is the unweighted average score in the region. Preferably, the country mean is calculated periodically (e.g. every three years) from historical data of raw scores on the respondent level in each country. In addition, the country standard deviation represents the variability of scores among the respondents. Preferably, the country standard deviation is calculated periodically (e.g. every three years) from historical data of raw scores on the respondent level in each country. Preferably, the country mean and country standard deviation should be recalculated for a market, or region, within a predefined time period, for example, but not limited to, every three years, as previously mentioned, although it need not be every three years. This will ensure a high level of accuracy in the determined reputation score of the present system and method.
One having ordinary skill in the art would appreciate that while an example of a “market” has been provided as being a geographical region, a different measurable grouping may be used and the present invention is not intended to be limited to markets only being geographical regions.
During data collection from the unique group it is also beneficial to hit demographically representative quotas. As an example, common demographic groups that are targeted to ensure a market demographic representative sample for surveying may include age and gender. Unfortunately, during data collection it is not always possible to hit target demographic quotas exactly. As an example, a target demographic quota may be sixty percent women and forty percent men, however, a resulting highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity (i.e., unique group) may not comprise sixty percent women. In these cases, the reputation perception score module 130 applies a demographic weight to the data collected from the unique group, where the demographic weight is dependent upon the actual unique group itself, so as to ensure that the unique group results are representative of the population targeted (block 228). If the unique group is representative of the population targeted, then no demographic weight is applied.
In accordance with one exemplary embodiment of the invention, a Random Iterative Method (RIM) weighting algorithm may be used to demographically weigh, and therefore match, the returned sample of the unique group, which may have already been culturally weighted, to the demographics of the population to which the study is intended. The algorithm is repeatedly applied to the data until the demographic weight converges. It is noted that a RIM weighting method allows a more precise analysis than a proportional weighting approach, although either method may be used, as well as other weighting methods known to those having ordinary skill in the art.
In accordance with an alternative embodiment of the invention, it should be noted that other weighting of data may be performed to account for other biases, such as, but not limited to, a data sources weighing. This would be beneficial when it is known that a specific data source tends to have a more positively biased or negatively biased audience that is used for surveying. Data source weighting takes this into account and applies a weight based on the source so as to normalize results.
The results after the weighting steps (blocks 226 and 228) are X number of rescaled, from zero to one hundred, emotional connection survey question returned ratings that have been weighted for cultural bias and demographic bias, for each of the categories of esteem reputation perception scores, admiration reputation perception scores, trust reputation perception scores, and feeling reputation perception scores, where X is the number of individuals within the unique group. These results are then aggregated within each category to provide a single aggregated reputation perception score within each of the four categories and a single final reputation perception score is derived from averaging these four (4) scores (block 230). It should be noted that additional or fewer categories may be implemented.
The resulting single final reputation perception score is then categorized into a normative scale, preferably of five categories (block 232), although less or more categories may be used. It is found that the use of five categories is ideal. For example, the five categories may be weak reputation, poor reputation, average reputation, strong reputation, and excellent reputation. Specific ranges of values between the zero to one hundred range for the reputation perception score may be assigned to each of the five categories, for example based on quantiles of a normal distribution, so that the received reputation perception score for an associated entity may have deeper meaning.
The present system and method provide for a great level of granularity of data, in addition to providing the overall single final aggregated reputation perception score, the reputation perception scores within each of the four categories, and the categorization of the single final aggregated reputation perception score into a normative scale. Specifically, it is recalled that the X number of original emotional connection survey question ratings are saved, the X number of rescaled emotional connection survey question ratings are saved, the X number of rescaled emotional connection survey question ratings that have been weighted for cultural and demographic bias for each of the four categories have been saved, the single aggregated reputation perception score within each of the four categories have been saved, the resulting single final reputation perception score has been saved, and the normative scale associated with the single final aggregated reputation perception score has been saved. This level of granularity in the data is very beneficial to the entity for which reputation determination is desired.
Returning to
Preferably there are more than twenty survey questions with the second set of focused survey questions. It is noted, however, that there may be more or fewer such survey questions within the set of practical thinking survey questions, depending upon a level of granularity needed to address aspects of drivers desired.
In accordance with the present invention, the second set practical thinking survey questions transmitted fall into one of seven categories. These categories are referred to herein as reputation drivers. Specifically, it was found that there are typically seven areas that people tend to care about when assessing the reputation of an entity. Those areas are referred to herein as reputation drivers, and include, for example, products and services, innovation, workplace, governance (also called conduct), citizenship, leadership, and performance.
The first reputation driver is Products & Services, which provides a perception of the general public on the quality and value of the entity's offerings and customer care. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity offers high quality products and services, with a seven rating representing strong agreement, and a one rating representing strong disagreement felt by the surveyed individual.
The second reputation driver is Innovation, which addresses the perception of the company being innovative in its offerings, first to market and adaptable to change. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity is an innovative company, with a seven rating representing strong agreement that the entity is an innovative company, and a one rating representing strong disagreement felt by the surveyed individual.
The third reputation driver is Workplace, which addresses the wellbeing of employees, Diversity, Equity and Inclusion, and workplace satisfaction. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity offers equal opportunity in the workplace, with a seven rating representing strong agreement that the entity does offer equal opportunity in the workplace, and a one rating representing strong disagreement felt by the surveyed individual.
The fourth reputation driver is Governance (also called Conduct), which addresses ethics, transparency and corporate responsibility. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity is fair in the way it does business, with a seven rating representing strong agreement that the entity is fair in the way it does business, and a one rating representing strong disagreement felt by the surveyed individual.
The fifth reputation driver is Citizenship, which addresses the company's contribution to making the world better by supporting good causes and contributing to the community. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity supports good causes, with a seven rating representing strong agreement that the entity does support good causes, and a one rating representing strong disagreement felt by the surveyed individual.
The sixth reputation driver is Leadership, which represents perceptions on entity's leadership and clear direction. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity has excellent managers, with a seven rating representing strong agreement that the entity does have excellent managers, and a one rating representing strong disagreement felt by the surveyed individual.
The seventh reputation driver is Performance, which addresses the financial performance of an entity and future growth prospects. As an example, the survey question may ask a surveyed individual to rate from one to seven whether the entity is a profitable company, with a seven rating representing strong agreement that the entity is a profitable company, and a one rating representing strong disagreement felt by the surveyed individual.
While it is preferred that the answered rating for each practical thinking survey question be between one and seven, one having ordinary skill in the art would appreciate that a different scale may be used that is smaller than a zero to one-hundred scale. This will be apparent in the following description since a conversion is performed to change from a smaller scale to a zero to one-hundred scale, as described herein.
The result of this step is X number of returned responses for each practical thinking survey question or factor, where X is the number of individuals within the unique group. Therefore, if there are two-thousand individuals within the unique group (the highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity), then assuming that all who received the practical thinking survey questions responded with ratings, there will be two-thousand responses for each practical thinking survey question, also referred to herein as the resulting practical thinking survey question rating. The resulting practical thinking survey question ratings may be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
As shown by block 246, each of the resulting practical thinking survey question ratings returned are then converted by the reputation engine 108 from a raw scale of one to seven, to a zero to one hundred scale, to provide a rescaled score for each of the resulting practical thinking survey question rating returned. This process may be performed by using an equation such as, but not limited to, the previously mentioned equation 1.
The result of this step is X number of factor scores for each practical thinking survey question, where X is the number of individuals within the unique group. Each of the factor scores may then be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
To ensure high accuracy in the final determined reputation of the reputation server 100, the present system and method overcomes “cultural bias” in the data by standardizing all factor scores against the aggregate distribution of all factor scores stored (block 248). A standardization formula is applied in each market to ensure that factor scores are comparable across different markets, where different weights are applied to specific data based on the specific market. As previously mentioned, it should be noted that a “market” may be considered a specific geographical region, for example, a country, or a smaller geographical region. The standardization formula takes into account a country mean and a country standard deviation to normalize that which may have been skewed merely as a product of cultural pride or bias in general. The standardized formula may include the previously mentioned equations 2 and 3. Preferably, the country mean and country standard deviation should be recalculated for a market, or region, within a predefined time period, for example, but not limited to, every three years. This will ensure a high level of accuracy in the determined factor score of the present system and method.
As previously mentioned, during data collection from the unique group it is also beneficial to hit demographically representative quotas. As an example, common demographic groups that are targeted to ensure a market demographic representative sample for surveying may include age and gender. Unfortunately, during data collection it is not always possible to hit target demographic quotas exactly. As an example, a target demographic quota may be sixty percent women and forty percent men, however, a resulting highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity (i.e., unique group) may not comprise sixty percent women. In these cases, the reputation driver score module 140 applies a demographic weight to the data collected from the unique group, where the demographic weight is dependent upon the actual unique group itself, so as to ensure that the unique group results are representative of the population targeted (block 250). If the unique group is representative of the population targeted, then no demographic weight is applied.
Also as previously mentioned, in accordance with one exemplary embodiment of the invention, a RIM weighting algorithm may be used to demographically weigh, and therefore match, the returned sample of the unique group, which may have already been culturally weighted, to the demographics of the population to which the study is intended. The algorithm is repeatedly applied to the data until the demographic weight converges. It is noted that a RIM weighting method allows a more precise analysis than a proportional weighting approach, although either method may be used, as well as other weighting methods known to those having ordinary skill in the art.
In accordance with an alternative embodiment of the invention, it should be noted that other weighting of data may be performed to account for other biases, such as, but not limited to, a data sources weighing. This would be beneficial when it is known that a specific data source tends to have a more positively biased or negatively biased audience that is used for surveying. Data source weighting takes this into account and applies a weight based on the source so as to normalize results.
The results after the weighting steps (blocks 248 and 250) are X number of rescaled, from zero to one hundred, practical thinking survey question returned ratings that have been weighted for cultural bias and demographic bias, for each of the factors, where X is the number of individuals within the unique group. These results are also referred to herein as weighted factor scores. The X number of resulting weighted factor scores for each factor are then aggregated within each factor to provide a single aggregated factor score for each factor (block 252). The result is one resulting reputation factor score for each factor.
Returning to
Referring to
Unsupervised learning clustering is then used by the reputation driver score module 140 (
To determine reputation driver scores the reputation driver score module 140 (
As a result of this step, the reputation server 100 now contains the previously determined reputation score, the multiple resulting factor scores, and the reputation driver scores. Returning to
To determine the driver weights, the driver score module 140 uses a supervised machine learning regression module, such as, but not limited to, linear regression, multivariate linear regression, random forest, or gradient boosting, to predict or inference the reputation score from the driver scores, where for the dependent variable, which is the variable we are seeking to predict, the reputation score is used, and for the input variable the driver scores are used. The result is the weight for each of the drivers, which allows for determining importance based on values of the weights. Specifically, a lower determined weight for a driver demonstrates that the specific driver is less important to the overall reputation of the entity. This allows the entity to determine which drivers are most important to increasing reputation of the entity. For ease of use, the resulting driver weights may be normalized so that each weight is a specific percentage of a total of one hundred. For example, out of seven weights, a first weight may be five percent, a second weight twenty percent, a third weight ten percent, a fourth weight thirty percent, a fifth weight five percent, and sixth weight eight percent, and a seventh weight twenty two percent. This would demonstrate that the fourth weight is the most important to the overall reputation of the entity, and therefore, the most value would be gained by the entity with investing time and effort into the fourth weight.
A similar weighting process as described above may be performed for weighting the reputation factors within each driver to derive for the entity a more granular level of insights on their reputation management.
A resulting report for an entity seeking its determined reputation and seeking to manage its reputation could include, per period of time, its overall reputation score and a list of the drivers and associated importance weights so as to provide guidance on which areas to invest additional time and money for maximum increase in reputation. Additional data may also be provided, such as, but not limited to, the factors scores and weights, benchmarking to competitors, industry classification, benchmarking to industry, touch-points between the stakeholder and the entity, media data and scores, ESG (environmental, social, and governance) perception scores, and various business outcomes scores (also referred to herein as behavioral connection scores). This data and the trend lines, benchmarking analysis, machine learning results, and reputation management insight and guidance related to reputation management may be delivered to the entity via a digital platform, via one or more means, such as, but not limited to, Power point, excel files, and one or more of many other means.
The memory 106 also contains a behavioral connection module 150. The behavioral connection module 150 is used to determine behavioral connection scores. Behavioral connection scores are a measure of the likelihood that the targeted population will perform a positive action on behalf of an entity.
The transmitted behavioral connection survey questions measure the likelihood that a member of the unique group will perform a positive action on behalf of the entity for which reputation is being measured. This may also be considered as the Business Outcomes of Reputation Management. Non-limiting examples of survey questions focused on determining a measure of the likelihood that the public will perform a positive action on behalf of an entity include questions asking how likely is it that the surveyed individual would recommend the entity, how likely it is that the surveyed individual would work for the entity, how likely it is that the surveyed individual would buy from the entity, how likely it is that the surveyed individual would invest in the entity, and more.
It is to be recalled that the unique group is the result of the data cleaning step, so that behavioral connection survey answers from the unique group are extremely similar (within a 95% confidence interval) to what would be returned by an entire desired population. More specifically, if for example a unique group is two thousand individuals, survey results from that unique group would be extremely similar (within 95% confidence interval), if not the same (with only a 5% margin of error) as that which would have been received if an entire population of more than one million people for which the survey was intended, were in fact surveyed.
While a specific number of behavioral connection survey questions is not required, it is preferred that there be about ten of such behavioral connection questions, each representing a particular desired business outcome, for example, 11 survey questions within the third set of focused survey questions. It is noted, however, that there may be more or fewer such survey questions within the set of behavioral connection survey questions, depending upon a level of granularity needed.
It is preferred that the answered ratings for each behavioral connection survey question be between one and seven, although one having ordinary skill in the art would appreciate that a different scale may be used that is smaller than a zero to one-hundred scale. This will be apparent in the following description since a conversion is performed to change from a smaller scale to a zero to one-hundred scale, as described herein.
The result of this step is X number of returned responses for each behavioral connection survey question or factor, where X is the number of individuals within the unique group. Therefore, if there are two-thousand individuals within the unique group (the highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity), then assuming that all who received the behavioral connection survey questions responded with ratings, there will be two-thousand responses for each behavioral connection survey question, also referred to herein as the resulting behavioral connection survey question ratings. The resulting behavioral connection survey question ratings may be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
As shown by block 304, each of the resulting behavioral connection survey question ratings returned are then converted by the reputation engine 108 from a raw scale of one to seven, to a zero to one hundred scale, to provide a rescaled score for each resulting behavioral connection survey question rating returned. This process may be performed by using an equation such as, but not limited to, the previously mentioned equation 1.
The result of this step is X number of behavioral connection scores for each behavioral connection survey question, where X is the number of individuals within the unique group. Each of the behavioral connection scores may then be stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1.
To ensure high accuracy in the final determined reputation of the reputation server 100, the present system and method overcomes “cultural bias” in the data by standardizing all behavioral connection scores against the aggregate distribution of all behavioral connection scores stored (block 306). A standardization formula is applied in each market to ensure that behavioral connection scores are comparable across different markets, where different weights are applied to specific data based on the specific market. As previously mentioned, it should be noted that a “market” may be considered a specific geographical region, for example, a country, or a smaller geographical region. The standardization formula takes into account a country mean and a country standard deviation to normalize that which may have been skewed merely as a product of cultural pride or bias in general. The standardized formula may include the previously mentioned equations 2 and 3.
Preferably, the country mean and country standard deviation should be recalculated for a market, or region, within a predefined time period, for example, but not limited to, every three years. This will ensure a high level of accuracy in the determined behavioral connection score of the present system and method.
As previously mentioned, during data collection from the unique group it is also beneficial to hit demographically representative quotas. As an example, common demographic groups that are targeted to ensure a market demographic representative sample for surveying may include age and gender. Unfortunately, during data collection it is not always possible to hit target demographic quotas exactly. As an example, a target demographic quota may be sixty percent women and forty percent men, however, a resulting highly reliable demographically representative sample of a survey population that is sized to provide a ninety-five percent or greater confidence interval, from people who are familiar with the surveyed entity (i.e., unique group) may not comprise sixty percent women. In these cases, the behavioral connection score module 150 applies a demographic weight to the data collected from the unique group, where the demographic weight is dependent upon the actual unique group itself, so as to ensure that the unique group results are representative of the population targeted (block 308). If the unique group is representative of the population targeted, then no demographic weight is applied.
Also as previously mentioned, in accordance with one exemplary embodiment of the invention, a RIM weighting algorithm may be used to demographically weigh, and therefore match, the returned sample of the unique group, which may have already been culturally weighted, to the demographics of the population to which the study is intended. The algorithm is repeatedly applied to the data until the demographic weight converges. It is noted that a RIM weighting method allows a more precise analysis than a proportional weighting approach, although either method may be used, as well as other weighting methods known to those having ordinary skill in the art.
In accordance with an alternative embodiment of the invention, it should be noted that other weighting of data may be performed to account for other biases, such as, but not limited to, a data sources weighing. This would be beneficial when it is known that a specific data source tends to have a more positively biased or negatively biased audience that is used for surveying. Data source weighting takes this into account and applies a weight based on the source so as to normalize results.
The results after the weighting steps are X number of rescaled, from zero to one hundred, behavioral connection survey question returned ratings that have been weighted for cultural bias and demographic bias, for each of the questions, where X is the number of individuals within the unique group. These results are also referred to herein as weighted behavioral connection scores. The X number of resulting weighted behavioral connection scores for each question are then aggregated within each individual question to provide a single aggregated behavioral connection score for each behavioral connection survey question (block 310). The result is one resulting behavioral connection score for each behavioral connection survey question. For example, if there are ten (10) behavioral connection survey questions, the result after the previously mentioned steps is one aggregate resulting behavioral connection score for each of the ten behavioral connection survey questions.
A resulting report for an entity seeking its determined reputation and seeking to manage its reputation could include, per period of time, its aggregate resulting behavioral connection score so as to provide guidance on which areas to invest additional time and money for maximum increase in reputation.
At this point, resulting data includes reputation scores, driver scores per driver, factor scores per factor, and behavioral connection (also called business outcome) scores per behavioral connection question, as well as the overall reputation score and a list of the drivers and associated importance weights (prioritizing the importance of each driver for the entity's reputation). In accordance with the present invention, additional actionable insights may be presented to the entity seeking its reputation score. Specifically, for each behavioral connection score, the present system and method can prioritize the importance of drivers and/or factors for predicting or inferencing each of the behavioral connections between the public and the entity for which reputation is sought (block 312). This allows the entity to be better able to determine what things, as identified by drivers and factors, to focus on in order to increase behavioral connection between the public represented by the unique group and the entity.
Prioritization of the importance of drivers for predicting or inferencing each of the behavioral connections can be performed by using a supervised machined learning regression model, such as, but not limited to, linear regression, multivariate linear regression, random forest, or gradient boosting, to predict or inference a known behavioral connection score from the driver scores, where the known behavioral connection score may be one previously derived as already described, or one received remotely or stored within the reputation server 100 storage device 104, or within one of the other databases 30a, 30b, 30c within the network 1. As a non-limiting example, in using a linear regression or a random forest regression, the behavioral connection question may be “are you likely to purchase a product from the entity”, and the scores, for instance from a two-thousand scores from a unique group of two-thousand individuals, may be the scores derived for this behavioral connection question after calculations and weightings. Another example of implementation would be using aggregated scores of reputation perception and drivers as inputs per each dependent set of behavioral connection aggregated scores, for example, 12 or 24 for each entity. Ideally, for increased accuracy and for detecting evolution of weights over time, the regression model is used tens or even hundreds of times, to derive an unknown weight, or hyperparameter, for each driver. Based on the weight derived for each driver, or a perception score, all perception scores (including reputation score, reputation drivers scores, ESG scores, brand scores, etc.) that are predictors of a single behavioral connection are prioritized by numerically sorting the perception drivers and other aggregated perception scores by weight value, with the largest perception score weight representing the perception score that has the greatest effect on the behavioral connection score (also called business outcome score). The prioritized perception scores weights may then be normalized to a zero to one-hundred percent range. An example of a result may be for a first behavioral connection question a first driver weight of forty percent, a second driver weight of ten percent, a third driver weight of fifteen percent, a fourth driver weight of twelve percent, and so on. This demonstrates that the first perception score, for example the score of the reputation driver ‘Products and Services’ has the greatest effect on improving the first behavioral connection score. The prioritization of the most important drivers and other perception scores for each of the business outcomes provide information to the entity (company or industry) on areas to invest in, and which activities to prioritize, to increase a certain business outcome, or to target a certain population of stakeholders.
The process for prioritizing the importance of factors for predicting or inferencing each of the behavioral connections can be performed in the same manner as the process used for prioritizing the importance of perception drivers, as previously mentioned, with supplementing the factors for the drivers. The result of this process is having prioritized drivers and prioritized factors that are important to predicting a specific behavioral connection. In accordance with the present system and method, the behavioral connection module 150 can then use supervised machine learning classification models to predict each of the behavioral connection scores from historical data, namely, the combination of already derived scores, including the reputation scores, factor scores, and driver scores, and possibly additional scores such as perception ESG (environmental, social, governance) scores, brand scores, etc.
In implementation, for example, block 352 works by training of a single supervised machine learning model for a single behavioral connection question first using as the input variables the reputation scores, factor scores, and driver scores, and as output dependent variable a single behavioral connection score for a responding unique group, with or without a time lag, for example of one month. This process may be performed for one entity or multiple entities for a set of unique groups. As a second step of block 352, this same process is then performed for the same set of entities and the same set of responding unique groups, using the same input variables, but using a second of the behavioral connection scores, which is associated with a second behavioral connection score provided by that set of unique groups for one or multiple entities. As a third step of block 352, this process is performed for a set of unique groups per one or multiple entities, until a supervised machine learning model is trained for each of the behavioral connection scores, resulting in a set of supervised machines learning trained models, per each of the behavioral connection business outcomes. When training a supervised machine learning model per one entity, for one behavioral connection business outcome, the model training may be performed tens or even hundreds of times to increase the accuracy of the model by tuning the hyperparameters. The first, second, and third steps are then repeated for a second entity, and so on, until it has been performed for all sets of respondent groups, per entity, resulting in Y sets of supervised machine learning trained models, each per entity. Therefore, if there are for example, one hundred entities, the result would be one hundred sets of supervised machine learning trained models, each per entity. When training a supervised machine learning model per unique group, or per a set of unique groups, for multiple entities, the first, second, and third steps are then repeated for a second unique group, and so on, until it has been performed for all unique groups, resulting in Z sets of supervised machine learning trained models, each per a unique group. Similarly, therefore, if there are, for example, ten sets of unique groups, the result would be ten sets of supervised machine learning trained models, each per a unique group. Another example of implementation may be training one machine learning model, per at least one behavioral connection business outcome, using at least one unique group and at least one entity. As previously mentioned each of these machine learning models may be trained with or without a time lag, where a time lag may be, as a nonlimiting example, of one month. The trained models are then saved within the reputation server 100 (block 354). The trained models are then may be used to predict each of the behavioral connections from an input set of perception scores to help entities manage their reputation and business outcomes.
A non-limiting example of block 352 in
Training of the models results in obtaining the relationship between the input variables and the output variables, also referred to as hyperparameters, or weights. Since the relationship between the input variables and output variables is known through having determined the hyperparameters, the present system and method can predict behavioral connection business outcome scores from newly received input variables associated with a new respondents' group who has responded to survey questions necessary to provide the seven driver scores, twenty factor scores, and one reputation score. Again, this is possible because the relationship between input variables and output variables is known, via the trained models, resulting in the hyperparameters, which can then be used when newly received input variables are provided, so as to predict resulting business outcome scores for the new respondent group associated with the newly received input variables, which are the scores calculated form all the respondents in the unique group in a time period, such as one month.
As a result of the abovementioned, after the supervised machine learning models are trained, the present system and method can predict business outcome scores for a new group of respondents as resulting from the input variables of the new respondents' group, including the driver scores, factor scores, and reputation score. This, in turn, decreases the number of survey questions that would need to be presented to a respondent, which decreases investment cost associated with surveys such as, but not limited to, fielding surveys focused on business outcomes, processing responses to the fielding, time associated with the fielding and processing, and more. As these machine learning models may be trained with a time lag, for example of one month, the behavioral connection business outcomes expected in a future time, for example next month, may be predicted from the input variables of reputation, drivers and factors, preparing the entity for expected stakeholder behaviors in the near future. The predicted behavioral connection value may be, but is not limited to, a score, a confidence interval, a range of scores, a quantile to which the score may belong, a bucket on a normative scale (for example, poor, weak, average, strong, or excellent), if the score is expected to be in the top 75% or below, or simply if the score is expected to go up or down next month as compared to the previous month.
At this point, resulting data includes, among other things, reputation scores and behavioral connection scores per business outcome (for example for 10 business outcomes), as well as the overall reputation score overtime trend and behavioral connections over time trends per each of the behavioral connection scores. With this information, correlations may be derived between the set of reputation scores over a period of time, for example 12 monthly scores, and each of the behavioral connection scores for the same period of time (for example, 12 monthly scores). In accordance with the present invention, additional actionable insights may be presented to the entity seeking its reputation score. Specifically, for each behavioral connection score, the present system and method can calculate the correlation, for example one may use the Pearson correlation, or Spearman correlation, between each of the behavioral connections set of scores over a period of time (for example 12 months) and between the reputation of the entity for which reputation is sought (block 312) over this period of 12 months. Calculating the correlation between the set of reputation scores over a period of time (for example 12 months) that were collected for a particular entity from a set of unique groups (for example 12) and a set of scores of a particular behavioral connection over the same period of time and the same set of unique groups would involve for example calculating the R2 value, where a value of R2 above a defined threshold (for example 0.8) would indicate a high correlation. As statistical methods of calculating the correlation between two set of numbers, or two trends, are known to those having ordinary skill in the art, we will not describe here further how a correlation may be derived. For purposes of the example, we can assume that the behavioral connection question is “how likely is it that you will purchase a product from this entity?” Of course, this is only an example and many other behavioral connection questions may be used. This allows the entity to better understand the correlation between its reputation over a set period of time (for example 12 months) and each of the business outcomes this entity is seeking. Thus, further understanding and being able to manage the influence of the reputation of the entity and the behavioral connection between the public represented by a certain unique group, or a set of unique groups, and the entity.
The prior description has been provided for determining reputations through use of surveys. In addition to use of surveys, the present invention may instead determine reputation from media and additional sources of data. Unlike in use of surveys, when determining reputation from media, questions cannot easily be asked and answered. Examples of media that may be used for determining reputation in accordance with the present invention include, but are not limited to, electronic news feeds, blogs, social media, video media, and much more.
For calculating reputation scores and weights from media data (and additional enrichment data sources), the present system and method uses Natural Language Processing (NLP) methods of machine learning (ML) and artificial intelligence (AI). One having ordinary skill in the art would appreciate that other analytics and data science methods, such as audio to text transcription, information extraction methods, and classic methods of text analytics or text mining, may instead be used by the present system and method to translate media into searchable text and derive quantities and measurements. In other implementations of the present invention, one may also use image recognition, voice recognition or video analysis, and sentiment analysis from audio, images, and video may be deployed.
In the current invention various information extraction and entity extraction NLP methods may be used. As a non-limiting example, one may use Named Entity Recognition (NER) to extract entity names from unstructured text and Keyword in Context (KWIC) methods to create summarized or short text extract around the extracted entity name. While other methods, both for entity extraction and short text summaries around an entity name or a keyword may be used, the present description focuses on using these as an example of the implementation.
One having ordinary skill in the art would appreciate that an entity-based sentiment is an emotion or feeling about the entity portrayed in media. A non-limiting example of such an emotion demonstrated within a media data may be that entity X is a “great company”, which would be a positive entity-based sentiment about entity X. Another non-limiting example of an emotion demonstrated within a media data may be that entity Y is “struggling to meet demands of its stakeholders”, which would be a negative entity-based sentiment about entity Y. Usually entity-based sentiment is measured per an at least one media data point, where some non-limiting examples of media data points may be an article, a news post, a blog, a social media post, a video, a podcast, or any other piece of information posted or shared via the media.
Media reputation driver scores are then determined using the entity-based sentiment. To do so, an NLP machine learning model is trained, as shown by block 550, after which driver classification of associated KWIC data is predicted (also called inferenced or inferred), as shown by block 600. As shown by block 630, entity media reputation driver scores are then determined. Media reputation driver weights are then determined, as shown by block 670. As shown by block 700, the media reputation scores and the media reputation driver scores are processed by a normative scale alignment algorithm that normalizes and aligns the entity media reputation scores and the entity media reputation driver scores to a normative scale of poor, weak, average, strong, excellent. The normative scale above is just an example, and one having ordinary skill in the art would appreciate that other normative scales may be used as well.
The NLP model is trained using sentiment-labeled text extracts or segments surrounding a desired keyword (for example, an entity name). As a nonlimiting example, the text segments surrounding a keyword may be referred to as key word in context (KWIC), but other short text around a keyword or a summarized text with keywords may be considered similarly. The KWIC data is annotated with a sentiment label, for example of 1 for positive sentiment, −1 for negative sentiment, and 0 for a neutral sentiment, and stored. One having ordinary skill in the art would appreciate that other labeling of sentiment, using numbers or characters, may be used as well. The NLP model is then trained using the sentiment annotated KWIC data, and the trained model is then stored.
As shown by block 408, the acquired data is then cleaned and enriched by the data cleaning module 120 (
Enrichment of the data is performed to increase robustness of data that will be made available through training of the model, for predicting entity-based sentiment, and for determining media reputation scores. Enrichment is performed by determining names of other parties that are related to the entity for which reputation is being sought so that those names may also be searched for within the acquired data. In addition, other names or acronyms for the entity may be searched such as a dba, and names of brands or products of the entity may be searched. For example, if corporation X is the entity for which reputation is sought, names of the president and CEO of corporation X may also be searched for within the acquired data, thereby making the acquired data more robust because this data is related to the entity for which reputation is sought. Enrichment data may also include trademarks of the entity, brand names of the entity, product names of the entity, and more, as long as the enrichment data is related to the entity for which reputation is sought.
Location of the entity name for which the reputation is sought, within the acquired data, is then determined, as shown by block 410. Specifically, the cleaned data is searched by the data acquisition module 122 to find location of the entity name for which reputation is sought.
The data acquisition module 122 (
Location of the enrichment data associated with the entity for which the reputation is sought, within the acquired data, is then determined, as shown by block 416. Specifically, the enrichment data is searched by the data acquisition module 122 to find location and content of the enrichment data.
The data acquisition module 122 (
The saved KWIC data, inclusive of the entity name KWIC data and the saved enrichment data KWIC data, is what is used by the reputation engine 108 (
The first eight steps of NLP model training may be performed daily, weekly, monthly, quarterly, or on another cycle so as to ensure that the KWIC data, inclusive of the entity KWIC data and the enrichment data KWIC data, is current.
As shown by block 422, the reputation engine 108 (
Returning to
As previously mentioned,
The keywords used during annotation are words that positively describe the entity, or the reputation drivers, or the reputation factors of the reputation drivers and separately, that negatively describe the entity, or the reputation drivers, or the reputation factors of the reputation drivers. Additional keywords that are used during the annotation process, for training a model to classify media data into media reputation drivers, are keywords that would associate a media data point as related to at least one reputation driver, or to at least one reputation factor of a reputation driver. These keywords may be stored within the storage device 104 (
Since additional keywords within a media data point may be associated with a specific reputation factor or reputation driver, finding a specific keyword within KWIC data also allows for labelling of the KWIC data with the specific reputation factor or driver associated with the found factor or driver keyword, thereby allowing for labelling of the KWIC data with one or more specific reputation factors or drivers.
As shown by block 434, after the annotation job instructions are written, so as to allow the media data management module 160 (
As shown by block 438, a determination is then made by the media data management module 160 (
With the NLP model trained and saved, the KWICs stored, and a sentiment value assigned to each stored KWIC of one, zero, or minus one, the media data management module 160 (
As previously mentioned,
Presence and location of the entity name for which the reputation is sought, within the acquired data, is then determined, as shown by block 456. Specifically, the cleaned data is searched by the data acquisition module 122 to extract the entity name and find location of the entity name for which reputation is sought, using for example NER (named entity recognition) or other text mining information extraction or search algorithms. The data acquisition module 122 (
Location of the enrichment data associated with the entity for which the reputation is sought, within the acquired data, is then determined, as shown by block 462. The data acquisition module 122 (
As shown by block 468, the new entity name KWIC data and new enrichment data KWIC data are then input into the trained NLP model resulting in the new KWICs receiving a positive sentiment value of +1, a negative sentiment value of −1, or a neutral sentiment value of 0, so that each new KWIC has an associated sentiment value. Of course, as previously mentioned, other numerical or alphabetical labels may be chosen to label positive, negative, and neutral, and a finer scale of positivity and negativity may be set. Here for simplicity we use as an example labeling of 1 for positive, −1 for negative, and 0 for neutral.
The step of calculating and storing media reputation scores (block 480), of
As shown by block 488, the raw media reputation score of the entity then has a weighting algorithm applied to it, where the weighting is based on text sources, volume, reach, and impact so as to allow for greater consideration of KWICs from sources that are potentially more important, having greater influence, and/or having a stronger reputation or validity within the industry. One having ordinary skill in the art would know different techniques for weighting data or data sources, and therefore, while the following provides one example of a method that may be used to weigh KWICs, it should be noted that this is not intended to limit the present invention to this method of weighting. It is noted that it may not be necessary to apply a weighting algorithm to the raw media reputation score. This step depends upon whether there is a desire to weight certain sources differently.
It will be recalled that the storage device 102 (
In accordance with an alternative embodiment of the invention, a quality step may be added to ensure that redundancy in storage and sentiment consideration of KWICs does not adversely affect the media data reputation score. More specifically, it should be noted that certain KWICs may contain both an entity name and enrichment data. The process of finding, capturing, and storing entity name KWICs and separately enrichment data KWICs may result in the entity name KWIC and enrichment data KWIC being the same KWIC. This is the purpose of the quality step. In accordance with the alternative embodiment of the invention, the media data management module 160 determines if there are redundantly stored KWICs, and if so, it determines if the sentiment value for the redundantly stored KWICs are the same. If the KWIC is redundantly stored and the sentiment value of the redundantly stored KWICs is the same, one of the KWICs is deleted. Alternatively, the KWIC may be redundantly stored and the sentiment value of the redundantly stored KWICs may differ. If this happens a mediating algorithm is usually invoked to determine the most correct sentiment value of the duplicated KWIC, and in some cases even a human quality assurance (QA) process may be initiated if necessary, and only one copy of the redundant KWIC with the final determined sentiment value is stored.
Weighting of the media reputation raw score may be performed in one of many different ways. For exemplary purposes only, the following describes a three-step process that may be used for weighting. In a first step, prominence weight features are created per KWIC, so that a weight may be applied for KWIC sources that are potentially more important, having greater influence, and/or having a stronger reputation or validity within the industry. For nonlimiting exemplary purposes, the following equations 7-9 may be used in this calculation.
Where, for example, a1=numerical value<1 (for example 0.5) and a2=1.
Then, additional weighting steps may be added. For example, a second step, where a Weighted Sentiment is created for each KWIC as:
Weighted Sentiment=*(Impact*Prominence Weight)*Sentiment (Eq. 8)
Where, for example, a3<1.
Then, for example, as a third and final step, the weighted media reputation score for an entity in a given time period, such as, but not limited to, a week or month, may be defined as:
As shown by block 490, the weighted media reputation score is then aligned to a normative scale. An example of a process that may be used to normalize the resulting weighted media reputation score and align it to a normative scale includes, but is not limited to, applying logarithmic transformation in the form of equation 10.
Using the distributions of perception reputation scores collected over a period of time in the reputation storage device, and media reputation scores calculated with the trained model and the process described above, an iterative process is run to tune the parameters of equation six to normalize the media reputation score to the normative scale. The result is a final zero to one-hundred entity media reputation score. The entity media reputation score may then be stored to the storage device 104 (
Similar to the perception reputation drivers, when using media data for assessing the reputation of an entity, there are a series of drivers of reputation that are used and considered in determining the media reputation of the entity for which reputation is sought from media data. In accordance with one exemplary embodiment of the invention, there may be seven drivers of reputation used in calculating media data reputation scores, including, for example, products and services, innovation, workplace, conduct (also called governance), citizenship, leadership, and performance. These reputation driver names are for illustration purposes only and synonyms, other or additional reputation drivers or reputation driver names may be used in a similar manner. The media reputation drivers' scores may be calculated on a predefined cadence or time-interval, such as, but not limited to, daily, weekly, monthly, or quarterly, or even at real time, or near real time.
For determining the media reputation driver scores using the entity-based sentiment, a text analytics or a text mining model is used. As a nonlimiting example we describe in this embodiment the use of a machine learning deep learning NLP model, but it is important to mention that in other implementations of the current invention other text analysis models may be used such as other supervised machine learning models, unsupervised machine learning models such as Latent Dirichlet Allocation (LDA), other clustering models, a semi-supervised model, a combination of unsupervised and supervised models, or other text analytics and text mining methods.
As previously mentioned, for determining the media reputation driver scores using the entity-based sentiment, an NLP machine learning model is trained, as shown by block 550 of
Location of the entity name for which the reputation is sought, within the acquired data, is then determined, as shown by block 556. Specifically, the cleaned data is searched by the data acquisition module 122 to find location of the entity name for which reputation is sought. The data acquisition module 122 (
Location of the enrichment data associated with the entity for which the reputation is sought, within the acquired data, is then determined, as shown by block 562. The data acquisition module 122 (
As shown by block 568, data annotation is then performed for reputation driver labeling, which is similar to the data annotation previously described for reputation factors, except now being provided for reputation drivers.
The keywords used during annotation for driver labelling are words associated with a specific reputation driver, where the word is related to at least one reputation driver and in its vicinity there may be a word that either positively describes the entity when considering the at least one reputation driver, or negatively describes the entity when considering the at least one reputation driver. Each reputation driver individually has specific keywords that are used to classify a media data point to the associated media reputation driver, and to determine whether the keyword related to each reputation driver was mentioned positively or negatively to describe the entity. These keywords may be stored within the storage device 104 (
As shown by block 584, after the annotation job instructions are written, so as to allow the media data management module 160 (
As shown by block 588, a determination is then made by the media data management module 160 (
Alternatively, if the F1 score does not satisfy the predefined threshold value, the media data management module 160 (
Returning to
With the NLP model already trained, reputation driver classification of associated KWIC data is predicted (which may also be called inferenced or inferred), as further illustrated by
Location of the entity name for which the reputation is sought, within the acquired data, is then determined, as shown by block 606. Specifically, the cleaned data is searched by the data acquisition module 122 to find location of the entity name for which reputation is sought. The data acquisition module 122 (
Location of the enrichment data associated with the entity for which the reputation is sought, within the acquired data, is then determined, as shown by block 612. The data acquisition module 122 (
As shown by block 618, the new entity name KWIC data and new enrichment data KWIC data are then input into the trained NLP model resulting in the new KWICs receiving a positive sentiment value of +1, a negative sentiment value of −1, or a neutral sentiment value of 0, so that each new KWIC has an associated sentiment value. The new KWIC data is then classified by the trained and stored NLP machine learning model by predicting (or inferring) the label of each KWIC within the new KWIC data as belonging to one or more of the media reputation drivers, or to none of the media reputation drivers. The predicted sentiment value and the media reputation driver labels are then stored per each KWIC together with the KWIC data and the entity name and the enrichment data within the storage device 102 (
The step of calculating and storing media reputation driver scores (block 630), of
As shown by block 638, the raw media reputation driver score of the entity then has a weighting algorithm applied to it, where the weighting is based on text sources, volume, reach, and impact so as to allow for greater consideration of KWICs from sources that are potentially more important, having greater influence, and/or having a stronger reputation or validity within the industry. One having ordinary skill in the art would know different techniques for weighting data or data sources, and therefore, while the following provides one example of a method that may be used to weigh KWICs, it should be noted that this is not intended to limit the present invention to this method of weighting. It is noted that it may not be necessary to apply a weighting algorithm to the raw media reputation driver score. This step depends upon whether there is a desire to weight certain data sources differently.
It will be recalled that the storage device 102 (
In accordance with an alternative embodiment of the invention, a quality step may be added to ensure that redundancy in storage and sentiment consideration of KWICs does not adversely affect the media reputation driver score. More specifically, it should be noted that certain KWICs may contain both an entity name and enrichment data. The process of finding, capturing, and storing entity name KWICs and separately enrichment data KWICs may result in the entity name KWIC and enrichment data KWIC being the same KWIC. This is the purpose of the quality step. In accordance with the alternative embodiment of the invention, the media data management module 160 (
Weighting of the media reputation driver raw score may be performed in one of many different ways. For exemplary purposes only, the following describes a three-step process that may be used for weighting. In a first step, prominence weight features are created per KWIC, so that a weight may be applied for KWIC sources that are potentially more important, having greater influence, and/or having a stronger reputation or validity within the industry. The previously provided equations 7, 8, 9 may be used in this calculation.
As shown by block 640, the weighted media reputation driver score is then aligned to a normative scale. An example of a process that may be used to align to a normative scale includes, but is not limited to, applying logarithmic transformation in the form of equation 10 previously presented.
Using the distributions of reputation driver scores collected over a period of time via the trained model, an iterative process is run to tune the parameters of equation 10 to normalize the media reputation driver score to the normative scale. The result is a final zero to one-hundred entity media reputation driver score. The entity media reputation driver score may then be stored to the storage device 104 (
Having the reputation perception score of the entity and the media reputation score of the entity provides a better overall reputation assessment and scoring that considers both surveys and media data. In addition, it is noted that the media reputation score may have an effect on the reputation perception score and the reputation perception score may have an effect on the media reputation score. The following description, as provided with regard to
As shown by block 704, the supervised machine learning model is then trained using the historical data of previously determined media reputation scores, media reputation driver scores, and weights of the media reputation driver scores as input parameters, and the perception reputation scores as the output dependent variable. It is noted that any combination, derivation, only a part of, or a statistical property of the media reputation scores, media reputation driver scores, and weights of the media reputation driver scores may also be used as input variables (also called features) to train the supervised machine learning model. Here, for simplicity, we describe one example of inputs in the implementation of the model training.
In implementation, training of the supervised machine learning model that is used to predict the effect of media reputation on perception reputation is performed by the following steps. It is noted that such training may be performed in another manner known by one having ordinary skill in the art, which would be used for training of a supervised machine learning model.
The supervised machine learning model is trained by randomly separating the historical data to a train set and a test set and tuning the hyperparameters of the model multiple times until a desired accuracy is reached. For example, in each iteration 80% of the historical data may be used as a train set and 20% of the historical data may be used as a test set. The separation of the data to a train set and test set is well known in the art and will not be further described here. Both the train set, and the test set, contain matching data sets per entity and a pre-defined time-point of media reputation scores and media reputation drivers scores as inputs and perception reputation scores as outputs. As a first step, the model is trained using the inputs of the media reputation scores and media driver scores, for one or multiple time-points with the single output set of the dependent variable of the perception reputation scores from a particular time point (for example a certain month), with or without a time lag, for example of one month. This process may be performed for one entity or multiple entities. As a second step in training the supervised machine learning model, this same process is then performed for the same entity or a set of entities using the input variables and output variables from another point in time (for example a consecutive month), and so forth, until the supervised machine learning model is trained for each of the entities, resulting in a set of supervised machines learning trained models, per each of the entities for which reputation is sought. To achieve a higher accuracy the models may be trained using as input variables time series of media reputation scores and media reputation driver scores per each set of output dependent variables of perception reputation scores. For example, the time series may be of three months of data. When training a supervised machine learning model per one entity, the model training may be performed tens or even hundreds of times to increase the accuracy of the model by tuning the hyperparameters. Some examples of supervised machine learning models that may be trained on time series of input variables of media reputation scores and media reputation drivers scores may include, but not limited to, a VAR (vector autoregression) and VER (vector error correction) models, or LSTM models. Preferably, training of the model continues until a desired level of accuracy is received, as shown by block 706. For example, training may continue until a seventy percent accuracy is obtained.
As shown by block 708, the trained model is then saved within the reputation server 100. Each of the trained supervised learning models are then used to predict the effect of media reputation scores on perception reputation scores, as would be understood by one having ordinary skill in the art. As shown by block 710, the trained models may then be used in accordance with a predefined time interval to predict impact of the media reputation of the entity on the same entity's perception reputation per time interval, such as, but not limited to, per month, to help entities manage their reputation.
As one nonlimiting example of implementation, the trained models may then be used to predict (which may be also called to inference or to infer) whether the impact of the media reputation on the perception reputation of an entity will be positive, negative or neutral in a certain period of time. For example, if the trained model predicts that the perception reputation score will increase above a pre-set threshold in the following month compared to the previous month it is considered that the media reputation will have a positive impact on the perception reputation in that month for this entity. Similarly, if the trained model predicts that for a certain entity and their media reputation scores of the last two months the perception reputation score will decrease below a pre-set threshold in the following month compared to the previous month it is considered that the media reputation will have a negative impact on the perception reputation in that month for this entity. If the trained model per entity in a certain time period does not predict neither statistically significant increase nor decrease of the entity's perception reputation score in the following month above or below the pre-set thresholds it is considered that the media reputation will have a neutral impact or will have no impact on the perception reputation score in the next month.
As shown by block 754, the supervised machine learning model is then trained using the historical data of previously calculated perception reputation scores and weights as input variables and historical previously determined media reputation scores as dependent output variables.
In implementation, training of the supervised machine learning model that is used to predict the impact of perception reputation on media reputation is performed by similar steps to that describe above for predicting media reputation impact on perception reputation per entity where now the perception reputation scores, the perception reputation driver scores and weights are used as the input variables and the media reputation score, with or without a time-lag, for example of one month, is used as the output dependent variable. This training may be performed as described previously per multiple entities or per one entity resulting in a multiple set of trained models per entity for which reputation is being sought.
It is noted that such training may be performed in another manner known by one having ordinary skill in the art, which would be used for training of a supervised machine learning model.
Preferably, training of the model continues until a desired level of accuracy is received, as shown by block 756. For example, training may continue until a seventy percent accuracy is obtained.
As shown by block 758, the trained model is then saved within the reputation server 100. Each of the trained supervised machine learning models are then used to predict (which may be also called to inference or to infer) the effect of perception reputation scores on media reputation scores per entity, as would be understood by one having ordinary skill in the art. As shown by block 760, the trained models may then be used in accordance with a predefined time interval to predict impact of the perception reputation on media reputation of the entity, such as, but not limited to, per month, to help entities manage their reputation.
As one nonlimiting example of implementation, the trained models may then be used to predict (which may be also called to inference or to infer) whether the impact of the perception reputation on the media reputation of an entity will be positive, negative or neutral in a certain period of time. For example, if the trained model predicts that for a certain entity and their perception reputation scores of the last two months the media reputation score will increase above a pre-set threshold in the following month compared to the previous month it is considered that the perception reputation will have a positive impact on the media reputation in that month for this entity. Similarly, if the trained model predicts that for a certain entity and their perception reputation scores of the last two months the media reputation score will decrease below a pre-set threshold in the following month compared to the previous month it is considered that the perception reputation will have a negative impact on the media reputation in that month for this entity. If the trained model per entity in a certain time period does not predict neither statistically significant increase nor decrease of the entity's media reputation score in the following month above or below the pre-set thresholds it is considered that the perception reputation will have a neutral impact, or will have no impact on the media reputation score in the next month.
In accordance with an alternative embodiment of the invention, the input variables (also called features) to the supervised machine learning model may be enhanced by using mathematical manipulations of the input variables, also referred to by one of ordinary skill in the art as feature engineering. Such mathematical manipulation may include, but is not limited to, statistical derivations of the input scores and weights, such as, but not limited to, average and standard deviation, as well as using scores of a range of lags of the time unit, such as, but not limited to, a few weeks, or a few months. The result of using the statistical derivations (engineered features) of the input scores and weights in addition to the scores and weights themselves may increase the accuracy of prediction of the impact of media reputation on perception reputation of the entity for which reputation is sought, or of the perception reputation on the media reputation of the entity for which reputation is sought in a certain period of time and hence enhancing the ability of the entity to manage their reputation with higher precision.
In accordance with one embodiment of the invention, different computation results may be provided or displayed to an entity seeking its media reputation. For example, the media management module may display per period of time, for example a day, week, month or quarter, the following nonlimiting list of media reputation results: media reputation scores, media reputation driver scores and weights, media reputation factor scores and weights, prevalence of media reputation drivers or factors within the media reputation data of the entity, media reputation scores weights and volumes by media channel, media reputation trends over a time period (for example, six months), impact of the media reputation on the perception reputation of the entity in a certain time period, the impact of perception reputation on the media reputation of the entity in a certain time period, combined overall reputation scores and weights form media and perception reputation, benchmarking information of the entity's media reputation and perception reputation comparing to other entities such as competitors, industries or regions. Additional components may include analyses of impact and reach of media data on entity's reputation, topic analysis of the media data that drove the media reputation scores in a certain period of time, deep dives into the data for root cause analyses, and providing the entity with actionable recommendations and insights derived from the media reputation and perception reputation analyses. These may help the entity manage their reputation over time, understand and possibly influence the effect of their reputation on their various business outcomes.
It is noted that media data may be collected daily from partnered data providers via an application programming interface (API), or by scraping, or web crawling, or by other means, for a particular entity, such as company, corporate, brand, person, or industry, and is saved as “full text” in the database, such as the storage device 104 described in
It is also noted that the described here derivation of entity-based sentiment and classification of media data into reputation drivers may be done on full text articles and other media data components in other implementations of the current invention.
This application is a continuation-in-part of U.S. patent application Ser. No. 17/176,271, filed Feb. 16, 2021, entitled “System and Method For Determining And Managing Reputation Of Entities And Industries,” and a continuation-in-part of U.S. patent application Ser. No. 17/306,397, filed May 3, 2021, entitled “System and Method For Determining and Managing Reputation of Entities and Industries Through Use Of Behavioral Connections,” both of which are incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | 17176271 | Feb 2021 | US |
Child | 17510656 | US | |
Parent | 17306397 | May 2021 | US |
Child | 17176271 | US | |
Parent | 17176271 | Feb 2021 | US |
Child | 17306397 | US |