Emotional states of individuals, also known as moods, are central to the expression of thoughts, ideas and opinions, and in turn impact attitudes and behavior. Human emotions and mood have been a well-studied research area in psychology. Generally speaking, moods are complex patterns of cognitive processes, physiological arousal, and behavioral reactions. Moods serve a variety of purposes. They rouse us to action and direct and sustain that action. They help us organize our experience by directing attention, and by influencing our perceptions of self, others, and our interpretation and memory of events. By intensifying experiences, moods help identify self-relevant events. In other words, moods play a critical role in our everyday lives, fundamentally directing our attention and responses to environment, framing our attitudes and impacting our social relationships.
Considerable research in psychology has defined and examined human emotion and mood, with basic moods encompassing positive experiences like joy and acceptance, negative experiences like anger and disgust, and others like anticipation that are less clearly positive or negative. Of notable importance is the role of intensity or activation in emotion and mood, defined as the psychophysiological arousal of the mood in response to an associated stimulus. Together with valence (i.e., the degree of positivity/negativity of a mood), these two attributes characterize the structure of affective experience.
Given the important role of moods in human behavior, a growing body of literature has emerged in the social network/media research community that aims to mine temporal and semantic trends of affect, or detect and classify sentiment in order to better understand polarity of human opinions on various topics and contexts. This is understandable as social media tools (such as Twitter™) continue to evolve as major platforms of human expression, allowing individuals across the globe to share their thoughts, ideas, opinions and events of interest with others. While such content sharing can be objective in nature, it can also reflect emotional states from personal (e.g., loneliness, depression) to global scales (e.g., thoughts about a political candidate, musings about a newly released product or the global economy). Understanding these emotional states, or moods, of individuals at a large scale, manifested via their shared content on social media will help better interpret and make sense of the behavior of millions of individuals.
The aforementioned social network research has been largely directed towards automatic classification of sentiment in online domains. Typically, machine learning techniques are employed in an attempt to label text as being in one of two broad, general categories—namely either text exhibiting a Positive Affect (PA) or text exhibiting a Negative Affect (NA).
Affective state classification embodiments described herein generally involve the training and use of a classifier which identifies an affect exhibited by a segment of text. In an exemplary embodiment of the classifier training, a plurality of segments of text, each of which exhibits a known affect, is input into a computing device. For each of these, features are extracted from the segment to produce a vector of word features. The word feature vectors are then employed to train a classifier to identify an affect exhibited by a later-input text segment. It is noted that the affect being identified is chosen from a group of affects, each of which corresponds to a different emotion or sentiment being expressed by a person authoring the segment of text. In addition, each affect relates more than just the valence of the emotion or sentiment being expressed.
As indicated above, once the classifier is trained, in an exemplary embodiment it is used to classify an affect exhibited by a segment of text. This generally involves using a computing device to input the trained classifier, as well as a segment of text. The trained classifier is then used to identify an affect exhibited by the inputted segment of text. The affect being identified is chosen from the aforementioned group of affects.
It is noted that this Summary is provided to introduce a selection of concepts, in a simplified form, that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
The specific features, aspects, and advantages of the disclosure will become better understood with regard to the following description, appended claims, and accompanying drawings where:
In the following description of affective state classification embodiments reference is made to the accompanying drawings which form a part hereof, and in which are shown, by way of illustration, specific embodiments in which the technique may be practiced. It is understood that other embodiments may be utilized and structural changes may be made without departing from the scope of the technique.
It is also noted that for the sake of clarity specific terminology will be resorted to in describing the affective state classification embodiments described herein and it is not intended for these embodiments to be limited to the specific terms so chosen. Furthermore, it is to be understood that each specific term includes all its technical equivalents that operate in a broadly similar manner to achieve a similar purpose. Reference herein to “one embodiment”, or “another embodiment”, or an “exemplary embodiment”, or an “alternate embodiment”, or “one implementation”, or “another implementation”, or an “exemplary implementation”, or an “alternate implementation” means that a particular feature, a particular structure, or particular characteristics described in connection with the embodiment or implementation can be included in at least one embodiment of the affective state classification. The appearances of the phrases “in one embodiment”, “in another embodiment”, “in an exemplary embodiment”, “in an alternate embodiment”, “in one tested embodiment”, “in one implementation”, “in another implementation”, “in an exemplary implementation”, “in an alternate implementation” in various places in the specification are not necessarily all referring to the same embodiment or implementation, nor are separate or alternative embodiments/implementations mutually exclusive of other embodiments/implementations. Yet furthermore, the order of process flow representing one or more embodiments or implementations of the affective state classification does not inherently indicate any particular order nor imply any limitations of the technique.
As indicated previously, social media tools have been gaining significant traction of late in emerging as platforms of human sentiment and affect expression. Sentiment and affect analysis can be useful in a number of scenarios, including marketing campaigns, monitoring responses to local and global happenings, and deciphering geographic and temporal mood trends. As social media tools become increasingly ubiquitous, such analyses of affect can also enable new information-seeking approaches; for instance, identifying search features given an affect attribute. Consequently, there is significant value to be derived from predicting and classifying human affect in social media.
To this end, the affective state classification embodiments described herein generally classify several human affective states in segments of text and in some embodiments classify affects in segments of social media text (e.g. social media posts). Starting with about 200 moods (as will be described in more detail later), mechanical turk studies are used to derive naturalistic signals from posts shared on social media about a variety of affects of individuals. This dataset is then deployed in an affect classification task.
However, classifying affective states in social media domains has several challenges. First, the variety of linguistic styles and emotional interpretations of millions of individuals impedes inferring affective states concretely; and at the same time constructing consistent features across shared content is challenging. Second, most standard machine learning techniques rely on availability of labeled data for training. Manual labeling of ground truth is possible and can be employed. However, an alternate method of gathering appropriate training examples is also available involving explicit mood words used as hashtags at the end of some social media posts.
More particularly, some affective state classification embodiments described herein generally employ an affect classifier of social media data that uses about 200 mood hashtags as a supervision ground truth signal for inferring affect. These explicit hashtag mood words are mapped to a number of affective states. In one embodiment, 11 different affective states of individuals are employed (as will be described in more detail later). The aforementioned mapping is derived via a series of mechanical turk studies. The affect-labeled posts are then used in a maximum entropy classification framework to predict the affective state, given a post.
Note that the affective state classification embodiments described herein go well beyond the standard Positive Affect (PA)/Negative Affect (NA) classification schemes (or “positive”, “negative” and “neutral” schemes). Thus, the described embodiments have the advantage of providing a rich classification framework that captures important nuances in mood expression.
However, before a more detailed description of the affective state classification embodiments is presented, it would be useful to understand how the expression of moods is associated with the behavior of individuals, e.g., their linguistic usage of moods, social ties, activity levels, interaction patterns and so on. As described previously, moods impact human behavior, responses to environmental stimuli, and social ties. Naturally, it is imperative to understand characteristics of moods on social media in order to decipher the collective behavior of large populations.
To this end, the first sections presented below will discuss the identification and study a variety of moods that frequent social media posts. The dimensions of valence and activation are employed to represent moods in the circumplex model, and the topology of this space is studied with respect to mood usage, network structure, activity, and participatory patterns. It is through this study that the aforementioned approximately 200 moods frequently found in social media are identified. In addition, these moods are analyzed in the context of behavioral attributes that define an individual's actions in social media, including mood usage levels, linguistic diversity of shared content, network structure, activity rates and participatory patterns (e.g., link sharing and conversational engagement).
Identifying social media moods begins with identifying representative mood words, i.e., signals that would indicate individuals' broad emotional states. These moods are then characterized by the two dimensions of valence and activation.
A large number of potential mood words were first pooled from a variety of sources. While these words bore the notion of positive/negative feelings, not all of them were appropriate as mood words. For example, “pretty” and “peace” both represent a positive affect, but are not convincingly moods. Hence, a filtering task was performed to identify mood-indicative words from the candidate list.
This filtering task progressed in two parallel phases. In one phase, two researchers (both fluent English speakers) were asked to rate each of these words on a Likert scale of 1-7, where 1 indicated “not a mood at all” and 7 indicated “absolutely a mood”. This resulted in a set of high quality base-line ratings. In the other parallel phase, a similar mood-rating collection task was set up using the framework provided by Amazon's Mechanical Turk (AMT) interface (http://aws.amazon.com/mturk/). AMT is a crowd sourcing Internet marketplace where a requester can co-ordinate the use of human intelligence to perform tasks. Like the researchers, the “turkers” were asked to rate the words on the 1-7 Liked scale. Each word was rated by 12 turkers, and only those turkers who had a greater than 95% approval rating and were from the United States were considered.
Using the ratings from the turkers and the researchers, a list was constructed of all those words where both the median turk rating and the median researcher rating was at least 4 (mid-point of the scale), and the standard deviation was less than or equal to 1. This gave a final set of 203 mood words that were agreed upon by both parties to be mood-indicative (examples include: excited, nervous, quiet, grumpy, depressed, patient, thankful, bored).
As indicated previously, the mood words can be characterized by the dimensions of valence and activation. Valence represents the degree of positivity/negativity of a mood, and activation represents the degree of psychophysiological arousal of the mood in response to an associated stimulus (with some emotions being more arousing than others). For example, depressed is higher in arousal than sad. Together valence and activation characterize the structure of affective experience, and their inter-relatedness aids in conceptualizing the affect.
Given the aforementioned list of representative moods, the next task is to determine the values of the valence and activation dimensions of each mood. For some words in the list, these mood attributes are conventionally known, having been computed after extensive and rigorous psychometric studies. As such these known mood attribute values can be adopted.
For the remaining words, another turk study was conducted. Like before, only those turkers who had at least 95% approval rating history and were from U.S. were considered. For a given mood word, each turker was asked to rate the valence and activation measures, on two different 1-10 Likert scales (1 indicated low valence/low activation, while 10 indicated high valence/high activation). The choice of this Likert scale was made to align with the scales used for words with the aforementioned adopted valence and activation values. Twenty-four ratings were collected per mood—12 each for valence and activation. Finally, for each mood, the ratings valence and activation were combined separately. In one embodiment, the mean ratings were employed as the final measures for the two attributes.
The outcome of the foregoing studies was a set of 203 mood words, with each word characterized by both valence and activation values.
A popular representation of moods is the circumplex model, which represents moods in 2-dimensional topology defined by valence (x-axis) and activation (y-axis). It is a spatial model in which affective concepts fall in a circle. When the aforementioned mood words are plotted using the circumplex model by their mean valence (x-axis) and mean activation (y-axis) ratings, several features of the resulting space are noteworthy. First, the number of different mood words are fairly equally distributed in the four quadrants Q1-Q4. Second, while the valence ratings cover almost the entire range between 1 and 10, there are fewer mood words with very high activation or very low activation, compared to valence. Also note that words of neutral valence (e.g., quiet) tend to be of lower activation. This is reasonable, as typically more extreme moods (e.g., infuriated) generate higher activation. Overall this circumplex model provides a fine-grained psychometric instrument for study of mood expression in social media.
In one tested embodiment, mood data was collected from the popular social media Twitter™. Because of its widespread use, Twitter™ can be seen as a large and reliable repository for observing the rich ensemble of moods. Mood data was collected on a full year's worth of Twitter™ posts posted in English. Since there were a considerable volume of posts that does not reflect moods, the first task was to eliminate as many non-mood-indicative posts as possible, while simultaneously avoiding labor-intensive manual labeling of posts with moods.
To tackle this challenge, it was observed that Twitter™ users share posts with hashtagged moods, often with the hashtag at the end of the post, which can serve as labels for constructing a mood dataset. For example, the hashtagged mood “#excited” is found in many posts. In this light, hashtags were used as labels to train classifiers (as will be described in more detail later). Posts which have one of the moods found the aforementioned 203 mood word lexicon in the form of a hashtag at the end of a post, were collected. By this process, the labeled mood dataset comprised about 10.6 million posts from about 4.1 million users. Finally, RT (“retweet”) posts were eliminated because there may be cases where a mood hashtag is added at the end to comment on the retweet—which is arguably different from the author associating a mood with the language they produce.
The collection of hashtag labeled data on moods can be relatively easy since it avoids manual annotations or computationally intensive machine learning, but how reliable is it to consider mood hashtags at the end of posts as true indicators of an individual's emotional state? To answer this question, responses were first gathered from a set of users via a turk study. The study intended to determine for how many cases a hashtagged mood word occurring at the end of a post truly captures an individual's mood (without the presence of external signals). Specifically, a Yes/No question was displayed alongside a post, to which the turker indicated whether the (highlighted) mood hashtag at the end of the post indeed reflected the author's sentiment. Like before, only U.S. turkers with greater than 95% approval rating history were considered. In addition, a requirement that the turker had to use Twitter™ at least five times a week was added (consuming content).
Separately, the quality of the mood data was also compared to a naïve method of spotting mood words anywhere in a post. Like before, a turk approach was used to determine for how many cases a mood word present anywhere in a post indicated the author's sentiment. For both the studies, this exercise was conducted over 100 posts per study, and each post was rated by 10 different turkers.
The studies indicated that in 83% of the cases, hashtagged moods at the end of posts indeed captured the users' moods; while for posts with moods present anywhere, only 58% captured the emotional states of the corresponding users (Fleiss-Kappa measures of inter-rater agreement for both studies were 0.68 and 0.64 respectively), thus providing a systematic verification of the quality of the labeled mood dataset.
Affect refers to the experience of feeling or emotion and is a key part of the process of an individual's interaction with environmental stimuli. The primary challenge in classifying affect lies in the unavailability of ground truth. The psychology literature indicates that there is an implicit relationship between the externally observed affect and the internal mood of an individual. When affect is detected by an individual (e.g., smile as an expression of joviality), it is characterized as an emotion or mood.
In this section, it is described how the previously-described mood word lexicon is mapped to a set of affective states. First, the set of affective states is defined. For this purpose, a version of the Positive and Negative Affect Schedule (PANAS) known as PANAS-X is adopted. PANAS-X defines 11 specific affects—namely “fear”, “sadness”, “guilt”, “hostility”, “joviality”, “self-assurance”, “attentiveness”, “shyness”, “fatigue”, “surprise”, and “serenity”. These 11 affects are employed in the classification process to be described shortly.
PANAS-X also includes a mapping of some mood words to particular affects. Approximately 60% of the aforementioned mood word lexicon can be mapped to affects based on the PANAS-X listings. A turk study was undertaken to establish mappings between some of the remaining mood words and the 11 affects. In this study, each turker was shown a set of 10 mood words and the set of 11 affects were listed with each. The turker was asked to select from the list the most appropriate affect that described the particular mood. Twelve ratings were collected per mood. Finally, the ratings for each mood word were combined, and for each word, the affect that received a majority rating was associated with the mood word. In one tested embodiment, the foregoing resulted in a mapping of 112 of the 203 mood words to an affect, where each mood word corresponded to one type of affect.
Two additional tables are also presented herein to provide a different perspective on the mood word-to-affect mappings.
The first study of mood exploration on social media post data is based on analyzing the circumplex model of moods in terms of the moods' usage frequencies (counts over all posts). It is noted that the usages of moods in each of the quadrants is considerably different (the differences between each pair of quadrants were found to be statistically significant based on independent sample t-tests: p<0.0001). The overall trend is that moods in Q3 (low valence, low activation) tend to be used extensively (sad, bored, annoyed, lazy), along with a small number of moods in Q1, of relatively higher valence and activation (happy, optimistic). Overall, usage frequencies of lower valence moods exceed those of higher valence moods.
It is hypothesized that the presence of a “broadcasting bias” lies behind these observations. Since individuals often use Twitter™ to broadcast their opinions and feelings on various topics, it is likely that the mood about some information needs to be of sufficiently low or high valence to be worth reporting to the audience. This appears to be particularly true with respect to positive valence terms, with mildly positive moods expressed only rarely. The observation that lower valence moods are shared more often might be due to individuals seeking social support from their audiences in response to various happenings externally as well as in their own lives. The observation that lower activation words dominate usage of these lower valence moods could reflect the simple fact that people experience these moods more than their high activation counterparts (people are bored more frequently than they are infuriated) or that it is less acceptable to express extreme negative moods.
Also explored was how the linguistic content associated with usage of various moods relates to their valence and activation. A circumplex model was employed to plot a mood's normalized entropy, defined as the entropy of the textual content (i.e., unigrams (which in one embodiment represent single words) over all posts associated with the mood), divided by the total number of posts expressing the mood. It is observed that moods on the right side of the circumplex model (Q1, Q4) tend to have higher entropy than left (Q2, Q3) (statistically significant based on an independent sample t-test). This indicates that while positive moods tend to be shared across a wide array of linguistic context (topics, events etc.), negative moods tend to be shared in a limited context, confined to limited topics.
In the next study, it was intended to investigate the relationship between the nature of moods expressed and how “social” an individual is, referred to as “sociality”. For the purposes of this description, sociality is defined as the ratio of the number of followers (inlinks) to the number of followees (outlinks) of an individual. It is assumed that individuals with a ratio close to 1 would be the most “social” since this implies roughly equal engagement on the part of those individuals in both outgoing and incoming social and information exchanges. When the ratio is significantly less than 1, it may indicate that the individual is not interesting enough to others and thereby possibly less “social” to the social media audience. On the other hand a high ratio much greater than 1 indicates that the individual is likely an elite user (e.g., a celebrity or news source) who typically is more of a broadcaster and thus also is not very “social” in the context of the social media. Nevertheless, it is noted that this measure of sociality does not incorporate the absolute number of followers and followees (i.e., two people with sociality ratios of 20/20 and 2000/2000 respectively would be considered equally ‘social’). This is acceptable as the sociality of an individual is likely to be a function of the #followers and #followees together, since they define the structure of his/her ego-network. Hence the ratio is more appropriate than the absolute values. However, other network-centric measures could be used to study the relationship between followers and followees as desired (e.g., clustering coefficient methods).
Here again, the circumplex model was used to represent relationship between moods and sociality. Each mood plotted was represented by a color square, indicating the mean #followers/#followees ratio of individuals who shared the mood. A red square denoted a low #followers/#followees ratio of the individuals who shared the mood, while blue denoted a high #followers/#followees ratio: both indicating that the individuals who share these moods are not very “social” according to the foregoing definition. Moods represented by green squares were the most “social”, with ratios of #followers/#followees close to 1. Moods of higher valence (Q1, Q4) tend to be green squares, indicating that the average individual sharing them has a #followers/#followees ratio close to 1—in other words is more “social”. On the other hand, moods of lower valence (Q2, Q3) are consistently red, i.e., individuals sharing them are not very “social”. In addition, blue square users are comparatively infrequent and tend to use moods from all quadrants. To the extent that these users are elite users, this indicates that, most times, they are reporting information on various events/topics in a rather objective and ‘mood-balanced’ manner. It is noted that to formalize these differences, the #followers/#followees ratios associated with all moods in Q1 through Q4 were compared using independent sample t-tests and the results were statistically significant: p<0.001.
These results indicate a relationship between the ego-network of individuals (i.e., the way they connect to others) and the expression of moods. While these results do show that there is positive correlation between one being more “social” and expressing moods of higher valence, they do not establish a causal relationship in any direction. The results, however, do indicate that positive moods appear to be associated with social interactions.
Next the relationship between an individual's activity and his/her mood expression is investigated. A measurement of how “active” s/he is in sharing posts is defined as: the number of posts shared per second since the time of the individual's account creation. It is conjectured that a highly active individual is likely to be more “social” as well, since s/he is interested in dissipating information to his/her audience consistently, and thereby remains connected to them.
Based on this definition, the circumplex model of moods represented by squares whose size is proportional to the mean rate of activity of all individuals who have shared the particular mood. The majority of the larger squares (or moods shared by highly “active” individuals) were found to lie in Q1 and Q4; in other words, high (or positive) valence moods are shared by highly active individuals (statistically significant based on independent sample t-tests between quadrant pairs). On the other hand, moods of high activation (in Q2) but low valence are shared primarily by individuals with a low activity rate. In general, this indicates that positive moods are associated more frequently with active individuals, while negative and high arousal moods appear to be shared more frequently by individuals with low activity. Interestingly, when combined with the previously described usage frequencies, this implies that a “tail” of users post relatively infrequently and express the bulk of the mildly negative moods (e.g., bored, lazy, annoyed, sad).
In the next study, two types of participatory patterns of individuals are examined in relation to expression of mood: sharing of (external) information via links, and conversational engagement of one individual with another via the “@-reply” feature in social media posts.
The interest lies in investigating how mood expression and participation interact in the context of external information sharing and conversational engagement, compared to similar contexts where moods are not shared. To this end, the “background probabilities” of link sharing (pli) and @-replies (prp) are computed, given as ratio of #links (or @-replies) in posts to the total number of posts. Thereafter, for each mood, the probability of the mood's co-occurrence in a post along with, first a link, and second a @-reply, is computed. A distribution of the number of moods that have occurrence probabilities less than or equal to the background probability, and those greater than the background probability (of links, @-replies), is then computed. Frequency histograms of the distributions are plotted against mood valence and activation. This shows that highly negative moods (low valence) tend to co-occur less frequently with a link or a @-reply, compared to the respective background probabilities. On the other hand, several highly positive moods (higher valence) co-occur more frequently than usual with a link. This is likely to indicate that when individuals choose to express moods in a link-bearing post, it is more frequently a positive than a negative mood. In general, mood expression, positive or negative, is quite low in posts with @-replies. It is also noted that low activation moods tend to co-occur less frequently with a link, compared to the respective background probability, while high activation moods appear to be relatively more frequently co-occurring with links. Finally, moods co-occurring with @-replies are overall quite infrequent, irrespective of their activation measures. In fact, a large number of moods co-occur with @-replies with a probability much lower than the respective background probability.
Based on these findings, it appears that there is a difference in the manner in which moods are shared in various participation contexts. For link sharing, when links and moods co-occur, those moods tend to be both positive and of moderate to high activation. Replies, on the other hand, show little in the way of mood expression.
From an applications perspective, the foregoing studies provide promising avenues to build mood classifiers useful in search and advertising, or to enable organizations track behavior of populations in health, socio-economics or urban development domains. The following sections described affective state classification embodiments that can be used in these pursuits.
The classifier is trained on segments of text with known affect. In one tested embodiment this involves employing social media posts (e.g., Twitter™ posts) as the text segments, each of which has a hashtag mood word associated with it that appears in the aforementioned mood words lexicon. The previously described mappings are then used to identify the affect associated with the hashtag mood word for each of the posts being used to train the classifier. However, it is noted that in general any segment of text that evokes a sentiment or emotion can be employed as long as the affect associated with each of the segments is known. For example, the text segments could be from a document and the segments could be sentences. In the following description, the example of employing social media posts as text segments is used. However, it is to be understood that other types of text segments can be used as well.
For each text segment of known effect, word features are extracted. In one tested embodiment, this involves, for each of the aforementioned social media posts being used to train the classifier, converting the text into a form conducive for feature extraction. For example, this conversion can include, any or all of, converting the characters of the post to lowercase, normalizing all numbers into a canonical from, removing any URLs, tokenizing the text, removing non-alphanumeric characters (e.g., commas, apostrophes), and so on. After conversion, word features are extracted from each post to form a vector of word features. In one embodiment, just single words are employed as the word features. In another embodiment, both single words and word pairs are employed as word features. It is noted that the addition of word pairs in the latter embodiment recognizes that in some cases an adjacent word can modify the meaning of a mood word. For example, the mood word “good” might be part of a word pair “not good”. Clearly, the word pair “not good” would correspond to a completely different affect than the single word “good”.
In view of the foregoing,
The foregoing computing program architecture can be advantageously used to implement the aforementioned classifier training. More particularly, with reference to
With regard to the foregoing process action of determining if the selected segment exhibits one of the aforementioned affects,
It is noted that after the word features are extracted from the text segments, in one embodiment, prior to using the aforementioned word feature vectors to train the classifier, the number of features in the feature-extracted text segments is reduced to improve the efficiency and accuracy of the classifier training procedure. In one implementation, features that occur fewer than a prescribed number of times (e.g., five) across all the feature-extracted text segments are removed. In another implementation, word features are reduced to a prescribed number (e.g., 50K) of the top features in terms of log likelihood ratio. More particularly, referring to
It is noted that both of the above-described feature reduction implementations can be employed together, or the feature reduction could use just one or the other implementation. In addition, other feature reduction methods can be implemented as well.
Further, in one embodiment, the word feature vectors are randomly split into a prescribed number of folds for cross-validation. In one implementation, the word feature vectors are split into three equal-sized folds, and two of the three folds (i.e., 66.6% of the word feature vectors) are designated as training vectors. The remaining third of the three folds (i.e., 33.3% of the word feature vectors) is designated as testing vectors. The classifier, once trained using the training vectors, is used to predict the affect of each of the testing vectors in a manner to be described shortly, so as to assess its accuracy.
With regard to training, each of the word feature vectors formed from the training segments is operated on by a classification process associated with the classifier being trained. In one embodiment, the classifier is a standard maximum entropy classifier that once trained using the training vectors assigns numerical weights to each word feature of the vector depending on how strongly each is deemed to correlate with the affect associated with the post. The combined weights of all the features in a post are deemed to be indicative of the degree to which it reflects a particular affect. However, other classifiers can be employed as well, such as, without limitation, a support vector machine (SVM) classifier or a naïve Bayes classifier. It is noted that in one tested embodiment, systematic parameter tuning of the classifier was not performed. Rather parameter values were selected based on prior performance on various affect classification tasks.
At runtime, a new text segment is presented, and word features are extracted from the new text segment using a feature extractor. These are passed into the trained classifier, which looks up the weights for the word features and combines them in a classifier-specific mathematical formulation. The output is a prediction of the affective class of the text segment, and a probability indicative of the likelihood the text segment exhibits the predicted affect.
The foregoing computing program architecture can be advantageously used to implement the aforementioned affect classification. More particularly, with reference to
The affective state classification embodiments described herein are operational within numerous types of general purpose or special purpose computing system environments or configurations.
For example,
To allow a device to implement the affective state classification embodiments described herein, the device should have a sufficient computational capability and system memory to enable basic computational operations. In particular, as illustrated by
In addition, the simplified computing device of
The simplified computing device of
Retention of information such as computer-readable or computer-executable instructions, data structures, program modules, etc., can also be accomplished by using any of a variety of the aforementioned communication media to encode one or more modulated data signals or carrier waves, or other transport mechanisms or communications protocols, and includes any wired or wireless information delivery mechanism. Note that the terms “modulated data signal” or “carrier wave” generally refer to a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. For example, communication media includes wired media such as a wired network or direct-wired connection carrying one or more modulated data signals, and wireless media such as acoustic, RF, infrared, laser, and other wireless media for transmitting and/or receiving one or more modulated data signals or carrier waves. Combinations of the any of the above should also be included within the scope of communication media.
Further, software, programs, and/or computer program products embodying some or all of the various affective state classification embodiments described herein, or portions thereof, may be stored, received, transmitted, or read from any desired combination of computer or machine readable media or storage devices and communication media in the form of computer executable instructions or other data structures.
Finally, the affective state classification embodiments described herein may be further described in the general context of computer-executable instructions, such as program modules, being executed by a computing device. Generally, program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types. The embodiments described herein may also be practiced in distributed computing environments where tasks are performed by one or more remote processing devices, or within a cloud of one or more devices, that are linked through one or more communications networks. In a distributed computing environment, program modules may be located in both local and remote computer storage media including media storage devices. Still further, the aforementioned instructions may be implemented, in part or in whole, as hardware logic circuits, which may or may not include a processor.
It is noted that any or all of the aforementioned embodiments throughout the description may be used in any combination desired to form additional hybrid embodiments. In addition, although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
This application claims the benefit of and priority to provisional U.S. patent application Ser. No. 61/831,139 filed Jun. 5, 2013.
Number | Date | Country | |
---|---|---|---|
61831139 | Jun 2013 | US |