While the invention is described herein by way of example for several embodiments and illustrative drawings, those skilled in the art will recognize that the invention is not limited to the embodiments or drawings described. It should be understood that the drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the invention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention. As used throughout this application, the word “may” is used in a permissive sense (i.e., meaning “having the potential to”), rather than the mandatory sense (i.e. meaning “must”).
The present invention will now be described in terms of specific, example embodiments. It is to be understood that the invention is not limited to the example embodiments disclosed. It should also be understood that not every feature of the presently disclosed apparatus, device and computer-readable code for facilitating advertising is necessary to implement the invention as claimed in any particular one of the appended claims. Various elements and features of devices are described to fully enable the invention. It should also be understood that throughout this disclosure, where a process or method is shown or described, the steps of the method may be performed in any order or simultaneously, unless it is clear from the context that one step depends on another being performed first.
The present inventors are now disclosing that it is useful to extract mood and/or personality data from digital media content (i.e. audio and/or video) and to target advertising content to one or more individuals associated with one or more ‘speaking parties’ of the digital media content in accordance with the extracted mood and/or personality data.
Embodiments of the present invention relate to a technique for provisioning advertisements in accordance with the context and/or content of voice content—including but not limited to voice content transmitted over a telecommunications network in the context of a multiparty conversation.
Certain examples of related to this technique are now explained in terms of exemplary use scenarios. After presentation of the use scenarios, various embodiments of the present invention will be described with reference to flow-charts and block diagrams. It is noted that the use scenarios relate to the specific case where the advertisements are presented ‘visually’ by the client device. In other examples, audio advertisements may be presented—for example, before, during or following a call or conversation.
Also, it is noted that the present use scenarios and many other examples relate to the case where the multi-party conversation is transmitted via a telecommunications network (e.g. circuit switched and/or packet switched). In another example, two or more people are conversing ‘in the same room’ and the conversation is recorded by a single microphones or plurality of microphones (and optionally one or more cameras) deployed ‘locally’ without any need for transmitting content of the conversation via a telecommunications network.
In the following examples, certain personality properties for a given user are detected from electronic media content of a multi-party conversation, and advertisements are targeted to one or more individuals associated with the ‘given user’ in accordance with the computed personality profiles.
As will be discussed with later figures, it is often useful to provide a service to advertisers (or those wishing to place ads) where the advertiser and/or ad-placer can specify how to which user personalities to target a given ad and/or how to target a given ad to different personalities. Furthermore, it may be useful, for example, to price service where the advertisement is distributed in accordance with the various personality profiles.
According to this scenario, a first user (i.e. ‘party 1’) of a desktop computer phones a second user (i.e. ‘party 2’) cellular telephone using VOIP software residing on the desktop, such as Skype® software in two distinct conversations: conversation a first conversation (
According to the example of
It is appreciated that
Thus, in some embodiments, it is advantageous to track the media content generated by a given user or speaker over multiple conversations in order to more accurately assess one or more personality characteristics of a personality profile for the given user or speaker.
In some embodiments, is hypothesized that if properties indicative of a ‘competitive’ personality are detected (i.e. from electronic media content of multi-party conversations) for a given user over time and/or in multiple conversations and/or in different situations, then the user is more likely to have a ‘competitive’ personality. Conversely, it is recognized that if these properties are only detected rarely and/or only in certain situations (for example, ‘competitive’ situations where the ‘baseline’ even for ‘non-competitive people is competitive—for example, sport-discussions), then it is less likely that the given user has the ‘competitive’ personality.
According to this scenario, a first user (i.e. ‘party 1’) of a desktop computer phones a second user (i.e. ‘party 2’) cellular telephone using VOIP software residing on the desktop, such as Skype® software in two distinct conversations: conversation a first conversation (
According to the example of
In one example, certain products or services may specifically be targeted to complainers. In another example, complainers are less likely to purchase certain products or services, so advertisement is targeted to all users except for those with a personality profile that includes ‘complainer.’
It is noted that
According to this scenario, a first user (i.e. ‘party 1’) of a desktop computer phones a second user (i.e. ‘party 2’) cellular telephone using VOIP software residing on the desktop, such as Skype® (software in two distinct conversations: conversation a first conversation (
According to the example of
In the example of
The aforementioned examples list to very specific personality traits, namely “competitiveness” (see
Below is a non-limiting list of various personality traits, each of which may be detected for a given speaker or speakers—in accordance with one or more personality traits, advertisement may be provided. In the list below, certain personality traits are contrasted with their opposite, though it is understood that this is not intended as a limitation.
a) Ambitious vs. Lazy
b) Passive vs. active
c) passionate vs. dispassionate
d) selfish vs. selfless
e) Norm Abiding vs. Adventurous
f) Creative or not
g) Risk averse vs. Risk taking
h) Optimist vs Pessimist
i) introvert vs. extrovert
j) thinking vs feeling
k) image conscious or not
l) impulsive or not
m) gregarious/anti-social
n) addictions—food, alcohol, drugs, sex
o) contemplative or not
p) intellectual or not
q) bossy or not
r) hedonistic or not
s) fear-prone or not
t) neat or sloppy
u) honest vs. untruthful
In some embodiments, individual speakers are given a numerical ‘score’ indicating a propensity to exhibiting a given personality trait. Alternatively or additionally, individual speakers are given a ‘score’ indicating a lack of exhibiting a given personality trait.
As noted above, presence or absence of ‘key words’ is just one exemplary technique for detecting a presence or absence of a given personality trait in a given speaker. In certain examples as shown with reference to
Thus, in one example related to video conferencing, the appearance of a dog may make a certain person draw back in fear, indicating that this individual is fear-prone.
In another example related to video conferencing, a person's appearance may indicate if the person is neat or sloppy.
As used herein, ‘providing’ of media or media content includes one or more of the following: (i) receiving the media content (for example, at a server cluster comprising at least one cluster, for example, operative to analyze the media content and/or at a proxy); (ii) sending the media content; (iii) generating the media content (for example, carried out at a client device such as a cell phone and/or PC); (iv) intercepting; and (v) handling media content, for example, on the client device, on a proxy or server.
As used herein, a ‘multi-party’ voice conversation includes two or more parties, for example, where each party communicated using a respective client device including but not limited to desktop, laptop, cell-phone, and personal digital assistant (PDA).
In one example, the electronic media content from the multi-party conversation is provided from a single client device (for example, a single cell phone or desktop). In another example, the media from the multi-party conversation includes content from different client devices.
Similarly, in one example, the media electronic media content from the multi-party conversation is from a single speaker or a single user. Alternatively, in another example, the media electronic media content from the multi-party conversation is from multiple speakers.
The electronic media content may be provided as streaming content. For example, streaming audio (and optionally video) content may be intercepted, for example, as transmitted a telecommunications network (for example, a packet switched or circuit switched network). Thus, in some embodiments, the conversation is monitored on an ongoing basis during a certain time period.
Alternatively or additionally, the electronic media content is pre-stored content, for example, stored in any combination of volatile and non-volatile memory.
As used herein, ‘providing at least one advertisement in accordance with a least one personality feature and/or a personality profile detectable from media content includes one or more of the following:
i) configuring a client device (i.e. a screen of a client device) to display advertisement such that display of the client device displays advertisement in accordance with the detectable at least one personality feature of the media content. This configuring may be accomplished, for example, by displaying a advertising message using an email client and/or a web browser and/or any other client residing on the client device; ii) sending or directing or targeting an advertisement to a client device in accordance with the at least one detectable personality feature of the media content (for example, from a client to a server, via an email message, an SMS or any other method);
iii) configuring an advertisement targeting database that indicates how or to whom or when advertisements should be sent, for example, using ‘snail mail to a targeted user—i.e. in this case the database is a mailing list.
Embodiments of the present invention relate to providing or targeting advertisement to an ‘one individual associated with a party of the multi-party voice conversation.’
In one example, this individual is actually a participant in the multi-party voice conversation. Thus, a user may be associated with a client device (for example, a desktop or cellphone) for speaking and participating in the multi-party conversation. According to this example, the user's client device is configured to present (i.e. display and or play audio content) the targeted advertisement.
In another example, the advertisement is ‘targeted’ or provided using SMS or email or any other technique. The ‘associated individual’ may thus include one or more of: a) the individual himself/herself; b) a spouse or relative of the individual (for example, as determined using a database); c) any other person for which there is an electronic record associating the other person with the participant in the multi-party conversation (for example, a neighbor as determined from a white pages database, a co-worker as determined from some purchasing ‘discount club’, a member of the same club or church or synagogue, etc).
In one example, a certain personality trait is detected in a given user (for example, the person is ‘impulsive’) from electronic media content of a multi-party conversation, and an advertisement is provided to an associated of the ‘impulsive’ person. This may be, for example, a spouse or a sibling of the impulsive person, even if the ‘associate’ that receives the advertisement does not participate in the multi-party conversation from which the ‘impulsive’ personality trait is detected.
Detailed Description of Block Diagrams and Flow Charts
In step S101, electronic digital media content including spoken or voice content (e.g. of a multi-party audio conversation) is provided—e.g. received and/or intercepted and/or handled.
In step S105, one or more aspects of electronic voice content (for example, content of multi-party audio conversation are analyzed), or context features are computed. Based on the results of the analysis, personality and/or mood traits may be determined.
This may be done in any one or more of a number of ways. In one example (see S159 of
In another example, the multi-party conversation is a ‘video conversation’ (i.e. voice plus video). In a specific example, if a conversation participant is dressed in an neat manner or a sloppy manner this may indicate whether or not the person is a perfectionist by nature. In another example, if a conversation participant exhibits certain body motions (for example, constantly shaking his/her knee, constantly pacing, etc) this may indicate a nervous and/or hyperactive disposition.
Other specific examples of specific implementations of step S105 will be discussed below, with reference to other figures.
In step S109, one or more operations are carried out to facilitate provisioning advertising in accordance with results of the analysis of step S105. (as noted throughout this disclosure, there are many examples where multiple conversations are analyzed over a period of time are analyzed in order to better ascertain the personality of a participant in the conversation).
One example of ‘facilitating the provisioning of advertising’ is using an ad server to serve advertisements to a user. Alternatively or additionally, another example of ‘facilitating the provisioning of advertising’ is using an aggregation service. More examples of provisioning advertisement(s) are described below.
It is noted that the aforementioned ‘use scenarios’ related to
It is also noted that the ‘use scenarios’ relate to the case where a multi-party conversation is monitored on an ongoing basis (i.e. S105 includes monitoring the conversation either in real-time or with some sort of time delay). Alternatively or additionally, the multi-party conversation may be saved in some sort of persistent media, and the conversation may be analyzed S105 ‘off line’.
In step S121, the media content is analyzed such that person-specific media content is associated with given specific parties. In one example, a VOIP “skype” conversation is analyzed—for example, see
Once it is determined which visual and/or audio content is generated by which participant, it is possible to associate different content C(P) with respective parties Pi of the conversation.
A Brief Discussion of How to Determine a Presence or Absence of a Personality Trait in a Person
For the present disclosure, “determining” or “generating” a “personality profile” includes determining a presence of at least one personality trait for a given person. In some embodiments, “determining” or “generating” a “personality profile” also includes determining an “absence” of at least one personality trait.
Typically, this is carried out in accordance with a “threshold” certainty for a presence or absence of the personality trait.
There are many situations where both “positive indications” as well as “negative indications” are present, and it may be necessary to “weigh” one against the other—for example, using a statistical model. Exemplary statistical models include but are not limited to C45 trees, neural networks, Markov models, linear regression, and the like.
If S179 the “positive” indications outweigh the “negative indications” (i.e. indicating the presence of the personality trait”) for example, according to some statistical model and according to some “threshold” indicative of a statistical significance (for example, established using a training set), the presence of the personality trait in the given person may be identified S181.
If S183 the “negative” indications outweigh the “positive indications” (i.e. indicating the absence of the personality trait”) for example, according to some statistical model and according to some “threshold” indicative of a statistical significance (for example, established using a training set), the absence of the personality trait in the given person may be identified S185.
A Brief Discussion of False Positives and False Negatives
It is noted that there are certain situations where some features “indicative of the presence or absence” of a given personality trait may be detectable, but nevertheless not enough features are present, or too many “contradictory” features (i.e. that contradict a given “present” or “absent” hypothesis) are present for the feature to be considered “present” (or “absent”). This issue has already been discussed with respect to
Thus, in one oversimplified example, if a person (i.e. a participant in a multi-person conversation—i.e. a potential “target”) exhibits feature “A” (i.e. this is detected in electronic audio and/or video media content generated by the person in the multi-person conversation) there is a 60% chance the “conversation-participant” is “correctly” associated with a given personality trait. If the person exhibits features “A” and “B” the probability is 80%. If the person exhibits features “A,” “B” and “C” but not feature “D” the probability is 90%. If the person exhibits features “A,” “B,” “C” and “D” the probability is 65%.
Thus, it is noted that any model for determining the presence or absence of any given personality trait my be associated with a rate of false positives and false negatives. If we require a “high threshold” (for example, requiring a probability of at least 80% before identifying the presence of personality trait, as in S181 of
In some embodiments, as will be discussed below with reference to element 916 of
A Discussion of Multiple Distinct Conversations and Time Profile Features of One or More Personality Traits
For the present disclosure, video and/or audio media content may be associated with a “time of generation”—i.e. the time the audio and/or visual signals are recorded, for example, during a multi-party voice and optionally video conversation. This “time of generation” may be known within some sort of tolerance—for example, within a few minutes or a few seconds or even less.
The beginning of a conversation may be defined as: (i) the time an audio and/or video “signal” is provided that a conversation is beginning—for example, a user saying “hi” or “hello”; and/or (ii) for the case of conversations that are transmitted over a switching network (for example, the Internet) between different terminal devices, the time that the audio and/or video stream connection between the different terminal devices residing at different locations over the switching network is established.
Similarly, the “end” of a conversation may be defined as: (i) the time an audio and/or video “signal” is provided that a conversation is ending—for example, a user saying “goodbye”; and/or (ii) for the case of conversations that are transmitted over a switching network (for example, the Internet) between different terminal devices, the time that the audio and/or video stream connection between the different terminal devices residing at different locations over the switching network is terminated.
For the present disclosure, the term “distinct multi-party conversations” (for example, between distinct user terminals of a communications network), refers to conversations where (a) each conversation has a length of at least 30 seconds; (b) the time gap (i.e. see
Some example of “distinct multi-party conversations” include (i) day-separated distinct multi-party conversations (i.e. conversations separated by a gap time of at least 24 hours); (ii) week-separated distinct multi-party conversations (i.e. conversations separated by a gap time of at least 7 days); (iii) month-separated distinct multi-party conversations (i.e. conversations separated by a gap time of at least 1 month).
For the present disclosure, a “long-term time profile” of one or more detected personality traits is either (I) detected separately for at least two distinct multi-party conversations that are at least day-separated multi-party conversations, or possibly week-separated or month-separated (i.e. every conversation individually indicates the presence or absence of the given personality trait beyond some sort of threshold—for example, the technique of
This is illustrated in
In some embodiments, “older detected features” (for example, associated with a conversation that is “previous” to the “most recent” conversation—for example, an at least day separated or week separated or month separated previous conversation) are given less weight (i.e. when categorizing a person has having a presence or absence of one or more personality features) in accordance with the “age” of the conversation—i.e. media content of a “newer” conversation is given greater weight when determining one or more personality features of a given person.
In one example, it may be decided to target people who historically are introverted, but who recently have become extroverted.
In another example, it may be decided to target people who, typically over a period of time, are introverted, but are having an “extroverted day” or react to a certain person in an extroverted manner.
Mood Deviations
For the present disclosure a “mood deviation” refers to the difference between the mood of an individual (for example, a participant in a multi-party conversation) (i) at a point in time, or during a given time interval (for example, a “short interval” of less than 30 minutes, or less than 10 minutes) and (ii) the person's historical moods or exhibited personality traits, for example, as observed in an earlier and
In one example, it is desired to target individuals who typically are introverted or soft-spoken at a time that they exhibit a period of extroversion or agitation, or some mood which contrasts a typical historical personality.
Each of determining the “current mood” as well as historical “personality traits” may carried out using some sort of statistical classifier model, for example, a configurable classifier model for minimizing false negatives or false positives.
Digital media content of a most recent or “current” conversation is analyzied to determine a presence or absence of a mood deviation or personality time trend function, for example, using a statistical classifier.
In some embodiments, it is advantageous to market advertising content in accordance with the personality profile of the user (or associated thereof) to whom the advertisement will be served.
For example, it may be determined that in some situations, “risk taking” individuals are an appropriate target audience. According to this example, a “seller” of electronic advertisement distribution services (or a party or mediator acting on behalf of the “seller”) will offer the “buyer” of such services (or a party acting on behalf of the “buyer”) the option to select a “target audience” in accordance with determined personality of a conversation-participant (i.e. a personality determined, at least in part, from the audio and/or video conversation generated by the conversation-participant).
Towards this end, it may, in some embodiments, be useful to provide an interface whereby the “buyer” can specify to a “seller” various directives for provisioning personality-targeted advertisements. One example of such a “personality-advertisement data-receiving user interface” 910 is provided in
Column 912 allows the user to select which personality features to target. In the example of
It is noted that the exemplary “personality-advertisement data-receiving user interface” 910 of
In one business scenario, it is important for a purchaser of advertisement placement services to serve a given advertisement to all individuals having a given personality trait, even if some advertisements are “wrongly” served to individuals with only certain indications (i.e. the purchaser is willing to “suffer” a certain number of false positives in order to minimize false negatives). In this example, the “hurdle” number of column 716 may be set to a relatively low number.
Conversely, in a different business scenario, it is important for a purchaser of advertisement placement services to target advertisement only to individuals that “beyond a doubt” exhibit the personality trait. In this example, the purchaser is willing to “miss” some possible individuals with the trait (i.e. increase more false negatives) while minimizing the number of false positives.
In the example of
Thus, it is noted that column 916 acts as a “noise filter”—the higher the number, the more “noise” or false positives are filtered out, but at the cost of potentially missing “signal” (i.e. false negatives).
Column 918 of the “personality-advertisement data-receiving user interface” interface 910 includes allows for the user representing the buyer to specify a required “time significance.” Thus, in the example of
It is noted that, in many business scenarios, the fee charged for advertisement placement may be influenced by the personality provide selected and/or the “noise filter” and/or “time filter” values. In the example of
In one example, the “price factor” is determined such that “more valuable” personality traits (for example, ambitious people) are priced higher. In another example, the “price factor” is determined by supply and demand from various “buyers.”
It is noted that the exemplary “personality-advertisement data-receiving user interface” 910 should not be construed as limiting, and is not a requirement. In some embodiments, the “buyer” and “seller” are represented by machines which neotiate with each other in an “interfaceless” manner, for example, using some data exchange XML or EDI-based protocol.
A Discussion of Various Business Methods
In step S201, an interface is presented for linking advertisement to personalities (for example, as in 910 of
In step S223, the advertisement is provided or targeted in accordance with the directives received in step S215 and the detected S219 personality traits or mood deviation functions.
Sometimes it may be convenient to store data about previous conversations and to associate this data with user account information. Thus, the system may determine from a first conversation (or set of conversations) specific data about a given user with a certain level of certainty.
Later, when the user engages in a second multi-party conversation, it may be advantageous to access the earlier-stored demographic data in order to provide to the user the most appropriate advertisement. Thus, there is no need for the system to re-profile the given user.
In another example, the earlier demographic profile may be refined in a later conversation by gathering more ‘input data points.’
In some embodiments, the user may be averse to giving ‘account information’—for example, because there is a desire not to inconvenience the user.
Nevertheless, it may be advantageous to maintain a ‘voice print’ database which would allow identifying a given user from his or her ‘voice print.’
Recognizing an identity of a user from a voice print is known in the art—the skilled artisan is referred to, for example, US 2006/0188076; US 2005/0131706; US 2003/0125944; and US 2002/0152078 each of which is incorporated herein by reference in entirety
Thus, content (i.e. voice content and optionally video content) of a multi-party conversation may be analyzed and one or more biometric parameters or features (for example, voice print or face ‘print’) are computed. The results of the analysis and optionally demographic data are stored and are associated with a user identity and/or voice print data.
During a second conversation, the identity of the user is determined and/or the user is associated with the previous conversation using voice print data based on analysis of voice and/or video content. At this point, the previous personality trait information of the user is available.
Then, the personality trait data may be refined by analyzing the second conversation.
This could allow for determining a personality trait with greater ‘clasification’ certainty (i.e. from ‘cumulative’ of different conversations) and/or determining a personality trait exhibited over a ‘long term’ (for example, at least a day, week or month) which provides a ‘time certainty.’
Discussion of Exemplary Apparatus
The exemplary system 100 may an input 110 for receiving one or more digitized audio and/or visual waveforms, a speech recognition engine 154 (for converting a live or recorded speech signal to a sequence of words), one or more feature extractor(s) 118, one or more advertisement targeting engine(s) 134, a historical data storage 142, and a historical data storage updating engine 150.
Exemplary implementations of each of the aforementioned components are described below.
It is appreciated that not every component in
In some embodiments, the media input 110 for receiving a digitized waveform is a streaming input. This may be useful for ‘eavesdropping’ on a multi-party conversation in substantially real time. In some embodiments, ‘substantially real time’ refers to refer time with no more than a pre-determined time delay, for example, a delay of at most 15 seconds, or at most 1 minute, or at most 5 minutes, or at most 30 minutes, or at most 60 minutes.
In
In yet another example, input 110 does not necessarily receive or handle a streaming signal. In one example, stored digital audio and/or video waveforms may be provided stored in non-volatile memory (including but not limited to flash, magnetic and optical media) or in volatile memory.
It is also noted, with reference to
In yet another example, two or more parties may converse over a ‘traditional’ circuit-switched phone network, and the audio sounds may be streamed to advertisement system 100 and/or provided as recording digital media stored in volatile and/or non-volatile memory.
It is noted that the feature extractors may employ any technique for feature extraction of media content known in the art, including but not limited to heuristically techniques and/or ‘statistical AI’ and/or ‘data mining techniques’ and/or ‘machine learning techniques’ where a training set is first provided to a classifier or feature calculation engine. The training may be supervised or unsupervised.
Exemplary techniques include but are not limited to tree techniques (for example binary trees), regression techniques, Hidden Markov Models, Neural Networks, and meta-techniques such as boosting or bagging. In specific embodiments, this statistical model is created in accordance with previously collected “training” data. In some embodiments, a scoring system is created. In some embodiments, a voting model for combining more than one technique is used.
Appropriate statistical techniques are well known in the art, and are described in a large number of well known sources including, for example, Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations by Ian H. Witten, Eibe Frank; Morgan Kaufmann, October 1999), the entirety of which is herein incorporated by reference.
It is noted that in exemplary embodiments a first feature may be determined in accordance with a different feature, thus facilitating ‘feature combining.’
In some embodiments, one or more feature extractors or calculation engine may be operative to effect one or more ‘classification operations’ for determining a personality trait and/or mood deviation.
Each element described in
Text Feature Extractor(s) 210
In one example, when a speaker uses a certain phrase, this may indicate a current desire or preference. For example, if a speaker says “I am quite angry” this may indicate a mood; if this happens frequently, this may indicate a personality trait—i.e. easily angered.
The speaker profile built from detecting these phrases, and optionally performing statistical analysis, may be useful for present or future provisioning of ads to the speaker or to another person associated with the speaker.
The phrase detector 260 may include, for example, a database of pre-determined words or phrases or regular expressions.
In one example, it is recognized that the computational cost associated with analyzing text to determine the appearance of certain regular phrases (i.e. from a pre-determined set) may increase with the size of the set of phrases.
In some embodiments, it may be useful to analyze frequencies of words (or word combinations) in a given segment of conversation using a language model engine 256.
For example, it is recognized that more educated people tend to use a different set of vocabulary in their speech than less educated people. Thus, it is possible to prepare predetermined conversation ‘training sets’ of more educated people and conversation ‘training sets’ of less educated people. For each training set, frequencies of various words may be computed. For each pre-determined conversation ‘training set,’ a language model of word (or word combination) frequencies may be constructed.
According to this example, when a segment of conversation is analyzed, it is possible (i.e. for a given speaker or speakers) to compare the frequencies of word usage in the analyzed segment of conversation, and to determine if the frequency table more closely matches the training set of more educated people or less educated people, in order to obtain demographic data (i.e. This principle may also be used for different conversation ‘types.’For example, conversations related to computer technologies would tend to provide an elevated frequency for one set of words, romantic conversations would tend to provide an elevated frequency for another set of words, etc. Thus, for different conversation types, or conversation topics, various training sets can be prepared. For a given segment of analyzed conversation, word frequencies (or word combination frequencies) can then be compared with the frequencies of one or more training sets.
The same principle described for word frequencies can also be applied to sentence structures—i.e. certain pre-determined demographic groups or conversation type may be associated with certain sentence structures. Thus, in some embodiments, a part of speech (POS) tagger 264 is provided.
A Discussion of
As with any feature detector or computation engine disclosed herein, speech delivery feature extractor 220 or any component thereof may be pre-trained with ‘training data’ from a training set.
This includes a conversation harmony level classifier (for example, determining if a conversation is friendly or unfriendly and to what extent) 452, a deviation feature calculation engine 456, a feature engine for demographic feature(s) 460, a feature engine for physiological status 464, a feature engine for conversation participants relation status 468 (for example, family members, business partners, friends, lovers, spouses, etc), conversation expected length classifier 472 (i.e. if the end of the conversation is expected within a ‘short’ period of time, the advertisement providing may be carried out differently than for the situation where the end of the conversation is not expected within a short period of time), conversation topic classifier 476, etc.
In another example, advertisement delivery engine 718 may decide a parameter for a delayed provisioning of advertisement—for example, 10 minutes after the conversation, several hours, a day, a week, etc.
In another example, the ad may be served in the context of a computer gaming environment. For example, games may speak when engaged in a multi-player computer game, and advertisements may be served in a manner that is integrated in the game environment. In one example, for a computer basketball game, the court or ball may be provisioned with certain ads determined in accordance with the content of the voice and/or video content of the conversation between games.
In the description and claims of the present application, each of the verbs, “comprise” “include” and “have”, and conjugates thereof, are used to indicate that the object or objects of the verb are not necessarily a complete listing of members, components, elements or parts of the subject or subjects of the verb.
All references cited herein are incorporated by reference in their entirety. Citation of a reference does not constitute an admission that the reference is prior art.
The articles “a” and “an” are used herein to refer to one or to more than one (i.e., to at least one) of the grammatical object of the article. By way of example, “an element” means one element or more than one element.
The term “including” is used herein to mean, and is used interchangeably with, the phrase “including but not limited” to.
The term “or” is used herein to mean, and is used interchangeably with, the term “and/or,” unless context clearly indicates otherwise.
The present invention has been described using detailed descriptions of embodiments thereof that are provided by way of example and are not intended to limit the scope of the invention. The described embodiments comprise different features, not all of which are required in all embodiments of the invention. Some embodiments of the present invention utilize only some of the features or possible combinations of the features. Variations of embodiments of the present invention that are described and embodiments of the present invention comprising different combinations of features noted in the described embodiments will occur to persons of the art.
This patent application claims the benefit of U.S. Provisional Patent Application No. 60/821,271 filed Aug. 3, 2006 by the present inventors.
Number | Date | Country | |
---|---|---|---|
60821271 | Aug 2006 | US |