The present invention relates to systems and methods for measuring biologically and behaviorally based responses to social media, locations, or experiences and providing instant and continuous feedback in response thereto. In addition, the system and methods of the invention are capable of monitoring stress levels and well-being. The system and methods of the invention may be implemented using a cloud-based infrastructure for remote monitoring.
There are many different kinds of audio, visual and audio-visual presentations and activities that people are exposed to every day. These presentations serve as sensory experiences that stimulate our senses and are known to result in biologically based responses that can be measured electronically and mechanically (for example, heart rate, respiration rate, blood pressure, and skin conductance).
Likewise, people now have the ability to provide instant and continuous feedback in response to various social media such as pictures, websites, and the like. Such feedback can be provided on computers, tablets, smart phones, and other devices that access the internet. For example, “like” is a way to give positive feedback or to connect with things a person is interested in on the popular social media site Facebook®. In particular, the “like” button on Facebook® is a button a user may click on after looking at most content on Facebook®, which is then reported in newsfeeds to “friends”. Websites unrelated to Facebook® may also use a “like” button that enables a website visitor to click on the button to let his/her friends know that they like the site. For example, after clicking on the website's “like” button, a pop-up will request login to Facebook® (or sign-up if not already a member) and a post on the user's Facebook® page will let his/her friends know that he/she likes the site. When used on a mobile device, such as a smart phone, the “like” button is merely an integrated hardware “Facebook®” button on the phone that does nothing more than take the user to Facebook® when the button is pressed.
Similarly, the “Pin It” button on a computer or mobile device allows users to grab images and videos from around the web and add them to an on-line pinboards created by the users. Other users can view the pinboards, comment, and “re-pin”.
Capabilities have also been introduced to allow people to use mobile devices to interact with their environment. For example, location-based social networking websites allow users to “check-in” at venues using a mobile website, text messaging, or a device-specific application by selecting from a list of venues the application locates nearby. The location is based on GPS hardware in the mobile device or the network location provided by the application. Each check-in awards the user points or other types of rewards.
Even with these advances in technology, the ability to measure and evaluate the user experience, effectiveness, and the usability of social media, locations, or experiences has been limited. In fact, current methodologies for measuring or evaluating user experience, effectiveness, and usability of websites and other interactive internet and software media has thus far been limited to traditional self-report, i.e., relying on the user to use the “like” button and to accurately reflect his/her actual response to the social media, which may be subject to error, bias, or low compliance.
Thus, a need in the art exists for a system and method that integrates passive biometric sensors into smart phones or other portable devices to collaborate with or eliminate the “like” button and replace it with a continuous stream of emotional responses across all experiences. A need also exists in the art for a biometrically enabled suite of applications that are built into smart phones, tablets, and other social media enabled devices to determine when a user unconsciously likes (or dislikes) their current experience, e.g., a web page, “app”, song, video, location, or other experience, and also to remotely monitor the user's stress levels and well-being.
The present invention is related to a system and method that integrates passive biometric sensors into smart phones or other portable devices to collaborate with or eliminate the “like” button and replace it with a continuous stream of emotional responses across all experiences. In one embodiment of the invention, biometric responses may be automatically detected using physiological signal measurement, via an explicit button on the mobile device, or via a continuous biometric data collector. In one embodiment, the biometric data collector is associated with the mobile device, such as a case that incorporates sensors to collect data from the hands of the user.
In another embodiment of the invention, the response may be automatically associated with specific content stimuli, not only in the mobile phone, but also in the surrounding environment of the consumer using a specialized content-associating system (predicting/associating by “proximity” of the response to the variety of content/activities).
In yet another embodiment, the system and method of the invention provides content recommendations by transforming emotional responses to music, video, and other content into more effective and pervasive recommendations. In still another embodiment, the system and method of the invention enhances advertisement targeting by combining biometric sensing and location-based data based on anticipated emotional responses. In yet another embodiment, the system and method of the invention add emotional intelligence to social networks by generating a complex dataset of interests from which users can share and connect with friends. In still another embodiment, the system and method of the invention continually performs background searches based on emotional reactions to any activity in order to push more relevant information to consumers.
The present invention is also directed to a system for determining the emotional response of a user to social media including: a plurality of biometric sensors in a first device operable to measure a plurality of biometric parameters for a user of the first device when exposed to a social media application, marketing or advertising applications, or a combination thereof; a computer system operable to receive data representative of the plurality of biometric parameters, wherein the computer system further includes a memory capable of storing the data, wherein the system is capable of determining an emotional response of the user, identifying the causation of the emotional response, and delivering the emotional response and the causation to a listening application for appropriate action.
In one embodiment, the plurality of biometric parameters include at least two of galvanic skin response, heart response, motion, skin temperature, breathing, EEG, EMG, pupil dilation, eye tracking, facial expressions, or a combination thereof. In another embodiment, the first device includes a device capable of accessing a social media application, a marketing or advertising application, or a combination thereof. For example, the first device may be selected from the group consisting of mobile devices, tablets, laptops, desktops, goggles, watches, and combinations thereof.
In another embodiment, the first device further includes image acquisition functionality, and wherein the image acquisition functionality includes static photographs, videos, or a combination thereof. For example, the image acquisition functionality may include at least one camera, and wherein the at least one camera is capable of acquiring a plurality of images including images in the user's view, images of the user, or a combination thereof. The system may also be capable of image recognition of the images in the user's view. In yet another embodiment, the system may be capable of processing facial recognition using the images, videos, or a combination thereof of the user.
The present invention is also directed to a system for determining the emotional state of a user to social media including: a first device capable of accessing a social media application marketing or advertising applications, or a combination thereof; a second device including at least one biometric sensor operatively connected to the first device, wherein the at least one biometric sensor is operable to measure at least one biometric parameter for a user of the first device when exposed to social media; a computer system operable to receive data representative of the at least one biometric parameter, wherein the computer system further includes a memory capable of storing the data, wherein the system is capable of determining an emotional response of the user to the social media, and wherein the system is capable of delivering the emotional response to a targeted application for appropriate action.
In one embodiment, the at least one biometric parameter includes galvanic skin response, heart response, motion, skin temperature, breathing, EEG, EMG, pupil dilation, eye tracking, facial recognition, or a combination thereof. In another embodiment, the second device includes at least two biometric sensors. In still another embodiment, the at least two biometric sensors are operable to measure at least two biometric parameters selected from the group consisting of galvanic skin response, heart response, motion, skin temperature, breathing, EEG, EMG, pupil dilation, eye tracking, facial recognition, and combinations thereof.
The first device may further include image acquisition functionality. In one embodiment, the image acquisition functionality includes at least one camera, and wherein the at least one camera is capable of acquiring a plurality of images including at least one image in the user's view, at least one image of the user, or a combination thereof. The system may be capable of image recognition of the at least one image in the user's view, facial recognition of the at least one images of the user, or a combination thereof.
In another embodiment, the system may be capable of collecting motion data to determine activity level, direction, or a combination thereof. In yet another embodiment, the system is capable of determining a current activity of the user. In still another embodiment, the system may be capable of feeding the current activity and the emotional state to listening applications, wherein the listening applications reside on the device or on a central server. The listening application may be capable of providing tailored content to the user based on the emotional state and current activity. The listening application may also be capable of processing feeds of the emotional state and current activity for social media.
Further features and advantages of the invention can be ascertained from the following detailed description that is provided in connection with the drawing(s) described below:
The present invention is directed to asocial media interface that will execute social media functions when pressed, rather than simply open a social media application. In particular, the system and method of the invention will accomplish at least one of the following:
The result of the system and method of the invention is the creation of a continuous stream of social media data that provides a richer and more usable experience than active and discreet “liking” in existing social media applications.
In addition to measuring relevant emotional responses, the system and method of the invention is also capable of computing overall stress and other vital signals of the user to establish the general stress levels and well-being of the user. In one embodiment of the invention, the system and method of the invention is implemented at least in part with a cloud-based infrastructure that allows remote monitoring of the user's vital signs and biological signatures for well-being and derived parameters thereof. This information can then be combined with the user's location and activities over time to further define the drivers of well-being (and the opposite thereof (i.e., stress)). In this aspect of the invention, the system may also include the creation of web/mobile interfaces for sharing the user's well-being and derived parameters thereof with the user's specified social network, specified contacts or groups, as well as with designated family or designated health care providers.
In one embodiment, the system and method of the invention incorporates at least one affinity button or software application on a mobile device, such as a smart phone, a tablet, or other device that is capable of accessing a social media application. For example, to determine the user's current activities, when the affinity button is pressed or the application is started, the device will use some or all software and sensors available to determine the possible set of activities the user may be engaged in. This includes, but is not limited to, all or a subset of the following:
(1) The current application being used;
(2) The user's current location to determine what is located at that location;
(3) Images and/or video within the user's view and/or images of the user;
(4) Neighboring devices; and/or
(5) Motion data.
In particular, with regard to (1), a software layer may determine if the mobile device is currently running any applications and which application is in the foreground. For example, in one embodiment, the process/application using the most central processing unit (cpu) in the current time (e.g., past 15 seconds, past 60 seconds, or another predetermined time interval) may be the application in the foreground. In this aspect, the following steps may be performed:
With regard to (2), the mobile device may use the cellular networks, accessible Wi-Fi and/or GPS to (a) determine the user's current location and (b) search to determine what activities may be available at that location. In particular, the determination of the available activities may favor businesses, parks, addresses in the users' address book, tourist spots, or other activities identifiable through the internet. For example, in one embodiment, locations at the GPS coordinates that have the most details online or in point-of-interest databases will be favored.
With regard to (3), the mobile device may acquire a plurality of snapshots and/or video upon a button press or, if programmed, automatically. The plurality of snapshots and/or video may include images in view of the user and images of the user. In one embodiment, the plurality of images/videos may then be run through image recognition software created for the application to determine the image in view (e.g., a product, location, logo, person, etc.). In another embodiment, the plurality of images/videos may then be run through existing software (such as Google Goggles or the like) to determine the image in view (e.g., a product, location, logo, person, etc.). Facial detection software may also be used to determine which specific facial-emotion is being expressed for the activity.
With regard to (4), the mobile device may use Bluetooth to determine whether there are any known devices in the area. In one embodiment, pressing the affinity button near a second user with the same capability on his/her mobile device will identify that device and add it to the identifiable device list as well as the second user's information to the primary user's phone's contact database.
With regard to (5), motion data may be obtained to determine whether the mobile device is manipulated in such a way to indicate that the user is attempting to capture an image. If such action is determined, the image recognition process described with respect to (3) will be given preference in the activity determination. Motion data may also be used to determine whether the mobile device is in “heavy” motion or “light” motion. In particular, if the mobile device is determined to be in “heavy” motion, i.e., walking briskly, running, and the like, preference will be given to activities requiring more action. Conversely, “light” movement or no movement may indicate less active current activities, such as web browsing or movie watching. Accordingly, preference will be given to less active activities. Motion data may also be used to determine the direction of the mobile device such that, in conjunction with (2) above, an appropriate landmark may be identified. In one embodiment, the motion data may be obtained using at least one of an accelerometer, a magnometer, a gyroscope, or a combination thereof.
With regard to (1)-(5), the system and method of the invention then predict the current activity using a set of likely rules. In particular, the system may use a set of likely rules to determine the most likely relevant activity in the mobile device and/or surrounding environment by monitoring the increases in “recent” time window in the (1) the cpu usage, (2) the memory usage, (3) the decibel (volume) usage, (3) the images being viewed, (4) the online social activity, (5) the online web page interactions, (6) the online or web game interactions, and other similar categories.
More specifically, in one embodiment of the invention, the system and method of the invention will also be capable of selecting a likely activity from the collected set. In particular, machine learning will be employed to determine which potential activity is the activity meant to be tagged via social media. Examples of machine learning in accordance with the present invention include, but are not limited to, the following:
Once an activity is determined, the system will then take the desired social media action(s). Suitable social media actions include, but are not limited to:
In one embodiment, (c) may include a web search based on any or all of the current possible activities. For example, a user is at a sports bar in Location A watching a Soda 1 commercial featuring Singer X with a couple of friends on his/her mobile device. During the commercial, the user holds up the phone to a Picture Z on the wall and presses the affinity button. The system will go through its normal process of determining the liked action, but will also (i) perform a web search on each item returning a search results page split into sections showing information on Location A, Picture Z, the sports bar, their friends, Singer Z and Soda 1 and (ii) flag all of these items in a social media list stored on the mobile device for future search and reference. In other words, (c) may be a form of social media in a natural environment bookmark.
Once these actions are taken, the information can be stored in a form of customer relationship management (CRM) database enabling advertisers to target messages to these users based on items they have flagged using this application. Over time, advertising within the app or across the operating system will become increasingly more directed at the user.
Biometrics may also be integrated into or associated with the device in order to ascertain the emotional state of the phone user. Biometrics include, but are not limited to galvanic skin response (GSR) to measure emotional arousal, heart response (approach/avoid), motion to determine activity level), skin temperature, breathing, EEG, EMG, pupil dilation, eye tracking, and facial recognition. The sensor(s) used to collect the biometrics may be integrated into the hardware platform through additional sensors.
In one embodiment, the biometric data collection is performed through a sensor-array case on the back, sides, and/or front of the existing mobile device. For example, as shown in
In another embodiment, the biometric data collector is incorporated into a wrist or arm band, such as those shown in
In yet another embodiment, the biometric data collector is incorporated into headgear, such as the hat illustrated in
In still another embodiment, the biometric data collector may be incorporated into a smart watch, as generally shown in
In another embodiment, the biometric data collector may be incorporated into a steering wheel of a vehicle. For example, as shown in
Furthermore, while additional sensors may be added directly through integrated hardware, a subset of the sensors may require an external monitoring device (due to sensor needs on other parts of the body). In this aspect of the invention, suitable sensors for collecting EEG and EMG include, but are not limited to, an armband, a pair of glasses, a watch, and similar sensors. In one embodiment, all of the remaining sensors are built into a secondary device. Some, such as a breathing sensor, may work more effectively in a secondary device that communicates with the device of the invention.
In an alternative embodiment, the system of the invention incorporates a device that is not a handheld device. In other words, instead of incorporating the sensors into a mobile phone, the device used is a pair of glasses with attached sensors to the scalp, the forehead, or the like, which are designed with all of the required sensors and equipment as most handheld technology devices. In one aspect, the glasses include a visible screen and sensors (e.g., sensors to the central-parietal regions of the head for EEG, to the forehead, chin or neck for the skin-conductance and to the neck for heart-rate, or other optimal locations on the face/neck of a person for these biological signals), that may have otherwise required a secondary device. The collection of biometric data distinguishes the glasses described in this embodiment from conventional eye tracking goggles.
The biometric data thus collected is intended to determine the emotional state of the user. This emotional state may then be used to activate the social media function with limited button presses. Thus, in one aspect of the invention, the affinity button or application may be used in conjunction with the biometric data. In this aspect, the mobile device (via sensors incorporated directly into the mobile device, into the biometric data collector holding the mobile device, and/or other external secondary devices) is capable of gathering biometric data to determine current physical and emotional state. Affinity buttons may be selected to perform the following steps:
The affinity buttons exist on the mobile device solely for training the system on the user's biometric profile. In particular, as a particular affinity button is pressed, a biometric snapshot is taken of the user to determine the emotional state. The snapshot may then be stored as a biometric fingerprint associated with both the self-described state (positive or negative) and the current activity. As the user builds this dataset, the system begins to suggest content (e.g., music, video, books), products, locations, and the like for future biometric states (without actual button presses). Indeed, at a predetermined stage, the button presses may be completely removed. For example, if every time a user presses the positive affinity button when their heart rate and GSR spike above a threshold while listening to Elvis music, when the user has similar future spikes, the system may recommend listening to similar music.
In addition, the affinity buttons assist in training the system/device. For example, the user presses the particular affinity buttons to train the system that biometric patterns are associated with liking and disliking. In one embodiment of the invention, the system will be able to continuously monitor the user's biometrics. When the system senses a biometric profile/signature pattern matching “like” or “dislike”, the system automatically takes the associated social media action, which would then execute the series of steps listed in the previous function on a continuous basis without the user ever having to press a button. The buttons or applications may be removed or disabled after the system is adequately trained.
In an alternate embodiment of the system, the training algorithm occurs completely outside the biometric device (using a simulated environment), or through predictions of valence and arousal, such as through the methods described in U.S. Pat. No. 8,296,172 and U.S. Patent Publication Nos. 2010/0211439 and 2010/0004977, the entire disclosures of which are incorporated herein by reference.
Alternatively, the system may be a device that has no affinity buttons and only relies on the biometric sensors to accomplish the end goal. In this aspect, no additional training is necessary (existing techniques as described in U.S. Pat. No. 8,296,172 and U.S. Patent Publication Nos. 2010/0211439 and 2010/0004977 for predicting valence, arousal, and engagement, and resonance may be used). In this embodiment, the system continuously determines emotional/biometric state such that whenever a specific emotional state of interest, such as an approach or avoid response, is identified, the series of activities previously described are followed, i.e., determining current activity and taking desired social media actions.
Other embodiments are within the scope and spirit of the invention. For example, functions described above can be implemented and/or automated using software, hardware, firmware, hardwiring, or combinations of any of these. Features implementing functions may also be physically located at various positions, including being distributed such that portions of functions are implemented at different physical locations.
Further, while the description above refers to the invention, the description may include more than one invention.
This application claims priority to U.S. Provisional Patent Application No. 61/603,528, filed Feb. 27, 2012, the entire disclosure of which is incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
2549836 | McIntyre et al. | Apr 1951 | A |
3490439 | Rolston | Jan 1970 | A |
3572322 | Wade | Mar 1971 | A |
3735753 | Pisarski | May 1973 | A |
3880144 | Coursin et al. | Apr 1975 | A |
3901215 | John | Aug 1975 | A |
3998213 | Price | Dec 1976 | A |
4075657 | Weinblatt | Feb 1978 | A |
4145122 | Rinard et al. | Mar 1979 | A |
4149716 | Scudder | Apr 1979 | A |
4201224 | John | May 1980 | A |
4279258 | John | Jul 1981 | A |
4411273 | John | Oct 1983 | A |
4417592 | John | Nov 1983 | A |
4537198 | Corbett | Aug 1985 | A |
4557270 | John | Dec 1985 | A |
4610259 | Cohen et al. | Sep 1986 | A |
4613951 | Chu | Sep 1986 | A |
4626904 | Lurie | Dec 1986 | A |
4632122 | Johansson et al. | Dec 1986 | A |
4683892 | Johansson et al. | Aug 1987 | A |
4686999 | Snyder et al. | Aug 1987 | A |
4695879 | Weinblatt | Sep 1987 | A |
4736751 | Gevins et al. | Apr 1988 | A |
4800888 | Itil et al. | Jan 1989 | A |
4802484 | Friedman et al. | Feb 1989 | A |
4846190 | John | Jul 1989 | A |
4870579 | Hey | Sep 1989 | A |
4885687 | Carey | Dec 1989 | A |
4894777 | Negishi et al. | Jan 1990 | A |
4913160 | John | Apr 1990 | A |
4955388 | Silberstein | Sep 1990 | A |
4967038 | Gevins et al. | Oct 1990 | A |
4973149 | Hutchinson | Nov 1990 | A |
4987903 | Keppel et al. | Jan 1991 | A |
5003986 | Finitzo et al. | Apr 1991 | A |
5010891 | Chamoun | Apr 1991 | A |
5038782 | Gevins et al. | Aug 1991 | A |
5052401 | Sherwin | Oct 1991 | A |
5083571 | Prichep | Jan 1992 | A |
RE34015 | Duffy | Aug 1992 | E |
5137027 | Rosenfeld | Aug 1992 | A |
5213338 | Brotz | May 1993 | A |
5226177 | Nickerson | Jul 1993 | A |
5243517 | Schmidt et al. | Sep 1993 | A |
5273037 | Itil et al. | Dec 1993 | A |
5291888 | Tucker | Mar 1994 | A |
5293867 | Oommen | Mar 1994 | A |
5295491 | Gevins | Mar 1994 | A |
5331544 | Lu et al. | Jul 1994 | A |
5339826 | Schmidt et al. | Aug 1994 | A |
5345281 | Taboada et al. | Sep 1994 | A |
5357957 | Itil et al. | Oct 1994 | A |
5363858 | Farwell | Nov 1994 | A |
5392788 | Hudspeth | Feb 1995 | A |
5406956 | Farwell | Apr 1995 | A |
5410609 | Kado et al. | Apr 1995 | A |
5436830 | Zatlman | Jul 1995 | A |
5447166 | Gevins | Sep 1995 | A |
5474082 | Junker | Dec 1995 | A |
5479934 | Imran | Jan 1996 | A |
5513649 | Gevins et al. | May 1996 | A |
5518007 | Becker | May 1996 | A |
5537618 | Boulton et al. | Jul 1996 | A |
5617855 | Waletzky et al. | Apr 1997 | A |
5655534 | Ilmoniemi | Aug 1997 | A |
5676138 | Zawlinski | Oct 1997 | A |
5676148 | Koo et al. | Oct 1997 | A |
5687322 | Deaton et al. | Nov 1997 | A |
5720619 | Fisslinger | Feb 1998 | A |
5724987 | Gevins et al. | Mar 1998 | A |
5729205 | Kwon | Mar 1998 | A |
5736986 | Sever, Jr. | Apr 1998 | A |
5740035 | Cohen et al. | Apr 1998 | A |
5762611 | Lewis et al. | Jun 1998 | A |
5771897 | Zufrin | Jun 1998 | A |
5774591 | Black et al. | Jun 1998 | A |
5787187 | Bouchard et al. | Jul 1998 | A |
5800351 | Mann | Sep 1998 | A |
5802208 | Podilchuk et al. | Sep 1998 | A |
5802220 | Black et al. | Sep 1998 | A |
5812642 | Leroy | Sep 1998 | A |
5817029 | Gevins et al. | Oct 1998 | A |
5842199 | Miller et al. | Nov 1998 | A |
5848399 | Burke | Dec 1998 | A |
5892566 | Bullwinkel | Apr 1999 | A |
5945863 | Coy | Aug 1999 | A |
5961332 | Joao | Oct 1999 | A |
5974262 | Fuller et al. | Oct 1999 | A |
5983129 | Cowan et al. | Nov 1999 | A |
5995868 | Dorfmeister et al. | Nov 1999 | A |
6001065 | DeVito | Dec 1999 | A |
6021346 | Ryu et al. | Feb 2000 | A |
6032129 | Greef et al. | Feb 2000 | A |
6052619 | John | Apr 2000 | A |
6088040 | Oda et al. | Jul 2000 | A |
6099319 | Zaltman et al. | Aug 2000 | A |
6120440 | Goknar | Sep 2000 | A |
6128521 | Marro et al. | Oct 2000 | A |
6154669 | Hunter et al. | Nov 2000 | A |
6155927 | Levasseur et al. | Dec 2000 | A |
6161030 | Levendowski et al. | Dec 2000 | A |
6170018 | Voll et al. | Jan 2001 | B1 |
6173260 | Slaney | Jan 2001 | B1 |
6175753 | Menkes et al. | Jan 2001 | B1 |
6182113 | Narayanaswami | Jan 2001 | B1 |
6190314 | Ark et al. | Feb 2001 | B1 |
6212502 | Ball et al. | Apr 2001 | B1 |
6228038 | Claessens | May 2001 | B1 |
6236885 | Hunter et al. | May 2001 | B1 |
6236975 | Boe et al. | May 2001 | B1 |
6254536 | DeVito | Jul 2001 | B1 |
6280198 | Calhoun et al. | Aug 2001 | B1 |
6286005 | Cannon | Sep 2001 | B1 |
6289234 | Mueller | Sep 2001 | B1 |
6292688 | Patton | Sep 2001 | B1 |
6299308 | Voronka et al. | Oct 2001 | B1 |
6301493 | Marro et al. | Oct 2001 | B1 |
6315569 | Zaltman | Nov 2001 | B1 |
6330470 | Tucker et al. | Dec 2001 | B1 |
6334778 | Brown | Jan 2002 | B1 |
6358201 | Childre et al. | Mar 2002 | B1 |
6370513 | Kolawa et al. | Apr 2002 | B1 |
6374143 | Berrang et al. | Apr 2002 | B1 |
6381481 | Levendowski et al. | Apr 2002 | B1 |
6398643 | Knowles et al. | Jun 2002 | B1 |
6422999 | Hill | Jul 2002 | B1 |
6434419 | Gevins et al. | Aug 2002 | B1 |
6435878 | Reynolds et al. | Aug 2002 | B1 |
6453194 | Hill | Sep 2002 | B1 |
6453241 | Bassett, Jr. et al. | Sep 2002 | B1 |
6487444 | Mimura | Nov 2002 | B2 |
6488617 | Katz | Dec 2002 | B1 |
6510340 | Jordan | Jan 2003 | B1 |
6520905 | Surve et al. | Feb 2003 | B1 |
6545685 | Dorbie | Apr 2003 | B1 |
6575902 | Burton | Jun 2003 | B1 |
6577329 | Flickner et al. | Jun 2003 | B1 |
6585521 | Obrador | Jul 2003 | B1 |
6594521 | Tucker | Jul 2003 | B2 |
6598006 | Honda et al. | Jul 2003 | B1 |
6609024 | Ryu et al. | Aug 2003 | B1 |
6648822 | Hamamoto et al. | Nov 2003 | B2 |
6652283 | Van Schaack et al. | Nov 2003 | B1 |
6654626 | Devlin et al. | Nov 2003 | B2 |
6662052 | Sarwal et al. | Dec 2003 | B1 |
6665560 | Becker et al. | Dec 2003 | B2 |
6678685 | McGill et al. | Jan 2004 | B2 |
6688890 | von Buegner | Feb 2004 | B2 |
6708051 | Durousseau | Mar 2004 | B1 |
6712468 | Edwards | Mar 2004 | B1 |
6754524 | Johnson, Jr. | Jun 2004 | B2 |
6757556 | Gopenathan et al. | Jun 2004 | B2 |
6788882 | Geer et al. | Sep 2004 | B1 |
6792304 | Silberstein | Sep 2004 | B1 |
6842877 | Robarts et al. | Jan 2005 | B2 |
6850252 | Hoffberg | Feb 2005 | B1 |
6852875 | Prakash | Feb 2005 | B2 |
6888457 | Wilkinson et al. | May 2005 | B2 |
6904408 | McCarthy et al. | Jun 2005 | B1 |
6950698 | Sarkela et al. | Sep 2005 | B2 |
6958710 | Zhang et al. | Oct 2005 | B2 |
6973342 | Swanson | Dec 2005 | B1 |
6993380 | Modarres | Jan 2006 | B1 |
7020508 | Stivoric et al. | Mar 2006 | B2 |
7043056 | Edwards et al. | May 2006 | B2 |
7113916 | Hill | Sep 2006 | B1 |
7120880 | Dryer et al. | Oct 2006 | B1 |
7130673 | Tolvanen-Laakso et al. | Oct 2006 | B2 |
7150715 | Collura et al. | Dec 2006 | B2 |
7164967 | Etienne-Cummings et al. | Jan 2007 | B2 |
7177675 | Suffin et al. | Feb 2007 | B2 |
7222071 | Neuhauser et al. | May 2007 | B2 |
7246081 | Hill | Jul 2007 | B2 |
7249708 | McConnell et al. | Jul 2007 | B2 |
7269590 | Hull et al. | Sep 2007 | B2 |
7272982 | Neuhauser et al. | Sep 2007 | B2 |
7286871 | Cohen | Oct 2007 | B2 |
7340060 | Tomkins et al. | Mar 2008 | B2 |
7359894 | Liebman et al. | Apr 2008 | B1 |
7391835 | Gross et al. | Jun 2008 | B1 |
7408460 | Crystal et al. | Aug 2008 | B2 |
7420464 | Fitzgerald et al. | Sep 2008 | B2 |
7443292 | Jensen et al. | Oct 2008 | B2 |
7460827 | Schuster et al. | Dec 2008 | B2 |
7463143 | Forr et al. | Dec 2008 | B2 |
7463144 | Crystal et al. | Dec 2008 | B2 |
7471987 | Crystal et al. | Dec 2008 | B2 |
7483835 | Neuhauser et al. | Jan 2009 | B2 |
7483844 | Takakura et al. | Jan 2009 | B2 |
7496400 | Hoskonen et al. | Feb 2009 | B2 |
7548774 | Kurtz et al. | Jun 2009 | B2 |
7551952 | Gevins et al. | Jun 2009 | B2 |
7592908 | Zhang et al. | Sep 2009 | B2 |
7614066 | Urdang et al. | Nov 2009 | B2 |
7623823 | Zito et al. | Nov 2009 | B2 |
7630757 | Dorfmeister et al. | Dec 2009 | B2 |
7636456 | Collins et al. | Dec 2009 | B2 |
7650793 | Jensen et al. | Jan 2010 | B2 |
7657523 | Ebanks | Feb 2010 | B2 |
7689272 | Farwell | Mar 2010 | B2 |
7697979 | Martinerie et al. | Apr 2010 | B2 |
7698238 | Barletta et al. | Apr 2010 | B2 |
7720351 | Levitan | May 2010 | B2 |
7729755 | Laken | Jun 2010 | B2 |
7797186 | Dybus | Sep 2010 | B2 |
7809420 | Hannula et al. | Oct 2010 | B2 |
7840248 | Fuchs et al. | Nov 2010 | B2 |
7840250 | Tucker | Nov 2010 | B2 |
7844484 | Arnett et al. | Nov 2010 | B2 |
7865394 | Calloway | Jan 2011 | B1 |
7892764 | Xiong et al. | Feb 2011 | B2 |
7895075 | Gettys et al. | Feb 2011 | B2 |
7908133 | Neuhauser | Mar 2011 | B2 |
7917366 | Levanon et al. | Mar 2011 | B1 |
7930199 | Hill | Apr 2011 | B1 |
7946974 | Lordereau | May 2011 | B2 |
7962315 | Jensen et al. | Jun 2011 | B2 |
7974889 | Raimbeault | Jul 2011 | B2 |
7988557 | Soderlund | Aug 2011 | B2 |
8014847 | Shastri et al. | Sep 2011 | B2 |
8027518 | Baker et al. | Sep 2011 | B2 |
8055722 | Hille | Nov 2011 | B2 |
8060795 | Bakekolo et al. | Nov 2011 | B2 |
8065203 | Chien et al. | Nov 2011 | B1 |
8069125 | Jung et al. | Nov 2011 | B2 |
8073707 | Teller et al. | Dec 2011 | B2 |
8079054 | Dhawan et al. | Dec 2011 | B1 |
8082215 | Jung et al. | Dec 2011 | B2 |
8086563 | Jung et al. | Dec 2011 | B2 |
8098152 | Zhang et al. | Jan 2012 | B2 |
8103328 | Turner et al. | Jan 2012 | B2 |
8135606 | Dupree | Mar 2012 | B2 |
8151298 | Begeja et al. | Apr 2012 | B2 |
8165916 | Hoffberg et al. | Apr 2012 | B2 |
8179604 | Prada Gomez et al. | May 2012 | B1 |
8200775 | Moore | Jun 2012 | B2 |
8209224 | Pradeep et al. | Jun 2012 | B2 |
8229469 | Zhang et al. | Jul 2012 | B2 |
8235725 | Hill | Aug 2012 | B1 |
8239030 | Hagedorn et al. | Aug 2012 | B1 |
8255267 | Breiter | Aug 2012 | B2 |
8270814 | Pradeep et al. | Sep 2012 | B2 |
8296172 | Marci et al. | Oct 2012 | B2 |
8300526 | Saito et al. | Oct 2012 | B2 |
8326002 | Hill | Dec 2012 | B2 |
8327395 | Lee | Dec 2012 | B2 |
8332883 | Lee | Dec 2012 | B2 |
8335715 | Pradeep et al. | Dec 2012 | B2 |
8386312 | Pradeep et al. | Feb 2013 | B2 |
8386313 | Pradeep et al. | Feb 2013 | B2 |
8388165 | Zhang | Mar 2013 | B2 |
8392250 | Pradeep et al. | Mar 2013 | B2 |
8392251 | Pradeep et al. | Mar 2013 | B2 |
8392253 | Pradeep et al. | Mar 2013 | B2 |
8392254 | Pradeep et al. | Mar 2013 | B2 |
8392255 | Pradeep et al. | Mar 2013 | B2 |
8396744 | Pradeep et al. | Mar 2013 | B2 |
8442429 | Hawit | May 2013 | B2 |
8467133 | Miller | Jun 2013 | B2 |
8473345 | Pradeep et al. | Jun 2013 | B2 |
8477425 | Border et al. | Jul 2013 | B2 |
8484081 | Pradeep et al. | Jul 2013 | B2 |
8494610 | Pradeep et al. | Jul 2013 | B2 |
8494905 | Pradeep et al. | Jul 2013 | B2 |
8533042 | Pradeep et al. | Sep 2013 | B2 |
8543446 | Richardson et al. | Sep 2013 | B2 |
8548852 | Pradeep et al. | Oct 2013 | B2 |
8560530 | Krichman et al. | Oct 2013 | B2 |
8561095 | Dimitrova et al. | Oct 2013 | B2 |
8600100 | Hill | Dec 2013 | B2 |
8635105 | Pradeep et al. | Jan 2014 | B2 |
8655428 | Pradeep et al. | Feb 2014 | B2 |
8655437 | Pradeep et al. | Feb 2014 | B2 |
8684742 | Siefert | Apr 2014 | B2 |
8700009 | Quy | Apr 2014 | B2 |
8764652 | Lee et al. | Jul 2014 | B2 |
8788372 | Kettner et al. | Jul 2014 | B2 |
8793715 | Weitzenfeld et al. | Jul 2014 | B1 |
8793727 | Serdiuk | Jul 2014 | B2 |
8856235 | Zhou et al. | Oct 2014 | B2 |
8874727 | Swahar | Oct 2014 | B2 |
20010013009 | Greening et al. | Aug 2001 | A1 |
20010020236 | Cannon | Sep 2001 | A1 |
20010029468 | Yamaguchi et al. | Oct 2001 | A1 |
20010056225 | DeVito | Dec 2001 | A1 |
20020053076 | Landesmann | May 2002 | A1 |
20020055857 | Mault | May 2002 | A1 |
20020056124 | Hay | May 2002 | A1 |
20020059577 | Lu et al. | May 2002 | A1 |
20020065826 | Bell et al. | May 2002 | A1 |
20020072952 | Hamzy et al. | Jun 2002 | A1 |
20020077534 | DuRousseau | Jun 2002 | A1 |
20020082902 | Ando et al. | Jun 2002 | A1 |
20020103429 | deCharms | Aug 2002 | A1 |
20020143627 | Barsade et al. | Oct 2002 | A1 |
20020155878 | Lert, Jr. et al. | Oct 2002 | A1 |
20020156842 | Signes et al. | Oct 2002 | A1 |
20020169665 | Hughes et al. | Nov 2002 | A1 |
20020188216 | Kayyali et al. | Dec 2002 | A1 |
20020188217 | Farwell | Dec 2002 | A1 |
20020193670 | Garfield et al. | Dec 2002 | A1 |
20030013981 | Gevins et al. | Jan 2003 | A1 |
20030036955 | Tanaka et al. | Feb 2003 | A1 |
20030037333 | Ghashghai et al. | Feb 2003 | A1 |
20030044050 | Clark | Mar 2003 | A1 |
20030059750 | Bindler et al. | Mar 2003 | A1 |
20030063222 | Creed et al. | Apr 2003 | A1 |
20030065524 | Giacchetti et al. | Apr 2003 | A1 |
20030073921 | Sohmer et al. | Apr 2003 | A1 |
20030081834 | Philomin et al. | May 2003 | A1 |
20030093792 | Labeeb et al. | May 2003 | A1 |
20030100998 | Brunner et al. | May 2003 | A2 |
20030104865 | Itkis et al. | Jun 2003 | A1 |
20030149344 | Nizan | Aug 2003 | A1 |
20030165270 | Endrikhovski et al. | Sep 2003 | A1 |
20030177488 | Smith et al. | Sep 2003 | A1 |
20030204412 | Brier | Oct 2003 | A1 |
20030208754 | Sridhar et al. | Nov 2003 | A1 |
20030233278 | Marshall | Dec 2003 | A1 |
20040001616 | Gutta et al. | Jan 2004 | A1 |
20040005143 | Tsuru et al. | Jan 2004 | A1 |
20040013398 | Miura et al. | Jan 2004 | A1 |
20040015608 | Ellis et al. | Jan 2004 | A1 |
20040055448 | Byon | Mar 2004 | A1 |
20040068431 | Smith et al. | Apr 2004 | A1 |
20040073129 | Caldwell et al. | Apr 2004 | A1 |
20040092809 | DeCharms | May 2004 | A1 |
20040098298 | Yin | May 2004 | A1 |
20040133081 | Teller et al. | Jul 2004 | A1 |
20040187167 | Maguire et al. | Sep 2004 | A1 |
20040193068 | Burton et al. | Sep 2004 | A1 |
20040210159 | Kibar | Oct 2004 | A1 |
20040219184 | Brown et al. | Nov 2004 | A1 |
20040220483 | Yeo et al. | Nov 2004 | A1 |
20040236623 | Gopalakrishnan | Nov 2004 | A1 |
20050010475 | Perkowski et al. | Jan 2005 | A1 |
20050041951 | Inoue et al. | Feb 2005 | A1 |
20050060312 | Curtiss et al. | Mar 2005 | A1 |
20050062637 | El Zabadani et al. | Mar 2005 | A1 |
20050071462 | Bodin et al. | Mar 2005 | A1 |
20050071865 | Martins | Mar 2005 | A1 |
20050076359 | Pierson et al. | Apr 2005 | A1 |
20050079474 | Lowe | Apr 2005 | A1 |
20050097594 | O'Donnell et al. | May 2005 | A1 |
20050107716 | Eaton et al. | May 2005 | A1 |
20050113649 | Bergantino | May 2005 | A1 |
20050143629 | Farwell | Jun 2005 | A1 |
20050154290 | Langleben | Jul 2005 | A1 |
20050177058 | Sobell | Aug 2005 | A1 |
20050197590 | Osorio et al. | Sep 2005 | A1 |
20050203798 | Jensen et al. | Sep 2005 | A1 |
20050216243 | Graham et al. | Sep 2005 | A1 |
20050223237 | Barletta et al. | Oct 2005 | A1 |
20050227233 | Buxton et al. | Oct 2005 | A1 |
20050240956 | Smith et al. | Oct 2005 | A1 |
20050256905 | Gruhl et al. | Nov 2005 | A1 |
20050267798 | Panara | Dec 2005 | A1 |
20050272017 | Neuhauser et al. | Dec 2005 | A1 |
20050273017 | Gordon | Dec 2005 | A1 |
20050273802 | Crystal et al. | Dec 2005 | A1 |
20050288954 | McCarthy et al. | Dec 2005 | A1 |
20050289582 | Tavares et al. | Dec 2005 | A1 |
20060003732 | Neuhauser et al. | Jan 2006 | A1 |
20060035707 | Nguyen et al. | Feb 2006 | A1 |
20060041548 | Parsons et al. | Feb 2006 | A1 |
20060042483 | Work et al. | Mar 2006 | A1 |
20060053110 | McDonald et al. | Mar 2006 | A1 |
20060069663 | Adar et al. | Mar 2006 | A1 |
20060075003 | Adams et al. | Apr 2006 | A1 |
20060093998 | Vertegaal | May 2006 | A1 |
20060111044 | Keller | May 2006 | A1 |
20060111644 | Guttag et al. | May 2006 | A1 |
20060129458 | Maggio | Jun 2006 | A1 |
20060149337 | John | Jul 2006 | A1 |
20060167376 | Viirre et al. | Jul 2006 | A1 |
20060168613 | Wood et al. | Jul 2006 | A1 |
20060168630 | Davies | Jul 2006 | A1 |
20060176289 | Horn | Aug 2006 | A1 |
20060190822 | Basson et al. | Aug 2006 | A1 |
20060256133 | Rosenberg | Nov 2006 | A1 |
20060257834 | Lee et al. | Nov 2006 | A1 |
20060259360 | Flinn et al. | Nov 2006 | A1 |
20060259371 | Perrier et al. | Nov 2006 | A1 |
20060293921 | McCarthy et al. | Dec 2006 | A1 |
20070016096 | McNabb | Jan 2007 | A1 |
20070038516 | Apple et al. | Feb 2007 | A1 |
20070048707 | Caamano et al. | Mar 2007 | A1 |
20070055169 | Lee et al. | Mar 2007 | A1 |
20070060830 | Le et al. | Mar 2007 | A1 |
20070060831 | Le et al. | Mar 2007 | A1 |
20070061720 | Kriger | Mar 2007 | A1 |
20070066874 | Cook | Mar 2007 | A1 |
20070066915 | Frei et al. | Mar 2007 | A1 |
20070066916 | Lemos | Mar 2007 | A1 |
20070067007 | Schulman et al. | Mar 2007 | A1 |
20070067305 | Ives | Mar 2007 | A1 |
20070078700 | Lenzmann et al. | Apr 2007 | A1 |
20070078706 | Datta et al. | Apr 2007 | A1 |
20070079331 | Datta et al. | Apr 2007 | A1 |
20070101360 | Gutta et al. | May 2007 | A1 |
20070106170 | Dunseath, Jr. et al. | May 2007 | A1 |
20070112460 | Kiselik | May 2007 | A1 |
20070135727 | Virtanen et al. | Jun 2007 | A1 |
20070135728 | Snyder et al. | Jun 2007 | A1 |
20070150916 | Begole et al. | Jun 2007 | A1 |
20070198510 | Ebanks | Aug 2007 | A1 |
20070214121 | Ebanks | Sep 2007 | A1 |
20070225585 | Washbon et al. | Sep 2007 | A1 |
20070225674 | Molnar et al. | Sep 2007 | A1 |
20070226760 | Neuhauser et al. | Sep 2007 | A1 |
20070235716 | Delic et al. | Oct 2007 | A1 |
20070238945 | Delic et al. | Oct 2007 | A1 |
20070239713 | Leblang | Oct 2007 | A1 |
20070250846 | Swix et al. | Oct 2007 | A1 |
20070250901 | McIntire et al. | Oct 2007 | A1 |
20070265507 | de Lemos | Nov 2007 | A1 |
20070282566 | Whitlow et al. | Dec 2007 | A1 |
20070294132 | Zhang et al. | Dec 2007 | A1 |
20070294705 | Gopalakrishnan | Dec 2007 | A1 |
20070294706 | Neuhauser et al. | Dec 2007 | A1 |
20080001600 | deCharms | Jan 2008 | A1 |
20080004940 | Rolleston Phillips | Jan 2008 | A1 |
20080010110 | Neuhauser et al. | Jan 2008 | A1 |
20080027345 | Kumada et al. | Jan 2008 | A1 |
20080040740 | Plotnick et al. | Feb 2008 | A1 |
20080059997 | Plotnick et al. | Mar 2008 | A1 |
20080065468 | Berg et al. | Mar 2008 | A1 |
20080065721 | Cragun | Mar 2008 | A1 |
20080081961 | Westbrook et al. | Apr 2008 | A1 |
20080082019 | Ludving et al. | Apr 2008 | A1 |
20080086356 | Glassman et al. | Apr 2008 | A1 |
20080091463 | Shakamuri | Apr 2008 | A1 |
20080091512 | Marci et al. | Apr 2008 | A1 |
20080097854 | Young | Apr 2008 | A1 |
20080109840 | Walter et al. | May 2008 | A1 |
20080125110 | Ritter | May 2008 | A1 |
20080133724 | Clark | Jun 2008 | A1 |
20080147488 | Tunick et al. | Jun 2008 | A1 |
20080147742 | Allen | Jun 2008 | A1 |
20080152300 | Knee et al. | Jun 2008 | A1 |
20080162182 | Cazares et al. | Jul 2008 | A1 |
20080195471 | Dube et al. | Aug 2008 | A1 |
20080204273 | Crystal et al. | Aug 2008 | A1 |
20080208072 | Fadem et al. | Aug 2008 | A1 |
20080214902 | Lee et al. | Sep 2008 | A1 |
20080218472 | Breen et al. | Sep 2008 | A1 |
20080221400 | Lee et al. | Sep 2008 | A1 |
20080221472 | Lee et al. | Sep 2008 | A1 |
20080221969 | Lee et al. | Sep 2008 | A1 |
20080222670 | Lee et al. | Sep 2008 | A1 |
20080222671 | Lee et al. | Sep 2008 | A1 |
20080228077 | Wilk et al. | Sep 2008 | A1 |
20080249865 | Angell et al. | Oct 2008 | A1 |
20080255949 | Genco et al. | Oct 2008 | A1 |
20080287821 | Jung | Nov 2008 | A1 |
20080295126 | Lee et al. | Nov 2008 | A1 |
20080306398 | Uchiyama et al. | Dec 2008 | A1 |
20090018996 | Hunt et al. | Jan 2009 | A1 |
20090024049 | Pradeep et al. | Jan 2009 | A1 |
20090024447 | Pradeep et al. | Jan 2009 | A1 |
20090024448 | Pradeep et al. | Jan 2009 | A1 |
20090024449 | Pradeep et al. | Jan 2009 | A1 |
20090024475 | Pradeep et al. | Jan 2009 | A1 |
20090024747 | Moses et al. | Jan 2009 | A1 |
20090025023 | Pradeep et al. | Jan 2009 | A1 |
20090025024 | Beser et al. | Jan 2009 | A1 |
20090030287 | Pradeep et al. | Jan 2009 | A1 |
20090030303 | Pradeep et al. | Jan 2009 | A1 |
20090030717 | Pradeep et al. | Jan 2009 | A1 |
20090030762 | Lee et al. | Jan 2009 | A1 |
20090030780 | York et al. | Jan 2009 | A1 |
20090030930 | Pradeep et al. | Jan 2009 | A1 |
20090036755 | Pradeep et al. | Feb 2009 | A1 |
20090036756 | Pradeep et al. | Feb 2009 | A1 |
20090037575 | Crystal et al. | Feb 2009 | A1 |
20090060240 | Coughlan et al. | Mar 2009 | A1 |
20090062629 | Pradeep et al. | Mar 2009 | A1 |
20090062679 | Tan et al. | Mar 2009 | A1 |
20090062680 | Sandford | Mar 2009 | A1 |
20090062681 | Pradeep et al. | Mar 2009 | A1 |
20090063255 | Pradeep et al. | Mar 2009 | A1 |
20090063256 | Pradeep et al. | Mar 2009 | A1 |
20090066722 | Kriger | Mar 2009 | A1 |
20090069652 | Lee et al. | Mar 2009 | A1 |
20090070219 | D'Angelo et al. | Mar 2009 | A1 |
20090070798 | Lee et al. | Mar 2009 | A1 |
20090082643 | Pradeep et al. | Mar 2009 | A1 |
20090082689 | Guttag et al. | Mar 2009 | A1 |
20090083129 | Pradeep et al. | Mar 2009 | A1 |
20090088610 | Lee et al. | Apr 2009 | A1 |
20090089830 | Chandratillake et al. | Apr 2009 | A1 |
20090094286 | Lee et al. | Apr 2009 | A1 |
20090094627 | Lee et al. | Apr 2009 | A1 |
20090094628 | Lee et al. | Apr 2009 | A1 |
20090094629 | Lee et al. | Apr 2009 | A1 |
20090097689 | Prest et al. | Apr 2009 | A1 |
20090099873 | Kurple | Apr 2009 | A1 |
20090112077 | Nguyen et al. | Apr 2009 | A1 |
20090112117 | Rewari | Apr 2009 | A1 |
20090119154 | Jung et al. | May 2009 | A1 |
20090131764 | Lee et al. | May 2009 | A1 |
20090132441 | Muller et al. | May 2009 | A1 |
20090133047 | Lee et al. | May 2009 | A1 |
20090138356 | Pomplun | May 2009 | A1 |
20090144780 | Toebes et al. | Jun 2009 | A1 |
20090150919 | Lee et al. | Jun 2009 | A1 |
20090150920 | Jones | Jun 2009 | A1 |
20090153328 | Otani et al. | Jun 2009 | A1 |
20090156907 | Jung et al. | Jun 2009 | A1 |
20090156955 | Jung et al. | Jun 2009 | A1 |
20090158308 | Weitzenfeld et al. | Jun 2009 | A1 |
20090163777 | Jung | Jun 2009 | A1 |
20090164132 | Jung et al. | Jun 2009 | A1 |
20090171164 | Jung et al. | Jul 2009 | A1 |
20090187467 | Fang et al. | Jul 2009 | A1 |
20090195392 | Zalewski | Aug 2009 | A1 |
20090214060 | Chuang et al. | Aug 2009 | A1 |
20090216611 | Leonard et al. | Aug 2009 | A1 |
20090221928 | Einav et al. | Sep 2009 | A1 |
20090248484 | Surendran et al. | Oct 2009 | A1 |
20090248496 | Hueter et al. | Oct 2009 | A1 |
20090248594 | Castleman et al. | Oct 2009 | A1 |
20090253996 | Lee et al. | Oct 2009 | A1 |
20090259137 | Delic et al. | Oct 2009 | A1 |
20090259509 | Landvater | Oct 2009 | A1 |
20090287108 | Levy | Nov 2009 | A1 |
20090292587 | Fitzgerald | Nov 2009 | A1 |
20090300672 | Van Gulik | Dec 2009 | A1 |
20090305006 | Steffen | Dec 2009 | A1 |
20090318773 | Jung et al. | Dec 2009 | A1 |
20090318826 | Green et al. | Dec 2009 | A1 |
20090327068 | Pradeep et al. | Dec 2009 | A1 |
20090327907 | Estrada et al. | Dec 2009 | A1 |
20090328089 | Pradeep et al. | Dec 2009 | A1 |
20100004977 | Marci et al. | Jan 2010 | A1 |
20100022821 | Dubi et al. | Jan 2010 | A1 |
20100039618 | De Lemos | Feb 2010 | A1 |
20100041962 | Causevic et al. | Feb 2010 | A1 |
20100042012 | Alhussiny | Feb 2010 | A1 |
20100060300 | Muller et al. | Mar 2010 | A1 |
20100094702 | Silberstein | Apr 2010 | A1 |
20100094869 | Ebanks | Apr 2010 | A1 |
20100121716 | Golan | May 2010 | A1 |
20100125219 | Harris et al. | May 2010 | A1 |
20100145176 | Himes | Jun 2010 | A1 |
20100145215 | Pradeep et al. | Jun 2010 | A1 |
20100145217 | Otto et al. | Jun 2010 | A1 |
20100153175 | Pearson et al. | Jun 2010 | A1 |
20100169153 | Hwacinski et al. | Jul 2010 | A1 |
20100169162 | Anderson et al. | Jul 2010 | A1 |
20100179881 | Wiederstein | Jul 2010 | A1 |
20100180029 | Fourman | Jul 2010 | A1 |
20100183279 | Pradeep et al. | Jul 2010 | A1 |
20100186031 | Pradeep et al. | Jul 2010 | A1 |
20100186032 | Pradeep et al. | Jul 2010 | A1 |
20100198042 | Popescu et al. | Aug 2010 | A1 |
20100211439 | Marci et al. | Aug 2010 | A1 |
20100214318 | Pradeep et al. | Aug 2010 | A1 |
20100215289 | Pradeep et al. | Aug 2010 | A1 |
20100218208 | Holden | Aug 2010 | A1 |
20100223094 | Cumming et al. | Sep 2010 | A1 |
20100228604 | Desai et al. | Sep 2010 | A1 |
20100228614 | Zhang et al. | Sep 2010 | A1 |
20100234752 | Sullivan et al. | Sep 2010 | A1 |
20100241580 | Schleier-Smith | Sep 2010 | A1 |
20100249538 | Pradeep et al. | Sep 2010 | A1 |
20100249636 | Pradeep et al. | Sep 2010 | A1 |
20100250325 | Pradeep et al. | Sep 2010 | A1 |
20100250458 | Ho | Sep 2010 | A1 |
20100257023 | Kendall et al. | Oct 2010 | A1 |
20100257052 | Zito et al. | Oct 2010 | A1 |
20100262477 | Hillerbrand et al. | Oct 2010 | A1 |
20100263005 | White | Oct 2010 | A1 |
20100268540 | Arshi et al. | Oct 2010 | A1 |
20100268573 | Jain et al. | Oct 2010 | A1 |
20100268720 | Spivack et al. | Oct 2010 | A1 |
20100269127 | Krug | Oct 2010 | A1 |
20100274152 | McPeck et al. | Oct 2010 | A1 |
20100274153 | Tucker et al. | Oct 2010 | A1 |
20100287152 | Hauser | Nov 2010 | A1 |
20100292998 | Bodlaender et al. | Nov 2010 | A1 |
20100306120 | Ciptawilangga | Dec 2010 | A1 |
20100317988 | Terada et al. | Dec 2010 | A1 |
20100318507 | Grant et al. | Dec 2010 | A1 |
20100325660 | Holden | Dec 2010 | A1 |
20100331661 | Nakagawa | Dec 2010 | A1 |
20100332283 | Ng et al. | Dec 2010 | A1 |
20110004089 | Chou | Jan 2011 | A1 |
20110015503 | Joffe et al. | Jan 2011 | A1 |
20110020778 | Forbes | Jan 2011 | A1 |
20110022965 | Lawrence et al. | Jan 2011 | A1 |
20110040155 | Guzak et al. | Feb 2011 | A1 |
20110040202 | Luo et al. | Feb 2011 | A1 |
20110043759 | Bushinsky | Feb 2011 | A1 |
20110046473 | Pradeep et al. | Feb 2011 | A1 |
20110046502 | Pradeep et al. | Feb 2011 | A1 |
20110046503 | Pradeep et al. | Feb 2011 | A1 |
20110046504 | Pradeep et al. | Feb 2011 | A1 |
20110047035 | Gidwani et al. | Feb 2011 | A1 |
20110047121 | Pradeep et al. | Feb 2011 | A1 |
20110059422 | Masaoka | Mar 2011 | A1 |
20110071874 | Schneersohn et al. | Mar 2011 | A1 |
20110076942 | Taveau | Mar 2011 | A1 |
20110084795 | Fukuyori | Apr 2011 | A1 |
20110085700 | Lee | Apr 2011 | A1 |
20110098593 | Low et al. | Apr 2011 | A1 |
20110105937 | Pradeep et al. | May 2011 | A1 |
20110106621 | Pradeep et al. | May 2011 | A1 |
20110106750 | Pradeep et al. | May 2011 | A1 |
20110119124 | Pradeep et al. | May 2011 | A1 |
20110119129 | Pradeep et al. | May 2011 | A1 |
20110119130 | Agan et al. | May 2011 | A1 |
20110124977 | Winarski | May 2011 | A1 |
20110131274 | Hille | Jun 2011 | A1 |
20110137894 | Narayanan et al. | Jun 2011 | A1 |
20110144519 | Causevic | Jun 2011 | A1 |
20110153390 | Harris | Jun 2011 | A1 |
20110153391 | Tenbrock | Jun 2011 | A1 |
20110153414 | Elvekrog et al. | Jun 2011 | A1 |
20110153423 | Elvekrog et al. | Jun 2011 | A1 |
20110161095 | Line et al. | Jun 2011 | A1 |
20110161163 | Carlson et al. | Jun 2011 | A1 |
20110161790 | Junior et al. | Jun 2011 | A1 |
20110191142 | Huang | Aug 2011 | A1 |
20110208515 | Neuhauser | Aug 2011 | A1 |
20110213670 | Strutton et al. | Sep 2011 | A1 |
20110224569 | Isenhart et al. | Sep 2011 | A1 |
20110237923 | Picht et al. | Sep 2011 | A1 |
20110237971 | Pradeep et al. | Sep 2011 | A1 |
20110246574 | Lento et al. | Oct 2011 | A1 |
20110248729 | Mueller et al. | Oct 2011 | A2 |
20110256520 | Siefert | Oct 2011 | A1 |
20110257502 | Lee | Oct 2011 | A1 |
20110257937 | Lee | Oct 2011 | A1 |
20110270620 | Pradeep et al. | Nov 2011 | A1 |
20110276504 | Pradeep et al. | Nov 2011 | A1 |
20110282231 | Pradeep et al. | Nov 2011 | A1 |
20110282232 | Pradeep et al. | Nov 2011 | A1 |
20110282749 | Pradeep et al. | Nov 2011 | A1 |
20110282880 | Krichman et al. | Nov 2011 | A1 |
20110296004 | Swahar | Dec 2011 | A1 |
20110298706 | Mann | Dec 2011 | A1 |
20110301431 | Greicius | Dec 2011 | A1 |
20110313849 | Brueck et al. | Dec 2011 | A1 |
20110319975 | Ho et al. | Dec 2011 | A1 |
20120002848 | Hill | Jan 2012 | A1 |
20120004899 | Arshi | Jan 2012 | A1 |
20120022391 | Leuthardt | Jan 2012 | A1 |
20120036004 | Pradeep et al. | Feb 2012 | A1 |
20120036005 | Pradeep et al. | Feb 2012 | A1 |
20120046993 | Hill | Feb 2012 | A1 |
20120054018 | Pradeep et al. | Mar 2012 | A1 |
20120072289 | Pradeep et al. | Mar 2012 | A1 |
20120072845 | John et al. | Mar 2012 | A1 |
20120072936 | Small et al. | Mar 2012 | A1 |
20120078065 | De Lemos et al. | Mar 2012 | A1 |
20120083668 | Pradeep et al. | Apr 2012 | A1 |
20120084139 | Pradeep et al. | Apr 2012 | A1 |
20120089552 | Chang et al. | Apr 2012 | A1 |
20120096363 | Barnes et al. | Apr 2012 | A1 |
20120108995 | Pradeep et al. | May 2012 | A1 |
20120114305 | Holden | May 2012 | A1 |
20120130800 | Pradeep et al. | May 2012 | A1 |
20120166252 | Walker | Jun 2012 | A1 |
20120173701 | Tenbrock | Jul 2012 | A1 |
20120203363 | McKenna et al. | Aug 2012 | A1 |
20120203559 | McKenna et al. | Aug 2012 | A1 |
20120239407 | Jain et al. | Sep 2012 | A1 |
20120245978 | Crystal et al. | Sep 2012 | A1 |
20120249797 | Haddick et al. | Oct 2012 | A1 |
20120254909 | Serdiuk | Oct 2012 | A1 |
20120272256 | Bedi | Oct 2012 | A1 |
20120284332 | Pradeep et al. | Nov 2012 | A1 |
20120290637 | Perantatos et al. | Nov 2012 | A1 |
20120296699 | Richardson et al. | Nov 2012 | A1 |
20120317198 | Patton et al. | Dec 2012 | A1 |
20130018949 | Pradeep | Jan 2013 | A1 |
20130022948 | Angell et al. | Jan 2013 | A1 |
20130024272 | Pradeep et al. | Jan 2013 | A1 |
20130046577 | Marci et al. | Feb 2013 | A1 |
20130060125 | Zeman | Mar 2013 | A1 |
20130094722 | Hill | Apr 2013 | A1 |
20130097715 | Fourman | Apr 2013 | A1 |
20130121591 | Hill | May 2013 | A1 |
20130124365 | Pradeep | May 2013 | A1 |
20130143185 | Liu et al. | Jun 2013 | A1 |
20130152506 | Pradeep | Jun 2013 | A1 |
20130166373 | Pradeep et al. | Jun 2013 | A1 |
20130183646 | Lusted et al. | Jul 2013 | A1 |
20130185140 | Pradeep et al. | Jul 2013 | A1 |
20130185141 | Pradeep et al. | Jul 2013 | A1 |
20130185142 | Pradeep et al. | Jul 2013 | A1 |
20130185144 | Pradeep et al. | Jul 2013 | A1 |
20130185145 | Pradeep et al. | Jul 2013 | A1 |
20130268279 | Srinivasan et al. | Oct 2013 | A1 |
20130280682 | Levine et al. | Oct 2013 | A1 |
20130304540 | Pradeep et al. | Nov 2013 | A1 |
20130311132 | Tobita | Nov 2013 | A1 |
20130332259 | Pradeep et al. | Dec 2013 | A1 |
20140025620 | Greenzeiger et al. | Jan 2014 | A1 |
20140067466 | Xiao et al. | Mar 2014 | A1 |
20140162225 | Hill | Jun 2014 | A1 |
20140214335 | Siefert | Jul 2014 | A1 |
20140221866 | Quy | Aug 2014 | A1 |
20140244345 | Sollis et al. | Aug 2014 | A1 |
20140278914 | Gurumoorthy et al. | Sep 2014 | A1 |
20140350431 | Hagedorn | Nov 2014 | A1 |
20150186923 | Gurumoorthy et al. | Jul 2015 | A1 |
Number | Date | Country |
---|---|---|
1087618 | Dec 2003 | EP |
1609418 | Dec 2005 | EP |
1374658 | Nov 1974 | GB |
2221759 | Feb 1990 | GB |
2001-147944 | May 2001 | JP |
2005-51654 | Feb 2005 | JP |
2005-160805 | Jun 2005 | JP |
2006-227994 | Aug 2006 | JP |
2006-305334 | Nov 2006 | JP |
2006-6355 | Jul 2007 | JP |
2004-22399 | Jul 2006 | KR |
95-18565 | Jul 1995 | WO |
97-17774 | May 1997 | WO |
97-40745 | Nov 1997 | WO |
97-41673 | Nov 1997 | WO |
02-100241 | Dec 2002 | WO |
02-102238 | Dec 2002 | WO |
2004-049225 | Jun 2004 | WO |
2008-064431 | Jun 2008 | WO |
2008-055078 | Jul 2008 | WO |
2008-077178 | Jul 2008 | WO |
2008-109694 | Sep 2008 | WO |
2008-109699 | Sep 2008 | WO |
2008-121651 | Oct 2008 | WO |
2008-137579 | Nov 2008 | WO |
2008-137581 | Nov 2008 | WO |
2008-141340 | Nov 2008 | WO |
2008-154410 | Dec 2008 | WO |
2009-018374 | Feb 2009 | WO |
2009-052833 | Apr 2009 | WO |
2011-055291 | May 2011 | WO |
2011-056679 | May 2011 | WO |
Entry |
---|
U.S. Appl. No. 12/749,376, filed Mar. 29, 2010. |
U.S. Appl. No. 12/426,259, filed Apr. 19, 2009. |
U.S. Appl. No. 13/089,752, filed Apr. 19, 2011. |
“ARF, AAAA and ANA Are Getting Emotional about Engagement”, Presentation, pp. 1-103 (2005). |
Boltz, M.G., “The cognitive processing of film and musical soundtracks”, Memory & Cognition, 32(7):1194-1205 (2004). |
Christie et al., “Autonomic specificity of discrete emotion and dimensions of affective space: a multivariate approach”, Int'l J. Psychophysiol., 51:143-153 (2004). |
Coombes et al., “Emotion and movement: Activation of defensive circuitry alters the magnitude of a sustained muscle contraction”, Neurosci. Lett., 396:192-196 (2006). |
Cryer et al., “Pull the plug on stress”, Harv. Bus. Rev., 81(7):102-107 (2003). |
Demaree et al., “Predicting facial valence to negative stimuli from resting RSA: Not a function of active emotion regulation”, Cognition and Emotion, 20(2):161-176 (2006). |
Ekman et al., “Autonomic Nervous System Activity Distinguishes among Emotions”, Science, 221(4616):1208-1210 (1983). |
Elton, C., “Measuring emotion at the symphony”, http://www.boston.com, pp. 1-3 (2006). |
Goldberg, C., “Getting wired could help predict emotions”, http://www.boston.com, pp. 1-4 (2005). |
Gomez et al., “Respiratory responses associated with affective processing of film stimuli”, Biol. Psychol., 68:223-235 (2005). |
Hall, B.F., “A New Approach to Measuring Advertising Effectiveness”, Article 1502a:1-17 (2001). |
Hall, B.F., “Advertising as a Factor of production”, Admap, pp. 30-32 (2003). |
Hall, B.F., “Is cognitive processing the right dimension”, Admap, pp. 37-39 (2003). |
Hall, B.F., “On Measuring the Power Communications”, JAR, pp. 1-11 (2004). |
Hall, B.F., “Research and strategy: a fall from grace”, Admap, pp. 2-4 (2003). |
Hall, B.F., “Review of Casting for Big Ideas, by Andrew Jaffe”, pp. 1-2 (2003). |
Hall, B.F., “Why Advertisers Do It”, pp. 1-5 (2003). |
Hubert, et al., “Autonomic, neuroendocrine, and subjective responses to emotion-inducing film stimuli”, Int'l J. Psychophysiol., 11:131-140 (1991). |
Levenson et al., “Emotion and Autonomic Nervous System Activity in the Minangkabau of West Sumatra”, J. Personality Soc. Psychol., 62(6):972-988 (1992). |
Marci et al., “The Effect of Emotional Distance on Pyschophysiologic Concordance and Perceived Empathy Between Patient and Interviewer”, Appl. Psychophysiol. Biofeedback, 31:115-129 (2006). |
McCraty et al., “Analysis of twenty-four hour heart rate variability in patients with panic disorder”, Biol. Psychol., 56(2):131-150 (2001). |
McCraty et al., “Electrophysiolocial Evidence of Intuition: Part 1. The Surprising Role of the Heart”, J. Altern. Complement. Med., 10(1):133-143 (2004). |
McCraty et al., “Electrophysiological Evidence of Intuition: Part 2. A System-Wide Process?”, J. Altern. Complement. Med., 10(2\0):325-336 (2004). |
McCraty et al., “Impact of a Workplace Stress Reduction Program on Blood Pressure and Emotional Health in Hypertensive Employees”, J. Altern. Complement. Med., 9(3):355-369 (2003). |
McCraty et al., “The Effects of Different Types of Music on Mood, Tension, and Mental Clarity”, Altern. Ther. Health Med., 4(1):75-84 (1998). |
McCraty et al., “The Effects of Emotions on Ahort-Term Power Spectrum Analysis of Heart RateVariability”, Am. J. Cardiol., 76(14):1089-1093 (1995). |
McCraty et al., “The Impact of a New Emotional Self-Management Program on Stress, Emotions, Heart Rate Variability, DHEA and Cortisol”, Intergr. Physiol. Behay. Sci., 33(2):151-170 (1998). |
McCraty et al., “The Impact of an Emotional Self-Management Skills Course on Psychosocial Functioning and Autonomic Recovery to Stress in Middle School Children”, Integr. Physiol. Behav. Sci., 34(4):246-268 (1999). |
Melillo, W., “Inside the consumer mind: What Neuroscience can tell us about marketing”, http://www.answerstream.com, pp. 1-13 (2006). |
Miller et al., “Influence of Specific Emotional States on Autonomic Reactivity and Pulmonary Function in Asthmatic Children”, J. Am. Acad. Child Adolescent Psychiatry, 36(5):669-677 (1997). |
Murphy et al., “The Heart Reinnervates After Transplantation”, Ann. Thorac. Surg., 69(6):1769-1781 (2000). |
Ranii, D., “Adding Science to Gut Check”, The News & Observer, pp. 1 (2005). |
Rosenberg, K., “Emotional R.O.I.”, The Hub, pp. 24-25 (2006). |
Tiller et al., “Cardiac Coherence: A New, Noninvasive Measure of Autonomic Nervous System Order”, Altern. Ther. Health Med., 2(1):52-65 (1996). |
“Topline: Emotional Response to Advertising”, MSW Research, pp. 1-6 (2005). |
Umetani et al., “Twenty-Four Hour Time Domain Heart Rate Variability and Heart Rate: Relations to Age and Gender Over Nine Decades”, J. Am. Coll. Cardiol., 31(3):593-601 (1998). |
Von Leupoldt et al., “Emotions in a Body Plethysmograph”, J. Psychophysiol., 18(4):170-176 (2004). |
Kallman, H. Effect of Blank Time on Picture Recognition. The American Journal of Psychology, vol. 97, No. 3, Autumn, 1984, pp. 399-406 [retrieved on Nov. 3, 2011]. Retrieved from the Internet: <URL: http://www.jstor.org/pss/1422527>. |
Larose, Daniel T., Data Mining Methods and Models, John Wiley & Sons, Inc., 2006. |
Han, Micheline Kamber Jiawei, Data Mining: Concepts and Techniques, Second Edition (The Morgan Kaufmann Series in Data Management Systems), Elsevier, Inc., 2006. |
Liu, Bing, Web Data Mining: Exploring Hyperlinks, Contents, and Usage Data (Data-Centric Systems and Applications), Springer-Verlag, 2007. |
Berry, Michael J.A. and Linoff, Gordon S., Data Mining Techniques: For Marketing, Sales, and Customer Relationship Management, John Wiley & Sons, Inc., 1997. |
Watching Ads Is Real Science Research Companies Monitor Physiological Reactions to Commercials to Determine Their Effectiveness.: [3 Star Edition] Bruce Horovitz Los Angeles Times. Orlando Sentinel [Orlando, Fla] Sep. 1, 1991; D1, 2pgs. |
Wearable feedback systems for rehabilitation Sung, Michael; Marci, Carl; Pentland, Alex. Journal of NeuroEngineering and Rehabilitation 2 (2005), 2pgs. |
International Search Report dated Mar. 24, 2008 of International Patent Application No. PCT/US2007/019398. |
International Search Report dated Nov. 9, 2010 of International Patent Application No. PCT/US2010/031375. |
International Search Report dated Oct. 21, 2010 of International Patent Application No. PCT/US2010/029162. |
International Search Report dated Nov. 22, 2011 of International Patent Application No. PCT/US2011/033050. |
Non-Final Office Action dated Oct. 18, 2010 of related U.S. Appl. No. 11/850,650. |
Final Office Action dated Jun. 8, 2011 of related U.S. Appl. No. 11/850,650. |
Non-Final Office Action dated Dec. 13, 2011 of related U.S. Appl. No. 11/850,650. |
Notice of Allowance dated Jul. 16, 2012 of related U.S. Appl. No. 11/850,650. |
Non-Final Office Action dated Dec. 21, 2011 of related U.S. Appl. No. 12/749,376. |
Final Office Action dated Oct. 4, 2012 of related U.S. Appl. No. 12/749,376. |
Non-Final Office Action dated Apr. 1, 2013 of related U.S. Appl. No. 12/749,376. |
Non-Final Office Action dated Apr. 25, 2012 of related U.S. Appl. No. 12/426,259. |
Final Office Action dated Dec. 19, 2012 of related U.S. Appl. No. 12/426,259. |
Non-Final Office Action dated Feb. 21, 2013 of related U.S. Appl. No. 13/657,432. |
Non-Final Office Action dated Feb. 13, 2013 of related U.S. Appl. No. 13/089,752. |
Australian Office Action dated Mar. 26, 2012 or corresponding Australian Patent Application No. 2007293092. |
European Office Action dated Nov. 11, 2011 of corresponding European Patent Application No. 10717912.9. |
European Supplementary Search Report dated Mar. 1, 2013 of corresponding European Patent Application No. 10717912.9. |
European Office Action dated Mar. 14, 2013 of corresponding European Patent Application No. 10717912.9. |
European Office Action dated Nov. 29, 2011 of corresponding European Patent Application No. 10717932.7. |
European Search Report dated Aug. 8, 2013 of corresponding European Patent Application No. 10717932.7. |
European Supplementary Search Report dated Aug. 27, 2013 of corresponding European Patent Application No. 10717932.7. |
Japanese Office Action dated Apr. 23, 2012 of corresponding Japanese Patent Application No. 2009-527401. |
Canadian Office Action dated Jul. 24, 2014 of related Canadian Patent Application No. 2,662,632. |
Brown, M. “Should My Advertising Stimulate an Emotional Response?” 2009, available at http://www.wpp.com/˜/media/sharedwpp/readingroom/marketing/millward—brown—emotional—response.pdf, 6 pages. |
Mehta, A. et al., “Reconsidering Recall and Emotion in Advertising,” Journal of Advertising Research, Mar. 2006, pp. 49-56, 8 pages. |
Micu, A. C. et al., “Measurable Emotions: How Television Ads Really Work: How the Patterns of Reactions to Commercials can Demonstrate Advertising Effectiveness”, Management Slant, 50(2), Jun. 2010; pp. 1-17, 18 pages. |
Cheung, Kwok-Wai, et al., “Mining Customer Product Ratings for Personalized Marketing,” Decision Support Systems 35, 2003, pp. 231-243, 13 pages. |
Jaimes, A., Sebe, N., Multimodal Human-Computer Interaction: A Survey, Computer Vision and Image Understanding 108, Oct.-Nov. 2007, pp. 116-134, 19 pages. |
Garton, L. et al., “Studying Online Social Networks”, Journal of Computer-Mediated Communication , 3(1), Jun. 1997, 29 pages. |
Darrow, Chester, “Psychological and psychophysiological significance of the electroencephalogram,” Psychological Review (May 1947) 157-168, 12 pages. |
Stamm, John, “On the Relationship between Reaction Time to Light and Latency of Blocking the Alpha Rhythm,” Electroencephalography and Clinical Neurophysiology (Feb. 1952), 61-68, 8 pages. |
Mizuki, Yashushi, et al., “Periodic Appearance of the Theta Rhythm in the Frontal Midline Area During Performance of a Mental Task,:” Electroencephalography and Clinical Neurophysiology (Aug. 1980), 345-351, 7 pages. |
Ekman, P., Friesen, W., Measuring Facial Movement, Environmental Psychology and Nonverbal Behavior, 1(1) (Fall 1976), pp. 56-75, 20 pages. |
Ekman, P., Friesen, W.V., Facial Action Coding System: A Technique for Measurement of Facial Movement, Consulting Psychologists Press, Palo Alto, Calif. (1978). (Book). |
Ekman, P., Friesen, W., Unmasking the Face—A Guide to Recognizing Emotions from Facial Clues, Prentice-Hall, Inc., Englewood Cliffs, N.J. (1979). (Book). |
Ekman, P., Friesen, W., Ancoli, S., Facial Signs of Emotional Experience, J. Personality & Social Psychology, 39(6) (Dec. 1980), pp. 1125-1134, 10 pages. |
Izard, C. E., The Maximally Discriminative Facial Movement Coding System, (Rev. ed.), Instructional Resources Center, University of Delaware, Newark, Del. (1983). (Book). |
Izard, C., Dougherty, L., Hembree, E., A System for Identifying Affect Expressions by Holistic Judgments (AFFEX), Instructional Resources Center, University of Delaware, Newark, Del. (1983). (Book). |
Jia, X., Nixon, M.S., Extending the Feature Set for Automatic Face Recognition, International Conference on Image Processing and Its Applications (Apr. 7-9, 1992), 6 pages. |
Lisetti, C., Nasoz, F., Using Noninvasive Wearable Computers to Recognize Human Emotions from Physiological Signals, EURASIP J. Applied Signal Processing, 11 (Sep. 2004), pp. 1672-1687, 16 pages. |
McClure, Samuel, et al., “Neural Correlates of Behavioral Preference for Culturally Familiar Drinks,” Neuron (Oct. 14, 2004), 379-387, 9 pages. |
Opitz, S. “Neuromarketing: An Introduction” PowerPoint Presentation (2008), available at http://www.powershow.com/view/94a7b-YzlmN/Neuromarketing—powerpoint—ppt—presentation (last accessed Oct. 14, 2015), 20 pages. |
Axis Communications, “Improve your merchandising effectiveness. Get the full picture with network video” (2008), available at :http://www.axis.com/files/user—scenarios/ap—ret—merchandising—311 07—en—0803—lo.pdf, 2 pages. |
Kamba, Tomonari, “The Krakatoa Chronicl—An Interactive, Personalized Newspaper on the Web,” available at: http://www.w3.org/Conferences/WWW4/Papers/93/ (last accessed Nov. 2, 2015), 15 pages. |
Ehrenberg et al. , “Understanding Brand Performance Measures: Using Dirichlet Benchmarks,” 2004, Journal of Business Research, vol. 57, pp. 1307-1325, 19 pages. |
Leeflang et al. , “Building Models for Marketing Decisions,” 2000, Springer Science + Business Media, pp. 192-235, 482-521, 86 pages. |
Bassi et al., “The Dirichlet Model: Analysis of a Market and Comparison of Estimation Procedures,” 2011, Marketing Bulletin, vol. 22, Technical Note 1, pp. 1-11, 11 pages. |
Bhattacharya, “Is your brand's loyalty too much, too little, or just right?: Explaining deviations in loyalty from the Dirichlet norm,” 1997, International Journal of Research in Marketing, vol. 14, pp. 421-435, 15 pages. |
Cohen, “Differentiated product demand analysis with a structured covariance probit: A Bayesian econometric approach,” 2009, PhD dissertation, University of Connecticut, pp. 1-184, 197 pages. |
Nikolaeva et al., “The Moderating Role of Consumer and Product Characteristics on the Value of Customized On-Line Recommendations,” 2006, International Journal of Electronic Commerce, vol. 11, No. 2, pp. 101-123, 24 pages. |
Ehrenberg, “New Brands and the Existing Market,” 1991, International Journal of Market Research, vol. 33, No. 4, 10 pages. |
Foxall, “The Substitutability of Brands,” 1999, Managerial and Decision Economics, vol. 20, pp. 241-257, 17 pages. |
Pammer, “Forecasting the Penetration of a New Product—A Bayesian Approach,” 2000, Journal of Business and Economic Statistics, vol. 18, No. 4, pp. 428-435, 8 pages. |
Rungie et al., “Calculation of Theoretical Brand Performance Measures from the Parameters of the Dirichlet Model,” 2004, Marketing Bulletin, Massey University, 15, Technical Note 2, pp. 1-19, 20 pages. |
Uncles et al., “Patterns of Buyer Behavior: Regularities, Models, and Extensions,” 1995, Marketing Science, vol. 14, No. 3, pp. G71-G78, 9 pages. |
Aaker et al., “Warmth in Advertising: Measurement, Impact, and Sequence Effects,” Journal of Consumer Research, vol. 12, No. 4, pp. 365-381, (Mar. 1986), 18 pages. |
Ambler, “Salience and Choice: Neural Correlates of Shopping Decisions,” Psychology & Marketing, vol. 21, No. 4, p. 247-261, Wiley Periodicals, Inc., doi: 10.1002/mar20004, (Apr. 2004), 16 pages. |
Bagozzi et al., “The Role of Emotions in Marketing,” Journal of the Academy of Marketing Science, vol. 27, No. 2, pp. 184-206, Academy of Marketing Science (1999), 23 pages. |
Belch et al., “Psychophysiological and Cognitive Response to Sex in Advertising,” Advances in Consumer Research, vol. 9, pp. 424-427, (1982), 6 pages. |
Blakeslee, “If You Have a ‘Buy Button’ in Your Brain, What Pushes It?” The New York Times, www.nytimes.com, (Oct. 19, 2004), 3 pages. |
Braeutigam, “Neuroeconomics-From neural systems to economic behaviour,” Brain Research Bulletin, vol. 67, pp. 355-360, Elsevier, (2005), 6 pages. |
Buschman, et al., “Top-Down versus Bottom-Up Control of Attention in the Prefrontal and Posterior Parietal Cortices,” Science, vol. 315, www.sciencemag.org/cgi/content/full/315/5820/1860, American Association for the Advancement of Science, (2007), 4 pages. |
Canolty, et al., “High Gamma Power Is Phase-Locked to Theta Oscillations in Human Neocortex,” Science, vol. 313, www.sciencemag.org, (Sep. 15, 2006), 3 pages. |
Canolty, et al., “The functional role of cross-frequency coupling,” Trends in Cognitive Sciences, Elsevier, (Nov. 2010) 11 pages. |
Crawford et al., “Self-generated happy and sad emotions in low and highly hypnotizable persons during waking and hypnosis: laterality and regional EEG activity differences,” International Journal of Psychophysiology, vol. 24, pp. 239-266, (Dec. 1996), 28 pages. |
D'Esposito, “From cognitive to neural models of working memory,” Philosophical Transitions of the Royal Society B, doi: 10.1098/rstb.2007.2086, (Mar. 30, 2007), 12 pages. |
Davidson, et al., “The functional neuroanatomy of emotion and affective style,” TRENDS in Cognitive Sciences, vol. 3, No. 1, (Jan. 1999), 11 pages. |
Desmet, “Measuring Emotions: Development and Application of an Instrument to Measure Emotional Responses to Products,” to be published in Funology: From Usability to Enjoyment, pp. 111-123, Kluwer Academic Publishers, (Blythe et al., eds., 2004), 13 pages. |
EEG Protocols, “Protocols for EEG Recording,” retrieved from the Internet on Aug. 23, 2011, http://www.q-metrx.com/EEGrecordingProtocols.pdf, (Nov. 13, 2007), 3 pages. |
Fogelson, et al., “Prefrontal cortex is critical for contextual processing: evidence from brain lesions,” Brain: A Journal of Neurology, vol. 132, pp. 3002-3010, doi:10.1093/brain/awp230, (Aug. 27, 2009), 9 pages. |
Fries, Pascal, “A mechanism for cognitive dynamics: neuronal communication through neuronal coherence,” Trends in Cognitive Sciences, vol. 9, No. 10, pp. 474-480, Elsevier B.V. www.sciencedirect.com, (Oct. 2005), 7 pages. |
Fuster, Joaquin M., “Cortex and Memory: Emergence of a New Paradigm,” Journal of Cognitive Neuroscience, vol. 21, No. 11, pp. 2047-2072, Massachusetts Institute of Technology, (Nov. 2009), 26 pages. |
Gazzaley et al., “Top-down Enhancement and Suppression of Magnitude and Speed of Neural Activity,” Journal of Cognitive Neuroscience, vol. 17, No. 3, pp. 507-517, Massachusetts Institute of Technology, (2005), 11 pages. |
Haq, Amber, “This Is Your Brain on Advertising,” Business Week, Market Research, (Oct. 8, 2007), 4 pages. |
Hartikainen et al., Manuscript Draft of “Emotionally arousing stimuli compete with attention to left hemispace,” NeuroReport, (Sep. 8, 2007), 26 pages. |
Hazlett et al., “Emotional Response to Television Commercials: Facial EMG vs. Self-Report,” Journal of Advertising Research, (Apr. 1999), 17 pages. |
Keren et al., “Saccadic spike potentials in gamma-band EEG: Characterization, detection and suppression,” NeuroImage, http://dx.doi:10.1016/j.neuroimage.2009.10.057, (Oct. 2009), 16 pages. |
Kishiyama et al., “Novelty Enhancements in Memory Are Dependent on Lateral Prefrontal Cortex,” The Journal of Neuroscience, pp. 8114-8118, Society for Neuroscience (Jun. 24, 2009), 5 pages. |
Knight, Robert T., “Contribution of human hippocampal region to novelty detection,” Nature, vol. 383, www.nature.com, (Sep. 19, 1996), 4 pages. |
Knight, Robert T., “Decreased Response to Novel Stimuli after Prefrontal Lesions in Man,” Electroencephalography and Clinical Neurophysiology, vol. 59, pp. 9-20, Elsevier Scientific Publishers Ireland, Ltd., (1984), 12 pages. |
Lee et al., “What is ‘neuromarketing’? A discussion and agenda for future research,” International Journal of Psychophysiology, vol. 63, pp. 199-204, Elsevier (2006), 6 pages. |
Lekakos, George, “Personalized Advertising Services Through Hybrid Recommendation Methods: The Case of Digital Interactive Television,” Department of Informatics, Cyprus University, (2004), 11 pages. |
Lewis et al., “Market Researchers make Increasing use of Brain Imaging,” ACNR, vol. 5, No. 3, pp. 36-37, (Jul./Aug. 2005), 2 pages. |
Lui et al., “Marketing Strategies in Virtual Worlds,” The Data Base for Advances in Information Systems, vol. 38, No. 4, pp. 77-80, (Nov. 2007), 4 pages. |
Makeig et al., “Dynamic Brain Sources of Visual Evoked Responses,” Science, vol. 295, www.sciencemag.org, (Jan. 25, 2002), 5 pages. |
Makeig et al., “Mining event-related brain dynamics,” TRENDS in Cognitive Sciences, vol. 8, No. 5, (May 2004), www.sciencedirect.com, 7 pages. |
Miltner et al., “Coherence of gamma-band EEG activity as a basis for associative learning,” Nature, vol. 397, www.nature.com, (Feb. 4, 1999), 3 pages. |
Moran et al. “Peak frequency in the theta and alpha bands correlates with human working memory capacity,” frontiers in Human Neuroscience, vol. 4, Article 200, www.frontiersin.org, (Nov. 11, 2010), 12 pages. |
Neurofocus—Neuroscientific Analysis for Audience Engagement, accessed on Jan. 8, 2010 at http://web.archive.org/web/20080621114525/www.neurofocus.com /BrandImage.htm, (2008), 2 pages. |
Nielsen, “Neuroinformatics in Functional Neuroimaging,” Informatics and Mathematical Modeling, Technical University of Denmark, (Aug. 30, 2002), 241 pages. |
Page et al., “Cognitive Neuroscience, Marketing and Research,” Congress 2006—Foresight—The Predictive Power of Research Conference Papers, ESOMAR Publications, (Sep. 17, 2006), 25 pages. |
Simon-Thomas et al, “Behavioral and Electrophysiological Evidence of a Right Hemisphere Bias for the Influence of Negative Emotion on Higher Cognition,” Journal of Cognitive Neuroscience, pp. 518-529, Massachusetts Institute of Technology (2005), 12 pages. |
Sutherland, Max, “Neuromarketing: What's it all about?” Retrieved from Max Sutherland's Weblog on Aug. 23, 2011, http://www.sutherlandsurvey.com/Column—pages/Neuromarketing—whats—it—all—about.htm, (Mar. 2007), 5 pages. |
Swick et al., “Contributions of Prefrontal Cortex to Recognition Memory: Electrophysiological and Behavioral Evidence,” Neuropsychology, vol. 13, No. 2, pp. 155-170, American Psychological Association, Inc. (1999), 16 pages. |
Voytek et al., “Shifts in gamma phase-amplitude coupling frequency from theta to alpha over posterior cortex during visual tasks,” Frontiers in Human Neuroscience, doi: 10.3389/fnhum.2010.00191, (Oct. 19, 2010), 9 pages. |
Wang, Xiao-Jing, “Neurophysiological and Computational Principles of Cortical Rhythms in Cognition,” Physiol Rev 90: pp. 1195-1268, American Physiological Society, www.prv.org, (2010), 75 pages. |
Woodman et al., “Serial Deployment of Attention During Visual Search,” Journal of Experimental Psychology: Human Perception and Performance, vol. 29, No. 1, pp. 121-138, American Physiological Association (2003), 18 pages. |
Yuval-Greenberg et al., “Transient Induced Gamma-Bands Response in EEG as a Manifestation of Miniature Saccades,” Neuron, vol. 58, pp. 429-441, Elsevier Inc. (May 8, 2008), 13 pages. |
Ziegenfuss, Jennifer S., “Neuromarketing: Advertising Ethical & Medical Technology,” The Brownstone Journal, vol. XII, Boston University, pp. 69-73, (May 2005), 9 pages. |
Zyga, Lisa, “A Baseball Cap That Can Read Your Mind,” PhysOrg.com, located at www.physorg.com/news130152277.html, (May 16, 2008), 11 pages. |
Ambler et al., “Ads on the Brain; A Neuro-Imaging Comparison of Cognitive and Affective Advertising Stimuli,” London Business School, Centre for Marketing Working Paper, No. 00-902, (Mar. 2000), 23 pages. |
Clifford, Stephanie, “Billboards That Look Back,” The New York Times, NYTimes.com, available at http://www.nytimes.com/2008/05/31/business/media/31billboard.html, (May 31, 2008), 4 pages. |
Bimler et al., “Categorical perception of facial expressions of emotion: Evidence from multidimensional scaling,” Cognition and Emotion, vol. 15(5), pp. 633-658 (Sep. 2001), 26 pages. |
De Gelder et al., “Categorical Perception of Facial Expressions: Categories and their Internal Structure,” Cognition and Emotion, vol. 11(1), pp. 1-23 (1997), 23 pages. |
Newell et al., “Categorical perception of familiar objects,” Cognition, vol. 85, Issue 2, pp. 113-143 (Sep. 2002), 31 pages. |
Padgett et al., “Categorical Perception in Facial Emotion Classification,” In Proceedings of the 18th Annual Conference of the Cognitive Science Society, pp. 249-253 (1996), 5 pages. |
Klimesch, Wolfgang, “EEG alpha and theta oscillations reflect cognitive and memory performance a review and analysis,” Brain Research Reviews, vol. 29, 169-195, (1999), 27 pages. |
Krakow et al., “Methodology: EEG-correlated fMRI,” Functional Imaging in the Epilepsies, (Lippincott Williams & Wilkins, 2000), 17 pages. |
Krugman, “Brain Wave Measures of Media Involvement,” Journal of Advertising Research vol. 11, No. 1, pp. 3-9 (Feb. 1971), 7 pages. |
Palva et al., “Phase Synchrony Among Neuronal Oscillations in the Human Cortex,” Journal of Neuroscience 25 (2005), 3962-3972, 11 pages. |
Lachaux et al., “Measuring Phase Synchrony in Brain Signals,” Human Brain Mapping 8 (1999), 194-208, 15 pages. |
Zhang, P., “Will You Use Animation on Your Web Pages?” Doing Business on the Internet: Opportunities and Pitfalls, C. Romm and F. Sudweeks (eds.), Spring-Verlag (1999), 17 pages. |
Coan et al., “Voluntary Facial Expression and Hemispheric Asymmetry Over the Frontal Cortex,” Psycophysiology (Nov. 2001), 912-924, 14 pages. |
Duchowski, “A Breadth-First Survey of Eye-tracking Applications,” Beahavior Research Methods, Instruments, and Computers (Nov. 2002), 455-470, 16 pages. |
Heo et al., “Wait! Why is it Not Moving? Attractive and Distractive Ocular Responses to Web Ads,” Paper presented to AEJMC, (Aug. 2001) Washington, DC, available at http://www.psu.edu/dept/medialab/researchpage/newabstracts/wait.html, 3 pages. |
Rothschild et al., “Predicting Memory for Components of TV Commercials from EEG,” Journal of Consumer Research (Mar. 1990), p. 472-478, 8 pages. |
Beaver, John D., et al., “Individual Differences in Reward Drive Predict Neural Responses to Images of Food”, J. of Neuroscience, (May 10, 2006), 5160-5166, 7 pages. |
Shandlen et al., “A Computational Analysis of the Relationship between Neuronal and Behavioral Responses to Visual Motion”, The Journal of Neuroscience, (Feb. 15, 1996) 1486-1510, 25 pages. |
Cassanello et al., “Neuronal Responses to Moving Targets in Monkey Frontal Eye Fields”, J Neurophysiol (Sep. 2008), 1544-1556, 16 pages. |
Bandari et al., “The Pulse of News in Social Media: Forecasting Popularity,” ICWSM, Feb. 2012, 9 pages. |
Rodriguez et al., “Social Networks for News Media Distribution,” STB-RL: Digital Library Research and Prototyping, Los Alamos National Laboratory, 2006, 6 pages. |
Number | Date | Country | |
---|---|---|---|
20130280682 A1 | Oct 2013 | US |
Number | Date | Country | |
---|---|---|---|
61603528 | Feb 2012 | US |