Content based selection and meta tagging of advertisement breaks

Information

  • Patent Grant
  • 11610223
  • Patent Number
    11,610,223
  • Date Filed
    Friday, May 28, 2021
    2 years ago
  • Date Issued
    Tuesday, March 21, 2023
    a year ago
Abstract
An example system to identify an advertisement to include in source material to increase an effectiveness of the advertisement includes an analyzer to determine one or more priming characteristics for a plurality of locations of a source material based on neuro-response data collected from a first subject exposed to the source material and a selector to identify an attribute of the advertisement, identify at least one of a temporal attribute or a spatial attribute for the plurality of locations, perform a comparison of the attribute of the advertisement to the at least one of the temporal attribute or the spatial attribute for the plurality of locations, select a first location of the plurality of locations for insertion of the advertisement based on the comparison and the priming characteristics, and transform the source material to include the advertisement at the first location.
Description
FIELD OF THE DISCLOSURE

The present disclosure relates to content based selection and meta-tagging of advertisement breaks.


BACKGROUND

Conventional systems for content selection and meta-tagging of advertisement breaks are limited or non-existent. Some conventional systems provide very rudimentary information for content selection through demographic information and statistical data. However, conventional systems are subject to semantic, syntactic, metaphorical, cultural, and interpretive errors.





BRIEF DESCRIPTION OF THE DRAWINGS

The disclosure may best be understood by reference to the following description taken in conjunction with the accompanying drawings, which illustrate particular examples.



FIG. 1 illustrates one example of a system for content selection and meta-tagging of advertisement breaks.



FIG. 2 illustrates examples of stimulus attributes that can be included in a stimulus attributes repository.



FIG. 3 illustrates examples of data models that can be used with a stimulus and response repository.



FIG. 4 illustrates one example of a query that can be used with the content selection and meta-tagging system.



FIG. 5 illustrates one example of a report generated using the content selection and meta-tagging system.



FIG. 6 illustrates one example of a technique for performing data analysis.



FIG. 7 illustrates one example of technique for content selection and meta-tagging of advertisement breaks.



FIG. 8 provides one example of a system that can be used to implement one or more mechanisms.





DETAILED DESCRIPTION

Reference will now be made in detail to some specific examples of the disclosure including the best modes contemplated by the inventors for carrying out the teachings of the disclosure. Examples of these specific examples are illustrated in the accompanying drawings. While the disclosure is described in conjunction with these specific examples, it will be understood that it is not intended to limit the disclosure to the described examples. On the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the disclosure as defined by the appended claims.


For example, the techniques and mechanisms of the present disclosure will be described in the context of particular types of data such as central nervous system, autonomic nervous system, and effector data. However, it should be noted that the techniques and mechanisms of the present disclosure apply to a variety of different types of data. It should be noted that various mechanisms and techniques can be applied to any type of stimuli. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present disclosure. Particular examples of the present disclosure may be implemented without some or all of these specific details. In other instances, well known process operations have not been described in detail in order not to unnecessarily obscure the present disclosure.


Various techniques and mechanisms of the present disclosure will sometimes be described in singular form for clarity. However, it should be noted that some examples include multiple iterations of a technique or multiple instantiations of a mechanism unless noted otherwise. For example, a system uses a processor in a variety of contexts. However, it will be appreciated that a system can use multiple processors while remaining within the scope of the present disclosure unless otherwise noted. Furthermore, the techniques and mechanisms of the present disclosure will sometimes describe a connection between two entities. It should be noted that a connection between two entities does not necessarily mean a direct, unimpeded connection, as a variety of other entities may reside between the two entities. For example, a processor may be connected to memory, but it will be appreciated that a variety of bridges and controllers may reside between the processor and memory. Consequently, a connection does not necessarily mean a direct, unimpeded connection unless otherwise noted.


Overview


Consequently, it is desirable to provide improved methods and apparatus for content selection and meta-tagging of advertisement (ad) breaks.


A system evaluates stimulus materials such as videos, imagery, web pages, text, etc., in order to determine resonance and priming levels for various products and services at different temporal and spatial locations including advertisement breaks in the stimulus materials. The stimulus materials are tagged with resonance and priming level information to allow intelligent selection of suitable advertisement content for insertion at various locations in the stimulus materials. Response data such as survey data and/or neuro-response data including Event Related Potential (ERP), Electroencephalography (EEG), Galvanic Skin Response (GSR), Electrocardiograms (EKG), Electrooculography (EOG), eye tracking, and facial emotion encoding data may be used to determine resonance and priming levels.


EXAMPLES

Conventional mechanisms for selecting advertising content rely on demographic information, statistical information, and survey based response collection. One problem with conventional mechanisms for selecting advertising is that they do not measure the inherent message resonance and priming for various products, services, and offerings that are attributable to the stimulus. They are also prone to semantic, syntactic, metaphorical, cultural, and interpretive errors thereby preventing the accurate and repeatable targeting of the audience.


Conventional systems do not use neuro-behavioral and neuro-physiological response blended manifestations in assessing the user response and do not elicit an individual customized neuro-physiological and/or neuro-behavioral response to the stimulus. Conventional systems also fail to blend multiple datasets, and blended manifestations of multi-modal responses, across multiple datasets, individuals and modalities, to reveal and validate the elicited measures of resonance and priming to allow for intelligent selection of advertising content.


In these respects, the content selection and meta-tagging of advertising break system according to the present disclosure substantially departs from the conventional concepts and designs of the prior art. According to various examples, it is recognized that advertisements for particular products, services, and offerings may be particularly effective when a subject is primed for the particular products, services, and offerings. For examples, an advertisement for cleaning supplies may be particularly effective after viewers watch a scene showing a dirty room, or an advertisement for a fuel efficient car may be particularly effective after viewers watch a documentary about high oil prices. In still other examples, an audio advertisement for packaged salads may be more effective after viewers hear a radio program about coronary disease, or a brand image for camping products may be more effective placed near a mural showing mountain scenery.


Consequently, the techniques and mechanisms of the present disclosure tag stimulus materials such as video, audio, web pages, printed materials, etc. with information indicating resonance and/or priming levels for various products, services and offerings. Meta-tags may be stored in a separate repository or in the stimulus material itself. In some examples, advertising content suitable for particular priming levels may be automatically selected based on meta-tags for introduction into the stimulus materials. In other examples, advertising content can be intelligently inserted based on priming levels for particular products and services. In some examples, advertising break slots can be sold or auctioned more efficiently based on priming levels and resonance. Advertisers can assess the value of particular slots based on priming levels and resonance.


According to various examples, the techniques and mechanisms of the present disclosure may use a variety of mechanisms such as survey based responses, statistical data, and/or neuro-response measurements such as central nervous system, autonomic nervous system, and effector measurements to improve content selection and meta-tagging of stimulus material. Some examples of central nervous system measurement mechanisms include Functional Magnetic Resonance Imaging (fMRI) and Electroencephalography (EEG). fMRI measures blood oxygenation in the brain that correlates with increased neural activity. However, current implementations of FMRI have poor temporal resolution of few seconds. EEG measures electrical activity associated with post synaptic currents occurring in the milliseconds range. Subcranial EEG can measure electrical activity with the most accuracy, as the bone and dermal layers weaken transmission of a wide range of frequencies. Nonetheless, surface EEG provides a wealth of electrophysiological information if analyzed properly. Even portable EEG with dry electrodes provides a large amount of neuro-response information.


Autonomic nervous system measurement mechanisms include Galvanic Skin Response (GSR), Electrocardiograms (EKG), pupillary dilation, etc. Effector measurement mechanisms include Electrooculography (EOG), eye tracking, facial emotion encoding, reaction time etc.


According to various examples, the techniques and mechanisms of the present disclosure intelligently blend multiple modes and manifestations of precognitive neural signatures with cognitive neural signatures and post cognitive neurophysiological manifestations to more accurately perform content selection and meta-tagging. In some examples, autonomic nervous system measures are themselves used to validate central nervous system measures. Effector and behavior responses are blended and combined with other measures. According to various examples, central nervous system, autonomic nervous system, and effector system measurements are aggregated into a measurement that allows content selection and meta-tagging of advertising breaks.


In particular examples, subjects are exposed to stimulus material and data such as central nervous system, autonomic nervous system, and effector data is collected during exposure. According to various examples, data is collected in order to determine a resonance measure that aggregates multiple component measures that assess resonance data. In particular examples, specific event related potential (ERP) analyses and/or event related power spectral perturbations (ERPSPs) are evaluated for different regions of the brain both before a subject is exposed to stimulus and each time after the subject is exposed to stimulus.


Pre-stimulus and post-stimulus differential as well as target and distracter differential measurements of ERP time domain components at multiple regions of the brain are determined (DERP). Event related time-frequency analysis of the differential response to assess the attention, emotion and memory retention (DERPSPs) across multiple frequency bands including but not limited to theta, alpha, beta, gamma and high gamma is performed. In particular examples, single trial and/or averaged DERP and/or DERPSPs can be used to enhance the resonance measure and determine priming levels for various products and services.


A variety of stimulus materials such as entertainment and marketing materials, media streams, billboards, print advertisements, text streams, music, performances, sensory experiences, etc. can be analyzed. According to various examples, enhanced neuro-response data is generated using a data analyzer that performs both intra-modality measurement enhancements and cross-modality measurement enhancements. According to various examples, brain activity is measured not just to determine the regions of activity, but to determine interactions and types of interactions between various regions. The techniques and mechanisms of the present disclosure recognize that interactions between neural regions support orchestrated and organized behavior. Attention, emotion, memory, and other abilities are not merely based on one part of the brain but instead rely on network interactions between brain regions.


The techniques and mechanisms of the present disclosure further recognize that different frequency bands used for multi-regional communication can be indicative of the effectiveness of stimuli. In particular examples, evaluations are calibrated to each subject and synchronized across subjects. In particular examples, templates are created for subjects to create a baseline for measuring pre and post stimulus differentials. According to various examples, stimulus generators are intelligent and adaptively modify specific parameters such as exposure length and duration for each subject being analyzed.


A variety of modalities can be used including EEG, GSR, EKG, pupillary dilation, EOG, eye tracking, facial emotion encoding, reaction time, etc. Individual modalities such as EEG are enhanced by intelligently recognizing neural region communication pathways. Cross modality analysis is enhanced using a synthesis and analytical blending of central nervous system, autonomic nervous system, and effector signatures. Synthesis and analysis by mechanisms such as time and phase shifting, correlating, and validating intra-modal determinations allow generation of a composite output characterizing the significance of various data responses to effectively perform content selection and meta-tagging.



FIG. 1 illustrates one example of a system for performing content selection and meta-tagging using central nervous system, autonomic nervous system, and/or effector measures. According to various examples, the content selection and meta-tagging system includes a stimulus presentation device 101. In particular examples, the stimulus presentation device 101 is merely a display, monitor, screen, etc., that displays stimulus material to a user. The stimulus material may be a media clip, a commercial, pages of text, a brand image, a performance, a magazine advertisement, a movie, an audio presentation, and may even involve particular tastes, smells, textures and/or sounds. The stimuli can involve a variety of senses and occur with or without human supervision. Continuous and discrete modes are supported. According to various examples, the stimulus presentation device 101 also has protocol generation capability to allow intelligent customization of stimuli provided to multiple subjects in different markets.


According to various examples, stimulus presentation device 101 could include devices such as televisions, cable consoles, computers and monitors, projection systems, display devices, speakers, tactile surfaces, etc., for presenting the stimuli including but not limited to advertising and entertainment from different networks, local networks, cable channels, syndicated sources, websites, internet content aggregators, portals, service providers, etc.


According to various examples, the subjects are connected to data collection devices 105. The data collection devices 105 may include a variety of neuro-response measurement mechanisms including neurological and neurophysiological measurements systems such as EEG, EOG, GSR, EKG, pupillary dilation, eye tracking, facial emotion encoding, and reaction time devices, etc. According to various examples, neuro-response data includes central nervous system, autonomic nervous system, and effector data. In particular examples, the data collection devices 105 include EEG 111, EOG 113, and GSR 115. In some instances, only a single data collection device is used. Data collection may proceed with or without human supervision.


The data collection device 105 collects neuro-response data from multiple sources. This includes a combination of devices such as central nervous system sources (EEG), autonomic nervous system sources (GSR, EKG, pupillary dilation), and effector sources (EOG, eye tracking, facial emotion encoding, reaction time). In particular examples, data collected is digitally sampled and stored for later analysis. In particular examples, the data collected could be analyzed in real-time. According to particular examples, the digital sampling rates are adaptively chosen based on the neurophysiological and neurological data being measured.


In one particular example, the content selection and meta-tagging system includes EEG 111 measurements made using scalp level electrodes, EOG 113 measurements made using shielded electrodes to track eye data, GSR 115 measurements performed using a differential measurement system, a facial muscular measurement through shielded electrodes placed at specific locations on the face, and a facial affect graphic and video analyzer adaptively derived for each individual.


In particular examples, the data collection devices are clock synchronized with a stimulus presentation device 101. In particular examples, the data collection devices 105 also include a condition evaluation subsystem that provides auto triggers, alerts and status monitoring and visualization components that continuously monitor the status of the subject, data being collected, and the data collection instruments. The condition evaluation subsystem may also present visual alerts and automatically trigger remedial actions. According to various examples, the data collection devices include mechanisms for not only monitoring subject neuro-response to stimulus materials, but also include mechanisms for identifying and monitoring the stimulus materials. For example, data collection devices 105 may be synchronized with a set-top box to monitor channel changes. In other examples, data collection devices 105 may be directionally synchronized to monitor when a subject is no longer paying attention to stimulus material. In still other examples, the data collection devices 105 may receive and store stimulus material generally being viewed by the subject, whether the stimulus is a program, a commercial, printed material, or a scene outside a window. The data collected allows analysis of neuro-response information and correlation of the information to actual stimulus material and not mere subject distractions.


According to various examples, the content selection and meta-tagging system also includes a data cleanser device 121. In particular examples, the data cleanser device 121 filters the collected data to remove noise, artifacts, and other irrelevant data using fixed and adaptive filtering, weighted averaging, advanced component extraction (like PCA, ICA), vector and component separation methods, etc. This device cleanses the data by removing both exogenous noise (where the source is outside the physiology of the subject, e.g. a phone ringing while a subject is viewing a video) and endogenous artifacts (where the source could be neurophysiological, e.g. muscle movements, eye blinks, etc.).


The artifact removal subsystem includes mechanisms to selectively isolate and review the response data and identify epochs with time domain and/or frequency domain attributes that correspond to artifacts such as line frequency, eye blinks, and muscle movements. The artifact removal subsystem then cleanses the artifacts by either omitting these epochs, or by replacing these epoch data with an estimate based on the other clean data (for example, an EEG nearest neighbor weighted averaging approach).


According to various examples, the data cleanser device 121 is implemented using hardware, firmware, and/or software. It should be noted that although a data cleanser device 121 is shown located after a data collection device 105 and before data analyzer 181, the data cleanser device 121 like other components may have a location and functionality that varies based on system implementation. For example, some systems may not use any automated data cleanser device whatsoever while in other systems, data cleanser devices may be integrated into individual data collection devices.


According to various examples, an optional data meta attributes repository 131 provides information on the stimulus material being presented. According to various examples, stimulus attributes include properties of the stimulus materials as well as purposes, presentation attributes, report generation attributes, etc. In particular examples, stimulus attributes include time span, channel, rating, media, type, etc. Stimulus attributes may also include positions of entities in various frames, object relationships, locations of objects and duration of display. Purpose attributes include aspiration and objects of the stimulus including excitement, memory retention, associations, etc. Presentation attributes include audio, video, imagery, and messages needed for enhancement or avoidance. Other attributes may or may not also be included in the stimulus attributes repository or some other repository.


The data cleanser device 121 and the stimulus attributes repository 131 pass data to the data analyzer 181. The data analyzer 181 uses a variety of mechanisms to analyze underlying data in the system to determine resonance. According to various examples, the data analyzer customizes and extracts the independent neurological and neuro-physiological parameters for each individual in each modality, and blends the estimates within a modality as well as across modalities to elicit an enhanced response to the presented stimulus material. In particular examples, the data analyzer 181 aggregates the response measures across subjects in a dataset.


According to various examples, neurological and neuro-physiological signatures are measured using time domain analyses and frequency domain analyses. Such analyses use parameters that are common across individuals as well as parameters that are unique to each individual. The analyses could also include statistical parameter extraction and fuzzy logic based attribute estimation from both the time and frequency components of the synthesized response.


In some examples, statistical parameters used in a blended effectiveness estimate include evaluations of skew, peaks, first and second moments, population distribution, as well as fuzzy estimates of attention, emotional engagement and memory retention responses.


According to various examples, the data analyzer 181 may include an intra-modality response synthesizer and a cross-modality response synthesizer. In particular examples, the intra-modality response synthesizer is configured to customize and extract the independent neurological and neurophysiological parameters for each individual in each modality and blend the estimates within a modality analytically to elicit an enhanced response to the presented stimuli. In particular examples, the intra-modality response synthesizer also aggregates data from different subjects in a dataset.


According to various examples, the cross-modality response synthesizer or fusion device blends different intra-modality responses, including raw signals and signals output. The combination of signals enhances the measures of effectiveness within a modality. The cross-modality response fusion device can also aggregate data from different subjects in a dataset.


According to various examples, the data analyzer 181 also includes a composite enhanced effectiveness estimator (CEEE) that combines the enhanced responses and estimates from each modality to provide a blended estimate of the effectiveness. In particular examples, blended estimates are provided for each exposure of a subject to stimulus materials. The blended estimates are evaluated over time to assess resonance characteristics. According to various examples, numerical values are assigned to each blended estimate. The numerical values may correspond to the intensity of neuro-response measurements, the significance of peaks, the change between peaks, etc. Higher numerical values may correspond to higher significance in neuro-response intensity. Lower numerical values may correspond to lower significance or even insignificant neuro-response activity. In other examples, multiple values are assigned to each blended estimate. In still other examples, blended estimates of neuro-response significance are graphically represented to show changes after repeated exposure.


According to various examples, the data analyzer 181 provides analyzed and enhanced response data to a data communication device. It should be noted that in particular instances, a data communication device is not necessary. According to various examples, the data communication device provides raw and/or analyzed data and insights. In particular examples, the data communication device may include mechanisms for the compression and encryption of data for secure storage and communication.


According to various examples, the data communication device transmits data using protocols such as the File Transfer Protocol (FTP), Hypertext Transfer Protocol (HTTP) along with a variety of conventional, bus, wired network, wireless network, satellite, and proprietary communication protocols. The data transmitted can include the data in its entirety, excerpts of data, converted data, and/or elicited response measures. According to various examples, the data communication device is a set top box, wireless device, computer system, etc. that transmits data obtained from a data collection device to a resonance estimator 185. In particular examples, the data communication device may transmit data even before data cleansing or data analysis. In other examples, the data communication device may transmit data after data cleansing and analysis.


In particular examples, the data communication device sends data to a resonance estimator 185. According to various examples, the resonance estimator 185 assesses and extracts resonance patterns. In particular examples, the resonance estimator 185 determines entity positions in various stimulus segments and matches position information with eye tracking paths while correlating saccades with neural assessments of attention, memory retention, and emotional engagement. In particular examples, the resonance estimator 185 also collects and integrates user behavioral and survey responses with the analyzed response data to more effectively estimate resonance.


According to various examples, the resonance estimator 185 provides data to a priming repository system 187. In particular examples, the priming repository system 187 associates meta-tags with various temporal and spatial locations in stimulus material, such as a television program, movie, video, audio program, print advertisement, etc. In some examples, every second of a show is associated with a set of meta-tags. In other examples, commercial or advertisement (ad) breaks are provided with a set of meta-tags that identify commercial or advertising content that would be most suitable for a particular break.


Pre-break content may identify categories of products and services that are primed at a particular point in a program. The content may also specify the level of priming associated with each category of product or service. For example, a movie may show old house and buildings. Meta-tags may be manually or automatically generated to indicate that commercials for home improvement products would be suitable for a particular advertisement break.


In some instances, meta-tags may include spatial and temporal information indicating where and when particular advertisements should be placed. For example, a documentary about wildlife that shows a blank wall in several scenes may include meta-tags that indicate a banner advertisement for nature oriented vacations may be suitable. The advertisements may be separate from a program or integrated into a program. According to various examples, the priming repository system 187 also identifies scenes eliciting significant audience resonance to particular products and services as well as the level and intensity of resonance.


A variety of data can be stored for later analysis, management, manipulation, and retrieval. In particular examples, the repository could be used for tracking stimulus attributes and presentation attributes, audience responses optionally could also be integrated into meta-tags.


As with a variety of the components in the system, the repository can be co-located with the rest of the system and the user, or could be implemented in a remote location. It could also be optionally separated into repository system that could be centralized or distributed at the provider or providers of the stimulus material. In other examples, the repository system itself is integrated into a library of stimulus materials such as a media library.



FIG. 2 illustrates examples of data models that may be provided with a stimulus attributes repository. According to various examples, a stimulus attributes data model 201 includes a channel 203, media type 205, time span 207, audience 209, and demographic information 211. A stimulus purpose data model 215 may include intents 217 and objectives 219. According to various examples, stimulus attributes data model 201 also includes spatial and temporal information 221 about entities and emerging relationships between entities.


According to various examples, another stimulus attributes data model 221 includes creation attributes 223, ownership attributes 225, broadcast attributes 227, and statistical, demographic and/or survey based identifiers 229 for automatically integrating the neuro-physiological and neuro-behavioral response with other attributes and meta-information associated with the stimulus.


According to various examples, a stimulus priming data model 231 includes fields for identifying advertisement breaks 233 and scenes 235 that can be associated with various priming levels 237 and audience resonance measurements 239. In particular examples, the data model 231 provides temporal and spatial information for ads, scenes, events, locations, etc. that may be associated with priming levels and audience resonance measurements. In some examples, priming levels for a variety of products, services, offerings, etc. are correlated with temporal and spatial information in source material such as a movie, billboard, advertisement, commercial, store shelf, etc. In some examples, the data model associates with each second of a show a set of meta-tags for pre-break content indicating categories of products and services that are primed. The level of priming associated with each category of product or service when the advertisement break is specified may also be provided. Audience resonance measurements and maximal audience resonance measurements for various scenes and advertisement breaks may be maintained and correlated with sets of products, services, offerings, etc.


The priming and resonance information may be used to select stimulus content suited for particular levels of priming and resonance. In some examples, the priming and resonance information may be used to more intelligently price advertising breaks based on value to advertisers.



FIG. 3 illustrates examples of data models that can be used for storage of information associated with tracking and measurement of resonance. According to various examples, a dataset data model 301 includes an experiment name 303 and/or identifier, client attributes 305, a subject pool 307, logistics information 309 such as the location, date, and time of testing, and stimulus material 311 including stimulus material attributes.


In particular examples, a subject attribute data model 315 includes a subject name 317 and/or identifier, contact information 321, and demographic attributes 319 that may be useful for review of neurological and neuro-physiological data. Some examples of pertinent demographic attributes include marriage status, employment status, occupation, household income, household size and composition, ethnicity, geographic location, sex, race. Other fields that may be included in data model 315 include shopping preferences, entertainment preferences, and financial preferences. Shopping preferences include favorite stores, shopping frequency, categories shopped, favorite brands. Entertainment preferences include network/cable/satellite access capabilities, favorite shows, favorite genres, and favorite actors. Financial preferences include favorite insurance companies, preferred investment practices, banking preferences, and favorite online financial instruments. A variety of subject attributes may be included in a subject attributes data model 315 and data models may be preset or custom generated to suit particular purposes.


According to various examples, data models for neuro-feedback association 325 identify experimental protocols 327, modalities included 329 such as EEG, EOG, GSR, surveys conducted, and experiment design parameters 333 such as segments and segment attributes. Other fields may include experiment presentation scripts, segment length, segment details like stimulus material used, inter-subject variations, intra-subject variations, instructions, presentation order, survey questions used, etc. Other data models may include a data collection data model 337. According to various examples, the data collection data model 337 includes recording attributes 339 such as station and location identifiers, the data and time of recording, and operator details. In particular examples, equipment attributes 341 include an amplifier identifier and a sensor identifier.


Modalities recorded 343 may include modality specific attributes like EEG cap layout, active channels, sampling frequency, and filters used. EOG specific attributes include the number and type of sensors used, location of sensors applied, etc. Eye tracking specific attributes include the type of tracker used, data recording frequency, data being recorded, recording format, etc. According to various examples, data storage attributes 345 include file storage conventions (format, naming convention, dating convention), storage location, archival attributes, expiry attributes, etc.


A preset query data model 349 includes a query name 351 and/or identifier, an accessed data collection 353 such as data segments involved (models, databases/cubes, tables, etc.), access security attributes 355 included who has what type of access, and refresh attributes 357 such as the expiry of the query, refresh frequency, etc. Other fields such as push-pull preferences can also be included to identify an auto push reporting driver or a user driven report retrieval system.



FIG. 4 illustrates examples of queries that can be performed to obtain data associated with content selection and meta-tagging. According to various examples, queries are defined from general or customized scripting languages and constructs, visual mechanisms, a library of preset queries, diagnostic querying including drill-down diagnostics, and eliciting what if scenarios. According to various examples, subject attributes queries 415 may be configured to obtain data from a neuro-informatics repository using a location 417 or geographic information, session information 421 such as testing times and dates, and demographic attributes 419. Demographics attributes include household income, household size and status, education level, age of kids, etc.


Other queries may retrieve stimulus material based on shopping preferences of subject participants, countenance, physiological assessment, completion status. For example, a user may query for data associated with product categories, products shopped, shops frequented, subject eye correction status, color blindness, subject state, signal strength of measured responses, alpha frequency band ringers, muscle movement assessments, segments completed, etc. Experimental design based queries may obtain data from a neuro-informatics repository based on experiment protocols 427, product category 429, surveys included 431, and stimulus provided 433. Other fields that may be used include the number of protocol repetitions used, combination of protocols used, and usage configuration of surveys.


Client and industry based queries may obtain data based on the types of industries included in testing, specific categories tested, client companies involved, and brands being tested. Response assessment based queries 437 may include attention scores 439, emotion scores, 441, retention scores 443, and effectiveness scores 445. Such queries may obtain materials that elicited particular scores.


Response measure profile based queries may use mean measure thresholds, variance measures, number of peaks detected, etc. Group response queries may include group statistics like mean, variance, kurtosis, p-value, etc., group size, and outlier assessment measures. Still other queries may involve testing attributes like test location, time period, test repetition count, test station, and test operator fields. A variety of types and combinations of types of queries can be used to efficiently extract data.



FIG. 5 illustrates examples of reports that can be generated. According to various examples, client assessment summary reports 501 include effectiveness measures 503, component assessment measures 505, and resonance measures 507. Effectiveness assessment measures include composite assessment measure(s), industry/category/client specific placement (percentile, ranking, etc.), actionable grouping assessment such as removing material, modifying segments, or fine tuning specific elements, etc, and the evolution of the effectiveness profile over time. In particular examples, component assessment reports include component assessment measures like attention, emotional engagement scores, percentile placement, ranking, etc. Component profile measures include time based evolution of the component measures and profile statistical assessments. According to various examples, reports include the number of times material is assessed, attributes of the multiple presentations used, evolution of the response assessment measures over the multiple presentations, and usage recommendations.


According to various examples, client cumulative reports 511 include media grouped reporting 513 of all stimulus assessed, campaign grouped reporting 515 of stimulus assessed, and time/location grouped reporting 517 of stimulus assessed. According to various examples, industry cumulative and syndicated reports 521 include aggregate assessment responses measures 523, top performer lists 525, bottom performer lists 527, outliers 529, and trend reporting 531. In particular examples, tracking and reporting includes specific products, categories, companies, brands.



FIG. 6 illustrates one example of content selection and meta-tagging. At 601, stimulus material is provided to multiple subjects. According to various examples, stimulus includes streaming video and audio. In particular examples, subjects view stimulus in their own homes in group or individual settings. In some examples, verbal and written responses are collected for use without neuro-response measurements. In other examples, verbal and written responses are correlated with neuro-response measurements. At 603, subject neuro-response measurements are collected using a variety of modalities, such as EEG, ERP, EOG, GSR, etc. At 605, data is passed through a data cleanser to remove noise and artifacts that may make data more difficult to interpret. According to various examples, the data cleanser removes EEG electrical activity associated with blinking and other endogenous/exogenous artifacts.


According to various examples, data analysis is performed. Data analysis may include intra-modality response synthesis and cross-modality response synthesis to enhance effectiveness measures. It should be noted that in some particular instances, one type of synthesis may be performed without performing other types of synthesis. For example, cross-modality response synthesis may be performed with or without intra-modality synthesis.


A variety of mechanisms can be used to perform data analysis. In particular examples, a stimulus attributes repository 131 is accessed to obtain attributes and characteristics of the stimulus materials, along with purposes, intents, objectives, etc. In particular examples, EEG response data is synthesized to provide an enhanced assessment of effectiveness. According to various examples, EEG measures electrical activity resulting from thousands of simultaneous neural processes associated with different portions of the brain. EEG data can be classified in various bands. According to various examples, brainwave frequencies include delta, theta, alpha, beta, and gamma frequency ranges. Delta waves are classified as those less than 4 Hz and are prominent during deep sleep. Theta waves have frequencies between 3.5 to 7.5 Hz and are associated with memories, attention, emotions, and sensations. Theta waves are typically prominent during states of internal focus.


Alpha frequencies reside between 7.5 and 13 Hz and typically peak around 10 Hz. Alpha waves are prominent during states of relaxation. Beta waves have a frequency range between 14 and 30 Hz. Beta waves are prominent during states of motor control, long range synchronization between brain areas, analytical problem solving, judgment, and decision making. Gamma waves occur between 30 and 60 Hz and are involved in binding of different populations of neurons together into a network for the purpose of carrying out a certain cognitive or motor function, as well as in attention and memory. Because the skull and dermal layers attenuate waves in this frequency range, brain waves above 75-80 Hz are difficult to detect and are often not used for stimuli response assessment.


However, the techniques and mechanisms of the present disclosure recognize that analyzing high gamma band (kappa-band: Above 60 Hz) measurements, in addition to theta, alpha, beta, and low gamma band measurements, enhances neurological attention, emotional engagement and retention component estimates. In particular examples, EEG measurements including difficult to detect high gamma or kappa band measurements are obtained, enhanced, and evaluated. Subject and task specific signature sub-bands in the theta, alpha, beta, gamma and kappa bands are identified to provide enhanced response estimates. According to various examples, high gamma waves (kappa-band) above 80 Hz (typically detectable with sub-cranial EEG and/or magnetoencephalograophy) can be used in inverse model-based enhancement of the frequency responses to the stimuli.


Various examples of the present disclosure recognize that particular sub-bands within each frequency range have particular prominence during certain activities. A subset of the frequencies in a particular band is referred to herein as a sub-band. For example, a sub-band may include the 40-45 Hz range within the gamma band. In particular examples, multiple sub-bands within the different bands are selected while remaining frequencies are band pass filtered. In particular examples, multiple sub-band responses may be enhanced, while the remaining frequency responses may be attenuated.


An information theory based band-weighting model is used for adaptive extraction of selective dataset specific, subject specific, task specific bands to enhance the effectiveness measure. Adaptive extraction may be performed using fuzzy scaling. Stimuli can be presented and enhanced measurements determined multiple times to determine the variation profiles across multiple presentations. Determining various profiles provides an enhanced assessment of the primary responses as well as the longevity (wear-out) of the marketing and entertainment stimuli. The synchronous response of multiple individuals to stimuli presented in concert is measured to determine an enhanced across subject synchrony measure of effectiveness. According to various examples, the synchronous response may be determined for multiple subjects residing in separate locations or for multiple subjects residing in the same location.


Although a variety of synthesis mechanisms are described, it should be recognized that any number of mechanisms can be applied—in sequence or in parallel with or without interaction between the mechanisms.


Although intra-modality synthesis mechanisms provide enhanced significance data, additional cross-modality synthesis mechanisms can also be applied. A variety of mechanisms such as EEG, Eye Tracking, GSR, EOG, and facial emotion encoding are connected to a cross-modality synthesis mechanism. Other mechanisms as well as variations and enhancements on existing mechanisms may also be included. According to various examples, data from a specific modality can be enhanced using data from one or more other modalities. In particular examples, EEG typically makes frequency measurements in different bands like alpha, beta and gamma to provide estimates of significance. However, the techniques of the present disclosure recognize that significance measures can be enhanced further using information from other modalities.


For example, facial emotion encoding measures can be used to enhance the valence of the EEG emotional engagement measure. EOG and eye tracking saccadic measures of object entities can be used to enhance the EEG estimates of significance including but not limited to attention, emotional engagement, and memory retention. According to various examples, a cross-modality synthesis mechanism performs time and phase shifting of data to allow data from different modalities to align. In some examples, it is recognized that an EEG response will often occur hundreds of milliseconds before a facial emotion measurement changes. Correlations can be drawn and time and phase shifts made on an individual as well as a group basis. In other examples, saccadic eye movements may be determined as occurring before and after particular EEG responses. According to various examples, time corrected GSR measures are used to scale and enhance the EEG estimates of significance including attention, emotional engagement and memory retention measures.


Evidence of the occurrence or non-occurrence of specific time domain difference event-related potential components (like the DERP) in specific regions correlates with subject responsiveness to specific stimulus. According to various examples, ERP measures are enhanced using EEG time-frequency measures (ERPSP) in response to the presentation of the marketing and entertainment stimuli. Specific portions are extracted and isolated to identify ERP, DERP and ERPSP analyses to perform. In particular examples, an EEG frequency estimation of attention, emotion and memory retention (ERPSP) is used as a co-factor in enhancing the ERP, DERP and time-domain response analysis.


EOG measures saccades to determine the presence of attention to specific objects of stimulus. Eye tracking measures the subject's gaze path, location and dwell on specific objects of stimulus. According to various examples, EOG and eye tracking is enhanced by measuring the presence of lambda waves (a neurophysiological index of saccade effectiveness) in the ongoing EEG in the occipital and extra striate regions, triggered by the slope of saccade-onset to estimate the significance of the EOG and eye tracking measures. In particular examples, specific EEG signatures of activity such as slow potential shifts and measures of coherence in time-frequency responses at the Frontal Eye Field (FEF) regions that preceded saccade-onset are measured to enhance the effectiveness of the saccadic activity data.


GSR typically measures the change in general arousal in response to stimulus presented. According to various examples, GSR is enhanced by correlating EEG/ERP responses and the GSR measurement to get an enhanced estimate of subject engagement. The GSR latency baselines are used in constructing a time-corrected GSR response to the stimulus. The time-corrected GSR response is co-factored with the EEG measures to enhance GSR significance measures.


According to various examples, facial emotion encoding uses templates generated by measuring facial muscle positions and movements of individuals expressing various emotions prior to the testing session. These individual specific facial emotion encoding templates are matched with the individual responses to identify subject emotional response. In particular examples, these facial emotion encoding measurements are enhanced by evaluating inter-hemispherical asymmetries in EEG responses in specific frequency bands and measuring frequency band interactions. The techniques of the present disclosure recognize that not only are particular frequency bands significant in EEG responses, but particular frequency bands used for communication between particular areas of the brain are significant. Consequently, these EEG responses enhance the EMG, graphic and video based facial emotion identification.


According to various examples, post-stimulus versus pre-stimulus differential measurements of ERP time domain components in multiple regions of the brain (DERP) are measured at 607. The differential measures give a mechanism for eliciting responses attributable to the stimulus. For example the messaging response attributable to an advertisement or the brand response attributable to multiple brands is determined using pre-resonance and post-resonance estimates


At 609, target versus distracter stimulus differential responses are determined for different regions of the brain (DERP). At 611, event related time-frequency analysis of the differential response (DERPSPs) are used to assess the attention, emotion and memory retention measures across multiple frequency bands. According to various examples, the multiple frequency bands include theta, alpha, beta, gamma and high gamma or kappa. At 613, priming levels and resonance for various products, services, and offerings are determined at different locations in the stimulus material. In some examples, priming levels and resonance are manually determined. In other examples, priming levels and resonance are automatically determined using neuro-response measurements. According to various examples, video streams are modified with different inserted advertisements for various products and services to determine the effectiveness of the inserted advertisements based on priming levels and resonance of the source material.


At 617, multiple trials are performed to enhance priming and resonance measures. In some examples, stimulus. In some examples, multiple trials are performed to enhance resonance measures.


In particular examples, the priming and resonance measures are sent to a priming repository 619. The priming repository 619 may be used to automatically select advertising suited for particular ad breaks.



FIG. 7 illustrates an example of a technique for estimating resonance. According to various examples, measurements from different modalities are obtained. According to various examples, measurements including Differential Event Related Potential (DERP), Differential Event Related Power Spectral Perturbations (DERPSPs), Pupilary Response, etc., are blended to obtain a combined measurement. In particular examples, each measurement may have to be aligned appropriately in order to allow blending. According to various examples, a resonance estimator includes mechanisms to use and blend different measures from across the modalities from the data analyzer. In particular examples, the data includes the DERP measures, DERPSPs, pupilary response, GSR, eye movement, coherence, coupling and lambda wave based response. Measurements across modalities are blended to elicit a synthesized measure of user resonance.


In particular examples, user resonance to attributes of stimulus material such as communication, concept, experience, message, images, genre, product categories, service categories, etc. are measured at 701. The attributes of the stimulus material are evaluated to identify ad categories and genres that are naturally primed as a consequence of the preceding content 703. The effectiveness of source material may be determined using a mechanism to weigh and combine the outputs of the data analyzer. According to various examples, a set of predetermined weights and nonlinear functions combine the outputs of the data analyzer to determine a hierarchy of the effectiveness of a set of categories for products and services that are primed by the source material or pre-break show content at 705. According to various examples, a set of predetermined weights and nonlinear functions combine the outputs of the data analyzer to determine scenes of a show of maximal effectiveness to perform a differential extraction of categories of products and services that are effectively primed by the source material at 707.


At 711, priming levels for various products and services are correlated with various ad breaks based on the number of scenes of maximal effectiveness, the number of categories of products and services primed by the pre break content, and the level of priming effectiveness for each category.


At 713, priming levels and resonance are maintained in a priming level repository. In some examples, the priming levels and resonance are written to the source material.


According to various examples, various mechanisms such as the data collection mechanisms, the intra-modality synthesis mechanisms, cross-modality synthesis mechanisms, etc. are implemented on multiple devices. However, it is also possible that the various mechanisms be implemented in hardware, firmware, and/or software in a single system. FIG. 8 provides one example of a system that can be used to implement one or more mechanisms. For example, the system shown in FIG. 8 may be used to implement a resonance measurement system.


According to particular examples, a system 800 suitable for implementing particular examples of the present disclosure includes a processor 801, a memory 803, an interface 811, and a bus 815 (e.g., a PCI bus). When acting under the control of appropriate software or firmware, the processor 801 is responsible for such tasks such as pattern generation. Various specially configured devices can also be used in place of a processor 801 or in addition to processor 801. The complete implementation can also be done in custom hardware. The interface 811 is typically configured to send and receive data packets or data segments over a network. Particular examples of interfaces the device supports include host bus adapter (HBA) interfaces, Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like.


In addition, various very high-speed interfaces may be provided such as fast Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as data synthesis.


According to particular examples, the system 800 uses memory 803 to store data, algorithms and program instructions. The program instructions may control the operation of an operating system and/or one or more applications, for example. The memory or memories may also be configured to store received data and process received data.


Because such information and program instructions may be employed to implement the systems/methods described herein, the present disclosure relates to tangible, machine readable media that include program instructions, state information, etc. for performing various operations described herein. Examples of machine-readable media include, but are not limited to, magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM disks and DVDs; magneto-optical media such as optical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM). Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.


Although the foregoing disclosure has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced within the scope of the appended claims. Therefore, the present examples are to be considered as illustrative and not restrictive and the disclosure is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Claims
  • 1. A system to identify an advertisement to include in source material to increase an effectiveness of the advertisement, the system comprising: memory;instructions in the system; andprocessor circuitry to execute the instructions to: determine one or more priming characteristics for a plurality of locations of a source material based on neuro-response data collected from a first subject exposed to the source material, the neuro-response data indicative of a resonance of the first subject to the source material associated with the plurality of locations;identify an attribute of the advertisement;identify at least one of a temporal attribute or a spatial attribute for the plurality of locations;perform a comparison of the attribute of the advertisement to the at least one of the temporal attribute or the spatial attribute for the plurality of locations;select a first location of the plurality of locations for insertion of the advertisement based on the comparison and the priming characteristics; andtransform the source material to include the advertisement at the first location.
  • 2. The system of claim 1, wherein the processor circuitry is to execute the instructions to further perform the comparison based on the resonance of the first subject.
  • 3. The system of claim 1, wherein the processor circuitry is to execute the instructions to tag the source material with one or more of the resonance or the priming characteristics.
  • 4. The system of claim 1, wherein the processor circuitry is to execute the instructions to: determine an effectiveness of the advertisement at the first location based on second neuro-response data collected from the first subject or a second subject exposed to the source material including the advertisement at the first location; andselect a second location of the source material to receive the advertisement or a second advertisement based on the effectiveness.
  • 5. The system of claim 1, wherein the attribute of the advertisement includes a format of the advertisement and the spatial attribute for the plurality of locations includes a format of the source material and wherein the processor circuitry is to execute the instructions to correlate the format of the advertisement with the format of the source material.
  • 6. A system to identify an advertisement to include in source material to increase an effectiveness of the advertisement, the system comprising: memory;instructions in the system; andprocessor circuitry to execute the instructions to: determine a first resonance of a first subject to a first portion of the source material and a second resonance of the first subject to a second portion of the source material based on first neuro-response data gathered from the first subject exposed to the source material, the first neuro-response data including two frequency bands of electroencephalographic data, the first resonance to be determined based on a coherence between the two frequency bands;identify a first priming level of the first portion of the source material based on a first attribute of the source material and the first resonance;identify a second priming level of the second portion of the source material based on a second attribute of the source material and the second resonance;retrieve (a) a first meta-tag representative of at least one of a temporal attribute or a spatial attribute of the first portion and (b) a second meta-tag representative at least one of a temporal attribute or a spatial attribute of the second portion;identify a correlation between (a) an attribute of the advertisement and (b) the first meta-tag or the second meta-tag;select (a) the first portion of the source material or (b) the second portion of the source material for inclusion of the advertisement based on the first priming level, the second priming level, and the correlation; andtransform the source material to include the advertisement at the first portion or the second portion.
  • 7. The system of claim 6, wherein the processor circuitry is to execute the instructions to: analyze second neuro-response data gathered from the first subject or a second subject exposed to the advertisement to determine a third resonance to the advertisement; andselect (a) the first portion of the source material or (b) the second portion of the source material for inclusion of the advertisement based on the third resonance.
  • 8. The system of claim 6, wherein the processor circuitry is to execute the instructions to appraise a financial value of the first portion of the source material and the second portion of the source material based on the first priming level and the second priming level.
  • 9. The system of claim 6, wherein the first portion of the source material and the second portion of the source material are temporally separate.
  • 10. The system of claim 6, wherein the first portion of the source material and the second portion of the source material are temporally concurrent, and the first portion of the source material is spaced a distance from the second portion of the source material.
  • 11. The system of claim 6, wherein the spatial attribute of the first meta-tag indicates a location in the source material where the advertisement is to be placed and the attribute of the advertisement is a format of the advertisement.
  • 12. The system of claim 6, wherein the advertisement is a product placement within a scene of the source material.
  • 13. The system of claim 6, wherein the second resonance is based on a second coherence between the two frequency bands.
  • 14. The system of claim 6, wherein the two frequency bands of electroencephalographic data are gathered from a frontal eye field region of a brain of the first subject.
  • 15. The system of claim 6, wherein the two frequency bands of electroencephalographic data are gathered prior to saccade onset.
  • 16. A system comprising: memory;instructions in the system; andprocessor circuitry to execute the instructions to: generate neuro-response data from electroencephalography data collected from a brain of a subject exposed to an advertisement, the neuro-response data representative of an interaction between a first frequency band of the electroencephalography data and a second frequency band of the electroencephalography data, the second frequency band being different than the first frequency band and occurring at a same time, the interaction between the first frequency band and the second frequency band being represented by a degree of phase synchrony corresponding to a repeating sequence of relative phase angles between a first pattern of oscillation in the first frequency band and a second pattern of oscillation in the second frequency band;identify one or more priming levels for one or more locations in a source material based on an attribute of the source material;determine resonance to the advertisement based on the degree of phase synchrony; andselect a first one of the one or more locations for inclusion of the advertisement based on the resonance and the one or more priming levels.
  • 17. The system of claim 16, wherein the electroencephalography data includes event related potential measurements at multiple regions of the brain, and the processor circuitry is to execute the instructions to determine differential measurements based on time domain components of the event related potential measurements.
  • 18. The system of claim 17, wherein the processor circuitry is to execute the instructions to perform event related time-frequency analysis of the differential measurements to assess at least one of attention, emotion or memory retention across multiple frequency bands.
  • 19. The system of claim 16, wherein at least one of the locations is an advertising break in the source material.
  • 20. The system of claim 16, wherein the processor circuitry is to execute the instructions to validate the neuro-response data with one or more of pupillary dilation data, galvanic skin response data, or electrocardiogram data.
CROSS-REFERENCE TO RELATED APPLICATIONS

This patent arises from a continuation of U.S. patent application Ser. No. 16/151,050 (now U.S. Pat. No. 11,023,920), titled “CONTENT BASED SELECTION AND META TAGGING OF ADVERTISEMENT BREAKS,” filed on Oct. 3, 2018, which is a continuation of U.S. patent application Ser. No. 13/730,564 (now U.S. Pat. No. 10,140,628), titled “CONTENT BASED SELECTION AND META TAGGING OF ADVERTISEMENT BREAKS,” filed on Dec. 28, 2012, which is a continuation of U.S. patent application Ser. No. 12/200,813 (now U.S. Pat. No. 8,392,255), titled “CONTENT BASED SELECTION AND META TAGGING OF ADVERTISEMENT BREAKS,” filed on Aug. 28, 2008, which claims the benefit under 35 U.S.C. § 119(e) of U.S. Patent Provisional Application 60/968,567, titled “CONTENT BASED SELECTION AND META TAGGING OF AD BREAKS,” filed on Aug. 29, 2007. U.S. patent application Ser. Nos. 16/151,050; 13/730,564; 12/200,813; and U.S. 60/968,567 are hereby incorporated by reference in their entireties.

US Referenced Citations (606)
Number Name Date Kind
2549836 Mcintyre Apr 1951 A
3490439 Rolston Jan 1970 A
3572322 Wade Mar 1971 A
3735753 Pisarski May 1973 A
3880144 Coursin et al. Apr 1975 A
3901215 John Aug 1975 A
3998213 Price Dec 1976 A
4075657 Weinblatt Feb 1978 A
4145122 Rinard et al. Mar 1979 A
4149716 Scudder Apr 1979 A
4201224 John May 1980 A
4279258 John Jul 1981 A
4411273 John Oct 1983 A
4417592 John Nov 1983 A
4537198 Corbett Aug 1985 A
4557270 John Dec 1985 A
4610259 Cohen et al. Sep 1986 A
4613951 Chu Sep 1986 A
4626904 Lurie Dec 1986 A
4632122 Johansson et al. Dec 1986 A
4683892 Johansson et al. Aug 1987 A
4686999 Snyder et al. Aug 1987 A
4695879 Weinblatt Sep 1987 A
4736751 Gevins et al. Apr 1988 A
4800888 Itil et al. Jan 1989 A
4802484 Friedman et al. Feb 1989 A
4846190 John Jul 1989 A
4859050 Borah et al. Aug 1989 A
4870579 Hey Sep 1989 A
4885687 Carey Dec 1989 A
4894777 Negishi et al. Jan 1990 A
4913160 John Apr 1990 A
4955388 Silberstein Sep 1990 A
4967038 Gevins et al. Oct 1990 A
4973149 Hutchinson Nov 1990 A
4987903 Keppel et al. Jan 1991 A
5003986 Finitzo et al. Apr 1991 A
5010891 Chamoun Apr 1991 A
5038782 Gevins et al. Aug 1991 A
5052401 Sherwin Oct 1991 A
5083571 Prichep Jan 1992 A
RE34015 Duffy Aug 1992 E
5137027 Rosenfeld Aug 1992 A
5213338 Brotz May 1993 A
5226177 Nickerson Jul 1993 A
5243517 Schmidt et al. Sep 1993 A
5273037 Itil et al. Dec 1993 A
5291888 Tucker Mar 1994 A
5293867 Oommen Mar 1994 A
5295491 Gevins Mar 1994 A
5331544 Lu et al. Jul 1994 A
5339826 Schmidt et al. Aug 1994 A
5345281 Taboada et al. Sep 1994 A
5357957 Itil et al. Oct 1994 A
5363858 Farwell Nov 1994 A
5392788 Hudspeth Feb 1995 A
5406956 Farwell Apr 1995 A
5410609 Kado et al. Apr 1995 A
5436830 Zaltman Jul 1995 A
5447166 Gevins Sep 1995 A
5450855 Rosenfeld Sep 1995 A
5474082 Junker Dec 1995 A
5479934 Imran Jan 1996 A
5513649 Gevins et al. May 1996 A
5518007 Becker May 1996 A
5537618 Boulton et al. Jul 1996 A
5550928 Lu et al. Aug 1996 A
5617855 Waletzky et al. Apr 1997 A
5655534 Ilmoniemi Aug 1997 A
5676138 Zawilinski Oct 1997 A
5676148 Koo et al. Oct 1997 A
5687322 Deaton et al. Nov 1997 A
5720619 Fisslinger Feb 1998 A
5724987 Gevins et al. Mar 1998 A
5726701 Needham Mar 1998 A
5729205 Kwon Mar 1998 A
5736986 Sever, Jr. Apr 1998 A
5740035 Cohen et al. Apr 1998 A
5762611 Lewis et al. Jun 1998 A
5771897 Zufrin Jun 1998 A
5774591 Black et al. Jun 1998 A
5787187 Bouchard et al. Jul 1998 A
5800351 Mann Sep 1998 A
5802208 Podilchuk et al. Sep 1998 A
5802220 Black et al. Sep 1998 A
5812642 Leroy Sep 1998 A
5817029 Gevins et al. Oct 1998 A
5842199 Miller et al. Nov 1998 A
5848399 Burke Dec 1998 A
5892566 Bullwinkel Apr 1999 A
5945863 Coy Aug 1999 A
5961332 Joao Oct 1999 A
5974262 Fuller et al. Oct 1999 A
5983129 Cowan et al. Nov 1999 A
5995868 Dorfmeister et al. Nov 1999 A
6001065 DeVito Dec 1999 A
6016475 Miller et al. Jan 2000 A
6021346 Ryu et al. Feb 2000 A
6032129 Greef et al. Feb 2000 A
6052619 John Apr 2000 A
6088040 Oda et al. Jul 2000 A
6099319 Zaltman et al. Aug 2000 A
6120440 Goknar Sep 2000 A
6128521 Marro et al. Oct 2000 A
6154669 Hunter et al. Nov 2000 A
6155927 Levasseur et al. Dec 2000 A
6161030 Levendowski et al. Dec 2000 A
6170018 Voll et al. Jan 2001 B1
6171239 Humphrey Jan 2001 B1
6173260 Slaney Jan 2001 B1
6175753 Menkes et al. Jan 2001 B1
6182113 Narayanaswami Jan 2001 B1
6190314 Ark et al. Feb 2001 B1
6212502 Ball et al. Apr 2001 B1
6228038 Claessens May 2001 B1
6236885 Hunter et al. May 2001 B1
6236975 Boe et al. May 2001 B1
6254536 DeVito Jul 2001 B1
6280198 Calhoun et al. Aug 2001 B1
6286005 Cannon Sep 2001 B1
6289234 Mueller Sep 2001 B1
6292688 Patton Sep 2001 B1
6299308 Voronka et al. Oct 2001 B1
6301492 Zonenshayn Oct 2001 B1
6301493 Marro et al. Oct 2001 B1
6315569 Zaltman Nov 2001 B1
6330470 Tucker et al. Dec 2001 B1
6334778 Brown Jan 2002 B1
6358201 Childre et al. Mar 2002 B1
6370513 Kolawa et al. Apr 2002 B1
6374143 Berrang et al. Apr 2002 B1
6381481 Levendowski et al. Apr 2002 B1
6398643 Knowles et al. Jun 2002 B1
6422999 Hill Jul 2002 B1
6434419 Gevins et al. Aug 2002 B1
6435878 Reynolds et al. Aug 2002 B1
6453194 Hill Sep 2002 B1
6453241 Bassett, Jr. et al. Sep 2002 B1
6487444 Mimura Nov 2002 B2
6488617 Katz Dec 2002 B1
6510333 Licata et al. Jan 2003 B1
6510340 Jordan Jan 2003 B1
6520905 Surve et al. Feb 2003 B1
6545685 Dorbie Apr 2003 B1
6575902 Burton Jun 2003 B1
6577329 Flickner et al. Jun 2003 B1
6585521 Obrador Jul 2003 B1
6594521 Tucker Jul 2003 B2
6598006 Honda et al. Jul 2003 B1
6609024 Ryu et al. Aug 2003 B1
6648822 Hamamoto et al. Nov 2003 B2
6652283 Van Schaack et al. Nov 2003 B1
6654626 Devlin et al. Nov 2003 B2
6662052 Sarwal et al. Dec 2003 B1
6665560 Becker et al. Dec 2003 B2
6678685 McGill et al. Jan 2004 B2
6688890 von Buegner Feb 2004 B2
6708051 Durousseau Mar 2004 B1
6712468 Edwards Mar 2004 B1
6754524 Johnson, Jr. Jun 2004 B2
6757556 Gopinathan et al. Jun 2004 B2
6788882 Geer et al. Sep 2004 B1
6792304 Silberstein Sep 2004 B1
6842877 Robarts et al. Jan 2005 B2
6850252 Hoffberg Feb 2005 B1
6852875 Prakash Feb 2005 B2
6888457 Wilkinson et al. May 2005 B2
6904408 McCarthy et al. Jun 2005 B1
6950698 Sarkela et al. Sep 2005 B2
6958710 Zhang et al. Oct 2005 B2
6973342 Swanson Dec 2005 B1
6993380 Modarres Jan 2006 B1
7020508 Stivoric et al. Mar 2006 B2
7043056 Edwards et al. May 2006 B2
7047550 Yasukawa et al. May 2006 B1
7113916 Hill Sep 2006 B1
7120880 Dryer et al. Oct 2006 B1
7130673 Tolvanen-Laakso et al. Oct 2006 B2
7150715 Collura et al. Dec 2006 B2
7164967 Etienne-Cummings et al. Jan 2007 B2
7177675 Suffin et al. Feb 2007 B2
7222071 Neuhauser et al. May 2007 B2
7246081 Hill Jul 2007 B2
7249708 McConnell et al. Jul 2007 B2
7272982 Neuhauser et al. Sep 2007 B2
7286871 Cohen Oct 2007 B2
7340060 Tomkins et al. Mar 2008 B2
7359894 Liebman et al. Apr 2008 B1
7391835 Gross et al. Jun 2008 B1
7394385 Franco, Jr. et al. Jul 2008 B2
7408460 Crystal et al. Aug 2008 B2
7420464 Fitzgerald et al. Sep 2008 B2
7443292 Jensen et al. Oct 2008 B2
7460827 Schuster et al. Dec 2008 B2
7463143 Forr et al. Dec 2008 B2
7463144 Crystal et al. Dec 2008 B2
7471987 Crystal et al. Dec 2008 B2
7483835 Neuhauser et al. Jan 2009 B2
7483844 Takakura et al. Jan 2009 B2
7496400 Hoskonen et al. Feb 2009 B2
7548774 Kurtz et al. Jun 2009 B2
7551952 Gevins et al. Jun 2009 B2
7592908 Zhang et al. Sep 2009 B2
7614066 Urdang et al. Nov 2009 B2
7623823 Zito et al. Nov 2009 B2
7630757 Dorfmeister et al. Dec 2009 B2
7636456 Collins et al. Dec 2009 B2
7641341 Weinblatt Jan 2010 B2
7650793 Jensen et al. Jan 2010 B2
7658327 Tuchman et al. Feb 2010 B2
7689272 Farwell Mar 2010 B2
7697979 Martinerie et al. Apr 2010 B2
7698238 Barletta et al. Apr 2010 B2
7720351 Levitan May 2010 B2
7729755 Laken Jun 2010 B2
7765564 Deng Jul 2010 B2
7774052 Burton et al. Aug 2010 B2
7797186 Dybus Sep 2010 B2
7809420 Hannula et al. Oct 2010 B2
7840248 Fuchs et al. Nov 2010 B2
7840250 Tucker Nov 2010 B2
7844484 Arnett et al. Nov 2010 B2
7865394 Calloway et al. Jan 2011 B1
7892764 Xiong et al. Feb 2011 B2
7895075 Gettys et al. Feb 2011 B2
7895625 Bryan et al. Feb 2011 B1
7908133 Neuhauser Mar 2011 B2
7917366 Levanon et al. Mar 2011 B1
7930199 Hill Apr 2011 B1
7962315 Jensen et al. Jun 2011 B2
7988557 Soderlund Aug 2011 B2
8014847 Shastri et al. Sep 2011 B2
8027518 Baker et al. Sep 2011 B2
8065203 Chien et al. Nov 2011 B1
8073707 Teller et al. Dec 2011 B2
8098152 Zhang et al. Jan 2012 B2
8099315 Amento et al. Jan 2012 B2
8126220 Greig Feb 2012 B2
8135606 Dupree Mar 2012 B2
8151298 Begeja et al. Apr 2012 B2
8165916 Hoffberg et al. Apr 2012 B2
8196168 Bryan et al. Jun 2012 B1
8200775 Moore Jun 2012 B2
8209224 Pradeep et al. Jun 2012 B2
8229469 Zhang et al. Jul 2012 B2
8235725 Hill Aug 2012 B1
8255267 Breiter Aug 2012 B2
8270814 Pradeep et al. Sep 2012 B2
8296172 Marci et al. Oct 2012 B2
8300526 Saito et al. Oct 2012 B2
8335715 Pradeep et al. Dec 2012 B2
8381244 King et al. Feb 2013 B2
8386312 Pradeep et al. Feb 2013 B2
8386313 Pradeep et al. Feb 2013 B2
8392250 Pradeep et al. Mar 2013 B2
8392251 Pradeep et al. Mar 2013 B2
8392253 Pradeep et al. Mar 2013 B2
8392254 Pradeep et al. Mar 2013 B2
8392255 Pradeep et al. Mar 2013 B2
8396744 Pradeep et al. Mar 2013 B2
8473345 Pradeep et al. Jun 2013 B2
8484081 Pradeep et al. Jul 2013 B2
8484801 Li et al. Jul 2013 B2
8494610 Pradeep et al. Jul 2013 B2
8494905 Pradeep et al. Jul 2013 B2
8533042 Pradeep et al. Sep 2013 B2
8548852 Pradeep et al. Oct 2013 B2
8561095 Dimitrova et al. Oct 2013 B2
8635105 Pradeep et al. Jan 2014 B2
8655428 Pradeep et al. Feb 2014 B2
8655437 Pradeep et al. Feb 2014 B2
8762202 Pradeep et al. Jun 2014 B2
8764652 Lee et al. Jul 2014 B2
8788372 Kettner et al. Jul 2014 B2
9336535 Pradeep et al. May 2016 B2
9560984 Pradeep et al. Feb 2017 B2
9886981 Pradeep et al. Feb 2018 B2
10068248 Knight et al. Sep 2018 B2
10127572 Pradeep et al. Nov 2018 B2
10140628 Pradeep et al. Nov 2018 B2
10173095 Gomberg et al. Jan 2019 B2
10189294 Raymond et al. Jan 2019 B2
10269036 Knight et al. Apr 2019 B2
10580031 Pradeep et al. Mar 2020 B2
10679241 Pradeep et al. Jun 2020 B2
11023920 Pradeep et al. Jun 2021 B2
20010013009 Greening et al. Aug 2001 A1
20010020236 Cannon Sep 2001 A1
20010029468 Yamaguchi et al. Oct 2001 A1
20010032140 Hoffman Oct 2001 A1
20010056225 DeVito Dec 2001 A1
20020053076 Landesmann May 2002 A1
20020055857 Mault May 2002 A1
20020056087 Berezowski et al. May 2002 A1
20020056124 Hay May 2002 A1
20020059577 Lu et al. May 2002 A1
20020065826 Bell et al. May 2002 A1
20020072952 Hamzy et al. Jun 2002 A1
20020077534 DuRousseau Jun 2002 A1
20020082902 Ando et al. Jun 2002 A1
20020103429 deCharms Aug 2002 A1
20020111796 Nemoto Aug 2002 A1
20020116042 Boling Aug 2002 A1
20020143627 Barsade et al. Oct 2002 A1
20020155878 Lert, Jr. et al. Oct 2002 A1
20020156842 Signes et al. Oct 2002 A1
20020169665 Hughes et al. Nov 2002 A1
20020178440 Agnihotri et al. Nov 2002 A1
20020188216 Kayyali et al. Dec 2002 A1
20020188217 Farwell Dec 2002 A1
20020193670 Garfield et al. Dec 2002 A1
20030013981 Gevins et al. Jan 2003 A1
20030036955 Tanaka et al. Feb 2003 A1
20030037333 Ghashghai et al. Feb 2003 A1
20030044050 Clark et al. Mar 2003 A1
20030059750 Bindler et al. Mar 2003 A1
20030063222 Creed et al. Apr 2003 A1
20030065524 Giacchetti et al. Apr 2003 A1
20030073921 Sohmer et al. Apr 2003 A1
20030081834 Philomin et al. May 2003 A1
20030093792 Labeeb et al. May 2003 A1
20030100998 Brunner et al. May 2003 A1
20030104865 Itkis et al. Jun 2003 A1
20030131351 Shapira Jul 2003 A1
20030149344 Nizan Aug 2003 A1
20030165270 Endrikhovski et al. Sep 2003 A1
20030177488 Smith et al. Sep 2003 A1
20030204412 Brier Oct 2003 A1
20030208754 Sridhar et al. Nov 2003 A1
20030233278 Marshall Dec 2003 A1
20040001616 Gutta et al. Jan 2004 A1
20040005143 Tsuru et al. Jan 2004 A1
20040013398 Miura et al. Jan 2004 A1
20040015608 Ellis et al. Jan 2004 A1
20040055448 Byon Mar 2004 A1
20040068431 Smith et al. Apr 2004 A1
20040073129 Caldwell et al. Apr 2004 A1
20040092809 DeCharms May 2004 A1
20040098298 Yin May 2004 A1
20040101212 Fedorovskaya et al. May 2004 A1
20040133081 Teller et al. Jul 2004 A1
20040187167 Maguire et al. Sep 2004 A1
20040193068 Burton et al. Sep 2004 A1
20040210159 Kibar Oct 2004 A1
20040219184 Brown et al. Nov 2004 A1
20040220483 Yeo et al. Nov 2004 A1
20040236623 Gopalakrishnan Nov 2004 A1
20050010475 Perkowski et al. Jan 2005 A1
20050041951 Inoue et al. Feb 2005 A1
20050043646 Viirre et al. Feb 2005 A1
20050060312 Curtiss et al. Mar 2005 A1
20050062637 El Zabadani et al. Mar 2005 A1
20050071462 Bodin et al. Mar 2005 A1
20050071865 Martins Mar 2005 A1
20050076359 Pierson et al. Apr 2005 A1
20050079474 Lowe Apr 2005 A1
20050097594 O'Donnell et al. May 2005 A1
20050107716 Eaton et al. May 2005 A1
20050108092 Campbell et al. May 2005 A1
20050113649 Bergantino May 2005 A1
20050132401 Boccon-Gibod et al. Jun 2005 A1
20050143629 Farwell Jun 2005 A1
20050149964 Thomas et al. Jul 2005 A1
20050154290 Langleben Jul 2005 A1
20050165766 Szabo Jul 2005 A1
20050177058 Sobell Aug 2005 A1
20050197590 Osorio et al. Sep 2005 A1
20050203798 Jensen et al. Sep 2005 A1
20050216071 Devlin et al. Sep 2005 A1
20050216243 Graham et al. Sep 2005 A1
20050223237 Barletta et al. Oct 2005 A1
20050227233 Buxton et al. Oct 2005 A1
20050240956 Smith et al. Oct 2005 A1
20050246002 Martinez Nov 2005 A1
20050256905 Gruhl et al. Nov 2005 A1
20050261980 Hadi Nov 2005 A1
20050267798 Panara Dec 2005 A1
20050272017 Neuhauser et al. Dec 2005 A1
20050273017 Gordon Dec 2005 A1
20050273802 Crystal et al. Dec 2005 A1
20050288954 McCarthy et al. Dec 2005 A1
20050289582 Tavares et al. Dec 2005 A1
20060003732 Neuhauser et al. Jan 2006 A1
20060009702 Iwaki et al. Jan 2006 A1
20060010470 Kurosaki et al. Jan 2006 A1
20060035707 Nguyen et al. Feb 2006 A1
20060041548 Parsons et al. Feb 2006 A1
20060042483 Work et al. Mar 2006 A1
20060053110 McDonald et al. Mar 2006 A1
20060069663 Adar et al. Mar 2006 A1
20060093998 Vertegaal May 2006 A1
20060094934 Shirai et al. May 2006 A1
20060111044 Keller May 2006 A1
20060111644 Guttag et al. May 2006 A1
20060129458 Maggio Jun 2006 A1
20060149337 John Jul 2006 A1
20060167376 Viirre et al. Jul 2006 A1
20060168613 Wood et al. Jul 2006 A1
20060168630 Davies Jul 2006 A1
20060176289 Horn Aug 2006 A1
20060189886 Jones et al. Aug 2006 A1
20060190822 Basson et al. Aug 2006 A1
20060218046 Carfi et al. Sep 2006 A1
20060256133 Rosenberg Nov 2006 A1
20060257834 Lee et al. Nov 2006 A1
20060259360 Flinn et al. Nov 2006 A1
20060259371 Perrier et al. Nov 2006 A1
20060293921 McCarthy et al. Dec 2006 A1
20070005752 Chawla et al. Jan 2007 A1
20070016096 McNabb Jan 2007 A1
20070038516 Apple et al. Feb 2007 A1
20070048707 Caamano et al. Mar 2007 A1
20070050256 Walker et al. Mar 2007 A1
20070055169 Lee et al. Mar 2007 A1
20070060830 Le et al. Mar 2007 A1
20070060831 Le et al. Mar 2007 A1
20070066874 Cook Mar 2007 A1
20070066915 Frei et al. Mar 2007 A1
20070066916 Lemos Mar 2007 A1
20070067007 Schulman et al. Mar 2007 A1
20070067305 Ives Mar 2007 A1
20070078700 Lenzmann et al. Apr 2007 A1
20070078706 Datta et al. Apr 2007 A1
20070079331 Datta et al. Apr 2007 A1
20070101360 Gutta et al. May 2007 A1
20070106170 Dunseath et al. May 2007 A1
20070112460 Kiselik May 2007 A1
20070135727 Virtanen et al. Jun 2007 A1
20070135728 Snyder et al. Jun 2007 A1
20070136753 Bovenschulte et al. Jun 2007 A1
20070150281 Hoff Jun 2007 A1
20070150916 Begole et al. Jun 2007 A1
20070209047 Hallberg et al. Sep 2007 A1
20070214471 Rosenberg Sep 2007 A1
20070225585 Washbon et al. Sep 2007 A1
20070225674 Molnar et al. Sep 2007 A1
20070226760 Neuhauser et al. Sep 2007 A1
20070235716 Delic et al. Oct 2007 A1
20070238945 Delic et al. Oct 2007 A1
20070244977 Atkins Oct 2007 A1
20070250846 Swix et al. Oct 2007 A1
20070250901 McIntire et al. Oct 2007 A1
20070265507 de Lemos Nov 2007 A1
20070282566 Whitlow et al. Dec 2007 A1
20070294132 Zhang et al. Dec 2007 A1
20070294705 Gopalakrishnan et al. Dec 2007 A1
20070294706 Neuhauser et al. Dec 2007 A1
20080001600 deCharms Jan 2008 A1
20080004940 Rolleston Phillips Jan 2008 A1
20080010110 Neuhauser et al. Jan 2008 A1
20080017883 Sarid et al. Jan 2008 A1
20080024725 Todd Jan 2008 A1
20080027345 Kumada et al. Jan 2008 A1
20080040740 Plotnick et al. Feb 2008 A1
20080043013 Gruttadauria et al. Feb 2008 A1
20080059997 Plotnick et al. Mar 2008 A1
20080065468 Berg et al. Mar 2008 A1
20080065721 Cragun Mar 2008 A1
20080081961 Westbrook et al. Apr 2008 A1
20080082019 Ludving et al. Apr 2008 A1
20080086356 Glassman et al. Apr 2008 A1
20080091463 Shakamuri Apr 2008 A1
20080091512 Marci et al. Apr 2008 A1
20080097854 Young Apr 2008 A1
20080109840 Walter et al. May 2008 A1
20080125110 Ritter May 2008 A1
20080133724 Clark Jun 2008 A1
20080147448 McLaughlin et al. Jun 2008 A1
20080147742 Allen Jun 2008 A1
20080152300 Knee et al. Jun 2008 A1
20080162182 Cazares et al. Jul 2008 A1
20080177197 Lee et al. Jul 2008 A1
20080195471 Dube et al. Aug 2008 A1
20080204273 Crystal et al. Aug 2008 A1
20080208072 Fadem et al. Aug 2008 A1
20080214902 Lee et al. Sep 2008 A1
20080218472 Breen et al. Sep 2008 A1
20080221400 Lee et al. Sep 2008 A1
20080221472 Lee et al. Sep 2008 A1
20080221969 Lee et al. Sep 2008 A1
20080222670 Lee et al. Sep 2008 A1
20080222671 Lee et al. Sep 2008 A1
20080228077 Wilk et al. Sep 2008 A1
20080249865 Angell et al. Oct 2008 A1
20080255949 Genco et al. Oct 2008 A1
20080263458 Altberg et al. Oct 2008 A1
20080295126 Lee et al. Nov 2008 A1
20080306398 Uchiyama et al. Dec 2008 A1
20090018996 Hunt et al. Jan 2009 A1
20090024049 Pradeep et al. Jan 2009 A1
20090024447 Pradeep et al. Jan 2009 A1
20090024448 Pradeep et al. Jan 2009 A1
20090024449 Pradeep et al. Jan 2009 A1
20090024475 Pradeep et al. Jan 2009 A1
20090024747 Moses et al. Jan 2009 A1
20090025023 Pradeep et al. Jan 2009 A1
20090025024 Beser et al. Jan 2009 A1
20090030287 Pradeep et al. Jan 2009 A1
20090030303 Pradeep et al. Jan 2009 A1
20090030717 Pradeep et al. Jan 2009 A1
20090030762 Lee et al. Jan 2009 A1
20090030780 York et al. Jan 2009 A1
20090030930 Pradeep et al. Jan 2009 A1
20090036755 Pradeep et al. Feb 2009 A1
20090036756 Pradeep et al. Feb 2009 A1
20090037575 Crystal et al. Feb 2009 A1
20090062629 Pradeep et al. Mar 2009 A1
20090062679 Tan et al. Mar 2009 A1
20090062681 Pradeep et al. Mar 2009 A1
20090063255 Pradeep et al. Mar 2009 A1
20090063256 Pradeep et al. Mar 2009 A1
20090070798 Lee et al. Mar 2009 A1
20090082643 Pradeep et al. Mar 2009 A1
20090082689 Guttag et al. Mar 2009 A1
20090083129 Pradeep et al. Mar 2009 A1
20090088610 Lee et al. Apr 2009 A1
20090112077 Nguyen et al. Apr 2009 A1
20090132441 Muller et al. May 2009 A1
20090138356 Pomplun May 2009 A1
20090153328 Otani et al. Jun 2009 A1
20090221928 Einav et al. Sep 2009 A1
20090248496 Hueter et al. Oct 2009 A1
20090253996 Lee et al. Oct 2009 A1
20090259509 Landvater Oct 2009 A1
20090271294 Hadi Oct 2009 A1
20090300672 Van Gulik Dec 2009 A1
20090305006 Steffen Dec 2009 A1
20090327068 Pradeep et al. Dec 2009 A1
20090328089 Pradeep et al. Dec 2009 A1
20100004977 Marci et al. Jan 2010 A1
20100022821 Dubi et al. Jan 2010 A1
20100060300 Muller et al. Mar 2010 A1
20100094702 Silberstein Apr 2010 A1
20100125219 Harris et al. May 2010 A1
20100145215 Pradeep et al. Jun 2010 A1
20100145217 Otto et al. Jun 2010 A1
20100180029 Fourman Jul 2010 A1
20100183279 Pradeep et al. Jul 2010 A1
20100186031 Pradeep et al. Jul 2010 A1
20100186032 Pradeep et al. Jul 2010 A1
20100198042 Popescu et al. Aug 2010 A1
20100214318 Pradeep et al. Aug 2010 A1
20100215289 Pradeep et al. Aug 2010 A1
20100228604 Desai et al. Sep 2010 A1
20100249538 Pradeep et al. Sep 2010 A1
20100249636 Pradeep et al. Sep 2010 A1
20100250325 Pradeep et al. Sep 2010 A1
20100257052 Zito et al. Oct 2010 A1
20100292998 Bodlaender et al. Nov 2010 A1
20110046473 Pradeep et al. Feb 2011 A1
20110046502 Pradeep et al. Feb 2011 A1
20110046503 Pradeep et al. Feb 2011 A1
20110046504 Pradeep et al. Feb 2011 A1
20110047121 Pradeep et al. Feb 2011 A1
20110059422 Masaoka Mar 2011 A1
20110105937 Pradeep et al. May 2011 A1
20110106621 Pradeep et al. May 2011 A1
20110106750 Pradeep et al. May 2011 A1
20110119124 Pradeep et al. May 2011 A1
20110119129 Pradeep et al. May 2011 A1
20110208515 Neuhauser Aug 2011 A1
20110237971 Pradeep et al. Sep 2011 A1
20110248729 Mueller et al. Oct 2011 A2
20110270620 Pradeep et al. Nov 2011 A1
20110276504 Pradeep et al. Nov 2011 A1
20110282231 Pradeep et al. Nov 2011 A1
20110282232 Pradeep et al. Nov 2011 A1
20110282749 Pradeep et al. Nov 2011 A1
20120036004 Pradeep et al. Feb 2012 A1
20120036005 Pradeep et al. Feb 2012 A1
20120046993 Hill Feb 2012 A1
20120054018 Pradeep et al. Mar 2012 A1
20120072289 Pradeep et al. Mar 2012 A1
20120083668 Pradeep et al. Apr 2012 A1
20120084139 Pradeep et al. Apr 2012 A1
20120108995 Pradeep et al. May 2012 A1
20120130800 Pradeep et al. May 2012 A1
20120245978 Jain et al. Sep 2012 A1
20120284112 Pradeep et al. Nov 2012 A1
20120284332 Pradeep et al. Nov 2012 A1
20120290409 Pradeep et al. Nov 2012 A1
20130024272 Pradeep et al. Jan 2013 A1
20130046577 Marci et al. Feb 2013 A1
20130124365 Pradeep May 2013 A1
20130124623 Munter May 2013 A1
20130152506 Pradeep Jun 2013 A1
20130166373 Pradeep et al. Jun 2013 A1
20130185140 Pradeep et al. Jul 2013 A1
20130185141 Pradeep et al. Jul 2013 A1
20130185142 Pradeep et al. Jul 2013 A1
20130185144 Pradeep et al. Jul 2013 A1
20130185145 Pradeep et al. Jul 2013 A1
20130304540 Pradeep et al. Nov 2013 A1
20130332259 Pradeep et al. Dec 2013 A1
20170039591 Knight et al. Feb 2017 A1
20180247332 Pradeep et al. Aug 2018 A1
20180341977 Knight et al. Nov 2018 A1
20190005532 Pradeep et al. Jan 2019 A1
20190034958 Pradeep et al. Jan 2019 A1
20190034959 Pradeep et al. Jan 2019 A1
20190139078 Pradeep et al. May 2019 A1
20190156352 Pradeep et al. May 2019 A1
20190220888 Knight et al. Jul 2019 A1
20190282153 Pradeep et al. Sep 2019 A1
20200005339 Pradeep et al. Jan 2020 A1
20200163571 Pradeep et al. May 2020 A1
Foreign Referenced Citations (32)
Number Date Country
1087618 Mar 2001 EP
1609418 Dec 2005 EP
1374658 Nov 1974 GB
2221759 Feb 1990 GB
2001147944 May 2001 JP
2005051654 Feb 2005 JP
2005-160805 Jun 2005 JP
2006006355 Jan 2006 JP
2006227994 Aug 2006 JP
2006-305334 Nov 2006 JP
200422399 Jul 2006 KR
95-018565 Jul 1995 WO
1997-017774 May 1997 WO
1997-040745 Nov 1997 WO
1997-041673 Nov 1997 WO
02-100241 Dec 2002 WO
02-102238 Dec 2002 WO
2004-049225 Jun 2004 WO
2006-009771 Jan 2006 WO
2008030831 Mar 2008 WO
2008055078 May 2008 WO
2008-064431 Jun 2008 WO
2008-077178 Jul 2008 WO
2008-109694 Sep 2008 WO
2008-109699 Sep 2008 WO
2008-121651 Oct 2008 WO
2008-137579 Nov 2008 WO
2008-137581 Nov 2008 WO
2008-141340 Nov 2008 WO
2008-154410 Dec 2008 WO
2009-018374 Feb 2009 WO
2009-052833 Apr 2009 WO
Non-Patent Literature Citations (589)
Entry
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Jul. 8, 2011, 16 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Jan. 7, 2011, 19 pages.
Office Action issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,221, dated Apr. 15, 2011, 24 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Jun. 9, 2011, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Dec. 27, 2010, 15 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Apr. 21, 2011, 10 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Dec. 3, 2010, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,240, dated Jun. 10, 2011, 12 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated May 26, 2011, 15 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated Dec. 9, 2010, 13 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Jan. 21, 2011, 16 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Oct. 28, 2010, 14 pages.
Notice of Panel Decision from Pre-Appeal Brief Review, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated May 31, 2011, 2 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Dec. 23, 2010, 14 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Jun. 9, 2011, 10 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Jul. 7, 2011, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Dec. 27, 2010, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Dec. 27, 2010, 14 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Jun. 9, 2011, 12 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Jun. 21, 2011, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Dec. 27, 2010, 17 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Jun. 14, 2011, 13 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Dec. 27, 2010, 17 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Jul. 6, 2011, 13 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Dec. 27, 2010, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Jun. 7, 2011, 10 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,069, dated Feb. 17, 2011, 32 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,069, dated Oct. 29, 2010, 21 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,315, dated May 4, 2011, 9 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Jun. 7, 2011, 9 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Jul. 18, 2011, 9 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Jul. 12, 2011, 15 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Aug. 10, 2011, 28 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,322, dated Aug. 23, 2011, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,069, dated Aug. 26, 2011, 33 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Sep. 2, 2011, 7 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Sep. 12, 2011, 12 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,851, dated Sep. 12, 2011, 7 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Sep. 29, 2011, 37 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,225, dated Oct. 3, 2011, 6 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Oct. 12, 2011, 27 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Oct. 13, 2011, 22 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Oct. 19, 2011, 21 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,315, dated Oct. 26, 2011, 41 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,240, dated Oct. 27, 2011, 39 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,221, dated Nov. 28, 2011, 44 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Dec. 7, 2011, 8 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Dec. 22, 2011, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated Dec. 22, 2011, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Dec. 22, 2011, 16 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Dec. 22, 2011, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Dec. 22, 2011, 15 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Dec. 22, 2011, 18 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Dec. 29, 2011, 18 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Jan. 3, 2012, 10 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Jan. 4, 2012, 10 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/544,921, dated Jan. 9, 2012, 13 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,302, dated Jan. 17, 2012, 11 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,225, dated Jan. 20, 2012, 12 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Jan. 24, 2012, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Feb. 1, 2012, 17 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/544,958, dated Feb. 10, 2012, 6 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,069, dated Feb. 14, 2012, 36 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,322, dated Feb. 14, 2012, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Feb. 16, 2012, 16 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Feb. 17, 2012, 22 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Feb. 17, 2012, 20 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Feb. 17, 2012, 15 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated Mar. 1, 2012, 7 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,851, dated Mar. 14, 2012, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Mar. 29, 2012, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/846,242, dated Mar. 29, 2012, 16 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Apr. 6, 2012, 6 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,315, dated Apr. 9, 2012, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/544,958, dated May 2, 2012, 15 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/357,302, dated May 7, 2012, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated May 8, 2012, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,696, dated May 15, 2012, 6 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/545,455, dated Jun. 13, 2012, 5 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,225, dated Jun. 15, 2012, 9 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/544,934, dated Jun. 18, 2012, 11 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Jun. 21, 2012, 10 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Jul. 10, 2012, 14 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Jul. 30, 2012, 15 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Aug. 3, 2012, 8 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Jun. 8, 2012, 12 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated May 23, 2012, 11 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Aug. 28, 2012, 3 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/545,455, dated Aug. 29, 2012, 11 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/778,828, dated Aug. 30, 2012, 9 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/778,810, dated Aug. 31, 2012, 12 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Sep. 17, 2012, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated Sep. 17, 2012, 11 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Sep. 17, 2012, 17 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Sep. 18, 2012, 18 pages.
Final Rejection, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Sep. 18, 2012, 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Sep. 19, 2012, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Sep. 19, 2012, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,213, dated Sep. 7, 2012, 9 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Sep. 26, 2012, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Sep. 27, 2012, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Sep. 28, 2012, 12 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Oct. 1, 2012, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/444,149, dated Oct. 4, 2012, 9 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,851, dated Oct. 4, 2012, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Oct. 5, 2012, 6 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,197, dated Oct. 16, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated Oct. 22, 2012, 5 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/846,242, dated Nov. 29, 2012, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Oct. 30, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Nov. 2, 2012, 5 pages.
Restriction Requirement, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,225, dated Nov. 2, 2012, 5 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,066, dated Nov. 13, 2012, 9 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Nov. 16, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Nov. 21, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated Nov. 23, 2012, 5 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/913,102, dated Dec. 7, 2012, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/544,958, dated Dec. 10, 2012, 16 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,197, dated Dec. 20, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Dec. 21, 2012, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Dec. 21, 2012, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated Dec. 21, 2012, 19 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Dec. 21, 2012, 12 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Dec. 21, 2012, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Dec. 21, 2012, 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Dec. 21, 2012, 9 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,213, dated Dec. 21, 2012, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated Dec. 26, 2012, 2 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Dec. 31, 2012, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Dec. 31, 2012, 10 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Jan. 4, 2013, 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Jan. 11, 2013, 11 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,225, dated Jan. 11, 2013, 11 pages.
Recertified IDS and Interview Summary, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,197, dated Jan. 16, 2013, 6 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,557, dated Jan. 29, 2013, 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,197, dated Jan. 29, 2013, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Jan. 31, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,863, dated Jan. 31, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Jan. 31, 2013, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,583, dated Feb. 1, 2013, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Feb. 1, 2013, 5pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Feb. 4, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,262, dated Feb. 5, 2013, 15 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/868,531, dated Feb. 5, 2013, 8 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/853,213, dated Feb. 5, 2013, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Feb. 15, 2013, 9 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Feb. 14, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/200,813, dated Sep. 20, 2012, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Apr. 16, 2013, 10 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Apr. 22, 2013, 11 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/058264, dated Sep. 29, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/058264, dated Aug. 1, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/058264, dated Aug. 1, 2008, 5 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/062273, dated Nov. 3, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/062273, dated Sep. 5, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/062273, dated Sep. 5, 2008, 4 pages.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/062275, dated Sep. 22, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/062275, dated Sep. 22, 2008, 6 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/063984, dated Nov. 17, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/063984, dated Sep. 29, 2008, 3 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/063984, dated Sep. 29, 2008, 4 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/063989, dated Nov. 17, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/063989, dated Jul. 17, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/063989, dated Jul. 17, 2008, 4 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection International Application No. PCT/US08/066166, dated Dec. 7, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/066166, dated Aug. 25, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/066166, dated Aug. 25, 2008, 6 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/071639, dated Feb. 2, 2010, 1 page.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/071639, dated Oct. 22, 2008, 4 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/074467, dated Mar. 2, 2010, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/074467, dated Nov. 17, 2008, 2 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US08/074467, dated Nov. 17, 2008, 4 pages.
International Preliminary Report of Patentability, issued by the International Bureau in connection with International Application No. PCT/US10/021535, dated Jul. 26, 2011, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US10/021535, dated Mar. 23, 2010, 3 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US10/021535, dated Mar. 23, 2010, 4 pages.
International Preliminary Report of Patentability, issued by the International Bureau in connection with International Application No. PCT/US09/065368, dated Jun. 23, 2011, 2 pages.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US09/065368, dated Jan. 21, 2010, 3 pages.
Written Opinion, issued by the International Searching Authority in connection with International Application No. PCT/US09/065368, dated Jan. 21, 2010, 7 pages.
International Preliminary Report on Patentability, issued by the International Bureau of WIPO in connection with International Application No. PCT/US08/062275, dated Nov. 3, 2009, 1 page.
International Search Report, issued by the International Searching Authority in connection with International Application No. PCT/US08/071639, dated Oct. 22, 2008, 3 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 08744383.4-2221/2130146, dated Jul. 27, 2011, 6 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 10173095.0-2221, dated Dec. 17, 2010, 3 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 10189294.1-2221, dated Mar. 21, 2011, 7 pages.
First Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880104982.1, dated Jan. 25, 2011, 15 pages (includes English translation).
English Translation of First Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 2008801015007, dated May 25, 2011, 8 pages.
English Translation of First Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880019166.0, dated Jul. 22, 2011, 16 pages.
English Translation of Decision of Rejection, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880104982.1, dated Sep. 23, 2011, 10 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 11006934.1-2221, dated Oct. 25, 2011, 5 pages.
English Translation of First Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 20080017883.X, dated Nov. 30, 2011, 16 pages.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 203176, dated Feb. 21, 2012, 2 pages.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 203177, dated Mar. 1, 2012, 2 pages.
English Translation of Second Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880101500.7, dated Apr. 5, 2012, 5 pages.
English Translation of Second Office Action issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880019166.0, dated Jun. 5, 2012, 8 pages.
English Translation of Second Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880104982.1, dated Jun. 29, 2012, 5 pages.
English Translation of Second Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880017883.X, dated Aug. 10, 2012, 9 pages.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 203176, dated Sep. 27, 2012, 1 page.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 203177, dated Sep. 27, 2012, 1 page.
Office Action, issued by the Japanese Patent Office in connection with Patent Application No. 2010-501190, dated Oct. 2, 2012, 10 pages (includes English translation).
English Translation of Office Action, issued by the Japanese Patent Office in connection with Patent Application No. 2010-506646, dated Oct. 23, 2012, 3 pages.
English Translation of Third Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880101500.7, dated Nov. 21, 2012, 5 pages.
English Translation of Office Action, issued by the Israeli Patent Office in connection with Patent Application No. 201187, dated Nov. 27, 2012, 2 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 08796890.5-2319/2170161, dated Dec. 7, 2012, 9 pages.
Notification to Grant Patent Right for Invention, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880104982.1, dated Jan. 14, 2013, 4 pages (includes English translation).
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 08770372.4-1265/2152155, dated Feb. 6, 2013, 7 pages.
English Translation of Third Office Action, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880017883.X, dated Mar. 18, 2013, 8 pages.
English translation of Notification to Grant Patent Right for Invention, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880101500.7, dated Apr. 3, 2013, 2 pages.
Aaker et al., “Warmth in Advertising: Measurement, Impact, and Sequence Effects,” Journal of Consumer Research, vol. 12, No. 4, pp. 365-381, (Mar. 1986), 18 pages.
Allen et al., “A Method for Removing Imaging Artifact from Continuous EEG Recorded during Functional MRI,” Neuroimage, vol. 12, 230-239, (Aug. 2000), 12 pages.
Ambler, “Salience and Choice: Neural Correlates of Shopping Decisions,” Psychology & Marketing, vol. 21, No. 4, p. 247-261, Wiley Periodicals, Inc., doi: 10.1002/mar20004, (Apr. 2004), 16 pages.
Ambler et al., “Ads on the Brain: A Neuro-Imaging Comparison of Cognitive and Affective Advertising Stimuli,” London Business School, Centre for Marketing Working Paper, No. 00-902, (Mar. 2000), 23 pages.
Bagozzi et al., “The Role of Emotions in Marketing,” Journal of the Academy of Marketing Science, vol. 27, No. 2, pp. 184-206, Academy of Marketing Science (1999), 23 pages.
Barcelo, et al., “Prefrontal modulation of visual processing in humans,” Nature Neuroscience, vol. 3, No. 4, Nature America, http//neurosci.nature.com, (Apr. 2000), 5 pages.
Barreto et al., “Physiologic Instrumentation for Real-time Monitoring of Affective State of Computer Users,” WSEAS International Conference on Instrumentation, Measurement, Control, Circuits and Systems (IMCCAS), (2004), 6 pages.
Belch et al., “Psychophysiological and cognitive Responses to Sex in Advertising,” Advances in Consumer Research, vol. 9, pp. 424-427, (1982), 6 pages.
Bimler et al., “Categorical perception of facial expressions of emotion: Evidence from multidimensional scaling,” Cognition and Emotion, vol. 15(5), pp. 633-658 (Sep. 2001), 26 pages.
Blakeslee, “If You Have a ‘Buy Button’ in Your Brain, What Pushes It?” The New York Times, www.nytimes.com, (Oct. 19, 2004), 3 pages.
Braeutigam, “Neuroeconomics—From neural systems to economic behavior,” Brain Research Bulletin, vol. 67, pp. 355-360, Elsevier, (2005), 6 pages.
Buschman, et al., “Top-Down versus Bottom-Up Control of Attention in the Prefrontal and posterior Parietal Cortices,” Science, vol. 315, www.sciencemag.org/cgi/content/full/315/5820/1860, American Association for the Advancement of Science, (2007), 4 pages.
Buschman, et al., “Serial, Covert Shifts of Attention during Visual Search Are Reflected by the Frontal Eye Fields and Correlated with Population Oscillations,” Neuron, vol. 63, pp. 386-396, Elsevier, (Aug. 13, 2009), 11 pages.
Canolty, et al., “High Gamma Power Is Phase-Locked to Theta Oscillations in Human Neocortex,” Science, vol. 313, www.sciencemag.org, (Sep. 15, 2006), 3 pages.
Cheng, et al. “Gender Differences in the Mu Rhythm of the Human Mirror-Neuron System,” PLos ONE, vol. 3, Issue 5, www.plosone.org, (May 2008), 7 pages.
Clemons, “Resonance Marketing in the Age of the Truly Informed Consumer: Creating Profits through Differentiation and Delight,” Wharton Information Strategy & Economics Blog 2, available at http://opim.wharton.npenn.edu/.abont.clemons/blogs/resonanceblog.pdf, (Mar. 28, 2007), 8 pages.
Clifford, “Billboards That Look Back,” The New York Times, NYTimes.com, available at http://www.nytimes.com/2008/05/31/business/media/31billboard.html, (May 31, 2008), 4 pages.
Crawford et al., “Self-generated happy and sad emotions in low and highly hypnotizable persons during waking and hypnosis: laterality and regional EEG activity differences,” International Journal of Psychophysiology, vol. 24, pp. 239-266, (Dec. 1996), 28 pages.
Davidson, et al., “The functional neuroanatomy of emotion and affective style,” TRENDS in Cognitive Sciences, vol. 3, No. 1, (Jan. 1999), 11 pages.
De Gelder et al., “Categorical Perception of Facial Expressions: Categories and their Internal Structure,” Cognition and Emotion, vol. 11(1), pp. 1-23 (1997), 23 pages.
D'Esposito, “From cognitive to neural models of working memory,” Phil. Trans. R. Soc. B, doi: 10.1098/rstb.2007.2086, (Mar. 30, 2007), 12 pages.
Desmet, “Measuring Emotions: Development and Application of an Instrument to Measure Emotional Responses to Products,” to be published in Funology: From Usability to Enjoyment, pp. 1-13, Kluwer Academic Publishers, (Blythe et al., eds., 2004), 13 pages.
Dien, et al., “Application of Repeated Measures ANOVA to High-Density ERP Datasets: A Review and Tutorial,” in Event-Related Potentials: A Methods Handbook pp. 57-82, (Todd C. Handy, ed., 2005), 14 pages.
Edgar, et al., “Digital Filters in ERP Research,” in Event-Related Potentials: A Methods Handbook pp. 85-113, (Todd C. Handy, ed., 2005), 15 pages.
EEG Protocols, “Protocols for EEG Recording,” retrieved from the Internet on Aug. 23, 2011, http://www.q-metrx.com/EEGrecordingProtocols.pdf, (Nov. 13, 2007), 3 pages.
Engel et al., “Dynamic Predictions: Oscillations and Synchrony in Top-down Processing,” Nature Reviews: Neuroscience, vol. 2, pp. 704-716, Macmillian Magazines Ltd., (Oct. 2001), 13 pages.
Filler, “MR Neurography and Diffusion Tensor Imaging. Origins, History & Clinical Impact of the first 50,000 Cases With an Assessment of Efficacy and Utility in a Prospective 5,000 Patent Study Group,” Institute for Nerve Medicine, (Nov. 7, 2008), 56 pages.
Fogelson, et al., “Prefrontal cortex is critical for contextual processing: evidence from brain lesions,” Brain: A Journal of Neurology, vol. 132, pp. 3002-3010, doi:10.1093/brain/awp230, (Aug. 27, 2009), 9 pages.
Friedman, et al., “Event-Related Potential (ERP) Studies of Memory Encoding and Retrieval: A Selective Review,” Microscopy Research and Technique 51:6-22, Wiley-Liss, Inc. (2000), 23 pages.
Fries, “A mechanism for cognitive dynamics: neuronal communication through neuronal coherence,” Trends in Cognitive Sciences, vol. 9, No. 10, pp. 474-480, Elsevier B.V. www.sciencedirect.com, (Oct. 2005), 7 pages.
Gaillard, “Problems and Paradigms in ERP Research,” Biological Psychology, Elsevier Science Publisher B.V. (1988), 10 pages.
Gargiulo et al., “A Mobile EEG System With Dry Electrodes,” (Nov. 2008), 4 pages.
Gazzaley et al., “Top-down Enhancement and Suppression of Magnitude and Speed of Neural Activity,” Journal of Cognitive Neuroscience, vol. 17, No. 3, pp. 507-517, Massachusetts Institute of Technology, (2005), 11 pages.
Griss et al., “Characterization of micromachined spiked biopotential electrodes”, Biomedical Engineering, IEEE Transactions (Jun. 2002), 8 pages.
Haq, “This Is Your Brain on Advertising,” BusinessWeek, Market Research, (Oct. 8, 2007), 4 pages.
Hartikainen et al., Manuscript Draft of “Emotionally arousing stimuli compete with attention to left hemispace,” NeuroReport, (Sep. 8, 2007), 26 pages.
Hazlett, et al., “Emotional Response to Television Commercials: Facial EMG vs. Self-Report,” Journal of Advertising Research, (Apr. 1999), 17 pages.
Herrmann, et al., “Mechanisms of human attention: event-related potentials and oscillations,” Neuroscience and Biobehavioral Reviews, pp. 465-476, Elsevier Science Ltd., www.elsvevier.com/locate/neubiorev, (2001), 12 pages.
Hopf, et al., “Neural Sources of Focused Attention in Visual Search,” Cerebral Cortex, 10:1233-1241, Oxford University Press, (Dec. 2000), 9 pages.
Jung et al., “Analysis and Visualization of Single-Trial Event-Related Potentials,” Human Brain Mapping vol. 14, 166-185 (2001), 20 pages.
Kay et al., “Identifying natural images from human brain activity,” Nature, vol. 452, pp. 352-356, Nature Publishing Group, (Mar. 20, 2008), 5 pages.
Keren, et al., “Saccadic spike potentials in gamma-band EEG: Characterization, detection and suppression,” NeuroImage, http://dx.doi:10.1016/j.neuroimage.2009.10.057, (Oct. 2009), 16 pages.
Kishiyama, et al., “Novelty Enhancements in Memory Are Dependent on Lateral Prefrontal Cortex,” The Journal of Neuroscience, pp. 8114-8118, Society for Neuroscience (Jun. 24, 2009), 5 pages.
Kishiyama, et al., “Socioeconomic Disparities Affect Prefrontal Function in Children,” Journal of Cognitive Neuroscience pp. 1106-1115, Massachusetts Institute of Technology, (2008), 10 pages.
Klimesch, “EEG alpha and theta oscillations reflect cognitive and memory performance a review and analysis,” Brain Research Reviews, vol. 29, 169-195, (1999), 27 pages.
Knight, “Contribution of human hippocampal region to novelty detection,” Nature, vol. 383, www.nature.com, (Sep. 19, 1996), 4 pages.
Knight, “Consciousness Unchained: Ethical Issues and the Vegetative and minimally Conscious State,” The American Journal of Bioethics, 8:9, 1-2, http://dx.doi.org/10.1080/15265160802414524, (Sep. 1, 2008), 3 pages.
Knight, et al., “Prefrontal cortex regulates inhibition and excitation in distributed neural networks,” Acta Psychologica vol. 101, pp. 159-178, Elsevier (1999), 20 pages.
Knight, “Decreased Response to Novel Stimuli after Prefrontal Lesions in Man,” Electroencephalography and Clinical Neurophysiology, vol. 59, pp. 9-20, Elsevier Scientific Publishers Ireland, Ltd., (1984), 12 pages.
Krakow et al., “Methodology: EEG-correlated fMRI,” Functional Imaging in the Epilepsies, (Lippincott Williams & Wilkins, 2000), 17 pages.
Krugman, “Brain Wave Measures of Media Involvement,” Journal of Advertising Research vol. 11, 3-9 (Feb. 1971), 7 pages.
Lachaux et al., “Measuring Phase Synchrony in Brain Signals,” Human Brain Mapping 8 (1999), 194-208, 15 pages.
Lee et al., “What is ‘neuromarketing’? a discussion and agenda for future research,” International Journal of Psychophysiology, vol. 63, pp. 199-204, Elsevier (2006), 6 pages.
Lekakos, “Personalized Advertising Services Through Hybrid Recommendation Methods: The Case of Digital Interactive Television,” Department of Informatics, Cyprus University, (2004), 11 pages.
Lewis et al., “Market Researchers make Increasing use of Brain Imaging,” ACNR, vol. 5, No. 3, pp. 36-37, (Jul./Aug. 2005), 2 pages.
Luck, et al., “The speed of visual attention in schizophrenia: Electrophysiological and behavioral evidence,” Schizophrenia Research, pp. 174-195, Elsevier B.V. www.sciencedirect.com, (2006), 22 pages.
Lui et al., “Marketing Strategies in Virtual Worlds,” The Data Base for Advances in Information Systems, vol. 38, No. 4, pp. 77-80, (Nov. 2007), 4 pages.
Makeig, et al., “Mining event-related brain dynamics,” TRENDS in Cognitive Sciences, vol. 8, No. 5, (May 2004), www.sciencedirect.com, 7 pages.
Makeig, et al., “Dynamic Brain Sources of Visual Evoked Responses,” Science, vol. 295, www.sciencemag.org, (Jan. 25, 2002), 5 pages.
The Mathworks, Inc., “MATLAB Data Analysis: Version 7,” p. 4-19 (2005), 3 pages.
Merriam-Webster Online Dictionary definition for “tangible,” available at http://www.merriam-webster.com/dictionary/tangible, 1 page.
Merriam Webster Online Dictionary, Definition of Virtual Reality, available at http://www.merriam-webster.com/dictionary/virtual%20reality, 2 page.
Miltner, et al., “Coherence of gamma-band EEG activity as a basis for associative learning,” Nature, vol. 397, www.nature.com, (Feb. 4, 1999), 3 pages.
Mosby's Dictionary of Medicine, Nursing, & Health Professions, 2009, Mosby, Inc., Definition of Alpha Wave, 1 page.
Mosby's Dictionary of Medicine, Nursing, & Health Professions, 2009, Mosby, Inc., Definition of Beta Wave, 1 page.
Neurofocus—Neuroscientific Analysis for Audience Engagement, accessed on Jan. 8, 2010 at http://web.archive.org/web/20080621114525/www.neurofocus.com /BrandImage.htm, (2008), 2 pages.
Newell et al., “Categorical perception of familiar objects,” Cognition, vol. 85, Issue 2, pp. 113-143 (Sep. 2002), 31 pages.
Nielsen, “Neuroinformatics in Functional Neuroimaging,” Informatics and Mathematical Modeling, Technical University of Denmark, (Aug. 30, 2002), 241 pages.
Oberman et al., “EEG evidence for mirror neuron Toward an analysis of the human qualities of activity during the observation of human and robot actions: interactive robots,” Neurocomputing 70 (2007) 2194-2203, 10 pages.
Osborne, “Embedded Watermarking for image Verification in Telemedicine,” Thesis submitted for the degree of Doctor of Philosophy, Electrical and Electronic Engineering, University of Adelaide (2005), 219 pages.
Padgett et al., “Categorical Perception in Facial Emotion Classification,” In Proceedings of the 18th Annual Conference of the Cognitive Science Society, pp. 249-253 (1996), 5 pages.
Page et al., “Cognitive Neuroscience, Marketing and Research,” Congress 2006—Foresight—The Predictive Power of Research Conference Papers, ESOMAR Publications, (Sep. 17, 2006), 25 pages.
Paller, et al., “Validating neural correlates of familiarity,” TRENDS in Cognitive Sciences, vol. 11, No. 6, www.sciencedirect.com, (May 2, 2007), 8 pages.
Palva et al., “Phase Synchrony Among Neuronal Oscillations in the Human Cortex,” Journal of Neuroscience 25 (2005), 3962-3972, 11 pages.
Picton, et al., “Guidelines for using human event-related potentials to study cognition: Recording standards and publication criteria,” Psychophysiology, pp. 127-152, Society for Psychophysiological Research, (2000), 26 pages.
Rizzolatti et al., “The Mirror-Neuron System,” Annu. Rev. Neurosci., vol. 27, pp. 169-192, (Mar. 5, 2004), 30 pages.
Ruchkin et al., “Modality-specific processing streams in verbal working memory: evidence from spatio-temporal patterns of brain activity,” Cognitive Brain Research, vol. 6, pp. 95-113, Elsevier, (1997), 19 pages.
Rugg, et al., “Event-related potentials and recognition memory,” TRENDS in Cognitive Sciences, vol. 11, No. 6, www.sciencedirect.com, (May 3, 2007), 7 pages.
Rugg, et al., “The ERP and cognitive psychology: conceptual issues,” (Sep. 1996), 7 pages.
“User monitoring,” Sapien Systems, available at http://web.archive.org/web/20030818043339/http:/www.sapiensystems.com/eye-tracking.html, (Aug. 18, 2003), 1 page.
Simon-Thomas, et al, “Behavioral and Electrophysiological Evidence of a Right Hemisphere Bias for the Influence of Negative Emotion on Higher Cognition,” Journal of Cognitive Neuroscience, pp. 518-529, Massachusetts Institute of Technology (2005), 12 pages.
Spencer, “Averaging, Detection, and Classification of Single-Trial ERPs,” in Event-Related Potentials: A Methods Handbook, pp. 209-227, (Todd C. Handy, ed., 2005), 10 pages.
Arousal in Sport, in Encyclopedia of Applied Psychology, vol. 1, p. 159, retrieved from Google Books, (Spielberger, ed., Elsevier Academic Press, 2004), 1 page.
Srinivasan, “High-Resolution EEG: Theory and Practice,” in Event-Related Potentials: A Methods Handbook, pp. 167-188, (Todd C. Handy, ed., 2005), 12 pages.
Summons to Attend Oral Proceedings Pursuant to Rule 115(1) EPC, issued by the European Patent Office in connection with European Patent Application No. 08744383.4, dated Dec. 11, 2018, 6 pages.
Sutherland, “Neuromarketing: What's it all about?” Retrieved from Max Sutherland's Weblog on Aug. 23, 2011, http://www.sutherlandsurvey.com/Column_pages/Neuromarketing_whats_it_all_-about.htm, (Mar. 2007), 5 pages.
Swick, et al., “Contributions of Prefrontal Cortex to Recognition Memory: Electrophysiological and Behavioral Evidence,” Neuropsychology, vol. 13, No. 2 pp. 155-170, American Psychological Association, Inc. (1999), 16 pages.
Taheri, et al., “A dry electrode for EEG recording,” Electroencephalography and clinical Neurophysiology, pp. 376-383, Elsevier Science Ireland Ltd. (1994), 8 pages.
Talsma, et al., “Methods for the Estimation and Removal of Artifacts and Overlap in ERP Waveforms,” in Event-Related Potentials: A Methods Handbook, pp. 115-148, (Todd C. Handy, ed., 2005), 22 pages.
Vogel, et al., “Electrophysiological Evidence for a Postperceptual Locus of Suppression During the Attentional Blink,” Journal of Experimental Psychology: Human Perception and Performance, vol. 24, No. 6, pp. 1656-1674, (1998), 19 pages.
“Functional magnetic resonance imaging,” retrieved online from Wikipedia, the Free Encyclopedia on Aug. 23, 2011, at http://en.wikipedia.org/w/index.php?title=Functional_magnetic_resonance_i-maging&oldid=319601772, (Oct. 13, 2009), 8 pages.
Woldorf, “Distortion of ERP averages due to overlap from temporally adjacent ERPs: Analysis and correction,” Psychophysiology, Society for Psychophysiological Research, Cambridge University Press (1993), 22 pages.
Woodman, et al., “Serial Deployment of Attention During Visual Search,” Journal of Experimental Psychology: Human Perception and Performance, vol. 29, No. 1, pp. 121-138, American Physiological Association (2003), 18 pages.
Yamaguchi, et al., “Rapid-Prefrontal-Hippocampal Habituation to Novel Events,” The Journal of Neuroscience, pp. 5356-5363, Society for Neuroscience, (Apr. 29, 2004), 8 pages.
Yap et al., “TIMER: Tensor Image Morphing for Elastic Registration,” NeuroImage, vol. 47, (May 3, 2009), 15 pages.
Yuval-Greenberg, et al., “Transient Induced Gamma-Bands Response in EEG as a Manifestation of Miniature Saccades,” Neuron, vol. 58, pp. 429-441, Elsevier Inc. (May 8, 2008), 13 pages.
Ziegenfuss, “Neuromarketing: Advertising Ethical & Medical Technology,” The Brownstone Journal, vol. XII, Boston University, pp. 69-73, (May 2005), 9 pages.
Zyga, “A Baseball Cap That Can Read Your Mind,” PhysOrg.com, located at www.physorg.com/news130152277.html, (May 16, 2008), 11 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Apr. 25, 2013, 34 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/444,149, dated May 2, 2013, 27 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated May 8, 2013, 4 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated May 8, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated May 8, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated May 8, 2013, 7 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/569,711, dated May 14, 2013, 6 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated May 17, 2013, 6 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/778,828, dated May 23, 2013, 25 pages.
Office Communication to Applicant, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated May 24, 2013, 2 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated May 28, 2013, 12 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated May 31, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,211, dated Jun. 3, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Jun. 3, 2013, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Jun. 11, 2013, 7 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/545,455, dated Jun. 11, 2013, 9 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/056,190, dated Jun. 13, 2013, 5 pages.
Office Communication to Applicant, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,388, dated Jun. 13, 2013, 2 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/135,074, dated Jun. 21, 2013, 5 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/105,774, dated Jun. 26, 2013, 10 pages.
English Translation of Office Action, issued by the Israeli Patent Office in connection with Patent Application No. 203176, dated Apr. 23, 2013, 1 page.
English Translation of Notice Prior to Allowance, issued by the Israeli Patent Office in connection with Patent Application No. 203176, dated Jun. 30, 2013, 1 page.
Merriam-Webster Online Dictionary, Definition for “Resonance,” available at http://www.merriam-webster.com/dictionary/resonance, 4 pages.
Enghoff, Sigurd, Thesis: “Moving ICA and Time-Frequency Analysis in Event-Related EEG Studies of Selective Attention,” Technical University of Denmark, (Dec. 1999), 54 pages.
Zhang, P., “Will You Use Animation on Your Web Pages?” Doing Business on the Internet: Opportunities and Pitfalls, C. Romm and F. Sudweeks (eds.), Spring-Verlag (1999), 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/182,874, dated Jul. 29, 2013, 6 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Sep. 12, 2013, 13 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Sep. 13, 2013, 7 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Sep. 17, 2013, 11 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/778,828, dated Oct. 8, 2013, 11 pages.
English Translation of Office Action, issued by the Japanese Patent Office in connection with Patent Application No. 2010-523112, dated Jul. 30, 2013, 2 pages.
Decision to Grant Patent, issued by the Japanese Patent Office in connection with Patent Application No. 2010-506646, dated Aug. 6, 2013, 4 pages (includes English translation).
English Translation of Decision on Rejection, issued by the Chinese Patent Office in connection with Patent Application No. 200880017883.X, dated Aug. 5, 2013, 13 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Oct. 23, 2013, 17 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Nov. 6, 2013, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/444,149, dated Nov. 19, 2013, 11 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Dec. 3, 2013, 16 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/199,596, dated Dec. 23, 2013, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/105,774, dated Jan. 16, 2014, 11 pages.
English Translation of Office Action, issued by the Japanese Patent Office in connection with Patent Application No. 2010-520159, dated Oct. 1, 2013, 2 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 08798799.6-1657/2180825, dated Nov. 4, 2013, 9 pages.
Coan et al., “Voluntary Facial Expression and Hemispheric Asymmetry Over the Frontal Cortex,” Psycophysiology (Nov. 2001), 912-924, 14 pages.
Duchowski, “A Breadth-First Survey of Eye-tracking Applications,” Behavior Research Methods, Instruments, and Computers (Nov. 2002), 455-470, 16 pages.
Heo et al., “Wait! Why is it Not Moving? Attractive and Distractive Ocular Responses to Web Ads,” Paper presented to AEJMC, (Aug. 2001) Washington, DC, available at http://www.psu.edu/dept/medialab/researchpage/newabstracts/wait.html, 3 pages.
Rothschild et al., “Predicting Memory for Components of TV Commercials from EEG,” Journal of Consumer Research (Mar. 1990), p. 472-478, 8 pages.
Beaver, John D., et al., “Individual Differences in Reward Drive Predict Neural Responses to Images of Food”, J. of Neuroscience, (May 10, 2006), 5160-5166, 7 pages.
Tapert, Susan F., et al., “Neural Response to Alcohol Stimuli in Adolescents With Alcohol Use Disorder”, Arch Gen Psychiatry (Jul. 2003), 727-735, 9 pages.
Shandlen, Michael N. et al., “A Computational Analysis of the Relationship between Neuronal and Behavioral Responses to Visual Motion”, The Journal of Neuroscience, (Feb. 15, 1996) 1486-1510, 25 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Jan. 30, 2014, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Jan. 31, 2014, 5 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Feb. 3, 2014, 15 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/444,149, dated Feb. 3, 2014, 5 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Feb. 4, 2014, 12 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Feb. 6, 2014, 17 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated Feb. 10, 2014, 14 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Feb. 10, 2014, 18 pages.
Mehta, A. et al., “Reconsidering Recall and Emotion in Advertising”, Journal of Advertising Research, (Mar. 2006), 49-56, 9 pages.
Cheung, Kwok-Wai, et al., “Mining Customer Product Ratings for Personalized Marketing,” Decision Support Systems 35 (2003) 231-243, 13 pages.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 201187, dated Apr. 2, 2013, 2 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated May 27, 2014, 8 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Jun. 2, 2014, 13 pages.
Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Jun. 5, 2014, 25 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Jul. 1, 2014, 16 pages.
Decision to Grant Patent, issued by the Japanese Patent Office in connection with Patent Application No. 2010-523112, dated Apr. 8, 2014, 4 pages (includes English translation).
Darrow, Chester, “Psychological and psychophysiological significance of the electroencephalogram,” Psychological Review (May 1947) 157-168, 12 pages.
Stamm, John, “On the Relationship between Reaction Time to Light and Latency of Blocking the Alpha Rhythm,” Electroencephalography and Clinical Neurophysiology (Feb. 1952), 61-68, 8 pages.
Mizuki, Yashushi, et al., “Periodic Appearance of the Theta Rhythm in the Frontal Midline Area During Performance of a Mental Task,:” Electroencephalography and Clinical Neurophysiology (Aug. 1980), 345-351, 7 pages.
Decision to Grant Patent, issued by the Korean Patent Office in connection with Patent Application No. 10-2009-7022551, dated Aug. 13, 2014, 3 pages (includes English translation).
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Jul. 23, 2014, 13 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Aug. 6, 2014, 18 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Aug. 14, 2014, 4 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated Aug. 15, 2014, 15 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Aug. 21, 2014, 20 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Sep. 4, 2014, 16 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/105,774, dated Sep. 18, 2014, 14 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Sep. 29, 2014, 21 pages.
Ekman, P., Friesen, W., Measuring Facial Movement, Environmental Psychology and Nonverbal Behavior, 1 (1) (Fall 1976), pp. 56-75, 20 pages.
Ekman, P., Friesen, W.V., Facial Action Coding System: A Technique for Measurement of Facial Movement, Consulting Psychologists Press, Palo Alto, Calif (1978), (Book, copy not provided.).
Ekman, P., Friesen, W., Unmasking the Face—A Guide to Recognizing Emotions from Facial Clues, Prentice-Hall, Inc., Englewood Cliffs, N.J. (1979), (Book, copy not provided.).
Ekman, P., Friesen, W., Ancoli, S., Facial Signs of Emotional Experience, J. Personality & Social Psychology, 39(6) (Dec. 1980), pp. 1125-1134, 10 pages.
Izard, C. E., The Maximally Discriminative Facial Movement Coding System, (Rev. ed.), Instructional Resources Center, University of Delaware, Newark, Del. (1983). (Book, copy not provided.).
Izard, C., Dougherty, L., Hembree, E., A System for Identifying Affect Expressions by Holistic Judgments (AFFEX), Instructional Resources Center, University of Delaware, Newark, Del. (1983). (Book, copy not provided).
Jia, X., Nixon, M.S., Extending the Feature Set for Automatic Face Recognition, International Conference on Image Processing and Its Applications (Apr. 7-9, 1992), 6 pages.
Lisetti, C., Nasoz, F., Using Noninvasive Wearable Computers to Recognize Human Emotions from Physiological Signals, EURASIP J. Applied Signal Processing, 11 (Sep. 2004), pp. 1672-1687, 16 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 14/177,698, dated Oct. 24, 2014, 13 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Dec. 22, 2014, 3 pages.
English Translation of Notification of Provisional Rejection, issued by the Korean Patent Office in connection with Patent Application No. 10-2010-7001406, dated Oct. 21, 2014, 1 page.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Feb. 20, 2015, 12 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Feb. 20, 2015, 52 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Mar. 6, 2015, 18 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,344, dated Apr. 9, 2015, 12 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 14/177,698, dated Apr. 24, 2015, 13 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated May 5, 2015, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/105,774, dated May 14, 2015, 15 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated May 14, 2015, 22 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated May 22, 2015, 6 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/249,512, dated Jun. 30, 2015, 36 pages.
Notification of Provisional Rejection, issued by the Korean Patent Office in connection with Patent Application No. 10-2010-7001406, dated Jun. 24, 2015, 9 pages (includes partial translation).
McClure, Samuel, et al., “Neural Correlates of Behavioral Preference for Culturally Familiar Drinks,” Neuron (Oct. 14, 2004), 379-387, 9 pages.
English Translation of Office Action, issued by the Israel Patent Office in connection with Patent Application No. 201187, dated Jun. 22, 2015, 4 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Jul. 30, 2015, 14 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Aug. 4, 2015, 29 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 14/177,698, dated Aug. 19, 2015, 12 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Sep. 11, 2015, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Sep. 2, 2015, 6 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Sep. 10, 2015, 15 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Sep. 16, 2015, 3 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/705,525, dated Sep. 30, 2015, 12 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Sep. 30, 2015, 6 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,344, dated Nov. 20, 2015, 28 pages.
Translation of Reexamination Decision, issued by the Chinese Patent Office in connection with Patent Application No. 200880017883.X, dated Nov. 13, 2015, 1 page.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Dec. 17, 2015, 14 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Dec. 18, 2015, 7 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 14/177,698, dated Jan. 14, 2016, 36 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Jan. 22, 2016, 38 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Feb. 3, 2016, 22 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Feb. 23, 2016, 24 pages.
English Translation of Notification of Provisional Rejection, issued by the Korean Patent Office in connection with Patent Application No. 10-2010-7001406, dated Jan. 26, 2016, 1 page.
English Translation of Notification to Grant Patent Right for Invention, issued by the State Intellectual Property Office of P.R. China in connection with Patent Application No. 200880017883.X, dated Feb. 3, 2016, 2 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Mar. 22, 2016, 27 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Mar. 30, 2016, 23 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Apr. 6, 2016, 3 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated Apr. 8, 2016, 7 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Apr. 21, 2016, 33 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated May 12, 2016, 61 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated May 20, 2016, 69 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/945,357, dated May 20, 2016, 22 pages.
M. Corbetta et al., “Control of Goal-Directed and Stimulus-Driven Attention in the Brain,” Nature Reviews Neuroscience, vol. 3, pp. 201-215 (Mar. 2002), 15 pages.
Becker, “A Study of Web Usability for Older Adults Seeking Online Health Resources,” ACM Transactions on Computer-Human Interaction, vol. 11, No. 4, pp. 387-406 (Dec. 2004), 20 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Jun. 17, 2016, 20 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Jul. 29, 2016, 67 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Jul. 27, 2016, 20 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Aug. 8, 2016, 3 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Aug. 16, 2016, 5 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Aug. 25, 2016, 61 pages.
Notification of Provisional Rejection, issued by the Korean Patent Office in connection with Patent Application No. 10-2010-7001406, dated Jul. 27, 2016, 4 pages (includes partial translation).
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/945,357, dated Nov. 1, 2016, 22 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Nov. 7, 2016, 3 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Nov. 14, 2016, 18 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Nov. 14, 2016, 10 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Nov. 29, 2016, 27 pages.
First Examination Report, issued by the European Patent Office in connection with European Application No. 08796890.5, dated Sep. 29, 2016, 4 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Dec. 15, 2016, 31 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,344, dated Jan. 26, 2017, 52 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Jan. 31, 2017, 25 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Feb. 9, 2017, 7 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Mar. 2, 2017, 14 pages.
English Translation of Notice Prior to Allowance, issued by the Israeli Patent Office in connection with Patent Application No. 201187, dated Feb. 14, 2017, 1 page.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/122,253, dated Mar. 31, 2017, 37 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Apr. 27, 2017, 45 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated May 25, 2017, 9 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Jun. 5, 2017, 39 pages.
European Patent Office, “Communication Pursuant to Article 94(3) EPC,” issued in connection with European Patent Application No. 08 744 383.4, dated Apr. 19, 2017, 6 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,344, dated Jun. 29, 2017, 38 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/945,357, dated Jul. 6, 2017, 17 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Jul. 13, 2017, 12 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Aug. 14, 2017, 38 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Aug. 18, 2017, 2017, 11 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Sep. 19, 2017, 43 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/113,870, dated Sep. 26, 2017, 51 pages.
Summons to Attend Oral Proceedings Pursuant to Rule 115(1) EPC, issued by the European Patent Office in connection with European Patent Application No. 08796890.5, dated Jul. 3, 2017, 3 pages.
English Translation of First Examination Report, issued by the Indian Patent Office in connection with Indian Patent Application No. 6145/CHENP/2009, dated Aug. 16, 2017, 6 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,344, dated Oct. 2, 2017, 5 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/945,357, dated Oct. 20, 2017, 16 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Oct. 26, 2017, 4 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Oct. 31, 2017, 2017, 68 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 15/299,752, dated Nov. 3, 2017, 131 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated Nov. 15, 2017, 49 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Dec. 6, 2017, 14 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Dec. 28, 2017, 23 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Jan. 29, 2018, 11 pages.
Ganel et al., “The Relationship Between fMRI Adapation and Repetition Priming,” NeuroImage, Jul. 18, 2006, pp. 1434-1440, 9 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated Jan. 30, 2014, 67 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated Aug. 13, 2014, 8 pp.. (Copy not provided as this is a Uspto document. Applicant will provide document upon request from Examiner.).
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated May 6, 2015, 15 pages.
Examiner's Answer, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated Feb. 18, 2016, 5 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated Mar. 1, 2018, 14 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,511, dated May 29, 2014, 8 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Feb. 12, 2015, 6 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Mar. 30, 2015, 6 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Jun. 15, 2015, 5 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Sep. 23, 2015, 16 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Sep. 22, 2015, 6 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297 , dated Nov. 27, 2015, 5 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated Feb. 2, 2018, 10 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,550, dated Mar. 27, 2018, 13 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 15/299,752, dated Apr. 17, 2017, 21 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/884,034, dated May 3, 2018, 46 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated May 18, 2018, 31 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,541, dated Jun. 27, 2018, 62 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/730,564, dated Jun. 27, 2018, 69 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Jun. 28, 2018, 22 pages.
Decision on Request for Rehearing, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,685, dated Aug. 10, 2018, 8 pages.
Extended European Search Report, issued by the European Patent Office in connection with European Application No. 08747389.8-1952, dated Sep. 25, 2015, 6 pages.
English Translation of First Examination Report, issued by the Indian Patent Office in connection with Indian Patent Application No. 4438/KOLNP/2009, dated Sep. 25, 2017, 7 pages.
Communication Under Rule 71(3) EPC, issued by the European Patent Office in connection with European Application No. 08796890.5, dated Mar. 16, 2018, 43 pages.
English Translation of First Examination Report, issued by the Indian Patent Office in connection with Indian Patent Application No. 4441/KOLNP/2009, dated May 21, 2018, 5 pages.
Knutson et al., “Neural Predictors of Purchases,” Neuron vol. 53 (Jan. 4, 2007), pp. 147-156, 10 pages.
Schaefer et al., “Neural Correlates of Culturally Familiar Brands of Car Manufacturers,” NeuroImage, vol. 31 (2006), pp. 861-865, 5 pages.
Aharon et al., “Beautiful Faces Have Variable Reward Value: fMRI and Behavioral Evidence,” Neuron, vol. 32 (2001), pp. 537-551, 15 pages.
Hall, Bmce F., “A New Model for Measuring Advertising Effectiveness,” Journal of Advertising Research, Mar.-Apr. 2002, 10 pages.
Kamba et al., “The Krakatoa Chronicle—An Interactive, Personalized, Newspaper on the Web,” available at: http://www.w3.org/Conferences/WWW4/Papers/93/ (last accessed Nov. 2, 2015), 15 pages.
Ehrenberg et al., “Understanding Brand Performance Measures: Using Dirichlet Benchmarks,” 2004, Journal of Business Research, vol. 57, pp. 1307-1325, 19 pages.
Leeflang et al., “Building Models for Marketing Decisions,” 2000, Springer Science + Business Media, pp. 192-235, 482-521, 86 pages.
Bhattacharya, “Is your brand's loyalty too much, too little, or just right?: Explaining deviations in loyalty from the Dirichlet norm,” 1997, International Journal of Research in Marketing, vol. 14, pp. 421-435, 15 pages.
McCraty et al., “Impact of a Workplace Stress Reduction Program on Blood Pressure and Emotional Health in Hypertensive Employees”, the Journal of Alternative and Complementary Medicine, vol. 9, No. 3, 2003, pp. 355-369, Mary Ann Liebert, Inc., 15 pages.
Nikolaeva et al., “The Moderating Role of Consumer and Product Characteristics on the Value of Customized On-Line Recommendations,” 2006, International Journal of Electronic Commerce, vol. 11, No. 2, pp. 101-123, 24 pages.
Ehrenberg, “New Brands and the Existing Market,” 1991, International Journal of Market Research, vol. 33, No. 4, 10 pages.
Foxall, “The Substitutability of Brands,” 1999, Managerial and Decision Economics, vol. 20, pp. 241-257, 17 pages.
Pammer et al., “Forecasting the Penetration of a New Product—A Bayesian Approach,” 2000, Journal of Business and Economic Statistics, vol. 18, No. 4, pp. 428-435, 8 pages.
Rungie et al., “Calculation of Theoretical Brand Performance Measures from the Parameters of the Dirichlet Model,” 2004, Marketing Bulletin, Massey University, 15, Technical Note 2, pp. 1-19, 20 pages.
Uncles et al., “Patterns of Buyer Behavior: Regularities, Models, and Extensions,” 1995, Marketing Science, vol. 14, No. 3, pp. G71-G78, 9 pages.
Boltz, “The cognitive processing of film and musical soundtracks,” Haverford College, Haverford, Pennsylvania, 2004, 32 (7), 1194-1205, 12 pages.
Christie et al., “Autonomic specificity of discrete emotion and dimensions of affective space: a multivariate approach,” International Journal of Psychophysiology, 51 (2004) 143-153, 11 pages.
Coombes et al., “Emotion and movement: Activation of defensive circuitry alters the magnitude of a sustained muscle contraction,” University of Florida, USA, Neuroscience Letters 396 (2006) 192-196, 5 pages.
Cryer et al. “Pull the Plug on Stress,” Harvard Business Review, Jul. 2003, 8 pages.
Demaree et al., “Predicting facial valence to negative stimuli from resting RSA: Not a function of active emotion regulation,” Cognition and Emotion vol. 20, Issue 2, 2006, pp. 161-176, published on Sep. 9, 2010, http://www.tandfonline.com/doi/abs/10.1080/02699930500260427, 6 pages. (Abstract provided.).
Elkman et al., “Autonomic Nervous System Activity Distinguishes among Emotions,” Science, New Series, vol. 221, No. 4616. (Sep. 16, 1983), pp. 1208-1210, http://links.jstor.org/sici?sici=0036-8075%2819830916%293%3A221%3A4616%3C-1208%3AANSADA%3E2.0.CO%3B2-H, 5 pages.
Elton, “Measuring emotion at the symphony,” The Boston Globe, Apr. 5, 2006, http://www.psych.mcgill.ca/labs/levitin/media/measuring_emotion_bos- ton.html, 3 pages.
Goldberg, “Getting wired could help predict emotions,” The Boston Globe, Jun. 13, 2005, http://www.boston.com/yourlife/health/mental/articles/2005/06/13/getting_-wired_could_help_predict_emotions/?page=full, 4 pages.
Gomez et al., “Respiratory Responses Associated with Affective Processing of Film Stimuli,” Biological Psychology, vol. 68, Issue 3, Mar. 2005, pp. 223-235, 2 pages. (Abstract provided.).
Hall, “Is cognitive processing the right dimension,” World Advertising Research Center, Jan. 2003, 3 pages.
Hall, “On Measuring the Power of Communications,” Journal of Advertising Research, 44, pp. 1-11, doi:10.1017/S0021849904040139, (2004), 1 page. (Abstract provided.).
Hall, “Research and strategy: a fall from grace,” ADMAP, Issue 443, pp. 18-20, 2003, 1 page. (Abstract provided).
Hubert et al., “Autonomic, neuroendocrine, and subjective responses to emotion-inducing film stimuli,” Int J Psychophysiol, Aug. 1991, 2 pages. (Abstract provided.).
Levenson et al., “Emotion and Autonomic Nervous System Activity in the Minangkabau of West Sumatra,” Department of Psychology, University of California, Berkeley, Journal of Personality and Social Psychology, 1992, 2 pages. (Abstract provided.).
Marci et al., “The effect of emotional distance on psychophysiologic concordance and perceived empathy between patient and interviewer,” Applied Psychophysiology and Biofeedback, Jun. 2006, vol. 31, issue 2, 31:115-129, 8 pages. (Abstract provided.).
McCraty et al., “Analysis of twenty-four hour heart rate variability in patients with panic disorder,” Biological Psychology, vol. 56, Issue 2, Jun. 2001, pp. 131-150, 1 page. (Abstract provided.).
McCraty et al., “Electrophysiological Evidence of Intuition: Part 1. The Surprising Role of the Heart,” The Journal of Alternative and Complementary Medicine, vol. 10, No. 1, 2004, pp. 133-143, Mary Ann Liebert, Inc., 12 pages.
McCraty et al., “Electrophysiological Evidence of Intuition: Part 2. A System-Wide Process?,” The Journal of Alternative and Complementary Medicine, vol. 10, No. 2, 2004, pp. 325-336, Mary Ann Liebert, Inc., 12 pages.
McCraty et al., “The Effects of Different Types of Music on Mood, Tension, and Mental Clarity,” Original Research, Alternative Therapies, Jan. 1998, vol. 4., No. 1, pp. 75-84, 10 pages.
McCraty et al., “The Effects of Emotions on Short-Term Power Spectrum Analysis of Heart Rate Variability,” American Journal of Cardiology, vol. 76, No. 14, Nov. 15, 1995, pp. 1089-1093, 6 pages.
McCraty et al., “The Impact of a New Emotional Self-Management Program on Stress, Emotions, Heart Rate Variability, DHEA and Cortisol,” Integrative Physiological and Behavioral Science, Apr.-Jun. 1998, vol. 33, No. 2, 151-170, 20 pages.
McCraty et al., “The Impact of an Emotional Self-Management Skills Course on Psychosocial Functioning and Autonomic Recovery to Stress in Middle School Children,” Integrative Physiological and Behavioral Science, Oct.-Dec. 1999, vol. 34, No. 4, 246-268, 23 pages.
Melillo, “Inside the Consumer Mind; What Neuroscience Can Tell Us About Marketing,” Adweek, Public Citizen's Commercial Alert, Jan. 16, 2006, http://www.adweek.com/news/advertising/inside-consumer-mind-83549, 8 pages.
Miller et al., “Influence of Specific Emotional States on Autonomic Reactivity and Pulmonary Function in Asthmatic Children,” Journal of the American Academy of Child & Adolescent Psychiatry, vol. 36, Issue 5, May 1997, pp. 669-677, 3 pages. (Abstract provided).
Murphy et al., “The Heart Reinnervates After Transplantation,” Official Journal of the Society of Thoracic Surgeons and the Southern Thoracic Surgical Association, Jun. 2000, vol. 69, Issue 6, pp. 1769-1781, 13 pages.
Rosenberg, “Emotional R.O.I.,” The Hub, May/Jun. 2006,pp. 24-25, 2 pages.
Tiller et al., “Cardiac Coherence: A New, Noninvasive Measure of Autonomic Nervous System Order,” Alternative Therapies, Jan. 1996, vol. 2, No. 1, 14 pages.
Umetani et al. “Twenty-Four Hour Time Domain Heart Rate Variability and Heart Rate: Relations to Age and Gender Over Nine Decades,” J Am Coll Cardiol, Mar. 1, 1998, pp. 593-601, 9 pages.
Von Leupoldt et al., “Emotions in a Body Plethysmograph,” Journal of Psychophysiology (2004), 18, pp. 170-176, 1 page. (Abstract provided.).
Kallman, “Effect of Blank Time on Picture Recognition,” The American Journal of Psychology, vol. 97, No. 3 (Autumn, 1984), pp. 399-406, 4 pages. (Abstract provided.).
Larose, Data Mining Methods and Models, Department of Mathematical Sciences, Central Connecticut State University, www.dbeBooks.com—An Ebook Library,published by John Wiley & Sons, Inc., 2006, 340 pages. (Book, copy not provided.).
Han et al., Data Mining: Concepts and Techniques, 2.sup.nd Edition, Elsevier, 2006, 772 pages. (Book, copy not provided.).
Liu et al., Web Data Mining: Exploring Hyperlinks, Contents, and Usage Data, Springer Science & Business Media, 2007, 532 pages, (Book, copy not provided.).
Berry et al., Data Mining Techniques: for Marketing, Sales, and Customer Support, Wiley Publishing Inc., Jun. 1997, 464 pages. (Book, copy not provided.).
Horovitz, “Watching Ads Is Real Science Research Companies Monitor Physiological Reactions to Commercials to Determine Their Effectiveness,” Los Angeles Times, Sep. 1, 1991, 3 pages.
Sung et al., “Wearable feedback systems for rehabilitation,” Journal of NeuroEngineering and Rehabilitation, Jun. 29, 2005, 12 pages.
Jaffe, Casting for Big Ideas, Adweek Magazine Series, Book 8, 2003, 256 page. (Book, copy not provided).
Hall, “Advertising as a Factor of Production,” ADMAP, 2003, pp. 47-49, 1 page. (Abstract provided.).
Ranii, “Adding Science to Gut Check,” The News & Observer, D3 (Apr. 6, 2005), 1 page. (Abstract provided.).
Landau et al., “Different Effects of Voluntary and Involuntary Attention on EEG Activity in the Gamma Band,” The Journal of Neuroscience, 27(44), Oct. 31, 2007, pp. 11986-11990, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/037,666, dated Dec. 6, 2018, 96 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/037,666, dated Oct. 4, 2018, 37 pages.
Advisory Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Oct. 1, 2018, 17 pages.
English Translation of Office Action, issued by the Israeli Patent Office in connection with Patent Application No. 201187, dated Apr. 23, 2014, 2 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/708,525, dated Oct. 5, 2018, 31 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/234,372, dated Sep. 19, 2018, 8 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 13/965,805, dated Sep. 10, 2018, 11 pages.
Advisory action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/608,660, dated Oct. 26, 2012, 3 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Dec. 28, 2018, 8 pages.
Decision on Request for Rehearing, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,372, dated Mar. 27, 2019, 7 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/410,380, dated Jul. 8, 2019, 12 pages.
Decision on Appeal, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/413,297, dated Sep. 24, 2019, 15 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 15/967,939, dated Oct. 4, 2019, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 15/989,987, dated Oct. 17, 2019, 8 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/183,131, dated Nov. 18, 2019, 5 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Jan. 22, 2020, 12 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 15/967,939, dated Jan. 29, 2020, 8 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/183,131, dated Mar. 24, 2020, 8 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/151,044, dated Apr. 2, 2020, 5 pages.
Intimation of Grant and Recordal of Patent, issued by the Indian Patent Office in connection with Indian Patent Application No. 4438/KOLNP/2009, dated Jul. 3, 2019, 1 page.
Intimation of Grant and Recordal of Patent, issued by the Indian Patent Office in connection with Indian Patent Application No. 4441/KOLNP/2009, dated Sep. 3, 2019, 1 page.
Hearing Notice issued by the Indian Patent Office in connection with Indian Patent Application No. 6145/CHENP/2009, dated Mar. 12, 2020, 2 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/151,044, dated Oct. 9, 2020, 8 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/790,160, dated Oct. 29, 2020, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/193,930, dated Nov. 18, 2020, 8 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Dec. 9, 2020, 8 pages.
Kim et al., “Design for an Interactive Television Advertising System,” Proceedings for the 39th Hawaii International Conference on System Sciences (2006), 9 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 12/546,586, dated Jul. 24, 2020, 14 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/193,930, dated Jul. 27, 2020, 12 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/421,864, dated Aug. 13, 2020, 24 pages.
Non-Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/151,050, dated Apr. 6, 2020, 13 pages.
Final Office Action, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/151,050, dated Oct. 7, 2020, 5 pages.
Notice of Allowance, issued by the United States Patent and Trademark Office in connection with U.S. Appl. No. 16/151,050, dated Jan. 22, 2021, 8 pages.
Related Publications (1)
Number Date Country
20210350407 A1 Nov 2021 US
Provisional Applications (1)
Number Date Country
60968567 Aug 2007 US
Continuations (3)
Number Date Country
Parent 16151050 Oct 2018 US
Child 17333787 US
Parent 13730564 Dec 2012 US
Child 16151050 US
Parent 12200813 Aug 2008 US
Child 13730564 US