The present invention relates to electronic reminders and, more particularly to, the intelligent generation, organization, triggering, and delivery of reminders and tasks in electronic to-do lists.
In accordance with some implementations, a method for triggering an action associated with a task is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The method includes receiving a task item associated with a triggering criterion, wherein the triggering criterion requires an occurrence of any of a plurality of communication events, wherein the plurality of communication events includes a telephone call. The method further includes detecting the occurrence of a first communication event of the plurality of communication events. The method further includes determining that the triggering criterion is satisfied. The method further includes, upon determining that the triggering criterion is satisfied, causing a notification associated with the task item to be presented to a user of the electronic device.
In some implementations, the first communication event is a telephone call. In some implementations, at least two of the plurality of communication events are of different communication types. In some implementations, the different communication types include a telephone call and an email. In some implementations, the different communication types include a telephone call and a text message. In some implementations, the different communication types include a telephone call and a video chat.
In some implementations, the first communication event includes content, and the triggering criterion requires that the content relate to a predefined subject. In some such implementations, the method further comprises determining that the triggering criterion is satisfied based at least in part on a determination that the content relates to the predefined subject. In some implementations, the content is text from an email or a text message.
In some implementations, the first communication event is a telephone call, and the method further comprises obtaining a text string corresponding to one or more utterances spoken during the telephone call by at least one participant in the telephone call, wherein the determination that the content relates to the predefined subject includes determining that one or more words in the text string relate to the predefined subject.
In some implementations, the triggering criterion further requires that the communication event be a communication with a specified person.
In some implementations, the first communication event is a text message, and causing the notification to be presented comprises causing an affordance relating to the first task item to be displayed on the electronic device near a representation of the text message. In some implementations, the affordance is a button that, when selected by the user, causes an action associated with the task item to be performed.
In some implementations, the first communication event is an email, and causing the notification to be presented comprises causing an affordance relating to the first task item to be displayed on the electronic device near a representation of the email. In some implementations, the affordance is a button that, when selected by the user, causes an action associated with the task item to be performed.
In some implementations, the first communication event is a telephone call, and causing the notification to be presented comprises causing an audio output relating to the first task item to be produced by the electronic device. In some implementations, the electronic device is a telephone, and the audio output is produced while the telephone is ringing. In some implementations, the audio output is produced during a telephone conversation between a user of the electronic device and a participant of the telephone call. In some implementations, the audio output is produced at a lower volume than a voice of the participant of the telephone call.
In accordance with some implementations, another method for triggering an action associated with a task is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The method includes providing a task item associated with a triggering criterion, wherein the triggering criterion requires an occurrence of any of a plurality of communication events, wherein at least two of the plurality of communication events are of different communication types. The method further includes detecting the occurrence of a first communication event of the plurality of communication events. The method further includes determining that the triggering criterion is satisfied. The method further includes, upon determining that the triggering criterion is satisfied, causing a notification associated with the task item to be presented to a user of the electronic device.
In some implementations, the triggering criterion requires that the first communication event be associated with a specified person. In some implementations, the first communication event is associated with the specified person when the specified person initiates the first communication event or receives the first communication event from the user.
In accordance with some implementations, yet another method for triggering an action associated with a task is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The method includes providing a task item associated with a triggering criterion, wherein the triggering criterion requires an occurrence of a telephone call at an electronic device. The method further includes detecting the occurrence of the telephone call. The method further includes, in response to detecting the occurrence of the telephone call, causing a notification associated with the task item to be presented to a user of the electronic device, wherein the notification is separate from a telephone call notification.
In accordance with some implementations, a computer-readable storage medium (e.g., a non-transitory computer readable storage medium) is provided, the computer-readable storage medium storing one or more programs for execution by one or more processors of an electronic device, the one or more programs including instructions for performing any of the methods described herein.
In accordance with some implementations, an electronic device (e.g., a portable electronic device) is provided that comprises means for performing any of the methods described herein.
In accordance with some implementations, an electronic device (e.g., a portable electronic device) is provided that comprises a processing unit configured to perform any of the methods described herein.
In accordance with some implementations, an electronic device (e.g., a portable electronic device) is provided that comprises one or more processors and memory storing one or more programs for execution by the one or more processors, the one or more programs including instructions for performing any of the methods described herein.
In accordance with some implementations, an information processing apparatus for use in an electronic device is provided, the information processing apparatus comprising means for performing any of the methods described herein.
People have devised numerous ways to remind themselves of certain tasks or events. Many people have and still use physical media to write on, such as sticky notes and calendars. With the ubiquity of electronic devices, many people have turned to computers to help manage their to-do lists and keep a record of upcoming events. Numerous reminder and to-do applications are available, both for desktop computers as well as handheld devices, such as laptop computers, tablet computers, and “smart” phones.
However, the timeliness and accuracy of a notification provided to a user of a reminder application depends almost entirely on input received from the user. For example, if a user enters a reminder with the wrong date for an important event, then the user might not receive a notification of the event until after the event has occurred. As another example, if a user provides a generic description of a task (e.g., “send him an email”) in a to-do application, then, when the user later reads the description, the user might not remember who “him” is and/or what the content of the email should be. In other words, when it comes to reminder and to-do applications, the old adage of “garbage in garbage out” is applicable.
Moreover, users do not always have the time or willingness to set a triggering condition for each and every event that should trigger a reminder notification, and reminder and to-do applications may not be able to monitor each type of event or condition that may be a useful trigger. For example, for a reminder such as “remind me to call my mom after I hear from my sister,” traditional reminder and to-do applications would not monitor different communications types so that a reminder notification could be issued regardless of how the user “heard from” her sister.
The approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.
In the drawings:
In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
Multiple techniques are provided below for assisting a user in managing his/her tasks. Tasks are represented as task items in a computing system. Specifically, approaches for generating task items, organizing task items, triggering the notification of tasks based on corresponding task items, and consuming task items are described in detail below.
With respect to generating a task item, a task item is generated based on input from a user. The input (whether voice or text) includes one or more details about a task while context of the input is used to determine one or more other details about the task. The one or more other details are not reflected in the input. Context may include data that is stored on a device of the user. For example, a user provides the following voice input: “Call George at 5 PM today.” The user may have a contact list that includes information about multiple people with the first name of George. However, based on the context of the input, such as a recent voice message from a phone number associated with a particular George, it can be determined which George the user intends to call. In this way, a user is not required to be specific about each detail of a task when providing input about the task.
With respect to triggering a notification of a task item, one or more characteristics of a device may be used to determine whether to trigger the notification. Thus, time is not the sole factor (if at all) of whether a notification should be provided to a user. Examples of such characteristics may include where the device is located, what the device is displaying or processing, and specific movements of the device. For example, the fact that the device is in a car or at work may trigger the generation of a reminder of a task. As another example, if the device is currently displaying web content, then a user of the device may be considered to be “online,” which status might trigger a notification of a task to be provided to the user. As yet another example, communication events, such as emails, telephone calls, text messages, and the like, can trigger a notification of a task to be provided to the user.
With respect to “consuming” task items, instead of simply providing a reminder of a task, the task is automated so that a user is not required to perform the steps typically required to perform the task. For example, a user may want to call a particular person at a particular time. When the particular time equals the current time, instead of simply reminding the user about the call, the call can be set up, ready to make without the user having to specify the person's phone number.
With respect to organizing task items, task items may be organized automatically or manually into multiple lists. Each list corresponds to a different attribute of a task item, such as the type of task, the type of criteria that is used to trigger a notification of a task, and the location of where the task is to be performed.
Again, a “task item” is an electronic data item that contains one or more details about a task to perform, whether by a user or automatically by a process. A task item is generated based on input from a user. A task item may be one of two types: tasks associated with reminders (“reminder task”) and tasks not associated with reminders (“non-reminder task”). A reminder task is associated with an action, such as a notification being generated and provided to a user, while a non-reminder task is not associated with any action. A non-reminder task may be associated with a “complete-by” date or time. However, the complete-by date or time does not trigger the creation of a notification or other action. In other words, while a reminder task is associated with one or more triggering criteria that, when satisfied, trigger an action, a non-reminder task is not. Thus, a “task item” may or may not be associated with one or more triggering criteria that, when satisfied, triggers an action.
Device 110 is any computing device that is capable of receiving input from a user and displaying information about tasks. Non-limiting examples of device 110 include a desktop computer and a handheld device, such as a laptop computer, a tablet computer, and a “smart” phone. In the illustrated embodiment, device 110 includes a task manager 112. Task manager 112 processes task items, both of the reminder type or of the non-reminder type. Task manager 112 may be implemented in software, hardware, or any combination of software and hardware.
Device 110 includes communication technology (e.g., wireless technology) for sharing information with other devices. Device 110 can include a variety of integrated user interface units or can be coupled to user interface units through one or more communication ports or data links of the device. Non-limiting examples of user interface units include a voice input unit (e.g., a microphone), physical input units (e.g., a keyboard, a mouse, a track ball, a rotary dial or wheel, a touchpad, or a touch-screen), and motion sensors (e.g., an accelerometer, magnetometer, or a gyroscope). Any of these user interface units can be implemented as an external unit that communicates with device 110 to provide user input using a wired or wireless communication technology. Examples of wired communication technology include Universal Serial Bus (USB) interface, FireWire interface, etc. Examples of wireless communication technology include Bluetooth, Wi-Fi, and WiMax, infrared. Through these user interface units, device 110 can receive physical or voice inputs from the user.
Device 110 includes one or more output units to present visual and audio information to a user. Non-limiting examples of output units include a display unit for displaying visual data and a speaker for playing audio.
Cloud 130 is implemented by one or more computing devices. Cloud 130 hosts multiple services, such as a NLP (natural language processing) service 132 and one or more other services 134A-N. NLP service 132 uses one or more models of real-world things that a user can talk about in order to make sense of what the user is trying to say. For example, NLP service 132 can determine, based on models and context, what a user may be referring to when the user uses terms like, “him,” “there,” or “that.” An example of how NLP service 132 might operate is described in U.S. patent application Ser. No. 12/987,982, referenced above.
NLP service 132 may employ numerous APIs to communicate with and initiate actions performed by the one or more other services 134A-N and, optionally, other services not hosted in cloud 130. For example, in response to voice data sent from device 110, where the voice data reflects the user command “Reserve two seats at Maribella's in San Jose at 7 PM tonight,” NLP service 132 makes an API call to an online reservation service provided by Maribella's restaurant to initiate the creation of two reservations at that restaurant for 7 PM. Thus, NLP service 132 allows many operations to be performed automatically without requiring a user of device 110 to manually input text data and interact with numerous applications.
Communication between device 110 and services hosted in cloud 130 is made possible via network 120. Network 120 may be implemented by any medium or mechanism that provides for the exchange of data between various computing devices. Examples of such a network include, without limitation, a network such as a Local Area Network (LAN), Wide Area Network (WAN), Ethernet or the Internet, or one or more terrestrial, satellite, or wireless links.
The network may include a combination of networks such as those described. Without limitation, the network may transmit data according to Transmission Control Protocol (TCP), User Datagram Protocol (UDP), and/or Internet Protocol (IP).
The following description includes numerous examples where both device 110 and cloud 130 take part in generating task items, organizing task items, triggering notifications of task items, and consuming task items, and examples where such operations are performed by either the device 110 or the cloud 130 alone. Instead, one or more of the techniques described herein may be implemented wholly on device 110 (making network 120 and cloud 130 unnecessary), wholly in cloud 130, or using any combination of device 110 and cloud 130.
Task items may be created on device 110 or in cloud 130 based on input received at device 110. Although not depicted, task items may be stored on device 110 or in cloud 130, or synchronized to both. If task items are stored in cloud 130, then task manager 112 may retrieve the task items in response to, for example, input from a user or the one or more triggering criteria associated with one or more task items being satisfied.
In the scenario where task items are created and stored in cloud 130, task manager 112 may be, primarily, a set of one or more user interfaces that display information about tasks. Thus, a task service (not shown) in cloud 130 would be responsible for maintaining task items and triggering any notifications when triggering events occur.
Alternatively, task manager 112 creates and stores task items on device 110. In this scenario, task manager 112 may be entirely responsible for maintaining task items and generating any notifications when triggering events occur. One advantage of this scenario is that device 110 may be operating in an “offline” mode where device 110 is not capable of communicating with any service hosted in cloud 130.
Further, in this scenario, device 110 may include a service like NLP service 132, which may be part of task manager 112 or may execute separately from task manager 112. Such a service acts as a conversational interface to allow a user to quickly and easily create tasks. Such a service may be implemented by a process that is continually executing in the background without requiring a user of device 110 to provide input to cause the service to execute. Thus, whenever device 110 starts up (or restarts), the service is automatically started.
Alternatively, information needed to create task items may be identified by NLP service 132 (i.e., hosted in cloud 130). Device 110 may include a user input interface that continuously executes in the background, identifies input (e.g., voice or text) from a user, and sends the input over network 120 to NLP service 132. Once NLP service 132 identifies task details in the input, NLP service 132 may send task information (e.g., a description of a task and a time to complete the task) (a) over network 120 to task manager 112, which creates and stores a task item based on the information or (b) to a task service in cloud 130 to create a task item based on the information.
Most of the examples provided herein involve NLP service 132 receiving input data from device 110, identifying details (about a task) reflected in the input data, and providing those details to task manager 112. However, embodiments of the invention are not limited to this scenario. Such examples may alternatively involve only device 110 or may involve device 110 as merely an input and display device where NLP service 132 and a task service in cloud 130 provide the primary functionality.
I. Generating Task Items Based on Context
According to an embodiment of the invention, a task item is generated based on input and context of the input. “Context” of input refers to data that is currently or recently (relative to input, from a user, that initiated the generation of a task item) displayed or processed at device 110. Thus, context data is not reflected in the input from the user. For example, a user of device 110 may provide the following voice input: “Send him an email about the project when I get home.” The pronoun “him” is ambiguous because it is not clear, from the input alone, to whom “him” refers. However, the context of the voice input may be that device 110 currently displays (or just recently displayed) an email from an individual named Jack Bauer where the email includes a request for a status update about a project named “Bunny.” Based on the voice input and the context, task manager 112 (or a task service in cloud 130) creates a task item that includes the description “Send Jack Bauer an email about Project Bunny” and that includes the triggering criterion of device 110 being at a geographical location that is at or near the user's home. When device 110 is at or near the user's home, task manager 112 causes a notification to be displayed on device 110 where the notification includes the description from the task item.
At step 220, a task item is generated for the task based on the input data. At step 230, one or more second attributes for the task are determined based on context data that is separate from the input. Although step 230 is depicted as occurring after step 220, step 230 may occur before step 220.
At step 240, the one or more first attributes and the one or more second attributes are stored in association with the task item.
The steps of process 200 may be performed by one or multiple devices. For example, the input in step 210 may be processed at device 110 to generate the task item. In this scenario, task manager 112 (or another process executing on device 110) identifies the context data associated with the input to determine the one or more second attributes, for the task, that are not identified in the input. Task manager 112 then stores the one or more second attributes in or in association with the task item.
Alternatively, in another scenario, device 110 sends the user input over network 120 to NLP service 132. NLP service 132 accepts, as input, context data associated with the input to determine the one or more second attributes, for the task, that are not identified in the input. Context data may have been sent to NLP service 132 prior to the input that initiates the generation of the task item (in step 220). NLP service 132 sends the one or more second attributes to task manager 112 (or a task service in cloud 130). Task manager 112 stores the one or more second attributes in or in association with a newly-generated task item.
Certain words or phrases may be used to cue NLP service 132 to communicate with manager 112. For example, user commands that begin with “Remind me . . . ” and “I need to . . . ” are used by NLP service 132 to determine to communicate with task manager 112. In response to detecting one of those user commands, NLP service 132 analyzes the input data (from device 110) and, optionally, context data for certain types of task details, such as a location, time, description, and/or action. NLP service 132 then determines to communicate with task manager 112 and sends, to task manager 112, the task details as part of the communication(s).
Context data associated with user input that initiates the generation of a task item may come from one of many sources. Non-limiting examples of context data include data that is or was displayed on device 110 (“display data”), data that is stored on or in association with device 110 (“personalized data”), data that is or was processed by device 110 (“process data”), data that was previously provided by a user of device 110 (“input data”), data that indicates the location of device 110 (“location data”).
The following is an example of display data, or data that is or was displayed on device 110. Device 110 displays a map that includes a marker associated with a specific location on the map. A user of device 110 then says, while the map is displayed or soon after the map was displayed, “I need to be there by 5 today.” NLP service 132 (or a voice analyzer on device 110) analyzes voice data that reflects the voice input. NLP service 132 analyzes data that is currently displayed on device 110 to determine what “there” refers to. NLP service 132 identifies the marker and the associated location and replaces “there” with the location. NLP service 132 sends, to task manager 112, task data that indicates 5 PM today as the completion time of the task and the specified location as the location of the task. Task manager 112 generates a task item based on the task data.
As another example of display data, device 110 displays an email that is from a particular sender and includes a subject line. A user of device 110 then says, “I need to email him about that subject in two hours.” Device 110 sends voice data that reflects this input and an image of what is displayed to NLP service 132. In response, NLP service 132 identifies the email address of the sender of the email and the subject of the email. NLP service 132 sends, to task manager 112, task data that indicates a time of two hours from the current time as the completion time of the task and
The following is an example of personalized data, or data that is stored on or in association with device 110. A user of device 110 says, “I will have lunch with Rachelle tomorrow at 12 noon.” Device 110 sends voice data that reflects this input to NLP service 132, which identifies “Rachelle” in the voice data. NLP service 132 looks up “Rachelle” in contact data or an “address book” (stored on device 110 or in cloud 130) and determines that the last name of Rachelle is Goodwin. NLP service 132 then causes “Rachelle Goodwin” to be associated with a task item that is generated for the task. In addition to or instead of being stored on device 110, personalized data may be stored in cloud 130, i.e., remote to device 110.
The following is an example of process data, or data that was recently processed by device 110. For example, a user of device 110 used device 110 as a phone to communicate with a friend. Device 110 keeps track of who the user recently spoke with. After ending the call, the user says, “Remind me to call her back in 30 minutes.” NLP service 132, in addition to analyzing the voice input, analyzes data that indicates who recently established a phone call with device 110 (e.g., the last five phone calls). NLP service 132 determines the phone number of the most recently established phone call with device 110. NLP service 132 then determines, based on contact data, that the phone number is associated with particular individual. NLP service 132 sends, to task manager 112, task data that indicates a task of calling, a time of 30 minutes from the current time as the completion time of the task, the name of the particular individual, and, optionally, the phone number of the particular individual. Task manager 112 generates a task item based on the task item.
The following is an example of input data, or data that was recently (e.g., the last 5 minutes) provided by a user of device 110. The input from the user may be text input or voice input. Device 110 or NLP service 132 keeps track of recently entered input and may use that input to determine the identity of certain terms reflected in current input. For example, a user of device 110 says, “Remind me to meet him there at 7 tonight.” NLP service 132 receives voice data that reflects that voice input and identifies the terms “him” and “there.” Although it is not clear who “him” is and where “there” is, NLP service 132 accesses input that was recently received from the user. Such recently-received input reflects the names “George Reed” (identified as a name of a person) and “Starbucks” (identified as a place). In response, NLP service 132 causes a task item to be generated where the task is to “Meet George Reed at Starbucks” where the time is 7 PM of the current day.
The following is example of location data, or data that indicates a location of device 110, whether current or past. A user of device 110 says, “Remind me to meet Julian here next Thursday for lunch.” Device 110 sends voice data that reflects this input to NLP service 132. NLP service 132 identifies the term “here” and, in response, determines where device 110 is currently located. The current location may be determined in numerous ways. For example, device 110 may provide, to NLP service 132, a geographical location, such as longitude and latitude coordinates. NLP service 132 may then determine, based on the coordinates, a name of the place or establishment that is located at those coordinates. NLP service 132 causes a name of the place or establishment to be associated with a task item for the task to meet Julian for lunch on the date indicated.
Alternatively, the user may say, “I need to meet Josh Peters tomorrow at the same place where I was last Thursday at noon.” Device 110 sends voice data that reflects this input to NLP service 132. NLP service identifies the phrase “at the same place where I was last Thursday at noon” and, in response, determines where device 110 was located last Thursday at noon. NLP service 132 accesses location history data (stored in cloud 130 or stored on device 110 and sent to NLP service 132) and determines where device 110 was located last Thursday at noon. The location history may indicate the name of a place or may consist of geographical coordinates. If geographical coordinates, then NLP service 132 determines a name of the place or establishment that is located at those coordinates. NLP service 132 causes that name to be associated with a task item for the task to meet Josh Peters on the date indicated.
Events that occur with respect to device 110 may also be used to create task items. Such events may fall into one or more categories (or types) of context data described above, such as display data, presentation data, and process data. For example, device 110 detects an incoming call and notifies the user of the call by causing a phone number or other identifying information about the call or caller to be displayed on a screen of device 110. In addition to this information, the display may include three selectable options: “Answer”, “Ignore”, and “Call Back Later.” If the user selects “Call Back Later”, then a task item is created where the task item identifies the caller and, optionally, a time of the call and/or a time to make a call to the caller. Also, the task item may be automatically categorized as a task of type “To Call.”
Many of the examples herein regarding generating task items include a user providing voice or text input that includes details about a task. Another non-limiting example of how a task item may be generated is a user selecting (or highlighting) text that is displayed on a screen of device 110. The selected text is considered context data. After the text is selected, the user may be presented with one or more options, one of which is a “Remind” option which, when selected, causes a task item to be generated. Task manager 112 generates the task item based on the information reflected in the selected text. Details of the task item may be also determined from other context data, such as a time or event to trigger a notification of the task.
In some situations, NLP service 132 is unable to determine one or more details about a task based on input received from device 110 and the context associated with the input. Thus, in an embodiment, NLP service 132 prompts a user of device 110 for further input to determine the one or more details. The one or more details may pertain to any attribute of a task item, such as the description of the task, the location of the task, the location of a reminder (if any), or the time of the task.
For example, NLP service 132 receives, from device 110, voice data that reflects a user's command to “Remind me to call Steve at 7.” NLP service 132 may have access to information (e.g., an address book) about numerous contacts, of the user, that have the name of Steve. Further, nothing in the address book can be used to disambiguate which of the Steve contacts to call. Therefore, NLP service 132 sends, to device 110, the following message to be displayed (or played audibly) by device 110: “Do you mean Steve Anderson, Steve Hanson, or Steve Jobs?” The user then provides, to device 110, voice or text input that indicates one of the three Steve contacts. In response, device 110 sends the corresponding voice or text data over network 120 to NLP service 132.
As another example, NLP service 132 receives, from device 110, voice data that reflects a user's command to “I need to pick up bread at Whole Foods.” in response, NLP service 132 performs a lookup of the nearest Whole Foods stores to (a) the current location of device 110 or (b) the user's home. There may be multiple Whole Foods stores that are near device 110's current location and near the user's home. Therefore, NLP service 132 sends, to device 110, the following message to be displayed by device 110: “Which Whole Food's? The one on Almaden Rd, Chester Expressway, or Green Street?” The user then provides, to device 110, voice or text input that indicates one of the three Whole Foods stores. In response, device 110 sends the corresponding voice or text data over network 120 to NLP service 132.
As another example, NLP service 132 receives, from device 110, voice data that reflects a user's command to “Remind me to text Jerry by 8.” In response, NLP service 132 determines, based on the voice data and the context of the input that Jerry is Jerry Wall, indicated in the user's contact list (or address book). However, it is unclear whether the user intended 8 AM or 8 PM as the time to send an SMS message to Jerry. Therefore, NLP service 132 sends, to device 110, the following message to be displayed by device 110: “Do you want to text Jerry Wall at 8 AM or 8 PM?” The user then provides, to device 110, voice or text input that selects one of the two times. In response, device 110 sends the corresponding voice or text data over network 120 to NLP service 132.
In an embodiment, NLP service 132 determines, based on input from a user of device 110, one or more categories to associate with a task item. The one or more categories may be one of many different categories, which may be virtually limitless. Non-limiting examples of categories with which a task item may be associated include things to purchase, things to do on vacation, things to do at work, and things to do while driving. Each category may be associated with a sub-category. For example, a “purchase category” may be divided into a grocery category indicating items to purchase at a grocery store, a book category indicating books to purchase, and a music category indicating songs to purchase.
For example, a user may provide the following voice input to device 110: “Remind me to get milk.” Device 110 sends voice data that reflects that input to NLP service 132. NLP service 132 determines that a task item should be created and that “get milk” should be the description associated with the task item. NLP service 132 may also determine that milk is a grocery item and that the task item should be associated with a grocery category and/or a purchase category. Thus, NLP service 132 may send, to task manager 112, category data that indicates one or more categories with which the task item (whether created by NLP service 132, by a task service in cloud 130, or by task manager 112) should be associated.
As will be described hereinafter, the one or more categories associated with each task item may be used to organize task items that belong to the same category and display, on device 110, task items of the same category. This will allow a user of device 110 to view task items by category, in addition to or instead of by completion time, by creation time, by trigger type (described hereinafter), by location, by type (e.g., reminder task v. non-reminder task), or by some other criterion.
II. Triggering Notifications of Task Items
As noted previously, a task item may be associated with one or more triggering criteria (or triggers) that, when satisfied, causes a notification to be presented to a user of device 110 or some other action to be performed. When one or more triggering criteria of a task item are satisfied, a notification (or other action) is “triggered.” Non-limiting examples of triggering criteria include time, location, relative travel time, context triggers, and exogenous triggers, each of which is described in more detail below.
The time of a time trigger may be an absolute time, a relative time, a recurring time, or a symbolic deadline. An example of an absolute time is Jun. 6, 2011, 9 AM Pacific Time. An example of a relative time is “10 minutes before the Patriots-Jets football game.” An example of a recurring time is “Every Thursday at 10 AM.” An example of a symbolic deadline is “end of business day”.
According to an embodiment of the invention, the location of device 110 is a triggering criterion associated with a task item. Such a triggering criterion is referred to herein as a “location trigger.” The location of device 110 may be determined in one of many ways. For example, the location of device 110 may be automatically determined based on Wi-Fi positioning, cell positioning, and/or GPS (global positioning system) positioning. Device 110 may determine its current location with or without input from a service in cloud 130.
In an embodiment, a user may provide input that indicates a label to be associated with a certain geographical location. For example, a user of device 110 may speak the following sentence, “I am home” or “I am at Whole Foods.” NLP service 132 may then associate the word “home” or phrase “Whole Foods” with the current location of device 110, as determined based on one of the three positioning methods mentioned previously. This association of a word with a location may be later leveraged to determine where “home” or “Whole Foods” is located.
A location trigger may not be associated with a specific geographic location or area. Instead, a location trigger may be associated with a place that is not limited to a specific geographic location or area. For example, a location trigger of a task item may be “on the road” or “while driving.” Device 110 (or a process executing on device 110) determines that the current location of device 110 is on a freeway or another busy road. Thus, this determination can be made regardless of the speed at which device 110 is moving or whether device 110 is paired with another device that would indicate that the user is traveling. Based on this determination, task manager 112 analyzes one or more task items to determine whether any task items are associated with the “on the road” or “while driving” location trigger.
As another example, a location trigger of a task item may be the user's car. Specifically, the user may have provided the following voice command: “Remind me to call my mom while driving.” NLP service 132 analyzes voice data that reflects that command and determines that “while driving” refers to the user's car. The user's car may have a Bluetooth-enabled component to allow device 110 to communicate with the user's car. When device 110 comes into range of a Bluetooth signal propagated by a Bluetooth-enabled component in the user's car, device 110 determines that device 110 is located in (or at least near) the user's car. In response to this determination, task manager 112 triggers the location trigger of the task item. Task manager 112 causes a reminder message to be displayed on device 110, where the reminder message informs the user to call his mother. The user may then provide a single tap or a voice response that causes a phone application executing on device 110 to initiate a call to a phone number associated with the user's mom.
While establishing a connection (or “pairing”) with another Bluetooth-enabled device is one example of pairing that can be used to determine device 110's location, other types of pairings are possible. For example, device 110 may detect certain network data during the evening and morning hours. The network data indicates one or more networks to which device 110 may connect. The network data may include the names of one or more networks or MAC addresses of one or more routers. Device 110 may then determine that whenever that network data is detected, device 110 is considered to be at the user's home. Thus, actual pairing is not required since pairing entails the establishment of a connection between device 110 and another device, such as a router. As another example, device 110 may detect a Wi-Fi signal on a train, subway, or bus. The Wi-Fi signal might indicate the type of transportation that corresponds to the Wi-Fi signal. Thus, device 110 might detect, based on the Wi-Fi signal, that its location is “on a train,” “in a subway,” or “on a bus.” If a triggering criterion of a task item indicates one or more of these locations, then an action associated with the task item may be triggered. Further, such “transit-oriented” locations may also be considered to be associated with specific contexts (described in more detail below), such as “in transit” or “while traveling.” Thus, detection by task manager 112 of such contexts may cause actions associated with certain task items to be performed.
The foregoing examples of location triggers can be categorized as “arrival triggers,” such as are found in user input to “Remind me to do X when I arrive at Y.” Another type of location trigger is a “departure trigger,” an example of which is found in the user command to “Remind me to do X when 1 leave work” or “ . . . when I leave here.” In an embodiment, in the departure trigger scenario, a minimum distance from the current location and the location of the departure is required before a particular departure trigger “fires.” Such a minimum distance may be helpful to avoid the performance of corresponding actions when there are false starts.
Additionally, a location trigger may be one of multiple conditions that trigger an action of a task item. Examples of user commands that include multiple conditions include “Remind me to do X when I get home or at 8 PM at the latest,” “Remind me to do X before 8 PM or when I leave, whichever is first,” and “Remind me to do X before 8 PM or while I am driving, whichever is first.”
In an embodiment, the location of device 110 and a time associated with a task item is used to provide a notification to a user of device 110. Thus, while the time may be one of the one or more triggering criteria associated with the task item, the location of device 110 may not be, at least explicitly so.
At step 310, task manager 112 determines a current location of device 110. At step 320, task manager 112 determines a location of a destination (or “destination location”) associated with (or identify by) a task item. At step 320, based on the distance between the two locations, task manager 112 determines a “travel time,” or the time it might take for the user of device 110 to travel to the destination location. At step 330, task manager 112 determines a “difference time,” or the difference between the current time and the time triggering criterion associated with the task item. At step 340, if the travel time is the same as or near the difference time, then task manager 112 provides a notification to the user. This notification acts as a reminder for the user to begin (if s/he has not already done so) traveling to the destination.
For example, a task item may be for a reminder to meet Sue at a particular restaurant at 2 PM. Task manager 112 determines the location of device 110 and the location of the particular restaurant. The location of the particular restaurant may be determined by initiating, e.g., an Internet search and identifying the closest restaurant, with the same name, to device 110's location. Alternatively, an address of the particular restaurant may already be stored in association with the task item. Based on the distance between device 110's location and the particular restaurant, task manager 112 determines how long it will take for the user of device 110 to travel to the particular restaurant (or “travel time”). When the travel time is the same as or near (e.g., within 10 minutes) the difference between the current time and the time trigger (i.e., 2 PM), then task manager 112 causes, to be displayed on device 110, a message that indicates that the user should leave soon to arrive at the particular restaurant at 2 PM.
In an embodiment, the time of when to leave for a destination changes based on the current location of device 110. For example, when the current location of device 110 is at location A and the destination is at location B, task manager 112 determines that the user should begin traveling 50 minutes before the time of a scheduled task. However, in response to detecting that the current location of device 110 is now at location C, task manager 112 determines that the user should begin traveling 20 minutes before the time of the scheduled task. For example, a user of device 110 may be at home at the beginning of the day and task manager 112 determines that it will take 50 minutes to travel, from the user's home, to the location of a dinner event in the evening. Later in the day, the user of device 110 travels to work, which is closer to the location of the dinner event. In response to device 110 being at a different location, task manager 112 determines that it will take 20 minutes to travel, from the user's work, to the location of the dinner event.
In an embodiment, the time of when to leave for a destination changes based on current traffic information. For example, at 2:30 PM, task manager 112 determines that the time of when a user of device 110 should leave for a restaurant is 5:00 PM. However, due to a car accident on a freeway that the user can take to arrive at the restaurant, the traffic slows considerably. Task manager 112 determines, at 3:30 PM, that the time of when the user should leave for the restaurant is 4:00 PM.
The action associated with action 430 may be triggered (or performed) in response to task manager 112 determining that the current time (indicated at the top of travel reminder) equals the time indicated by start time 450. In the illustrated example, action 430 is a map-related action where task manager 112 causes a map to be generated at start time 450 and displayed to the user of device 110. The map includes an indication of the address of location 460, an indication of the user's current location, or both. Instead of automatically causing the map to be displayed at start time 450, task manager 112 might first cause a message to be displayed on device 110, wherein the message includes an option to generate the map. If the user selects the option (e.g., through voice input or tapping on the screen), then task manager 112 causes the map to be generated and displayed.
A reminder setting may be in an “on” or “off” mode. In
As indicated previously, task manager 112 might change start time 450 in response to changes in device 110's location. Thus, while start time 450 may indicate “3:30 PM Today” when device 110 is located at the user's home in the morning, start time 450 may indicate “4:20 PM Today” when device 110 is located at the user's work office in the afternoon.
In an embodiment, task manager 112 checks for changes in computed start time 450 in response to significant changes in device 110's location. Significant changes in location may be determined as a result of other events that are already being computed. For example, device 110 might already process events when it transitions between cell towers, and these events could trigger the re-computation of a change in location and, therefore, in an updated start time 450. Other non-limiting examples of events that indicate a potential significant change in location are changes in Wi-Fi signatures detected, the computation of accurate GPS locations for some other application (such as maps or navigation), a power cycle event, turning on or off radios on the device, alerts based on accelerometer signals, and the receipt of text messages or push notifications that contain location information.
In an embodiment, task manager 112 combines strategies for detecting significant event changes. For example, in a low power/low resolution mode, task manager 112 only checks for significant location changes every N minutes or only when some periodic computation occurs, such as checking for incoming data. In a high power/high resolution mode, task manager 112 uses cell tower positioning and/or GPS. A combined strategy might run the low power solution by default and then invoke the high power solution when the estimated start time is soon or when other events occur (for example, a change in Wi-Fi or Bluetooth signatures is detected).
In an embodiment, a travel reminder or start time item in a travel reminder may be associated with one or more modes of transportation. Non-limiting examples of modes of transportation include driving a car, riding a bus, bicycling, and walking. A default transportation mode may be driving a car. For example, task manager 112 may provide the option for a user to view start time 450 in a “car” mode, a “bus” mode, a “bike” mode, a “walking” mode, or multiple modes simultaneously. Depending on the current mode(s) selected for start time 450, the start time may vary widely. For example, in
In a related embodiment, a task item is associated with both a location and a date/time and a notification of the task may be triggered by either the user (or, rather, the user's device) being at the location or by the date/time. For example, if the user's device is at the location, (either on the date or regardless of the date), then a notification is triggered. If the user has not arrived at the location on the day indicated by the date (or at the location by the time), then the time is used as a “last resort” for triggering a notification.
As described previously, time and location are examples of types of triggering criteria associated with a task item. Another type of triggering criteria associated with a task item is context. A “context trigger” refers to one or more characteristics of device 110 other than simply the device 110's location. Thus, like context triggers, travel triggers and travel time triggers also refer to one or more characteristics of device 110.
Context triggers may be categorized into one of two types: data-oriented context triggers and spatial-oriented context triggers. Non-limiting examples of data and/or events on which data-oriented context triggers can depend include the kind or type of data that device 110 is displaying on its screen (e.g., video), the specific application(s) or type of application(s) that are currently executing on device 110 (e.g., a texting application or a music application), the type of input that device 110 is receiving from a user (e.g., voice, text, selections, button presses), the occurrence of (or a property of) a communication event (e.g., receiving or initiating a telephone call, text message, or email, etc.), and the type of network connections available to device 110 (e.g., Wi-Fi or cellular network).
For example, a user command that device 110 receives may be “I need to email Bob when I am surfing the Internet.” The phrase “surfing the Internet” is presumed to mean that when the user is interacting with a web browser on device 110, the context of device 110 (or of the user) is “when online.” In response to determining the context of the device or of the user, a reminder will be sent to the user to inform the user to email Bob. Additionally another reminder may be provided to the user for any other task items that are associated with the “when online” context trigger.
As another example, a user command that device 110 receives may be “Remind me to email Jane Smith when I have a Wi-Fi connection.” In response to device 110 detecting a Wi-Fi signal that does not require a password or that requires a password accessible to device 110, task manager 112 causes a notification to be displayed on a screen of device 110, where the notification indicates that Jane is to email Jane Smith.
Another type of data-oriented context trigger is a communication based trigger. Communication based triggers detect the occurrence of and/or a property of a communication event that occurs at or is detectable by device 110. Communication events include, for example, telephone calls, emails, text messages (e.g., SMS messages, Internet chat messages, IMESSAGE messages), videotelephony calls (e.g., FACETIME video chat sessions), and the like. These communication events can be used as triggering criteria for reminder notifications. In particular, by monitoring various communication types, such as the examples listed above, the usefulness and relevance of reminder notifications can be increased. For example, triggering criteria can be selected to more closely match the user's intent in setting the reminder. For example, when a user asks to be reminded to mention something to a person “the next time I talk to her,” it may more accurately reflect the user's intent to trigger the reminder before or during any communication event with that person-regardless of the communication type-rather than just monitoring for a communication event of a single type (e.g., a telephone call).
In some implementations, communication based triggers are configured to detect the occurrence of a communication event with a certain person or group of people that the user identifies when creating the task item. For example, a user may issue a command to device 110 to “remind me to ask my Mom about Thanksgiving plans next time I talk to her.” A task item can then be created that will detect the next communication event between the user and his mother.
Several types of triggering criteria can be established for communication based triggers. For example, triggering criteria can be based on the type of communication event (e.g., telephone calls, emails, etc.), the identity of a party to a communication event (e.g., a sender of an email, a participant in a telephone call), the initiator of a communication event (e.g., who called whom), the content of a communication event (e.g., whether the communication relates to a particular subject or includes particular words), or any combination of these or other factors. Thus, the statement “next time I talk to her” in the example above may be interpreted as a condition on the party to the communication event (i.e., the user's mother) and on the type of communication event (i.e., that the communication type must include a voice communication component). As discussed below, that statement could also be interpreted without regard to the type of communication event, such that the reminder is triggered when any type of communication is detected between the user and his mother. With respect to conditions on the initiator of the communication, the statement “next time I talk to her” may suggest that the user did not intend to constrain the triggering event based on who initiated the communication, and the reminder can be triggered regardless of who called whom, for example. On the other hand, the command “next time I call her,” may be interpreted such that the reminder is only triggered if the user initiates the communication with his mother. Further examples and explanations of communication based triggers and various triggering constraints are discussed below.
In some implementations, the specific type or types of communication events that trigger a reminder depend on the particular words in the user input that established the reminder. For example, in order to create a task item with a communication based trigger, a user may issue the following command to device 110 (e.g., as a voice input): “Remember to text Mom when I am talking to my sister Sarah.” As noted above, in some implementations, the phrase “when I am talking to my sister Sarah” is presumed to mean that a reminder should be provided to the user when a voice communication (e.g., a phone call) is established with Sarah. In some implementations, however, the command is interpreted to mean that a reminder should be triggered when any type of communication event with Sarah is detected, such as a phone call, an email, a text message, a videochat, etc. Thus, in some implementations, device 110 monitors communication events of multiple different types in order to determine whether a triggering criterion has been met. Given the various ways in which people communicate, and the numerous communication types available on the device 110, triggering a single notification with multiple types of communications provides for flexible and convenient task reminders. Indeed, it would be a nuisance if a user had to create a different rule or triggering criterion for each communication type, such as one reminder to be triggered based on an email, another based on a phone call, another based on a text message, and so on, especially when all of the reminders would have identical content.
In some cases, though, users will wish to specify that only particular communication types are to trigger a particular notification. For example, a user may request a notification to be triggered “the next time I call James,” “the next time I email James,” or “the next time I text with James,” etc. In these cases, tasks will be created with triggers limited to phone calls, emails, or text messages, respectively. In some implementations, users may specify a particular group of communication types to trigger a particular notification. For example, a user may request a notification for a task to be triggered “the next time I text or email with James.” Accordingly, this notification will be triggered by a text message or email communication event with a contact named “James.”
In some implementations, NLP service 132 interprets the particular terms in a task item creation command to determine what communication types to monitor as triggering criteria for that task item. For example, in some implementations, NLP service 132 interprets phrases such as “talk to,” “speak with,” “chat with,” and “hear from” as establishing a trigger based on any communication type, and not only phone calls. Specifically, many users may understand these terms to relate to several types of communication (e.g., phone, text, etc.), rather than to just a single type (e.g., phone). Accordingly, when a command is received that uses these terms, the triggering criteria will not be limited to only one communication type. Moreover, interpreting such terms to include different communication types may reflect a user's intent more accurately: unless a user specifies a particular communication type (e.g., by reciting words such as “call” or “email” or “text”), it may be that the user actually intends the reminder to be triggered upon the occurrence of any of the various possible communication types. On the other hand, in some implementations, NLP service 132 interprets terms such as “talk,” “speak,” and/or “chat” as establishing a trigger requiring a voice-based communication event, such as a telephone call or a video chat.
In some implementations, where the user's command does not explicitly recite the communication types that are to trigger a notification, or when it is otherwise unclear from the user's command, device 110 (in conjunction with NLP service 132, discussed above) engages in a virtual dialogue with the user to disambiguate the user's input. In some implementations, NLP service 132 prompts a user for further input to determine additional details about the task item. For example, if a user requests a reminder to be triggered “next time I talk to James,” device 110 may respond by presenting to the user a question such as “do you mean when you are on the phone with James?” or “shall I also remind you when you email or text with James?” Such questions may be presented audibly, visually, or both. The user can then provide an answer to the question, which device 110 and/or NLP service 132 will interpret to determine what types of communication events will trigger that particular notification. These and other techniques can be used to disambiguate other aspects of a user's input as well, such as ambiguous names in user inputs (e.g., where multiple people named “James” appear in the user's contacts), and the like.
As noted above, communication based triggers can also depend on which party to a communication event initiated the communication. For example, a triggering criterion of “next time I call my wife” indicates that the notification should be triggered only when the user initiates a telephone call with his wife, but not when the user's wife calls the user. On the other hand, communication based triggers can also be agnostic to which party initiated the communication. For example, in some implementations, a triggering criterion of “next time I talk to my wife” will be satisfied regardless of whether the user called (or emailed, texted, etc.) his wife, or whether the wife called (or emailed, texted, etc.) the user.
In some implementations, communication based triggers are configured to detect the occurrence of communication events of a particular communication type without regard to whom the communication is with. For example, a user may issue the command “Remind me to call my mom next time I am on the phone.” The phrase “on the phone” can be presumed to mean that when the user is using device 110 as a phone, a notification will be sent to the user to inform the user to call his/her mom. This notification will be triggered regardless of the other party to the phone call. In another example, the command “Remind me to email Bob next time I'm sending emails” results in a triggering criterion that will be satisfied when the user sends any email, regardless of the recipient.
In some implementations, communication based triggers are conditioned on the content of a communication event. For example, a notification can be triggered only when a particular subject, topic, word, etc., is detected in the content of a communication. The content of a communication event can be textual (e.g., the body of a text message or email), audio (e.g., words spoken during a telephone call or video chat), or any other content (e.g., information relating to attachments to emails or text messages). Various techniques may be used to determine whether particular content satisfies a triggering criterion. In some implementations, natural language processing techniques are used to determine subjects, topics, words, etc., in a communication. For example, NLP service 132 can analyze a transcript of a telephone call or the body of an email to determine whether it satisfies the triggering criterion (e.g., whether the communication relates to “Project Bunny,” or discusses the stock price of Apple Inc.).
Communication based triggers can be used to trigger notifications of various types. For example, notifications can include static content and/or dynamic content. In some implementations, static content includes text or other fixed content (e.g., image, sounds, files, etc.) that is specified by the user. For example, a task item may be created in response to the input “remind me to tell Sherri about the dinner plans when she calls.” When Sherri calls, then, the notification will include text such as “Tell Sherri about the dinner plans.” Dynamic content, on the other hand, includes information that is obtained, identified, updated, or retrieved at the time the notification is triggered. For example a task item may be created in response to a user input to “tell me the score of the last Packers game when I talk to Dad.” In this example, device 110 (or any appropriate combination of devices or systems described herein) looks up the score of the most recent Packers game when a communication event with the user's father is detected, and notifies the user of the score. Other dynamic content that can be looked up or searched includes emails, web content, stock prices, sports scores/statistics, weather reports, news stories, and the like. For example, an input that creates a task item may include a request for dynamic content such as “find all recent emails from my boss when he calls,” or “get the weather in Florida when my parents call,” or “search the web for news on Apple, Inc. when I'm texting with Tim.” Such content is then presented to the user when the appropriate communication event is detected.
In some implementations, the content (either static or dynamic content, as described above) is provided at the same device at which the communication event is detected. Continuing an example from above, when a phone call from Sherri is detected at a user's smart phone, the text of the reminder can be presented in a popup window (or spoken by an audio output unit) on the smart phone. In some implementations, the content is provided at a different device. Continuing another example from above, when a phone call from a user's boss is detected on a user's phone, recent emails from the user's boss are displayed on a computer that is separate from the phone (e.g., a laptop, desktop, or tablet computer).
Attention is directed to
A task item associated with a triggering criterion is received, wherein the triggering criterion requires an occurrence of any of a plurality of communication events including a telephone call (1802). For example, the triggering criterion may be the occurrence of a telephone call, and a plurality of communication events are monitored in order to detect the occurrence of the telephone call. In some implementations, the triggering criterion requires that the communication event be a communication with a specified person (1804).
In some implementations, at least two of the plurality of communication events are of different communication types (1806). For example, in some implementations, communication events of different types (e.g., phone, email, text message, etc.) are monitored in order to determine whether a communication event satisfying the triggering criterion occurs. In some implementations, the different communication types include a telephone call and an email (1808). In some implementations, the different communication types include a telephone call and a text message (1810). In some implementations, the different communication types include a telephone call and a videochat (1812).
The occurrence of a first communication event of the plurality of communication events is then detected (1814). In some implementations, the first communication event is a telephone call (1816).
Subsequently, it is determined that the triggering criterion is satisfied (1818). For example, in some implementations, detecting the occurrence of a first communication event (1806) satisfies the triggering criterion. In some implementations, other conditions must also be met in order to determine that the triggering criterion is satisfied. For example, as described below, a triggering criteria may be satisfied only upon a determination that content in a communication event relates to a predefined subject.
In response to determining that the triggering criterion is satisfied, a notification associated with the task item is caused to be presented to a user of the electronic device (1820). In some implementations, such as when the method 1800 is performed at least partially on device 110 (e.g., a smart phone, laptop/desktop/tablet computer, etc.), causing the notification to be presented to a user comprises any of displaying text on a screen of the device, outputting audio from an audio output unit, and the like. In some implementations, the notification includes information associated with the task item (1822). For example, if the task item includes information (e.g., text) that is to be presented as a reminder (e.g., “ask Mom about Thanksgiving plans”), that information can be presented to the user-either visually, audibly, or both—when the triggering criterion is satisfied. In some implementations, where information is presented to the user audibly, it is presented such that only the user can hear it. For example, a voice output including the words “ask Mom about Thanksgiving plans” may be presented to a user during a telephone call with his mother such that only the user can hear the output.
In some implementations, information associated with the task item includes dynamic information, as described above. The dynamic content is then obtained, retrieved, or identified when the notification is presented to the user. For example, the task item may be associated with a request to “show me the stock price for Apple Inc. when Tim calls.” When the triggering criterion is satisfied for this task item (e.g., “Tim” called the user), the current stock price for Apple Inc. is retrieved and presented to the user as part of the notification.
Step (1820) of method 1800 continues on
In some implementations, the task item includes information that determines whether an affordance is to be displayed, and what action should be taken in response to a selection of the affordance. For example the task item may have been created in response to a user input to “send a text to Mom when I talk to Sarah.” When a communication event with Sarah is detected, a button may be presented in association with the text “Send Text to Mom.” In some implementations, selection of the button causes a predefined text message to be sent to the user's mother, or causes a text message editing application or region to appear so that the user can compose or edit a text message.
In some implementations, the first communication event is an email, and causing the notification to be presented comprises causing an affordance relating to the first task item to be displayed on the electronic device near a representation of the email (1826). In some implementations, the affordance is a button that, when selected by the user, causes an action associated with the task item to be performed (1828).
In implementations where a draft communication (such as a draft email, text message, or any other draft communication that includes fields) is presented to a user, one or more of the fields of the draft are pre-populated with information from or related to the task item. For example, if a draft text message is presented to the user in response to a task item to “send a text message” to a recipient, the “to” field may be pre-populated with a phone number, email address, or other identifier of the recipient. Moreover, the “message” field (e.g., for the body of the text message) may even be pre-populated. As another example, if a draft email is presented to the user, the “to” field may be pre-populated with the email address of the intended recipient, and the “subject” and “message” fields may also be pre-populated if such information is available. The name and/or contact information of the recipient and/or the content for the “subject” and “message” field of a text message or email may be included in the task item. As noted above, such a draft may be presented to a user in response to the user selecting an affordance presented in a notification item, or in response to a triggering criterion being satisfied and without the user separately selecting an affordance.
In some implementations, the first communication event is a telephone call, and causing the notification to be presented comprises causing an audio output relating to the first task item to be produced by the electronic device (1830). The audio output can be any audio, including synthesized or recorded voice output, tones or beeps, music, and the like. In some implementations, the audio output is produced during a telephone-notification period (e.g., while the telephone is ringing) (1832). For example, the audio output (e.g., a synthesized voice output) is produced during the time in which a phone is ringing to alert a user to an incoming telephone call. In some implementations, the ringing of the incoming call is muted or its volume is reduced so that the audio output can be heard by a user. In some implementations, the audio output (e.g., a synthesized voice output) is produced during a ringing period of an outgoing call. For example, a synthesized voice saying “don't forget to ask about Thanksgiving plans” may briefly interrupt the ringing sound of an outgoing call when it is detected that the user is calling his mother. In some implementations, the ringing is muted while the audio is being presented to the user. In some implementations, the volume of the ringing is lowered, but is still audible, while the audio is being presented to the user. In some implementations, if the notification period ends while audio is being presented, such as if the recipient picks up the phone, the audio output is terminated.
In some implementations, once a user initiates a communication event that satisfies a triggering criterion, actual execution of the communication event can be delayed until the audio output is presented. For example, if the user initiates a telephone call that satisfies a triggering criterion (e.g., by entering a phone number and pressing a “call” button or touchscreen element), the audio output is presented to the user before the phone call is actually placed. Accordingly, the audio output (e.g., “don't forget to ask about Thanksgiving plans”) can be presented to the user in its entirety without being interrupted by the recipient answering the phone, and without interrupting the user's conversation.
In some implementations, the audio output is produced during a telephone conversation between a user of the electronic device and a participant of the telephone call (1834). For example, an audio output, such as a tone, beep, or the like, can be output during a telephone conversation to alert a user that a notification has been activated. This can prompt the user to look at a screen of the telephone. Or, a synthesized voice could be output by the device, during the telephone conversation for the user to hear. In some implementations, the audio output is configured so that the other party to the telephone conversation does not hear the output. In some implementations, the audio output is produced at a lower volume than a voice of the participant of the telephone call (1836).
Method 1800 continues on
In some implementations, the triggering criterion requires that the content of a first communication event relate to a predefined subject, and the first communication event is a telephone call (1844). In some implementations, a text string corresponding to one or more utterances spoken during the telephone call by a participant in the telephone call is obtained (1846). In some implementations, it is determined whether one or more words in the text string relate to the predefined subject (1848). In some implementations, NLP service 132 converts voice data obtained from one or more of the participants in a telephone conversation into text. The text is then processed (e.g., by NLP service 132 and/or device 110) to determine whether one or more words in the text string relate to the predefined subject. For example, if a triggering criterion for a notification is “when I talk to Jack about Project Bunny,” a conversation between the user and Jack can be analyzed to determine whether either party says the words “Project Bunny,” or says other words that may be indicative of that subject.
Attention is directed to
A task item associated with a triggering criterion is provided, wherein the triggering criterion requires an occurrence of any of a plurality of communication events, and wherein at least two of the plurality of communication events are of different communication types (1902). In some implementations, the triggering criterion requires that the communication event be a communication with a specified person (1904). In some implementations, the first communication event is associated with the specified person when the specified person initiates the first communication event or receives the first communication event from the user (1906). For example, a communication event may be associated with a specified person regardless of who initiated the communication event (e.g., regardless of who called whom, who emailed whom, etc.). In some implementations, the different communication types include a telephone call and an email (1908). In some implementations, the different communication types include a telephone call and a text message (1910). In some implementations, the different communication types include a telephone call and a video chat (1912).
The occurrence of a first communication event of the plurality of communication events is detected (1914). For example, an incoming or outgoing telephone call, email, text message, or the like is detected. It is determined that the triggering criterion is satisfied (1916). For example, in some implementations, just detecting the occurrence of a first communication event (1914) satisfies the triggering criterion. In some implementations, however, other conditions must be met in order to determine that the triggering criterion is satisfied. For example, as described above, a triggering criteria may be satisfied only upon a determination that content in a communication event relates to a predefined subject. In response to determining that the triggering criterion is satisfied, a notification associated with the task item is caused to be presented to a user of the electronic device (1918). In some implementations, the notification includes information associated with the task item (1920). For example, if the task item includes text (e.g., information) that is to be presented as a reminder (e.g., “ask Mom about Thanksgiving plans”), that text can be presented to the user—either visually, audibly, or both—when the triggering criterion is satisfied.
Attention is directed to
A task item associated with a triggering criterion is provided, wherein the triggering criterion requires an occurrence of a telephone call at an electronic device (2002). In some implementations, the triggering criterion requires that the communication event be a communication with a specified person. The occurrence of the telephone call is detected (2004). In response to detecting the occurrence of the telephone call, a notification associated with the task item is caused to be presented to a user of the electronic device, wherein the notification is separate from a telephone call notification (2006). In some implementations, the notification includes text that is displayed on a screen of a smart phone during or before a telephone call notification (e.g., a ringtone, vibration, or other call notification technique). In some implementations, the notification includes audio content that is presented to the user during or before a telephone call notification.
Referring now to spatial-oriented context triggers, non-limiting examples of spatial-oriented context triggers include the speed at which device 110 is moving (e.g., over 30 mph indicating driving, or less than 3 mph indicating walking), a direction (absolute or relative) at which device 110 is moving, and a set of movements of device 110 (e.g., short vertical movements while moving continuously in a horizontal direction). In other words, device 110 may be configured to detect how device 110 is moving through space.
For example, device 110 (or rather a process executing on device 110) determines, based on detecting changes in its location over a period of time, that device 110 is moving at 60 mph. Based on this information, device 110 determines that the device's context is “while driving” or “on the road.” Task manager 112 analyzes one or more task items to determine whether any task items are associated with a “while driving” or “on the road” context trigger. If a task item is associated with a “while driving” or “on the road” context trigger, then an action (e.g., displaying a notification) associated with the task item is performed.
As another example, device 110 determines, based on detecting changes in its location over a period of time, that device 110 is moving towards his home over a certain period of time (e.g., 5 minutes). Based on this information, device 110 determines that the context is “on my way home.” Task manager 112 analyzes one or more task items to determine whether any task items are associated with a “on my way home” context trigger. If a task item is associated with a “on my way home” context trigger, then an action (e.g., displaying a notification) associated with the task item is performed.
As another example, device 110 includes an accelerator that detects certain repetitive movements. Device 110 may determine, based on these repetitive movements over a period of time, that the user of device 110 might be running at a slow pace. Based on this determination, device 110 determines that the context is “while jogging.” Task manager 112 analyzes one or more task items to determine whether any task items are associated with a “while jogging” or “while walking” context trigger. If a task item is associated with a “while jogging” or “while walking” context trigger, then an action (e.g., displaying a notification) associated with the task item is performed.
As another example, device 110 might detect that it has not moved for a period of time (e.g., 3 hours). A user of device 110 might be interested in being alert and non-movement of device 110 might indicate that the user is asleep. Thus, the user might issue the command, “Alert me if the phone doesn't move for 3 hours.”
In addition to data-oriented and spatial-oriented triggers, other kinds of triggers may be based on any sensor on device 110. Device 110 may include multiple sensors, such as temperature sensors and light sensors. For example, device 110 might include a thermometer for detecting the outside temperature or an internal temperature of device 110. Thus, a user of device 110 might issue the command, “Remind me to call Harold when it reaches 100 degrees.”
Another type of triggering criteria that may be associated with a task item is exogenous criteria. An “exogenous trigger” is a triggering criterion that depends on one or more factors that exist outside and separate from device 110 and the user of device 110. Such factors may be considered “events” that occur with respect to devices other than device 110 or with respect to data that is stored on one or more devices other than device 110. Non-limiting examples of exogenous triggers include social location, social proximity, standing queries, and local events.
An example of a social location trigger is when a friend or associate of the user of device 110 arrives or leaves a certain location. For example, a user command that initiated the creation of a task item may have been “Notify me if Sarah leaves the mall.” Thus, the location of Sarah (or Sarah's mobile device) is an essential factor in setting off this type of exogenous trigger. Specifically, task manager 112 determines the current location of Sarah's device. The current location of Sarah's device may be provided by a cloud service (e.g., in cloud 130) to which both Sarah's device and device 110 are subscribed. Device 110 receives, from the cloud service, updates as to the location of Sarah's device. Task manager 112 uses that location information to determine whether the social location trigger should be activated. A similar user command is “Remind me when my daughter gets home.”
An example of a social proximity trigger is when a friend or associate of the user of device 110 is within a certain distance of the user (or device 110). For example, a user command that initiated the creation of a task item may have been “Remind me to call George when he is within 100 feet of me.” Thus, the location of George (or George's mobile device) is an essential factor in setting off this exogenous trigger. Specifically, task manager 112 or another process executing on device 110 compares the current location of device 110 with the current location of George's device to determine the distance that separates the two devices. Alternatively, George's device may transmit its location to a cloud service to which both George's device and device 110 are subscribed. Device 110 receives, from the cloud service, updates as to a distance between George's device and device 110. Task manager 112 uses that distance information to determine whether the social proximity trigger should be activated.
An example of a standing query trigger is when a webpage mentions a particular term or phrase, such as a company name. To detect this, a standing query is generated and issued continuously (e.g., once a day). For example, a user command that initiated the creation of a task item may have been “Tell me when cnn.com mentions Berkman Industries.” Task manager 112 or another process executing on device 110 issues a search query (e.g., to a search engine) and receives results. When task manager 112 determines that the results include a webpage from cnn.com that includes the name “Berkman Industries,” task manager 112 provides a notification to the user of device 110.
An example of a local event trigger is when a certain local event occurs. To detect this, task manager 112 receives data from an online service. Task manager 112 (or a task service in cloud 130) may periodically send a request to the online service (via one or more communication protocols). Alternatively, task manager 112 may subscribe with the online service to receive information about certain events. For example, a user command that initiated the creation of a task item may have been “Tell me when Beatles tickets go on sale at Shoreline.” In response, task manager 112, another process executing on device 110, or NLP service 132 sends a subscription request to an online ticket service to receive a notification when Beatles tickets for a performance at Shoreline Amphitheatre become available for purchase. When task manager 112 is determines Beatles tickets are available for purchase, task manager 112 provides a notification to the user of device 110.
As another example, a user might be interested in knowing when the surf is up. Thus, the user might issue the command, “Remind me an hour before the surf is up.” Task service 112 (or a task service in cloud 130) might regularly issue a query of a surfing site or might subscribe for alerts from the surfing site.
Based on the foregoing, the types and examples of exogenous triggers are virtually endless. As long as task manager 112 (or a task service in cloud 130) can make a determination about an event that occurs separate from device 110, that event can be used to trigger the performance of an action associated with a task item.
III. Consuming Task Items (Active Payloads)
A task item is “consumed” when an action associated with the task item is performed. Such an action may be a notification that is displayed (or played, if the notification is an audio notification) on device 110. In addition to or instead of providing a notification to a user of device 110, other possible actions include initiating a phone call or a search query, sending an HTTP request (that includes a Uniform Resource Location (URL)), sending an email or a text (SMS) message, causing an application to execute, and causing a purchase to be made on the user's behalf. Such actions that can be associated with task items are referred to as “active payloads.” The processing of an active payload causes some action to be performed, whether by task manager 112 or by another process, whether local or remote to device 110. In other words, instead of simply notifying the user of a task associated with a task item, task manager 112 (or a service in cloud 130) can automate the action part of the task item.
As alluded to above, causing an action to be performed may involve task manager 112 causing another application or process to perform the action. The calling or invoking of the other application (e.g., via an API of the other application) may be performed with or without further input, as indicated in the following examples.
The types of “other” applications can vary greatly. Non-limiting examples of applications that might be available on device 110 include a phone application, an email application, a Web browser, a music player application, a media player application, a music download application, an image processing application, a geopositioning application, a contacts application, an SMS application, a video game application, and a text processing application.
For example, a user of device 110 says aloud, “Remind me to call her back this afternoon.” This voice input is converted into voice data that device 110 sends (along with context data) over network 120 to NLP service 132. NLP service 132 analyzes the voice data and the context data to determine that “her” refers to Marilyn Merlot. NLP service 132 determines that “afternoon” is 2 PM (whether based on context data, a pre-defined setting, or prior history) and determines a phone number for Marilyn Merlot based on a contacts list (or address book), associated with the user, that includes one or more phone numbers for Marilyn Merlot. The contacts list may be stored on device 110 or in cloud 130. NLP 132 sends, to task manager 112 (or to a task service in cloud 130), reminder data used to generate a task item. The reminder data includes the date of “Today”, time of 2 PM, and an instruction to call Marilyn Merlot using a particular phone number. When task manager 112 determines that the current time is 2 PM, task manager 112 may cause a message to be displayed that prompts the user to call Marilyn Merlot. The message may include a “Later” button and a “Call Now” button. If the user selects the “Later” button, then task manager 112 will send the message again later in the afternoon (e.g., in 1 hour). If the user selects the “Call Now” button, then task manager 112 initiates a call to Marilyn Merlot. This initiation may involve task manager 112 making an API call to a phone application (not shown) executing on device 110 and passing the phone number as an argument of the API call. The phone application then uses the phone number to call a device associated with the phone number.
As another example, a user of device 110 says aloud, “Text Lindsay that I love her at 5 o'clock.” This voice input is converted into voice data that device 110 sends over network 120 to NLP service 132. NLP service 132 analyzes the voice data to determine that a cell phone number of Lindsay is necessary and that “5 o'clock” refers to 5 PM of the current day. Task manager 112 (or a task service in cloud 130) creates a task item that includes the following data items: (1) a completion time of 5 PM today, (2) an action of sending a text (or SMS) message, (3) a number of Lindsay's cell phone, and (4) a text string of “I love you” that will be part of the text message. In response to determining that the current time is 5 PM, task manager 112 analyzes the task item to determine the action that needs to be performed. Task manager 112 then causes a text message that includes the text string associated with the task item to be sent to Lindsay's cell phone. This step may comprise task manager 112 invoking an API call of a texting application (not shown) executing on device 110, where the text string (“I love you”) is an argument of the API call.
As another example, a user of device 110 says aloud, “Show me directions on how to get to Rachel's Restaurant in San Jose when I leave the office.” This voice input is converted into voice data that device 110 sends over network 120 to NLP service 132. NLP service 132 analyzes the voice data to determine that a cell phone number of Lindsay is necessary and that “5 o'clock” refers to 5 PM of the current day. Task manager 112 (or a task service in cloud 130) creates a task item that includes the following data items: (1) a location trigger of leaving the user's office and (2) an action of displaying instructions (and, optionally, a map) on how to arrive at Rachel's Restaurant from the user's office. In response to determining that the user of device 110 has left his/her office, task manager 112 analyzes the task item to determine the action that needs to be performed. Task manager 112 then causes (without further input from the user) a travel directions request to be sent to a travel directions service. The travel directions request includes the name of the restaurant, any address information of the restaurant, or both. The travel directions service may be hosted on device 110 or on another device (not shown).
As another example, a user of device 110 says aloud, “Order a cheese only pizza at Pizza Heaven in San Jose, home delivered, 30 minutes before the Bulls-Pacers game starts.” This voice input is converted into voice data that device 110 sends over network 120 to NLP service 132. NLP service 132 analyzes the voice data to determine that a Bulls-Pacers game starts at 6 PM local time; thus, the time trigger is 5:30 PM local time. NLP service 132 also determines that Pizza Heaven in San Jose allows online ordering. Task manager 112 (or a task service in cloud 130) creates a task item that includes the following data items: (1) a time trigger of 5:30 PM and (2) an action of ordering a cheese only pizza from Pizza Heaven with home delivery as an option. In response to determining that the current time is 5:30 PM, task manager 112 (or a task service in cloud 13) analyzes the task item to determine the action that needs to be performed. Task manager 112 then causes a pizza order request to be sent to Pizza Heaven's online ordering service. The pizza order request includes the pizza type of cheese only, the delivery option of home delivery, and the user's home address. The pizza order request may be in the form of an API call to the online ordering service, where arguments of the API call include indications of cheese only topping, home delivery, and the user's home address. Alternatively, before causing the pizza order required to be sent, task manager 112 may formulate a message that is displayed on (or played by) device 110, where the message informs the user about this task. If the user provides affirmative input, then task manager 112 causes the pizza request order to be sent. If the user provides negative input, then no pizza request order is sent.
As another example, a user of device 110 says aloud, “Play my classical station on Pandora at 3 PM tomorrow.” The time of “3 PM tomorrow” coincides with the end of a scheduled exam for the user. This voice input is converted into voice data that device 110 sends over network 120 to NLP service 132. NLP service 132 analyzes the voice data to determine a date and time that the intended action is going to be performed. Task manager 112 (or a task service in cloud 130) creates a task item that includes the following data items: (1) a time trigger of 3 PM with a date that identifies the following day and (2) an action of playing a classical “station” of the user's Pandora music application, where the classical station was established by the user and associated with the user's Pandora account. In response to determining that the current time is 3:00 PM on the proper date, task manager 112 (or a task service in cloud 13) analyzes the task item to determine the action that needs to be performed. Task manager 112 then causes the Pandora music application (not shown) to begin executing on device 110 and to “play” the user's classical station. Task manager 112 may cause the classical station to play by invoking an API call to the Pandora music application, where an argument of the API call includes an indication of a classical station.
In some embodiments, active payloads are processed automatically based on one or more triggering events (e.g., time, location, etc.) without providing reminders, notifications, or otherwise requesting further permission from the user. The device 110 can thus automatically take actions on behalf of the user without the user needing to provide further input. For example, task items can cause the device 110 to send a pre-written email or text message to a recipient in response to a triggering event. A user could say to the device 110 “tell my wife I'm almost there when I get off the freeway.” The device 110 (and/or any associated services such as NLP service 132) can create a task item with an active payload that will send a message (such as a text message or email) saying “I'm almost there” to the user's wife when the user reaches a particular location (e.g., an exit ramp near the user's house). In some embodiments, a recording of the user's voice input corresponding to the message “I'm almost there” can be delivered to the recipient via voicemail, or as an attachment to an email or text message. Also, other triggering events can also cause active payloads to be processed automatically and without user intervention, such as time, date, receipt of messages from others, etc.
The active payloads described above enable the device 110 to initiate certain actions based on a triggering criterion of a task item. For example, if a task item is a reminder to call a co-worker at a 2:00 PM, the device 110 may automatically prompt the user with a “Call Now” button at 2:00 PM. In some cases, though, it is beneficial to configure the task items so that the active payload can be processed without a triggering event. Specifically, a user may create task items that do not have triggers (such as a time or location trigger), or a user may wish to complete a task item before its triggering condition is met. To accommodate this, in some embodiments, an action associated with a task item is performed in response to a user selecting the task item itself. By linking the active payload to the task item (or an input associated with the task item), the active payload can be processed at the user's request without requiring other triggering conditions to be met. Further, it allows a user who is viewing a task list to take advantage of active payloads by simply selecting the task item from the list at their own convenience. For example, a user may view a task list to review his outstanding tasks, and see a task item that he could conveniently complete right away. The user can then simply select the task item (e.g., the description “call Greg Almond”) in order to process the active payload (e.g., initiate a phone call to Greg Almond) and complete the task item.
In order to be selectable by a user, tasks list items may be associated with an activation region that, when selected by a user, causes the action associated with the task to be performed. Activation regions are any area of a graphical display that can be selected by pressing, clicking, touching, or otherwise selecting the area of the display. In some embodiments, the activation region coincides with or overlaps graphical or textual elements that are displayed on a screen (e.g., a touch screen) of the device 110. In some embodiments, the activation region coincides with or overlaps a title or description of the task item, such that when a user selects the displayed text of the title or description, the device 110 performs an action associated with that task item. Task items that may include selectable text in accordance with some embodiments are shown in
In some embodiments, the activation region coincides with or overlaps a graphical element that is displayed on the touch screen in association with the task list item. Graphical elements may be displayed next to, in line with, above, below, or otherwise near the task list item, indicating to the user that the action associated with that task list item will be performed upon selection of the graphical element.
Examples of graphical elements displayed in conjunction with task items are shown in
Similarly, graphic 632 includes a graphical depiction of a letter, which can be made user-selectable by overlaying an activation region on the graphic 632. When the graphic 632 is selected by a user (e.g., by pressing, tapping, touching, etc.), the device 110 will perform actions associated with sending a text or email message to the recipient “Pablo Marc.” For example, the device 110 may display a prompt with a text input region to a user, in which the user can compose a message addressed to Pablo Marc.
In some embodiments, a user can select a task item using a voice input. Voice activation may be used alone or in conjunction with activation regions as described above. In some embodiments, a task item is selected when the device 110 detects a spoken utterance that matches a title, description, or indicator of the task item. For example, a user viewing the task list in
In some embodiments, when a task item is selected, an action associated with that task is performed without further user intervention. For example, if the user selects the task item “call John Appleseed,” the device 110 will initiate a telephone call to John Appleseed. In some embodiments, selection of a task item causes a prompt to be displayed to the user, asking the user for confirmation that the device 110 should perform an action associated with that task item. This confirmation or permission step can help prevent inadvertent actions if a user selects a task list item by mistake. For example, upon selection of the task item “call John Appleseed,” the device 110 may display a prompt with selectable options of “Call Now” or “Cancel.” Similarly, upon selection of the task item “reply to Pablo Marc,” the device 110 may provide an empty text composition area (where the user can compose the body of an email or text message to be sent to Pablo Marc) and selectable options of “Send” or “Cancel.”
Attention is directed to
In some embodiments, the task item is not associated with a triggering criterion (e.g., a time or location trigger) that determines when to trigger a notification to complete the task, or when the action is to be performed. The task item “call Greg Almond” in
The identified action can be any action that the device 110 can perform, including those actions discussed above in reference to active payloads. In some embodiments, the action is causing an email or a text message (e.g., including a previously drafted message) to be sent over a network to a recipient that is associated with the task item. In some embodiments, the action is displaying a text composition area. In some embodiments, text input by a user into the text composition area is then sent as an email or a text message over a network to a recipient that is associated with the task item. In some embodiments, the text composition area is associated with an email address or telephone number of a recipient associated with the task item. In some embodiments, the action is initiating a phone call to a device associated with a phone number that is associated with the task item. In some embodiments, the action is initiating a search based on a search query or URL that is associated with the task item.
In some embodiments, causing an action to be performed (such as one of the actions described above) comprises a first process causing a second process to perform the first action. In some embodiments, the first process causing the second process to perform the action comprises the first process invoking an API call of the second process, wherein the API call comprises one or more attributes associated with the task item
The method 1700 further includes displaying, in a task list, the task item and an associated activation region (1706). (See also
The method 1700 further includes receiving a user selection of the activation region (1708). The user may select an activation region by clicking, touching, pressing, and the like. In some embodiments, the method 1700 includes after receiving a user selection of the activation region and prior to performing the action, displaying a prompt requesting permission to perform the action (1710), and receiving a permission input from the user (1712). Requiring a user to read and/or respond to a prompt with permission to proceed can help prevent the device 110 from taking actions that the user did not intend.
The method 1700 further includes performing the action in response to receiving the user selection of the activation region (1714). Some examples of actions that may be performed are described above with reference to step (1704).
In some embodiments, the method 1700 further includes, in response to performing the action, marking the task item as complete (1724).
In an embodiment, task manager 112 (or a task service in cloud 130) “marks” the task item as complete in response to detecting that a task item is consumed. In other words, a task item may be associated with a complete or an incomplete status. Task manager 112 may provide an interface for a user to view task items managed by task manager 112 and determine whether a task item is complete or not. Task manager 112 may provide an option for a user of device 110 to view all completed task items. The completed task items may be ordered based on when the task items were created, consumed (or completed), or some other criteria.
In some embodiments, the task manager 112 (or another component of the device 110) marks a task item as complete once it has detected that the task has actually been performed. In some embodiments, if a task item has an active payload, the task manager 112 determines that a task as been performed when an action associated with the active payload is performed. For example, if a task item is a reminder to make a telephone call to a person, the task manager 112 marks the task item as complete when, after the task item is selected by a user, the device 110 initiates a telephone to that person.
In some embodiments, task items are marked as complete based on one or more rules that identify that a task has been or is likely to have been performed. In particular, the task manager 112 may be able to monitor various aspects of the device 110, such as network communication traffic (e.g., voice, internet protocol, etc.), API calls, and the like, to determine whether tasks have been performed. In one example, if a task item is a reminder to send an email or text message, the task manager 112 may detect whether, after selection of the task item, an email or text message is actually sent by the device 110. Once the task manager 112 detects that the email or text message has been sent, the task item related to that action is marked as complete.
In another example, if a task item is a reminder to call someone, the task manager 112 may detect whether a telephone call has been initiated. In some cases, this requires not only detecting whether a telephone call was initiated, but also whether the telephone call was successful. For example, a user may initiate a telephone call, but the other party may not pickup, or the call may go to voicemail. In these cases, it may be inappropriate to mark the task as complete. Accordingly, several different aspects of the device 110 may be monitored in order to determine whether a telephone call task item has been successfully completed.
In some embodiments, the task manager 112 monitors a communication interface (e.g., communication interface 1618) of the device 110 to detect whether a telephone call has been initiated after the user has selected the task item. In some embodiments, the task manager 112 detects whether the telephone call results in a successful connection. In some embodiments, the task manager 112 monitors for voice input from the user that is characteristic of a telephone call. The task manager 112 can use any of this information (and/or other information), alone or in combination, to determine whether the phone call was successfully completed.
In some embodiments, the task manager 112 can also determine that a task was not successfully completed in order to prevent the marking of a task item as complete. Specifically, a user may take certain actions on the device 110 that are inconsistent with the completion of a recently selected task item. For example, if a user selects a task item that initiates a phone call, but selects an “end call” button after a short amount of time (or before the call is answered by a recipient), the task manager 112 can determine that the task item should not be marked as complete.
Indeed, any type of cancellation of a process that is necessary for completion of a task may be detected in order to determine that a task has not been completed. In some embodiments, device 110 includes one or more control elements that are configured to cancel current operations, or that have an effect of cancelling certain operations. The control element may be an activation region of a touch screen, a power button, a control button, a switch, or the like. Button 634 (
In some embodiments, if a user activates (e.g., by touching, clicking, pressing, etc) a particular control element after a task item is selected, but before the task is completed, the task manager 112 will recognize that the task item should not be marked as complete. Returning to the email example from above, if a task item is selected that causes the display of a text input area, and the user selects a control element that has the effect of terminating the message before it is sent, the task manager 112 will recognize that the task (i.e., sending an email to a particular recipient) was not completed.
Detecting events that indicate both task completion and task non-completion allows the task manager 112 to accurately and automatically determine whether task items should be marked as complete. This helps to further automate the task list functions of the device 110, as the user does not need to return to the task list and manually mark task items as complete. Moreover, it prevents the task manager 112 from marking task items as complete when they should not be.
Additionally or alternatively, task items that are consumed (i.e., completed) are deleted from storage. For example, task manager 112 deletes, from storage on device 110, any task items that have been consumed. The deletion of a task item may occur a certain period of time (e.g., 1 month) after the corresponding task has been completed to allow a user of device 110 to review recently-consumed task items. If a task service in cloud 130 manages task items that are stored in cloud 130, then that task service may delete consumed task items.
Attention is directed to
In some embodiments, the method 1700 includes detecting one or more events associated with the performing of the action (1716), and determining whether the one or more events satisfy one or more conditions indicating that the action has been completed (1718). The different types of actions are described above with reference to step (1704).
In some embodiments, the action is sending an email, and the one or more conditions are satisfied upon determining that an email has been sent by the electronic device. In some embodiments, the action is initiating a telephone call, and the one or more conditions are satisfied upon receiving a telephone call initiation request. In some embodiments, the action is initiating a telephone call, and the one or more conditions are satisfied upon detecting a telephone call connection event. In embodiments where the action is initiating a telephone call, the one or more conditions may be satisfied upon detecting a voice input that is characteristic of a telephone call.
In some embodiments, the method 1700 further includes detecting one or more events associated with a failure to perform the action (1720), and determining whether the one or more events satisfy one or more conditions indicating that the action has not been completed (1722). This helps to ensure that tasks are not marked as complete where they were aborted after the task item was selected but before it was able to be completed. In some embodiments, the one or more conditions are satisfied upon receiving a user selection of a control element on the electronic device. In some embodiments, the control element is selected from the group consisting of an activation region of a touch screen, a power button, a control button, and a switch. In some embodiments where the action is initiating a telephone call, the one or more conditions include not detecting a voice input within a predetermined time after selection of the activation region.
In an embodiment, when a task item is created, only some details of the corresponding task may be known and stored in association with the task item. Other details regarding the description, address (if any), trigger, and/or action may be determined later, whether automatically or via a manual process.
For example, device 110 sends, to NLP service 132, voice data that reflects a user command to “Call Sarah at 5.” NLP service 132 determines that 5 PM of the current day is a time trigger and causes task manager 112 (or a task service in cloud 130) to create a task item with that time trigger. However, an action item associated with the task item is “Call Sarah” without any indication of a phone number. NLP service 132 has not yet determined who Sarah is and, thus, what phone number to use to call her. Instead, those details are determined later; for example, when the current time is 5 PM and the action is triggered or sometime before the trigger activates. At 5 PM, task manager 112 sends the action item “Call Sarah” (whether in text form or audio form) to NLP service 132 or another service to identify information about a particular Sarah (if there are many) and to determine a phone number for Sarah. When a phone number for Sarah is determined, task manager 112 (or another process) causes a phone application on device 110 to initiate a call using the phone number. In this example, the disambiguation of (a) the identity of an individual and (b) a phone number for that individual is delayed until after the task item is generated.
As another example, device 110 sends, to NLP service 132, voice data that reflects a user command to “Check the weather in San Jose tomorrow morning.” NLP service 132 determines that 7 AM of the next day is a time trigger and causes task manager 112 (or a task service in cloud 130) to create a task item with that time trigger. However, an action item associated with the task item is “Check the weather in San Jose” without any indication of how to perform the action. NLP service 132 has not yet interpreted that portion of the user command to determine how the weather in San Jose is to be checked. Instead, those details are determined later; for example, when the current time is 7 AM of the next day and the action is triggered or sometime before the trigger activates. At 7 AM of the next day, task manager 112 sends the action item “Check the weather in San Jose” (whether in text form or audio form) to NLP service 132 or another service to identify how the weather in San Jose is to be checked. In response, NLP service 132 or another service retrieves information about the weather in San Jose and provides that information to device 110 to be displayed. In this example, the determination of how the action is to be performed is delayed until after the task item is generated.
As alluded to previously, for a task item that is associated with an action that is more than a mere notification, instead of performing the action, a user of device 110 is first alerted of a task and the user is allowed to respond with an affirmative or negative response. For example, an action of a task item is to email Jane Smith about Project Knuckles. Task manager 112 causes, to be displayed on device 110, a message that indicates that the user of device 110 is suppose to email Jane Smith. The user may press a physical or graphical button that indicates an affirmative response. Alternatively, the user may speak the command, “Do it” or “Yes” indicating an affirmative response. In response to the input (whether via a touch screen of device 110, a keyboard selection, or voice input), task manager 112 causes an email application on device 110 to compose an email message addressed to Jane Smith with a subject line that refers to Project Knuckles. Alternatively, the user may decide to be reminded later of the task to email Jane Smith. Thus, in response to the notification, the user provides input (via device 110) that indicates that s/he would like to email Jane Smith some time later, such as in one hour or the next day. Such input may be the user saying “Remind me later” or simply “later.”
In an embodiment, when the action is to respond to an act of communication such as an email message, task manager 112 stores the context of the communication at the time of task creation and retrieves the context at the time of performing the action. The context of communication might be, in various embodiments, a Universal Resource Identifier or other reference to the context or a copy of the data of the context. For example, task manager 112 stores a reference to or copy of the email message that is to be replied to. When the action is performed, the contents of the email message can be recreated just as if the user had performed a reply when initially reading it. Other examples of context data that can be stored and retrieved in this manner include without limitation text messages, documents, web pages, voicemail messages, photographs, audio recordings, and videos.
As another example, an action of a task item is to call George Burt. In response to determining to trigger the action to call, task manager 112 provides an indication that a reminder is available for a user of device 110. The indication may be device 110 buzzing/shaking, generating an audible noise, and/or displaying a notification message. Without holding device 110, the user says aloud, “Read it.” In response to task manager 112 (or another process) processing this input, device 110 plays an audible version of the following statement: “Reminder . . . call George Burt.” The audible version may be based on a playback of the original input from the user or may reflect a computer-generated voice. If the user decides to call George Burt, then the user may simply say, “Okay” or “Do it,” which causes a phone application on device 110 to call George Burt. If the user decides not to call George Burt, then the user may say, “Ignore” or “remind me later.”
IV. Organizing Task Items Using Lists
According to an embodiment of the invention, a task item may be associated with one or more lists. A list is a set of one or more task items that are associated with (or belong to) the same category. Lists are ways that a user of device 110 can view task items in an organized way. The different lists allow the user to intelligently and intuitively browse the tasks that s/he would like to perform (or have performed on his/her behalf).
When a new task item is created, task manager 112 (or a service in cloud 130) identifies one or more attributes associated with the new task item and assigns the new task item to one or more lists. For example, if the new task item includes the action “to call,” then task manager 112 (or other process) adds the new task item to a To Call list. Similarly, if the new task item includes a certain context and a particular location, then task manager 112 might identify the context and/or the particular location and add the new task item to a location list and/or a context list. Alternatively, a user might manually identify one or more of the lists, which are described in detail below, to which a new task item is to be added.
Lists may be characterized as one of three types: built-in or predefined list, smart list, or custom list. Today list 510, All To Do list 520, and Completed list 590 are examples of built-in or pre-defined lists.
Smart lists are based on different characteristics or attributes that a task item might have, such as an action (e.g., call, email, text, alert), a location, and/or a context in which the action is to be performed. Examples of smart lists include By Action lists, By Location lists, and By Context lists. In Car list 540, To Call list 550, and To Email list 560 are examples of By Action lists. Other examples of By Actions lists might include a To Text list, a To Lookup list, and a To Visit list.
Examples of custom lists include lists that are based on categories identified by NLP service 132 and lists that are created by a user. Groceries list 570 and To Buy list 580 are examples of custom lists. Another example of a custom list is a wine list (not shown) that includes a list of the user's favorite wines.
Returning to the lists depicted in
All Lists view 500 also includes a “+” image that when selected, allows a user of device 110 to create another custom list so that current and/or future task items can be added thereto.
For example, the third task item in section 610 is to “pick up Chloe” at 5:00 PM. The icon to the right of that description is an image of a compass, indicating that the action associated with this task item is to generate travel directions to help guide the user of device 110 to the intended destination, which is Pinewood School in this example.
As another example, the second task item in section 620 is to “call John Appleseed.” The icon to the right of that description is an image of a phone, indicating that the action associated with this task item is to call John Appleseed. The image adjacent to the phone image is of a car, indicating that the user of device 110 is to call John Appleseed when the user is in a car or while the user is traveling.
As another example, the last task item in section 620 is to “reply to Pablo Marc.” The icon to the right of that description is an image of an envelope, indicating that the action associated with this task item is to send an email to Pablo Marc. View 600 also indicates that this task item is overdue, or rather, that the originally-scheduled time to email Pablo Marc has passed.
Description item 710 contains a high-level description of the task (“Call John Appleseed”) and includes details about the subject matter (“Discuss the almond deal”). Selection of description item 710 may allow a user of device 110 to edit the description.
Action item 720 contains a description of the action (“Call”) and includes which phone (“mobile”) of John Appleseed to use. Selection of action item 720 may allow the user of device 110 to view the phone number associated with John Appleseed and/or provide other contact options, such as another phone number associated with John Appleseed, an email address of John Appleseed, etc. Furthermore, selection of the phone icon in action item 720 may cause task manager 112 to initiate a call phone to John Appleseed right then instead of waiting for the one or more triggering criteria associated with the task item to be satisfied.
Reminder item 730 indicates the type of trigger (“when in car”) that, when detected, will cause the action to be performed, or at least an alert about the task. Selection of reminder item 730 may allow a user to change the type of reminder.
List assignment item 740 indicates the list to which the task item belongs, which is the “Nut to Crack Project” list in this example. This list is an example of a customized list. Selection of list assignment item 740 may cause device 110 to display multiple task items that belong to the “Nut to Crack Project” list.
Some of the task items referenced in view 800 have been completed. Such completed task items are shown with a lighter gray image to the left of the corresponding description. Task items that have been completed may be distinguished from not-yet-completed task items by other techniques, such as check marks.
In the example depicted in
Each location indicated in view 900 is associated with a different location list. Each location list may be associated with one or more task items. For example, the “Home” location may be associated with four task items (which may be displayed on user selected of the “Home” location) while the “Atherton Dry Cleaning” location may be associated with just one task item.
Because the locations indicated in view 900 are ordered based on distance from the current location of device 110, when the current location of device 110 changes, the location indicators may be re-ordered, some may be removed from view 900, and others not currently displayed in view 900 may appear in view 900. For example, if device 110 is currently located in a store that is next to the Whole Foods store identified by the second location indicated in view 900, then, if device 110 displays view 900, that Whole Foods location indicator will be at the top of the list.
As indicated above, view 900 includes a “Home” location and a “Work” location. The association of a location labeled “Home” (or “Work”) with a particular address may be made in numerous ways. For example, many mobile devices store profile information about a user of the mobile device. This information is referred to as a “me card.” A me card typically stores a user's home address and the user's work address. Thus, task manager 112 (or another process) analyzes the me card that is stored on device 110 to determine a home address and a work address (if any) of the user.
In an embodiment, a radius is associated with a particular location and any task items that are associated with a location that is within the distance indicated by the radius is considered to be associated with the particular location. For example, a radius associated with a home of a user of device 110 is 2 miles. If a task item is associated with a park and the park is within 2 miles from the home, then the task item is associated with a “home” list, along with other task items that are associated with the home.
As noted previously, a location list is an example of a smart list. In an embodiment, any task item that is associated with a location (e.g., as part of the one or more triggering criteria) is automatically associated with a location list that is associated with the same location as the location of the task item. Task manager 112 (or a task service in cloud 130) may maintain multiple location lists.
Location List view 1000 also includes a map icon 1002 which, when selected, causes task manager 112 to communicate with a map application that generates a map of the location associated with the map icon. In this example, a map of the user's home would be generated.
The grocery items identified in Location List view 1050 was associated with the Whole Foods grocery list in response to input from a user of device 110. For example, a user spoke the following command: “Add almond milk to my grocery list” or “Remember to pick up almond milk at Whole Foods near my house.” Device 110 transmits voice data that reflects this command to NLP service 132. NLP service 132 determines, based on the voice data, that the user intends to purchase almond milk. NLP service 132 may cause task manager 112 to (a) create a task item for the task of purchasing almond milk and add the task item to the Whole Foods list or (b) simply add “almond milk” to the Whole Foods list.
Location List view 1050 also includes a map icon 1052 which, when selected, causes task manager 112 to communicate with a map application that generates a map of the location associated with the map icon. In this example, a map of the Whole Foods store identified by the displayed address would be generated.
As noted previously, By Location lists, By Action lists, and By Context lists are examples of smart lists.
View 1100 contains task items that are associated with tasks that are to be performed in a specific context, i.e., the “In Car” context. The task items in the In Car list may be associated with different actions, such as calling and getting directions.
In contrast, view 1110, depicted in
View 1120 indicates two data items that are contained in (or associated with) the “call Bob” task item: a description item and an action item. The action item indicates that multiple contacts are known as “Bob.” As a result, the action item includes a call button that is disabled, whereas the call buttons associated with the other task items in view 1110 are not disabled. Selection of the action item may initiate a process for disambiguating the identity of“Bob.” For example, selection of the action item may cause task manager 112 to display a list of names, each of which have the name of Bob or Robert. In this way, the disambiguation of an identity or of a phone number may occur much later than the creation of the corresponding task item.
View 1130, depicted in
In an embodiment, an “email” task item is created from an email application that is separate from task manager 112. The email application may invoke an API call of task manager 112 to create a task item whose action is to email, where the action includes an active payload that includes an email address and a subject.
As noted previously, custom lists are one of the three main types of lists, including built-in lists and smart lists. Examples of custom lists indicated above include Grocery list 570 and To Buy list 580 (referenced in
Alternatively, NLP service 132 may determine, based on input data (whether voice or text) received from device 110, a specific list to associate with a task item. For example, voice data may reflect a user command to “I need to write a proposal for the Nut to Crack Project.” NLP service 132 determines that “write a proposal” is the task and that “Nut to Crack Project” is the name of a list, which task manager 112 may or may not have yet created. NLP service 132 then sends, to task manager 112, the description (“write proposal”) and the name of a possible list to which the to-be-created task item may be added (“Nut to Crack Project”). Task manager 112 determines whether there is a list that has the same or similar name as “Nut to Crack Project.” If so, then task manager 112 creates a new task item and associates the task item with that list. If not, then task manager 112 creates a new list with that name, creates a new task item, and associates that task item with the new list.
As noted previously, a list may contain items that are not tasks. Such “non-task” are referred to as “notes” that consist only of a description.
Also as noted previously, NLP service 132 may be configured to recognize list names so that task manager 112 can easily assign tasks and notes to the appropriate list(s).
In an embodiment, calendar events created in the context of a calendar application are used to create task items that are managed by task manager 112. The calendar application may be part of task manager 112 or may be separately executing applications. For example, the calendar application might be configured to send newly-created calendar events to task manager 112, e.g., via one or more API calls that cause task manager 112 to create a task item based on the details of a calendar event, such as a description, a date, a location (if any), a duration (if any), and a reminder (if any). Alternatively, task manager 112 might provide a calendar service that allows a user to view a calendar and create events that are associated with a specific date and time or set of dates. Upon creation of events, task manager 112 also creates task items for the events.
In either scenario, if a calendar event that is created and maintained by the calendar service is associated with a location, then a task item that is generated based on the calendar event might also be associated with the location. In that case, task manager 112 might automatically associate the task item with a location list, such as the location list in view 1000 of
While the foregoing description includes four main approaches (generating task items, organizing task items, triggering notifications, and consuming task items), each of these approaches may be implemented individually or may be used together, as noted in many of the examples. For example, natural language processing may be used to generate a task item, but none of the approaches described herein for processing the task item (i.e., organizing the task item, triggering a notification, and consuming the task item) are used. As another example, natural language processing may be used to generate a task item and an approach for organizing the task item as described herein may be used, but none of the approaches for triggering a notification or consuming the task item described herein are used. As another example, none of the approaches for generating and organizing task items and triggering a notification is used, but the approach for consuming the task item as described herein is used.
According to one embodiment, the techniques described herein are implemented by one or more special-purpose computing devices. The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, portable computer systems, handheld devices, networking devices or any other device that incorporates hard-wired and/or program logic to implement the techniques.
For example,
Computer system 1600 also includes a main memory 1606, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 1602 for storing information and instructions to be executed by processor 1604. Main memory 1606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 1604. Such instructions, when stored in non-transitory storage media accessible to processor 1604, render computer system 1600 into a special-purpose machine that is customized to perform the operations specified in the instructions.
Computer system 1600 further includes a read only memory (ROM) 1608 or other static storage device coupled to bus 1602 for storing static information and instructions for processor 1604. A storage device 1610, such as a magnetic disk or optical disk, is provided and coupled to bus 1602 for storing information and instructions.
Computer system 1600 may be coupled via bus 1602 to a display 1612, such as a cathode ray tube (CRT), for displaying information to a computer user. An input device 1614, including alphanumeric and other keys, is coupled to bus 1602 for communicating information and command selections to processor 1604. Another type of user input device is cursor control 1616, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 1604 and for controlling cursor movement on display 1612. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
Computer system 1600 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 1600 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 1600 in response to processor 1604 executing one or more sequences of one or more instructions contained in main memory 1606. Such instructions may be read into main memory 1606 from another storage medium, such as storage device 1610. Execution of the sequences of instructions contained in main memory 1606 causes processor 1604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
The term “storage media” as used herein refers to any non-transitory media that store data and/or instructions that cause a machine to operation in a specific fashion. Such storage media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 1610. Volatile media includes dynamic memory, such as main memory 1606. Common forms of storage media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge.
Storage media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 1602. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
Various forms of media may be involved in carrying one or more sequences of one or more instructions to processor 1604 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 1600 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 1602. Bus 1602 carries the data to main memory 1606, from which processor 1604 retrieves and executes the instructions. The instructions received by main memory 1606 may optionally be stored on storage device 1610 either before or after execution by processor 1604.
Computer system 1600 also includes a communication interface 1618 coupled to bus 1602. Communication interface 1618 provides a two-way data communication coupling to a network link 1620 that is connected to a local network 1622. For example, communication interface 1618 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 1618 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented. In any such implementation, communication interface 1618 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
Network link 1620 typically provides data communication through one or more networks to other data devices. For example, network link 1620 may provide a connection through local network 1622 to a host computer 1624 or to data equipment operated by an Internet Service Provider (ISP) 1626. ISP 1626 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 1628. Local network 1622 and Internet 1628 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 1620 and through communication interface 1618, which carry the digital data to and from computer system 1600, are example forms of transmission media.
Computer system 1600 can send messages and receive data, including program code, through the network(s), network link 1620 and communication interface 1618. In the Internet example, a server 1630 might transmit a requested code for an application program through Internet 1628, ISP 1626, local network 1622 and communication interface 1618.
The received code may be executed by processor 1604 as it is received, and/or stored in storage device 1610, or other non-volatile storage for later execution.
In the foregoing specification, embodiments of the invention have been described with reference to numerous specific details that may vary from implementation to implementation. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. The sole and exclusive indicator of the scope of the invention, and what is intended by the applicants to be the scope of the invention, is the literal and equivalent scope of the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction.
The present application is a continuation of U.S. patent application Ser. No. 13/729,597, filed Dec. 28, 2012, entitled “Active Transport Based Notifications” which is a Continuation-in-Part of U.S. patent application Ser. No. 13/251,118, filed Sep. 30, 2011, entitled “Performing Actions Associated with Task Items that Represent Tasks to Perform,” which in turn claims priority to U.S. Provisional Patent Application No. 61/493,201, filed Jun. 3, 2011, entitled “Generating and Processing Data Items that Represent Tasks to Perform.” These applications are incorporated by reference for all purposes as if fully set forth herein. This application is related to U.S. patent application Ser. No. 12/479,477, filed Jun. 5, 2009, entitled “Contextual Voice Commands,” the entire contents of which are hereby incorporated by reference as if fully set forth herein. This application is related to U.S. patent application Ser. No. 12/987,982, filed Jan. 10, 2011, entitled “Intelligent Automated Assistant,” the entire contents of which are hereby incorporated by reference as if fully set forth herein. This application is related to U.S. patent application Ser. No. 13/251,088, filed Sep. 30, 2011, entitled “Generating and Processing Task Items that Represent Tasks to Perform,” the entire contents of which are hereby incorporated by reference as if fully set forth herein. This application is related to U.S. patent application Ser. No. 13/251,104, filed Sep. 30, 2011, entitled “Triggering Notifications Associated with Tasks Items that Represent Task to Perform,” the entire contents of which are hereby incorporated by reference as if fully set forth herein. This application is related to U.S. patent application Ser. No. 13/251,127, filed Sep. 30, 2011, entitled “Organizing Task Items that Represent Tasks to Perform,” the entire contents of which are hereby incorporated by reference as if fully set forth herein.
Number | Name | Date | Kind |
---|---|---|---|
5530861 | Diamant et al. | Jun 1996 | A |
5546538 | William et al. | Aug 1996 | A |
6081786 | Barry et al. | Jun 2000 | A |
6177905 | Welch | Jan 2001 | B1 |
6526351 | Whitham | Feb 2003 | B2 |
6680675 | Suzuki | Jan 2004 | B1 |
6795710 | Creemer | Sep 2004 | B1 |
6957076 | Hunzinger | Oct 2005 | B2 |
7084758 | Cole | Aug 2006 | B1 |
7248900 | Deeds | Jul 2007 | B2 |
7302686 | Togawa | Nov 2007 | B2 |
7315809 | Xun | Jan 2008 | B2 |
7315818 | Stevens et al. | Jan 2008 | B2 |
7318020 | Kim | Jan 2008 | B1 |
7319957 | Robinson et al. | Jan 2008 | B2 |
7321783 | Kim | Jan 2008 | B2 |
7322023 | Shulman et al. | Jan 2008 | B2 |
7324833 | White et al. | Jan 2008 | B2 |
7324947 | Jordan et al. | Jan 2008 | B2 |
7328155 | Endo et al. | Feb 2008 | B2 |
7328250 | Wang et al. | Feb 2008 | B2 |
7333998 | Heckerman et al. | Feb 2008 | B2 |
7337108 | Florencio et al. | Feb 2008 | B2 |
7345670 | Armstrong | Mar 2008 | B2 |
7345671 | Robbin et al. | Mar 2008 | B2 |
7349953 | Lisitsa et al. | Mar 2008 | B2 |
7353139 | Burrell et al. | Apr 2008 | B1 |
7356748 | Taleb | Apr 2008 | B2 |
7359493 | Wang et al. | Apr 2008 | B1 |
7359671 | Richenstein et al. | Apr 2008 | B2 |
7359851 | Tong et al. | Apr 2008 | B2 |
7360158 | Beeman | Apr 2008 | B1 |
7362738 | Taube et al. | Apr 2008 | B2 |
7363227 | Mapes-Riordan et al. | Apr 2008 | B2 |
7363586 | Briggs et al. | Apr 2008 | B1 |
7365260 | Kawashima | Apr 2008 | B2 |
7366461 | Brown | Apr 2008 | B1 |
7369984 | Fairweather | May 2008 | B2 |
7369993 | Atal | May 2008 | B1 |
7373291 | Garst | May 2008 | B2 |
7373612 | Risch et al. | May 2008 | B2 |
7376556 | Bennett | May 2008 | B2 |
7376632 | Sadek et al. | May 2008 | B1 |
7376645 | Bernard | May 2008 | B2 |
7378963 | Begault et al. | May 2008 | B1 |
7379874 | Schmid et al. | May 2008 | B2 |
7380203 | Keely et al. | May 2008 | B2 |
7383170 | Mills et al. | Jun 2008 | B2 |
7386110 | Petrunka et al. | Jun 2008 | B2 |
7386438 | Franz et al. | Jun 2008 | B1 |
7386449 | Sun et al. | Jun 2008 | B2 |
7386799 | Clanton et al. | Jun 2008 | B1 |
7389224 | Elworthy | Jun 2008 | B1 |
7389225 | Jensen et al. | Jun 2008 | B1 |
7392185 | Bennett | Jun 2008 | B2 |
7394947 | Li et al. | Jul 2008 | B2 |
7398209 | Kennewick et al. | Jul 2008 | B2 |
7401300 | Nurmi | Jul 2008 | B2 |
7403938 | Harrison et al. | Jul 2008 | B2 |
7403941 | Bedworth et al. | Jul 2008 | B2 |
7404143 | Freelander et al. | Jul 2008 | B2 |
7409337 | Potter et al. | Aug 2008 | B1 |
7409347 | Bellegarda | Aug 2008 | B1 |
7412389 | Yang | Aug 2008 | B2 |
7412470 | Masuno et al. | Aug 2008 | B2 |
7415100 | Cooper et al. | Aug 2008 | B2 |
7415469 | Singh et al. | Aug 2008 | B2 |
7418382 | Maes | Aug 2008 | B1 |
7418389 | Chu et al. | Aug 2008 | B2 |
7418392 | Mozer et al. | Aug 2008 | B1 |
7426467 | Nashida et al. | Sep 2008 | B2 |
7426468 | Coifman et al. | Sep 2008 | B2 |
7427024 | Gazdzinski et al. | Sep 2008 | B1 |
7428541 | Houle | Sep 2008 | B2 |
7430508 | Williamson et al. | Sep 2008 | B2 |
7433869 | Gollapudi | Oct 2008 | B2 |
7433921 | Ludwig et al. | Oct 2008 | B2 |
7436947 | Wadler et al. | Oct 2008 | B2 |
7441184 | Frerebeau et al. | Oct 2008 | B2 |
7443316 | Lim | Oct 2008 | B2 |
7444589 | Zellner | Oct 2008 | B2 |
7447360 | Li et al. | Nov 2008 | B2 |
7447624 | Fuhrmann et al. | Nov 2008 | B2 |
7447635 | Konopka et al. | Nov 2008 | B1 |
7447637 | Grant et al. | Nov 2008 | B1 |
7451081 | Gajic et al. | Nov 2008 | B1 |
7454351 | Jeschke et al. | Nov 2008 | B2 |
7460652 | Chang | Dec 2008 | B2 |
7461043 | Hess | Dec 2008 | B2 |
7467087 | Gillick et al. | Dec 2008 | B1 |
7467164 | Marsh | Dec 2008 | B2 |
7472061 | Alewine et al. | Dec 2008 | B1 |
7472065 | Aaron et al. | Dec 2008 | B2 |
7475010 | Chao | Jan 2009 | B2 |
7475015 | Epstein et al. | Jan 2009 | B2 |
7475063 | Datta et al. | Jan 2009 | B2 |
7477238 | Fux et al. | Jan 2009 | B2 |
7477240 | Yanagisawa | Jan 2009 | B2 |
7478037 | Strong | Jan 2009 | B2 |
7478091 | Mojsilovic et al. | Jan 2009 | B2 |
7478129 | Chemtob | Jan 2009 | B1 |
7479948 | Kim et al. | Jan 2009 | B2 |
7479949 | Jobs et al. | Jan 2009 | B2 |
7483832 | Tischer | Jan 2009 | B2 |
7483894 | Cao | Jan 2009 | B2 |
7487089 | Mozer | Feb 2009 | B2 |
7487093 | Mutsuno et al. | Feb 2009 | B2 |
7490034 | Finnigan et al. | Feb 2009 | B2 |
7490039 | Shaffer et al. | Feb 2009 | B1 |
7493251 | Gao et al. | Feb 2009 | B2 |
7493560 | Kipnes et al. | Feb 2009 | B1 |
7496498 | Chu et al. | Feb 2009 | B2 |
7496512 | Zhao et al. | Feb 2009 | B2 |
7499923 | Kawatani | Mar 2009 | B2 |
7502738 | Kennewick et al. | Mar 2009 | B2 |
7505795 | Lim et al. | Mar 2009 | B1 |
7508324 | Suraqui | Mar 2009 | B2 |
7508373 | Lin et al. | Mar 2009 | B2 |
7516123 | Betz et al. | Apr 2009 | B2 |
7519327 | White | Apr 2009 | B2 |
7519398 | Hirose | Apr 2009 | B2 |
7522927 | Fitch et al. | Apr 2009 | B2 |
7523036 | Akabane et al. | Apr 2009 | B2 |
7523108 | Cao | Apr 2009 | B2 |
7526466 | Au | Apr 2009 | B2 |
7526738 | Ording et al. | Apr 2009 | B2 |
7528713 | Singh et al. | May 2009 | B2 |
7529671 | Rockenbeck et al. | May 2009 | B2 |
7529676 | Koyama | May 2009 | B2 |
7535997 | McQuaide, Jr. et al. | May 2009 | B1 |
7536029 | Choi et al. | May 2009 | B2 |
7536565 | Girish et al. | May 2009 | B2 |
7538685 | Cooper et al. | May 2009 | B1 |
7539619 | Seligman et al. | May 2009 | B1 |
7539656 | Fratkina et al. | May 2009 | B2 |
7541940 | Upton | Jun 2009 | B2 |
7542967 | Hurst-Hiller et al. | Jun 2009 | B2 |
7542971 | Thione et al. | Jun 2009 | B2 |
7543232 | Easton, Jr. et al. | Jun 2009 | B2 |
7546382 | Healey et al. | Jun 2009 | B2 |
7546529 | Reynar et al. | Jun 2009 | B2 |
7548895 | Pulsipher | Jun 2009 | B2 |
7552045 | Barliga et al. | Jun 2009 | B2 |
7552055 | Lecoeuche | Jun 2009 | B2 |
7555431 | Bennett | Jun 2009 | B2 |
7555496 | Lantrip et al. | Jun 2009 | B1 |
7558381 | Ali et al. | Jul 2009 | B1 |
7558730 | Davis et al. | Jul 2009 | B2 |
7559026 | Girish et al. | Jul 2009 | B2 |
7561069 | Horstemeyer | Jul 2009 | B2 |
7562007 | Hwang | Jul 2009 | B2 |
7562032 | Abbosh et al. | Jul 2009 | B2 |
7565104 | Brown et al. | Jul 2009 | B1 |
7565380 | Venkatachary | Jul 2009 | B1 |
7568151 | Bargeron et al. | Jul 2009 | B2 |
7571092 | Nieh | Aug 2009 | B1 |
7571106 | Cao et al. | Aug 2009 | B2 |
7577522 | Rosenberg | Aug 2009 | B2 |
7580551 | Srihari et al. | Aug 2009 | B1 |
7580576 | Wang et al. | Aug 2009 | B2 |
7580839 | Tamura et al. | Aug 2009 | B2 |
7584092 | Brockett et al. | Sep 2009 | B2 |
7584093 | Potter et al. | Sep 2009 | B2 |
7584278 | Rajarajan et al. | Sep 2009 | B2 |
7584429 | Fabritius | Sep 2009 | B2 |
7593868 | Margiloff et al. | Sep 2009 | B2 |
7596269 | King et al. | Sep 2009 | B2 |
7596499 | Anguera et al. | Sep 2009 | B2 |
7596606 | Codignotto | Sep 2009 | B2 |
7596765 | Almas | Sep 2009 | B2 |
7599918 | Shen et al. | Oct 2009 | B2 |
7603349 | Kraft et al. | Oct 2009 | B1 |
7603381 | Burke et al. | Oct 2009 | B2 |
7606444 | Erol et al. | Oct 2009 | B1 |
7609179 | Diaz-Gutierrez et al. | Oct 2009 | B2 |
7610258 | Yuknewicz et al. | Oct 2009 | B2 |
7613264 | Wells et al. | Nov 2009 | B2 |
7614008 | Ording | Nov 2009 | B2 |
7617094 | Aoki et al. | Nov 2009 | B2 |
7620407 | Donald et al. | Nov 2009 | B1 |
7620549 | Di Cristo et al. | Nov 2009 | B2 |
7620894 | Kahn | Nov 2009 | B1 |
7623119 | Autio et al. | Nov 2009 | B2 |
7624007 | Bennett | Nov 2009 | B2 |
7627481 | Kuo et al. | Dec 2009 | B1 |
7630900 | Strom | Dec 2009 | B1 |
7630901 | Omi | Dec 2009 | B2 |
7633076 | Huppi et al. | Dec 2009 | B2 |
7634409 | Kennewick et al. | Dec 2009 | B2 |
7634413 | Kuo et al. | Dec 2009 | B1 |
7634718 | Nakajima | Dec 2009 | B2 |
7634732 | Blagsvedt et al. | Dec 2009 | B1 |
7636657 | Ju et al. | Dec 2009 | B2 |
7640158 | Detlef et al. | Dec 2009 | B2 |
7640160 | Di Cristo et al. | Dec 2009 | B2 |
7643990 | Bellegarda | Jan 2010 | B1 |
7647225 | Bennett et al. | Jan 2010 | B2 |
7649454 | Singh et al. | Jan 2010 | B2 |
7649877 | Vieri et al. | Jan 2010 | B2 |
7653883 | Hotelling et al. | Jan 2010 | B2 |
7656393 | King et al. | Feb 2010 | B2 |
7657424 | Bennett | Feb 2010 | B2 |
7657828 | Lucas et al. | Feb 2010 | B2 |
7657844 | Gibson et al. | Feb 2010 | B2 |
7657849 | Chaudhri et al. | Feb 2010 | B2 |
7660715 | Thambiratnam | Feb 2010 | B1 |
7663607 | Hotelling et al. | Feb 2010 | B2 |
7664558 | Lindahl et al. | Feb 2010 | B2 |
7664638 | Cooper et al. | Feb 2010 | B2 |
7668710 | Doyle | Feb 2010 | B2 |
7669134 | Christie et al. | Feb 2010 | B1 |
7672841 | Bennett | Mar 2010 | B2 |
7672952 | Isaacson et al. | Mar 2010 | B2 |
7673238 | Girish et al. | Mar 2010 | B2 |
7673251 | Wibisono | Mar 2010 | B1 |
7673340 | Cohen et al. | Mar 2010 | B1 |
7676026 | Baxter, Jr. | Mar 2010 | B1 |
7676365 | Hwang et al. | Mar 2010 | B2 |
7676463 | Thompson et al. | Mar 2010 | B2 |
7679534 | Kay et al. | Mar 2010 | B2 |
7680649 | Park | Mar 2010 | B2 |
7681126 | Roose | Mar 2010 | B2 |
7683886 | Willey | Mar 2010 | B2 |
7683893 | Kim | Mar 2010 | B2 |
7684985 | Dominach et al. | Mar 2010 | B2 |
7684990 | Caskey et al. | Mar 2010 | B2 |
7684991 | Stohr et al. | Mar 2010 | B2 |
7689245 | Cox et al. | Mar 2010 | B2 |
7689408 | Chen et al. | Mar 2010 | B2 |
7689409 | Heinecke | Mar 2010 | B2 |
7689412 | Wu et al. | Mar 2010 | B2 |
7689421 | Li et al. | Mar 2010 | B2 |
7693715 | Hwang et al. | Apr 2010 | B2 |
7693717 | Kahn et al. | Apr 2010 | B2 |
7693719 | Chu et al. | Apr 2010 | B2 |
7693720 | Kennewick et al. | Apr 2010 | B2 |
7698131 | Bennett | Apr 2010 | B2 |
7702500 | Blaedow | Apr 2010 | B2 |
7702508 | Bennett | Apr 2010 | B2 |
7703091 | Martin et al. | Apr 2010 | B1 |
7706510 | Ng | Apr 2010 | B2 |
7707026 | Liu | Apr 2010 | B2 |
7707027 | Balchandran et al. | Apr 2010 | B2 |
7707032 | Wang et al. | Apr 2010 | B2 |
7707221 | Dunning et al. | Apr 2010 | B1 |
7707226 | Tonse | Apr 2010 | B1 |
7707267 | Lisitsa et al. | Apr 2010 | B2 |
7710262 | Ruha | May 2010 | B2 |
7711129 | Lindahl et al. | May 2010 | B2 |
7711550 | Feinberg et al. | May 2010 | B1 |
7711565 | Gazdzinski | May 2010 | B1 |
7711672 | Au | May 2010 | B2 |
7712053 | Bradford et al. | May 2010 | B2 |
7716056 | Weng et al. | May 2010 | B2 |
7716216 | Harik et al. | May 2010 | B1 |
7720674 | Kaiser et al. | May 2010 | B2 |
7720683 | Vermeulen et al. | May 2010 | B1 |
7721226 | Barabe et al. | May 2010 | B2 |
7721301 | Wong et al. | May 2010 | B2 |
7724242 | Hillis et al. | May 2010 | B2 |
7724696 | Parekh | May 2010 | B1 |
7725307 | Bennett | May 2010 | B2 |
7725318 | Gavalda et al. | May 2010 | B2 |
7725320 | Bennett | May 2010 | B2 |
7725321 | Bennett | May 2010 | B2 |
7725838 | Williams | May 2010 | B2 |
7729904 | Bennett | Jun 2010 | B2 |
7729916 | Coffman et al. | Jun 2010 | B2 |
7734461 | Kwak et al. | Jun 2010 | B2 |
7735012 | Naik | Jun 2010 | B2 |
7739588 | Reynar et al. | Jun 2010 | B2 |
7742953 | King et al. | Jun 2010 | B2 |
7743188 | Haitani et al. | Jun 2010 | B2 |
7747616 | Yamada et al. | Jun 2010 | B2 |
7752152 | Paek et al. | Jul 2010 | B2 |
7756708 | Cohen et al. | Jul 2010 | B2 |
7756868 | Lee | Jul 2010 | B2 |
7756871 | Yacoub et al. | Jul 2010 | B2 |
7757173 | Beaman | Jul 2010 | B2 |
7757182 | Elliott et al. | Jul 2010 | B2 |
7761296 | Bakis et al. | Jul 2010 | B1 |
7763842 | Hsu et al. | Jul 2010 | B2 |
7774202 | Spengler et al. | Aug 2010 | B2 |
7774204 | Mozer et al. | Aug 2010 | B2 |
7774388 | Runchey | Aug 2010 | B1 |
7777717 | Fux et al. | Aug 2010 | B2 |
7778432 | Larsen | Aug 2010 | B2 |
7778595 | White et al. | Aug 2010 | B2 |
7778632 | Kurlander et al. | Aug 2010 | B2 |
7778830 | Davis et al. | Aug 2010 | B2 |
7779353 | Grigoriu et al. | Aug 2010 | B2 |
7779356 | Griesmer | Aug 2010 | B2 |
7779357 | Naik | Aug 2010 | B2 |
7783283 | Kuusinen et al. | Aug 2010 | B2 |
7783486 | Rosser et al. | Aug 2010 | B2 |
7788590 | Taboada et al. | Aug 2010 | B2 |
7788663 | Illowsky et al. | Aug 2010 | B2 |
7796980 | McKinney et al. | Sep 2010 | B1 |
7797265 | Brinker et al. | Sep 2010 | B2 |
7797269 | Rieman et al. | Sep 2010 | B2 |
7797331 | Theimer et al. | Sep 2010 | B2 |
7797629 | Fux et al. | Sep 2010 | B2 |
7801721 | Rosart et al. | Sep 2010 | B2 |
7801728 | Ben-David et al. | Sep 2010 | B2 |
7801729 | Mozer | Sep 2010 | B2 |
7805299 | Coifman | Sep 2010 | B2 |
7809550 | Barrows | Oct 2010 | B1 |
7809565 | Coifman | Oct 2010 | B2 |
7809569 | Attwater et al. | Oct 2010 | B2 |
7809570 | Kennewick et al. | Oct 2010 | B2 |
7809610 | Cao | Oct 2010 | B2 |
7809744 | Nevidomski et al. | Oct 2010 | B2 |
7813729 | Lee et al. | Oct 2010 | B2 |
7818165 | Carlgren et al. | Oct 2010 | B2 |
7818176 | Freeman et al. | Oct 2010 | B2 |
7818215 | King et al. | Oct 2010 | B2 |
7818291 | Ferguson et al. | Oct 2010 | B2 |
7818672 | Mccormack et al. | Oct 2010 | B2 |
7822608 | Cross, Jr. et al. | Oct 2010 | B2 |
7823123 | Sabbouh | Oct 2010 | B2 |
7826945 | Zhang et al. | Nov 2010 | B2 |
7827047 | Anderson et al. | Nov 2010 | B2 |
7831246 | Smith et al. | Nov 2010 | B1 |
7831423 | Schubert | Nov 2010 | B2 |
7831426 | Bennett | Nov 2010 | B2 |
7831432 | Bodin et al. | Nov 2010 | B2 |
7835504 | Donald et al. | Nov 2010 | B1 |
7836437 | Kacmarcik et al. | Nov 2010 | B2 |
7840348 | Kim et al. | Nov 2010 | B2 |
7840400 | Lavi et al. | Nov 2010 | B2 |
7840447 | Kleinrock et al. | Nov 2010 | B2 |
7840581 | Ross et al. | Nov 2010 | B2 |
7840912 | Elias et al. | Nov 2010 | B2 |
7844394 | Kim | Nov 2010 | B2 |
7848924 | Nurminen et al. | Dec 2010 | B2 |
7848926 | Goto et al. | Dec 2010 | B2 |
7853444 | Wang et al. | Dec 2010 | B2 |
7853445 | Bachenko et al. | Dec 2010 | B2 |
7853574 | Kraenzel et al. | Dec 2010 | B2 |
7853577 | Sundaresan et al. | Dec 2010 | B2 |
7853664 | Wang et al. | Dec 2010 | B1 |
7853900 | Nguyen et al. | Dec 2010 | B2 |
7865817 | Ryan et al. | Jan 2011 | B2 |
7869998 | Di Fabbrizio et al. | Jan 2011 | B1 |
7869999 | Amato et al. | Jan 2011 | B2 |
7870118 | Jiang et al. | Jan 2011 | B2 |
7870133 | Krishnamoorthy et al. | Jan 2011 | B2 |
7873149 | Schultz et al. | Jan 2011 | B2 |
7873519 | Bennett | Jan 2011 | B2 |
7873654 | Bernard | Jan 2011 | B2 |
7877705 | Chambers et al. | Jan 2011 | B2 |
7880730 | Robinson et al. | Feb 2011 | B2 |
7881283 | Cormier et al. | Feb 2011 | B2 |
7881936 | Longe et al. | Feb 2011 | B2 |
7885390 | Chaudhuri et al. | Feb 2011 | B2 |
7885844 | Cohen et al. | Feb 2011 | B1 |
7886233 | Rainisto et al. | Feb 2011 | B2 |
7889101 | Yokota | Feb 2011 | B2 |
7889184 | Blumenberg et al. | Feb 2011 | B2 |
7889185 | Blumenberg et al. | Feb 2011 | B2 |
7890330 | Ozkaragoz et al. | Feb 2011 | B2 |
7890652 | Bull et al. | Feb 2011 | B2 |
7895039 | Brahe et al. | Feb 2011 | B2 |
7895531 | Radtke et al. | Feb 2011 | B2 |
7899666 | Varone | Mar 2011 | B2 |
7904297 | Mirkovic et al. | Mar 2011 | B2 |
7908287 | Katragadda | Mar 2011 | B1 |
7912289 | Kansal et al. | Mar 2011 | B2 |
7912699 | Saraclar et al. | Mar 2011 | B1 |
7912702 | Bennett | Mar 2011 | B2 |
7912720 | Hakkani-Tur et al. | Mar 2011 | B1 |
7912828 | Bonnet et al. | Mar 2011 | B2 |
7913185 | Benson et al. | Mar 2011 | B1 |
7916979 | Simmons | Mar 2011 | B2 |
7917367 | Di Cristo et al. | Mar 2011 | B2 |
7917497 | Harrison et al. | Mar 2011 | B2 |
7920678 | Cooper et al. | Apr 2011 | B2 |
7920682 | Byrne et al. | Apr 2011 | B2 |
7920857 | Lau et al. | Apr 2011 | B2 |
7925525 | Chin | Apr 2011 | B2 |
7925610 | Elbaz et al. | Apr 2011 | B2 |
7929805 | Wang et al. | Apr 2011 | B2 |
7930168 | Weng et al. | Apr 2011 | B2 |
7930183 | Odell et al. | Apr 2011 | B2 |
7930197 | Ozzie et al. | Apr 2011 | B2 |
7936339 | Marggraff et al. | May 2011 | B2 |
7936861 | Martin et al. | May 2011 | B2 |
7936863 | John et al. | May 2011 | B2 |
7937075 | Zellner | May 2011 | B2 |
7941009 | Li et al. | May 2011 | B2 |
7945294 | Zhang et al. | May 2011 | B2 |
7945470 | Cohen et al. | May 2011 | B1 |
7949529 | Weider et al. | May 2011 | B2 |
7949534 | Davis et al. | May 2011 | B2 |
7949752 | Lange et al. | May 2011 | B2 |
7953679 | Chidlovskii et al. | May 2011 | B2 |
7957975 | Burns et al. | Jun 2011 | B2 |
7958136 | Curtis et al. | Jun 2011 | B1 |
7962179 | Huang | Jun 2011 | B2 |
7974835 | Balchandran et al. | Jul 2011 | B2 |
7974844 | Sumita | Jul 2011 | B2 |
7974972 | Cao | Jul 2011 | B2 |
7975216 | Woolf et al. | Jul 2011 | B2 |
7983478 | Liu et al. | Jul 2011 | B2 |
7983915 | Knight et al. | Jul 2011 | B2 |
7983917 | Kennewick et al. | Jul 2011 | B2 |
7983919 | Conkie | Jul 2011 | B2 |
7983997 | Allen et al. | Jul 2011 | B2 |
7984062 | Dunning et al. | Jul 2011 | B2 |
7986431 | Emori et al. | Jul 2011 | B2 |
7987151 | Schott et al. | Jul 2011 | B2 |
7987244 | Lewis et al. | Jul 2011 | B1 |
7991614 | Washio et al. | Aug 2011 | B2 |
7992085 | Wang-Aryattanwanich et al. | Aug 2011 | B2 |
7996228 | Miller et al. | Aug 2011 | B2 |
7996589 | Schultz et al. | Aug 2011 | B2 |
7996769 | Fux et al. | Aug 2011 | B2 |
7996792 | Anzures et al. | Aug 2011 | B2 |
7999669 | Singh et al. | Aug 2011 | B2 |
8000453 | Cooper et al. | Aug 2011 | B2 |
8005664 | Hanumanthappa | Aug 2011 | B2 |
8005679 | Jordan et al. | Aug 2011 | B2 |
8006180 | Tunning et al. | Aug 2011 | B2 |
8014308 | Gates et al. | Sep 2011 | B2 |
8015006 | Kennewick et al. | Sep 2011 | B2 |
8015011 | Nagano et al. | Sep 2011 | B2 |
8015144 | Zheng et al. | Sep 2011 | B2 |
8018431 | Zehr et al. | Sep 2011 | B1 |
8019271 | Izdepski | Sep 2011 | B1 |
8019604 | Ma | Sep 2011 | B2 |
8020104 | Robarts et al. | Sep 2011 | B2 |
8024195 | Mozer et al. | Sep 2011 | B2 |
8024415 | Horvitz et al. | Sep 2011 | B2 |
8027836 | Baker et al. | Sep 2011 | B2 |
8031943 | Chen et al. | Oct 2011 | B2 |
8032383 | Bhardwaj et al. | Oct 2011 | B1 |
8036901 | Mozer | Oct 2011 | B2 |
8037034 | Plachta et al. | Oct 2011 | B2 |
8041557 | Liu | Oct 2011 | B2 |
8041570 | Mirkovic et al. | Oct 2011 | B2 |
8041611 | Kleinrock et al. | Oct 2011 | B2 |
8042053 | Darwish et al. | Oct 2011 | B2 |
8046363 | Cha et al. | Oct 2011 | B2 |
8046374 | Bromwich et al. | Oct 2011 | B1 |
8050500 | Batty et al. | Nov 2011 | B1 |
8054180 | Scofield et al. | Nov 2011 | B1 |
8055502 | Clark et al. | Nov 2011 | B2 |
8055708 | Chitsaz et al. | Nov 2011 | B2 |
8056070 | Goller et al. | Nov 2011 | B2 |
8060824 | Brownrigg, Jr. et al. | Nov 2011 | B2 |
8064753 | Freeman | Nov 2011 | B2 |
8065143 | Yanagihara | Nov 2011 | B2 |
8065155 | Gazdzinski | Nov 2011 | B1 |
8065156 | Gazdzinski | Nov 2011 | B2 |
8068604 | Leeds et al. | Nov 2011 | B2 |
8069046 | Kennewick et al. | Nov 2011 | B2 |
8069422 | Sheshagiri et al. | Nov 2011 | B2 |
8073681 | Baldwin et al. | Dec 2011 | B2 |
8073695 | Hendricks et al. | Dec 2011 | B1 |
8077153 | Benko et al. | Dec 2011 | B2 |
8078473 | Gazdzinski | Dec 2011 | B1 |
8082153 | Coffman et al. | Dec 2011 | B2 |
8082498 | Salamon et al. | Dec 2011 | B2 |
8090571 | Elshishiny et al. | Jan 2012 | B2 |
8095364 | Longe et al. | Jan 2012 | B2 |
8099289 | Mozer et al. | Jan 2012 | B2 |
8099395 | Pabla et al. | Jan 2012 | B2 |
8099418 | Inoue et al. | Jan 2012 | B2 |
8103510 | Sato | Jan 2012 | B2 |
8107401 | John et al. | Jan 2012 | B2 |
8112275 | Kennewick et al. | Feb 2012 | B2 |
8112280 | Lu | Feb 2012 | B2 |
8117037 | Gazdzinski | Feb 2012 | B2 |
8117542 | Radtke et al. | Feb 2012 | B2 |
8121413 | Hwang et al. | Feb 2012 | B2 |
8121837 | Agapi et al. | Feb 2012 | B2 |
8122094 | Kotab | Feb 2012 | B1 |
8122353 | Bouta | Feb 2012 | B2 |
8130929 | Wilkes et al. | Mar 2012 | B2 |
8131557 | Davis et al. | Mar 2012 | B2 |
8135115 | Hogg, Jr. et al. | Mar 2012 | B1 |
8138912 | Singh et al. | Mar 2012 | B2 |
8140335 | Kennewick et al. | Mar 2012 | B2 |
8140567 | Padovitz et al. | Mar 2012 | B2 |
8145489 | Freeman et al. | Mar 2012 | B2 |
8150694 | Kennewick et al. | Apr 2012 | B2 |
8150700 | Shin et al. | Apr 2012 | B2 |
8155956 | Cho et al. | Apr 2012 | B2 |
8156005 | Vieri | Apr 2012 | B2 |
8160877 | Nucci et al. | Apr 2012 | B1 |
8160883 | Lecoeuche | Apr 2012 | B2 |
8165321 | Paquier et al. | Apr 2012 | B2 |
8165886 | Gagnon et al. | Apr 2012 | B1 |
8166019 | Lee et al. | Apr 2012 | B1 |
8166032 | Sommer et al. | Apr 2012 | B2 |
8170790 | Lee et al. | May 2012 | B2 |
8175872 | Kristjansson et al. | May 2012 | B2 |
8179370 | Yamasani et al. | May 2012 | B1 |
8188856 | Singh et al. | May 2012 | B2 |
8190359 | Bourne | May 2012 | B2 |
8190596 | Nambiar et al. | May 2012 | B2 |
8195467 | Mozer et al. | Jun 2012 | B2 |
8195468 | Kennewick et al. | Jun 2012 | B2 |
8200489 | Baggenstoss | Jun 2012 | B1 |
8200495 | Braho et al. | Jun 2012 | B2 |
8201109 | Van Os et al. | Jun 2012 | B2 |
8204238 | Mozer | Jun 2012 | B2 |
8205788 | Gazdzinski et al. | Jun 2012 | B1 |
8209183 | Patel et al. | Jun 2012 | B1 |
8213911 | Williams et al. | Jul 2012 | B2 |
8219115 | Nelissen | Jul 2012 | B1 |
8219406 | Yu et al. | Jul 2012 | B2 |
8219407 | Roy et al. | Jul 2012 | B1 |
8219608 | alSafadi et al. | Jul 2012 | B2 |
8224649 | Chaudhari et al. | Jul 2012 | B2 |
8228299 | Maloney et al. | Jul 2012 | B1 |
8233919 | Haag et al. | Jul 2012 | B2 |
8234111 | Lloyd et al. | Jul 2012 | B2 |
8239206 | LeBeau et al. | Aug 2012 | B1 |
8239207 | Seligman et al. | Aug 2012 | B2 |
8244712 | Serlet et al. | Aug 2012 | B2 |
8250071 | Killalea et al. | Aug 2012 | B1 |
8254829 | Kindred et al. | Aug 2012 | B1 |
8255216 | White | Aug 2012 | B2 |
8255217 | Stent et al. | Aug 2012 | B2 |
8260247 | Lazaridis et al. | Sep 2012 | B2 |
8260617 | Dhanakshirur et al. | Sep 2012 | B2 |
8270933 | Riemer et al. | Sep 2012 | B2 |
8271287 | Kermani | Sep 2012 | B1 |
8275621 | Alewine et al. | Sep 2012 | B2 |
8279171 | Hirai et al. | Oct 2012 | B2 |
8280438 | Barbera | Oct 2012 | B2 |
8285546 | Reich | Oct 2012 | B2 |
8285551 | Gazdzinski | Oct 2012 | B2 |
8285553 | Gazdzinski | Oct 2012 | B2 |
8290777 | Nguyen et al. | Oct 2012 | B1 |
8290778 | Gazdzinski | Oct 2012 | B2 |
8290781 | Gazdzinski | Oct 2012 | B2 |
8296124 | Holsztynska et al. | Oct 2012 | B1 |
8296145 | Clark et al. | Oct 2012 | B2 |
8296146 | Gazdzinski | Oct 2012 | B2 |
8296153 | Gazdzinski | Oct 2012 | B2 |
8296380 | Kelly et al. | Oct 2012 | B1 |
8296383 | Lindahl | Oct 2012 | B2 |
8300801 | Sweeney et al. | Oct 2012 | B2 |
8301456 | Gazdzinski | Oct 2012 | B2 |
8311189 | Champlin et al. | Nov 2012 | B2 |
8311834 | Gazdzinski | Nov 2012 | B1 |
8311835 | Lecoeuche | Nov 2012 | B2 |
8311838 | Lindahl et al. | Nov 2012 | B2 |
8312017 | Martin et al. | Nov 2012 | B2 |
8321786 | Lunati et al. | Nov 2012 | B2 |
8326627 | Kennewick et al. | Dec 2012 | B2 |
8332205 | Krishnan et al. | Dec 2012 | B2 |
8332218 | Cross et al. | Dec 2012 | B2 |
8332224 | Di Cristo et al. | Dec 2012 | B2 |
8332748 | Karam | Dec 2012 | B1 |
8335689 | Wittenstein et al. | Dec 2012 | B2 |
8340975 | Rosenberger | Dec 2012 | B1 |
8345665 | Vieri et al. | Jan 2013 | B2 |
8346563 | Hjelm et al. | Jan 2013 | B1 |
8352183 | Thota et al. | Jan 2013 | B2 |
8352268 | Naik et al. | Jan 2013 | B2 |
8352272 | Rogers et al. | Jan 2013 | B2 |
8355919 | Silverman et al. | Jan 2013 | B2 |
8359234 | Vieri | Jan 2013 | B2 |
8370145 | Endo et al. | Feb 2013 | B2 |
8370158 | Gazdzinski | Feb 2013 | B2 |
8371503 | Gazdzinski | Feb 2013 | B2 |
8374871 | Ehsani et al. | Feb 2013 | B2 |
8375320 | Kotler et al. | Feb 2013 | B2 |
8380504 | Peden et al. | Feb 2013 | B1 |
8380507 | Herman et al. | Feb 2013 | B2 |
8381107 | Rottler et al. | Feb 2013 | B2 |
8381135 | Hotelling et al. | Feb 2013 | B2 |
8386485 | Kerschberg et al. | Feb 2013 | B2 |
8386926 | Matsuoka | Feb 2013 | B1 |
8391844 | Lamiraux et al. | Mar 2013 | B2 |
8396714 | Rogers et al. | Mar 2013 | B2 |
8401163 | Kirchhoff et al. | Mar 2013 | B1 |
8406745 | Upadhyay et al. | Mar 2013 | B1 |
8423288 | Stahl et al. | Apr 2013 | B2 |
8428758 | Naik et al. | Apr 2013 | B2 |
8433778 | Shreesha et al. | Apr 2013 | B1 |
8442821 | Vanhoucke | May 2013 | B1 |
8447612 | Gazdzinski | May 2013 | B2 |
8452597 | Bringert et al. | May 2013 | B2 |
8457959 | Kaiser | Jun 2013 | B2 |
8458115 | Cai et al. | Jun 2013 | B2 |
8458278 | Christie et al. | Jun 2013 | B2 |
8464150 | Davidson et al. | Jun 2013 | B2 |
8473289 | Jitkoff et al. | Jun 2013 | B2 |
8479122 | Hotelling et al. | Jul 2013 | B2 |
8484027 | Murphy | Jul 2013 | B1 |
8489599 | Bellotti | Jul 2013 | B2 |
8498857 | Kopparapu et al. | Jul 2013 | B2 |
8514197 | Shahraray et al. | Aug 2013 | B2 |
8515750 | Lei et al. | Aug 2013 | B1 |
8521513 | Millett et al. | Aug 2013 | B2 |
8521531 | Kim | Aug 2013 | B1 |
8527276 | Senior et al. | Sep 2013 | B1 |
8537033 | Gueziec | Sep 2013 | B2 |
8543398 | Strope et al. | Sep 2013 | B1 |
8560229 | Park et al. | Oct 2013 | B1 |
8571851 | Tickner et al. | Oct 2013 | B1 |
8583416 | Huang et al. | Nov 2013 | B2 |
8583511 | Hendrickson | Nov 2013 | B2 |
8589869 | Wolfram | Nov 2013 | B2 |
8589911 | Sharkey et al. | Nov 2013 | B1 |
8595004 | Koshinaka | Nov 2013 | B2 |
8600743 | Lindahl et al. | Dec 2013 | B2 |
8600930 | Sata et al. | Dec 2013 | B2 |
8606568 | Tickner et al. | Dec 2013 | B1 |
8620659 | Di Cristo et al. | Dec 2013 | B2 |
8620662 | Bellegarda | Dec 2013 | B2 |
8626681 | Jurca et al. | Jan 2014 | B1 |
8638363 | King et al. | Jan 2014 | B2 |
8639516 | Lindahl et al. | Jan 2014 | B2 |
8645137 | Bellegarda et al. | Feb 2014 | B2 |
8645138 | Weinstein et al. | Feb 2014 | B1 |
8654936 | Tofighbakhsh et al. | Feb 2014 | B1 |
8655646 | Lee et al. | Feb 2014 | B2 |
8655901 | Li et al. | Feb 2014 | B1 |
8660843 | Falcon et al. | Feb 2014 | B2 |
8660849 | Gruber et al. | Feb 2014 | B2 |
8660970 | Fiedorowicz | Feb 2014 | B1 |
8661112 | Creamer et al. | Feb 2014 | B2 |
8661340 | Goldsmith et al. | Feb 2014 | B2 |
8670979 | Gruber et al. | Mar 2014 | B2 |
8675084 | Bolton et al. | Mar 2014 | B2 |
8676904 | Lindahl et al. | Mar 2014 | B2 |
8677377 | Cheyer et al. | Mar 2014 | B2 |
8681950 | Mack et al. | Mar 2014 | B2 |
8682667 | Haughay et al. | Mar 2014 | B2 |
8687777 | Lavian et al. | Apr 2014 | B1 |
8688446 | Yanagihara et al. | Apr 2014 | B2 |
8688453 | Joshi et al. | Apr 2014 | B1 |
8695074 | Saraf et al. | Apr 2014 | B2 |
8696364 | Cohen | Apr 2014 | B2 |
8706472 | Ramerth et al. | Apr 2014 | B2 |
8706474 | Blume et al. | Apr 2014 | B2 |
8706503 | Cheyer et al. | Apr 2014 | B2 |
8713119 | Lindahl et al. | Apr 2014 | B2 |
8713418 | King et al. | Apr 2014 | B2 |
8719006 | Bellegarda et al. | May 2014 | B2 |
8719014 | Wagner et al. | May 2014 | B2 |
8731610 | Appaji | May 2014 | B2 |
8731912 | Tickner et al. | May 2014 | B1 |
8731942 | Cheyer et al. | May 2014 | B2 |
8739208 | Rodriguez et al. | May 2014 | B2 |
8744852 | Seymour et al. | Jun 2014 | B1 |
8760537 | Johnson et al. | Jun 2014 | B2 |
8762145 | Ouchi et al. | Jun 2014 | B2 |
8762156 | Chen et al. | Jun 2014 | B2 |
8762469 | Lindahl et al. | Jun 2014 | B2 |
8768693 | Lempel et al. | Jul 2014 | B2 |
8768702 | Boettcher et al. | Jul 2014 | B2 |
8775154 | Clinchant et al. | Jul 2014 | B2 |
8775931 | Fux et al. | Jul 2014 | B2 |
8781456 | Prociw | Jul 2014 | B2 |
8781841 | Wang | Jul 2014 | B1 |
8798255 | Lubowich et al. | Aug 2014 | B2 |
8798995 | Edara et al. | Aug 2014 | B1 |
8799000 | Guzzoni et al. | Aug 2014 | B2 |
8805690 | LeBeau et al. | Aug 2014 | B1 |
8812302 | Xiao et al. | Aug 2014 | B2 |
8838457 | Cerra et al. | Sep 2014 | B2 |
8855915 | Furuhata et al. | Oct 2014 | B2 |
8861925 | Ohme | Oct 2014 | B1 |
8862252 | Rottier et al. | Oct 2014 | B2 |
8868111 | Kahn et al. | Oct 2014 | B1 |
8868409 | Mengibar et al. | Oct 2014 | B1 |
8880405 | Cerra et al. | Nov 2014 | B2 |
8886534 | Nakano et al. | Nov 2014 | B2 |
8886540 | Cerra et al. | Nov 2014 | B2 |
8886541 | Friedlander | Nov 2014 | B2 |
8892446 | Cheyer et al. | Nov 2014 | B2 |
8893023 | Perry et al. | Nov 2014 | B2 |
8898568 | Bull et al. | Nov 2014 | B2 |
8903716 | Chen et al. | Dec 2014 | B2 |
8909693 | Frissora et al. | Dec 2014 | B2 |
8930176 | Li et al. | Jan 2015 | B2 |
8930191 | Gruber et al. | Jan 2015 | B2 |
8938394 | Faaborg et al. | Jan 2015 | B1 |
8938688 | Bradford et al. | Jan 2015 | B2 |
8942986 | Cheyer et al. | Jan 2015 | B2 |
8943423 | Merrill et al. | Jan 2015 | B2 |
8972240 | Brockett et al. | Mar 2015 | B2 |
8972432 | Shaw et al. | Mar 2015 | B2 |
8972878 | Mohler et al. | Mar 2015 | B2 |
8983383 | Haskin | Mar 2015 | B1 |
8989713 | Doulton | Mar 2015 | B2 |
8990235 | King et al. | Mar 2015 | B2 |
8994660 | Neels et al. | Mar 2015 | B2 |
8996350 | Dub et al. | Mar 2015 | B1 |
8996376 | Fleizach et al. | Mar 2015 | B2 |
8996381 | Mozer et al. | Mar 2015 | B2 |
8996639 | Faaborg et al. | Mar 2015 | B1 |
9009046 | Stewart | Apr 2015 | B1 |
9020804 | Barbaiani et al. | Apr 2015 | B2 |
9026425 | Nikoulina et al. | May 2015 | B2 |
9031834 | Coorman et al. | May 2015 | B2 |
9037967 | Al-Jefri et al. | May 2015 | B1 |
9043208 | Koch et al. | May 2015 | B2 |
9049255 | MacFarlane et al. | Jun 2015 | B2 |
9049295 | Cooper et al. | Jun 2015 | B1 |
9053706 | Jitkoff et al. | Jun 2015 | B2 |
9058811 | Wang et al. | Jun 2015 | B2 |
9063979 | Chiu et al. | Jun 2015 | B2 |
9070366 | Mathias et al. | Jun 2015 | B1 |
9071701 | Donaldson et al. | Jun 2015 | B2 |
9076448 | Bennett et al. | Jul 2015 | B2 |
9076450 | Sadek et al. | Jul 2015 | B1 |
9081411 | Kalns et al. | Jul 2015 | B2 |
9081482 | Zhai et al. | Jul 2015 | B1 |
9082402 | Yadgar et al. | Jul 2015 | B2 |
9098467 | Blanksteen et al. | Aug 2015 | B1 |
9101279 | Ritchey et al. | Aug 2015 | B2 |
9112984 | Sejnoha et al. | Aug 2015 | B2 |
9117447 | Gruber et al. | Aug 2015 | B2 |
9123338 | Sanders et al. | Sep 2015 | B1 |
9171541 | Kennewick et al. | Oct 2015 | B2 |
9171546 | Pike | Oct 2015 | B1 |
9190062 | Haughay | Nov 2015 | B2 |
9208153 | Zaveri et al. | Dec 2015 | B1 |
9218809 | Bellegarda | Dec 2015 | B2 |
9218819 | Stekkelpak et al. | Dec 2015 | B1 |
9223537 | Brown et al. | Dec 2015 | B2 |
9255812 | Maeoka et al. | Feb 2016 | B2 |
9258604 | Bilobrov et al. | Feb 2016 | B1 |
9262612 | Cheyer | Feb 2016 | B2 |
9286910 | Li et al. | Mar 2016 | B1 |
9292487 | Weber | Mar 2016 | B1 |
9292489 | Sak et al. | Mar 2016 | B1 |
9299344 | Braho et al. | Mar 2016 | B2 |
9300718 | Khanna | Mar 2016 | B2 |
9305543 | Fleizach et al. | Apr 2016 | B2 |
9305548 | Kennewick et al. | Apr 2016 | B2 |
9311912 | Swietlinski et al. | Apr 2016 | B1 |
9313317 | LeBeau et al. | Apr 2016 | B1 |
9318108 | Gruber et al. | Apr 2016 | B2 |
9325809 | Barros et al. | Apr 2016 | B1 |
9330659 | Ju et al. | May 2016 | B2 |
9330720 | Lee | May 2016 | B2 |
9338493 | Van Os et al. | May 2016 | B2 |
9349368 | LeBeau et al. | May 2016 | B1 |
9361084 | Costa | Jun 2016 | B1 |
9367541 | Servan et al. | Jun 2016 | B1 |
9377871 | Waddell et al. | Jun 2016 | B2 |
9378740 | Rosen et al. | Jun 2016 | B1 |
9380155 | Reding et al. | Jun 2016 | B1 |
9390726 | Smus et al. | Jul 2016 | B1 |
9401147 | Jitkoff et al. | Jul 2016 | B2 |
9406224 | Sanders et al. | Aug 2016 | B1 |
9412392 | Lindahl | Aug 2016 | B2 |
9423266 | Clark et al. | Aug 2016 | B2 |
9436918 | Pantel et al. | Sep 2016 | B2 |
9437186 | Liu et al. | Sep 2016 | B1 |
9437189 | Epstein et al. | Sep 2016 | B2 |
9454957 | Mathias et al. | Sep 2016 | B1 |
9484021 | Mairesse et al. | Nov 2016 | B1 |
9495129 | Fleizach et al. | Nov 2016 | B2 |
9501741 | Cheyer et al. | Nov 2016 | B2 |
9502025 | Kennewick et al. | Nov 2016 | B2 |
9508028 | Bannister et al. | Nov 2016 | B2 |
9510044 | Pereira et al. | Nov 2016 | B1 |
9524355 | Forbes et al. | Dec 2016 | B2 |
9535906 | Lee et al. | Jan 2017 | B2 |
9536527 | Carlson | Jan 2017 | B1 |
9547647 | Badaskar | Jan 2017 | B2 |
9548050 | Gruber et al. | Jan 2017 | B2 |
9569549 | Jenkins et al. | Feb 2017 | B1 |
9575964 | Yadgar et al. | Feb 2017 | B2 |
9578173 | Sanghavi et al. | Feb 2017 | B2 |
9607612 | Deleeuw | Mar 2017 | B2 |
9620113 | Kennewick et al. | Apr 2017 | B2 |
9620126 | Chiba | Apr 2017 | B2 |
9626955 | Fleizach et al. | Apr 2017 | B2 |
9633004 | Giuli et al. | Apr 2017 | B2 |
9633660 | Haughay | Apr 2017 | B2 |
9652453 | Mathur et al. | May 2017 | B2 |
9658746 | Cohn et al. | May 2017 | B2 |
9668121 | Naik et al. | May 2017 | B2 |
9672725 | Dotan-Cohen et al. | Jun 2017 | B2 |
9691378 | Meyers et al. | Jun 2017 | B1 |
9697827 | Lilly et al. | Jul 2017 | B1 |
9720907 | Bangalore et al. | Aug 2017 | B2 |
9734839 | Adams | Aug 2017 | B1 |
9741343 | Miles et al. | Aug 2017 | B1 |
20020004703 | Gaspard, II | Jan 2002 | A1 |
20020007300 | Slatter | Jan 2002 | A1 |
20020067308 | Robertson | Jun 2002 | A1 |
20020086680 | Hunzinger | Jul 2002 | A1 |
20020154160 | Hosokawa | Oct 2002 | A1 |
20030140088 | Robinson et al. | Jul 2003 | A1 |
20030177046 | Socha-Leialoha | Sep 2003 | A1 |
20040167652 | Ishii | Aug 2004 | A1 |
20040193420 | Kennewick et al. | Sep 2004 | A1 |
20050015772 | Saare et al. | Jan 2005 | A1 |
20050108074 | Bloechl et al. | May 2005 | A1 |
20050221876 | Van Bosch et al. | Oct 2005 | A1 |
20060061488 | Dunton | Mar 2006 | A1 |
20060077055 | Basir | Apr 2006 | A1 |
20060136280 | Cho et al. | Jun 2006 | A1 |
20060141990 | Zak et al. | Jun 2006 | A1 |
20060156252 | Sheshagiri et al. | Jul 2006 | A1 |
20060156307 | Kunjithapatham et al. | Jul 2006 | A1 |
20070006098 | Krumm et al. | Jan 2007 | A1 |
20070014280 | Cormier et al. | Jan 2007 | A1 |
20070027732 | Hudgens | Feb 2007 | A1 |
20070036117 | Taube et al. | Feb 2007 | A1 |
20070036286 | Champlin et al. | Feb 2007 | A1 |
20070050191 | Weider et al. | Mar 2007 | A1 |
20070052586 | Horstemeyer | Mar 2007 | A1 |
20070143163 | Weiss et al. | Jun 2007 | A1 |
20070185754 | Schmidt | Aug 2007 | A1 |
20070276714 | Beringer | Nov 2007 | A1 |
20080001785 | Elizarov et al. | Jan 2008 | A1 |
20080005053 | Pulsipher | Jan 2008 | A1 |
20080010050 | Fux et al. | Jan 2008 | A1 |
20080010355 | Vieri et al. | Jan 2008 | A1 |
20080010605 | Frank et al. | Jan 2008 | A1 |
20080012950 | Lee et al. | Jan 2008 | A1 |
20080013751 | Hiselius | Jan 2008 | A1 |
20080015863 | Agapi et al. | Jan 2008 | A1 |
20080015864 | Ross et al. | Jan 2008 | A1 |
20080016575 | Vincent et al. | Jan 2008 | A1 |
20080021708 | Bennett et al. | Jan 2008 | A1 |
20080021886 | Wang-Aryattanwanich et al. | Jan 2008 | A1 |
20080022208 | Morse | Jan 2008 | A1 |
20080027711 | Rajendran et al. | Jan 2008 | A1 |
20080027726 | Hansen et al. | Jan 2008 | A1 |
20080031475 | Goldstein | Feb 2008 | A1 |
20080033719 | Hall et al. | Feb 2008 | A1 |
20080033723 | Jang et al. | Feb 2008 | A1 |
20080034032 | Healey et al. | Feb 2008 | A1 |
20080034044 | Bhakta et al. | Feb 2008 | A1 |
20080034081 | Marshall et al. | Feb 2008 | A1 |
20080036743 | Westerman et al. | Feb 2008 | A1 |
20080040339 | Zhou et al. | Feb 2008 | A1 |
20080042970 | Liang et al. | Feb 2008 | A1 |
20080043936 | Liebermann | Feb 2008 | A1 |
20080043943 | Sipher et al. | Feb 2008 | A1 |
20080046239 | Boo | Feb 2008 | A1 |
20080046250 | Agapi et al. | Feb 2008 | A1 |
20080046422 | Lee et al. | Feb 2008 | A1 |
20080046820 | Lee et al. | Feb 2008 | A1 |
20080046948 | Verosub | Feb 2008 | A1 |
20080048908 | Sato | Feb 2008 | A1 |
20080050027 | Bashyam et al. | Feb 2008 | A1 |
20080052063 | Bennett et al. | Feb 2008 | A1 |
20080052073 | Goto et al. | Feb 2008 | A1 |
20080052077 | Bennett et al. | Feb 2008 | A1 |
20080052080 | Narayanan et al. | Feb 2008 | A1 |
20080052262 | Kosinov et al. | Feb 2008 | A1 |
20080055194 | Baudino et al. | Mar 2008 | A1 |
20080056459 | Vallier et al. | Mar 2008 | A1 |
20080056579 | Guha | Mar 2008 | A1 |
20080057922 | Kokes et al. | Mar 2008 | A1 |
20080059190 | Chu et al. | Mar 2008 | A1 |
20080059200 | Puli | Mar 2008 | A1 |
20080059876 | Hantler et al. | Mar 2008 | A1 |
20080062141 | Chaudhri | Mar 2008 | A1 |
20080065382 | Gerl et al. | Mar 2008 | A1 |
20080065387 | Cross et al. | Mar 2008 | A1 |
20080071529 | Silverman et al. | Mar 2008 | A1 |
20080071544 | Beaufays et al. | Mar 2008 | A1 |
20080071742 | Yang et al. | Mar 2008 | A1 |
20080072143 | Assadollahi | Mar 2008 | A1 |
20080075296 | Lindahl et al. | Mar 2008 | A1 |
20080076972 | Dorogusker et al. | Mar 2008 | A1 |
20080077310 | Murlidar et al. | Mar 2008 | A1 |
20080077384 | Agapi et al. | Mar 2008 | A1 |
20080077386 | Gao et al. | Mar 2008 | A1 |
20080077391 | Chino et al. | Mar 2008 | A1 |
20080077393 | Gao et al. | Mar 2008 | A1 |
20080077406 | Ganong | Mar 2008 | A1 |
20080077859 | Schabes et al. | Mar 2008 | A1 |
20080079566 | Singh et al. | Apr 2008 | A1 |
20080080411 | Cole | Apr 2008 | A1 |
20080082332 | Mallett et al. | Apr 2008 | A1 |
20080082338 | O'Neil et al. | Apr 2008 | A1 |
20080082390 | Hawkins et al. | Apr 2008 | A1 |
20080082576 | Bodin et al. | Apr 2008 | A1 |
20080082651 | Singh et al. | Apr 2008 | A1 |
20080084974 | Dhanakshirur | Apr 2008 | A1 |
20080085689 | Zellner | Apr 2008 | A1 |
20080091406 | Baldwin et al. | Apr 2008 | A1 |
20080091426 | Rempel et al. | Apr 2008 | A1 |
20080091428 | Bellegarda | Apr 2008 | A1 |
20080091443 | Strope et al. | Apr 2008 | A1 |
20080096531 | McQuaide | Apr 2008 | A1 |
20080096533 | Manfredi et al. | Apr 2008 | A1 |
20080096726 | Riley et al. | Apr 2008 | A1 |
20080097937 | Hadjarian | Apr 2008 | A1 |
20080098302 | Roose | Apr 2008 | A1 |
20080098480 | Henry et al. | Apr 2008 | A1 |
20080100579 | Robinson et al. | May 2008 | A1 |
20080101584 | Gray et al. | May 2008 | A1 |
20080103774 | White | May 2008 | A1 |
20080109222 | Liu | May 2008 | A1 |
20080109402 | Wang et al. | May 2008 | A1 |
20080114480 | Harb | May 2008 | A1 |
20080114598 | Prieto et al. | May 2008 | A1 |
20080114604 | Wei et al. | May 2008 | A1 |
20080114714 | Vemuri | May 2008 | A1 |
20080114841 | Lambert | May 2008 | A1 |
20080115084 | Scott et al. | May 2008 | A1 |
20080118143 | Gordon et al. | May 2008 | A1 |
20080119953 | Reed | May 2008 | A1 |
20080120102 | Rao | May 2008 | A1 |
20080120112 | Jordan et al. | May 2008 | A1 |
20080120196 | Reed et al. | May 2008 | A1 |
20080120311 | Reed | May 2008 | A1 |
20080120312 | Reed | May 2008 | A1 |
20080120330 | Reed | May 2008 | A1 |
20080120342 | Reed et al. | May 2008 | A1 |
20080122796 | Jobs et al. | May 2008 | A1 |
20080124695 | Myers et al. | May 2008 | A1 |
20080126075 | Thorn et al. | May 2008 | A1 |
20080126077 | Thorn | May 2008 | A1 |
20080126091 | Clark et al. | May 2008 | A1 |
20080126093 | Sivadas | May 2008 | A1 |
20080126100 | Grost et al. | May 2008 | A1 |
20080126491 | Portele et al. | May 2008 | A1 |
20080129520 | Lee | Jun 2008 | A1 |
20080130867 | Bowen | Jun 2008 | A1 |
20080131006 | Oliver | Jun 2008 | A1 |
20080132221 | Willey et al. | Jun 2008 | A1 |
20080132295 | Horowitz | Jun 2008 | A1 |
20080133215 | Sarukkai | Jun 2008 | A1 |
20080133228 | Rao | Jun 2008 | A1 |
20080133230 | Herforth et al. | Jun 2008 | A1 |
20080133241 | Baker et al. | Jun 2008 | A1 |
20080133245 | Proulx et al. | Jun 2008 | A1 |
20080133479 | Zelevinsky et al. | Jun 2008 | A1 |
20080133956 | Fadell | Jun 2008 | A1 |
20080140413 | Millman et al. | Jun 2008 | A1 |
20080140415 | Shostak | Jun 2008 | A1 |
20080140416 | Shostak | Jun 2008 | A1 |
20080140652 | Millman et al. | Jun 2008 | A1 |
20080140657 | Azvine et al. | Jun 2008 | A1 |
20080140702 | Reed et al. | Jun 2008 | A1 |
20080141125 | Ghassabian et al. | Jun 2008 | A1 |
20080141180 | Reed et al. | Jun 2008 | A1 |
20080141182 | Barsness et al. | Jun 2008 | A1 |
20080146245 | Appaji | Jun 2008 | A1 |
20080146290 | Sreeram et al. | Jun 2008 | A1 |
20080147408 | Da Palma et al. | Jun 2008 | A1 |
20080147411 | Dames et al. | Jun 2008 | A1 |
20080147874 | Yoneda et al. | Jun 2008 | A1 |
20080150900 | Han | Jun 2008 | A1 |
20080154577 | Kim et al. | Jun 2008 | A1 |
20080154599 | Muschett et al. | Jun 2008 | A1 |
20080154600 | Tian et al. | Jun 2008 | A1 |
20080154603 | Oddo | Jun 2008 | A1 |
20080154612 | Evermann et al. | Jun 2008 | A1 |
20080154828 | Antebi et al. | Jun 2008 | A1 |
20080155547 | Weber et al. | Jun 2008 | A1 |
20080157867 | Krah | Jul 2008 | A1 |
20080161113 | Hansen et al. | Jul 2008 | A1 |
20080162120 | Mactavish et al. | Jul 2008 | A1 |
20080162137 | Saitoh et al. | Jul 2008 | A1 |
20080162471 | Bernard | Jul 2008 | A1 |
20080163119 | Kim et al. | Jul 2008 | A1 |
20080163131 | Hirai et al. | Jul 2008 | A1 |
20080165144 | Forstall et al. | Jul 2008 | A1 |
20080165980 | Pavlovic et al. | Jul 2008 | A1 |
20080165994 | Caren et al. | Jul 2008 | A1 |
20080167013 | Novick et al. | Jul 2008 | A1 |
20080167858 | Christie et al. | Jul 2008 | A1 |
20080167876 | Bakis et al. | Jul 2008 | A1 |
20080168052 | Ott et al. | Jul 2008 | A1 |
20080168144 | Lee | Jul 2008 | A1 |
20080168366 | Kocienda et al. | Jul 2008 | A1 |
20080172698 | Berger et al. | Jul 2008 | A1 |
20080183473 | Nagano et al. | Jul 2008 | A1 |
20080186960 | Kocheisen et al. | Aug 2008 | A1 |
20080189099 | Friedman et al. | Aug 2008 | A1 |
20080189106 | Low et al. | Aug 2008 | A1 |
20080189110 | Freeman et al. | Aug 2008 | A1 |
20080189114 | Fail et al. | Aug 2008 | A1 |
20080189360 | Kiley et al. | Aug 2008 | A1 |
20080189606 | Rybak | Aug 2008 | A1 |
20080195312 | Aaron et al. | Aug 2008 | A1 |
20080195388 | Bower et al. | Aug 2008 | A1 |
20080195391 | Marple et al. | Aug 2008 | A1 |
20080195601 | Ntoulas et al. | Aug 2008 | A1 |
20080195630 | Exartier et al. | Aug 2008 | A1 |
20080195940 | Gail et al. | Aug 2008 | A1 |
20080200142 | Abdel-Kader et al. | Aug 2008 | A1 |
20080201000 | Heikkila et al. | Aug 2008 | A1 |
20080201306 | Cooper et al. | Aug 2008 | A1 |
20080201375 | Khedouri et al. | Aug 2008 | A1 |
20080201434 | Holmes et al. | Aug 2008 | A1 |
20080204379 | Perez-Noguera | Aug 2008 | A1 |
20080207176 | Brackbill et al. | Aug 2008 | A1 |
20080208585 | Ativanichayaphong et al. | Aug 2008 | A1 |
20080208587 | Ben-David et al. | Aug 2008 | A1 |
20080208864 | Cucerzan et al. | Aug 2008 | A1 |
20080212796 | Denda | Sep 2008 | A1 |
20080219641 | Sandrew et al. | Sep 2008 | A1 |
20080221866 | Katragadda et al. | Sep 2008 | A1 |
20080221879 | Cerra et al. | Sep 2008 | A1 |
20080221880 | Cerra et al. | Sep 2008 | A1 |
20080221887 | Rose et al. | Sep 2008 | A1 |
20080221889 | Cerra et al. | Sep 2008 | A1 |
20080221903 | Kanevsky et al. | Sep 2008 | A1 |
20080222118 | Scian et al. | Sep 2008 | A1 |
20080226130 | Kansal et al. | Sep 2008 | A1 |
20080228463 | Mori et al. | Sep 2008 | A1 |
20080228485 | Owen | Sep 2008 | A1 |
20080228490 | Fischer et al. | Sep 2008 | A1 |
20080228495 | Cross et al. | Sep 2008 | A1 |
20080228496 | Yu et al. | Sep 2008 | A1 |
20080228928 | Donelli et al. | Sep 2008 | A1 |
20080229185 | Lynch | Sep 2008 | A1 |
20080229218 | Maeng | Sep 2008 | A1 |
20080235017 | Satomura et al. | Sep 2008 | A1 |
20080235023 | Kennewick et al. | Sep 2008 | A1 |
20080235024 | Goldberg et al. | Sep 2008 | A1 |
20080235027 | Cross | Sep 2008 | A1 |
20080240569 | Tonouchi | Oct 2008 | A1 |
20080242280 | Shapiro et al. | Oct 2008 | A1 |
20080242322 | Scott et al. | Oct 2008 | A1 |
20080242363 | Onda et al. | Oct 2008 | A1 |
20080243501 | Hafsteinsson et al. | Oct 2008 | A1 |
20080243834 | Rieman et al. | Oct 2008 | A1 |
20080244390 | Fux et al. | Oct 2008 | A1 |
20080244446 | Lefevre et al. | Oct 2008 | A1 |
20080247519 | Abella et al. | Oct 2008 | A1 |
20080247529 | Barton et al. | Oct 2008 | A1 |
20080248797 | Freeman et al. | Oct 2008 | A1 |
20080249770 | Kim et al. | Oct 2008 | A1 |
20080249778 | Barton et al. | Oct 2008 | A1 |
20080253577 | Eppolito | Oct 2008 | A1 |
20080254419 | Cohen | Oct 2008 | A1 |
20080254425 | Cohen et al. | Oct 2008 | A1 |
20080255837 | Kahn et al. | Oct 2008 | A1 |
20080255842 | Simhi et al. | Oct 2008 | A1 |
20080255845 | Bennett | Oct 2008 | A1 |
20080256613 | Grover | Oct 2008 | A1 |
20080259022 | Mansfield et al. | Oct 2008 | A1 |
20080261572 | Tsui et al. | Oct 2008 | A1 |
20080262828 | Och et al. | Oct 2008 | A1 |
20080262838 | Nurminen et al. | Oct 2008 | A1 |
20080262846 | Burns et al. | Oct 2008 | A1 |
20080263139 | Martin | Oct 2008 | A1 |
20080267416 | Goldstein et al. | Oct 2008 | A1 |
20080270118 | Kuo et al. | Oct 2008 | A1 |
20080270138 | Knight et al. | Oct 2008 | A1 |
20080270139 | Shi et al. | Oct 2008 | A1 |
20080270140 | Hertz et al. | Oct 2008 | A1 |
20080270151 | Mahoney et al. | Oct 2008 | A1 |
20080270344 | Yurick et al. | Oct 2008 | A1 |
20080273672 | Didcock et al. | Nov 2008 | A1 |
20080277473 | Kotlarsky et al. | Nov 2008 | A1 |
20080281510 | Shahine | Nov 2008 | A1 |
20080281582 | Hsu et al. | Nov 2008 | A1 |
20080288259 | Chambers et al. | Nov 2008 | A1 |
20080288460 | Poniatowski et al. | Nov 2008 | A1 |
20080292112 | Valenzuela et al. | Nov 2008 | A1 |
20080294418 | Cleary et al. | Nov 2008 | A1 |
20080294517 | Hill | Nov 2008 | A1 |
20080294651 | Masuyama et al. | Nov 2008 | A1 |
20080294981 | Balzano et al. | Nov 2008 | A1 |
20080298563 | Rondeau et al. | Dec 2008 | A1 |
20080298766 | Wen et al. | Dec 2008 | A1 |
20080299523 | Chai et al. | Dec 2008 | A1 |
20080300857 | Barbaiani et al. | Dec 2008 | A1 |
20080300871 | Gilbert | Dec 2008 | A1 |
20080300877 | Gilbert et al. | Dec 2008 | A1 |
20080300878 | Bennett | Dec 2008 | A1 |
20080301567 | Martin et al. | Dec 2008 | A1 |
20080303645 | Seymour et al. | Dec 2008 | A1 |
20080306727 | Thurmair et al. | Dec 2008 | A1 |
20080312909 | Hermansen et al. | Dec 2008 | A1 |
20080312928 | Goebel et al. | Dec 2008 | A1 |
20080313335 | Jung et al. | Dec 2008 | A1 |
20080316183 | Westerman et al. | Dec 2008 | A1 |
20080319735 | Kambhatla et al. | Dec 2008 | A1 |
20080319738 | Liu et al. | Dec 2008 | A1 |
20080319753 | Hancock | Dec 2008 | A1 |
20080319763 | Di Fabbrizio et al. | Dec 2008 | A1 |
20080319783 | Yao et al. | Dec 2008 | A1 |
20090003115 | Lindahl et al. | Jan 2009 | A1 |
20090005012 | Van Heugten | Jan 2009 | A1 |
20090005891 | Batson et al. | Jan 2009 | A1 |
20090006096 | Li et al. | Jan 2009 | A1 |
20090006097 | Etezadi et al. | Jan 2009 | A1 |
20090006099 | Sharpe et al. | Jan 2009 | A1 |
20090006100 | Badger et al. | Jan 2009 | A1 |
20090006343 | Platt et al. | Jan 2009 | A1 |
20090006345 | Platt et al. | Jan 2009 | A1 |
20090006488 | Lindahl et al. | Jan 2009 | A1 |
20090006671 | Batson et al. | Jan 2009 | A1 |
20090007001 | Morin et al. | Jan 2009 | A1 |
20090011709 | Akasaka et al. | Jan 2009 | A1 |
20090012748 | Beish et al. | Jan 2009 | A1 |
20090012775 | El Hady et al. | Jan 2009 | A1 |
20090018828 | Nakadai et al. | Jan 2009 | A1 |
20090018829 | Kuperstein | Jan 2009 | A1 |
20090018834 | Cooper et al. | Jan 2009 | A1 |
20090018835 | Cooper et al. | Jan 2009 | A1 |
20090018839 | Cooper et al. | Jan 2009 | A1 |
20090018840 | Lutz et al. | Jan 2009 | A1 |
20090022329 | Mahowald | Jan 2009 | A1 |
20090024595 | Chen | Jan 2009 | A1 |
20090028435 | Wu et al. | Jan 2009 | A1 |
20090030800 | Grois | Jan 2009 | A1 |
20090030978 | Johnson et al. | Jan 2009 | A1 |
20090043580 | Mozer et al. | Feb 2009 | A1 |
20090043583 | Agapi et al. | Feb 2009 | A1 |
20090043763 | Peng | Feb 2009 | A1 |
20090044094 | Rapp et al. | Feb 2009 | A1 |
20090048821 | Yam et al. | Feb 2009 | A1 |
20090048841 | Pollet et al. | Feb 2009 | A1 |
20090048845 | Burckart et al. | Feb 2009 | A1 |
20090049067 | Murray | Feb 2009 | A1 |
20090055168 | Wu et al. | Feb 2009 | A1 |
20090055175 | Terrell et al. | Feb 2009 | A1 |
20090055179 | Cho et al. | Feb 2009 | A1 |
20090055186 | Lance et al. | Feb 2009 | A1 |
20090055380 | Peng et al. | Feb 2009 | A1 |
20090055381 | Wu et al. | Feb 2009 | A1 |
20090055648 | Kim et al. | Feb 2009 | A1 |
20090058823 | Kocienda | Mar 2009 | A1 |
20090058860 | Fong et al. | Mar 2009 | A1 |
20090060351 | Li et al. | Mar 2009 | A1 |
20090060472 | Bull et al. | Mar 2009 | A1 |
20090063974 | Bull et al. | Mar 2009 | A1 |
20090064031 | Bull et al. | Mar 2009 | A1 |
20090070097 | Wu et al. | Mar 2009 | A1 |
20090070102 | Maegawa | Mar 2009 | A1 |
20090070109 | Didcock et al. | Mar 2009 | A1 |
20090070114 | Staszak | Mar 2009 | A1 |
20090074214 | Bradford et al. | Mar 2009 | A1 |
20090076792 | Lawson-Tancred | Mar 2009 | A1 |
20090076796 | Daraselia | Mar 2009 | A1 |
20090076798 | Oh et al. | Mar 2009 | A1 |
20090076819 | Wouters et al. | Mar 2009 | A1 |
20090076821 | Brenner et al. | Mar 2009 | A1 |
20090076825 | Bradford et al. | Mar 2009 | A1 |
20090077165 | Rhodes et al. | Mar 2009 | A1 |
20090077464 | Goldsmith et al. | Mar 2009 | A1 |
20090079622 | Seshadri et al. | Mar 2009 | A1 |
20090083034 | Hernandez et al. | Mar 2009 | A1 |
20090083035 | Huang et al. | Mar 2009 | A1 |
20090083036 | Zhao et al. | Mar 2009 | A1 |
20090083037 | Gleason et al. | Mar 2009 | A1 |
20090083047 | Lindahl et al. | Mar 2009 | A1 |
20090089058 | Bellegarda | Apr 2009 | A1 |
20090091537 | Huang et al. | Apr 2009 | A1 |
20090092239 | Macwan et al. | Apr 2009 | A1 |
20090092260 | Powers | Apr 2009 | A1 |
20090092261 | Bard | Apr 2009 | A1 |
20090092262 | Costa et al. | Apr 2009 | A1 |
20090094029 | Koch et al. | Apr 2009 | A1 |
20090094033 | Mozer et al. | Apr 2009 | A1 |
20090097634 | Nambiar et al. | Apr 2009 | A1 |
20090097637 | Boscher et al. | Apr 2009 | A1 |
20090098903 | Donaldson et al. | Apr 2009 | A1 |
20090100049 | Cao | Apr 2009 | A1 |
20090100454 | Weber | Apr 2009 | A1 |
20090102637 | Hsu | Apr 2009 | A1 |
20090104898 | Harris | Apr 2009 | A1 |
20090106026 | Ferrieux | Apr 2009 | A1 |
20090106376 | Tom et al. | Apr 2009 | A1 |
20090106397 | O'Keefe | Apr 2009 | A1 |
20090112572 | Thorn | Apr 2009 | A1 |
20090112576 | Jackson et al. | Apr 2009 | A1 |
20090112592 | Candelore et al. | Apr 2009 | A1 |
20090112677 | Rhett | Apr 2009 | A1 |
20090112892 | Cardie et al. | Apr 2009 | A1 |
20090119587 | Allen et al. | May 2009 | A1 |
20090123021 | Jung et al. | May 2009 | A1 |
20090123071 | Iwasaki | May 2009 | A1 |
20090125477 | Lu et al. | May 2009 | A1 |
20090125602 | Bhatia et al. | May 2009 | A1 |
20090125947 | Ibaraki | May 2009 | A1 |
20090128505 | Partridge et al. | May 2009 | A1 |
20090132253 | Bellegarda | May 2009 | A1 |
20090132255 | Lu | May 2009 | A1 |
20090132378 | Othmer et al. | May 2009 | A1 |
20090137286 | Luke et al. | May 2009 | A1 |
20090138263 | Shozakai et al. | May 2009 | A1 |
20090138430 | Nambiar et al. | May 2009 | A1 |
20090138736 | Chin | May 2009 | A1 |
20090138828 | Schultz et al. | May 2009 | A1 |
20090144036 | Jorgensen et al. | Jun 2009 | A1 |
20090144049 | Haddad et al. | Jun 2009 | A1 |
20090144428 | Bowater et al. | Jun 2009 | A1 |
20090144609 | Liang et al. | Jun 2009 | A1 |
20090146848 | Ghassabian | Jun 2009 | A1 |
20090150147 | Jacoby et al. | Jun 2009 | A1 |
20090150156 | Kennewick et al. | Jun 2009 | A1 |
20090152349 | Bonev et al. | Jun 2009 | A1 |
20090153288 | Hope et al. | Jun 2009 | A1 |
20090154669 | Wood et al. | Jun 2009 | A1 |
20090157382 | Bar | Jun 2009 | A1 |
20090157384 | Toutanova et al. | Jun 2009 | A1 |
20090157401 | Bennett | Jun 2009 | A1 |
20090157693 | Palahnuk | Jun 2009 | A1 |
20090158200 | Palahnuk et al. | Jun 2009 | A1 |
20090158323 | Bober et al. | Jun 2009 | A1 |
20090158423 | Orlassino et al. | Jun 2009 | A1 |
20090160803 | Hashimoto | Jun 2009 | A1 |
20090163243 | Barbera | Jun 2009 | A1 |
20090164301 | O'Sullivan et al. | Jun 2009 | A1 |
20090164441 | Cheyer | Jun 2009 | A1 |
20090164655 | Pettersson et al. | Jun 2009 | A1 |
20090164937 | Alviar et al. | Jun 2009 | A1 |
20090167508 | Fadell et al. | Jul 2009 | A1 |
20090167509 | Fadell et al. | Jul 2009 | A1 |
20090171578 | Kim et al. | Jul 2009 | A1 |
20090171662 | Huang et al. | Jul 2009 | A1 |
20090171664 | Kennewick et al. | Jul 2009 | A1 |
20090172108 | Singh | Jul 2009 | A1 |
20090172542 | Girish et al. | Jul 2009 | A1 |
20090174667 | Kocienda et al. | Jul 2009 | A1 |
20090174677 | Gehani et al. | Jul 2009 | A1 |
20090177300 | Lee | Jul 2009 | A1 |
20090177461 | Ehsani et al. | Jul 2009 | A1 |
20090177966 | Chaudhri | Jul 2009 | A1 |
20090182445 | Girish et al. | Jul 2009 | A1 |
20090182549 | Anisimovich et al. | Jul 2009 | A1 |
20090182702 | Miller | Jul 2009 | A1 |
20090183070 | Robbins | Jul 2009 | A1 |
20090187402 | Scholl | Jul 2009 | A1 |
20090187577 | Reznik et al. | Jul 2009 | A1 |
20090187950 | Nicas et al. | Jul 2009 | A1 |
20090190774 | Wang et al. | Jul 2009 | A1 |
20090191895 | Singh et al. | Jul 2009 | A1 |
20090192782 | Drewes | Jul 2009 | A1 |
20090192787 | Roon | Jul 2009 | A1 |
20090192798 | Basson et al. | Jul 2009 | A1 |
20090197619 | Colligan et al. | Aug 2009 | A1 |
20090198497 | Kwon | Aug 2009 | A1 |
20090204402 | Marhawa et al. | Aug 2009 | A1 |
20090204409 | Mozer et al. | Aug 2009 | A1 |
20090204478 | Kaib et al. | Aug 2009 | A1 |
20090204596 | Brun et al. | Aug 2009 | A1 |
20090204601 | Grasset | Aug 2009 | A1 |
20090204620 | Thione et al. | Aug 2009 | A1 |
20090210230 | Schwarz et al. | Aug 2009 | A1 |
20090210232 | Sanchez et al. | Aug 2009 | A1 |
20090213134 | Stephanick et al. | Aug 2009 | A1 |
20090215466 | Ahl et al. | Aug 2009 | A1 |
20090215503 | Zhang et al. | Aug 2009 | A1 |
20090216396 | Yamagata | Aug 2009 | A1 |
20090216528 | Gemello et al. | Aug 2009 | A1 |
20090216540 | Tessel et al. | Aug 2009 | A1 |
20090216569 | Bonev et al. | Aug 2009 | A1 |
20090216704 | Zheng et al. | Aug 2009 | A1 |
20090219166 | MacFarlane | Sep 2009 | A1 |
20090221274 | Venkatakrishnan et al. | Sep 2009 | A1 |
20090222257 | Sumita et al. | Sep 2009 | A1 |
20090222270 | Likens et al. | Sep 2009 | A2 |
20090222488 | Boerries et al. | Sep 2009 | A1 |
20090224867 | O'Shaughnessy et al. | Sep 2009 | A1 |
20090228126 | Spielberg et al. | Sep 2009 | A1 |
20090228273 | Wang et al. | Sep 2009 | A1 |
20090228277 | Bonforte et al. | Sep 2009 | A1 |
20090228281 | Singleton et al. | Sep 2009 | A1 |
20090228439 | Manolescu et al. | Sep 2009 | A1 |
20090228792 | Van Os et al. | Sep 2009 | A1 |
20090228842 | Westerman et al. | Sep 2009 | A1 |
20090233264 | Rogers et al. | Sep 2009 | A1 |
20090234638 | Ranjan et al. | Sep 2009 | A1 |
20090234651 | Basir et al. | Sep 2009 | A1 |
20090234655 | Kwon | Sep 2009 | A1 |
20090235280 | Tannier et al. | Sep 2009 | A1 |
20090239202 | Stone | Sep 2009 | A1 |
20090239552 | Churchill et al. | Sep 2009 | A1 |
20090240485 | Dalal et al. | Sep 2009 | A1 |
20090241054 | Hendricks | Sep 2009 | A1 |
20090241760 | Georges | Oct 2009 | A1 |
20090247237 | Mittleman et al. | Oct 2009 | A1 |
20090248182 | Logan et al. | Oct 2009 | A1 |
20090248395 | Alewine et al. | Oct 2009 | A1 |
20090248402 | Ito et al. | Oct 2009 | A1 |
20090248420 | Basir et al. | Oct 2009 | A1 |
20090248422 | Li et al. | Oct 2009 | A1 |
20090248456 | Fahmy et al. | Oct 2009 | A1 |
20090249198 | Davis et al. | Oct 2009 | A1 |
20090249247 | Tseng et al. | Oct 2009 | A1 |
20090252350 | Seguin | Oct 2009 | A1 |
20090253457 | Seguin | Oct 2009 | A1 |
20090253463 | Shin et al. | Oct 2009 | A1 |
20090254339 | Seguin | Oct 2009 | A1 |
20090254345 | Fleizach et al. | Oct 2009 | A1 |
20090254819 | Song et al. | Oct 2009 | A1 |
20090254823 | Barrett | Oct 2009 | A1 |
20090259475 | Yamagami et al. | Oct 2009 | A1 |
20090259969 | Pallakoff | Oct 2009 | A1 |
20090265171 | Davis | Oct 2009 | A1 |
20090265368 | Crider et al. | Oct 2009 | A1 |
20090271109 | Lee et al. | Oct 2009 | A1 |
20090271175 | Bodin et al. | Oct 2009 | A1 |
20090271176 | Bodin et al. | Oct 2009 | A1 |
20090271178 | Bodin et al. | Oct 2009 | A1 |
20090271188 | Agapi et al. | Oct 2009 | A1 |
20090271189 | Agapi et al. | Oct 2009 | A1 |
20090274315 | Carnes et al. | Nov 2009 | A1 |
20090278804 | Rubanovich et al. | Nov 2009 | A1 |
20090281789 | Waibel et al. | Nov 2009 | A1 |
20090284482 | Chin | Nov 2009 | A1 |
20090286514 | Lichorowic et al. | Nov 2009 | A1 |
20090287583 | Holmes | Nov 2009 | A1 |
20090290718 | Kahn et al. | Nov 2009 | A1 |
20090292987 | Sorenson | Nov 2009 | A1 |
20090296552 | Hicks et al. | Dec 2009 | A1 |
20090298474 | George | Dec 2009 | A1 |
20090298529 | Mahajan | Dec 2009 | A1 |
20090299745 | Kennewick et al. | Dec 2009 | A1 |
20090299849 | Cao et al. | Dec 2009 | A1 |
20090300391 | Jessup et al. | Dec 2009 | A1 |
20090300488 | Salamon et al. | Dec 2009 | A1 |
20090304198 | Herre et al. | Dec 2009 | A1 |
20090305203 | Okumura et al. | Dec 2009 | A1 |
20090306967 | Nicolov et al. | Dec 2009 | A1 |
20090306969 | Goud et al. | Dec 2009 | A1 |
20090306979 | Jaiswal et al. | Dec 2009 | A1 |
20090306980 | Shin | Dec 2009 | A1 |
20090306981 | Cromack et al. | Dec 2009 | A1 |
20090306985 | Roberts et al. | Dec 2009 | A1 |
20090306988 | Chen et al. | Dec 2009 | A1 |
20090306989 | Kaji | Dec 2009 | A1 |
20090307162 | Bui et al. | Dec 2009 | A1 |
20090307201 | Dunning et al. | Dec 2009 | A1 |
20090307584 | Davidson et al. | Dec 2009 | A1 |
20090307594 | Kosonen et al. | Dec 2009 | A1 |
20090309352 | Walker et al. | Dec 2009 | A1 |
20090313014 | Shin et al. | Dec 2009 | A1 |
20090313020 | Koivunen | Dec 2009 | A1 |
20090313023 | Jones | Dec 2009 | A1 |
20090313026 | Coffman et al. | Dec 2009 | A1 |
20090313544 | Wood et al. | Dec 2009 | A1 |
20090313564 | Rottler et al. | Dec 2009 | A1 |
20090316943 | Frigola Munoz et al. | Dec 2009 | A1 |
20090318119 | Basir et al. | Dec 2009 | A1 |
20090318198 | Carroll | Dec 2009 | A1 |
20090319257 | Blume et al. | Dec 2009 | A1 |
20090319266 | Brown et al. | Dec 2009 | A1 |
20090319342 | Shilman et al. | Dec 2009 | A1 |
20090326923 | Yan et al. | Dec 2009 | A1 |
20090326936 | Nagashima | Dec 2009 | A1 |
20090326938 | Marila et al. | Dec 2009 | A1 |
20090326949 | Douthitt et al. | Dec 2009 | A1 |
20090327977 | Bachfischer et al. | Dec 2009 | A1 |
20100004918 | Lee et al. | Jan 2010 | A1 |
20100004930 | Strope et al. | Jan 2010 | A1 |
20100004931 | Ma et al. | Jan 2010 | A1 |
20100005081 | Bennett | Jan 2010 | A1 |
20100010803 | Ishikawa et al. | Jan 2010 | A1 |
20100010814 | Patel | Jan 2010 | A1 |
20100013760 | Hirai et al. | Jan 2010 | A1 |
20100013796 | Abileah et al. | Jan 2010 | A1 |
20100017212 | Attwater et al. | Jan 2010 | A1 |
20100017382 | Katragadda et al. | Jan 2010 | A1 |
20100019834 | Zerbe et al. | Jan 2010 | A1 |
20100023318 | Lemoine | Jan 2010 | A1 |
20100023320 | Di Cristo et al. | Jan 2010 | A1 |
20100023331 | Duta et al. | Jan 2010 | A1 |
20100026526 | Yokota | Feb 2010 | A1 |
20100030549 | Lee et al. | Feb 2010 | A1 |
20100030928 | Conroy et al. | Feb 2010 | A1 |
20100031143 | Rao et al. | Feb 2010 | A1 |
20100036653 | Kim et al. | Feb 2010 | A1 |
20100036655 | Cecil et al. | Feb 2010 | A1 |
20100036660 | Bennett | Feb 2010 | A1 |
20100036928 | Granite et al. | Feb 2010 | A1 |
20100037183 | Miyashita et al. | Feb 2010 | A1 |
20100042400 | Block et al. | Feb 2010 | A1 |
20100042576 | Roettger et al. | Feb 2010 | A1 |
20100046842 | Conwell et al. | Feb 2010 | A1 |
20100049498 | Cao et al. | Feb 2010 | A1 |
20100049514 | Kennewick et al. | Feb 2010 | A1 |
20100050064 | Liu et al. | Feb 2010 | A1 |
20100054512 | Solum | Mar 2010 | A1 |
20100057435 | Kent et al. | Mar 2010 | A1 |
20100057443 | Di Cristo et al. | Mar 2010 | A1 |
20100057457 | Ogata et al. | Mar 2010 | A1 |
20100057461 | Neubacher et al. | Mar 2010 | A1 |
20100057643 | Yang | Mar 2010 | A1 |
20100058200 | Jablokov et al. | Mar 2010 | A1 |
20100060646 | Unsal et al. | Mar 2010 | A1 |
20100063804 | Sato et al. | Mar 2010 | A1 |
20100063825 | Williams et al. | Mar 2010 | A1 |
20100063961 | Guiheneuf et al. | Mar 2010 | A1 |
20100064113 | Lindahl et al. | Mar 2010 | A1 |
20100064218 | Bull et al. | Mar 2010 | A1 |
20100064226 | Stefaniak et al. | Mar 2010 | A1 |
20100066546 | Aaron | Mar 2010 | A1 |
20100066684 | Shahraray et al. | Mar 2010 | A1 |
20100067723 | Bergmann et al. | Mar 2010 | A1 |
20100067867 | Lin et al. | Mar 2010 | A1 |
20100070281 | Conkie et al. | Mar 2010 | A1 |
20100070521 | Clinchant et al. | Mar 2010 | A1 |
20100070899 | Hunt et al. | Mar 2010 | A1 |
20100071003 | Bychkov et al. | Mar 2010 | A1 |
20100076760 | Kraenzel et al. | Mar 2010 | A1 |
20100076993 | Klawitter et al. | Mar 2010 | A1 |
20100077350 | Lim et al. | Mar 2010 | A1 |
20100079501 | Ikeda et al. | Apr 2010 | A1 |
20100080398 | Waldmann | Apr 2010 | A1 |
20100080470 | Deluca et al. | Apr 2010 | A1 |
20100081456 | Singh et al. | Apr 2010 | A1 |
20100081487 | Chen et al. | Apr 2010 | A1 |
20100082286 | Leung | Apr 2010 | A1 |
20100082327 | Rogers et al. | Apr 2010 | A1 |
20100082328 | Rogers et al. | Apr 2010 | A1 |
20100082329 | Silverman et al. | Apr 2010 | A1 |
20100082333 | Al-Shammari | Apr 2010 | A1 |
20100082346 | Rogers et al. | Apr 2010 | A1 |
20100082347 | Rogers et al. | Apr 2010 | A1 |
20100082348 | Silverman et al. | Apr 2010 | A1 |
20100082349 | Bellegarda et al. | Apr 2010 | A1 |
20100082567 | Rosenblatt et al. | Apr 2010 | A1 |
20100082970 | Lindahl et al. | Apr 2010 | A1 |
20100086152 | Rank et al. | Apr 2010 | A1 |
20100086153 | Hagen et al. | Apr 2010 | A1 |
20100086156 | Rank et al. | Apr 2010 | A1 |
20100088020 | Sano et al. | Apr 2010 | A1 |
20100088093 | Lee et al. | Apr 2010 | A1 |
20100088100 | Lindahl | Apr 2010 | A1 |
20100094632 | Davis et al. | Apr 2010 | A1 |
20100098231 | Wohlert et al. | Apr 2010 | A1 |
20100100212 | Lindahl et al. | Apr 2010 | A1 |
20100100371 | Yuezhong et al. | Apr 2010 | A1 |
20100100384 | Ju et al. | Apr 2010 | A1 |
20100100385 | Davis et al. | Apr 2010 | A1 |
20100100816 | Mccloskey et al. | Apr 2010 | A1 |
20100103776 | Chan | Apr 2010 | A1 |
20100103779 | Kakirde | Apr 2010 | A1 |
20100106486 | Hua et al. | Apr 2010 | A1 |
20100106498 | Morrison et al. | Apr 2010 | A1 |
20100106500 | McKee et al. | Apr 2010 | A1 |
20100106503 | Farrell et al. | Apr 2010 | A1 |
20100114856 | Kuboyama | May 2010 | A1 |
20100114887 | Conway et al. | May 2010 | A1 |
20100121637 | Roy et al. | May 2010 | A1 |
20100125456 | Weng et al. | May 2010 | A1 |
20100125458 | Franco et al. | May 2010 | A1 |
20100125460 | Mellott et al. | May 2010 | A1 |
20100125811 | Moore et al. | May 2010 | A1 |
20100131269 | Park et al. | May 2010 | A1 |
20100131273 | Aley-Raz et al. | May 2010 | A1 |
20100131498 | Linthicum et al. | May 2010 | A1 |
20100131899 | Hubert | May 2010 | A1 |
20100138215 | Williams | Jun 2010 | A1 |
20100138224 | Bedingfield, Sr. | Jun 2010 | A1 |
20100138416 | Bellotti | Jun 2010 | A1 |
20100138680 | Brisebois et al. | Jun 2010 | A1 |
20100138759 | Roy | Jun 2010 | A1 |
20100138798 | Wilson et al. | Jun 2010 | A1 |
20100142740 | Roerup | Jun 2010 | A1 |
20100145694 | Ju et al. | Jun 2010 | A1 |
20100145700 | Kennewick et al. | Jun 2010 | A1 |
20100145707 | Ljolje et al. | Jun 2010 | A1 |
20100146442 | Nagasaka et al. | Jun 2010 | A1 |
20100150321 | Harris et al. | Jun 2010 | A1 |
20100153114 | Shih et al. | Jun 2010 | A1 |
20100153115 | Klee et al. | Jun 2010 | A1 |
20100153448 | Harpur et al. | Jun 2010 | A1 |
20100161311 | Massuh | Jun 2010 | A1 |
20100161313 | Karttunen | Jun 2010 | A1 |
20100161337 | Pulz et al. | Jun 2010 | A1 |
20100161554 | Datuashvili et al. | Jun 2010 | A1 |
20100164897 | Morin et al. | Jul 2010 | A1 |
20100169075 | Raffa et al. | Jul 2010 | A1 |
20100169093 | Washio | Jul 2010 | A1 |
20100169097 | Nachman et al. | Jul 2010 | A1 |
20100169098 | Patch | Jul 2010 | A1 |
20100171713 | Kwok et al. | Jul 2010 | A1 |
20100174544 | Heifets | Jul 2010 | A1 |
20100175066 | Paik | Jul 2010 | A1 |
20100179932 | Yoon et al. | Jul 2010 | A1 |
20100179991 | Lorch et al. | Jul 2010 | A1 |
20100180218 | Boston et al. | Jul 2010 | A1 |
20100185448 | Meisel | Jul 2010 | A1 |
20100185949 | Jaeger | Jul 2010 | A1 |
20100191520 | Gruhn et al. | Jul 2010 | A1 |
20100197359 | Harris | Aug 2010 | A1 |
20100199180 | Brichter et al. | Aug 2010 | A1 |
20100199215 | Seymour et al. | Aug 2010 | A1 |
20100204986 | Kennewick et al. | Aug 2010 | A1 |
20100211199 | Naik et al. | Aug 2010 | A1 |
20100211379 | Gorman et al. | Aug 2010 | A1 |
20100211644 | Lavoie et al. | Aug 2010 | A1 |
20100216509 | Riemer et al. | Aug 2010 | A1 |
20100217604 | Baldwin et al. | Aug 2010 | A1 |
20100222033 | Scott et al. | Sep 2010 | A1 |
20100222098 | Garg | Sep 2010 | A1 |
20100223055 | Mclean | Sep 2010 | A1 |
20100223056 | Kadirkamanathan et al. | Sep 2010 | A1 |
20100223131 | Scott et al. | Sep 2010 | A1 |
20100225599 | Danielsson et al. | Sep 2010 | A1 |
20100225809 | Connors et al. | Sep 2010 | A1 |
20100227642 | Kim et al. | Sep 2010 | A1 |
20100228540 | Bennett | Sep 2010 | A1 |
20100228549 | Herman et al. | Sep 2010 | A1 |
20100228691 | Yang et al. | Sep 2010 | A1 |
20100229082 | Karmarkar et al. | Sep 2010 | A1 |
20100229100 | Miller et al. | Sep 2010 | A1 |
20100231474 | Yamagajo et al. | Sep 2010 | A1 |
20100235167 | Bourdon | Sep 2010 | A1 |
20100235341 | Bennett | Sep 2010 | A1 |
20100235729 | Kocienda et al. | Sep 2010 | A1 |
20100235732 | Bergman | Sep 2010 | A1 |
20100235770 | Ording et al. | Sep 2010 | A1 |
20100235780 | Westerman et al. | Sep 2010 | A1 |
20100241418 | Maeda et al. | Sep 2010 | A1 |
20100250542 | Fujimaki | Sep 2010 | A1 |
20100250599 | Schmidt et al. | Sep 2010 | A1 |
20100255858 | Juhasz | Oct 2010 | A1 |
20100257160 | Cao | Oct 2010 | A1 |
20100257478 | Longe et al. | Oct 2010 | A1 |
20100262599 | Nitz | Oct 2010 | A1 |
20100268537 | Al-Telmissani | Oct 2010 | A1 |
20100268539 | Xu et al. | Oct 2010 | A1 |
20100269040 | Lee | Oct 2010 | A1 |
20100274753 | Liberty et al. | Oct 2010 | A1 |
20100277579 | Cho et al. | Nov 2010 | A1 |
20100278320 | Arsenault et al. | Nov 2010 | A1 |
20100278453 | King | Nov 2010 | A1 |
20100280983 | Cho et al. | Nov 2010 | A1 |
20100281034 | Petrou et al. | Nov 2010 | A1 |
20100286984 | Wandinger et al. | Nov 2010 | A1 |
20100286985 | Kennewick et al. | Nov 2010 | A1 |
20100287514 | Cragun et al. | Nov 2010 | A1 |
20100290632 | Lin | Nov 2010 | A1 |
20100293460 | Budelli | Nov 2010 | A1 |
20100295645 | Falldin et al. | Nov 2010 | A1 |
20100299133 | Kopparapu et al. | Nov 2010 | A1 |
20100299138 | Kim | Nov 2010 | A1 |
20100299142 | Freeman et al. | Nov 2010 | A1 |
20100302056 | Dutton et al. | Dec 2010 | A1 |
20100304342 | Zilber | Dec 2010 | A1 |
20100304705 | Hursey et al. | Dec 2010 | A1 |
20100305807 | Basir et al. | Dec 2010 | A1 |
20100305947 | Schwarz et al. | Dec 2010 | A1 |
20100311395 | Zheng et al. | Dec 2010 | A1 |
20100312547 | Van Os et al. | Dec 2010 | A1 |
20100312566 | Odinak et al. | Dec 2010 | A1 |
20100318366 | Sullivan et al. | Dec 2010 | A1 |
20100318576 | Kim | Dec 2010 | A1 |
20100322438 | Siotis | Dec 2010 | A1 |
20100324709 | Starmen | Dec 2010 | A1 |
20100324895 | Kurzweil et al. | Dec 2010 | A1 |
20100324896 | Attwater et al. | Dec 2010 | A1 |
20100324905 | Kurzweil et al. | Dec 2010 | A1 |
20100325131 | Dumais et al. | Dec 2010 | A1 |
20100325158 | Oral et al. | Dec 2010 | A1 |
20100325573 | Estrada et al. | Dec 2010 | A1 |
20100325588 | Reddy et al. | Dec 2010 | A1 |
20100330908 | Maddern et al. | Dec 2010 | A1 |
20100332220 | Hursey et al. | Dec 2010 | A1 |
20100332224 | Mäkelä et al. | Dec 2010 | A1 |
20100332235 | David | Dec 2010 | A1 |
20100332236 | Tan | Dec 2010 | A1 |
20100332280 | Bradley et al. | Dec 2010 | A1 |
20100332348 | Cao | Dec 2010 | A1 |
20100332428 | Mchenry et al. | Dec 2010 | A1 |
20100332976 | Fux et al. | Dec 2010 | A1 |
20100333030 | Johns | Dec 2010 | A1 |
20100333163 | Daly | Dec 2010 | A1 |
20110002487 | Panther et al. | Jan 2011 | A1 |
20110004475 | Bellegarda | Jan 2011 | A1 |
20110006876 | Moberg et al. | Jan 2011 | A1 |
20110009107 | Guba et al. | Jan 2011 | A1 |
20110010178 | Lee et al. | Jan 2011 | A1 |
20110010644 | Merrill et al. | Jan 2011 | A1 |
20110015928 | Odell et al. | Jan 2011 | A1 |
20110016150 | Engstrom et al. | Jan 2011 | A1 |
20110016421 | Krupka et al. | Jan 2011 | A1 |
20110018695 | Bells et al. | Jan 2011 | A1 |
20110021211 | Ohki | Jan 2011 | A1 |
20110021213 | Carr | Jan 2011 | A1 |
20110022292 | Shen et al. | Jan 2011 | A1 |
20110022388 | Wu et al. | Jan 2011 | A1 |
20110022393 | Waller et al. | Jan 2011 | A1 |
20110022394 | Wide et al. | Jan 2011 | A1 |
20110022472 | Zon et al. | Jan 2011 | A1 |
20110022952 | Wu et al. | Jan 2011 | A1 |
20110029616 | Wang et al. | Feb 2011 | A1 |
20110030067 | Wilson | Feb 2011 | A1 |
20110033064 | Johnson et al. | Feb 2011 | A1 |
20110034183 | Haag et al. | Feb 2011 | A1 |
20110035144 | Okamoto et al. | Feb 2011 | A1 |
20110035434 | Lockwood | Feb 2011 | A1 |
20110038489 | Visser et al. | Feb 2011 | A1 |
20110040707 | Theisen et al. | Feb 2011 | A1 |
20110045841 | Kuhlke et al. | Feb 2011 | A1 |
20110047072 | Ciurea | Feb 2011 | A1 |
20110047149 | Vaananen | Feb 2011 | A1 |
20110047161 | Myaeng et al. | Feb 2011 | A1 |
20110047266 | Yu et al. | Feb 2011 | A1 |
20110047605 | Sontag et al. | Feb 2011 | A1 |
20110050591 | Kim et al. | Mar 2011 | A1 |
20110050592 | Kim et al. | Mar 2011 | A1 |
20110054647 | Chipchase | Mar 2011 | A1 |
20110054894 | Phillips et al. | Mar 2011 | A1 |
20110054901 | Qin et al. | Mar 2011 | A1 |
20110055256 | Phillips et al. | Mar 2011 | A1 |
20110060584 | Ferrucci et al. | Mar 2011 | A1 |
20110060587 | Phillips et al. | Mar 2011 | A1 |
20110060589 | Weinberg et al. | Mar 2011 | A1 |
20110060807 | Martin et al. | Mar 2011 | A1 |
20110064387 | Mendeloff et al. | Mar 2011 | A1 |
20110065456 | Brennan et al. | Mar 2011 | A1 |
20110066366 | Ellanti et al. | Mar 2011 | A1 |
20110066468 | Huang et al. | Mar 2011 | A1 |
20110066634 | Phillips et al. | Mar 2011 | A1 |
20110072492 | Mohler et al. | Mar 2011 | A1 |
20110076994 | Kim et al. | Mar 2011 | A1 |
20110077943 | Miki et al. | Mar 2011 | A1 |
20110080260 | Wang et al. | Apr 2011 | A1 |
20110081889 | Gao et al. | Apr 2011 | A1 |
20110082688 | Kim et al. | Apr 2011 | A1 |
20110083079 | Farrell et al. | Apr 2011 | A1 |
20110087491 | Wittenstein et al. | Apr 2011 | A1 |
20110087685 | Lin et al. | Apr 2011 | A1 |
20110090078 | Kim et al. | Apr 2011 | A1 |
20110092187 | Miller | Apr 2011 | A1 |
20110093261 | Angott | Apr 2011 | A1 |
20110093265 | Stent et al. | Apr 2011 | A1 |
20110093271 | Bernard et al. | Apr 2011 | A1 |
20110099000 | Rai et al. | Apr 2011 | A1 |
20110103682 | Chidlovskii et al. | May 2011 | A1 |
20110105097 | Tadayon et al. | May 2011 | A1 |
20110106736 | Aharonson et al. | May 2011 | A1 |
20110106892 | Nelson et al. | May 2011 | A1 |
20110110502 | Daye et al. | May 2011 | A1 |
20110111724 | Baptiste | May 2011 | A1 |
20110112827 | Kennewick et al. | May 2011 | A1 |
20110112837 | Kurki-Suonio et al. | May 2011 | A1 |
20110112838 | Adibi | May 2011 | A1 |
20110112921 | Kennewick et al. | May 2011 | A1 |
20110116610 | Shaw et al. | May 2011 | A1 |
20110119049 | Ylonen | May 2011 | A1 |
20110119051 | Li et al. | May 2011 | A1 |
20110119623 | Kim | May 2011 | A1 |
20110119715 | Chang et al. | May 2011 | A1 |
20110123004 | Chang et al. | May 2011 | A1 |
20110125498 | Pickering et al. | May 2011 | A1 |
20110125540 | Jang et al. | May 2011 | A1 |
20110125701 | Nair et al. | May 2011 | A1 |
20110130958 | Stahl et al. | Jun 2011 | A1 |
20110131036 | DiCristo et al. | Jun 2011 | A1 |
20110131038 | Oyaizu et al. | Jun 2011 | A1 |
20110131045 | Cristo et al. | Jun 2011 | A1 |
20110137636 | Srihari et al. | Jun 2011 | A1 |
20110141141 | Kankainen | Jun 2011 | A1 |
20110143726 | de Silva | Jun 2011 | A1 |
20110143811 | Rodriguez | Jun 2011 | A1 |
20110144857 | Wingrove et al. | Jun 2011 | A1 |
20110144901 | Wang | Jun 2011 | A1 |
20110144973 | Bocchieri et al. | Jun 2011 | A1 |
20110144999 | Jang et al. | Jun 2011 | A1 |
20110145718 | Ketola et al. | Jun 2011 | A1 |
20110151830 | Blanda et al. | Jun 2011 | A1 |
20110153209 | Geelen | Jun 2011 | A1 |
20110153322 | Kwak et al. | Jun 2011 | A1 |
20110153324 | Ballinger et al. | Jun 2011 | A1 |
20110153329 | Moorer | Jun 2011 | A1 |
20110153330 | Yazdani et al. | Jun 2011 | A1 |
20110153373 | Dantzig et al. | Jun 2011 | A1 |
20110154193 | Creutz et al. | Jun 2011 | A1 |
20110157029 | Tseng | Jun 2011 | A1 |
20110161072 | Terao et al. | Jun 2011 | A1 |
20110161076 | Davis et al. | Jun 2011 | A1 |
20110161079 | Gruhn et al. | Jun 2011 | A1 |
20110161309 | Lung et al. | Jun 2011 | A1 |
20110161852 | Vainio et al. | Jun 2011 | A1 |
20110166851 | LeBeau et al. | Jul 2011 | A1 |
20110167350 | Hoellwarth | Jul 2011 | A1 |
20110175810 | Markovic et al. | Jul 2011 | A1 |
20110179002 | Dumitru et al. | Jul 2011 | A1 |
20110179372 | Moore et al. | Jul 2011 | A1 |
20110183650 | Mckee et al. | Jul 2011 | A1 |
20110184721 | Subramanian et al. | Jul 2011 | A1 |
20110184730 | LeBeau et al. | Jul 2011 | A1 |
20110184736 | Slotznick | Jul 2011 | A1 |
20110184737 | Nakano et al. | Jul 2011 | A1 |
20110184768 | Norton et al. | Jul 2011 | A1 |
20110184789 | Kirsch | Jul 2011 | A1 |
20110185288 | Gupta et al. | Jul 2011 | A1 |
20110191108 | Friedlander | Aug 2011 | A1 |
20110191271 | Baker et al. | Aug 2011 | A1 |
20110191344 | Jin et al. | Aug 2011 | A1 |
20110195758 | Damale et al. | Aug 2011 | A1 |
20110196670 | Dang et al. | Aug 2011 | A1 |
20110197128 | Assadollahi et al. | Aug 2011 | A1 |
20110199312 | Okuta | Aug 2011 | A1 |
20110201385 | Higginbotham et al. | Aug 2011 | A1 |
20110201387 | Paek et al. | Aug 2011 | A1 |
20110202526 | Lee et al. | Aug 2011 | A1 |
20110205149 | Tom et al. | Aug 2011 | A1 |
20110208511 | Sikstrom et al. | Aug 2011 | A1 |
20110208524 | Haughay | Aug 2011 | A1 |
20110209088 | Hinckley et al. | Aug 2011 | A1 |
20110212717 | Rhoads et al. | Sep 2011 | A1 |
20110216093 | Griffin | Sep 2011 | A1 |
20110218806 | Alewine et al. | Sep 2011 | A1 |
20110218855 | Cao et al. | Sep 2011 | A1 |
20110219018 | Bailey et al. | Sep 2011 | A1 |
20110223893 | Lau et al. | Sep 2011 | A1 |
20110224972 | Millett et al. | Sep 2011 | A1 |
20110228913 | Cochinwala et al. | Sep 2011 | A1 |
20110231182 | Weider et al. | Sep 2011 | A1 |
20110231184 | Kerr | Sep 2011 | A1 |
20110231188 | Kennewick et al. | Sep 2011 | A1 |
20110231218 | Tovar | Sep 2011 | A1 |
20110231432 | Sata et al. | Sep 2011 | A1 |
20110231474 | Locker et al. | Sep 2011 | A1 |
20110238191 | Kristjansson et al. | Sep 2011 | A1 |
20110238407 | Kent | Sep 2011 | A1 |
20110238408 | Larcheveque et al. | Sep 2011 | A1 |
20110238676 | Liu et al. | Sep 2011 | A1 |
20110239111 | Grover | Sep 2011 | A1 |
20110242007 | Gray et al. | Oct 2011 | A1 |
20110244888 | Ohki | Oct 2011 | A1 |
20110246471 | Rakib et al. | Oct 2011 | A1 |
20110249144 | Chang | Oct 2011 | A1 |
20110250570 | Mack et al. | Oct 2011 | A1 |
20110257966 | Rychlik | Oct 2011 | A1 |
20110258188 | Abdalmageed et al. | Oct 2011 | A1 |
20110260829 | Lee | Oct 2011 | A1 |
20110260861 | Singh et al. | Oct 2011 | A1 |
20110264643 | Cao | Oct 2011 | A1 |
20110264999 | Bells et al. | Oct 2011 | A1 |
20110274303 | Filson et al. | Nov 2011 | A1 |
20110276595 | Kirkland et al. | Nov 2011 | A1 |
20110276598 | Kozempel | Nov 2011 | A1 |
20110276944 | Bergman et al. | Nov 2011 | A1 |
20110279368 | Klein et al. | Nov 2011 | A1 |
20110282663 | Talwar et al. | Nov 2011 | A1 |
20110282888 | Koperski et al. | Nov 2011 | A1 |
20110282906 | Wong | Nov 2011 | A1 |
20110283189 | McCarty | Nov 2011 | A1 |
20110288852 | Dymetman et al. | Nov 2011 | A1 |
20110288855 | Roy | Nov 2011 | A1 |
20110288861 | Kurzweil et al. | Nov 2011 | A1 |
20110288863 | Rasmussen | Nov 2011 | A1 |
20110288866 | Rasmussen | Nov 2011 | A1 |
20110288917 | Wanek et al. | Nov 2011 | A1 |
20110298585 | Barry | Dec 2011 | A1 |
20110301943 | Patch | Dec 2011 | A1 |
20110302162 | Xiao et al. | Dec 2011 | A1 |
20110306426 | Novak et al. | Dec 2011 | A1 |
20110307241 | Waibel et al. | Dec 2011 | A1 |
20110307491 | Fisk et al. | Dec 2011 | A1 |
20110307810 | Hilerio et al. | Dec 2011 | A1 |
20110313775 | Laligand et al. | Dec 2011 | A1 |
20110313803 | Friend et al. | Dec 2011 | A1 |
20110314003 | Ju et al. | Dec 2011 | A1 |
20110314032 | Bennett et al. | Dec 2011 | A1 |
20110314404 | Kotler | Dec 2011 | A1 |
20110320187 | Motik et al. | Dec 2011 | A1 |
20120002820 | Leichter | Jan 2012 | A1 |
20120005602 | Anttila et al. | Jan 2012 | A1 |
20120008754 | Mukherjee et al. | Jan 2012 | A1 |
20120010886 | Razavilar | Jan 2012 | A1 |
20120011138 | Dunning et al. | Jan 2012 | A1 |
20120013609 | Reponen et al. | Jan 2012 | A1 |
20120015629 | Olsen et al. | Jan 2012 | A1 |
20120016658 | Wu et al. | Jan 2012 | A1 |
20120016678 | Gruber et al. | Jan 2012 | A1 |
20120019400 | Patel et al. | Jan 2012 | A1 |
20120020490 | Leichter | Jan 2012 | A1 |
20120022787 | LeBeau et al. | Jan 2012 | A1 |
20120022857 | Baldwin et al. | Jan 2012 | A1 |
20120022860 | Lloyd et al. | Jan 2012 | A1 |
20120022868 | LeBeau et al. | Jan 2012 | A1 |
20120022869 | Lloyd et al. | Jan 2012 | A1 |
20120022870 | Kristjansson et al. | Jan 2012 | A1 |
20120022872 | Gruber et al. | Jan 2012 | A1 |
20120022874 | Lloyd et al. | Jan 2012 | A1 |
20120022876 | LeBeau et al. | Jan 2012 | A1 |
20120022967 | Bachman et al. | Jan 2012 | A1 |
20120023088 | Cheng et al. | Jan 2012 | A1 |
20120023095 | Wadycki et al. | Jan 2012 | A1 |
20120023462 | Rosing et al. | Jan 2012 | A1 |
20120029661 | Jones et al. | Feb 2012 | A1 |
20120029910 | Medlock et al. | Feb 2012 | A1 |
20120034904 | LeBeau et al. | Feb 2012 | A1 |
20120035907 | Lebeau et al. | Feb 2012 | A1 |
20120035908 | Lebeau et al. | Feb 2012 | A1 |
20120035924 | Jitkoff et al. | Feb 2012 | A1 |
20120035925 | Friend et al. | Feb 2012 | A1 |
20120035926 | Ambler | Feb 2012 | A1 |
20120035931 | LeBeau et al. | Feb 2012 | A1 |
20120035932 | Jitkoff et al. | Feb 2012 | A1 |
20120035935 | Park et al. | Feb 2012 | A1 |
20120036556 | LeBeau et al. | Feb 2012 | A1 |
20120039539 | Boiman et al. | Feb 2012 | A1 |
20120041752 | Wang et al. | Feb 2012 | A1 |
20120042014 | Desai et al. | Feb 2012 | A1 |
20120042343 | Laligand et al. | Feb 2012 | A1 |
20120053815 | Montanari et al. | Mar 2012 | A1 |
20120053829 | Agarwal et al. | Mar 2012 | A1 |
20120053945 | Gupta et al. | Mar 2012 | A1 |
20120056815 | Mehra | Mar 2012 | A1 |
20120059655 | Cartales | Mar 2012 | A1 |
20120059813 | Sejnoha et al. | Mar 2012 | A1 |
20120062473 | Xiao et al. | Mar 2012 | A1 |
20120066212 | Jennings | Mar 2012 | A1 |
20120066581 | Spalink | Mar 2012 | A1 |
20120075054 | Ge et al. | Mar 2012 | A1 |
20120077479 | Sabotta et al. | Mar 2012 | A1 |
20120078611 | Soltani et al. | Mar 2012 | A1 |
20120078624 | Yook et al. | Mar 2012 | A1 |
20120078627 | Wagner | Mar 2012 | A1 |
20120078635 | Rothkopf et al. | Mar 2012 | A1 |
20120078747 | Chakrabarti et al. | Mar 2012 | A1 |
20120082317 | Pance et al. | Apr 2012 | A1 |
20120083286 | Kim et al. | Apr 2012 | A1 |
20120084086 | Gilbert et al. | Apr 2012 | A1 |
20120084634 | Wong et al. | Apr 2012 | A1 |
20120088219 | Briscoe et al. | Apr 2012 | A1 |
20120089331 | Schmidt et al. | Apr 2012 | A1 |
20120101823 | Weng et al. | Apr 2012 | A1 |
20120108166 | Hymel | May 2012 | A1 |
20120108221 | Thomas et al. | May 2012 | A1 |
20120116770 | Chen et al. | May 2012 | A1 |
20120117499 | Mori et al. | May 2012 | A1 |
20120124126 | Alcazar et al. | May 2012 | A1 |
20120128322 | Shaffer et al. | May 2012 | A1 |
20120130709 | Bocchieri et al. | May 2012 | A1 |
20120136572 | Norton | May 2012 | A1 |
20120136855 | Ni et al. | May 2012 | A1 |
20120136985 | Popescu et al. | May 2012 | A1 |
20120137367 | Dupont et al. | May 2012 | A1 |
20120149342 | Cohen et al. | Jun 2012 | A1 |
20120149394 | Singh et al. | Jun 2012 | A1 |
20120150544 | McLoughlin et al. | Jun 2012 | A1 |
20120150580 | Norton | Jun 2012 | A1 |
20120158293 | Burnham | Jun 2012 | A1 |
20120158399 | Tremblay et al. | Jun 2012 | A1 |
20120158422 | Burnham et al. | Jun 2012 | A1 |
20120159380 | Kocienda et al. | Jun 2012 | A1 |
20120163710 | Skaff et al. | Jun 2012 | A1 |
20120166196 | Ju et al. | Jun 2012 | A1 |
20120166959 | Hilerio et al. | Jun 2012 | A1 |
20120173222 | Wang et al. | Jul 2012 | A1 |
20120173244 | Kwak et al. | Jul 2012 | A1 |
20120173464 | Tur et al. | Jul 2012 | A1 |
20120174121 | Treat et al. | Jul 2012 | A1 |
20120179457 | Newman et al. | Jul 2012 | A1 |
20120179467 | Williams | Jul 2012 | A1 |
20120185237 | Gajic et al. | Jul 2012 | A1 |
20120185480 | Ni et al. | Jul 2012 | A1 |
20120185781 | Guzman et al. | Jul 2012 | A1 |
20120191461 | Lin et al. | Jul 2012 | A1 |
20120192096 | Bowman et al. | Jul 2012 | A1 |
20120197743 | Grigg et al. | Aug 2012 | A1 |
20120197995 | Caruso | Aug 2012 | A1 |
20120197998 | Kessel et al. | Aug 2012 | A1 |
20120201362 | Crossan et al. | Aug 2012 | A1 |
20120209654 | Romagnino et al. | Aug 2012 | A1 |
20120209853 | Desai et al. | Aug 2012 | A1 |
20120209874 | Wong et al. | Aug 2012 | A1 |
20120210266 | Jiang et al. | Aug 2012 | A1 |
20120214141 | Raya et al. | Aug 2012 | A1 |
20120214517 | Singh et al. | Aug 2012 | A1 |
20120215762 | Hall et al. | Aug 2012 | A1 |
20120221339 | Wang et al. | Aug 2012 | A1 |
20120221552 | Reponen et al. | Aug 2012 | A1 |
20120223889 | Medlock et al. | Sep 2012 | A1 |
20120223936 | Aughey et al. | Sep 2012 | A1 |
20120232885 | Barbosa et al. | Sep 2012 | A1 |
20120232886 | Capuozzo et al. | Sep 2012 | A1 |
20120232906 | Lindahl et al. | Sep 2012 | A1 |
20120233266 | Hassan et al. | Sep 2012 | A1 |
20120239661 | Giblin | Sep 2012 | A1 |
20120239761 | Linner et al. | Sep 2012 | A1 |
20120242482 | Elumalai et al. | Sep 2012 | A1 |
20120245719 | Story, Jr. et al. | Sep 2012 | A1 |
20120245939 | Braho et al. | Sep 2012 | A1 |
20120245941 | Cheyer | Sep 2012 | A1 |
20120245944 | Gruber et al. | Sep 2012 | A1 |
20120246064 | Balkow | Sep 2012 | A1 |
20120250858 | Iqbal et al. | Oct 2012 | A1 |
20120252367 | Gaglio et al. | Oct 2012 | A1 |
20120252540 | Kirigaya | Oct 2012 | A1 |
20120253785 | Hamid et al. | Oct 2012 | A1 |
20120254143 | Varma et al. | Oct 2012 | A1 |
20120254152 | Park et al. | Oct 2012 | A1 |
20120254290 | Naaman | Oct 2012 | A1 |
20120259615 | Morin et al. | Oct 2012 | A1 |
20120265528 | Gruber et al. | Oct 2012 | A1 |
20120265535 | Bryant-Rich et al. | Oct 2012 | A1 |
20120265806 | Blanchflower et al. | Oct 2012 | A1 |
20120271625 | Bernard | Oct 2012 | A1 |
20120271634 | Lenke | Oct 2012 | A1 |
20120271635 | Ljolje | Oct 2012 | A1 |
20120271640 | Basir | Oct 2012 | A1 |
20120271676 | Aravamudan et al. | Oct 2012 | A1 |
20120275377 | Lehane et al. | Nov 2012 | A1 |
20120284015 | Drewes | Nov 2012 | A1 |
20120284027 | Mallett et al. | Nov 2012 | A1 |
20120290291 | Shelley et al. | Nov 2012 | A1 |
20120290300 | Lee et al. | Nov 2012 | A1 |
20120295708 | Hernandez-Abrego et al. | Nov 2012 | A1 |
20120296649 | Bansal et al. | Nov 2012 | A1 |
20120296654 | Hendrickson et al. | Nov 2012 | A1 |
20120296891 | Rangan | Nov 2012 | A1 |
20120297348 | Santoro | Nov 2012 | A1 |
20120303369 | Brush et al. | Nov 2012 | A1 |
20120303371 | Labsky et al. | Nov 2012 | A1 |
20120304124 | Chen et al. | Nov 2012 | A1 |
20120309363 | Gruber et al. | Dec 2012 | A1 |
20120310642 | Cao et al. | Dec 2012 | A1 |
20120310649 | Cannistraro et al. | Dec 2012 | A1 |
20120310652 | O'Sullivan | Dec 2012 | A1 |
20120310922 | Johnson et al. | Dec 2012 | A1 |
20120311478 | Van Os et al. | Dec 2012 | A1 |
20120311583 | Gruber et al. | Dec 2012 | A1 |
20120311584 | Gruber et al. | Dec 2012 | A1 |
20120311585 | Gruber et al. | Dec 2012 | A1 |
20120316862 | Sultan et al. | Dec 2012 | A1 |
20120316875 | Nyquist et al. | Dec 2012 | A1 |
20120316878 | Singleton et al. | Dec 2012 | A1 |
20120317194 | Tian | Dec 2012 | A1 |
20120317498 | Logan et al. | Dec 2012 | A1 |
20120321112 | Schubert et al. | Dec 2012 | A1 |
20120324391 | Tocci et al. | Dec 2012 | A1 |
20120327009 | Fleizach | Dec 2012 | A1 |
20120329529 | van der Raadt | Dec 2012 | A1 |
20120330660 | Jaiswal | Dec 2012 | A1 |
20120330661 | Lindahl | Dec 2012 | A1 |
20120330990 | Chen et al. | Dec 2012 | A1 |
20130002716 | Walker et al. | Jan 2013 | A1 |
20130005405 | Prociw | Jan 2013 | A1 |
20130006633 | Grokop et al. | Jan 2013 | A1 |
20130006637 | Kanevsky et al. | Jan 2013 | A1 |
20130006638 | Lindahl | Jan 2013 | A1 |
20130007648 | Gamon et al. | Jan 2013 | A1 |
20130010575 | He et al. | Jan 2013 | A1 |
20130013313 | Shechtman et al. | Jan 2013 | A1 |
20130013319 | Grant et al. | Jan 2013 | A1 |
20130018659 | Chi | Jan 2013 | A1 |
20130024576 | Dishneau et al. | Jan 2013 | A1 |
20130027875 | Zhu et al. | Jan 2013 | A1 |
20130030787 | Cancedda et al. | Jan 2013 | A1 |
20130030789 | Dalce | Jan 2013 | A1 |
20130030804 | Zavaliagko et al. | Jan 2013 | A1 |
20130030815 | Madhvanath et al. | Jan 2013 | A1 |
20130030955 | David | Jan 2013 | A1 |
20130031162 | Willis et al. | Jan 2013 | A1 |
20130031476 | Coin et al. | Jan 2013 | A1 |
20130033643 | Kim et al. | Feb 2013 | A1 |
20130035086 | Chardon et al. | Feb 2013 | A1 |
20130035942 | Kim et al. | Feb 2013 | A1 |
20130035961 | Yegnanarayanan | Feb 2013 | A1 |
20130041647 | Ramerth et al. | Feb 2013 | A1 |
20130041654 | Walker et al. | Feb 2013 | A1 |
20130041661 | Lee et al. | Feb 2013 | A1 |
20130041665 | Jang et al. | Feb 2013 | A1 |
20130041667 | Longe et al. | Feb 2013 | A1 |
20130041968 | Cohen et al. | Feb 2013 | A1 |
20130046544 | Kay et al. | Feb 2013 | A1 |
20130050089 | Neels et al. | Feb 2013 | A1 |
20130054550 | Bolohan | Feb 2013 | A1 |
20130054609 | Rajput et al. | Feb 2013 | A1 |
20130054613 | Bishop | Feb 2013 | A1 |
20130054675 | Jenkins et al. | Feb 2013 | A1 |
20130054706 | Graham et al. | Feb 2013 | A1 |
20130055099 | Yao et al. | Feb 2013 | A1 |
20130055147 | Vasudev et al. | Feb 2013 | A1 |
20130063611 | Papakipos et al. | Mar 2013 | A1 |
20130066832 | Sheehan et al. | Mar 2013 | A1 |
20130067307 | Tian et al. | Mar 2013 | A1 |
20130073286 | Bastea-Forte et al. | Mar 2013 | A1 |
20130073346 | Chun et al. | Mar 2013 | A1 |
20130080152 | Brun et al. | Mar 2013 | A1 |
20130080162 | Chang et al. | Mar 2013 | A1 |
20130080167 | Mozer | Mar 2013 | A1 |
20130080177 | Chen | Mar 2013 | A1 |
20130080251 | Dempski | Mar 2013 | A1 |
20130082967 | Hillis et al. | Apr 2013 | A1 |
20130085755 | Bringert et al. | Apr 2013 | A1 |
20130085761 | Bringert et al. | Apr 2013 | A1 |
20130090921 | Liu et al. | Apr 2013 | A1 |
20130091090 | Spivack et al. | Apr 2013 | A1 |
20130095805 | Lebeau et al. | Apr 2013 | A1 |
20130096909 | Brun et al. | Apr 2013 | A1 |
20130096917 | Edgar et al. | Apr 2013 | A1 |
20130097566 | Berglund | Apr 2013 | A1 |
20130097682 | Zeljkovic et al. | Apr 2013 | A1 |
20130100268 | Mihailidis et al. | Apr 2013 | A1 |
20130103391 | Millmore et al. | Apr 2013 | A1 |
20130103405 | Namba et al. | Apr 2013 | A1 |
20130106742 | Lee et al. | May 2013 | A1 |
20130110505 | Gruber et al. | May 2013 | A1 |
20130110515 | Guzzoni et al. | May 2013 | A1 |
20130110518 | Gruber et al. | May 2013 | A1 |
20130110519 | Cheyer et al. | May 2013 | A1 |
20130110520 | Cheyer et al. | May 2013 | A1 |
20130110943 | Menon et al. | May 2013 | A1 |
20130111330 | Staikos et al. | May 2013 | A1 |
20130111348 | Gruber et al. | May 2013 | A1 |
20130111487 | Cheyer et al. | May 2013 | A1 |
20130111581 | Griffin et al. | May 2013 | A1 |
20130115927 | Gruber et al. | May 2013 | A1 |
20130117022 | Chen et al. | May 2013 | A1 |
20130124189 | Baldwin et al. | May 2013 | A1 |
20130132084 | Stonehocker et al. | May 2013 | A1 |
20130132089 | Fanty et al. | May 2013 | A1 |
20130132871 | Zeng et al. | May 2013 | A1 |
20130141551 | Kim | Jun 2013 | A1 |
20130142317 | Reynolds | Jun 2013 | A1 |
20130142345 | Waldmann | Jun 2013 | A1 |
20130144594 | Bangalore et al. | Jun 2013 | A1 |
20130144616 | Bangalore et al. | Jun 2013 | A1 |
20130151339 | Kim et al. | Jun 2013 | A1 |
20130152092 | Yadgar et al. | Jun 2013 | A1 |
20130154811 | Ferren et al. | Jun 2013 | A1 |
20130157629 | Lee et al. | Jun 2013 | A1 |
20130158977 | Senior | Jun 2013 | A1 |
20130165232 | Nelson et al. | Jun 2013 | A1 |
20130166303 | Chang et al. | Jun 2013 | A1 |
20130166442 | Nakajima et al. | Jun 2013 | A1 |
20130170738 | Capuozzo et al. | Jul 2013 | A1 |
20130172022 | Seymour et al. | Jul 2013 | A1 |
20130176244 | Yamamoto et al. | Jul 2013 | A1 |
20130176592 | Sasaki | Jul 2013 | A1 |
20130179440 | Gordon | Jul 2013 | A1 |
20130183944 | Mozer et al. | Jul 2013 | A1 |
20130185059 | Riccardi et al. | Jul 2013 | A1 |
20130185074 | Gruber et al. | Jul 2013 | A1 |
20130185081 | Cheyer et al. | Jul 2013 | A1 |
20130185336 | Singh et al. | Jul 2013 | A1 |
20130187850 | Schulz et al. | Jul 2013 | A1 |
20130191117 | Atti et al. | Jul 2013 | A1 |
20130197911 | Wei et al. | Aug 2013 | A1 |
20130204813 | Master et al. | Aug 2013 | A1 |
20130204897 | McDougall | Aug 2013 | A1 |
20130207898 | Sullivan et al. | Aug 2013 | A1 |
20130218553 | Fujii et al. | Aug 2013 | A1 |
20130218560 | Hsiao et al. | Aug 2013 | A1 |
20130222249 | Pasquero et al. | Aug 2013 | A1 |
20130225128 | Gomar | Aug 2013 | A1 |
20130231917 | Naik | Sep 2013 | A1 |
20130234947 | Kristensson et al. | Sep 2013 | A1 |
20130235987 | Arroniz-Escobar et al. | Sep 2013 | A1 |
20130238647 | Thompson | Sep 2013 | A1 |
20130244615 | Miller et al. | Sep 2013 | A1 |
20130246048 | Nagase et al. | Sep 2013 | A1 |
20130246050 | Yu et al. | Sep 2013 | A1 |
20130246329 | Pasquero et al. | Sep 2013 | A1 |
20130253911 | Petri et al. | Sep 2013 | A1 |
20130253912 | Medlock et al. | Sep 2013 | A1 |
20130275117 | Winer | Oct 2013 | A1 |
20130275138 | Gruber et al. | Oct 2013 | A1 |
20130275164 | Gruber et al. | Oct 2013 | A1 |
20130275199 | Proctor, Jr. et al. | Oct 2013 | A1 |
20130275625 | Taivalsaari et al. | Oct 2013 | A1 |
20130275875 | Gruber et al. | Oct 2013 | A1 |
20130275899 | Schubert et al. | Oct 2013 | A1 |
20130283168 | Brown et al. | Oct 2013 | A1 |
20130289991 | Eshwar et al. | Oct 2013 | A1 |
20130289993 | Rao et al. | Oct 2013 | A1 |
20130289994 | Newman et al. | Oct 2013 | A1 |
20130291015 | Pan | Oct 2013 | A1 |
20130297317 | Lee et al. | Nov 2013 | A1 |
20130297319 | Kim | Nov 2013 | A1 |
20130297348 | Cardoza et al. | Nov 2013 | A1 |
20130300645 | Fedorov | Nov 2013 | A1 |
20130303106 | Martin | Nov 2013 | A1 |
20130304479 | Teller et al. | Nov 2013 | A1 |
20130304758 | Gruber et al. | Nov 2013 | A1 |
20130304815 | Puente et al. | Nov 2013 | A1 |
20130305119 | Kern et al. | Nov 2013 | A1 |
20130307855 | Lamb et al. | Nov 2013 | A1 |
20130307997 | O'Keefe et al. | Nov 2013 | A1 |
20130308922 | Sano et al. | Nov 2013 | A1 |
20130311997 | Gruber et al. | Nov 2013 | A1 |
20130316746 | Miller et al. | Nov 2013 | A1 |
20130322634 | Bennett et al. | Dec 2013 | A1 |
20130325436 | Wang et al. | Dec 2013 | A1 |
20130325443 | Begeja et al. | Dec 2013 | A1 |
20130325447 | Levien et al. | Dec 2013 | A1 |
20130325448 | Levien et al. | Dec 2013 | A1 |
20130325481 | Van Os et al. | Dec 2013 | A1 |
20130325484 | Chakladar et al. | Dec 2013 | A1 |
20130325967 | Parks et al. | Dec 2013 | A1 |
20130325979 | Mansfield et al. | Dec 2013 | A1 |
20130329023 | Suplee, III et al. | Dec 2013 | A1 |
20130331127 | Sabatelli et al. | Dec 2013 | A1 |
20130332159 | Federighi et al. | Dec 2013 | A1 |
20130332162 | Keen | Dec 2013 | A1 |
20130332164 | Nalk | Dec 2013 | A1 |
20130332168 | Kim et al. | Dec 2013 | A1 |
20130332400 | González | Dec 2013 | A1 |
20130339256 | Shroff | Dec 2013 | A1 |
20130346068 | Solem et al. | Dec 2013 | A1 |
20130346347 | Patterson et al. | Dec 2013 | A1 |
20140006012 | Zhou et al. | Jan 2014 | A1 |
20140006025 | Krishnan et al. | Jan 2014 | A1 |
20140006027 | Kim et al. | Jan 2014 | A1 |
20140006030 | Fleizach et al. | Jan 2014 | A1 |
20140006153 | Thangam et al. | Jan 2014 | A1 |
20140012580 | Ganong et al. | Jan 2014 | A1 |
20140012586 | Rubin et al. | Jan 2014 | A1 |
20140019116 | Lundberg et al. | Jan 2014 | A1 |
20140019133 | Bao et al. | Jan 2014 | A1 |
20140028735 | Williams et al. | Jan 2014 | A1 |
20140032453 | Eustice et al. | Jan 2014 | A1 |
20140033071 | Gruber et al. | Jan 2014 | A1 |
20140035823 | Khoe et al. | Feb 2014 | A1 |
20140039894 | Shostak | Feb 2014 | A1 |
20140040274 | Aravamudan et al. | Feb 2014 | A1 |
20140040748 | Lemay et al. | Feb 2014 | A1 |
20140040801 | Patel et al. | Feb 2014 | A1 |
20140040918 | Li et al. | Feb 2014 | A1 |
20140046934 | Zhou et al. | Feb 2014 | A1 |
20140047001 | Phillips et al. | Feb 2014 | A1 |
20140052680 | Nitz et al. | Feb 2014 | A1 |
20140052791 | Chakra et al. | Feb 2014 | A1 |
20140053082 | Park et al. | Feb 2014 | A1 |
20140053210 | Cheong et al. | Feb 2014 | A1 |
20140057610 | Olincy et al. | Feb 2014 | A1 |
20140059030 | Hakkani-Tur et al. | Feb 2014 | A1 |
20140067361 | Nikoulina et al. | Mar 2014 | A1 |
20140067371 | Liensberger | Mar 2014 | A1 |
20140067402 | Kim | Mar 2014 | A1 |
20140068751 | Last et al. | Mar 2014 | A1 |
20140074466 | Sharifi et al. | Mar 2014 | A1 |
20140074470 | Jansche et al. | Mar 2014 | A1 |
20140074472 | Lin et al. | Mar 2014 | A1 |
20140074815 | Plimton | Mar 2014 | A1 |
20140078065 | Akkok et al. | Mar 2014 | A1 |
20140080428 | Rhoads et al. | Mar 2014 | A1 |
20140081633 | Badaskar et al. | Mar 2014 | A1 |
20140082501 | Bae et al. | Mar 2014 | A1 |
20140086458 | Rogers et al. | Mar 2014 | A1 |
20140087711 | Geyer et al. | Mar 2014 | A1 |
20140088961 | Woodward et al. | Mar 2014 | A1 |
20140095171 | Lynch et al. | Apr 2014 | A1 |
20140095172 | Cabaco et al. | Apr 2014 | A1 |
20140095173 | Lynch et al. | Apr 2014 | A1 |
20140096209 | Saraf et al. | Apr 2014 | A1 |
20140098247 | Rao et al. | Apr 2014 | A1 |
20140108017 | Mason et al. | Apr 2014 | A1 |
20140114554 | Lagassey | Apr 2014 | A1 |
20140118155 | Bowers et al. | May 2014 | A1 |
20140122059 | Patel et al. | May 2014 | A1 |
20140122086 | Kapur et al. | May 2014 | A1 |
20140122136 | Jayanthi | May 2014 | A1 |
20140122153 | Truitt | May 2014 | A1 |
20140134983 | Jung et al. | May 2014 | A1 |
20140135036 | Bonanni et al. | May 2014 | A1 |
20140136187 | Wolverton et al. | May 2014 | A1 |
20140136195 | Abdossalami et al. | May 2014 | A1 |
20140136212 | Kwon et al. | May 2014 | A1 |
20140136946 | Matas | May 2014 | A1 |
20140142923 | Jones et al. | May 2014 | A1 |
20140142935 | Lindahl et al. | May 2014 | A1 |
20140143550 | Ganong, III et al. | May 2014 | A1 |
20140143721 | Suzuki et al. | May 2014 | A1 |
20140146200 | Scott et al. | May 2014 | A1 |
20140152577 | Yuen et al. | Jun 2014 | A1 |
20140155031 | Lee et al. | Jun 2014 | A1 |
20140157422 | Livshits et al. | Jun 2014 | A1 |
20140163951 | Nikoulina et al. | Jun 2014 | A1 |
20140163953 | Parikh | Jun 2014 | A1 |
20140163954 | Joshi et al. | Jun 2014 | A1 |
20140163981 | Cook et al. | Jun 2014 | A1 |
20140164476 | Thomson | Jun 2014 | A1 |
20140164532 | Lynch et al. | Jun 2014 | A1 |
20140164533 | Lynch et al. | Jun 2014 | A1 |
20140169795 | Clough | Jun 2014 | A1 |
20140173460 | Kim | Jun 2014 | A1 |
20140180499 | Cooper et al. | Jun 2014 | A1 |
20140180689 | Kim et al. | Jun 2014 | A1 |
20140180697 | Torok et al. | Jun 2014 | A1 |
20140181865 | Koganei | Jun 2014 | A1 |
20140188477 | Zhang | Jul 2014 | A1 |
20140195230 | Han et al. | Jul 2014 | A1 |
20140195233 | Bapat | Jul 2014 | A1 |
20140195244 | Cha et al. | Jul 2014 | A1 |
20140195251 | Zeinstra et al. | Jul 2014 | A1 |
20140195252 | Gruber et al. | Jul 2014 | A1 |
20140198048 | Unruh et al. | Jul 2014 | A1 |
20140203939 | Harrington et al. | Jul 2014 | A1 |
20140207439 | Venkatapathy et al. | Jul 2014 | A1 |
20140207446 | Klein et al. | Jul 2014 | A1 |
20140207468 | Bartnik | Jul 2014 | A1 |
20140207582 | Flinn et al. | Jul 2014 | A1 |
20140214429 | Pantel | Jul 2014 | A1 |
20140214537 | Yoo et al. | Jul 2014 | A1 |
20140218372 | Missig et al. | Aug 2014 | A1 |
20140222436 | Binder et al. | Aug 2014 | A1 |
20140222678 | Sheets et al. | Aug 2014 | A1 |
20140223377 | Shaw et al. | Aug 2014 | A1 |
20140223481 | Fundament | Aug 2014 | A1 |
20140230055 | Boehl | Aug 2014 | A1 |
20140232656 | Pasquero et al. | Aug 2014 | A1 |
20140236595 | Gray | Aug 2014 | A1 |
20140236986 | Guzman | Aug 2014 | A1 |
20140237042 | Ahmed et al. | Aug 2014 | A1 |
20140244248 | Arisoy et al. | Aug 2014 | A1 |
20140244254 | Ju et al. | Aug 2014 | A1 |
20140244257 | Colibro et al. | Aug 2014 | A1 |
20140244258 | Song et al. | Aug 2014 | A1 |
20140244263 | Pontual et al. | Aug 2014 | A1 |
20140244268 | Abdelsamie et al. | Aug 2014 | A1 |
20140244271 | Lindahl | Aug 2014 | A1 |
20140244712 | Walters et al. | Aug 2014 | A1 |
20140245140 | Brown et al. | Aug 2014 | A1 |
20140247383 | Dave et al. | Sep 2014 | A1 |
20140247926 | Gainsboro et al. | Sep 2014 | A1 |
20140249817 | Hart et al. | Sep 2014 | A1 |
20140249821 | Kennewick et al. | Sep 2014 | A1 |
20140250046 | Winn et al. | Sep 2014 | A1 |
20140257815 | Zhao et al. | Sep 2014 | A1 |
20140258857 | Dykstra-Erickson et al. | Sep 2014 | A1 |
20140267022 | Kim | Sep 2014 | A1 |
20140267599 | Drouin et al. | Sep 2014 | A1 |
20140272821 | Pitschel et al. | Sep 2014 | A1 |
20140274203 | Ganong et al. | Sep 2014 | A1 |
20140274211 | Sejnoha et al. | Sep 2014 | A1 |
20140278343 | Tran | Sep 2014 | A1 |
20140278349 | Grieves et al. | Sep 2014 | A1 |
20140278379 | Coccaro et al. | Sep 2014 | A1 |
20140278390 | Kingsbury et al. | Sep 2014 | A1 |
20140278391 | Braho et al. | Sep 2014 | A1 |
20140278394 | Bastyr et al. | Sep 2014 | A1 |
20140278406 | Tsumura et al. | Sep 2014 | A1 |
20140278413 | Pitschel et al. | Sep 2014 | A1 |
20140278429 | Ganong, III | Sep 2014 | A1 |
20140278435 | Ganong et al. | Sep 2014 | A1 |
20140278443 | Gunn et al. | Sep 2014 | A1 |
20140278513 | Prakash et al. | Sep 2014 | A1 |
20140280138 | Li et al. | Sep 2014 | A1 |
20140280292 | Skinder | Sep 2014 | A1 |
20140280353 | Delaney et al. | Sep 2014 | A1 |
20140280450 | Luna | Sep 2014 | A1 |
20140281983 | Xian et al. | Sep 2014 | A1 |
20140282003 | Gruber et al. | Sep 2014 | A1 |
20140282007 | Fleizach | Sep 2014 | A1 |
20140282045 | Ayanam et al. | Sep 2014 | A1 |
20140282201 | Pasquero et al. | Sep 2014 | A1 |
20140282586 | Shear et al. | Sep 2014 | A1 |
20140282743 | Howard et al. | Sep 2014 | A1 |
20140288990 | Moore et al. | Sep 2014 | A1 |
20140289508 | Wang | Sep 2014 | A1 |
20140297267 | Spencer et al. | Oct 2014 | A1 |
20140297281 | Togawa et al. | Oct 2014 | A1 |
20140297284 | Gruber et al. | Oct 2014 | A1 |
20140297288 | Yu et al. | Oct 2014 | A1 |
20140304605 | Ohmura et al. | Oct 2014 | A1 |
20140310001 | Kains et al. | Oct 2014 | A1 |
20140310002 | Nitz et al. | Oct 2014 | A1 |
20140316585 | Boesveld et al. | Oct 2014 | A1 |
20140317030 | Shen et al. | Oct 2014 | A1 |
20140317502 | Brown et al. | Oct 2014 | A1 |
20140324884 | Lindahl et al. | Oct 2014 | A1 |
20140337048 | Brown et al. | Nov 2014 | A1 |
20140337266 | Wolverton et al. | Nov 2014 | A1 |
20140337371 | Li | Nov 2014 | A1 |
20140337438 | Govande et al. | Nov 2014 | A1 |
20140337751 | Lim et al. | Nov 2014 | A1 |
20140337814 | Kains et al. | Nov 2014 | A1 |
20140342762 | Hajdu et al. | Nov 2014 | A1 |
20140344627 | Schaub et al. | Nov 2014 | A1 |
20140344687 | Durham et al. | Nov 2014 | A1 |
20140350924 | Zurek et al. | Nov 2014 | A1 |
20140350933 | Bak et al. | Nov 2014 | A1 |
20140351741 | Medlock et al. | Nov 2014 | A1 |
20140351760 | Skory et al. | Nov 2014 | A1 |
20140358519 | Mirkin et al. | Dec 2014 | A1 |
20140358523 | Sheth et al. | Dec 2014 | A1 |
20140361973 | Raux et al. | Dec 2014 | A1 |
20140365209 | Evermann | Dec 2014 | A1 |
20140365214 | Bayley | Dec 2014 | A1 |
20140365216 | Gruber et al. | Dec 2014 | A1 |
20140365226 | Sinha | Dec 2014 | A1 |
20140365227 | Cash et al. | Dec 2014 | A1 |
20140365407 | Brown et al. | Dec 2014 | A1 |
20140365880 | Bellegarda | Dec 2014 | A1 |
20140365885 | Carson et al. | Dec 2014 | A1 |
20140365895 | Paulson et al. | Dec 2014 | A1 |
20140370817 | Luna | Dec 2014 | A1 |
20140370841 | Roberts et al. | Dec 2014 | A1 |
20140372112 | Xue et al. | Dec 2014 | A1 |
20140372356 | Bilal et al. | Dec 2014 | A1 |
20140372931 | Zhai et al. | Dec 2014 | A1 |
20140379334 | Fry | Dec 2014 | A1 |
20150003797 | Schmidt | Jan 2015 | A1 |
20150006148 | Goldszmit et al. | Jan 2015 | A1 |
20150006157 | Andrade Silva et al. | Jan 2015 | A1 |
20150006176 | Pogue et al. | Jan 2015 | A1 |
20150006178 | Peng et al. | Jan 2015 | A1 |
20150006184 | Marti et al. | Jan 2015 | A1 |
20150006199 | Snider et al. | Jan 2015 | A1 |
20150012271 | Peng et al. | Jan 2015 | A1 |
20150019219 | Tzirkel-Hancock et al. | Jan 2015 | A1 |
20150019221 | Lee et al. | Jan 2015 | A1 |
20150031416 | Wells et al. | Jan 2015 | A1 |
20150033219 | Breiner et al. | Jan 2015 | A1 |
20150039292 | Suleman et al. | Feb 2015 | A1 |
20150039299 | Weinstein et al. | Feb 2015 | A1 |
20150039305 | Huang | Feb 2015 | A1 |
20150040012 | Faaborg et al. | Feb 2015 | A1 |
20150045003 | Vora et al. | Feb 2015 | A1 |
20150045068 | Soffer et al. | Feb 2015 | A1 |
20150046537 | Rakib | Feb 2015 | A1 |
20150050633 | Christmas et al. | Feb 2015 | A1 |
20150058013 | Pakhomov et al. | Feb 2015 | A1 |
20150058018 | Georges et al. | Feb 2015 | A1 |
20150058785 | Ookawara | Feb 2015 | A1 |
20150065200 | Namgung et al. | Mar 2015 | A1 |
20150066494 | Salvador et al. | Mar 2015 | A1 |
20150066496 | Deoras et al. | Mar 2015 | A1 |
20150066506 | Romano et al. | Mar 2015 | A1 |
20150066516 | Nishikawa et al. | Mar 2015 | A1 |
20150067485 | Kim et al. | Mar 2015 | A1 |
20150067822 | Randall | Mar 2015 | A1 |
20150073788 | Allauzen et al. | Mar 2015 | A1 |
20150073804 | Senior et al. | Mar 2015 | A1 |
20150074524 | Nicholson et al. | Mar 2015 | A1 |
20150082229 | Ouyang et al. | Mar 2015 | A1 |
20150088511 | Bharadwaj et al. | Mar 2015 | A1 |
20150088514 | Typrin | Mar 2015 | A1 |
20150088522 | Hendrickson et al. | Mar 2015 | A1 |
20150088523 | Schuster | Mar 2015 | A1 |
20150095031 | Conkie et al. | Apr 2015 | A1 |
20150095278 | Flinn et al. | Apr 2015 | A1 |
20150100316 | Williams et al. | Apr 2015 | A1 |
20150100537 | Grieves et al. | Apr 2015 | A1 |
20150100983 | Pan | Apr 2015 | A1 |
20150106093 | Weeks et al. | Apr 2015 | A1 |
20150113407 | Hoffert et al. | Apr 2015 | A1 |
20150120723 | Deshmukh et al. | Apr 2015 | A1 |
20150121216 | Brown et al. | Apr 2015 | A1 |
20150127350 | Agiomyrgiannakis | May 2015 | A1 |
20150133109 | Freeman et al. | May 2015 | A1 |
20150134334 | Sachidanandam et al. | May 2015 | A1 |
20150135085 | Shoham et al. | May 2015 | A1 |
20150135123 | Carr et al. | May 2015 | A1 |
20150142420 | Sarikaya et al. | May 2015 | A1 |
20150142438 | Dai et al. | May 2015 | A1 |
20150142447 | Kennewick et al. | May 2015 | A1 |
20150142851 | Gupta et al. | May 2015 | A1 |
20150148013 | Baldwin et al. | May 2015 | A1 |
20150149177 | Kalns et al. | May 2015 | A1 |
20150149182 | Kalns et al. | May 2015 | A1 |
20150149354 | Mccoy | May 2015 | A1 |
20150149469 | Xu et al. | May 2015 | A1 |
20150154185 | Waibel | Jun 2015 | A1 |
20150161370 | North et al. | Jun 2015 | A1 |
20150161989 | Hsu et al. | Jun 2015 | A1 |
20150162001 | Kar et al. | Jun 2015 | A1 |
20150163558 | Wheatley | Jun 2015 | A1 |
20150169336 | Harper et al. | Jun 2015 | A1 |
20150170664 | Doherty et al. | Jun 2015 | A1 |
20150172463 | Quast et al. | Jun 2015 | A1 |
20150178388 | Winnemoeller et al. | Jun 2015 | A1 |
20150179176 | Ryu et al. | Jun 2015 | A1 |
20150185964 | Stout | Jul 2015 | A1 |
20150186012 | Coleman et al. | Jul 2015 | A1 |
20150186110 | Kannan | Jul 2015 | A1 |
20150186155 | Brown et al. | Jul 2015 | A1 |
20150186156 | Brown et al. | Jul 2015 | A1 |
20150186351 | Hicks et al. | Jul 2015 | A1 |
20150187355 | Parkinson et al. | Jul 2015 | A1 |
20150187369 | Dadu et al. | Jul 2015 | A1 |
20150189362 | Lee et al. | Jul 2015 | A1 |
20150193379 | Mehta | Jul 2015 | A1 |
20150193391 | Khvostichenko et al. | Jul 2015 | A1 |
20150193392 | Greenblatt et al. | Jul 2015 | A1 |
20150194152 | Katuri et al. | Jul 2015 | A1 |
20150195379 | Zhang et al. | Jul 2015 | A1 |
20150195606 | McDevitt | Jul 2015 | A1 |
20150199077 | Zuger et al. | Jul 2015 | A1 |
20150199960 | Huo et al. | Jul 2015 | A1 |
20150199965 | Leak et al. | Jul 2015 | A1 |
20150199967 | Reddy et al. | Jul 2015 | A1 |
20150201064 | Bells et al. | Jul 2015 | A1 |
20150205858 | Xie et al. | Jul 2015 | A1 |
20150208226 | Kuusilinna et al. | Jul 2015 | A1 |
20150212791 | Kumar et al. | Jul 2015 | A1 |
20150213796 | Waltermann et al. | Jul 2015 | A1 |
20150220507 | Mohajer et al. | Aug 2015 | A1 |
20150221304 | Stewart | Aug 2015 | A1 |
20150221307 | Shah et al. | Aug 2015 | A1 |
20150227633 | Shapira | Aug 2015 | A1 |
20150228281 | Raniere | Aug 2015 | A1 |
20150234636 | Barnes, Jr. | Aug 2015 | A1 |
20150234800 | Patrick et al. | Aug 2015 | A1 |
20150242091 | Lu et al. | Aug 2015 | A1 |
20150243278 | Kibre et al. | Aug 2015 | A1 |
20150243283 | Halash et al. | Aug 2015 | A1 |
20150245154 | Dadu et al. | Aug 2015 | A1 |
20150248651 | Akutagawa et al. | Sep 2015 | A1 |
20150248886 | Sarikaya et al. | Sep 2015 | A1 |
20150254057 | Klein et al. | Sep 2015 | A1 |
20150254058 | Klein et al. | Sep 2015 | A1 |
20150254333 | Fife et al. | Sep 2015 | A1 |
20150255071 | Chiba | Sep 2015 | A1 |
20150256873 | Klein et al. | Sep 2015 | A1 |
20150261496 | Faaborg et al. | Sep 2015 | A1 |
20150269139 | McAteer et al. | Sep 2015 | A1 |
20150277574 | Jain et al. | Oct 2015 | A1 |
20150278370 | Stratvert et al. | Oct 2015 | A1 |
20150279358 | Kingsbury et al. | Oct 2015 | A1 |
20150279360 | Mengibar et al. | Oct 2015 | A1 |
20150281380 | Wang et al. | Oct 2015 | A1 |
20150286627 | Chang et al. | Oct 2015 | A1 |
20150287401 | Lee et al. | Oct 2015 | A1 |
20150287409 | Jang | Oct 2015 | A1 |
20150288629 | Choi et al. | Oct 2015 | A1 |
20150294086 | Kare et al. | Oct 2015 | A1 |
20150294516 | Chiang | Oct 2015 | A1 |
20150295915 | Xiu | Oct 2015 | A1 |
20150302855 | Kim et al. | Oct 2015 | A1 |
20150302856 | Kim et al. | Oct 2015 | A1 |
20150302857 | Yamada | Oct 2015 | A1 |
20150309997 | Lee et al. | Oct 2015 | A1 |
20150310858 | Li et al. | Oct 2015 | A1 |
20150310862 | Dauphin et al. | Oct 2015 | A1 |
20150310879 | Buchanan et al. | Oct 2015 | A1 |
20150312182 | Langholz | Oct 2015 | A1 |
20150317069 | Clements et al. | Nov 2015 | A1 |
20150317310 | Eiche et al. | Nov 2015 | A1 |
20150324041 | Varley et al. | Nov 2015 | A1 |
20150324334 | Lee et al. | Nov 2015 | A1 |
20150331664 | Osawa et al. | Nov 2015 | A1 |
20150331711 | Huang et al. | Nov 2015 | A1 |
20150332667 | Mason | Nov 2015 | A1 |
20150339049 | Kasemset et al. | Nov 2015 | A1 |
20150339391 | Kang et al. | Nov 2015 | A1 |
20150340040 | Mun et al. | Nov 2015 | A1 |
20150340042 | Sejnoha et al. | Nov 2015 | A1 |
20150341717 | Song et al. | Nov 2015 | A1 |
20150347086 | Liedholm et al. | Dec 2015 | A1 |
20150347381 | Bellegarda | Dec 2015 | A1 |
20150347382 | Dolfing et al. | Dec 2015 | A1 |
20150347385 | Flor et al. | Dec 2015 | A1 |
20150347393 | Futrell et al. | Dec 2015 | A1 |
20150347733 | Tsou et al. | Dec 2015 | A1 |
20150347985 | Gross et al. | Dec 2015 | A1 |
20150348547 | Paulik et al. | Dec 2015 | A1 |
20150348548 | Piernot et al. | Dec 2015 | A1 |
20150348549 | Giuli et al. | Dec 2015 | A1 |
20150348551 | Gruber et al. | Dec 2015 | A1 |
20150348554 | Orr et al. | Dec 2015 | A1 |
20150350031 | Burks et al. | Dec 2015 | A1 |
20150352999 | Bando et al. | Dec 2015 | A1 |
20150355879 | Beckhardt et al. | Dec 2015 | A1 |
20150370531 | Faaborg | Dec 2015 | A1 |
20150370780 | Wang et al. | Dec 2015 | A1 |
20150371639 | Foerster et al. | Dec 2015 | A1 |
20150371665 | Naik et al. | Dec 2015 | A1 |
20150373183 | Woolsey et al. | Dec 2015 | A1 |
20150382047 | Napolitano et al. | Dec 2015 | A1 |
20150382079 | Lister et al. | Dec 2015 | A1 |
20160004690 | Bangalore et al. | Jan 2016 | A1 |
20160014476 | Caliendo, Jr. et al. | Jan 2016 | A1 |
20160019886 | Hong | Jan 2016 | A1 |
20160026258 | Ou et al. | Jan 2016 | A1 |
20160027431 | Kurzweil et al. | Jan 2016 | A1 |
20160028666 | Li | Jan 2016 | A1 |
20160029316 | Mohan et al. | Jan 2016 | A1 |
20160034811 | Paulik et al. | Feb 2016 | A1 |
20160042735 | Vibbert et al. | Feb 2016 | A1 |
20160042748 | Jain et al. | Feb 2016 | A1 |
20160048666 | Dey et al. | Feb 2016 | A1 |
20160055422 | Li | Feb 2016 | A1 |
20160062605 | Agarwal et al. | Mar 2016 | A1 |
20160063998 | Krishnamoorthy et al. | Mar 2016 | A1 |
20160070581 | Soon-Shiong | Mar 2016 | A1 |
20160071516 | Lee et al. | Mar 2016 | A1 |
20160071521 | Haughay | Mar 2016 | A1 |
20160077794 | Kim et al. | Mar 2016 | A1 |
20160078860 | Paulik et al. | Mar 2016 | A1 |
20160080165 | Ehsani et al. | Mar 2016 | A1 |
20160086116 | Rao et al. | Mar 2016 | A1 |
20160091967 | Prokofieva et al. | Mar 2016 | A1 |
20160092447 | Venkataraman et al. | Mar 2016 | A1 |
20160093291 | Kim | Mar 2016 | A1 |
20160093298 | Naik et al. | Mar 2016 | A1 |
20160093301 | Bellegarda et al. | Mar 2016 | A1 |
20160093304 | Kim et al. | Mar 2016 | A1 |
20160094979 | Naik et al. | Mar 2016 | A1 |
20160117386 | Ajmera et al. | Apr 2016 | A1 |
20160119338 | Cheyer | Apr 2016 | A1 |
20160125048 | Hamada | May 2016 | A1 |
20160125071 | Gabbai | May 2016 | A1 |
20160132484 | Nauze et al. | May 2016 | A1 |
20160139662 | Dabhade | May 2016 | A1 |
20160147725 | Patten et al. | May 2016 | A1 |
20160148610 | Kennewick, Jr. et al. | May 2016 | A1 |
20160155442 | Kannan et al. | Jun 2016 | A1 |
20160155443 | Khan et al. | Jun 2016 | A1 |
20160162456 | Munro et al. | Jun 2016 | A1 |
20160163312 | Naik et al. | Jun 2016 | A1 |
20160170966 | Kolo | Jun 2016 | A1 |
20160173578 | Sharma et al. | Jun 2016 | A1 |
20160173960 | Snibbe et al. | Jun 2016 | A1 |
20160179462 | Bjorkengren | Jun 2016 | A1 |
20160180844 | Vanblon et al. | Jun 2016 | A1 |
20160182410 | Janakiraman et al. | Jun 2016 | A1 |
20160188181 | Smith | Jun 2016 | A1 |
20160188738 | Gruber et al. | Jun 2016 | A1 |
20160189717 | Kannan et al. | Jun 2016 | A1 |
20160210981 | Lee | Jul 2016 | A1 |
20160212488 | Os et al. | Jul 2016 | A1 |
20160217784 | Gelfenbeyn et al. | Jul 2016 | A1 |
20160224540 | Stewart et al. | Aug 2016 | A1 |
20160224774 | Pender | Aug 2016 | A1 |
20160225372 | Cheung et al. | Aug 2016 | A1 |
20160240187 | Fleizach et al. | Aug 2016 | A1 |
20160247061 | Trask et al. | Aug 2016 | A1 |
20160253312 | Rhodes | Sep 2016 | A1 |
20160259656 | Sumner et al. | Sep 2016 | A1 |
20160260431 | Newendorp et al. | Sep 2016 | A1 |
20160260433 | Sumner et al. | Sep 2016 | A1 |
20160260436 | Lemay et al. | Sep 2016 | A1 |
20160266871 | Schmid et al. | Sep 2016 | A1 |
20160267904 | Biadsy et al. | Sep 2016 | A1 |
20160275941 | Bellegarda et al. | Sep 2016 | A1 |
20160275947 | Li et al. | Sep 2016 | A1 |
20160282956 | Ouyang et al. | Sep 2016 | A1 |
20160284199 | Dotan-Cohen et al. | Sep 2016 | A1 |
20160286045 | Shaltiel et al. | Sep 2016 | A1 |
20160299685 | Zhai et al. | Oct 2016 | A1 |
20160299882 | Hegerty et al. | Oct 2016 | A1 |
20160299883 | Zhu et al. | Oct 2016 | A1 |
20160307566 | Bellegarda | Oct 2016 | A1 |
20160314788 | Jitkoff et al. | Oct 2016 | A1 |
20160314792 | Alvarez et al. | Oct 2016 | A1 |
20160321261 | Spasojevic et al. | Nov 2016 | A1 |
20160322045 | Hatfeild et al. | Nov 2016 | A1 |
20160322050 | Wang et al. | Nov 2016 | A1 |
20160328893 | Cordova et al. | Nov 2016 | A1 |
20160336007 | Hanazawa | Nov 2016 | A1 |
20160336010 | Lindahl | Nov 2016 | A1 |
20160336024 | Choi et al. | Nov 2016 | A1 |
20160337299 | Lane et al. | Nov 2016 | A1 |
20160337301 | Rollins et al. | Nov 2016 | A1 |
20160342685 | Basu et al. | Nov 2016 | A1 |
20160351190 | Binder et al. | Dec 2016 | A1 |
20160357304 | Hatori et al. | Dec 2016 | A1 |
20160357728 | Bellegarda et al. | Dec 2016 | A1 |
20160357861 | Carlhian et al. | Dec 2016 | A1 |
20160357870 | Hentschel et al. | Dec 2016 | A1 |
20160358598 | Williams et al. | Dec 2016 | A1 |
20160358600 | Nallasamy et al. | Dec 2016 | A1 |
20160358619 | Ramprashad et al. | Dec 2016 | A1 |
20160359771 | Sridhar | Dec 2016 | A1 |
20160360039 | Sanghavi et al. | Dec 2016 | A1 |
20160364378 | Futrell et al. | Dec 2016 | A1 |
20160371250 | Rhodes | Dec 2016 | A1 |
20160378747 | Orr et al. | Dec 2016 | A1 |
20160379641 | Liu et al. | Dec 2016 | A1 |
20170004824 | Yoo et al. | Jan 2017 | A1 |
20170019987 | Dragone et al. | Jan 2017 | A1 |
20170031576 | Saoji et al. | Feb 2017 | A1 |
20170040002 | Basson et al. | Feb 2017 | A1 |
20170055895 | Des Jardins et al. | Mar 2017 | A1 |
20170060853 | Lee et al. | Mar 2017 | A1 |
20170068423 | Napolitano et al. | Mar 2017 | A1 |
20170068513 | Stasior et al. | Mar 2017 | A1 |
20170068670 | Orr et al. | Mar 2017 | A1 |
20170083179 | Gruber et al. | Mar 2017 | A1 |
20170091168 | Bellegarda et al. | Mar 2017 | A1 |
20170092270 | Newendorp et al. | Mar 2017 | A1 |
20170092278 | Evermann et al. | Mar 2017 | A1 |
20170105190 | Logan et al. | Apr 2017 | A1 |
20170116989 | Yadgar et al. | Apr 2017 | A1 |
20170127124 | Wilson et al. | May 2017 | A9 |
20170132199 | Vescovi et al. | May 2017 | A1 |
20170140644 | Hwang et al. | May 2017 | A1 |
20170161018 | Lemay et al. | Jun 2017 | A1 |
20170169819 | Mese et al. | Jun 2017 | A1 |
20170178619 | Naik et al. | Jun 2017 | A1 |
20170178626 | Gruber et al. | Jun 2017 | A1 |
20170180499 | Gelfenbeyn et al. | Jun 2017 | A1 |
20170186429 | Giuli et al. | Jun 2017 | A1 |
20170193083 | Bhatt et al. | Jul 2017 | A1 |
20170199874 | Patel et al. | Jul 2017 | A1 |
20170227935 | Su et al. | Aug 2017 | A1 |
20170230709 | Van Os et al. | Aug 2017 | A1 |
20170242653 | Lang et al. | Aug 2017 | A1 |
20170243468 | Dotan-Cohen et al. | Aug 2017 | A1 |
20170256256 | Wang et al. | Sep 2017 | A1 |
20170263248 | Gruber et al. | Sep 2017 | A1 |
20170285915 | Napolitano et al. | Oct 2017 | A1 |
20170316775 | Le et al. | Nov 2017 | A1 |
20170316782 | Haughay et al. | Nov 2017 | A1 |
20170323637 | Naik | Nov 2017 | A1 |
20170345411 | Raitio et al. | Nov 2017 | A1 |
20170346949 | Sanghavi et al. | Nov 2017 | A1 |
20170352346 | Paulik et al. | Dec 2017 | A1 |
20170352350 | Booker et al. | Dec 2017 | A1 |
20170357478 | Piersol et al. | Dec 2017 | A1 |
20170357632 | Pagallo et al. | Dec 2017 | A1 |
20170357633 | Wang et al. | Dec 2017 | A1 |
20170357637 | Nell et al. | Dec 2017 | A1 |
20170357640 | Bellegarda et al. | Dec 2017 | A1 |
20170357716 | Bellegarda et al. | Dec 2017 | A1 |
20170358300 | Laurens et al. | Dec 2017 | A1 |
20170358301 | Raitio et al. | Dec 2017 | A1 |
20170358304 | Castillo et al. | Dec 2017 | A1 |
20170358305 | Kudurshian et al. | Dec 2017 | A1 |
20180007538 | Naik et al. | Jan 2018 | A1 |
20180012596 | Piernot et al. | Jan 2018 | A1 |
20180033431 | Newendorp et al. | Feb 2018 | A1 |
20180090143 | Saddler et al. | Mar 2018 | A1 |
20180108346 | Paulik et al. | Apr 2018 | A1 |
20200302356 | Gruber et al. | Sep 2020 | A1 |
Number | Date | Country |
---|---|---|
2694314 | Aug 2010 | CA |
2792412 | Jul 2011 | CA |
2666438 | Jun 2013 | CA |
101162153 | Apr 2008 | CN |
101174366 | May 2008 | CN |
101179754 | May 2008 | CN |
101183525 | May 2008 | CN |
101188644 | May 2008 | CN |
101228503 | Jul 2008 | CN |
101233741 | Jul 2008 | CN |
101246020 | Aug 2008 | CN |
101271689 | Sep 2008 | CN |
101277501 | Oct 2008 | CN |
101281745 | Oct 2008 | CN |
101292282 | Oct 2008 | CN |
101297541 | Oct 2008 | CN |
101325756 | Dec 2008 | CN |
101416471 | Apr 2009 | CN |
101420471 | Apr 2009 | CN |
101427244 | May 2009 | CN |
101448340 | Jun 2009 | CN |
101453498 | Jun 2009 | CN |
101499156 | Aug 2009 | CN |
101500041 | Aug 2009 | CN |
101535983 | Sep 2009 | CN |
101547396 | Sep 2009 | CN |
101557432 | Oct 2009 | CN |
101604521 | Dec 2009 | CN |
101632316 | Jan 2010 | CN |
101636736 | Jan 2010 | CN |
101673544 | Mar 2010 | CN |
101751387 | Jun 2010 | CN |
101847405 | Sep 2010 | CN |
101894547 | Nov 2010 | CN |
101939740 | Jan 2011 | CN |
101951553 | Jan 2011 | CN |
102137193 | Jul 2011 | CN |
102160043 | Aug 2011 | CN |
102246136 | Nov 2011 | CN |
202035047 | Nov 2011 | CN |
102282609 | Dec 2011 | CN |
202092650 | Dec 2011 | CN |
102368256 | Mar 2012 | CN |
102498457 | Jun 2012 | CN |
102629246 | Aug 2012 | CN |
102682771 | Sep 2012 | CN |
102685295 | Sep 2012 | CN |
102693725 | Sep 2012 | CN |
102792320 | Nov 2012 | CN |
102870065 | Jan 2013 | CN |
102917004 | Feb 2013 | CN |
103035240 | Apr 2013 | CN |
103038728 | Apr 2013 | CN |
104284257 | Jan 2015 | CN |
104423625 | Mar 2015 | CN |
104516522 | Apr 2015 | CN |
102008024258 | Nov 2009 | DE |
1892700 | Feb 2008 | EP |
1912205 | Apr 2008 | EP |
1939860 | Jul 2008 | EP |
1944997 | Jul 2008 | EP |
651543 | Sep 2008 | EP |
1909263 | Jan 2009 | EP |
1335620 | Mar 2009 | EP |
2069895 | Jun 2009 | EP |
2081185 | Jul 2009 | EP |
2094032 | Aug 2009 | EP |
2096840 | Sep 2009 | EP |
2107553 | Oct 2009 | EP |
2109295 | Oct 2009 | EP |
1720375 | Jul 2010 | EP |
2205010 | Jul 2010 | EP |
2309491 | Apr 2011 | EP |
2329348 | Jun 2011 | EP |
2400373 | Dec 2011 | EP |
2431842 | Mar 2012 | EP |
2551784 | Jan 2013 | EP |
2555536 | Feb 2013 | EP |
2575128 | Apr 2013 | EP |
2733598 | May 2014 | EP |
2801890 | Nov 2014 | EP |
2801972 | Nov 2014 | EP |
2930715 | Oct 2015 | EP |
2938022 | Oct 2015 | EP |
2940556 | Nov 2015 | EP |
2911201 | Jul 2008 | FR |
2445436 | Jul 2008 | GB |
2445667 | Jul 2008 | GB |
10-69578 | Mar 1998 | JP |
2001-101100 | Apr 2001 | JP |
2002-41624 | Feb 2002 | JP |
2002-41731 | Feb 2002 | JP |
2002091892 | Mar 2002 | JP |
2002-281562 | Sep 2002 | JP |
2004-171216 | Jun 2004 | JP |
2005-80094 | Mar 2005 | JP |
2005-223782 | Aug 2005 | JP |
2006-107438 | Apr 2006 | JP |
2006-107445 | Apr 2006 | JP |
2006-155368 | Jun 2006 | JP |
2006-166118 | Jun 2006 | JP |
2006-309457 | Nov 2006 | JP |
2007-193794 | Aug 2007 | JP |
2008-009120 | Jan 2008 | JP |
2008-21002 | Jan 2008 | JP |
2008-26381 | Feb 2008 | JP |
2008-39928 | Feb 2008 | JP |
2008-58813 | Mar 2008 | JP |
2008-064687 | Mar 2008 | JP |
2008-90545 | Apr 2008 | JP |
2008-97003 | Apr 2008 | JP |
2008-134949 | Jun 2008 | JP |
2008-158510 | Jul 2008 | JP |
2008-526101 | Jul 2008 | JP |
2008-185693 | Aug 2008 | JP |
2008-198022 | Aug 2008 | JP |
2008-217468 | Sep 2008 | JP |
2008-228129 | Sep 2008 | JP |
2008-233678 | Oct 2008 | JP |
2008-236448 | Oct 2008 | JP |
2008-252161 | Oct 2008 | JP |
2008-268684 | Nov 2008 | JP |
2008-269480 | Nov 2008 | JP |
2008-271481 | Nov 2008 | JP |
2008-275731 | Nov 2008 | JP |
2008-299221 | Dec 2008 | JP |
2009-2850 | Jan 2009 | JP |
2009-503623 | Jan 2009 | JP |
2009-36999 | Feb 2009 | JP |
2009-505142 | Feb 2009 | JP |
2009-47920 | Mar 2009 | JP |
2009-069062 | Apr 2009 | JP |
2009-98490 | May 2009 | JP |
2009-110300 | May 2009 | JP |
2009-134409 | Jun 2009 | JP |
2009-140444 | Jun 2009 | JP |
2009-186989 | Aug 2009 | JP |
2009-193448 | Aug 2009 | JP |
2009-193457 | Aug 2009 | JP |
2009-193532 | Aug 2009 | JP |
2009-205367 | Sep 2009 | JP |
2009-223840 | Oct 2009 | JP |
2009-294913 | Dec 2009 | JP |
2009-294946 | Dec 2009 | JP |
2010-66519 | Mar 2010 | JP |
2010-78979 | Apr 2010 | JP |
2010-108378 | May 2010 | JP |
2010-518475 | May 2010 | JP |
2010-518526 | May 2010 | JP |
2010-157207 | Jul 2010 | JP |
2010-224236 | Oct 2010 | JP |
4563106 | Oct 2010 | JP |
2010-535377 | Nov 2010 | JP |
2010-287063 | Dec 2010 | JP |
2011-33874 | Feb 2011 | JP |
2011-41026 | Feb 2011 | JP |
2011-45005 | Mar 2011 | JP |
2011-59659 | Mar 2011 | JP |
2011-81541 | Apr 2011 | JP |
2011-525045 | Sep 2011 | JP |
2011-238022 | Nov 2011 | JP |
2011-250027 | Dec 2011 | JP |
2012-014394 | Jan 2012 | JP |
2012-508530 | Apr 2012 | JP |
2012-089020 | May 2012 | JP |
2012-116442 | Jun 2012 | JP |
2012-147063 | Aug 2012 | JP |
2012-518847 | Aug 2012 | JP |
2013-37688 | Feb 2013 | JP |
2013-511214 | Mar 2013 | JP |
2013-65284 | Apr 2013 | JP |
2013-73240 | Apr 2013 | JP |
2013-513315 | Apr 2013 | JP |
2013-080476 | May 2013 | JP |
2013-517566 | May 2013 | JP |
2013-134430 | Jul 2013 | JP |
2013-527947 | Jul 2013 | JP |
2013-528012 | Jul 2013 | JP |
2013-156349 | Aug 2013 | JP |
2013-200423 | Oct 2013 | JP |
2013-205999 | Oct 2013 | JP |
2013-238936 | Nov 2013 | JP |
2014-10688 | Jan 2014 | JP |
2014-026629 | Feb 2014 | JP |
2014-60600 | Apr 2014 | JP |
2014-72586 | Apr 2014 | JP |
2014-077969 | May 2014 | JP |
2014-124332 | Jul 2014 | JP |
2014-145842 | Aug 2014 | JP |
2014-150323 | Aug 2014 | JP |
2014-222514 | Nov 2014 | JP |
2015-501022 | Jan 2015 | JP |
2015-41845 | Mar 2015 | JP |
2015-94848 | May 2015 | JP |
2015-519675 | Jul 2015 | JP |
2015-528140 | Sep 2015 | JP |
2015-528918 | Oct 2015 | JP |
2016-119615 | Jun 2016 | JP |
10-2006-0127647 | Dec 2006 | KR |
10-0801227 | Feb 2008 | KR |
10-0810500 | Mar 2008 | KR |
10-2008-0033070 | Apr 2008 | KR |
10-0819928 | Apr 2008 | KR |
10-2008-0049647 | Jun 2008 | KR |
10-2008-0059332 | Jun 2008 | KR |
10-2008-0109322 | Dec 2008 | KR |
10-2009-0001716 | Jan 2009 | KR |
10-2009-0028464 | Mar 2009 | KR |
10-2009-0030117 | Mar 2009 | KR |
10-2009-0086805 | Aug 2009 | KR |
10-0920267 | Oct 2009 | KR |
10-2009-0122944 | Dec 2009 | KR |
10-2009-0127961 | Dec 2009 | KR |
10-2009-0129192 | Dec 2009 | KR |
10-2010-0015958 | Feb 2010 | KR |
10-2010-0048571 | May 2010 | KR |
10-2010-0053149 | May 2010 | KR |
10-2010-0119519 | Nov 2010 | KR |
10-2011-0043644 | Apr 2011 | KR |
10-1032792 | May 2011 | KR |
10-2011-0068490 | Jun 2011 | KR |
10-2011-0072847 | Jun 2011 | KR |
10-2011-0086492 | Jul 2011 | KR |
10-2011-0100620 | Sep 2011 | KR |
10-2011-0113414 | Oct 2011 | KR |
10-2011-0115134 | Oct 2011 | KR |
10-2012-0020164 | Mar 2012 | KR |
10-2012-0031722 | Apr 2012 | KR |
10-1178310 | Aug 2012 | KR |
10-2012-0120316 | Nov 2012 | KR |
10-2012-0137435 | Dec 2012 | KR |
10-2012-0137440 | Dec 2012 | KR |
10-2012-0138826 | Dec 2012 | KR |
10-2012-0139827 | Dec 2012 | KR |
10-1193668 | Dec 2012 | KR |
10-2013-0035983 | Apr 2013 | KR |
10-1334342 | Nov 2013 | KR |
10-2013-0131252 | Dec 2013 | KR |
10-2013-0133629 | Dec 2013 | KR |
10-2014-0147557 | Dec 2014 | KR |
10-2015-0043512 | Apr 2015 | KR |
10-2016-0010523 | Jan 2016 | KR |
2349970 | Mar 2009 | RU |
2353068 | Apr 2009 | RU |
2364917 | Aug 2009 | RU |
200801988 | Jan 2008 | TW |
I301373 | Sep 2008 | TW |
M348993 | Jan 2009 | TW |
200943903 | Oct 2009 | TW |
201018258 | May 2010 | TW |
201027515 | Jul 2010 | TW |
201028996 | Aug 2010 | TW |
201110108 | Mar 2011 | TW |
2011-42823 | Dec 2011 | TW |
201227715 | Jul 2012 | TW |
201245989 | Nov 2012 | TW |
201312548 | Mar 2013 | TW |
2007036762 | Apr 2007 | WO |
2008030970 | Mar 2008 | WO |
2008071231 | Jun 2008 | WO |
2008085742 | Jul 2008 | WO |
2008098900 | Aug 2008 | WO |
2008109835 | Aug 2008 | WO |
2008120036 | Oct 2008 | WO |
2008130095 | Oct 2008 | WO |
2008140236 | Nov 2008 | WO |
2008142472 | Nov 2008 | WO |
2008153639 | Dec 2008 | WO |
2009009240 | Jan 2009 | WO |
2009016631 | Feb 2009 | WO |
2009017280 | Feb 2009 | WO |
2009075912 | Jun 2009 | WO |
2009104126 | Aug 2009 | WO |
2009156438 | Dec 2009 | WO |
2009156978 | Dec 2009 | WO |
2010054373 | May 2010 | WO |
2010075623 | Jul 2010 | WO |
2010100937 | Sep 2010 | WO |
2010141802 | Dec 2010 | WO |
2011057346 | May 2011 | WO |
2011060106 | May 2011 | WO |
2011088053 | Jul 2011 | WO |
2011093025 | Aug 2011 | WO |
2011116309 | Sep 2011 | WO |
2011133543 | Oct 2011 | WO |
2011150730 | Dec 2011 | WO |
2011163350 | Dec 2011 | WO |
2011088053 | Jan 2012 | WO |
2012019637 | Feb 2012 | WO |
2012129231 | Sep 2012 | WO |
2012135157 | Oct 2012 | WO |
2012154317 | Nov 2012 | WO |
2012155079 | Nov 2012 | WO |
2012167168 | Dec 2012 | WO |
2013009578 | Jan 2013 | WO |
2013022135 | Feb 2013 | WO |
2013022223 | Feb 2013 | WO |
2013048880 | Apr 2013 | WO |
2013049358 | Apr 2013 | WO |
2013163113 | Oct 2013 | WO |
2013169842 | Nov 2013 | WO |
2013173504 | Nov 2013 | WO |
2013173511 | Nov 2013 | WO |
2013184953 | Dec 2013 | WO |
2013184990 | Dec 2013 | WO |
2014003138 | Jan 2014 | WO |
2014022148 | Feb 2014 | WO |
2014028797 | Feb 2014 | WO |
2014031505 | Feb 2014 | WO |
2014066352 | May 2014 | WO |
2014078965 | May 2014 | WO |
2014096506 | Jun 2014 | WO |
2014124332 | Aug 2014 | WO |
2014137074 | Sep 2014 | WO |
2014138604 | Sep 2014 | WO |
2014143959 | Sep 2014 | WO |
2014144579 | Sep 2014 | WO |
2014159581 | Oct 2014 | WO |
2014197336 | Dec 2014 | WO |
2014200728 | Dec 2014 | WO |
2014204659 | Dec 2014 | WO |
2015018440 | Feb 2015 | WO |
2015030796 | Mar 2015 | WO |
2015041892 | Mar 2015 | WO |
2015084659 | Jun 2015 | WO |
2015094169 | Jun 2015 | WO |
2015094369 | Jun 2015 | WO |
2015099939 | Jul 2015 | WO |
2015116151 | Aug 2015 | WO |
2015151133 | Oct 2015 | WO |
2015157013 | Oct 2015 | WO |
2015183401 | Dec 2015 | WO |
2015200207 | Dec 2015 | WO |
2016028946 | Feb 2016 | WO |
2016033257 | Mar 2016 | WO |
2016057268 | Apr 2016 | WO |
2016075081 | May 2016 | WO |
2016144982 | Sep 2016 | WO |
2016209444 | Dec 2016 | WO |
2017044260 | Mar 2017 | WO |
2017044629 | Mar 2017 | WO |
2017053311 | Mar 2017 | WO |
Entry |
---|
Oral Hearing Minutes received for U.S. Appl. No. 13/251,104, dated Feb. 28, 2019, 18 pages. |
Decision on Appeal received for U.S. Appl. No. 13/251,104, dated Mar. 29, 2019, 39 pages. |
Extended European Search Report received for European Patent Application No. 18213462.7, dated Mar. 21, 2019, 12 pages. |
Office Action received for Australian Patent Application No. 2018204265, dated Mar. 20, 2019, 3 pages. |
Notice of Allowance received for U.S. Appl. No. 13/894,354, dated Aug. 30, 2018, 13 pages. |
Preliminary Opinion received for European Patent Application No. 12727027.0, dated Aug. 17, 2018, 22 pages. |
Notice of Allowance received for Korean Patent Application No. 10-2019-7031839, dated Jan. 27, 2020, 6 pages (3 pages of English Translation and 3 pages of Official Copy). |
Notice of Allowance received for Korean Patent Application No. 10-2018-7031467, dated Jul. 30, 2019, 8 pages (5 pages of English Translation and 3 pages of Official Copy). |
Decision on Appeal received for U.S. Appl. No. 13/251,118, dated Jun. 3, 2019, 20 pages. |
Office Action received for Japanese Patent Application No. 2017-068594, dated May 10, 2019, 8 pages (4 pages of English Translation and 4 pages of Official Copy). |
Board Decision received for Chinese Patent Application No. 201280027176.5, dated Nov. 12, 2018, 22 pages (2 pages of English Translation and 20 pages of Official Copy). |
Office Action received for Korean Patent Application No. 10-2018-7031467, dated Jan. 22, 2019, 6 pages (2 pages of English Translation and 4 pages of Official Copy). |
Decision to Refuse received for European Patent Application No. 13726938.7, dated Jul. 11, 2019, 19 pages. |
Result of Consultation received for European Patent Application No. 13726938.7, dated Jun. 18, 2019, 5 pages. |
Office Action received for European Patent Application No. 18213462.7, dated Jan. 22, 2020, 20 pages. |
Office Action received for Japanese Patent Application No. 2017-068594, dated Jan. 7, 2020, 12 pages (6 pages of English translation and 6 pages of Official copy). |
Applicant Initiated Interview Summary received for U.S. Appl. No. 15/193,971, dated Nov. 4, 2019, 3 pages. |
Decision to Refuse received for European Patent Application No. 12727027.0, dated Oct. 18, 2018, 29 pages. |
Minutes of Meeting received for European Patent Application No. 12727027.0, dated Oct. 18, 2018, 3 pages. |
Notice of Allowance received for U.S. Appl. No. 13/251,088, dated Oct. 30, 2018, 7 pages. |
Extended European Search Report received for European Patent Application No. 12727027.0, dated Sep. 26, 2014, 7 pages. |
Extended European Search Report received for European Patent Application No. 13726938.7, dated Dec. 14, 2015, 8 pages. |
International Preliminary Report on Patentability received for PCT Patent Application No. PCT/US2012/040571, dated Dec. 19, 2013, 10 pages. |
International Preliminary Report on Patentability received for PCT Patent Application No. PCT/US2013/041233, dated Nov. 27, 2014, 9 pages. |
International Search Report and Written Opinion received for PCT Patent Application No. PCT/US2012/040571, dated Nov. 16, 2012, 14 pages. |
International Search Report and Written Opinion received for PCT Patent Application No. PCT/US2013/041233, dated Nov. 22, 2013, 11 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/193,971, dated Jan. 10, 2019, 22 pages. |
Notice of Allowance received for U.S. Appl. No. 13/894,354, dated Jan. 8, 2019, 13 pages. |
Summons to Attend Oral Proceedings received for European Patent Application No. 13726938.7, mailed on Jan. 15, 2019, 8 pages. |
Decision on Appeal received for U.S. Appl. No. 13/251,127, dated Aug. 30, 2019, 28 pages. |
Non-Final Office Action received for U.S. Appl. No. 15/193,971, dated Sep. 19, 2019, 16 pages. |
Advisory action received for U.S. Appl. No. 13/251,104, dated Nov. 2, 2015, 3 pages. |
Advisory Action received for U.S. Appl. No. 13/894,354, dated Mar. 24, 2016, 4 pages. |
“Alexa, Turn Up the Heat!”, Smartthings Samsung [online], Available online at https://web.archive.org/web/20160329142041/https://blog.smartthings.com/news/smartthingsupdates/alexa-turn-up-the-heat/, Mar. 3, 2016, 3 pages. |
“Ask Alexa—Things That Are Smart Wiki”, Available online at <URL:http://thingsthataresmart.wiki/index.php?title=Ask_Alexa&oldid=4283>, [retrieved from internet on Aug. 2, 2017], Jun. 8, 2016, pp. 1-31. |
Board Opinion received for Chinese Patent Application No. 201280027176.5, dated Feb. 12, 2018, 10 pages (1 page of English Translation and 9 pages of Official Copy). |
Board Opinion received for Chinese Patent Application No. 201280027176.5, dated Jun. 14, 2018, 12 pages (2 pages of English Translation and 10 pages of Official Copy). |
“DIRECTV™ Voice”, Now Part of the DIRECTTV Mobile App for Phones, Sep. 18, 2013, 5 pages. |
Final Office Action received for U.S. Appl. No. 13/251,088, dated Aug. 2, 2013, 16 pages. |
Final Office Action received for U.S. Appl. No. 13/251,088, dated May 20, 2015, 26 pages. |
Final Office Action received for U.S. Appl. No. 13/251,104, dated Aug. 14, 2013, 39 pages. |
Final Office Action received for U.S. Appl. No. 13/251,104, dated Jul. 10, 2015, 36 pages. |
Final Office Action received for U.S. Appl. No. 13/251,127, dated Mar. 25, 2013, 46 pages. |
Final Office Action received for U.S. Appl. No. 13/251,127, dated May 5, 2016, 60 pages. |
Final Office Action received for U.S. Appl. No. 13/894,354, dated Oct. 7, 2015, 25 pages. |
“Headset Button Controller v7.3 APK Full APP Download for Android”, Blackberry, iPhone, Jan. 27, 2014, 11 pages. |
“Hear voice from Google Translate”, Available on URL:https://www.youtube.com/watch?v=18AvMhFqD28, Jan. 28, 2011, 1 page. |
“Interactive Voice”, available at <http://www.helloivee.com/company/>, retrieved on Feb. 10, 2014, 2 pages. |
“Meet Ivee, Your Wi-Fi Voice Activated Assistant”, available at <http://www.helloivee.com/>, retrieved on Feb. 10, 2014, 8 pages. |
“Mel Scale”, Wikipedia the Free Encyclopedia, Last modified on Oct. 13, 2009 and retrieved on Jul. 28, 2010, available at <http://en.wikipedia.org/wiki/Mel_scale>, 2 pages. |
“Minimum Phase”, Wikipedia the free Encyclopedia, Last modified on Jan. 12, 2010 and retrieved on Jul. 28, 2010, available at <http://en.wikipedia.org/wiki/Minimum_phase>, 8 pages. |
Mobile Speech Solutions, Mobile Accessibility, SVOX AG Product Information Sheet, available at Khttp://www.svox.com/site/bra840604/con782768/mob965831936.aSQ?osLang=1>, Sep. 27, 2012, 1 page. |
Non-Final Office Action received for U.S. Appl. No. 13/251,088, dated Dec. 4, 2014, 19 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,088, dated Jan. 31, 2013, 17 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,088, dated Jun. 19, 2014, 16 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,104, dated Jan. 28, 2014, 48 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,104, dated Nov. 28, 2012, 32 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,104, dated Oct. 9, 2014, 52 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,127, dated Jan. 16, 2014, 48 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,127, dated Jul. 17, 2015, 53 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,127, dated Nov. 5, 2014, 42 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,127, dated Nov. 8, 2012, 18 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/894,354, dated Apr. 9, 2015, 20 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/894,354, dated Jan. 13, 2017, 28 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/894,354, dated Jul. 3, 2017, 32 pages. |
Notice of Acceptance received for Australian Patent application No. 2012261958, dated Apr. 6, 2016, 3 pages. |
Notice of Acceptance received for Australian Patent application No. 2013262803, dated Jul. 8, 2016, 3 pages. |
Notice of Acceptance received for Australian Patent application No. 2016204091, dated Feb. 28, 2018, 3 pages. |
Notice of Allowance received for Chinese Patent Application No. 201380028060.8, dated Jul. 10, 2018, 2 pages (1 page of English Translation and 1 page of Official Copy only). |
Notice of Allowance received for Japanese Patent Application No. 2014-513765, dated May 8, 2017, 3 pages (Official Copy only) (See Communication under 37 CFR § 1.98(a) (3)). |
Notice of Allowance received for Japanese Patent application No. 2015-512807, dated Jul. 28, 2017, 3 pages (Official Copy only) (See Communication under 37 CFR § 1.98(a) (3)). |
Notice of Allowance received for Korean Patent Application No. 10-2013-7034856, dated Mar. 22, 2017, 4 pages (1 page of English Translation and 3 pages of Official Copy). |
Notice of Allowance received for Korean Patent Application No. 10-2014-7033076, dated Oct. 31, 2016, 4 pages (1 page of English Translation and 3 pages of official copy). |
Notice of Allowance received for Korean Patent Application No. 10-2017-7017149, dated Jul. 30, 2018, 4 pages (1 page of English Translation and 3 pages of Official Copy). |
Notice of Allowance received for U.S. Appl. No. 13/251,088, dated Jun. 8, 2018, 8 pages. |
Office Action received for Australian Patent Application No. 2012261958, dated Mar. 10, 2016, 3 pages. |
Office Action received for Australian Patent Application No. 2012261958, dated Mar. 27, 2015, 3 pages. |
Office Action received for Australian Patent Application No. 2013262803, dated Aug. 6, 2015, 3 pages. |
Office Action received for Australian Patent Application No. 2016204091, dated Mar. 3, 2017, 3 pages. |
Office Action received for Chinese Patent Application No. 201280027176.5, dated Aug. 26, 2016, 12 pages (3 pages of English Translation and 9 pages of Official Copy). |
Office Action received for Chinese Patent Application No. 201280027176.5, dated Dec. 29, 2015, 22 pages (12 pages of English Translation and 10 pages of Official copy). |
Office Action received for Chinese Patent Application No. 201280027176.5, dated Feb. 15, 2017, 9 pages (1 page of English Translation and 8 pages of official copy). |
Office Action received for Chinese Patent Application No. 201380028060.8, dated Dec. 14, 2017, 6 pages (3 pages of English Translation and 3 pages of official copy). |
Office Action received for Chinese Patent Application No. 201380028060.8, dated Feb. 8, 2017, 14 pages (3 pages of English Translation and 11 pages of Official Copy). |
Office Action received for European Patent Application No. 12727027.0, dated Mar. 8, 2017, 6 pages. |
Office Action received for European Patent Application No. 13726938.7, dated Jan. 11, 2018, 7 pages. |
Office Action received for Japanese Patent Application No. 2014-513765, dated Dec. 2, 2016, 7 pages (3 pages of English Translation and 4 pages of Official Copy). |
Office Action received for Japanese Patent Application No. 2014-513765, dated Jan. 5, 2016, 8 pages (4 pages of English Translation and 4 pages of official copy). |
Office Action received for Japanese Patent Application No. 2014-513765, dated Jan. 19, 2015, 8 pages (5 pages of English Translation and 3 pages of official copy). |
Office Action received for Japanese Patent Application No. 2015-512807, dated Nov. 2, 2015, 5 pages (Official Copy Only) (See Communication under 37 CFR § 1.98(a) (3)). |
Office Action received for Japanese Patent Application No. 2015-512807, dated Oct. 3, 2016, 9 pages (4 pages of English Translation and 5 pages of official copy). |
Office Action received for Japanese Patent Application No. 2017-068594, dated May 7, 2018, 10 pages (5 pages of English Translation and 5 pages of Official Copy). |
Office Action received for Korean Patent Application No. 10-2013-7034856, dated Apr. 29, 2016, 7 pages (3 pages of English Translation and 4 pages of Official Copy). |
Office Action received for Korean Patent Application No. 10-2013-7034856, dated Jun. 13, 2015, 13 pages (8 pages of English Translation and 5 pages of Official Copy). |
Office Action received for Korean Patent Application No. 10-2014-7033076, dated Dec. 23, 2015, 9 pages (4 pages of English Translation and 5 pages of official copy). |
Office Action received for Korean Patent Application No. 10-2017-7017149, dated Sep. 21, 2017, 10 pages (4 pages of English Translation and 6 pages of Official Copy). |
“SmartThings +Amazon Echo”, Smartthings Samsung [online], Available online at <https://web.archive.org/web/20160509231428/https://blog.smartthings.com/featured/alexa-turn-on-my-smartthings/>, Aug. 21, 2015, 3 pages. |
“Speaker Recognition”, Wikipedia, The Free Enclyclopedia, Nov. 2, 2010, 4 pages. |
Summons to Attend Oral Proceeding received for European Patent Application No. 12727027.0, dated Apr. 3, 2018, 15 pages. |
Supplementary European Search Report received for European Patent Application No. 12727027.0, dated Oct. 14, 2014, 1 page. |
“The world of Virtual Assistants—more SemTech . . . ”, End of Business as Usual—Glenn's External blog, Online Available at <https://web.archive.org/web/20091101840940/http://glennas.wordpress.com/2009/10/17/the-world-of-virtual-assistants-more-semtech/>, Oct. 17, 2009, 5 pages. |
Adium, “AboutAdium—Adium X—Trac”, available at Khttp://web.archive.org/web/20070819113247/http://trac.adiumx.com/wiki/AboutAdium>, retrieved on Nov. 25, 2011, 2 pages. |
Alfred App, “Alfred”, available at <http://www.alfredapp.com/>, retrieved on Feb. 8, 2012, 5 pages. |
Anania, Peter, “Amazon Echo with Home Automation (Smartthings)”, Available online at https://www.youtube.com/watch?v=LMW6aXmsWNE, Dec. 20, 2015, 1 page. |
api.ai, “Android App Review—Speaktoit Assistant”, Available at <https://www.youtube.com/watch?v=myE498nyfGw>, Mar. 30, 2011, 3 pages. |
Apple, “VoiceOver”, available at <http://www.apple.com/accessibility/voiceover/>, May 19, 2014, 3 pages. |
Apple Computer, “Knowledge Navigator”, published by Apple Computer no later than 2008, as depicted in Exemplary Screenshots from video entitled ‘Knowledge Navigator’, 2008, 7 pages. |
Asakura et al., “What LG thinks; How the TV should be in the Living Room”, HiVi, vol. 31, No. 7 (Jul. 2013), Stereo Sound Publishing, Inc., Jun. 17, 2013, pp. 68-71 (Official Copy Only). (See Communication under 37 CFR § 1.98(a) (3)). |
Berry et al., “PTIME: Personalized Assistance for Calendaring”, ACM Transactions on Intelligent Systems and Technology, vol. 2, No. 4, Article 40, Jul. 2011, pp. 1-22. |
Bertulucci, Jeff, “Google Adds Voice Search to Chrome Browser”, PC World, Jun. 14, 2011, 5 pages. |
Bocchieri et al., “Use of Geographical Meta-Data in ASR Language and Acoustic Models”, IEEE International Conference on Acoustics Speech and Signal Processing, 2010, pp. 5118-5121. |
Butcher, Mike, “EVI Arrives in Town to go Toe-to-Toe with Siri”, TechCrunch, Jan. 23, 2012, 2 pages. |
Cambria et al., “Jumping NLP Curves: A Review of Natural Language Processing Research”, IEEE Computational Intelligence Magazine, 2014, vol. 9, May 2014, pp. 48-57. |
Caraballo et al., “Language Identification Based on a Discriminative Text Categorization Technique”, Iberspeech 2012—Vii Jornadas en Tecnologia del Habla and Iii Iberiansl Tech Workshop, Nov. 21, 2012, pp. 1-10. |
Castleos, “Whole House Voice Control Demonstration”, available online at : https://www.youtube.com/watch?v=9SRCoxrZ_W4, Jun. 2, 2012, 26 pages. |
Chamberlain, Kim, “Quick Start Guide Natural Reader”, available online at <http://atrc.colostate.edu/files/quickstarts/Natural_Reader_Quick_Start_Guide.>, Apr. 2008, 5 pages. |
Chen, Yi, “Multimedia Siri Finds and Plays Whatever You Ask For”, PSFK Report, Feb. 9, 2012, 9 pages. |
Cheyer, Adam, “About Adam Cheyer”, available at <http://www.adam.cheyer.com/about.html>, retrieved on Sep. 17, 2012, 2 pages. |
Choi et al., “Acoustic and Visual Signal based Context Awareness System for Mobile Application”, IEEE Transactions on Consumer Electronics, vol. 57, No. 2, May 2011, pp. 738-746. |
Colt, Sam, “Here's One Way Apple's Smartwatch Could Be Better Than Anything Else”, Business Insider, Aug. 21, 2014, pp. 1-4. |
Deedeevuu, “Amazon Echo Alarm Feature”, Available online at https://www.youtube.com/watch?v=fdjU8eRLk7c, Feb. 16, 2015, 1 page. |
Elliott et al., “Annotation Suggestion and Search for Personal Multimedia Objects on the Web”, CIVR, Jul. 7-9, 2008, pp. 75-84. |
Erol et al., “Multimedia Clip Generation From Documents for Browsing on Mobile Devices”, IEEE Transactions on Multimedia, vol. 10, No. 5, Aug. 2008, 13 pages. |
Evi, “Meet Evi: The One Mobile Application that Provides Solutions for your Everyday Problems”, Feb. 2012, 3 pages. |
Exhibit 1, “Natural Language Interface Using Constrained Intermediate Dictionary of Results”, List of Publications Manually Reviewed for the Search of U.S. Pat. No. 7,177,798, Mar. 22, 2013, 1 page. |
Filipowicz, Luke, “How to use the Quick Type Keyboard in iOS 8”, available online at <https://www.imore.com/comment/568232>, Oct. 11, 2014, pp. 1-17. |
Findlater et al., “Beyond QWERTY: Augmenting Touch-Screen Keyboards with Multi-Touch Gestures for Non-Alphanumeric Input”, CHI '12, Austin, Texas, USA, May 5-10, 2012, 4 pages. |
Finkel et al., “Joint Parsing and Named Entity Recognition”, Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the ACL, Jun. 2009, pp. 326-334. |
Gannes, Liz, “Alfred App Gives Personalized Restaurant Recommendations”, AllThingsD, Jul. 18, 2011, pp. 1-3. |
Gomez et al., “Mouth Gesture and Voice Command Based Robot Command Interface”, IEEE International Conference on Robotics and Automation, May 12-17, 2009, pp. 333-338. |
Gruber, Tom, “Big Think Small Screen: How Semantic Computing in the Cloud will Revolutionize the Consumer Experience on the Phone”, Keynote Presentation at Web 3.0 Conference, Jan. 2010, 41 pages. |
Gruber, Tom, “Despite Our Best Efforts, Ontologies are not the Problem”, AAAI Spring Symposium, Available online at <http://tomgruber.org/writing/aaai-ss08.htm>, Mar. 2008, pp. 1-40. |
Gruber, Tom, “Intelligence at the Interface: Semantic Technology and the Consumer Internet Experience”, Presentation at Semantic Technologies Conference, Available online at <http://tomgruber.org/writing/semtech08.htm>, May 20, 2008, pp. 1-40. |
Gruber, Tom, “Siri, A Virtual Personal Assistant-Bringing Intelligence to the Interface”, Semantic Technologies Conference, Jun. 16, 2009, 21 pages. |
Gruber, Thomas R., et al., U.S. Appl. No. 61/186,414, filed Jun. 12, 2009 titled “System and Method for Semantic Auto-Completion” 13 pages (Copy Not Attached). |
Guay, Matthew, “Location-Driven Productivity with Task Ave”, available at <http://iphone.appstorm.net/reviews/productivity/location-driven-productivity-with-task-ave/>, Feb. 19, 2011, 7 pages. |
Guim, Mark, “Howto Set a Person-Based Reminder with Cortana”, available at <http://www.wpcentral.com/how-to-person-based-reminder-cortana>, Apr. 26, 2014, 15 pages. |
Hardwar, Devindra, “Driving App Waze Builds its own Siri for Hands-Free Voice Control”, Available online at <http://venturebeat.com/2012/02/09/driving-app-waze-builds-its-own-siri-for-hands-free-voice-control/>, retrieved on Feb. 9, 2012, 4 pages. |
Hashimoto, Yoshiyuki, “Simple Guide for iPhone Siri, Which Can Be Operated with Your Voice”, Shuwa System Co., Ltd., vol. 1, Jul. 5, 2012, pp. 8, 130, 131. |
id3.org, “id3v2.4.0—Frames”, available at <http://id3.org/id3v2.4.0-frames?action=print>, retrieved on Jan. 22, 2015, 41 pages. |
Iowegian International, “FIR Filter Properties, DSPGuru, Digital Signal Processing Central”, available at <http://www.dspguru.com/dsp/faq/fir/properties> retrieved on Jul. 28, 2010, 6 pages. |
Jawaid et al., “Machine Translation with Significant Word Reordering and Rich Target-Side Morphology”, WDS'11 Proceedings of Contributed Papers, Part I, 2011, pp. 161-166. |
Jiang et al., “A Syllable-based Name Transliteration System”, Proc. of the 2009 Named Entities Workshop, Aug. 7, 2009, pp. 96-99. |
Jonsson et al., “Proximity-based Reminders Using Bluetooth”, 2014 IEEE International Conference on Pervasive Computing and Communications Demonstrations, 2014, pp. 151-153. |
Jouvet et al., “Evaluating Grapheme-to-phoneme Converters in Automatic Speech Recognition Context”, IEEE, 2012, pp. 4821-4824. |
Kane et al., “Slide Rule: Making Mobile Touch Screens Accessible to Blind People Using Multi-Touch Interaction Techniques”, ASSETS, Oct. 13-15, 2008, pp. 73-80. |
Karn, Ujjwal, “An Intuitive Explanation of Convolutional Neural Networks”, The Data Science Blog, Aug. 11, 2016, 23 pages. |
Kazmucha, Allyson, “How to Send Map Locations Using iMessage”, iMore.com, Available at <http://www.imore.com/how-use-imessage-share-your-location-your-iphone>, Aug. 2, 2012, 6 pages. |
Kickstarter, “Ivee Sleek: Wi-Fi Voice-Activated Assistant”, available at <https://www.kickstarter.com/projects/ivee/ivee-sleek-wi-fi-voice-activated-assistant>, retrieved on Feb. 10, 2014, 13 pages. |
Knownav, “Knowledge Navigator”, YouTube Video available at <http://www.youtube.com/watch?v=QRH8eimU_20>, Apr. 29, 2008, 1 page. |
Lewis, Cameron, “Task Ave for iPhone Review”, Mac Life, Available at <http://www.maclife.com/article/reviews/task_ave_iphone_review>, Mar. 3, 2011, 5 pages. |
Mactech, “Keystrokes 3.5 for Mac OS X Boosts Word Prediction”, available at <http://www.mactech.com/news/?p=1007129>, retrieved on Jan. 7, 2008, 3 pages. |
Majerus, Wesley, “Cell Phone Accessibility for your Blind Child”, Retrieved from the Internet <URL:https://web.archive.org/web/20100210001100/https://nfb.org/images/nfb/publications/fr/fr28/3/fr280314.htm>, 2010, pp. 1-5. |
Martins et al., “Extracting and Exploring the Geo-Temporal Semantics of Textual Resources”, Semantic Computing, IEEE International Conference, 2008, pp. 1-9. |
Mhatre et al., “Donna Interactive Chat-bot acting as a Personal Assistant”, International Journal of Computer Applications (0975-8887), vol. 140, No. 10, Apr. 2016, 6 pages. |
Microsoft, “Turn on and Use Magnifier”, available at <http://www.microsoft.com/windowsxp/using/accessibility/magnifierturnon.mspx>, retrieved on Jun. 6, 2009. |
Miller, Chance, “Google Keyboard Updated with New Personalized Suggestions Feature”, available at <http://9to5google.com/2014/03/19/google-keyboard-updated-with-new-personalized-suggestions-feature/>, Mar. 19, 2014, 4 pages. |
Morrison, Jonathan, “iPhone 5 Siri Demo”, Online Available at Khttps://www.youtube.com/watch?v=_wHWwG5lhWc>, Sep. 21, 2012, 3 pages. |
Morton, Philip, “Checking if an Element is Hidden”, StackOverflow, Available at Khttp://stackoverflow.com/questions/178325/checking-if-an-element-is-hidden>, Oct. 7, 2008, 12 pages. |
My Cool Aids, “What's New”, available at <http://www.mycoolaids.com/>, 2012, 1 page. |
Myers, Brad A., “Shortcutter for Palm”, available at <http://www.cs.cmu.edu/˜pebbles/v5/shortcutter/palm/index.html>, retrieved on Jun. 18, 2014, 10 pages. |
Nakazawa et al., “Detection and Labeling of Significant Scenes from TV program based on Twitter Analysis”, Proceedings of the 3rd Forum on Data Engineering and Information Management (DEIM 2011 proceedings), IEICE Data Engineering Technical Group. Available online at: http://db-event.jpn.org/deim2011/proceedings/pdf/f5-6.pdf, Feb. 28, 2011, 10 pages (Official Copy Only). (See Communication under 37 CFR § 1.98(a) (3)). |
Naone, Erica, “TR10: Intelligent Software Assistant”, Technology Review, Mar.-Apr. 2009, 2 pages. |
Navigli, Roberto, “Word Sense Disambiguation: A Survey”, ACM Computing Surveys, vol. 41, No. 2, Feb. 2009, 69 pages. |
NDTV, “Sony Smartwatch 2 Launched in India for Rs. 14,990”, available at <http://gadgets.ndtv.com/others/news/sony-smartwatch-2-launched-in-india-for-rs-14990-420319>, Sep. 18, 2013, 4 pages. |
Ng, Simon, “Google's Task List Now Comes to Iphone”, SimonBlog, Available at <http://www.simonblog.com/2009/02/04/googles-task-list-now-comes-to-iphone/>, Feb. 4, 2009, 3 pages. |
Nozawa, Naoki et al., “iPhone 4S Perfect Manual”, vol. 1, First Edition, Nov. 11, 2011, 5 pages. |
Osxdaily, “Get a List of Siri Commands Directly from Siri”, Available at Khttp://osxdaily.com/2013/02/05/list-siri-commands/>, Feb. 5, 2013, 15 pages. |
Pan et al., “Natural Language Aided Visual Query Building for Complex Data Access”, In proceeding of: Proceedings of the Twenty-Second Conference on Innovative Applications of Artificial Intelligence, XP055114607, Jul. 11, 2010, pp. 1821-1826. |
Pathak et al., “Privacy-preserving Speech Processing: Cryptographic and Stringmatching Frameworks Show Promise”, In: IEEE signal processing magazine, retrieved from <http://www.merl.com/publications/docs/TR2013-063.pdf>, Feb. 13, 2013, 16 pages. |
Patra et al., “A Kernel-Based Approach for Biomedical Named Entity Recognition”, Scientific World Journal, vol. 2013, 2013, pp. 1-7. |
Phoenix Solutions, Inc., “Declaration of Christopher Schmandt Regarding the MIT Galaxy System”, West Interactive Corp., A Delaware Corporation, Document 40, Jul. 2, 2010, 162 pages. |
Powell, Josh, “Now You See Me . . . Show/Hide Performance”, available at http://www.learningjquery.com/2010/05/now-you-see-me-showhide-performance, May 4, 2010, 3 pages. |
Rios, Mafe, “New bar search for Facebook”, Youtube, available at “https://www.youtube.com/watch?v=vwgN1WbvCas”, 1 page. |
Routines, “SmartThings Support”, Available online at <https://web.archive.Org/web/20151207165701/https://support.smartthings.com/hc/en-us/articles/205380034-Routines>, 2015, 2 pages. |
Sarawagi, Sunita, “CRF Package Page”, available at <http://crf.sourceforge.net/>, retrieved on Apr. 6, 2011, 2 pages. |
Selfrifge et al., “Interact: Tightly-coupling Multimodal Dialog with an Interactive Virtual Assistant”, International Conference on Multimodal Interaction, ACM, Nov. 9, 2015, pp. 381-382. |
Simonite, Tom, “One Easy Way to Make Siri Smarter”, Technology Review, Oct. 18, 2011, 2 pages. |
Spivack, Nova, “Sneak Preview of Siri—Part Two—Technical Foundations—Interview with Tom Gruber, CTO of Siri”, Online Available at <https://web.archive.org/web/20100114234454/http://www.twine.com/item/12vhy39k4-22m/interview-with-tom-gruber-of-siri>, Jan. 14, 2010, 5 pages. |
SRI, “SRI Speech: Products: Software Development Kits: EduSpeak”, available at Khttp://web.archive.org/web/20090828084033/http://www.speechatsri.com/products/eduspeak>shtml, retrieved on Jun. 20, 2013, 2 pages. |
Stent et al., “Geo-Centric Language Models for Local Business Voice Search”, AT&T Labs—Research, 2009, pp. 389-396. |
Sullivan, Danny, “How Google Instant's Autocomplete Suggestions Work”, available at <http://searchengineland.com/how-google-instant-autocomplete-suggestions-work-62592>, Apr. 6, 2011, 12 pages. |
Sundaram et al., “Latent Perceptual Mapping with Data-Driven Variable-Length Acoustic Units for Template-Based Speech Recognition”, ICASSP 2012, Mar. 2012, pp. 4125-4128. |
Sundermeyer et al., “From Feedforward to Recurrent LSTM Neural Networks for Language Modeling”, IEEE Transactions to Audio, Speech, and Language Processing, 2015, vol. 23, Mar. 2015, pp. 517-529. |
Tanaka, Tatsuo, “Next Generation IT Channel Strategy Through “Experience Technology””, Intellectual Resource Creation, Japan, Nomura Research Institute Ltd. vol. 19, No. 1, Dec. 20, 2010, 17 pages. (Official Copy only) (See Communication under 37 CFR § 1.98(a) (3)). |
TextnDrive, “Text'nDrive App Demo-Listen and Reply to your Messages by Voice while Driving!”, YouTube Video available at <http://www.youtube.com/watch?v=WaGfzoHsAMw>, Apr. 27, 2010, 1 page. |
Tofel, Kevin C., “SpeakTolt: A Personal Assistant for Older iPhones, iPads”, Apple News, Tips and Reviews, Feb. 9, 2012, 7 pages. |
Tucker, Joshua, “Too Lazy to Grab Your TV Remote? Use Siri Instead”, Engadget, Nov. 30, 2011, 8 pages. |
Tur et al., “The CALO Meeting Assistant System”, IEEE Transactions on Audio, Speech and Language Processing, vol. 18, No. 6, Aug. 2010, pp. 1601-1611. |
Tur et al., “The CALO Meeting Speech Recognition and Understanding System”, Proc. IEEE Spoken Language Technology Workshop, 2008, 4 pages. |
Vlingo InCar, “Distracted Driving Solution with Vlingo InCar”, YouTube Video, Available online at <http://www.youtube.com/watch?v=Vqs8XfXxgz4>, Oct. 2010, 2 pages. |
Vlingo, “Vlingo Launches Voice Enablement Application on Apple App Store”, Press Release, Dec. 3, 2008, 2 pages. |
Vodafone Deutschland, “Samsung Galaxy S3 Tastatur Spracheingabe”, Available online at—“https://www.youtube.com/watch?v=6kOd6Gr8uFE”, Aug. 22, 2012, 1 page. |
Voiceassist, “Send Text, Listen to and Send E-Mail by Voice”, YouTube Video, Available online at <http://www.youtube.com/watch?v=0tEU61nHHA4>, Jul. 30, 2009, 1 page. |
VoiceontheGo, “Voice on the Go (BlackBerry)”, YouTube Video, available online at <http://www.youtube.com/watch?v=pJqpWgQS98w>, Jul. 27, 2009, 1 page. |
Wikipedia, “Acoustic Model”, available at <http://en.wikipedia.org/wiki/AcousticModel>, retrieved on Sep. 14, 2011, 2 pages. |
Wikipedia, “Language Model”, available at <http://en.wikipedia.org/wiki/Language_model>, retrieved on Sep. 14, 2011, 4 pages. |
Wikipedia, “Speech Recognition”, available at <http://en.wikipedia.org/wiki/Speech_recognition>, retrieved on Sep. 14, 2011, 12 pages. |
Wilson, Mark, “New iPod Shuffle Moves Buttons to Headphones, Adds Text to Speech”, available at <http://gizmodo.com/5167946/new-ipod-shuffle-moves-buttons-to-headphones-adds-text-to-speech>, Mar. 11, 2009, 12 pages. |
X.Ai, “How it Works”, May 2016, 6 pages. |
Xiang et al., “Correcting Phoneme Recognition Errors in Learning Word Pronunciation through Speech Interaction”, Speech Communication, vol. 55, No. 1, Jan. 1, 2013, pp. 190-203. |
Xu, Yuhong, “Policy optimization of dialogue management in spoken dialogue system for out-of-domain utterances”, 2016 International Conference on Asian Language Processing (IALP), IEEE, Nov. 21, 2016, pp. 10-13. |
Xu et al., “Speech-Based Interactive Games for Language Learning: Reading, Translation, and Question-Answering”, Computational Linguistics and Chinese Language Processing, vol. 14, No. 2, Jun. 2009, pp. 133-160. |
Yan et al., “A Scalable Approach to Using DNN-Derived Features in GMM-HMM Based Acoustic Modeling for LVCSR”, InInterspeech, 2013, pp. 104-108. |
Young et al., “The Hidden Information State model: A practical framework for POMDP-based spoken dialogue management”, Computer Speech & Language, vol. 24, Issue 2, 2010, pp. 150-174. |
Zainab, “Google Input Tools Shows Onscreen Keyboard in Multiple Languages [Chrome]”, available at <http://www.addictivetips.com/internet-tips/google-input-tools-shows-multiple-language-onscreen-keyboards-chrome/>, Jan. 3, 2012, 3 pages. |
Zangerle et al., “Recommending #-Tag in Twitter”, Proceedings of the Workshop on Semantic Adaptive Socail Web, 2011, pp. 1-12. |
Zhang et al., “Research of Text Classification Model Based on Latent Semantic Analysis and Improved HS-SVM”, Intelligent Systems and Applications (ISA), 2010 2nd International Workshop, May 22-23, 2010, 5 pages. |
Zhong et al., “JustSpeak: Enabling Universal Voice Control on Android”, W4A'14, Proceedings of the 11th Web for All Conference, No. 36, Apr. 7-9, 2014, 8 pages. |
Advisory Action received for U.S. Appl. No. 13/729,597, dated Oct. 16, 2015, 5 pages. |
Final Office Action received for U.S. Appl. No. 13/251,118, dated un. 13, 2013, 35 pages. |
Final Office Action received for U.S. Appl. No. 13/251,118, dated Oct. 2, 2015, 27 pages. |
Final Office Action received for U.S. Appl. No. 13/729,597, dated Jun. 26, 2015, 27 pages. |
Final Office Action received for U.S. Appl. No. 13/729,597, dated Jun. 29, 2017, 49 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,118, dated Dec. 7, 2012, 34 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,118, dated Jan. 30, 2015, 20 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,118, dated Jun. 6, 2014, 33 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/729,597, dated Dec. 18, 2014, 19 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/729,597, dated Dec. 21, 2016, 30 pages. |
Notice of Allowance received for U.S. Appl. No. 13/729,597, dated Apr. 18, 2018, 12 pages. |
Notice of Allowance received for U.S. Appl. No. 13/729,597, dated Nov. 30, 2017, 10 pages. |
Non-Final Office Action received for U.S. Appl. No. 13/251,118, dated Oct. 4, 2019, 22 pages. |
Notice of Allowance received for Japanese Patent Application No. 2020-082199, dated Sep. 4, 2020, 4 pages (1 page of English Translation and 3 pages of Official Copy). |
Office Action received for Australian Patent Application No. 2020220088, dated Sep. 14, 2020, 6 pages. |
Notice of Allowance received for U.S. Appl. No. 13/251,118, dated Mar. 26, 2020, 17 pages. |
Eustice, et al., “A universal information appliance”, IBM Systems Journal, vol. 38, No. 4, 1999, pp. 575-601. |
Applicant Initiated Interview Summary received for U.S. Appl. No. 15/193,971, dated Jun. 15, 2020, 3 pages. |
Notice of Allowance received for Korean Patent Application No. 10-2020-7010042, dated Apr. 30, 2020, 5 pages (2 pages of English Translation and 3 pages of Official Copy). |
Notice of Allowance received for U.S. Appl. No. 15/193,971, dated Jul. 14, 2020, 15 pages. |
Sow et al., “Uncovering the to-dos hidden in your in-box”, IBM Systems Journal, vol. 45, No. 4, 2006, pp. 739-757. |
Final Office Action received for U.S. Appl. No. 15/193,971, dated Mar. 19, 2020, 18 pages. |
Notice of Acceptance received for Australian Patent Application No. 2018204265, dated Feb. 27, 2020, 3 pages. |
Notice of Acceptance received for Australian Patent Application No. 2020202534, dated May 21, 2020, 3 pages. |
Notice of Allowance received for Japanese Patent Application No. 2017-068594, dated Aug. 7, 2020, 4 pages (1 page of English Translation and 3 pages of Official Copy). |
Notice of Acceptance received for Australian Patent Application No. 2020220088, dated Feb. 4, 2021, 3 pages. |
Jiangwei606, “[Zhuan] Play ”Zhuan“ Siri-Siri function excavation”, Available online at https://www.feng.com/post/3255659, Nov. 12, 2011, 30 pages (17 pages of English Translation and 13 pages of Official Copy). |
Office Action received for Chinese Patent Application No. 201811108851.2, dated Feb. 3, 2021, 20 pages (8 pages of English Translation and 12 pages of Official Copy). |
Office Action received for Australian Patent Application No. 2020220088, dated Nov. 17, 2020, 7 pages. |
Notice of Allowance received for Chinese Patent Application No. 201811108851.2, dated Jun. 3, 2021, 2 pages (1 page of English Translation and 1 page of Official Copy). |
Notice of Allowance received for U.S. Appl. No. 16/895,944, dated Jun. 4, 2021, 24 pages. |
Notice of Allowance received for Japanese Patent Application No. 2020-167209, dated Jul. 12, 2021, 4 pages (1 page of English Translation and 3 pages of Official Copy). |
Number | Date | Country | |
---|---|---|---|
20190014450 A1 | Jan 2019 | US |
Number | Date | Country | |
---|---|---|---|
61493201 | Jun 2011 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13729597 | Dec 2012 | US |
Child | 16036730 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13251118 | Sep 2011 | US |
Child | 13729597 | US |