This application is related to co-pending U.S. patent application Ser. No. 11/270,407, filed Nov. 9, 2005, and entitled, “ADAPTIVE TASK FRAMEWORK”, co-pending U.S. patent application Ser. No. 11/270,393, filed Nov. 9, 2005, and entitled, “ADAPTIVE TASK FRAMEWORK”, and, co-pending U.S. patent application Ser. No. 11/290,076, filed Nov. 30, 2005, and entitled, “ADAPTIVE SEMANTIC REASONING ENGINE”. The entirety of the aforementioned applications is hereby incorporated by reference.
Human languages are rich and complicated, including huge vocabularies with complex grammar and contextual meaning. Machine interpretation of human language, even in a very limited way, is an extremely complex task and continues to be the subject of extensive research. Providing users with the ability to communicate their desires to an automated system without requiring users to learn a machine specific language or grammar would decrease learning costs and greatly improve system usability. However, users become quickly frustrated when automated systems and machines are unable to interpret user input correctly, resulting in unexpected results.
Natural language input can be useful for a wide variety of applications, including virtually every software application with which humans are intended to interact. Typically, during natural language processing the natural language input is separated into tokens and mapped to one or more actions provided by the software application. Each application can have a unique set of actions. Consequently, it can be both time-consuming and repetitive for software developers to draft code to interpret natural language input and map the input to the appropriate action for each application.
The Internet in particular has provided users with a mechanism for obtaining information regarding any suitable subject matter. For example, various websites are dedicated to posting text, images, and video relating to world, national, and/or local news. A user with knowledge of a Uniform Resource Locator (URL) associated with one of such websites can simply enter the URL into a web browser to be provided with the website and access content thereon. Another conventional manner of locating desired information from the Internet is through utilization of a search engine. For instance, a user can enter a word or series of words into a search field and thereafter initiate the search engine (e.g., through depression of a button, one or more keystrokes, voice commands, . . . ). The search engine then utilizes search algorithms to locate websites related to the word or series of words entered by the user into the search field, and the user can then select one of the websites returned by the search engine to review content therein.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
A task system and method are provided. Task(s) are useful in modeling actions user(s) perform, for example, on the web. Examples of tasks include buying plane tickets, purchasing books, reserving cars, buying stocks, checking accounts, or checking the weather in a particular city. Task(s) can be executed on a variety of websites and individual user(s) can have individual preferences as to which site best matches their needs.
The system provides an automated approach for task creation, maintenance and/or execution. Furthermore, by distributing the processes of task creation, maintenance and/or execution to end-user(s), the entire task system can improve without the guidance of a central authority.
The system includes a browser that receives search results and task(s) associated with a query from a search engine. The system further includes a browser helper object that binds to the browser at runtime. The browser helper object provides information (e.g., to the search engine) associated with a user's action with respect to the search results and/or at least one task.
The system automatically generates and/or executes task(s) using Browser Helper Object(s) (BHOs). BHOs are objects that bind to the browser at runtime and behave as if they were part of the browser. The web browser is an application program that is capable of displaying a web page (e.g., Internet Explorer).
Thus, the BHO is a piece of code that effectively becomes part of the browser through the extensibility of the browser. BHOs are capable of accessing an object model corresponding to a schema of the source code (e.g., HTML) that is created when a web page is loaded. Additionally, BHOs can insert value(s) into form(s) (e.g., HTML form(s)). BHOs are further capable of communicating with other component(s).
In one example, the browser and the browser helper object(s) communicate directly with a semantic reasoning component. In another example, the browser and/or browser helper object(s) communicate with a search engine that in turn communicates with the semantic reasoning component. In a client-server environment, the semantic reasoning component can be a component of a task server. In a distributed processing environment, at least a portion of the semantic reasoning component can be resident on a user's computer system.
When a website is loaded, the BHO can walk an object model representation of the source code (e.g., HTML) and determine, for example, that there are INPUT box(es), for item(s) such as “Going to:” and “Leaving from:” fields and/or that there are SELECT box(e)s for the time of day containing elements such as “Morning”, “Noon”, “Evening” and “Anytime”. The BHO can insert a value into one or more of the input/selection box(es).
The BHO can communicate with the search engine to facilitate various function(s). For example, based on information received from the BHO, the search engine can log an original query entered to the browser. The search engine can further log website(s) within a time window that the user has visited and entered information contained in the original query.
The BHO can receive a uniform resource locator (URL) containing semantic information and fill-in value(s) on a form associated with the URL. For example, the query “flight from Boston to Seattle” can return a link to a website with some parameters such as tcy=“Boston” and fcy=“Seattle. The BHO can interpret this information and fill in information value(s) based, at least in part, upon the original query. The BHO can further store a query entered into a search page so that it can be used for logging purposes. Finally, the BHO can further appropriately display results received from the search engine.
The BHO can provide information regarding a URL selected by a user. This information can be employed by the semantic reasoning component to update information retrieval and/or query classification model(s). The BHO can provide information regarding a website manually navigated to be a user. This information can be employed by the semantic reasoning engine to create new task(s). Finally, if information from the query is included in a form filled in by a user, the BHO can provide information regarding a form filled by the user. Slot filling model(s) of the semantic reasoning component can be updated based on this information.
To the accomplishment of the foregoing and related ends, certain illustrative aspects are described herein in connection with the following description and the annexed drawings. These aspects are indicative, however, of but a few of the various ways in which the principles of the claimed subject matter may be employed and the claimed subject matter is intended to include all such aspects and their equivalents. Other advantages and novel features of the claimed subject matter may become apparent from the following detailed description when considered in conjunction with the drawings.
The claimed subject matter is now described with reference to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the claimed subject matter. It may be evident, however, that the claimed subject matter may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to facilitate describing the claimed subject matter.
As used in this application, the terms “component,” “handler,” “model,” “system,” and the like are intended to refer to a computer-related entity, either hardware, a combination of hardware and software, software, or software in execution. For example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration, both an application running on a server and the server can be a component. One or more components may reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers. Also, these components can execute from various computer readable media having various data structures stored thereon. The components may communicate via local and/or remote processes such as in accordance with a signal having one or more data packets (e.g., data from one component interacting with another component in a local system, distributed system, and/or across a network such as the Internet with other systems via the signal). Computer components can be stored, for example, on computer readable media including, but not limited to, an ASIC (application specific integrated circuit), CD (compact disc), DVD (digital video disk), ROM (read only memory), floppy disk, hard disk, EEPROM (electrically erasable programmable read only memory) and memory stick in accordance with the claimed subject matter.
Referring to
The process of creating a comprehensive task list can be daunting. The sheer number of sites that exist covering a wide range of services can be prohibitive to employ a manual method for creating tasks and maintaining them. The system 100 provides an automated approach for task creation, maintenance and/or execution. Furthermore, by distributing the processes of task creation, maintenance and/or execution to end user(s), the entire task system can improve without the guidance of a central authority.
The system 100 automatically generates task(s) and/or executes them using Browser Helper Object(s) (BHOs) 110. BHOs 110 are objects that bind to a browser 120 at runtime and behave as if they were part of the browser 120. The BHO 110 provides information associated with a user's action with respect to search results and/or task(s), as discussed below. The web browser 120 is an application program that is capable of displaying a web page (e.g., Internet Explorer).
The BHO 110 is a piece of code that effectively becomes part of the browser 120 through the extensibility of the browser 120. BHOs 110 are capable of accessing an object model corresponding to a schema of the source code (e.g., HTML) that is created when a web page is loaded. Additionally, BHOs 110 can insert value(s) into form(s) (e.g., HTML form(s)). BHOs 110 are further capable of communicating with other component(s, as discussed below.
In one example, the browser 110 and the browser helper object(s) 120 communicate directly with a semantic reasoning component 130. In another example, as illustrated in
In a client-server environment, the semantic reasoning component 130 can be a component of a task server (not shown). In a distributed processing environment, at least a portion of the semantic reasoning component 130 can be resident on a user's computer system (not shown).
When a website is loaded, the BHO 110 can walk an object model representation of the source code (e.g., HTML) and determine, for example, that there are INPUT box(es), for item(s) such as “Going to:” and “Leaving from:” fields and/or that there are SELECT box(e)s for the time of day containing elements such as “Morning”, “Noon”, “Evening” and “Anytime”. The BHO 110 can insert a value into one or more of the input/selection box(es). For example, the BHO 110 can insert a value of “Boston” into an INPUT box named “tcy”, if the BHO 110 is notified that Value(tcy)=“Boston”.
The BHO 110 can communicate with the search engine 210 to facilitate various function(s). For example, based on information received from the BHO 110, the search engine can log an original query entered to the browser 110 (e.g., “flight to Boston from Seattle”). The search engine 210 can further log website(s) within a time window that the user has visited and entered “Boston” or “Seattle” as values in the web form (e.g., Site=Expedia.com). Based upon information received from the BHO 110, the search engine 210 can also log value(s) on the form (e.g., tcy=“Boston”, fcy=“Seattle”). In one example, the information provided to search engine 210 by the BHO 110 does not contain any information which is constituted as personally identifiable.
The BHO 110 can receive a uniform resource locator (URL) containing semantic information and fill-in value(s) on a form associated with the URL. For example, the query “flight from Boston to Seattle” can return a link to a website with some parameters such as tcy=“Boston” and fcy=“Seattle. The BHO 110 can interpret this information and fill in information value(s) based, at least in part, upon the original query. The BHO 110 can further store a query entered into a search page so that it can be used for logging purposes. Finally, the BHO 110 can further appropriately display results received from the search engine 210.
Semantic Reasoning Component 130
Providing Search Results for Query(ies)
The semantic reasoning component 130 is responsible for providing search results for queries. A first step returns a list of potential tasks using a combination of Information Retrieval (IR) and Query Classification (QC) applications. This step can return a list of likely tasks such as Expedia.com, Travelocity.com, Orbitz.com, etc. A second step takes the query and potential task (Expedia.com) and attempts to fill in parameter value(s) in the task such as determining that the variable “tcy” should have the value “Boston”.
Logging of User Data
The semantic reasoning component 130 can further log data returned from BHOs 110. For example, the semantic reasoning component 130 can log the task selected by the user for a model employed by the QC application (QC model). Additionally, the semantic reasoning component 130 can log site(s) the user goes to and enters information contained in the original query. The semantic reasoning component 130 can log value(s) entered on the web page (e.g., according to privacy laws and policy).
Creating New Task(s)
Additionally, the semantic reasoning component 130 can create new task(s) based, at least in part, upon information provided by the BHOs 110. For example, when a user manually visits a new site after entering a search query the data can be provided by the BHO 110 to the semantic reasoning component 130 (e.g., via the search engine 210) that logs the data. If the site has not been seen before, a new task can be created.
Updating the Reasoning and/or Scoring Models
The semantic reasoning component 130 can include a QC model employed for task retrieval and/or a slot-filling model (as discussed below). One or both of these models can be updated periodically to reflect user feedback.
Overview of the Semantic Reasoning Component 130
The semantic reasoning component 130 provides the search engine 210 with a standardized method for interpreting natural language input. Additionally, the semantic reasoning component 130 can provide application developer(s) with a standard manner of defining the tasks the search engine 210 is capable of performing. A task, as used herein, describes and defines a fundamental unit of action relevant to user. The semantic reasoning component 130 enables the search engine 210 to define and manage tasks. Task(s) may be defined using extended markup language (XML), databases, text files or in any other suitable manner.
The search engine 210 can receive any manner of natural language input (e.g., handwritten text, tablet input, speech and typed text). The search engine 210 can process the natural language input to generate a query for processing by the semantic reasoning component 130 (e.g., a simple string of text characters). The semantic reasoning component 130 selects one or more tasks based, at least in part, upon the query. The semantic reasoning component 130 can provide the task along with associated metadata and slot information to describe an action to the search engine 210.
Overall, the semantic reasoning component 130 is responsible for:
a. Receiving an input query;
b. Finding a set of likely tasks given the input query;
c. Filling out slot values given a task and the input query;
d. Retrieving the most likely tasks with the most like slot values; and,
e. (optionally) Receiving feedback data to learn from user feedback
Referring to
The task 300 can include an entity component 310. The entity component 310 can include one or more named entities. A named entity, as used herein, is a token that is known to have a specific meaning. The named entity can be task specific or can be utilized with multiple tasks. The task 300 can include a named entity (NE) recognizer component 312. The NE recognizer component 312 can include one or more recognizers capable of matching tokens or portions of the natural language input to the entities included in the entity component 310. The NE recognizers 312 are capable of recognizing tokens corresponding to the named entities contained within the entities component 310. These tokens have a specific task meaning. Recognizers may be general or may be specific to a certain category of tokens. For example, a city recognizer may include a list of names (e.g., Seattle, Boston). Similarly, a date recognizer may be capable of recognizing and interpreting dates, such as “Jun. 14, 2005.” The software developer may define certain recognizers when specifying a task.
The task 300 can also include a keyword component 314. The keyword component 314 can include one or more keywords. Keywords can be used to select a task 300 from a set of tasks. For example, the “BookFlight” task keyword component 314 can include keywords such as “Book Flight,” “airline” and the like. The keywords can be determine by the software developer or automatically generated by the semantic reasoning component 130. In addition, the semantic reasoning component 130 can add additional keywords to the keyword component 314 based upon natural language input, user actions and/or user feedback. Furthermore, the keywords may be weighted, such that the presence of certain keywords in the query is more likely to surface certain tasks. Such weight can also be used to rank or order a selected group of tasks.
The task 300 can also include a slot component 308 that specifies or defines slots for information required for the task. The slot component 308 can provide a mechanism for defining parameters used by the task. For example, a task that books airline flights may include slots for the arrival city, the departure city, the flight date and time. The slot component 308 can include any integer number of slots, from zero to N. Typically, information from the natural language input is used to fill the slots.
Turning next to
The slot 400 can also include an annotation component 406. The annotation component 406 can include one or more annotations. Annotations are tokens that mark or indicate the significance of other tokens. The annotation component 406 identifies an annotation token and uses that information to interpret other tokens within the natural language input. For example, the token “from” when contained within a natural language input string that maps to a “BookFlight” task indicates that the token that follows is likely to contain the name of the departure city. Annotations may appear either before or after the relevant token. For example, the token “departure city” when contained within a natural language input string that maps to a “BookFlight” task indicates that the token that precedes it is likely to contain the name of the departure city. Consequently, the phrase “leaving from Boston” and “Boston departure city” can both be interpreted to fill the departure city slot with the value “Boston.” Annotations which appear before the token are called pre-indicators, while annotations which follow the relevant token are called post-indicators. The annotation component 406 can recognize task system defined annotations as well as task specific annotations.
Next, referring to
Tasks can be generated by one or more applications or tasks can be generated automatically by the task framework 500. In addition, the task framework 500 may update or modify tasks generated by application(s). The task component 502 can be a flat file, a database or any other structure suitable for containing the data for one or more tasks.
The task framework 500 can include a task retrieval component 504. The task retrieval component 504 uses the query to select one or more tasks from the collection of tasks contained within the task component 502. The task retrieval component 504 may determine the appropriate task to be retrieved from the task component 502 based upon keywords in the query. The collection of tasks in the task component 502 can be indexed based upon the task keywords. The tokens contained within the query can be used to select an appropriate task or set of tasks. The application can also include additional information with the query. For example, the application could pass user context information to the framework to be used in the selection of the appropriate task. The task retrieval component 504 can use a variety of methodologies to select appropriate tasks. The task retrieval component 504 can be trained to improve performance based upon user actions and responses to the selected tasks.
In addition, the task framework 500 can include a slot-filling component 506. The slot-filling component 506 can be responsible for providing the best matching of the list of tokens from the natural language input or query with the task parameters. Typically, a slot-filling component 506 can receive a list of tokens and one or more tasks. The slot-filling component 506 can generate one or more possible mappings of the tokens to the slots of the task. The slot-filling component 506 can generate a score or rank for each of the possible mappings of tokens to task slots. The slot-filling component 506 can use a mathematical model, algorithm or function to calculate a score or rank for mappings. The slot-filling component 506 can utilize a heuristic function, a hidden Markov model, a Naïve Bayes based model, Maximum Entropy/Minimum Divergence Models (MEMD), blending strategies, linear discriminative models or any combination thereof to calculate a score for a mapping of tokens to a task.
The slot-filling component 506 can include a method responsible for taking the natural language input, culture information, a list of tokens, a list of named entities, a task and a predetermined maximum number of desired solutions. Culture information can include information such as the writing system and formatting utilized by the relevant culture. Named entities identify tokens with a specific meaning to the slot-filling system (e.g., Boston). The slot-filling component 506 can produce a list of up to the maximum number of requested semantic solutions with a semantic solution representing a mapping of tokens to slots that can be used by the search engine 210.
Optionally, the task framework 500 can also include a logging component 508. Tasks can pass information or feedback to the task framework 500 after completion of the task or during task processing. The logging component 508 stores the feedback information. This information can be used to train the task framework 500 and improve system performance. The feedback from tasks can include user actions. The task framework 500 can include a defined intent interface to facilitate feedback.
In addition, the task framework 500 or the slot-filling component 506 can include one or more GlobalRecognizers that provide the ability to recognize tokens that have special meaning to the task system in general. For example, the token “Boston” has special meaning as the city of Boston, Mass. The GlobalRecognizers property provides a set of recognizer components that identify special tokens, making them available throughout the entire system and across multiple tasks. For example, there may be several tasks that utilize “city,” “date” or “number” entities. Entities are a mechanism for providing type information. For example the “city” entity includes a set of annotations (e.g., “city,” “place,” and “town”). Occurrences of the annotations within the list of tokens indicate the likelihood of a “city” entity. GlobalRecognizers allows such entities or special tokens to be defined once rather than for each individual task.
In summary, keywords are terms that might be used to surface a task. Slots are parameter values that may or may not be filled by the user Query. Slots are uniquely specified by their Name and Type.
Additionally, preIndicators are words that might disambiguate slots by occurring before a value “to Boston” would prefer the “Arrival City” slot over the “Departure City” slot even though Boston maps to CITY and can be a value for either slot. PostIndicators are words that might disambiguate slots by occurring before a value “from Boston” would prefer the “Departure City” slot over the “Arrival City” slot even though Boston maps to CITY and can be a value for either slot. Consider the example of Table 1:
Given the schema of Table 1, the following queries match the ReserveFlight Task:
Additionally, as discussed previously, the semantic reasoning component 130 can employ user feedback to learn from user behavior such that if users start entering queries such as “departing Boston for Seattle” to mean “Departure City”=“Boston” and “Arrival City”=“Seattle”. The semantic reasoning component 130 will automatically learn the pattern “departing <Departure City> for <Arrival City>” without needing to explicitly add new Pre or Post indicators to the task definition.
User Perspective
Returning to
Referring back to
Continuing with
If there exists information entered into the form that was contained in the original query, the BHO 110 can send the information back to the semantic reasoning component 130 (e.g., in accordance with the privacy laws and policies). For example, if the user has typed “Boston” into a field named “originCity” then the query, URL of the website, the field name, and field value can be sent back to the semantic reasoning component 130.
If the URL is not contained in the database of tasks, the semantic reasoning component 130 can create a new task. The URL can be normalized to prevent millions of tasks from being created where the web pages are effectively the same but each individual URL may be unique due to cookies and/or other information supplied randomly. Thereafter, the object model (e.g., HTML object model) can be walked and a task created containing, for example:
Additionally, the model(s) for slot filling can be updated accordingly. For example, a new parameter can be created, if it doesn't already exist. Further, the model(s) for slot filling can learn a pattern such as “from X” will put X in the “originCity” field.
Next, referring to
For purposes of explanation, task(s) can be limited to form-enabled task(s), that is, task(s) that have a form as their input parameter(s). The output of the task extractor 800 is task object(s) that can be stored in a task store 810 (e.g., to be utilized by the semantic reasoning component 130).
Since task object(s) generally require field(s) of keyword(s) and a description of the task, a significant function of the task extractor 800 is discovering semantic information about the task (e.g., the functionality of the task for end user(s)). Additionally, the task extractor 800 can discover a description user(s) would type in when they want to perform this task. In one example, this functionality can be obtained by obtaining information from the form (e.g., HTML form) and its context. In another example, a query probing technique can be employed.
Given the seed website, which is identified as containing common task(s) beforehand, a crawler 820 can first crawl the web page(s) under this seed website and write them into a data store 830. Also, the crawler 820 can record linkage between different web pages in another table, for the reason that the links to and from a web page may induce some semantic information for tasks on the web page.
Next, a form filter 840 and, optionally, a schema probe 850 can employ the web page information stored in the data store 830. The form filter 840 can extract form(s) (e.g., HTML form(s)) from raw text (e.g., HTML raw text). Furthermore, optionally, the form filter 840 can filter out form(s) having the same functionality and/or pointing to the same action, which is often the case for web pages under the same website. A simple example for this situation is that there would be many pages having a particular search engine's form, but it is desirable to only have a single task object in the task store 810.
The form filter 840 passes filtered form(s) to the parser 860 which then extracts structured information of the form(s), including action URI (universal resource identifier), method, input type, etc. Moreover, the default value for INPUT in HTML form(s) may provide information about the slot entity.
Next, a context semantic extractor 870 captures semantic information other than that in HTML tags. Both slot-level information as well as task-level information can be captured. For example, with respect to word(s) that appear immediately in front of an INPUT element are highly likely to bring in slot-level information for the INPUT element. On the other hand, the TITLE of the web page and/or words immediately before or after the form may provide task-level information. However, in one example, it is likely that simply extracting information in certain contextual positions does not yield acceptable performance. In this case, substantially all the data on the web page can be employed as a richer context, which at the same time brings in a quantity of unwanted noise. A weighted importance model for data on the same web page can be introduced to address this noise-filtering/relevant information extraction issue. Importance can depend, for example, on the distance from the Form, or the IDF of that word, etc.
The optional schema probe 850 can provide information about the entities for slots. The schema probe 850 can automatically generate query and obtain feedback and/or more description information about slot entities.
For each form, a task generator 880 can collect the task information from the parser 860, the context semantic extractor 870, and, optionally, the schema probe, to create a task object. Thereafter, the task generator 880 can store the generated task object in the task store 810.
It is to be appreciated that the system 100, the browser 110, the browser help object(s) 120, the semantic reasoning component 130, the system 200, the search engine 210, the task extractor 800, the task store 810, the crawler 820, the data store 830, the form filter 840, the schema probe 850, the parser 860, the context semantic extractor 870 and/or the task object generator 880 can be computer components as that term is defined herein.
Turning briefly to
The claimed subject matter may be described in the general context of computer-executable instructions, such as program modules, executed by one or more components. Generally, program modules include routines, programs, objects, data structures, etc. that perform particular tasks or implement particular abstract data types. Typically the functionality of the program modules may be combined or distributed as desired in various embodiments.
Referring to
If the determination at 912 is NO, at 928, the user manually navigates to a website. At 932, the browser helper object provides information regarding the website (e.g., URL) to the search engine and/or semantic reasoning component.
At 924, the user enters information into a form associated with the website. At 936, the user submits the information. At 940, if any information from the original query is included in the form, the browser helper object provided information regarding the form filled by the user, for example, to the search engine and/or semantic reasoning component.
Turning to
At 1120, the search results and at least one task are provided to the user. At 1124, click-through information is obtained from a browser helper object. At 1128, the click-through information is provided to the semantic reasoning component.
At 1136, information regarding a user web action is received from the browser helper object. At 1140, the user web action is logged. Finally, at 1144, information regarding the user web action is provided to the semantic reasoning component.
Next referring to
At 1350, click-through information is received, for example, from the search engine and/or a browser helper object. At 1360, model(s) that perform task retrieval are updated based, at least in part, upon the click-through information.
At 1370, information regarding a user's web action is received, for example, from the search engine and/or a browser helper object. At 1380, model(s) that find best parameter value given the query are updated based, at least in part, upon the user web action.
In order to provide additional context for various aspects of the claimed subject matter,
With reference to
The system bus 1418 can be any of several types of bus structure(s) including the memory bus or memory controller, a peripheral bus or external bus, and/or a local bus using any variety of available bus architectures including, but not limited to, an 8-bit bus, Industrial Standard Architecture (ISA), Micro-Channel Architecture (MSA), Extended ISA (EISA), Intelligent Drive Electronics (IDE), VESA Local Bus (VLB), Peripheral Component Interconnect (PCI), Universal Serial Bus (USB), Advanced Graphics Port (AGP), Personal Computer Memory Card International Association bus (PCMCIA), and Small Computer Systems Interface (SCSI).
The system memory 1416 includes volatile memory 1420 and nonvolatile memory 1422. The basic input/output system (BIOS), containing the basic routines to transfer information between elements within the computer 1412, such as during start-up, is stored in nonvolatile memory 1422. By way of illustration, and not limitation, nonvolatile memory 1422 can include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable ROM (EEPROM), or flash memory. Volatile memory 1420 includes random access memory (RAM), which acts as external cache memory. By way of illustration and not limitation, RAM is available in many forms such as synchronous RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDR SDRAM), enhanced SDRAM (ESDRAM), Synchlink DRAM (SLDRAM), and direct Rambus RAM (DRRAM).
Computer 1412 also includes removable/nonremovable, volatile/nonvolatile computer storage media.
It is to be appreciated that
A user enters commands or information into the computer 1412 through input device(s) 1436. Input devices 1436 include, but are not limited to, a pointing device such as a mouse, trackball, stylus, touch pad, keyboard, microphone, joystick, game pad, satellite dish, scanner, TV tuner card, digital camera, digital video camera, web camera, and the like. These and other input devices connect to the processing unit 1414 through the system bus 1418 via interface port(s) 1438. Interface port(s) 1438 include, for example, a serial port, a parallel port, a game port, and a universal serial bus (USB). Output device(s) 1440 use some of the same type of ports as input device(s) 1436. Thus, for example, a USB port may be used to provide input to computer 1412, and to output information from computer 1412 to an output device 1440. Output adapter 1442 is provided to illustrate that there are some output devices 1440 like monitors, speakers, and printers among other output devices 1440 that require special adapters. The output adapters 1442 include, by way of illustration and not limitation, video and sound cards that provide a means of connection between the output device 1440 and the system bus 1418. It should be noted that other devices and/or systems of devices provide both input and output capabilities such as remote computer(s) 1444.
Computer 1412 can operate in a networked environment using logical connections to one or more remote computers, such as remote computer(s) 1444. The remote computer(s) 1444 can be a personal computer, a server, a router, a network PC, a workstation, a microprocessor based appliance, a peer device or other common network node and the like, and typically includes many or all of the elements described relative to computer 1412. For purposes of brevity, only a memory storage device 1446 is illustrated with remote computer(s) 1444. Remote computer(s) 1444 is logically connected to computer 1412 through a network interface 1448 and then physically connected via communication connection 1450. Network interface 1448 encompasses communication networks such as local-area networks (LAN) and wide-area networks (WAN). LAN technologies include Fiber Distributed Data Interface (FDDI), Copper Distributed Data Interface (CDDI), Ethernet/IEEE 802.3, Token Ring/IEEE 802.5 and the like. WAN technologies include, but are not limited to, point-to-point links, circuit switching networks like Integrated Services Digital Networks (ISDN) and variations thereon, packet switching networks, and Digital Subscriber Lines (DSL).
Communication connection(s) 1450 refers to the hardware/software employed to connect the network interface 1448 to the bus 1418. While communication connection 1450 is shown for illustrative clarity inside computer 1412, it can also be external to computer 1412. The hardware/software necessary for connection to the network interface 1448 includes, for exemplary purposes only, internal and external technologies such as, modems including regular telephone grade modems, cable modems and DSL modems, ISDN adapters, and Ethernet cards.
What has been described above includes examples of the claimed subject matter. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the claimed subject matter, but one of ordinary skill in the art may recognize that many further combinations and permutations of the claimed subject matter are possible. Accordingly, the claimed subject matter is intended to embrace all such alterations, modifications and variations that fall within the spirit and scope of the appended claims. Furthermore, to the extent that the term “includes” is used in either the detailed description or the claims, such term is intended to be inclusive in a manner similar to the term “comprising” as “comprising” is interpreted when employed as a transitional word in a claim.
Number | Name | Date | Kind |
---|---|---|---|
4736296 | Katayama et al. | Apr 1988 | A |
4965763 | Zamora | Oct 1990 | A |
4974191 | Amirghodsi et al. | Nov 1990 | A |
5208816 | Seshardi et al. | May 1993 | A |
5477451 | Brown et al. | Dec 1995 | A |
5577241 | Spencer | Nov 1996 | A |
5625814 | Luciw | Apr 1997 | A |
5636036 | Ashbey | Jun 1997 | A |
5696962 | Kupiec | Dec 1997 | A |
5748974 | Johnson | May 1998 | A |
5752244 | Rose et al. | May 1998 | A |
5754173 | Hiura et al. | May 1998 | A |
5754174 | Carpenter et al. | May 1998 | A |
5794259 | Kikinis | Aug 1998 | A |
5832459 | Cameron et al. | Nov 1998 | A |
5855015 | Shoham | Dec 1998 | A |
5892900 | Ginter et al. | Apr 1999 | A |
5999948 | Nelson et al. | Dec 1999 | A |
6055528 | Evans | Apr 2000 | A |
6078914 | Redfern | Jun 2000 | A |
6088700 | Larsen et al. | Jul 2000 | A |
6118939 | Nack et al. | Sep 2000 | A |
6212494 | Boguraev | Apr 2001 | B1 |
6278996 | Richardson et al. | Aug 2001 | B1 |
6314398 | Junqua et al. | Nov 2001 | B1 |
6513006 | Howard et al. | Jan 2003 | B2 |
6643620 | Contolini et al. | Nov 2003 | B1 |
6675159 | Lin et al. | Jan 2004 | B1 |
6678677 | Roux et al. | Jan 2004 | B2 |
6678694 | Zimmerman et al. | Jan 2004 | B1 |
6690390 | Walters et al. | Feb 2004 | B1 |
6724403 | Santoro et al. | Apr 2004 | B1 |
6751606 | Fries et al. | Jun 2004 | B1 |
6816857 | Weissman et al. | Nov 2004 | B1 |
6850252 | Hoffberg | Feb 2005 | B1 |
6901399 | Corston et al. | May 2005 | B1 |
6904402 | Wang et al. | Jun 2005 | B1 |
6910003 | Arnold et al. | Jun 2005 | B1 |
6947923 | Cha et al. | Sep 2005 | B2 |
7020607 | Adachi | Mar 2006 | B2 |
7020658 | Hill | Mar 2006 | B1 |
7328199 | Ramsey et al. | Feb 2008 | B2 |
7523099 | Egnor et al. | Apr 2009 | B1 |
20020032680 | Garber | Mar 2002 | A1 |
20020042793 | Choi | Apr 2002 | A1 |
20020045463 | Chen et al. | Apr 2002 | A1 |
20020049750 | Venkatram | Apr 2002 | A1 |
20020059132 | Quay | May 2002 | A1 |
20020065959 | Kim et al. | May 2002 | A1 |
20020101448 | Sanderson | Aug 2002 | A1 |
20020124115 | McLean et al. | Sep 2002 | A1 |
20020143949 | Rajarajan et al. | Oct 2002 | A1 |
20020152190 | Biebesheimer et al. | Oct 2002 | A1 |
20030023598 | Janakiraman et al. | Jan 2003 | A1 |
20030078766 | Appelt | Apr 2003 | A1 |
20030084035 | Emerick | May 2003 | A1 |
20030120700 | Boudnik et al. | Jun 2003 | A1 |
20030135584 | Roberts et al. | Jul 2003 | A1 |
20030222912 | Fairweather | Dec 2003 | A1 |
20040030556 | Bennett | Feb 2004 | A1 |
20040030697 | Cochran et al. | Feb 2004 | A1 |
20040030710 | Shadle | Feb 2004 | A1 |
20040034652 | Hofmann et al. | Feb 2004 | A1 |
20040111419 | Cook et al. | Jun 2004 | A1 |
20040117395 | Gong et al. | Jun 2004 | A1 |
20040122674 | Bangalore et al. | Jun 2004 | A1 |
20040130572 | Bala | Jul 2004 | A1 |
20040148154 | Acero et al. | Jul 2004 | A1 |
20040181749 | Chellapilla et al. | Sep 2004 | A1 |
20040220893 | Spivack | Nov 2004 | A1 |
20040236580 | Bennett | Nov 2004 | A1 |
20040250255 | Kraiss et al. | Dec 2004 | A1 |
20040260534 | Pak et al. | Dec 2004 | A1 |
20040260689 | Colace | Dec 2004 | A1 |
20040267725 | Harik | Dec 2004 | A1 |
20050027666 | Beck et al. | Feb 2005 | A1 |
20050028133 | Ananth et al. | Feb 2005 | A1 |
20050034098 | DeSchryver et al. | Feb 2005 | A1 |
20050044058 | Matthews et al. | Feb 2005 | A1 |
20050049852 | Chao | Mar 2005 | A1 |
20050049874 | Coffman et al. | Mar 2005 | A1 |
20050065995 | Milstein et al. | Mar 2005 | A1 |
20050075859 | Ramsey | Apr 2005 | A1 |
20050075878 | Balchandran et al. | Apr 2005 | A1 |
20050078805 | Mills et al. | Apr 2005 | A1 |
20050080625 | Bennett et al. | Apr 2005 | A1 |
20050080782 | Ratnaparkhi et al. | Apr 2005 | A1 |
20050086059 | Bennett | Apr 2005 | A1 |
20050114854 | Padisetty et al. | May 2005 | A1 |
20050131672 | Dalal et al. | Jun 2005 | A1 |
20050132380 | Chow | Jun 2005 | A1 |
20050137939 | Calabria et al. | Jun 2005 | A1 |
20050144064 | Calabria et al. | Jun 2005 | A1 |
20050144065 | Calabria et al. | Jun 2005 | A1 |
20050187818 | Zito et al. | Aug 2005 | A1 |
20050192992 | Reed | Sep 2005 | A1 |
20050193055 | Angel et al. | Sep 2005 | A1 |
20050216356 | Pearce et al. | Sep 2005 | A1 |
20050228744 | McHale | Oct 2005 | A1 |
20050246726 | Labrou | Nov 2005 | A1 |
20050257148 | Goodman et al. | Nov 2005 | A1 |
20050283473 | Rousso et al. | Dec 2005 | A1 |
20060005114 | Louch et al. | Jan 2006 | A1 |
20060005207 | Louch et al. | Jan 2006 | A1 |
20060059434 | Boss et al. | Mar 2006 | A1 |
20060064302 | Ativanichayaphong et al. | Mar 2006 | A1 |
20060064411 | Gross et al. | Mar 2006 | A1 |
20060107219 | Ahya et al. | May 2006 | A1 |
20060129463 | Zicherman | Jun 2006 | A1 |
20060156248 | Chaudhri et al. | Jul 2006 | A1 |
20060179404 | Yolleck et al. | Aug 2006 | A1 |
20060206835 | Chaudhri et al. | Sep 2006 | A1 |
20070027850 | Chan et al. | Feb 2007 | A1 |
20070038614 | Guha | Feb 2007 | A1 |
20070038934 | Fellman | Feb 2007 | A1 |
20070067217 | Schachter et al. | Mar 2007 | A1 |
20070100688 | Book | May 2007 | A1 |
20070101279 | Chaudhri et al. | May 2007 | A1 |
20070101297 | Forstall et al. | May 2007 | A1 |
20070106495 | Ramsey et al. | May 2007 | A1 |
20070106496 | Ramsey et al. | May 2007 | A1 |
20070124263 | Katariya et al. | May 2007 | A1 |
20070130124 | Ramsey et al. | Jun 2007 | A1 |
20070130134 | Ramsey et al. | Jun 2007 | A1 |
20070192179 | Van Luchene | Aug 2007 | A1 |
20070209013 | Ramsey et al. | Sep 2007 | A1 |
Number | Date | Country |
---|---|---|
1 571 578 | Sep 2005 | EP |
1 580 666 | Sep 2005 | EP |
0129823 | Apr 2001 | WO |
03001413 | Jan 2003 | WO |
WO 2004017230 | Feb 2004 | WO |
WO 2005036365 | Apr 2005 | WO |
Number | Date | Country | |
---|---|---|---|
20070130186 A1 | Jun 2007 | US |