The following relates to data processing systems and processes, and more particularly relates to systems and processes for processing record actions for multiple tenants in a multi-tenant database system environment.
Modern software development is evolving away from the client-server model toward “cloud”-based processing systems that provide access to data and services via the Internet or other networks. In contrast to prior systems that hosted networked applications on dedicated server hardware, the cloud computing model allows applications to be provided over the network “as a service” supplied by an infrastructure provider. The infrastructure provider typically abstracts the underlying hardware and other resources used to deliver a customer-developed application so that the customer no longer needs to operate and support dedicated server hardware. The cloud computing model can often provide substantial cost savings to the customer over the life of the application because the customer no longer needs to provide dedicated network infrastructure, electrical and temperature controls, physical security and other logistics in support of dedicated server hardware.
Multi-tenant platforms can facilitate effective collaboration between users, including between users in different tenants. One way that multi-tenant platforms facilitate effective collaboration is by providing for the controlled sharing of data between tenants. For example, one organization may elect to share various types of product information with other tenants as a way to facilitate cooperation on a host of issues. Multi-tenant platforms may facilitate such sharing data though the controlled sharing of selected records. When a selected record is shared, a connection is made from the originating tenant to the target tenant, allowing the target tenant to access the record.
One issue that arises with such sharing of records occurs when such records are updated or otherwise changed. Specifically, when a record is changed there is typically a need to update that record for all users with which that record has been shared. When large numbers of records are shared with large numbers of users, such updating of records may be very resource intensive. In some cases, the wait for such an updating to occur may be significant. Furthermore, users with large numbers of shared records and correspondingly large numbers of updates to shared records can hog resources and delay processing for other users. Thus, there remains a continuing need for systems methods for sharing and updating of records between tenants that improve the processing performance of such sharing.
Exemplary embodiments will hereinafter be described in conjunction with the following drawing figures, wherein like numerals denote like elements, and
According to various exemplary embodiments, systems and methods are provided to effectively processing record actions in multi-tenant database system. This facilitates improved sharing of records between users of different tenants in the multi-tenant system.
Turning now to
Database 130 is any sort of repository or other data storage system capable of storing and managing data 132 associated with any number of tenants. Database 130 may be implemented using any type of conventional database server hardware. In various embodiments, database 130 shares processing hardware 104 with server 102. In other embodiments, database 130 is implemented using separate physical and/or virtual database server hardware that communicates with server 102 to perform the various functions described herein.
Server 102 is implemented using one or more actual and/or virtual computing systems that collectively provide a dynamic application platform 110 for generating virtual applications 128A-C. Server 102 operates with any sort of conventional computing hardware 104, such as any processor 105, memory 106, input/output features 107 and the like. Processor 105 may be implemented using one or more of microprocessors, microcontrollers, processing cores and/or other computing resources spread across any number of distributed or integrated systems, including any number of “cloud-based” or other virtual systems. Memory 106 represents any non-transitory short or long term storage capable of storing programming instructions for execution on processor 105, including any sort of random access memory (RAM), read only memory (ROM), flash memory, magnetic or optical mass storage, and/or the like. Input/output features 107 represent conventional interfaces to networks (e.g., to network 145, or any other local area, wide area or other network), mass storage, display devices, data entry devices and/or the like. In a typical embodiment, application platform 110 gains access to processing resources, communications interfaces and other features of hardware 104 using any sort of conventional or proprietary operating system 108. As noted above, server 102 may be implemented using a cluster of actual and/or virtual servers operating in conjunction with each other, typically in association with conventional network communications, cluster management, load balancing and other features as appropriate.
As noted above, a “tenant” generally refers to a group of users that shares access to common data within database 130. Tenants may represent customers, customer departments, business or legal organizations, and/or any other entities that maintain data for particular sets of users within system 100. Although multiple tenants may share access to a common server 102 and database 130, the particular data and services provided from server 102 to each tenant can be securely isolated from those provided to other tenants, as described more fully below. The multi-tenant architecture therefore allows different sets of users to share functionality without necessarily sharing each other's data 132.
Furthermore, the multi-tenant architecture may be configured to allow selective sharing of data between users. In such implementations, individual records in data 132 may be shared with other users, thus facilitating various types of cooperation between tenants in the multi-tenant system. For example, one organization may elect to share various types of sales and product information in data 132 with other tenants as a way to facilitate cooperation between customers and clients. The server 102 may be configured to facilitate sharing of such on a record by record basis. When a selected record is shared, a connection is made from the originating tenant to the target tenant, allowing the target tenant to access the record. The connection allows the target tenant to access a copy of the record, and also provides for the updating of the copy when the originating tenant updates the record.
Data 132 may be organized and formatted in any manner to support multi-tenant application platform 110. In various embodiments, data 132 is suitably organized into a relatively small number of large data tables to maintain a semi-amorphous “heap”-type format. Data 132 can then be organized as needed for a particular virtual application 129A-C. In various embodiments, conventional data relationships are established using any number of pivot tables 234 that establish indexing, uniqueness, relationships between entities, and/or other aspects of conventional database organization as desired.
Further data manipulation and report formatting is generally performed at run-time using a variety of meta-data constructs. Metadata within a universal data directory (UDD) 236, for example, can be used to describe any number of forms, reports, workflows, user access privileges, business logic and other constructs that are common to multiple tenants. Tenant-specific formatting, functions and other constructs may be maintained as tenant-specific metadata 238A-B for each tenant, as desired. Rather than forcing data 132 into an inflexible global structure that is common to all tenants and applications, then, database 130 is organized to be relatively amorphous, with tables 234 and metadata 236-238 providing additional structure on an as-needed basis. To that end, application platform 110 suitably uses tables 234 and/or metadata 236, 238 to generate “virtual” components of applications 129A-C to logically obtain, process, and present the relatively amorphous data 132 from database 130.
Application platform 110 is any sort of software application or other data processing engine that generates virtual applications 129A-C that provide data and/or services to client devices 140A-B. Virtual applications 129A-C are typically generated at run-time in response to queries received from client devices 140A-B. In the example illustrated in
Runtime application generator 220 dynamically builds and executes virtual applications 129A-C in response to specific requests received from client devices 140A-B. Virtual applications 129A-C created by tenants are typically constructed in accordance with tenant-specific metadata 238, which describes the particular tables, reports, interfaces and/or other features of the particular application. In various embodiments, each virtual application 129A-C generates dynamic web content that can be served to a browser or other client program 142A-B associated with client device 140A-B, as appropriate.
Application generator 220 suitably interacts with query generator 214 to efficiently obtain multi-tenant data 132 from database 130 as needed. In a typical embodiment, query generator 214 considers the identity of the user requesting a particular function, and then builds and executes queries to database 130 using system-wide metadata 236, tenant specific metadata 238, pivot tables 234 and/or any other available resources. Query generator 214 in this example therefore maintains security of the multi-tenant database 130 by ensuring that queries are consistent with access privileges granted to the user that initiated the request.
Data processing engine 212 performs bulk processing operations on data 132 such as uploads or downloads, updates, online transaction processing and/or the like. In many embodiments, less urgent bulk processing of data 132 can be scheduled to occur as processing resources become available, thereby giving priority to more urgent data processing by query generator 214, search engine 216, virtual applications 129A-C and/or the like. Again, the various components, modules and inter-relationships of other application platforms 120 may vary from the particular examples described herein.
In operation, then, developers use application platform 110 to create data-driven virtual applications 129A-C for the tenants that they support. Such applications 129A-C may make use of interface features such as tenant-specific screens 224, universal screens 222 or the like. Any number of tenant-specific and/or universal objects 226 may also be available for integration into tenant-developed applications 129A-C. Data 132 associated with each application 129A-C is provided to database 130, as appropriate, and stored until requested, along with metadata 138 that describes the particular features (e.g., reports, tables, functions, etc.) of tenant-specific application 129A-C until needed.
Data and services provided by server 102 can be retrieved using any sort of personal computer, mobile telephone, tablet or other network-enabled client device 140 on network 145. Typically, the user operates a conventional browser or other client program 242 to contact server 102 via network 145 using, for example, the hypertext transport protocol (HTTP) or the like. The user typically authenticates his or her identity to the server 102 to obtain a session identification (“SessionlD”) that identifies the user in subsequent communications with server 102. When the identified user requests access to a virtual application 129A-C, application generator 220 suitably creates the application at run time based upon metadata 236 and 238, as appropriate. Query generator 214 suitably obtains the requested data 132 from database 130 as needed to populate the tables, reports or other features of virtual application 129A-C. As noted above, the virtual application 129A-C may contain Java, ActiveX or other content that can be presented using conventional client software 142A-B running on client device 140A-B; other embodiments may simply provide dynamic web or other content that can be presented and viewed by the user, as desired.
The virtual applications 129A-C may access the common database 130. Using any number of client devices 140A-B, users may the virtual applications 129A-C and the common database 130 via network 145. Again, the common platform 10 provides secure access to contact database in database 130 for each of the various tenants subscribing to system 100.
In some embodiments, a plurality of such servers 102 may be communicably coupled to the multi-tenant database 130 and have access to data 132 and tenant metadata 238. In certain embodiments, each server 132 is configured to handle requests for any user associated with any organization that is a tenant. Because it may desirable to be able to add and remove servers 132 from the server pool at any time for any reason, there is preferably no server affinity for a user and/or organization to a specific server 132. In this manner, system 200 is multi-tenant, wherein system 200 handles storage of, and access to, different objects, data and applications across disparate users and organizations.
As an example of storage, one tenant might be a company that employs a sales force where each salesperson uses system 200 to manage their sales process. Thus, a user might maintain contact data, leads data, customer follow-up data, performance data, goals and progress data, etc., all applicable to that user's personal sales process (e.g., in tenantl meta data 238A). In an example of a multi-tenant system arrangement, since all of the data and the applications to access, view, modify, report, transmit, calculate, etc., can be maintained and accessed by a user system having nothing more than network access, the user can manage his or her sales efforts and cycles from any of many different user systems. For example, if a salesperson is visiting a customer and the customer has Internet access in their lobby, the salesperson can obtain critical updates as to that customer while waiting for the customer to arrive in the lobby.
While each user's data might be separate from other users' data regardless of the employers of each user, some data might be organization-wide data shared or accessible by a plurality of users or all of the users for a given organization that is a tenant. Thus, there might be some data structures managed by system 102 that are allocated at the tenant level while other data structures might be managed at the user level. Because a multi-tenant system might support multiple tenants including possible competitors, the multi-tenant system should have security protocols that keep data, applications, and application use separate. Also, because many tenants may opt for access to an multi-tenant system rather than maintain their own system, redundancy, up-time, and backup are additional functions that may be implemented in the multi-tenant system. In addition to user-specific data and tenant specific data, multi-tenant system 100 might also maintain system level data usable by multiple tenants or other data. Such system level data might include industry reports, news, postings, and the like that are sharable among tenants.
In certain embodiments, client devices 140A-B communicate with servers 102 to request and update system-level and tenant-level data that may require sending one or more queries to the multi-tenant database 130.
In one embodiment, the database 130 may generally be viewed as a collection of objects, such as a set of logical tables, containing data fitted into predefined categories. A “table” is one representation of a data object, and may be used herein to simplify the conceptual description of objects and custom objects. It should be understood that “table” and “object” may be used interchangeably herein. Each table generally contains one or more data categories logically arranged as columns or fields in a viewable schema. Each row or record of a table contains an instance of data for each category defined by the fields. For example, a CRM database may include a table that describes a customer with fields for basic contact information such as name, address, phone number, fax number, etc. Another table might describe a purchase order, including fields for information such as customer, product, sale price, date, etc. In some multi-tenant database systems, standard entity tables might be provided for use by all tenants. For CRM database applications, such standard entities might include tables for Account, Contact, Lead, and Opportunity data, each containing pre-defined fields. It should be understood that the word “entity” may also be used interchangeably herein with “object” and “table”.
In some multi-tenant database systems, tenants may be allowed to create and store custom objects, or they may be allowed to customize standard entities or objects, for example by creating custom fields for standard objects, including custom index fields. U.S. patent application Ser. No. 10/817,161, filed Apr. 2, 2004, entitled “Custom Entities and Fields in a Multi-Tenant Database System”, and which is hereby incorporated herein by reference, teaches systems and methods for creating custom objects as well as customizing standard objects in a multi-tenant database system. In certain embodiments, for example, all custom entity data rows are stored in a single multi-tenant physical table, which may contain multiple logical tables per organization. It is transparent to customers that their multiple “tables” are in fact stored in one large table or that their data may be stored in the same table as the data of other customers.
As noted above, multiple tenants may share access to a common server 102 and database 130. And while data and services provided from server 102 to each tenant can be securely isolated from those provided to other tenants, in some cases it may be desirable to facilitate sharing of data between tenants. In such embodiments records may be shared by allowing “source” tenants to send selected records to selected other tenants. These “target” tenants have the ability to select which records they will accept. When such records are accepted copies of the accepted records are inserted into the target tenant's data. Furthermore, “connections” between these newly inserted records and the original source records are created.
These connections between the original record and the new record in the target tenant data allow updates in the source tenant's records to also be made in the receiving tenant's records. For example, when a source tenant updates a record containing product information with new specifications, those new specifications may be automatically updated in the target tenant's record as long as the connection between records remains. Likewise, new records, or deletions of records will likewise create changes in the receiving tenant's data. It should be noted that in a typical implementation millions of records may be shared between tenants. Furthermore, each record may be shared to and received by multiple different target tenants. With such large numbers of records being created, shared and updated, significant processing resources may be required for even relatively small record updates.
This is particularly true for updates of records that are shared with many different target tenants. However, other such record actions may also require significant processing resources depending largely upon the number of records and connections. In general, the processing of such record actions involves the creation and processing of “events” that define the type of parameters of the record action to be performed. Furthermore, these events have been commonly processed simply in the order created, without regard to the other related events.
To facilitate the performance these resource intensive record actions, the embodiments described here provide systems and methods that utilize a message queue and an event queue to selectively process record actions. The message queue is configured to store messages that correspond to records actions, and the event queue is configured to store corresponding events. So configured, record actions such as record updates are processed in bulk by selecting a plurality of messages in the message queue, identifying a plurality of events corresponding to the plurality of selected messages, and batch processing the plurality of events to execute the plurality of record actions. As will be described in greater detail below, this technique for processing record actions allows for large number of record updates to be performed without excessively interfering with other processing actions. For example, by batch processing record actions by source tenant and/or connection together in one batch, and then processing record actions for other tenants and/or other connections in later batches. As will be described in greater detail below, this selective batch processing of record actions based on tenant and connection can conserve computing resources and improve response times.
Turning now to
With respect to these embodiments, the multi-tenant database system may include any service that relies on a database system that is accessible over a network, in which various elements of hardware and software of the database system may be shared by one or more customers (e.g. tenants). For instance, a given application server may simultaneously process requests for a great number of tenants using different threads, and a given database may store data for a potentially much greater number of customers.
To this end, the message queue 302 and data queue 304 of the multi-tenant on-demand database system may include any data structure capable of queuing unprocessed messages to be processed by a server provided in the multi-tenant on-demand database system. In one embodiment, the message queue 302 may store the unprocessed messages in the order by which they have been added to the message queue 302.
As one example, the message queue 302 and event queue 304 may each include a table and the unprocessed messages and events may be stored in rows of the corresponding table.
Further, it should be noted that while the unprocessed messages in the message queue 302 are described as corresponding to events that generate specific record actions, such messages may also correspond to other tasks that are to be performed by one of the aforementioned multi-tenant database system. One specific example is in the preprocess messages used for record updates, which are described in more detail below. As other examples, the messages may include other types of actions with respect to creating, deleting, and modifying content, controlling access to contact, and other various actions that are performed in a multi-tenant database.
As noted above, the message queue 302 is configured to store messages that correspond to events, including events that initiate records actions relating to a plurality of records in a multi-tenant database. In accordance with the embodiments of the invention, these messages include record action messages and preprocess messages. When a user initiates a record action, a corresponding message with a message ID is inserted in the message queue 302.
Likewise, the event queue 304 is configured to store events that correspond to records actions relating to a plurality of records in a multi-tenant database. In accordance with the embodiments of the invention, these events may include a variety of record action events. As was noted above, when a user initiates a record action, a corresponding message is inserted in the message queue 302. Likewise, one or more corresponding events are inserted into the event queue 304. In one embodiment each message includes a message ID and each event in the event queue 304 includes an entry identifying a key of corresponding message in the message queue 302.
As noted above, multiple corresponding events may be inserted into the event queue 304 for each message in the message queue 302. This one-to-many relationship reduces the number of messages that are required to be in the message queue 302. This reduces the number of messages inserted into the message queue 302, and thus improves the performance of message queue 302. Specifically, reducing the number of messages in the message queue 302 prevents the queue 302 from flooding and decreases the time needed to query the queue 302.
It should be noted that the number of events that may be created for each message may be limited by processing constraints. In such an embodiment, each event in the event queue 304 that corresponds to the same message in the message queue 302 would have the same message ID. Thus, all of the events corresponding to the same message in the message queue 302 may be easily identified by using the message ID for each event.
The message queue picker 308 is configured to select messages in the message queue 302 for processing. In general, the message queue picker 308 is configured to bulk select messages that have common features that facilitate efficient batch processing. For example, the message queue picker 308 may be configured to select a plurality of messages from the message queue 302 that correspond to the same source tenant. In other embodiments, the message queue picker 308 may be configured to select a plurality of messages from the message queue 302 that correspond to the same source tenant and connection (i.e., to the same target tenant). In yet other embodiments, a combination may be used such that the message queue picker 308 selects those messages that correspond to both the same source tenant, same object/record type (i.e., for attachment records only) and the same connection to the target tenant.
The message queue handler 306 is called to process messages in the message queue 302. The message queue handler 306 is implemented with the ability to process multiple messages by de-queuing them as a batch using a bulk de-queue feature. As noted above, each message in the message queue 302 includes a message ID. The message queue handler 306 is configured to use the message ID to find corresponding events in the event queue 304. Additionally, the message ID identifies the corresponding source tenant and connection. The events corresponding to the batch of de-queued messages may then be processed by calling the appropriate function.
As described above, the message queue picker 308 is configured to select messages in the message queue 302 for processing. This generally involves a bulk selection of messages that have common features that lend themselves to batch processing.
In one embodiment the message queue picker 308 selects messages as follows. First, the message queue picker 308 selects the top or next message in the message queue 302. This would correspond to the oldest message in the queue when the message queue 302 is implemented with a first-in first-out data structure. Of course, in other embodiments the first message would not necessarily correspond to the oldest message in the queue. For example, in embodiments where certain messages are given priority over some older messages.
Then, with the first message selected the tenant corresponding to the first message is identified. As was noted above, in a typical embodiment the record action messages in the message queue 302 each identify a corresponding source tenant. Then, with the corresponding source tenant identified, the next step is to identify the connection corresponding to the first message. Again, in a typical embodiment the record action messages in the message queue 302 each identify a connection from the source record to a target tenant
With the source tenant and connection of the first message identified, the next step is to identify other messages in the message queue that have the same source tenant are connected to the same target tenant. All such messages having the same tenant and connection as the first message (but different message IDs) may then be bulk de-queued by the message queue handler 306. It should be noted that in some cases not all messages identified as having the same corresponding tenant and connection may be de-queued at once. For example, the bulk de-queue functionality may be limited in the number of messages that may be bulk de-queued, or the size of the de-queued messages may be similarly limited.
With the plurality of messages de-queued, the events in the event queue 304 corresponding to those messages selected messages may then be identified and processed as a batch. In general, this may be done by passing the events to method in the core processing system. Events that are not successfully processed by the system may be re-enqueued into the event queue 304 along with a corresponding message in the message queue 302 for further attempts. Such messages and events may be re-enqueued several times as needed.
Thus, the system 300 is able to identify, select and process those messages that correspond to the same target tenant and connection. Such a procedure minimizes the overhead that would otherwise be required to set up multiple API sessions for the same connection. Specifically, updating records in other tenants typically involves logging into the target tenant and setting up metal data before calling an appropriate function to initiate the action. Thus, the system 300 is able to improve performance by selecting and processing messages such that events corresponding to the same tenant and connection and processed together. This may improve performance in the multi-tenant database system by reducing the need to repeatedly log into and call functions for the same target tenant.
In one embodiment, the system 300 also includes a preprocess message handler 310 and a preprocess message picker 312. In general, the preprocess message handler 310 and preprocess message picker 312 serve to reduce the delay apparent to the user that may otherwise occur when the system 300 is used to update large numbers of connected records.
The preprocess message handler 310 and preprocess message picker 312 are preferably implemented to perform specific types of record actions that update multiple records. In an update action, the records have been previously created and the connection between shared records established. However, because of the sheer number of possible shared records and the resulting number of connections, update actions may be very resource intensive and take a noticeable amount of time to complete. To alleviate the resulting delay in returning control to the user, the system 300 uses the preprocess message handler 310 and preprocess message picker 312 to initiate update actions that follow when shared records are updated.
Specifically, when an update is initiated by a user a special preprocess message is created in the message queue. The preprocess message includes an ID that identifies the record being updated, but would typically not include a specific connection ID, as one preprocess message will typically correspond to multiple connections for the record.
The preprocess message picker 312 selects preprocess messages from the message queue 302 and passes the selected preprocess messages to the preprocess message handler 310 for execution.
Selected preprocess messages are executed by the preprocess message handler 310. Specifically, the preprocess message handler 310 creates messages and events corresponding to the record updates. Specifically, the preprocess message handler 310 identifies connections corresponding to the record update and creates a message in the message queue 302 and an event in the event queue 304 for each such record update action. It should be noted that with the creation of these messages and events that control may be returned to the user after a minimal delay. Thus, control may be returned to the user after the events are created, but before all the created events corresponding to the updates are actually executed. In most embodiments there can be a significant reduction in the amount of processing that must be performed before control is returned to the user.
The created messages and events may be then processed by the message queue handler 306 and the message queue picker 308 in the same way that typical record action messages and events executed. Thus, they will be selected by the message queue picker 308 and performed by the message queue handler 306 according to the corresponding source tenant and connection. The corresponding events are thus processed and the records updated. Again, this processing of the multiple events may be accomplished in the background after the returning of control to the user.
Thus, the system 300 is able to efficiently identify and process tasks in an event queue of a multi-tenant system. It should be noted one source of improvement is the ability to use multiple different servers, each with multiple different threads to process events in the event queues. This may greatly improve the performance of the system.
Generally speaking, the various functions and features of system 300 may be carried out with any sort of hardware, software and/or firmware logic that is stored and/or executed on any platform. Some or all of the functionality of system 300 may be carried out, for example, by logic executing within system 100 in
The term “exemplary” is used herein to represent one example, instance or illustration that may have any number of alternates. Any implementation described herein as “exemplary” should not necessarily be construed as preferred or advantageous over other implementations.
Although several exemplary embodiments have been presented in the foregoing description, it should be appreciated that a vast number of alternate but equivalent variations exist, and the examples presented herein are not intended to limit the scope, applicability, or configuration of the invention in any way. To the contrary, various changes may be made in the function and arrangement of the various features described herein without departing from the scope of the claims and their legal equivalents.
Number | Name | Date | Kind |
---|---|---|---|
5577188 | Zhu | Nov 1996 | A |
5608872 | Schwartz et al. | Mar 1997 | A |
5649104 | Carleton et al. | Jul 1997 | A |
5715450 | Ambrose et al. | Feb 1998 | A |
5761419 | Schwartz et al. | Jun 1998 | A |
5819038 | Carleton et al. | Oct 1998 | A |
5821937 | Tonelli et al. | Oct 1998 | A |
5831610 | Tonelli et al. | Nov 1998 | A |
5873096 | Lim et al. | Feb 1999 | A |
5918159 | Fomukong et al. | Jun 1999 | A |
5963953 | Cram et al. | Oct 1999 | A |
5983227 | Nazem et al. | Nov 1999 | A |
6092083 | Brodersen et al. | Jul 2000 | A |
6169534 | Raffel et al. | Jan 2001 | B1 |
6178425 | Brodersen et al. | Jan 2001 | B1 |
6189011 | Lim et al. | Feb 2001 | B1 |
6216133 | Masthoff | Apr 2001 | B1 |
6216135 | Brodersen et al. | Apr 2001 | B1 |
6233617 | Rothwein et al. | May 2001 | B1 |
6236978 | Tuzhilin | May 2001 | B1 |
6266669 | Brodersen et al. | Jul 2001 | B1 |
6288717 | Dunkle | Sep 2001 | B1 |
6295530 | Ritchie et al. | Sep 2001 | B1 |
6324568 | Diec et al. | Nov 2001 | B1 |
6324693 | Brodersen et al. | Nov 2001 | B1 |
6336137 | Lee et al. | Jan 2002 | B1 |
D454139 | Feldcamp et al. | Mar 2002 | S |
6367077 | Brodersen et al. | Apr 2002 | B1 |
6393605 | Loomans | May 2002 | B1 |
6405220 | Brodersen et al. | Jun 2002 | B1 |
6411949 | Schaffer | Jun 2002 | B1 |
6434550 | Warner et al. | Aug 2002 | B1 |
6446089 | Brodersen et al. | Sep 2002 | B1 |
6535909 | Rust | Mar 2003 | B1 |
6549908 | Loomans | Apr 2003 | B1 |
6553563 | Ambrose et al. | Apr 2003 | B2 |
6560461 | Fomukong et al. | May 2003 | B1 |
6574635 | Stauber et al. | Jun 2003 | B2 |
6577726 | Huang et al. | Jun 2003 | B1 |
6601087 | Zhu et al. | Jul 2003 | B1 |
6604117 | Lim et al. | Aug 2003 | B2 |
6604128 | Diec | Aug 2003 | B2 |
6609150 | Lee et al. | Aug 2003 | B2 |
6621834 | Scherpbier et al. | Sep 2003 | B1 |
6654032 | Zhu et al. | Nov 2003 | B1 |
6665648 | Brodersen et al. | Dec 2003 | B2 |
6665655 | Warner et al. | Dec 2003 | B1 |
6684438 | Brodersen et al. | Feb 2004 | B2 |
6711565 | Subramaniam et al. | Mar 2004 | B1 |
6724399 | Katchour et al. | Apr 2004 | B1 |
6728702 | Subramaniam et al. | Apr 2004 | B1 |
6728960 | Loomans et al. | Apr 2004 | B1 |
6732095 | Warshavsky et al. | May 2004 | B1 |
6732100 | Brodersen et al. | May 2004 | B1 |
6732111 | Brodersen et al. | May 2004 | B2 |
6754681 | Brodersen et al. | Jun 2004 | B2 |
6763351 | Subramaniam et al. | Jul 2004 | B1 |
6763501 | Zhu et al. | Jul 2004 | B1 |
6768904 | Kim | Jul 2004 | B2 |
6782383 | Subramaniam et al. | Aug 2004 | B2 |
6804330 | Jones et al. | Oct 2004 | B1 |
6826565 | Ritchie et al. | Nov 2004 | B2 |
6826582 | Chatterjee et al. | Nov 2004 | B1 |
6826745 | Coker et al. | Nov 2004 | B2 |
6829655 | Huang et al. | Dec 2004 | B1 |
6842748 | Warner et al. | Jan 2005 | B1 |
6850895 | Brodersen et al. | Feb 2005 | B2 |
6850949 | Warner et al. | Feb 2005 | B2 |
6907566 | McElfresh et al. | Jun 2005 | B1 |
7100111 | McElfresh et al. | Aug 2006 | B2 |
7269590 | Hull et al. | Sep 2007 | B2 |
7281022 | Gruhl et al. | Oct 2007 | B2 |
7340411 | Cook | Mar 2008 | B2 |
7373599 | McElfresh et al. | May 2008 | B2 |
7406501 | Szeto et al. | Jul 2008 | B2 |
7426557 | Gruhl et al. | Sep 2008 | B2 |
7454509 | Boulter et al. | Nov 2008 | B2 |
7599935 | La Rotonda et al. | Oct 2009 | B2 |
7603331 | Tuzhilin et al. | Oct 2009 | B2 |
7620655 | Larsson et al. | Nov 2009 | B2 |
7644122 | Weyer et al. | Jan 2010 | B2 |
7668861 | Steven | Feb 2010 | B2 |
7685091 | Boone et al. | Mar 2010 | B2 |
7698160 | Beaven et al. | Apr 2010 | B2 |
7747648 | Kraft et al. | Jun 2010 | B1 |
7827208 | Bosworth et al. | Nov 2010 | B2 |
7853881 | Aly Assal et al. | Dec 2010 | B1 |
7856445 | Gross | Dec 2010 | B2 |
7933843 | von Groll et al. | Apr 2011 | B1 |
7945653 | Zuckerberg et al. | May 2011 | B2 |
7970895 | Gruhl et al. | Jun 2011 | B2 |
7974983 | Goeldi | Jul 2011 | B2 |
7991764 | Rathod | Aug 2011 | B2 |
8005896 | Cheah | Aug 2011 | B2 |
8010539 | Blair-Goldensohn et al. | Aug 2011 | B2 |
8073850 | Hubbard et al. | Dec 2011 | B1 |
8082301 | Ahlgren et al. | Dec 2011 | B2 |
8095413 | Beaven | Jan 2012 | B1 |
8095594 | Beaven et al. | Jan 2012 | B2 |
8103611 | Tuzhilin et al. | Jan 2012 | B2 |
8150913 | Cheah | Apr 2012 | B2 |
8166013 | Bandaru et al. | Apr 2012 | B2 |
8275836 | Beaven et al. | Sep 2012 | B2 |
20010044791 | Richter et al. | Nov 2001 | A1 |
20020072951 | Lee et al. | Jun 2002 | A1 |
20020082892 | Raffel | Jun 2002 | A1 |
20020129352 | Brodersen et al. | Sep 2002 | A1 |
20020140731 | Subramaniam et al. | Oct 2002 | A1 |
20020143997 | Huang et al. | Oct 2002 | A1 |
20020162090 | Parnell et al. | Oct 2002 | A1 |
20020165742 | Robins | Nov 2002 | A1 |
20030004971 | Gong | Jan 2003 | A1 |
20030018705 | Chen et al. | Jan 2003 | A1 |
20030018830 | Chen et al. | Jan 2003 | A1 |
20030066031 | Laane et al. | Apr 2003 | A1 |
20030066032 | Ramachandran et al. | Apr 2003 | A1 |
20030069936 | Warner et al. | Apr 2003 | A1 |
20030070000 | Coker et al. | Apr 2003 | A1 |
20030070004 | Mukundan et al. | Apr 2003 | A1 |
20030070005 | Mukundan et al. | Apr 2003 | A1 |
20030074418 | Coker et al. | Apr 2003 | A1 |
20030120675 | Stauber et al. | Jun 2003 | A1 |
20030151633 | George et al. | Aug 2003 | A1 |
20030159136 | Huang et al. | Aug 2003 | A1 |
20030187921 | Diec et al. | Oct 2003 | A1 |
20030189600 | Gune et al. | Oct 2003 | A1 |
20030204427 | Gune et al. | Oct 2003 | A1 |
20030206192 | Chen et al. | Nov 2003 | A1 |
20030225730 | Warner et al. | Dec 2003 | A1 |
20040001092 | Rothwein et al. | Jan 2004 | A1 |
20040010489 | Rio et al. | Jan 2004 | A1 |
20040015981 | Coker et al. | Jan 2004 | A1 |
20040027388 | Berg et al. | Feb 2004 | A1 |
20040128001 | Levin et al. | Jul 2004 | A1 |
20040186860 | Lee et al. | Sep 2004 | A1 |
20040193510 | Catahan et al. | Sep 2004 | A1 |
20040199489 | Barnes-Leon et al. | Oct 2004 | A1 |
20040199536 | Barnes-Leon et al. | Oct 2004 | A1 |
20040199543 | Braud et al. | Oct 2004 | A1 |
20040249854 | Barnes-Leon et al. | Dec 2004 | A1 |
20040260534 | Pak et al. | Dec 2004 | A1 |
20040260659 | Chan et al. | Dec 2004 | A1 |
20040268299 | Lei et al. | Dec 2004 | A1 |
20050050555 | Exley et al. | Mar 2005 | A1 |
20050091098 | Brodersen et al. | Apr 2005 | A1 |
20050256905 | Gruhl et al. | Nov 2005 | A1 |
20050256949 | Gruhl et al. | Nov 2005 | A1 |
20060168013 | Wilson et al. | Jul 2006 | A1 |
20060200444 | Bracho et al. | Sep 2006 | A1 |
20060284873 | Forrest et al. | Dec 2006 | A1 |
20070214097 | Parsons et al. | Sep 2007 | A1 |
20070233656 | Bunescu et al. | Oct 2007 | A1 |
20070239452 | Madhavan et al. | Oct 2007 | A1 |
20080010270 | Gross | Jan 2008 | A1 |
20080082399 | Noble et al. | Apr 2008 | A1 |
20080082586 | Jasik et al. | Apr 2008 | A1 |
20080147487 | Hirshberg | Jun 2008 | A1 |
20080154798 | Valz | Jun 2008 | A1 |
20080209320 | Mawhinney et al. | Aug 2008 | A1 |
20080228695 | Sifry et al. | Sep 2008 | A1 |
20080288516 | Hadfield | Nov 2008 | A1 |
20090018996 | Hunt et al. | Jan 2009 | A1 |
20090138427 | Kalavade | May 2009 | A1 |
20090204507 | Cockayne et al. | Aug 2009 | A1 |
20090222743 | Hadfield | Sep 2009 | A1 |
20090319518 | Koudas et al. | Dec 2009 | A1 |
20100063877 | Soroca et al. | Mar 2010 | A1 |
20100063959 | Doshi et al. | Mar 2010 | A1 |
20100070485 | Parsons et al. | Mar 2010 | A1 |
20100088234 | Moore et al. | Apr 2010 | A1 |
20100094878 | Soroca et al. | Apr 2010 | A1 |
20100125502 | Solomon et al. | May 2010 | A1 |
20100217757 | Fujioka | Aug 2010 | A1 |
20100241576 | Beeri | Sep 2010 | A1 |
20100274815 | Vanasco | Oct 2010 | A1 |
20110041153 | Simon et al. | Feb 2011 | A1 |
20110055194 | Ghosh | Mar 2011 | A1 |
20110209067 | Bogess et al. | Aug 2011 | A1 |
20110246816 | Hsieh et al. | Oct 2011 | A1 |
20110258204 | Hubbard et al. | Oct 2011 | A1 |
20120296920 | Sahni et al. | Nov 2012 | A1 |
Entry |
---|
Linear Combination—Wikipedia (http://en.wikipedia.org/wiki/Linear—Combination (accessed Jul. 16, 2012). |
Final Office Action for U.S. Appl. No. 12/437,418, mailed Apr. 16, 2012. |
Final Office Action for U.S. Appl. No. 12/356,429, mailed May 2, 2012. |
Notice of Allowance for U.S. Appl. No. 12/819,402, mailed May 14, 2012. |
Non-Final Office Action for U.S. Appl. No. 13/453,802, mailed Jun. 8, 2012. |
U.S. Appl. No. 13/478,002, filed May 22, 2012. |
United States Patent Office, US Non-Final Office Action for U.S. Appl. No. 12/278,277 dated Sep. 14, 2012. |
USPTO, Notice of Allowance, for U.S. Appl. No. 13/526,246, mailed Feb. 14, 2013. |
USPTO, Notice of Allowance, for U.S. Appl. No. 12/356,429, mailed Feb. 15, 2013. |
Number | Date | Country | |
---|---|---|---|
20120254221 A1 | Oct 2012 | US |