The present disclosure relates to digital data processing, and more particularly, to methods, apparatus, and systems for processing events. The teachings herein have application, by way of non-limiting example, to rapid processing of events comprising large volumes of data between producer and consumer nodes in a cluster.
External systems generate growing numbers of events, such as social media posts and notifications. Furthermore, legacy systems can generate large volumes of events for processing by applications. Customers are increasingly looking to build or leverage enterprise applications that are capable of receiving and processing high volumes of events in short periods of time from these external or legacy systems. These systems can be used in a variety of applications ranging from health care to automotive repair. The applications can facilitate a range of business operations, from marketing and manufacturing to distribution to technical support. For example, an application can implement data-processing workflows to support processing transactional data ranging from customer service requests received by retail and banking enterprises, to routing and resolution of health care claims by insurance enterprises.
Computers can be networked into a cluster that includes two or more nodes capable of exchanging events or messages. The cluster can include one or more producer nodes and one or more consumer nodes. The producer nodes can transmit, or publish, events and the consumer nodes can receive, or subscribe to, events. Traditional techniques for exchanging and processing messages include a publish-subscribe model. In a publish-subscribe model, publishers or producers do not transmit messages direct to an individual subscriber or consumer. Instead, publishers transmit a message broadly to multiple subscribers at once. The subscribers generically register interest in certain messages and thereby receive only messages of interest.
However, the publish-subscribe model suffers from reduced throughput because a system uses valuable capacity in the network to provide a given message to multiple consumer nodes, in case more than one consumer node registers interest in the message. Furthermore, the publish-subscribe architecture generally uses an intermediate middleware node sometimes referred to as a message broker or message bus for message delivery, using additional resources in the cluster. A still further downside of traditional message processing systems is consumer nodes generally receive messages in any order, instead of preserving the order in which the producer nodes read the messages from a message stream for faster message routing and processing.
An object of this invention is to provide improved systems and methods for digital data processing. A more particular object is to provide improved systems and methods for event processing.
A further object is to provide such improved systems and methods as facilitate deployment of enterprise applications in a networked cluster of producer nodes and consumer nodes.
Yet a still further object is to provide such improved systems and methods as better utilize computing and networking resources upon addition or removal of producer nodes and consumer nodes to or from the cluster.
The foregoing are among the objects attained by the invention which provides, in one aspect, a digital data processing system and method for event processing.
In some embodiments, the digital data processing system includes a producer node in communicative coupling with one or more consumer nodes and with a sharding map, for example over a network. The producer node can be configured to receive at least one event stream including a plurality of events, and determine a sharding key associated with an event among the plurality of events in the event stream. The producer node can be further configured to identify, based on the sharding map, a producer event buffer associated with a producer channel on the producer node for transmitting the event to a corresponding consumer event buffer associated with a consumer channel on a consumer node among the one or more consumer nodes. The sharding map can correlate the sharding key for the event with the producer channel. The producer node can be further configured to provide the event to the producer event buffer associated with the producer channel in order to transmit the event to the corresponding consumer event buffer associated with the consumer channel on the consumer node.
According to related aspects of the invention, the producer node can be further configured to initialize a plurality of producer channels. The producer node can reference a channel map that correlates an event in the event stream with one or more consumer channels on the one or more consumer nodes. The producer node can create at least one producer channel based on the channel map. The producer channel can be communicatively coupled with a corresponding consumer channel among the one or more consumer channels. The producer node can update the sharding map to correlate the sharding key for the event with the at least one producer channel created based on the channel map.
In related aspects of the invention, the sharding map can correlate the sharding key for the event with the producer channel based on a partition space. The partition space can be determined using a partition criterion based on a count of consumer nodes available to process the plurality of events in the event stream.
In further related aspects of the invention, the producer node can be configured to update the sharding map in response to detecting an update to the channel map. For example, the producer node can be configured to update the sharding map by redistributing the partition space. The redistribution of the partition space can be based on determining an updated partition size for the existing producer channels based on a count of available consumer channels, and assigning an overflow portion of the partition space based on the updated partition size to a new producer channel. Alternatively, the redistribution of the partition space can be based on querying the one or more consumer nodes to identify resources available to the one or more consumer nodes, and weighting the partitions assigned to each producer channel based on the identified resources.
Related aspects of the invention provide a producer node that can be further configured to adjust the rate of providing events to the producer event buffer, so as to manipulate the rate of events processed on the corresponding consumer node by the consumer event buffer, in order to improve throughput between the producer node and the consumer node.
In related aspects of the invention, the producer node can be further configured to bundle event transport metadata with the plurality of events in the event stream for transmission to the consumer node, and trigger the consumer node to execute a rule identified by the event transport metadata for processing the plurality of events.
According to related aspects of the invention, the producer node can be further configured to update a plurality of producer channels in response to detecting an update to the sharding map or a channel map by the consumer node. The update to the sharding map or the channel map can include adding a consumer node to the system, removing a consumer node from the system, adding one or more consumer channels to the consumer node, or removing one or more consumer channels from the consumer node.
In further related aspects of the invention, the producer node can be further configured to transmit a message to the one or more consumer nodes in response to detecting an the update to the sharding map or the channel map. The transmitted message can trigger the one or more consumer nodes to determine a delta that tracks the update to the sharding map or the channel map, identify data to be moved to a different consumer channel or a different consumer node, and copy the data to be moved to a cluster change data map. The data copied to the cluster change data map can trigger the consumer node to copy the moved data from the cluster change data map, clear the copied data from the cluster change data map, and update a status of the consumer node to active in a consumer map that tracks the one or more consumer nodes in the cluster. The producer node can be further configured to resume transmission of data to the consumer node in response to detecting the updated status of the consumer node in the consumer map.
The foregoing and other aspects of the invention are evident in the text that follows and in the drawings.
Various objects, features, and advantages of the present disclosure can be more fully appreciated with reference to the following detailed description when considered in connection with the following drawings, in which like reference numerals identify like elements. The following drawings are for the purpose of illustration only and are not intended to be limiting of the invention, the scope of which is set forth in the claims that follow.
The present systems and methods allow for rapid processing of large volumes of discretely generated data referred to herein as events. In some embodiments, the event processing systems and methods can be general enough to be used as a generic message passing service in a cluster of one or more producer nodes and consumer nodes. (The event processing system and related methods, as discussed here and elsewhere herein, are referred to elsewhere herein simply as “the event processing.”) The event processing can exhibit and support both horizontal as well as vertical scaling to achieve speed and robustness in current and future client installations.
The event processing systems and methods involve receiving an event stream having one or more events. A producer node in a cluster determines a sharding key for a received event (i.e., a key for grouping events for routing). The producer node uses a sharding map to correlate the sharding key for the event with a producer channel. The producer node provides the event to a producer event buffer associated with the producer channel. The producer event buffer transmits the received event directly to a corresponding consumer event buffer associated with a consumer channel on a consumer node. The event processing leverages a paired relationship established between one or more producer channels on the producer node, and one or more consumer channels on the consumer node to generate enhanced throughput compared with traditional messaging systems.
The event processing systems and methods have utility in, for example, rapidly processing discussions on social media. Today, many people use and communicate ideas through social media. The topics and ideas shared can vary greatly, however topics specific to an industry or business are potentially germane to an event processing system of the type referred to above (and described further herein) employed by a given enterprise, say, in that industry or business. As users of social media publish discussions, these discussions can form the basis for grass roots marketing and decision-based campaigns, e.g., by that enterprise. In some embodiments, such event processing allows the enterprise to tap into these theme-based discussions to “mine” information important to their industry or specific business. The event processing provides an ability to tap into this vast source of data and distill the data into a meaningful form. To achieve this goal, some embodiments of the event processing system include a producer node that is able to connect to a “fire-hose” of events, or data, streaming from event services. Non-limiting example event services include microblogs such as Twitter, personal social networks such as Facebook, and professional social networks such as LinkedIn. In some embodiments, the event processing systems and methods provide a mechanism to identify events declaratively in which the enterprise is interested, once an administrator configures and establishes a connection. Further embodiments of the event processing allow the enterprise to define declarative rules that describe how to process those events of interest.
The event processing systems and methods described herein also have utility in rapidly processing events from traditional data warehouses or so-called “Big Data” systems that generate large volumes of data in short periods of time. A non-limiting example includes processing insurance claim data using declarative rules. In this case a large volume of insurance claim data can flow into the event processing system. The event processing routes the events from event services to producer nodes and consumer nodes for processing, transformation, and potentially storage.
System Architecture
Non-limiting examples of event services 102a-c include microblog 102a, personal social network 102b, and professional social network 102c, all of the type known in the art as adapted in accord with the teachings hereof. In this regard, event services 102a-c can be implemented, for example, on one or more digital data processing systems in the conventional manner known in the art, again, as adapted in accord with the teachings hereof. Event services 102a-c provide an event stream to producer nodes 104a-b. In some embodiments, the event stream can include events aggregated from multiple event sources.
Producer nodes 104a-b communicate with event services 102a-c over network 108. Producer node 104a includes event processor 110a, and producer node 104b includes event processor 110b. Event processors 110a-b are configured to receive events from one or more of event services 102a-c, process, and route those events to one or more of consumer nodes 106a-c. In some embodiments, producer nodes 104a-b can participate in and communicate with the overall cluster of illustrated nodes in support of client interactions, while the producer nodes themselves remain isolated and hidden from end users who are using consumer nodes 106a-c. That is, event processing system 100 allows administrators to configure a system in which the end users are only aware that system 100 receives events from event services 102a-c and routes the received events to consumer nodes 106a-c, without requiring knowledge of architectural or implementation details of—or even the existence of—producer nodes 104a-b.
Consumer nodes 106a-c receive the routed events from producer nodes 102a-b over network 108. Consumer node 106a includes event processor 112a, and consumer node 106b includes event processor 112b. Event processors 112a-b are configured to receive events from producer nodes 104a-b and process the events. Consumer nodes 106a-c participate and communicate with the overall cluster of illustrated nodes supporting client interactions and can be allocated for connection to and communication with end users. In some embodiments, consumer nodes 106a-c process received events in accordance with event transport metadata that is bundled with the event. For example, the event transport metadata can trigger consumer nodes 106a-c to process received events by performing sentiment analysis on the received events, grouping or aggregating the received events for further processing, displaying user interface updates about the received events, or storing the received events in a database.
In some embodiments, producer nodes 104a-b can include one or more producer digital data processors, and consumer nodes 106a-c can include one or more consumer digital data processors. The producer and consumer digital data processors can be digital processors of the type commercially available in the marketplace suitable for operation in event processing system 100 and adapted in accord with the teachings hereof, for example, utilizing rules forming applications executing in one or more rules engines, e.g. as discussed elsewhere herein. Though producer nodes 104a-b and consumer nodes 106a-c can be typically implemented in server-class computers such as a minicomputer, producer nodes 104a-b and consumer nodes 106a-c may also be implemented in desktop computers, workstations, laptop computers, tablet computers, personal digital assistants (PDAs) or other suitable apparatus adapted based on the systems and methods described herein. The producer digital data processor and consumer digital data processor include central processing, memory, storage using a non-transitory computer-readable medium (e.g., a magnetic disk, solid state drive, or other storage medium), and input/output units and other constituent components (not shown) of the type conventional in the art that are programmed or otherwise configured in accord with the teachings hereof.
Network 108 can include one or more networks of the type commercially available in the marketplace or otherwise suitable for supporting communication between event services 102a-c, producer nodes 104a-b, and consumer nodes 106a-c in accord with the teachings hereof. Network 108 can be wired or wireless, a cellular network, a Local Area Network (LAN), a Wireless LAN (WLAN), a Metropolitan Area Network (MAN), a Wireless MAN (WMAN), a Wide Area Network (WAN), a Wireless WAN (WWAN), a Personal Area Network (PAN), a Wireless PAN (WPAN), or a network operating in accordance with existing IEEE 802.11, 802.11a, 802.11b, 802.11g, 802.11n, 802.16, 802.16d, 802.16e, 802.16m standards or future versions or derivatives of the above standards.
An event service 102a-c publishes an incoming event into the cluster, and event processing system 114 ultimately designates a consumer node 106a-c in the cluster to handle or process the event (e.g., using one of producer nodes 104a-b). Some embodiments of event processing system 114 can support additional features. For example, producer nodes 104a-b and consumer nodes 106a-c can process events in the order in which event services 102a-c produce the events. An event key associated with events can identify that an event belongs to a particular group of related events. Lastly, event processing system 114 can allow an administrator to configure the system so that the same consumer node 106a-c processes events that have the same event key (e.g., via one or more of event processors 110a-b on producer nodes 104a-b).
In some embodiments, event processors 110a-b on producer nodes 104a-b process events using embedded Hazelcast logic (hereinafter, simply, “Hazelcast”) 124a-b, and event processors 112a-c process events using embedded Hazelcast 126a-c. As those skilled in the art will appreciate, Hazelcast refers to a publicly available third-party architecture and processing mechanism for in-memory object caching. Event processing system 114 can leverage Hazelcast to enable distributed object caching and locking, and node-to-node messaging. A producer node or consumer node using Hazelcast can become what the Hazelcast architecture refers to as a Hazelcast node. Hazelcast allows automatic discovery of other Hazelcast nodes using discovery mechanisms such as multicast or direct Internet Protocol (IP) discovery.
Producer nodes 104a-b use interfaces 120a-b to receive events from event services 102a-c over network 134. In some embodiments, interfaces 120a-b use embedded Hazelcast 124a-b to connect producer nodes 104a-b with event services 102a-c and consumer nodes 106a-c, for example using a discovery mechanism based on TCP/IP.
In some embodiments, producer nodes 104a-b and consumer nodes 106a-c process events using channel map 116 and sharding map 118. Channel map 116 correlates an event in the event stream from event services 102a-c with consumer nodes 106a-c. Specifically, channel map 116 identifies consumer channels on consumer nodes 106a-c to which to route and process events. Sharding map 118 correlates a sharding key for an event (e.g., a key for grouping or clustering events for routing) with a producer channel on producer nodes 104a-b to which to route and process events. As a consequence of the foregoing and the actions below, event processing system 114 creates a paired relationship between producer channels on producer nodes 104a-b and consumer channels on consumer nodes 106a-c to enable rapid processing of large volumes of events from event services 102a-c. In some embodiments, channel map 116 and sharding map 118 may be distributed in the cluster. For example, further embodiments of channel map 116 and sharding map 118 may be implemented in event processing system 114 as Hazelcast distributed maps.
In some embodiments, producer node 104a uses embedded Hazelcast 124a to read and write entries from channel map 116 and sharding map 118 over interface 120a, and producer node 104b uses embedded Hazelcast 124b to read and write from channel map 116 and sharding map 118 over interface 120b. Similarly, consumer nodes 106a-c use embedded Hazelcast 126a-c to read and write from channel map 116 and sharding map 118 over interfaces 122a-c.
In some embodiments, an enterprise can deploy event processing system 114 in support of enterprise applications executing locally on or remote to the cluster of illustrated nodes. Such enterprise applications can include specialized software or hardware used within a specific industry or business function (e.g., human resources, finance, healthcare, telecommunications, insurance, etc.). Alternatively, the enterprise applications can include cross-industry applications (e.g., project management), or other types of software or hardware applications.
In some embodiments, rules may define the enterprise applications. Producer nodes 104a-b and consumer nodes 106a-c can be in communication with rules engine 128. Rules engine 128 can be in communication with rules base 130 and transactional database 132. As the application executes on a producer digital data processor (e.g., producer nodes 104a-b) or a consumer digital data processor (e.g., consumer nodes 106a-c), event processing system 114 may retrieve any portion of the rules that define the application from rules base 130 and process or execute the rules in response to requests or events signaled to or detected by the producer digital data processors or consumer digital data processors at run-time, (e.g., using rules engine 128).
Rules base 130 can include a rules base of the type known in the art (albeit configured in accord with the teachings hereof) for storing rules (e.g., scripts, logic, controls, instructions, metadata, etc.) and other application-related information in tables, database records, database objects, and so forth. Preferred rules and rules bases can be of the type described in U.S. Pat. No. 5,826,250, entitled “Rules Bases and Methods of Access Thereof” and U.S. Pat. No. 7,640,222, entitled “Rules Base Systems and Methods with Circumstance Translation,” the entire contents of both of which are incorporated by reference herein in their entirety. In other embodiments, rules and rules bases that are architected or operated differently may be used as well.
Some embodiments may utilize multiple rules bases. For example, rules base 130 may be an enterprise-wide rules base in communication with rules engine 128, and domain-specific rules bases may be accessible to producer nodes 104a-b or consumer nodes 106a-c via network 134. If multiple rules bases are provided in a given embodiment, the rules bases may be of like architecture and operation or may differ in architecture and operation as well.
In some embodiments, rules may comprise meta-information structures. For example, the rules can include data elements or method elements. The method elements can be procedural or declarative. For example, method elements in a rule may be procedural insofar as the rule comprises one or more of a series of ordered steps. Declarative elements in a rule may set forth (i.e., declare) a relation between variables or values (e.g., a loan rate calculation or a decision-making criterion). Alternatively, declarative elements may declare a desired computation or result without specifying how the computations should be performed or how the result should be achieved. In one non-limiting example, a declarative portion of a rule may declare a desired result of retrieving a specified value without specifying a data source for the value or a particular query language for such retrieval (e.g., SQL, CQL, .QL, etc.). In other cases, the declarative portion of a meta-information structure may comprise declarative programming language statements (e.g., SQL). Still other types of declarative meta-information structures are possible.
While some rules may comprise meta-information structures that are wholly procedural and other rules may comprise meta-information structures that are wholly declarative, event processing system 114 can also include rules that comprise both procedural and declarative meta-information structures. That is, such rules can have meta-information structure portions that are declarative, as well as meta-information structure portions that are procedural. Furthermore, rules of the illustrated embodiments that comprise meta-information structures may also reference or incorporate other rules. Those other rules may themselves in turn reference or incorporate still other rules. As a result, editing such a rule may affect one or more rules that incorporate it (if any).
An advantage of rules that comprise meta-information structures over conventional rules is that meta-information structures provide administrators with flexibility to apply code-based or model-driven techniques in development and modification of applications or computing platforms. Particularly, like models in a model-driven environment, meta-information structures comprise data elements that can be used to define aspects of a complex system at a higher level of abstraction than source code written in programming languages such as Java or C++. On the other hand, administrators may also embed programming language statements into meta-information structures if the administrators deem that to be the most efficient design for the system being developed or modified. At run-time, rules engine 128 can convert the data elements of the meta-information structures along with programming language statements (if any) automatically into executable code for the application.
Thus, in some embodiments rules may be the primary artifacts that get created, stored (e.g., in rules base 130) or otherwise manipulated to define or modify the overall functionality of rules-based applications. The applications may automate or manage various types of work in different business domains at run-time. By way of non-limiting example, rules stored in rules base 130 may be configured to define aspects of an application. For example, rules can define the user interface, decision logic, integration framework, process definition, data model, reports, or security settings of a given application.
Transactional database 132 can include databases of the type known in the art (albeit configured in accord with the teachings hereof) for storing corporate, personal, governmental, or other data. Rules such as in rules base 130 may generate, update, transform, delete, store, or retrieve the data (herein collectively referred to as “processing” the data). Example data may include financial data; customer records; personal data; design-time, development-time, or runtime data related to an application; or other types of data. Transactional database 132 may store the data in tables, database records, or database objects, for example.
Transactional database 132 may be present in any given embodiment. Conversely, some embodiments may use multiple transactional databases, e.g., an enterprise-wide database on producer nodes 104a-b and branch-office specific databases on consumer nodes 106a-c, by way of non-limiting example. If multiple transactional databases are provided in a given embodiment, the transactional databases may be of like architecture and operation; though, they may have differing architecture or operation, as well.
Rules engine 128 can be of the type conventionally known in the art (albeit configured in accord with the teachings hereof) for use in processing or executing rules from rules base 130 to process data in (or for storage to) transactional database 132, e.g. in connection with events signaled to or detected by rules engine 128. Preferred such rules engines are of the type described in U.S. Pat. No. 5,826,250, entitled “Rules Bases and Methods of Access Thereof,” U.S. Pat. No. 7,640,222, entitled “Rules Base Systems and Methods with Circumstance Translation,” and U.S. Pat. No. 8,250,525, entitled “Proactive Performance Management For Multi-User Enterprise Software Systems,” all of which are incorporated by reference in their entirety herein. Rules engine 128 may be implemented in a single software program, multiple software programs or modules, or a combination of software modules or programs. Rules engine 128 may comprise programming instructions, scripts, or rules (e.g., rules stored in rules base 130) or a combination thereof.
Some embodiments of rules engine 128 may execute on or over multiple digital data processors. For example, event processing system 114 may invoke rules engine 128 for execution on a single digital data processor (e.g., a digital data processor on a producer node 104a-b or a consumer node 106a-c). Subsequently, event processing system 114 may apportion, distribute, or execute portions of rules engine 128 (or, potentially, the entirety of rules engine 128) over multiple digital data processors.
Other ways of implementing or executing rules engine 128 are also possible. By way of non-limiting example, rules engine 128 may have additional distinct components or portions that can be apportioned and distributed separately. Non-limiting example components can include a data access component for processing data during rule execution, a session management component for keeping track of activity across sessions of interaction with a digital data processor, or a performance monitoring component for monitoring and interacting with various system resources or event logs to manage performance thresholds.
Finally, network 134 can include one or more networks for supporting communication between event services 102a-c, producer nodes 104a-b, consumer nodes 106a-c, and rules engine 128. Network 134 can be wired or wireless, a cellular network, a Local Area Network (LAN), a Wireless LAN (WLAN), a Metropolitan Area Network (MAN), a Wireless MAN (WMAN), a Wide Area Network (WAN), a Wireless WAN (WWAN), a Personal Area Network (PAN), a Wireless PAN (WPAN), or a network operating in accordance with existing IEEE 802.11, 802.11a, 802.11b, 802.11g, 802.11n, 802.16, 802.16d, 802.16e, 802.16m standards or future versions or derivatives of the above standards.
The progression begins with event services 102a-c serving an event stream. For example, the event stream may be a series of social media microblog posts, or tweets, from microblog service Twitter. Each individual social media post can be considered an event to the event processing system. In some embodiments, an event can contain data and metadata. Non-limiting example data includes the contents of the event. An example tweet may include contents such as “I got great service at this restaurant today!” Events may also be collected into groups of related events. For example, a series of social media posts can describe a progression of events at a party. Events can also include metadata. As used herein, the term metadata refers to attributes that the event processing considers and processes separately from the underlying data in the event. Non-limiting example metadata includes the author of the social media post, a date of the social media post, an event stream type, and an event stream name.
In some embodiments, the event stream type can describe a source of the event stream. Non-limiting example event stream types can include social media services, such as Twitter, Facebook, or LinkedIn. In some embodiments, the event stream name can describe a specific instance of a stream type. The event processing supports processing multiple event streams in the cluster. The event stream name allows the event processing to address named event stream instances separately. Non-limiting example event stream names can include Finance Department, Human Resources Department, Department A, or Department B.
Producer nodes 104a-b receive the event stream from event sources 102a-c. Producer nodes 104a-b proceed to shard, or partition, the event stream into individual events for routing to consumer nodes 106a-c using the mechanisms discussed in still further detail below. In addition to the existing data and metadata associated with the events (e.g., contents, author, stream type, stream name), producer nodes 104a-b can also generate metadata. For example, producer nodes 104a-b can generate and bundle event transport metadata with received events. The event transport metadata can identify rules for the consumer node to execute upon receiving the sharded event, for example in conjunction with a rules engine. Using these identified rules, the event transport metadata can instruct the consumer node how to process the received event.
In some embodiments, in processing and routing the received event, the producer node determines various identifiers (e.g., using an event processor). For example, the producer node determines a producer channel for routing the event to a given consumer node, using a sharding key associated with the event. The sharding key groups or clusters events for routing from the producer node to the consumer node. In some embodiments, the producer node determines the sharding key based on the contents of the event. The event processing system creates a paired relationship between the producer channel and a corresponding consumer channel on a consumer node. In some embodiments, the producer node can modulate the rate of providing events to the producer channel, as a manner of flow control to improve throughput to the consumer node.
Consumer nodes 106a-c receive and process the transmitted event from producer nodes 104a-b. The data and metadata associated with the received event remain the same between consumer nodes 106a-c and producer nodes 104a-b. In some embodiments, consumer nodes 106a-c retrieve the bundled event transport metadata from the event. The event transport metadata identifies for consumer nodes 106a-c how to process the received event. In further embodiments, the event transport metadata can identify rules that are accessible to consumer nodes 106a-c. Non-limiting example rules can include rules to perform sentiment analysis on the received events, group or aggregate the received events for further processing, display user interface elements associated with or about the received events, or store the received events in a database. Consumer nodes 106a-c then execute the rules to process the received event, potentially in conjunction with a rules engine.
The producer node then determines a sharding key for each individual event (step 294). In some embodiments, the producer node determines the sharding key as follows. An administrator initially configures a partition space for a given event stream. As used herein, the term partition space refers to a universe or number of chunks or partitions into which a given event stream will be divided. For example, an event stream of stream type Twitter and stream name Department A might have a partition space of 100. Another event stream of stream type Facebook and stream name Department B might have a partition space of 200. Each event in the event stream can have an associated event key. In some embodiments, the producer node can determine the event key based on the contents of the event. For example, the event key can be a hash code or fingerprint of the contents of the event. The producer node can determine the sharding key by computing a modulo of the event key for an individual event with the partition space for a given event stream. As a result, the sharding key will be a value between 0 and the partition space minus one. For example, for a partition space of 100, the computation of the sharding key results in a value from 0-99. For a partition space of 200, the computation results in a value from 0-199.
The producer node identifies a producer event buffer associated with a producer channel using the sharding key and sharding map (step 296). The sharding map correlates a sharding key with a producer channel. In some embodiments, prior to receiving an event the producer node initializes and creates these producer channels using a list of consumer channels provided by the channel map. The producer node initializes a producer event buffer for each producer channel. The producer node also creates a paired relationship between the producer channel on the producer node and a corresponding consumer channel on a consumer node. The consumer channel also has its own associated consumer event buffer.
The producer node then provides the event to the producer event buffer associated with the producer channel identified by the sharding map (step 298). The producer node has previously configured a paired relationship between the producer channel identified by the sharding map and a corresponding consumer channel on a desired consumer node. Accordingly, providing the event to the producer event buffer allows the producer node to transmit the received event to the corresponding consumer event buffer associated with the consumer channel on the consumer node. In some embodiments, the producer node can modulate the rate of providing events to the producer channel, so as to improve throughput between the producer node and consumer node using dynamic flow control.
Some embodiments of the event processing system associate each event stream from the event services with a given number of channels. Channels represent resources available on producer nodes and consumer nodes to receive, transmit, and process events. For example, a first event stream for stream type Twitter and stream name “Department A” may have three channels available to process the event stream. In some embodiments the event processing system implements the three channels as three producer channels on one or more producer nodes, in a paired relationship with three consumer channels on one or more consumer nodes. Similarly, a second event stream for stream type Twitter and stream name “Department B” may have four channels available which the event processing system may implement as four producer channels on one or more producer nodes, in a paired relationship with four consumer channels on one or more consumer nodes. Channel map 116 illustrates this example arrangement of three and four consumer channels 302a-c, 304a-d.
In some embodiments, channel map 116 tracks the following information per consumer channel:
Channel map 116 allows the event processing to process multiple event services, such as multiple Twitter streams for different named instances of event services (e.g., different stream type and stream name). Channel map 116 tracks a first set of entries for an event stream having stream type Twitter and stream name Department A, and a second set of entries for stream type Twitter and stream name Department B. Channel map 116 indicates that three consumer channels are available to process the first event stream with stream type Twitter and stream name Department A. Consumer channels 0-2 (302a-c) show a status of Active. Furthermore, consumer channels 0-2 (302a-c) are allocated across two different consumer nodes. The first consumer node has node ID U1234567 and has two consumer channels 0-1 (302a-b) active for processing events from the event stream with stream type Twitter and stream name Department A. The second consumer node has node ID 055555555 and has one consumer channel 2 (302c) active for processing events from the event stream for stream type Twitter and stream name Department A.
Channel map 116 also indicates that the first consumer node with node ID U1234567 has four consumer channels in total. The first two consumer channels 0-1 (302a-b) are active for processing events from the first event stream, for stream type Twitter and stream name Department A. The remaining two consumer channels 3-4 (304a-b) are active for processing events from the second event stream, for stream type Twitter and stream name Department B. Similarly, channel map 116 indicates that the second consumer node with node ID 055555555 has three consumer channels in total. The first consumer channel 2 (302c) is active for processing events from the first event stream, for stream type Twitter and stream name Department A. The remaining two consumer channels 5-6 (304c-d) are active for processing events from the second event stream, for stream type Twitter and stream name Department B.
The event processing uses channel map 116 to discover dynamically what event streams are active and supported in the cluster. Furthermore, channel map 116 allows the event processing to query the channel map for an event stream of interest (e.g., by stream type and stream name), and discover a list of consumer channels and consumer nodes that are available to process events from that event stream of interest (e.g., have an Active status).
Some embodiments of channel map 116 also support elasticity in the cluster. The elasticity can include horizontal and vertical scaling. Horizontal scaling refers to support by the event processing for dynamically adding or removing producer and consumer nodes to or from the cluster. Vertical scaling refers to support for dynamically adding or removing consumer channels to or from an individual consumer node, or producer channels to or from an individual producer node. In some embodiments, when a consumer node first comes online and joins the cluster (i.e., horizontal scaling), the consumer node registers its supported event streams and consumer channels with channel map 116. In further embodiments, the producer nodes and consumer nodes can register event listeners and item listeners with channel map 116. In response to a notification of a change to channel map 116, the producer nodes or consumer nodes can react to the changes. Example changes to channel map 116 include adding new channels, removing channels, or updating node IDs, channel names, status, or other fields within a channel. Example reactions to these changes can include a producer node creating corresponding producer channels in response to receiving notifications of new consumer channel entries in channel map 116 from a consumer node. For example, a new consumer node may join the cluster (e.g., horizontal scaling), or an existing consumer node may add support for additional consumer channels (e.g., vertical scaling). In either case, the producer and consumer nodes in the cluster can react accordingly to changes to channel map 116 in response to update notifications from the event listeners and item listeners.
Sharding map 118 correlates partitions for a given event with a producer channel on the producer node. In some embodiments, an administrator configures a partition space for each event stream of interest. The partition space defines a count of discrete chunks that events will be broken into, for a given event stream. In some embodiments, the event system can determine the partition space according to a partition criterion. For example, the producer node can identify based on the channel map a count of consumer nodes available to process the event stream. Upon detecting a relatively higher count of consumer nodes available, the event system can suggest increasing the partition space, to reflect increased capacity for processing the events in the event stream. Upon detecting a relatively lower count of consumer nodes, the event system can suggest decreasing the partition space to reflect decreased capacity for event processing. Alternatively, in some embodiments the partition space can have a predefined default value of 100. Based on the default value of 100, the producer node can partition received events into a value between 0-99 using sharding map 118. For example, if the partition space has a value of 50, the producer node can assign received events a sharding key between 0-49. If the partition space has a value of 200, the producer node can assign a sharding key between 0-199.
When the producer node receives an event stream from one or more event services, the producer node uses sharding map 118 to determine a sharding key for individual events in the event stream and route the events to a producer channel for transmission. In some embodiments, the producer node determines the sharding key as follows. The producer node determines an event key for an individual event. For example, the producer node may determine the event key based on the contents of the event, such as determining a hash code (or fingerprint) of the event. The producer node determines the appropriate partition by calculating a modulo of the event key with the partition space. The producer node looks up the resulting partition in partition-to-channel index 310. For example, if the resulting partition has a value of three, partition-to-channel index 312a would apply and the producer node would provide the event to producer channel 0 (314a). If, instead, the resulting partition has a value of 52, partition-to-channel index 312b would apply and the producer node would provide the event to producer channel 1 (314b). If the resulting partition has a value of 49, partition-to-channel index 312c would apply and the producer node would provide the event to producer channel 2 (314c).
In some embodiments, the event processing system creates and populates sharding map 118 based on channel map 116 according to a sharding algorithm executing, e.g., on one or more of the nodes. With reference to
In one embodiment, the sharding algorithm evenly distributes the partition space uniformly across available producer channels. For example, if the partition space is 100 and there is one producer channel active, the partition-to-channel index would reflect:
In response to a notification that a consumer node has added another active consumer channel, the producer node can expand dynamically and create another producer channel (i.e., vertical scaling). In this embodiment of the sharding algorithm, the partition-to-channel index would become:
The producer node can also populate the channel-partitions map and the partition-to-channel index using these results from the sharding algorithm.
Another embodiment of the sharding algorithm populates sharding map 118 by distributing the partition space dynamically using an overflow portion in response to updates to the available consumer channels. In this embodiment, the sharding algorithm determines an updated partition size based on an updated count of available consumer channels, and assigns overflow portions of the partition space and the updated partition size to a newly added producer channel.
For example, when a consumer node comes online and joins the cluster, the consumer node may advertise one active consumer channel (channel 0) for an event stream of interest. When the producer node receives a notification of the newly active consumer channel, the producer node updates sharding map 118 by assigning all partition values of the partition space to a corresponding new producer channel:
Subsequently, the consumer node may elect to expand the number of consumer channels available, for example by adding a second consumer channel. In response to a notification of the newly active consumer channel, the producer node may redistribute the partition space. The producer node determines an updated partition size based on an updated count of available consumer channels. In some embodiments, the producer node calculates
pz=ps/c Equation (1)
where pz refers to the updated partition size, ps refers to the size of the partition space, and c refers to the updated number of consumer channels. For example, if the consumer node expands the number of consumer channels available to two (c=2), the producer node calculates
pz=ps/c
50=100/2
The producer node assigns partitions to existing producer channels based on the updated partition size pz, and determines an overflow portion of the existing producer channel based on the updated partition size pz and the previous partitions assigned to the producer channel. For example, the producer node dynamically redistributes the partition space as follows:
The producer node updates the sharding map accordingly to reflect the redistributed partition space.
The consumer node may elect to expand further the number of consumer channels available by adding a third consumer channel. The producer node again determines an updated partition size based on an updated count of available consumer channels (c=3):
pz=ps/c
33=100/3
The producer node assigns partitions to existing producer channels based on the updated partition size pz, and distributes the overflow portions of the existing producer channels to the new producer channel.
The producer node proceeds to update sharding map 118 using these values for channel partitions map 306 and partition-to-channel index 310.
In further embodiments, the sharding algorithm assigns partitions to producer channels by determining statistics using channel map 116. For example, the event processing may query the consumer nodes in the cluster to identify available resources for a given consumer node. Non-limiting example resources can include central processing unit (CPU) processor speed or number of cores, available memory or memory speed, available storage capacity or storage read/write speed, network connection speed, a count of existing consumer channels, or combinations thereof. Based on the available resources and on channel map 116, the event processing may determine weighted statistics to assign partitions to producer channels in sharding map 118. For example, if the event processing determines that consumer node U1234567 contains a faster CPU or more memory than other consumer nodes, the event processing may consult channel map 116 to identify consumer channels on consumer node U1234567 (e.g., using a producer node). The sharding algorithm may then assign more partitions to producer channels in sharding map 118 that correspond to consumer channels identified as having more available resources in channel map 118.
Channel-partitions map 306 tracks a related mapping as partition-to-channel index 310. In some embodiments, channel-partitions map 306 can support elasticity in the cluster, such as horizontal or vertical scaling. Channel-partitions map 306 allows sharding map 118 to identify what partitions are associated with a given producer channel and reassign the associated partitions quickly to expand or contract the cluster. For example, channel-partitions map 306 tracks that producer channel 0 can process partitions 0-32 (index 308a), producer channel 1 can process partitions 50-82 (index 308b), and producer channel 2 can process partitions 33-49 and 83-99 (index 308c).
In some embodiments, when consumer node 106a comes online and joins the cluster, the consumer node consults its configuration to identify event streams of interest. For example, the configuration may contain a collection of entries from an administrator that list stream types and stream names of event streams for processing by consumer node 106a. Consumer node 106a creates a set of event listeners 402 for each event stream entry in the configuration. Consumer node 106a further creates corresponding consumer event buffers 404. For example, consumer node 106a can create one consumer event buffer per consumer channel in the set of event listeners 402. In some embodiments the number of event listeners 402 in the set is equal to the value defined in the configuration, or defaults to five. Consumer node 106a further creates requestor pool 406 and pairs one requester with each consumer event buffer.
To create the paired relationship between consumer channels and producer channels, the consumer node registers the newly created consumer channels into the channel map. In some embodiments, the channel map is a Hazelcast distributed map for registering the event listeners and consumer node 106a uses embedded Hazelcast 126a over TCP/IP socket 410 for reading and writing updates to and from the channel map. When one or more producer nodes come online and join the cluster, the producer nodes can then read the channel map to determine how many producer channels to create. In this manner, based on the channel map the producer nodes are able to create and update the sharding map used to transmit events to the various consumer nodes in the cluster.
The event listener uses a consumer event buffer 416 to buffer received events 412. In some embodiments, consumer event buffers 416 are implemented using ring buffer data structures provided by a Disruptor platform. As those skilled in the art will appreciate, the Disruptor platform refers to a publicly available third-party architecture and processing mechanism for high-speed parallel processing, also referred to as concurrency. In further embodiments the ring buffers can be configured with a single thread input and a single thread output (e.g., using event consumer executors 418). This single thread usage provides enhanced throughput over a traditional usage of the Disruptor platform.
Regarding traditional usage, the Disruptor architecture provides components that process volumes of data concurrently, with low latency. The Disruptor architecture implements a data structure referred to herein as a ring buffer. A ring buffer performs similarly to a queue. The event processing system can push data onto and pop data off the ring buffer for processing. Disruptor ring buffers can be used similarly to a network of queues that can operate without needing locks. Locks provide thread safety by avoiding unexpected data corruption due to parallel processing, but also add significant and unwanted complexity. Traditionally the Disruptor pattern is designed for a system to use multiple threads to read from and write to ring buffers. However, using the Disruptor architecture in the traditional manner lowers throughput and reduces performance.
In general, the speed at which the event processing delivers events is affected by the ability to process and serialize the event for delivery. Traditional implementations have projected throughput in the range of 15,000-20,000 events per second. However, this traditional throughput is well below the steady state throughput desired to provide a performant high speed event processing service. To increase throughput, traditional systems can also use multi-threading architectures and aggregate structures. When considering a traditional multi-threading solution, thread contention is one problem to solve. Traditionally, a simplistic approach can use a Java queue with one thread pushing messages onto the queue and one or more threads pulling messages off the queue. To do this, the simplistic approach needs to consider efficient use of the queue as well as thread interactions between pushing messages in and pulling messages off. Instead of a traditional queue, the Disruptor platform delivers higher performance when accessing data by pushing in and pulling off the data from the ring buffer. However, traditional implementations using Disruptor ring buffers use multiple threads to interact with the ring buffers.
In contrast to traditional implementations, in the event processing system the single thread input and single thread output from consumer event buffers 416 (and the corresponding producer event buffers) help differentiate consumer node 106a. Furthermore, the paired consumer channel and producer channel and sharding algorithm applied by the producer node help the event processing system increase throughput over fourfold (e.g., 100,000 messages or more per second) when compared with traditional implementations.
The consumer node passes the received events to event consumer executors 418. Some embodiments of event consumer executors 418 identify a rule associated with received event 412. For example, event transport metadata that the producer node bundles with the event can help identify the appropriate rule for consumer node 106a. Consumer node 106a can then make a call 420 to execute the corresponding rule. For example, event consumer executors 418 can call into the rules engine to get a pre-allocated requestor object dedicated to execution of the rule associated with the received event.
An event service can begin to transmit events to producer node 104a via event streams 502. In some embodiments, the event processing helps the events to arrive in the cluster in the order in which the event service presented the events. When producer node 104a comes online and joins the cluster, producer node 104a retrieves and reads the channel map from the cluster. In some embodiments, the channel map is a Hazelcast distributed map and producer node 104a uses embedded Hazelcast 124a to read the channel map over TCP/IP socket 510. Producer node 104a uses the channel map to determine updates to the sharding map and identify what producer channels to create. The channel map contains information about the consumer nodes in the cluster and which event streams and consumer channels the consumer nodes support. In some embodiments, producer node 104a uses the sharding map in combination with the channel map to determine the consumer node destination of a given event. For example, producer node 104a first receives event stream 502. Producer node 104a shards, or partitions, event stream 502 into individual events (e.g., using event sharding module 504). After sharding event stream 502 into individual events, producer node 104a determines a sharding key for each event based on an event key generated upon receiving the event (e.g., also using event sharding module 504). Producer node 104a looks up the corresponding producer channel among channel pool 506, based on the sharding key in the sharding map. The sharding map identifies a producer channel for transmitting the event. The sharding map helps producer node 104a ultimately deliver a given event consistently to the same producer channel that corresponds to the sharding key and partition determined by the producer node. The paired relationship between the producer channel and consumer channel consequently help the producer node ultimately deliver the event to the same consumer channel on the same consumer node for a given event stream.
Producer node 104a shards 516 the received event stream into individual events according to the following algorithm. Producer node 104a divides the received event stream into individual events. For a given event 512, producer node 104a determines an event key for the event. In some embodiments, the event key is determined based on the contents of event 512. For example, the event key may be a hash code or fingerprint of the contents of event 512. Producer node 104a computes the sharding key by determining a modulo of the event key with the partition space. Producer node 104a identifies a producer channel for transmitting the event, by retrieving the corresponding sharding key in the sharding map for the event stream. Event publisher 514 places event 512 into producer event buffer 518 in the producer channel identified by the sharding map for the event stream. Producer event buffer 518 then routes event 512 from the producer channel to the consumer channel on the corresponding consumer node in the cluster. Upon receiving event 512, the consumer node buffers event 512 for execution.
One aspect that contributes to enhanced throughput is a symbiotic paired relationship between producer channels 314a-c and consumer channels 608a-c. Specifically, some embodiments of the event processing system leverage a symbiotic paired relationship between producer channels 314a-c and producer event buffers 518 on producer node 104a, and consumer channels 608a-c and consumer event buffers 416 on consumer node 106a. Event processing system 600 achieves massive speed because producer node 104a separates and buffers received events so that a single producer event buffer 518 has a direct TCP/IP connection to a corresponding single consumer channel 608a-c and consumer event buffer 416. In some embodiments, this symbiotic paired relationship also allows event processing system 600 to include flow control per paired producer channel 314a-c and consumer channel 608a-c. For example, event processing system 600 can affect the rate at which a consumer channel 608a-c receives and processes events, and match or modulate the speed on a producer event buffer 604a to achieve the desired rate indirectly on a corresponding consumer channel 608a-c. Further embodiments of the flow control can be dynamic, for example by altering the flow rate periodically (e.g., about every thousand messages). The result of this flow control is a steady state in which a producer event buffer 518 and a consumer event buffer 416 on a corresponding consumer channel 608a-c are operating at a related rate to achieve enhanced throughput.
Event processing system 600 begins by receiving an event stream from event service 102a. For example, an external source of the event stream can be a Twitter feed. Some embodiments of event processing system 600 can use an application programming interface (API) for retrieving social media data to retrieve the event stream of interest, such as an API provided by Gnip, Inc. of Boulder, Colo., United States. Event processing system 600 illustrates a cluster including producer node 104a in communication with event service 102a and consumer node 106a. Consumer node 106a announces an ability to accept event streams for Twitter events for Dept. A (e.g., stream type=Twitter and stream name=Dept. A). In this example an administrator has configured consumer node 106a to have three active consumer channels (channel 0 (608a), channel 1 (608b), channel 2 (608c)) for event stream Twitter/Dept. A. Consumer channels 0-2 (608a-c) receive and process events from producer node 104a. Producer node 104a is in communication with event service 102a serving the event stream for Twitter/Dept. A in the cluster.
Consumer node 106a contains three event receivers 414 associated with consumer channels 0-2 (608a-c). Each event receiver 414 has an associated consumer event buffer 416 along with an event consumer executor 418. Event consumer executor 418 pops events off consumer event buffers 416 for execution. In some embodiments, consumer event buffers 416 can be ring buffers using the Disruptor architecture.
Event processing system 600 illustrates symbiotic paired relationships on producer node 104a. Producer node 104a contains a producer channel 0-2 (314a-c) and producer event buffer 518 for each corresponding consumer channel 0-2 (608a-c). Producer event buffers 518 are in a paired relationship with each corresponding consumer channel 0-2 (608a-c). Each producer ring buffer 518 has a producer mechanism 520 that pops events off producer event buffers 518 for transmission to the corresponding consumer channel 0-2 (608a-c).
Event publisher 514 contains sharding logic 516 to shard, or partition, the received event stream into individual events including event 610. Event publisher 514 directs the sharded event 610 to one of producer channels 314a-c which directs the event in turn to an associated producer event buffer 518. Event publisher 514 determines the desired producer channel 314a-c using the sharding map based on a sharding key for event 610.
Producer node 104a passes the sharded event to producer channel 2 (314c) which inserts the event into the associated producer event buffer (step E-3). The event winds its way through the producer event buffer. An extraction thread in the associated event producer extracts the event (step E-4). The extraction thread serializes the event and transmits the event to consumer channel 2 (608c), for example using TCP/IP (step E-5). The receiver associated with consumer channel 2 (608c) receives the event (step E-6) and inserts the received event into the consumer event buffer associated with consumer channel 2 (608c). The consumer event buffer buffers the received event waiting to be processed (step E-7). The event makes its way around the consumer event buffer and the associated event consumer executor retrieves the event using an extraction thread for execution (step E-8). Consumer node 106a executes processing associated with the event (step E-9). In some embodiments, consumer node 106a converts the event using internal data structures and uses event transport metadata bundled with the event to identify rules of interest for processing the event using an associated rule engine to execute the rules. Consumer node 106a then makes call 420 to execute the identified rule. In some embodiments, the identified rule can be an activity rule that identifies an activity for event processing system 600 to perform.
Producer node 104a uses the sharding map and partition space to determine a sharding key for a given event. The sharding key identifies the producer channel 314a-c to receive the sharded event. In some embodiments, the sharding algorithm of computing a modulo of the event key with the partition space determines the sharding key. For example, given an event key of E-109 and a partition space of 100, one embodiment of the sharding algorithm would return a sharding key of 109% 100=9. Producer 104a can look up the sharding key in the sharding map to identify the appropriate producer channel 314a-c for the event. With reference to
Event publisher 514 provides the sharded events to producer channels 0-2 (314a-c) in order, in groups 704a-c (also illustrated in white, black, and gray). Producer channels 0-2 (314a-c) use producer event buffers 518 and event producers 520 to transmit the sharded events. The sharded events travel over TCP/IP to corresponding consumer channels 0-2 (608a-c) on consumer node 106a. Event receivers 414 receive the sharded events into consumer event buffers 416. Event consumer executors 418 extract the received events from event stream 702. Event consumer executors 418 then make calls 420 to execute any rules identified by event transport metadata bundled with the received events. Consumer node 106a may execute the identified rules on a rules engine that is remote to consumer node 106a and producer node 104a, or the rules engine may be locally installed on the consumer node or producer node.
Example Use Cases
An example company GoCars Corp may process large volumes of social media events with enhanced throughput using event processing system 800. GoCars Corp is launching a new marketing campaign that introduces a new type of navigation system for an existing line of cars called “GoCars.” GoCars Corp would like to analyze events such as Twitter messages to see what people are saying about its new offering.
Producer node 104a begins by making a connection to event service 102a. For example, event service 102a may be a social media feed such as a Twitter feed. Producer node 104a proceeds to retrieve events from event service 102a. In some embodiments producer node 104a determines the appropriate consumer node for processing the events based on parsing a subject of the event to identify a predetermined string, and based on a location of the author of the social media post. For example, producer node 104a may parse the event to detect the text “GoCars.”
If producer node 104a identifies the event to be of interest (e.g., if the event contains the predetermined string), producer node 104a determines the sharding key for the event based on a location of the author of the social media post. For example, producer node 104a may determine a zip code associated with the author (e.g., by aggregating zip code data from an external system). Event processing system 800 may configure the channel map and sharding map to assign consumer channels and consumer nodes based on zip code. For example, event processing system 800 may use embedded Hazelcast 124a, 126a-b to configure the channel map and sharding map to transmit events associated with even zip codes to consumer channels on consumer node 106a and odd zip codes to consumer channels on consumer node 106b. In further embodiments, event transport metadata bundled with the event may contain information on how consumer nodes 106a-b should process the event when received.
After receiving the event, producer node 104a determines an associated consumer channel for the event based on the sharding map and on a sharding key for the event. In this example, the sharding map associates the event with a consumer channel on consumer node 106a, based on the zip code. Producer node 104a provides the received event to a producer event buffer in a producer channel. The producer channel uses TCP/IP socket connection 510 to transmit the received event directly to a consumer event buffer associated with a corresponding consumer channel on consumer node 106a. Consumer node 106a receives the event via TCP/IP socket 410 and the consumer event buffer. Consumer node 106a unpacks the event and executes a rule identified by the event transport metadata. In some embodiments, the event transport metadata may identify a rule that causes consumer node 106a to perform sentiment analysis on the received event. Sentiment analysis can use machine learning and decisioning techniques in conjunction with analytic models to identify positive or negative sentiment in the contents of the received event. The event transport metadata may identify a further rule that causes consumer node 106a to store the event to persistent storage, based on the positive or negative result of the sentiment analysis. For example, consumer node 106a may store events with positive sentiment into a marketing database for customer highlights, and events with negative sentiment into a customer database for customer service follow-ups.
In this example, a customer has an external source system that provides line item details for invoices (e.g., medical claims for health insurance). An administrator deploys a rules-based application into a cluster. Consumer nodes 106a-b collate and compute aggregate information about item detail records associated with line items. The event service includes high speed database 902 capable of producing result sets that can be iterated at speeds in the range of 100,000 rows per second. In this application, an incoming record may have an overall case number that groups individual item detail records together. Item detail records may also have a field or indicator that tracks “end sub-group record,” for example. Upon receiving an item detail record with its “end sub-group record” indicator set to true, the application may be configured to collect or group item detail records and write out the set as a single record, for example to database 904.
Producer node 104a connects to an event service such as database 902. Some embodiments of producer node 104a use an event publisher to retrieve events, for example using a database specific query. Producer node 104a creates an event for each row retrieved from database 902 using the database-specific query. In some embodiments, producer node 104a determines the sharding key for each event based on an identified field from the database. For example, the sharding key in conjunction with the sharding map may determine that producer node 104a should route a given event to consumer node 106a. The sharding map directs the event to a producer channel corresponding to a consumer channel associated with consumer node 106a. Producer node 104a uses TCP/IP socket connection 510 to transmit the event directly to consumer node 106a.
Consumer node 106a receives the event via TCP/IP socket 410 into its consumer event buffer from producer node 104a. In some embodiments, consumer node 106a unpacks the received event including bundled event transport metadata. For example, the event transport metadata may identify a rule that checks whether the received event is part of an “existing” set or a “new” set in database 904. Consumer node 106a executes the rule identified by the event transport metadata. The event transport metadata may further identify or define rules creating a structure containing a “new” bucket, or retrieving an “existing” bucket based on the event ID or case ID. In further embodiments, if the event has its “end sub-group” record configured to true, event processing system 900 may group together the events in the bucket in a common format and store the grouped events in database 904 to allow further analysis and processing, potentially by follow-on applications.
In the illustrated example, a company has deployed an external source system that pushes information required by an customer service representative who is online. The company may have a call center where customer service representatives use a rules-based application in communication with event processing system 1000 to respond to end users who call in with questions and answer those questions. In this call center, the application notifies a customer service representative that a call has come in that the representative will respond to, and routes the call to the representative. In addition, the application sends a query to an external system requesting additional information about the incoming caller. If the external system identifies information for consideration by the representative, the external system delivers the information to an entry point (e.g., using event service 102a). The delivered information requires routing to the correct representative. In the cluster illustrated in
The customer service representative begins by using the application deployed by the company on computer 1002. For example, the representative may initiate a new case to track an issue faced by a caller. In some embodiments, event processing system 1000 may present a user interface for a customer service application on computer 1002, in communication with a rules engine and rules base (not shown). The representative on computer 1002 may be assigned to consumer node 106a. In some embodiments, the rules engine and rules base may be deployed on one or more remote systems in communication with producer node 104a or consumer nodes 106a-b. In other embodiments, the rules engine and rules base may be deployed locally on producer node 104a or consumer nodes 106a-b.
In connection with the call received by the representative on consumer node 106a, producer node 104a receives information from event service 102a. For example, event service 102a may be an external information service that provides information about the caller. Producer node may create an event that packages the information from event service 102a. Event processing system 1000 may send a query over TCP/IP socket connection 510 to identify the consumer node associated with the representative helping the caller. The query results may identify the representative and associated customer node 106a.
In some embodiments, producer node 104a may use embedded Hazelcast client 124a to update the sharding map to reflect that received events associated with the caller and representative will route to consumer node 106a. In some embodiments, the representative and caller ID information for the caller may be included as event metadata. Accordingly, when producer node 104a receives the event from event service 102a, producer node 104a provides the event to the correct producer channel based on the sharding map, which transmits the event over TCP/IP socket connection 510 directly to a consumer channel on consumer node 106a as intended. The corresponding consumer event buffer on consumer node 106a receives the event over TCP/IP socket 410. Consumer node 106a processes the event and reads the bundled event transport metadata. The event transport metadata may identify or define a rule to notify the representative on computer 1002 of the information from the external information service. In some embodiments, the application may notify the representative by displaying a pop-up window in a web browser on computer 1002.
Elasticity
For example, method 1100 supports adding a new consumer node to the cluster. When the new consumer node comes online and joins the cluster, the new consumer node begins by registering in a consumer map, and updates the channel map to reflect the event streams and corresponding consumer channels that the new consumer node supports (step 1110). In some embodiments, the consumer map may be a distributed Hazelcast map. When registering in the consumer map, the new consumer node may register initially with a status of Pending.
In response to detecting updates to the channel map, the producer nodes update their configurations (step 1120). For example, the producer nodes update their producer channels based on the updated consumer channel configuration in the channel map. The producer nodes also update the sharding map to assign sharding keys to the updated producer channels. Finally, the producer nodes proceed to buffer events for transmission over the updated producer channels.
In response to detecting updates to the sharding map, all consumer nodes in the cluster determine a consumer-node-wide delta reflecting data to be moved, and copy the data to be moved into a cluster change data map (step 1130). In some embodiments, the cluster change data map may be a distributed Hazelcast map. In response to detecting updates to the cluster change data map, the new consumer node copies the moved data, clears the moved data from the cluster change data map and from the consumer node that previously stored the data, and updates the status to Active for the new consumer node in the consumer map (step 1140). In response to detecting status updates in the consumer map, the producer nodes begin providing data to the producer channel for transmission to the corresponding consumer channels on the new consumer node (step 1150).
Event processing system 1200 supports horizontal scalability by expanding dynamically upon addition of a new consumer node. The addition of consumer node 1202 to the cluster can distribute and potentially increase the event processing capacity of the cluster. The process of adding consumer node 1202 involves a number of changes and updates in the cluster. In some embodiments, there are two sets of changes: one set for consumer nodes and a companion set for producer nodes. Although the changes are described sequentially, in some embodiments some steps can happen in parallel and in different threads, to streamline the addition of consumer node 1202.
New consumer node 1202 begins by coming online and joining the cluster. In general, as part of the startup process new consumer node 1202 orchestrates the rebalancing of the cluster and identifies itself as ready when the process is complete. Consumer node 1202 continues by configuring event stream support. For example, consumer node 1202 reads its configuration to identify supported event streams by stream type and stream name. Consumer node 1202 proceeds to initiate a cluster lock for re-sharding. In some embodiments, the cluster lock on re-sharding prevents initiation of additional re-sharding sessions while the current session is in progress. Consumer node 1202 then registers itself with consumer map 1206. Consumer map 1206 tracks the universe of consumer nodes in the cluster, and producer map 1208 tracks the universe of producer nodes in the cluster. In some embodiments, consumer node 1202 registers with initial status 1212 of Pending.
Based on the event stream support, consumer node 1202 creates corresponding consumer channels and consumer ring buffers to reflect the supported event streams. Consumer node 1202 determines updates to the channel map based on the consumer channels. Consumer node 1202 then initiates a wait loop, waiting for related activity on other consumer and producer nodes in the cluster to signal finalization. In some embodiments, consumer node 1202 registers event listeners or item listeners on cluster change data map 1210 to receive update notifications for changes to the cluster change data map.
Each producer node 104a reacts to changes in the channel map by new consumer node 1202. In some embodiments, producer node 104a reacts by transmitting an end event on all existing producer channels to all corresponding consumer channels in communication with producer node 104a. Producer node 104a proceeds to update the producer channels. For example, producer node 104a creates and initializes new producer event buffers corresponding to the updated producer channels. Producer node 104a next updates the sharding map to reflect the updated producer channels. In some embodiments, producer node 104a pauses production of events to copy the new sharding map, and resumes production of events based on the new sharding map. For events targeted to new consumer node 1202, producer node 104a buffers those events until status 1212 of consumer node 1202 becomes Active in consumer map 1206.
After producer node 104a provides an end event on all producer channels for transmission to all consumer channels, consumer nodes 106a-b, 1202 corresponding to the consumer channels process the received end event. Although the processing of the end event is described herein with reference to consumer node 106a, the processing is the same for all consumer nodes in the cluster. For example, consumer node 106a triggers all consumer event buffers corresponding to all consumer channels to process the received end event as follows. Consumer node 106a reads the updated channel map, and creates a consumer-node-wide delta reflecting the updates. As each consumer event buffer receives the end event to process, the consumer event buffer copies its data to be moved to cluster change data map 1210. Once the consumer event buffer has completed copying the data to move, the consumer event buffer resumes processing any events from the corresponding producer event buffer that have been buffered.
In response to a notification of updates to cluster change data map 1210, new consumer node 1202 continues to wait for all moved partitions to appear in cluster change data map 1210. Once the moved partitions all appear in cluster change data map 1210, consumer node 1202 copies the data from cluster change data map 1210 and clears the contents it copied from cluster change data map 1210 and from the consumer nodes that previously stored the copied data. Upon completion of copying the data, consumer node 1202 updates its status to Active in consumer map 1206 and clears the cluster lock on re-sharding. Lastly, upon receiving a notification of the status update for consumer node 1202 in consumer map 1206, producer node 104a begins providing events to the producer channels associated with consumer node 1202, and the producer channels begin transmitting the events directly to consumer node 1202.
The elasticity support in event processing system 1200 has thus far discussed adding a consumer node to the cluster. Method 1200 also supports the following dynamic rebalancing: removing consumer node 1202, adding consumer channels to a consumer node in support of a given event stream, and removing consumer channels from a consumer node for the event stream.
With reference to
In some embodiments, event processing system 1200 supports adding consumer channels to an existing consumer node in support of vertical scalability. In contrast to adding a new consumer node, the update to the channel map comprises changing (i.e., increasing) the number of consumer channels supported for a given event stream in an existing updated consumer node. Upon receiving a notification of an update to the channel map, event processing system 1200 rebalances the cluster by processing the channel map and sharding map in a similar manner as for addition of a new consumer node. That is, the updated consumer node updates the channel map to reflect the increased consumer channels for event streams of interest (step 1110). In response to detecting updates to the channel map, producer node 104a creates producer channels corresponding to the new consumer channels created on the existing consumer node, and updates the sharding map to reflect the new producer channels (step 1120). In response to detecting updates to the sharding map, the consumer node determines a delta that reflects data previously assigned to other consumer channels or other consumer nodes, and copies the data to be moved into the cluster change map (step 1130). In response to detecting the updates to the cluster change map, the existing updated consumer node copies the moved data and clears the moved data from the cluster change map and from the previous consumer node storing the data prior to moving (step 1140). After copying the moved data, producer node 104a begins providing data to the new producer channels for transmission to the updated consumer channels (step 1150).
In some embodiments, event processing system 1200 supports removing consumer channels from an existing consumer node. In this case the update to the channel map comprises changing (i.e., decreasing) the number of consumer channels supported for a given event stream. Event processing system 1200 processes the updates to the channel map in a similar manner as with removal of an existing consumer node. That is, the updated consumer node updates the channel map to reflect the decreased consumer channels for event streams of interest (step 1110). In response to detecting updates to the channel map, producer node 104a removes producer channels corresponding to the removed consumer channels created on the existing consumer node, and updates the sharding map to reflect the removed producer channels (step 1120). In response to detecting updates to the sharding map, the consumer node determines a delta that reflects data to be assigned to other consumer channels on the consumer node or on other consumer nodes, and copies the data to be moved into the cluster change map (step 1130). In response to detecting the updates to the cluster change map, the existing updated consumer node copies the moved data and clears the moved data from the cluster change map and from the previous consumer node that stored the data prior to moving (step 1140). After copying the moved data, producer node 104a begins providing data to the new producer channels for transmission to the updated consumer channels (step 1150).
The event processing systems and methods described herein address a technical problem of rapidly routing and processing large volumes of discrete events in a networked cluster. The event processing provides a technical contribution that involves determining a sharding key for a received event from an event stream, and using a sharding map to correlate the sharding key for the event with a producer channel. A producer node provides the received event to a producer event buffer associated with the producer channel, and the producer event buffer transmits the event to a corresponding consumer event buffer associated with a consumer channel on a consumer node. The sharding map, sharding key, and paired producer channels and consumer channels, among other aspects, add specific limitations other than what is well-understood, routine, and conventional in the field. The event processing further provides significantly more than an abstract idea by improving the functioning of the producer and consumer digital data processors themselves. For example, the event processing achieves improved throughput in the cluster by leveraging the paired relationship between producer channels on the producer node and consumer channels on the consumer node. The event processing also supports dynamic rebalancing of the system in response to adding or removing producer or consumer nodes, or adding or removing producer or consumer channels to or from producer or consumer nodes. These enhancements elevate the producer and consumer digital data processors beyond their conventional functioning.
Other embodiments are within the scope and spirit of the event processing systems and methods. For example, the event processing functionality described above can be implemented using software, hardware, firmware, hardwiring, or combinations of any of these. One or more digital data processors operating in accordance with instructions may implement the functions associated with event processing in accordance with the present disclosure as described above. If such is the case, it is within the scope of the event processing systems and methods that such instructions may be stored on one or more non-transitory computer-readable storage media (e.g., a magnetic disk, solid state drive, or other storage medium). Additionally, as described earlier, modules implementing functions may also be physically located at various positions, including being distributed such that portions of functions are implemented at different physical locations.
The event processing systems and methods are not to be limited in scope by the specific embodiments described herein. Indeed, other various embodiments of and modifications to the event processing, in addition to those described herein, will be apparent to those of ordinary skill in the art from the foregoing description and accompanying drawings. Thus, such other embodiments and modifications are intended to fall within the scope of the event processing systems and methods described herein. Furthermore, although the event processing has been described herein in the context of a particular implementation in a particular environment for a particular purpose, those of ordinary skill in the art will recognize that its usefulness is not limited thereto and that the event processing may be beneficially implemented in any number of environments for any number of purposes.
This application claims the benefit of U.S. Provisional Application No. 62/062,515 entitled “Internal Message Processing with Enhanced Throughput,” filed Oct. 10, 2014, which is incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4047059 | Rosenthal | Sep 1977 | A |
4344142 | Diehr, II et al. | Aug 1982 | A |
4602168 | Single | Jul 1986 | A |
4607232 | Gill, Jr. | Aug 1986 | A |
4659944 | Miller, Sr. et al. | Apr 1987 | A |
4701130 | Whitney et al. | Oct 1987 | A |
4866634 | Reboh et al. | Sep 1989 | A |
4884217 | Skeirik et al. | Nov 1989 | A |
4895518 | Arnold et al. | Jan 1990 | A |
4930071 | Tou et al. | May 1990 | A |
4953106 | Gansner et al. | Aug 1990 | A |
5062060 | Kolnick | Oct 1991 | A |
5077491 | Heck et al. | Dec 1991 | A |
5093794 | Howie et al. | Mar 1992 | A |
5119465 | Jack et al. | Jun 1992 | A |
5129043 | Yue | Jul 1992 | A |
5136184 | Deevy | Aug 1992 | A |
5136523 | Landers | Aug 1992 | A |
5140671 | Hayes et al. | Aug 1992 | A |
5193056 | Boes | Mar 1993 | A |
5199068 | Cox | Mar 1993 | A |
5204939 | Yamazaki et al. | Apr 1993 | A |
5228116 | Harris et al. | Jul 1993 | A |
5259766 | Sack et al. | Nov 1993 | A |
5262941 | Saladin et al. | Nov 1993 | A |
5267175 | Hooper | Nov 1993 | A |
5267865 | Lee et al. | Dec 1993 | A |
5270920 | Pearse et al. | Dec 1993 | A |
5276359 | Chiang | Jan 1994 | A |
5276885 | Milnes et al. | Jan 1994 | A |
5291394 | Chapman | Mar 1994 | A |
5291583 | Bapat | Mar 1994 | A |
5295256 | Bapat | Mar 1994 | A |
5297279 | Bannon et al. | Mar 1994 | A |
5301270 | Steinberg et al. | Apr 1994 | A |
5310349 | Daniels et al. | May 1994 | A |
5311422 | Loftin et al. | May 1994 | A |
5326270 | Ostby et al. | Jul 1994 | A |
5333254 | Robertson | Jul 1994 | A |
5337407 | Bates et al. | Aug 1994 | A |
5339390 | Robertson et al. | Aug 1994 | A |
5374932 | Wyschogrod et al. | Dec 1994 | A |
5379366 | Noyes | Jan 1995 | A |
5379387 | Carlstedt et al. | Jan 1995 | A |
5381332 | Wood | Jan 1995 | A |
5386559 | Eisenberg et al. | Jan 1995 | A |
5395243 | Lubin et al. | Mar 1995 | A |
5412756 | Bauman et al. | May 1995 | A |
5421011 | Camillone et al. | May 1995 | A |
5421730 | Lasker, III et al. | Jun 1995 | A |
5446397 | Yotsuyanagi | Aug 1995 | A |
5446885 | Moore et al. | Aug 1995 | A |
5450480 | Man et al. | Sep 1995 | A |
5463682 | Fisher et al. | Oct 1995 | A |
5473732 | Chang | Dec 1995 | A |
5477170 | Yotsuyanagi | Dec 1995 | A |
5481647 | Brody et al. | Jan 1996 | A |
5499293 | Behram et al. | Mar 1996 | A |
5504879 | Eisenberg et al. | Apr 1996 | A |
5512849 | Wong | Apr 1996 | A |
5519618 | Kastner et al. | May 1996 | A |
5537590 | Amado | Jul 1996 | A |
5542024 | Balint et al. | Jul 1996 | A |
5542078 | Martel et al. | Jul 1996 | A |
5548506 | Srinivasan | Aug 1996 | A |
5561740 | Barrett et al. | Oct 1996 | A |
5579223 | Raman | Nov 1996 | A |
5579486 | Oprescu et al. | Nov 1996 | A |
5586311 | Davies et al. | Dec 1996 | A |
5596752 | Knudsen et al. | Jan 1997 | A |
5597312 | Bloom et al. | Jan 1997 | A |
5608789 | Fisher et al. | Mar 1997 | A |
5611076 | Durtlinger et al. | Mar 1997 | A |
5627979 | Chang et al. | May 1997 | A |
5630127 | Moore et al. | May 1997 | A |
5649192 | Stucky | Jul 1997 | A |
5655118 | Heindel et al. | Aug 1997 | A |
5664206 | Murow et al. | Sep 1997 | A |
5675753 | Hansen et al. | Oct 1997 | A |
5678039 | Hinks et al. | Oct 1997 | A |
5689663 | Williams | Nov 1997 | A |
5715450 | Ambrose et al. | Feb 1998 | A |
5732192 | Malin et al. | Mar 1998 | A |
5754740 | Fukuoka et al. | May 1998 | A |
5761063 | Jannelle et al. | Jun 1998 | A |
5761673 | Bookman et al. | Jun 1998 | A |
5765140 | Knudson et al. | Jun 1998 | A |
5768480 | Crawford, Jr. et al. | Jun 1998 | A |
5788504 | Rice et al. | Aug 1998 | A |
5795155 | Morrel-Samuels | Aug 1998 | A |
5809212 | Shasha | Sep 1998 | A |
5815415 | Bentley et al. | Sep 1998 | A |
5819243 | Rich et al. | Oct 1998 | A |
5819257 | Monge et al. | Oct 1998 | A |
5822780 | Schutzman | Oct 1998 | A |
5825260 | Ludwig et al. | Oct 1998 | A |
5826077 | Blakeley et al. | Oct 1998 | A |
5826239 | Du et al. | Oct 1998 | A |
5826250 | Trefler | Oct 1998 | A |
5826252 | Wolters, Jr. et al. | Oct 1998 | A |
5829983 | Koyama et al. | Nov 1998 | A |
5831607 | Brooks | Nov 1998 | A |
5832483 | Barker | Nov 1998 | A |
5841435 | Dauerer et al. | Nov 1998 | A |
5841673 | Kobayashi et al. | Nov 1998 | A |
5864865 | Lakis | Jan 1999 | A |
5873096 | Lim et al. | Feb 1999 | A |
5875334 | Chow et al. | Feb 1999 | A |
5875441 | Nakatsuyama et al. | Feb 1999 | A |
5880614 | Zinke et al. | Mar 1999 | A |
5880742 | Rao et al. | Mar 1999 | A |
5886546 | Hwang | Mar 1999 | A |
5890146 | Wavish et al. | Mar 1999 | A |
5890166 | Eisenberg et al. | Mar 1999 | A |
5892512 | Donnelly et al. | Apr 1999 | A |
5907490 | Oliver | May 1999 | A |
5907837 | Ferrel et al. | May 1999 | A |
5909213 | Martin | Jun 1999 | A |
5910748 | Reffay et al. | Jun 1999 | A |
5911138 | Li et al. | Jun 1999 | A |
5918222 | Fukui et al. | Jun 1999 | A |
5920717 | Noda | Jul 1999 | A |
5930795 | Chen et al. | Jul 1999 | A |
5945852 | Kosiec | Aug 1999 | A |
5974441 | Rogers et al. | Oct 1999 | A |
5974443 | Jeske | Oct 1999 | A |
5978566 | Plank et al. | Nov 1999 | A |
5983267 | Shklar et al. | Nov 1999 | A |
5983369 | Bakoglu et al. | Nov 1999 | A |
5987415 | Breese et al. | Nov 1999 | A |
5990742 | Suzuki | Nov 1999 | A |
5995948 | Whitford et al. | Nov 1999 | A |
5995958 | Xu | Nov 1999 | A |
6008673 | Glass et al. | Dec 1999 | A |
6008808 | Almeida et al. | Dec 1999 | A |
6012098 | Bayeh et al. | Jan 2000 | A |
6020768 | Lim | Feb 2000 | A |
6023704 | Gerard et al. | Feb 2000 | A |
6023714 | Hill et al. | Feb 2000 | A |
6023717 | Argyroudis | Feb 2000 | A |
6028457 | Tihanyi | Feb 2000 | A |
6037890 | Glass et al. | Mar 2000 | A |
6044373 | Gladney et al. | Mar 2000 | A |
6044466 | Anand et al. | Mar 2000 | A |
6078982 | Du et al. | Jun 2000 | A |
6085188 | Bachmann et al. | Jul 2000 | A |
6085198 | Skinner et al. | Jul 2000 | A |
6091226 | Amano | Jul 2000 | A |
6092036 | Hamann | Jul 2000 | A |
6092083 | Brodersen et al. | Jul 2000 | A |
6094652 | Faisal | Jul 2000 | A |
6098172 | Coss et al. | Aug 2000 | A |
6105035 | Monge et al. | Aug 2000 | A |
6108004 | Medl | Aug 2000 | A |
6122632 | Botts et al. | Sep 2000 | A |
6125363 | Buzzeo et al. | Sep 2000 | A |
6130679 | Chen et al. | Oct 2000 | A |
6137797 | Bass et al. | Oct 2000 | A |
6144997 | Lamming et al. | Nov 2000 | A |
6151595 | Pirolli et al. | Nov 2000 | A |
6151624 | Teare et al. | Nov 2000 | A |
6154738 | Call | Nov 2000 | A |
6167441 | Himmel | Dec 2000 | A |
6177932 | Galdes et al. | Jan 2001 | B1 |
6185516 | Hardin et al. | Feb 2001 | B1 |
6185534 | Breese et al. | Feb 2001 | B1 |
6192371 | Schultz | Feb 2001 | B1 |
6194919 | Park | Feb 2001 | B1 |
6212502 | Ball et al. | Apr 2001 | B1 |
6216135 | Brodersen et al. | Apr 2001 | B1 |
6233332 | Anderson et al. | May 2001 | B1 |
6233617 | Rothwein et al. | May 2001 | B1 |
6240417 | Eastwick et al. | May 2001 | B1 |
6243713 | Nelson et al. | Jun 2001 | B1 |
6246320 | Monroe | Jun 2001 | B1 |
6275073 | Tokuhiro | Aug 2001 | B1 |
6275790 | Yamamoto et al. | Aug 2001 | B1 |
6281896 | Alimpich et al. | Aug 2001 | B1 |
6282547 | Hirsch | Aug 2001 | B1 |
6300947 | Kanevsky | Oct 2001 | B1 |
6304259 | DeStefano | Oct 2001 | B1 |
6308163 | Du et al. | Oct 2001 | B1 |
6310951 | Wineberg et al. | Oct 2001 | B1 |
6311324 | Smith et al. | Oct 2001 | B1 |
6313834 | Lau et al. | Nov 2001 | B1 |
6314415 | Mukherjee | Nov 2001 | B1 |
6324693 | Brodersen et al. | Nov 2001 | B1 |
6330554 | Altschuler et al. | Dec 2001 | B1 |
6338074 | Poindexter et al. | Jan 2002 | B1 |
6341277 | Coden et al. | Jan 2002 | B1 |
6341293 | Hennessey | Jan 2002 | B1 |
6344862 | Williams et al. | Feb 2002 | B1 |
6349238 | Gabbita et al. | Feb 2002 | B1 |
6351734 | Lautzenheiser et al. | Feb 2002 | B1 |
6356286 | Lawrence | Mar 2002 | B1 |
6356897 | Gusack | Mar 2002 | B1 |
6359633 | Balasubramaniam et al. | Mar 2002 | B1 |
6366299 | Lanning et al. | Apr 2002 | B1 |
6369819 | Pitkow et al. | Apr 2002 | B1 |
6370537 | Gilbert et al. | Apr 2002 | B1 |
6380910 | Moustakas et al. | Apr 2002 | B1 |
6380947 | Stead | Apr 2002 | B1 |
6381738 | Choi et al. | Apr 2002 | B1 |
6389460 | Stewart et al. | May 2002 | B1 |
6389510 | Chen et al. | May 2002 | B1 |
6393605 | Loomans | May 2002 | B1 |
6396885 | Ding et al. | May 2002 | B1 |
6405211 | Sokol et al. | Jun 2002 | B1 |
6405251 | Bullard et al. | Jun 2002 | B1 |
6415259 | Wolfinger et al. | Jul 2002 | B1 |
6415283 | Conklin | Jul 2002 | B1 |
6418448 | Sarkar | Jul 2002 | B1 |
6421571 | Spriggs et al. | Jul 2002 | B1 |
6426723 | Smith et al. | Jul 2002 | B1 |
6429870 | Chen et al. | Aug 2002 | B1 |
6430571 | Doan et al. | Aug 2002 | B1 |
6430574 | Stead | Aug 2002 | B1 |
6437799 | Shinomi et al. | Aug 2002 | B1 |
6446065 | Nishioka et al. | Sep 2002 | B1 |
6446089 | Brodersen et al. | Sep 2002 | B1 |
6446200 | Ball et al. | Sep 2002 | B1 |
6446256 | Hyman et al. | Sep 2002 | B1 |
6448964 | Isaacs et al. | Sep 2002 | B1 |
6453038 | McFarlane et al. | Sep 2002 | B1 |
6463346 | Flockhart et al. | Oct 2002 | B1 |
6463440 | Hind et al. | Oct 2002 | B1 |
6469715 | Carter et al. | Oct 2002 | B1 |
6469716 | Carter et al. | Oct 2002 | B1 |
6473467 | Wallace et al. | Oct 2002 | B1 |
6473748 | Archer | Oct 2002 | B1 |
6493331 | Walton et al. | Dec 2002 | B1 |
6493399 | Xia et al. | Dec 2002 | B1 |
6493731 | Jones et al. | Dec 2002 | B1 |
6493754 | Rosborough et al. | Dec 2002 | B1 |
6496812 | Campaigne et al. | Dec 2002 | B1 |
6496833 | Goldberg et al. | Dec 2002 | B1 |
6502239 | Zgarba et al. | Dec 2002 | B2 |
6509898 | Chi et al. | Jan 2003 | B2 |
6513018 | Culhane | Jan 2003 | B1 |
6526440 | Bharat | Feb 2003 | B1 |
6526457 | Birze | Feb 2003 | B1 |
6529217 | Maguire, III et al. | Mar 2003 | B1 |
6529899 | Kraft et al. | Mar 2003 | B1 |
6529900 | Patterson et al. | Mar 2003 | B1 |
6530079 | Choi et al. | Mar 2003 | B1 |
6532474 | Iwamoto et al. | Mar 2003 | B2 |
6539374 | Jung | Mar 2003 | B2 |
6542912 | Meltzer et al. | Apr 2003 | B2 |
6546381 | Subramanian et al. | Apr 2003 | B1 |
6546406 | DeRose et al. | Apr 2003 | B1 |
6549904 | Ortega et al. | Apr 2003 | B1 |
6556226 | Gould et al. | Apr 2003 | B2 |
6556983 | Altschuler et al. | Apr 2003 | B1 |
6556985 | Karch | Apr 2003 | B1 |
6559864 | Olin | May 2003 | B1 |
6560592 | Reid et al. | May 2003 | B1 |
6560649 | Mullen et al. | May 2003 | B1 |
6567419 | Yarlagadda | May 2003 | B1 |
6571222 | Matsumoto et al. | May 2003 | B1 |
6577769 | Kenyon et al. | Jun 2003 | B1 |
6583800 | Ridgley et al. | Jun 2003 | B1 |
6584464 | Warthen | Jun 2003 | B1 |
6584569 | Reshef et al. | Jun 2003 | B2 |
6594662 | Sieffert et al. | Jul 2003 | B1 |
6597381 | Eskridge et al. | Jul 2003 | B1 |
6597775 | Lawyer et al. | Jul 2003 | B2 |
6598043 | Baclawski | Jul 2003 | B1 |
6606613 | Altschuler et al. | Aug 2003 | B1 |
6625657 | Bullard | Sep 2003 | B1 |
6629138 | Lambert et al. | Sep 2003 | B1 |
6636850 | Lepien | Oct 2003 | B2 |
6636901 | Sudhakaran et al. | Oct 2003 | B2 |
6643638 | Xu | Nov 2003 | B1 |
6643652 | Helgeson et al. | Nov 2003 | B2 |
6661889 | Flockhart et al. | Dec 2003 | B1 |
6661908 | Suchard et al. | Dec 2003 | B1 |
6678679 | Bradford | Jan 2004 | B1 |
6678773 | Marietta et al. | Jan 2004 | B2 |
6678882 | Hurley et al. | Jan 2004 | B1 |
6684261 | Orton et al. | Jan 2004 | B1 |
6690788 | Bauer et al. | Feb 2004 | B1 |
6691067 | Ding et al. | Feb 2004 | B1 |
6691230 | Bardon | Feb 2004 | B1 |
6701314 | Conover et al. | Mar 2004 | B1 |
6711565 | Subramaniam et al. | Mar 2004 | B1 |
6721747 | Lipkin | Apr 2004 | B2 |
6728702 | Subramaniam et al. | Apr 2004 | B1 |
6728852 | Stoutamire | Apr 2004 | B1 |
6732095 | Warshaysky et al. | May 2004 | B1 |
6732111 | Brodersen et al. | May 2004 | B2 |
6748422 | Morin et al. | Jun 2004 | B2 |
6750858 | Rosenstein | Jun 2004 | B1 |
6751663 | Farrell et al. | Jun 2004 | B1 |
6754475 | Harrison et al. | Jun 2004 | B1 |
6756994 | Tlaskal | Jun 2004 | B1 |
6763351 | Subramaniam et al. | Jul 2004 | B1 |
6771706 | Ling et al. | Aug 2004 | B2 |
6772148 | Baclawski | Aug 2004 | B2 |
6772350 | Belani et al. | Aug 2004 | B1 |
6778971 | Altschuler et al. | Aug 2004 | B1 |
6782091 | Dunning, III | Aug 2004 | B1 |
6785341 | Walton et al. | Aug 2004 | B2 |
6788114 | Krenzke et al. | Sep 2004 | B1 |
6792420 | Chen et al. | Sep 2004 | B2 |
RE38633 | Srinivasan | Oct 2004 | E |
6804330 | Jones et al. | Oct 2004 | B1 |
6807632 | Carpentier et al. | Oct 2004 | B1 |
6810429 | Walsh et al. | Oct 2004 | B1 |
6820082 | Cook et al. | Nov 2004 | B1 |
6829655 | Huang et al. | Dec 2004 | B1 |
6831668 | Cras et al. | Dec 2004 | B2 |
6836275 | Arquie et al. | Dec 2004 | B1 |
6839682 | Blume et al. | Jan 2005 | B1 |
6847982 | Parker et al. | Jan 2005 | B2 |
6851089 | Erickson et al. | Feb 2005 | B1 |
6856575 | Jones | Feb 2005 | B2 |
6856992 | Britton et al. | Feb 2005 | B2 |
6859787 | Fisher et al. | Feb 2005 | B2 |
6865546 | Song | Mar 2005 | B1 |
6865566 | Serrano-Morales et al. | Mar 2005 | B2 |
6865575 | Smith et al. | Mar 2005 | B1 |
6867789 | Allen et al. | Mar 2005 | B1 |
6918222 | Lat et al. | Jul 2005 | B2 |
6920615 | Campbell et al. | Jul 2005 | B1 |
6925457 | Britton et al. | Aug 2005 | B2 |
6925609 | Lucke | Aug 2005 | B1 |
6927728 | Vook et al. | Aug 2005 | B2 |
6934702 | Faybishenko et al. | Aug 2005 | B2 |
6940917 | Menon et al. | Sep 2005 | B2 |
6944644 | Gideon | Sep 2005 | B2 |
6954737 | Kalantar et al. | Oct 2005 | B2 |
6956845 | Baker et al. | Oct 2005 | B2 |
6959432 | Crocker | Oct 2005 | B2 |
6961725 | Yuan et al. | Nov 2005 | B2 |
6965889 | Serrano-Morales et al. | Nov 2005 | B2 |
6966033 | Gasser et al. | Nov 2005 | B1 |
6976144 | Trefler et al. | Dec 2005 | B1 |
6978719 | Sebata et al. | Dec 2005 | B2 |
6985912 | Mullins et al. | Jan 2006 | B2 |
6991153 | Silverbrook et al. | Jan 2006 | B2 |
7020869 | Abrari et al. | Mar 2006 | B2 |
7020882 | Lewallen | Mar 2006 | B1 |
7028225 | Maso et al. | Apr 2006 | B2 |
7031901 | Abu | Apr 2006 | B2 |
7035808 | Ford | Apr 2006 | B1 |
7058367 | Luo et al. | Jun 2006 | B1 |
7058637 | Britton et al. | Jun 2006 | B2 |
7064766 | Beda et al. | Jun 2006 | B2 |
7073177 | Foote et al. | Jul 2006 | B2 |
7076558 | Dunn | Jul 2006 | B1 |
7089193 | Newbold | Aug 2006 | B2 |
7103173 | Rodenbusch et al. | Sep 2006 | B2 |
7124145 | Surasinghe | Oct 2006 | B2 |
7139999 | Bowman-Amuah | Nov 2006 | B2 |
7143116 | Okitsu et al. | Nov 2006 | B2 |
7171145 | Takeuchi et al. | Jan 2007 | B2 |
7171415 | Kan et al. | Jan 2007 | B2 |
7174514 | Subramaniam et al. | Feb 2007 | B2 |
7178109 | Hewson et al. | Feb 2007 | B2 |
7194380 | Barrow et al. | Mar 2007 | B2 |
7194690 | Guillermo et al. | Mar 2007 | B2 |
7289793 | Norwood et al. | Oct 2007 | B2 |
RE39918 | Slemmer | Nov 2007 | E |
7302417 | Iyer | Nov 2007 | B2 |
7318020 | Kim | Jan 2008 | B1 |
7318066 | Kaufman et al. | Jan 2008 | B2 |
7334039 | Majkut et al. | Feb 2008 | B1 |
7343295 | Pomerance | Mar 2008 | B2 |
7353229 | Vilcauskas, Jr. et al. | Apr 2008 | B2 |
7398391 | Carpentier et al. | Jul 2008 | B2 |
7406475 | Dorne et al. | Jul 2008 | B2 |
7412388 | Dalal et al. | Aug 2008 | B2 |
7415731 | Carpentier et al. | Aug 2008 | B2 |
7505827 | Boddy et al. | Mar 2009 | B1 |
7526481 | Cusson et al. | Apr 2009 | B1 |
7536294 | Stanz et al. | May 2009 | B1 |
7555645 | Vissapragada | Jun 2009 | B2 |
7574494 | Mayemia et al. | Aug 2009 | B1 |
7596504 | Hughes et al. | Sep 2009 | B2 |
7640222 | Trefier | Dec 2009 | B2 |
7647417 | Taneja | Jan 2010 | B1 |
7665063 | Hofmann et al. | Feb 2010 | B1 |
7685013 | Gendler | Mar 2010 | B2 |
7689447 | Aboujaoude et al. | Mar 2010 | B1 |
7711919 | Trefler et al. | May 2010 | B2 |
7779395 | Chotin et al. | Aug 2010 | B1 |
7783596 | Smolen et al. | Aug 2010 | B2 |
7787609 | Flockhart et al. | Aug 2010 | B1 |
7791559 | Piasecki | Sep 2010 | B2 |
7818506 | Shepstone et al. | Oct 2010 | B1 |
7844594 | Holt et al. | Nov 2010 | B1 |
7870244 | Chong et al. | Jan 2011 | B2 |
7889896 | Roehrig et al. | Feb 2011 | B2 |
7937690 | Casey | May 2011 | B2 |
7971180 | Kreamer et al. | Jun 2011 | B2 |
7974714 | Hoffberg | Jul 2011 | B2 |
7983895 | McEntee et al. | Jul 2011 | B2 |
8001519 | Conallen et al. | Aug 2011 | B2 |
8037329 | Leech et al. | Oct 2011 | B2 |
8073802 | Trefler | Dec 2011 | B2 |
8250525 | Khatutsky | Aug 2012 | B2 |
8335704 | Trefler et al. | Dec 2012 | B2 |
8386960 | Eismann et al. | Feb 2013 | B1 |
8468492 | Frenkel | Jun 2013 | B1 |
8479157 | Trefler et al. | Jul 2013 | B2 |
8516193 | Clinton et al. | Aug 2013 | B1 |
8739044 | Varadarajan | May 2014 | B1 |
8744999 | Clarke et al. | Jun 2014 | B2 |
8843435 | Trefler et al. | Sep 2014 | B1 |
8863008 | Chan et al. | Oct 2014 | B2 |
8880487 | Clinton et al. | Nov 2014 | B1 |
8924335 | Trefler et al. | Dec 2014 | B1 |
8959480 | Trefler et al. | Feb 2015 | B2 |
9026733 | Clinton et al. | May 2015 | B1 |
9189361 | Khatutsky | Nov 2015 | B2 |
9195936 | Chase | Nov 2015 | B1 |
9270743 | Frenkel | Feb 2016 | B2 |
9495340 | Powell et al. | Nov 2016 | B2 |
9678719 | Frenkel | Jun 2017 | B1 |
20010013799 | Wang | Aug 2001 | A1 |
20010035777 | Wang et al. | Nov 2001 | A1 |
20010047355 | Anwar | Nov 2001 | A1 |
20010049682 | Vincent et al. | Dec 2001 | A1 |
20010052108 | Bowman-Amuah | Dec 2001 | A1 |
20010054064 | Kannan | Dec 2001 | A1 |
20020010855 | Reshef et al. | Jan 2002 | A1 |
20020013804 | Gideon | Jan 2002 | A1 |
20020029161 | Brodersen et al. | Mar 2002 | A1 |
20020042831 | Capone et al. | Apr 2002 | A1 |
20020049603 | Mehra et al. | Apr 2002 | A1 |
20020049715 | Serrano-Morales et al. | Apr 2002 | A1 |
20020049788 | Lipkin et al. | Apr 2002 | A1 |
20020054152 | Palaniappan et al. | May 2002 | A1 |
20020059566 | Delcambre et al. | May 2002 | A1 |
20020065912 | Catchpole et al. | May 2002 | A1 |
20020070972 | Windl et al. | Jun 2002 | A1 |
20020073337 | Ioele et al. | Jun 2002 | A1 |
20020083063 | Egolf | Jun 2002 | A1 |
20020091677 | Sridhar | Jul 2002 | A1 |
20020091678 | Miller et al. | Jul 2002 | A1 |
20020091710 | Dunham et al. | Jul 2002 | A1 |
20020091835 | Lentini et al. | Jul 2002 | A1 |
20020093537 | Bocioned et al. | Jul 2002 | A1 |
20020107684 | Gao | Aug 2002 | A1 |
20020118688 | Jagannathan | Aug 2002 | A1 |
20020120598 | Shadmon et al. | Aug 2002 | A1 |
20020120627 | Mankoff | Aug 2002 | A1 |
20020120762 | Cheng et al. | Aug 2002 | A1 |
20020133502 | Rosenthal et al. | Sep 2002 | A1 |
20020177232 | Melker et al. | Nov 2002 | A1 |
20020178232 | Ferguson | Nov 2002 | A1 |
20020181692 | Flockhart et al. | Dec 2002 | A1 |
20020184610 | Chong et al. | Dec 2002 | A1 |
20020186826 | Hsu et al. | Dec 2002 | A1 |
20020198935 | Crandall et al. | Dec 2002 | A1 |
20030001894 | Boykin et al. | Jan 2003 | A1 |
20030004934 | Qian | Jan 2003 | A1 |
20030004951 | Chokshi | Jan 2003 | A1 |
20030009239 | Lombardo et al. | Jan 2003 | A1 |
20030014399 | Hansen et al. | Jan 2003 | A1 |
20030037145 | Fagan | Feb 2003 | A1 |
20030050834 | Caplan | Mar 2003 | A1 |
20030050927 | Hussam | Mar 2003 | A1 |
20030050929 | Bookman et al. | Mar 2003 | A1 |
20030061209 | Raboczi et al. | Mar 2003 | A1 |
20030065544 | Elzinga et al. | Apr 2003 | A1 |
20030066031 | Laane | Apr 2003 | A1 |
20030074352 | Raboczi et al. | Apr 2003 | A1 |
20030074369 | Scheutze et al. | Apr 2003 | A1 |
20030084401 | Abel et al. | May 2003 | A1 |
20030093279 | Malah et al. | May 2003 | A1 |
20030098991 | Laverty et al. | May 2003 | A1 |
20030109951 | Hsiung et al. | Jun 2003 | A1 |
20030115281 | McHenry et al. | Jun 2003 | A1 |
20030135358 | Lissauer et al. | Jul 2003 | A1 |
20030152212 | Burok et al. | Aug 2003 | A1 |
20030154380 | Richmond et al. | Aug 2003 | A1 |
20030191626 | Al-Onaizan et al. | Oct 2003 | A1 |
20030198337 | Lenard | Oct 2003 | A1 |
20030200254 | Wei | Oct 2003 | A1 |
20030200371 | Abujbara | Oct 2003 | A1 |
20030202617 | Casper | Oct 2003 | A1 |
20030222680 | Jaussi | Dec 2003 | A1 |
20030229529 | Mui et al. | Dec 2003 | A1 |
20030229544 | Veres et al. | Dec 2003 | A1 |
20040003043 | Rajamony et al. | Jan 2004 | A1 |
20040021686 | Barberis | Feb 2004 | A1 |
20040024603 | Mahoney et al. | Feb 2004 | A1 |
20040034651 | Gupta et al. | Feb 2004 | A1 |
20040049479 | Dorne et al. | Mar 2004 | A1 |
20040049509 | Keller et al. | Mar 2004 | A1 |
20040049580 | Boyd et al. | Mar 2004 | A1 |
20040054610 | Amstutz et al. | Mar 2004 | A1 |
20040064552 | Chong et al. | Apr 2004 | A1 |
20040068517 | Scott | Apr 2004 | A1 |
20040088199 | Childress et al. | May 2004 | A1 |
20040103014 | Teegan et al. | May 2004 | A1 |
20040117759 | Ripped et al. | Jun 2004 | A1 |
20040122652 | Andrews et al. | Jun 2004 | A1 |
20040133416 | Fukuoka et al. | Jul 2004 | A1 |
20040133876 | Sproule | Jul 2004 | A1 |
20040139021 | Reed et al. | Jul 2004 | A1 |
20040145607 | Alderson | Jul 2004 | A1 |
20040147138 | Vaartstra | Jul 2004 | A1 |
20040148152 | Horikawa | Jul 2004 | A1 |
20040148586 | Gilboa | Jul 2004 | A1 |
20040162812 | Lane et al. | Aug 2004 | A1 |
20040162822 | Papanyan et al. | Aug 2004 | A1 |
20040167765 | Abu | Aug 2004 | A1 |
20040205672 | Bates et al. | Oct 2004 | A1 |
20040220792 | Gallanis et al. | Nov 2004 | A1 |
20040236566 | Simske | Nov 2004 | A1 |
20040243587 | Nuyens et al. | Dec 2004 | A1 |
20040268221 | Wang | Dec 2004 | A1 |
20040268299 | Lei et al. | Dec 2004 | A1 |
20050027563 | Fackler et al. | Feb 2005 | A1 |
20050027871 | Bradley et al. | Feb 2005 | A1 |
20050039191 | Hewson et al. | Feb 2005 | A1 |
20050044198 | Okitsu et al. | Feb 2005 | A1 |
20050050000 | Kwok et al. | Mar 2005 | A1 |
20050055330 | Britton et al. | Mar 2005 | A1 |
20050059566 | Brown et al. | Mar 2005 | A1 |
20050060372 | DeBettencourt et al. | Mar 2005 | A1 |
20050071211 | Flockhart et al. | Mar 2005 | A1 |
20050096959 | Kumar et al. | May 2005 | A1 |
20050104628 | Tanzawa et al. | May 2005 | A1 |
20050125683 | Matsuyama et al. | Jun 2005 | A1 |
20050132048 | Kogan et al. | Jun 2005 | A1 |
20050138162 | Byrnes | Jun 2005 | A1 |
20050144023 | Aboujaoude et al. | Jun 2005 | A1 |
20050165823 | Ondrusek et al. | Jul 2005 | A1 |
20050198021 | Wilcox et al. | Sep 2005 | A1 |
20050216235 | Butt et al. | Sep 2005 | A1 |
20050222889 | Lai et al. | Oct 2005 | A1 |
20050228875 | Monitzer et al. | Oct 2005 | A1 |
20050234882 | Bennett et al. | Oct 2005 | A1 |
20050267770 | Banavar et al. | Dec 2005 | A1 |
20050288920 | Green et al. | Dec 2005 | A1 |
20060004845 | Kristiansen et al. | Jan 2006 | A1 |
20060015388 | Flockhart et al. | Jan 2006 | A1 |
20060020783 | Fisher | Jan 2006 | A1 |
20060041861 | Trefler et al. | Feb 2006 | A1 |
20060053125 | Scott | Mar 2006 | A1 |
20060063138 | Loff et al. | Mar 2006 | A1 |
20060064486 | Baron et al. | Mar 2006 | A1 |
20060064667 | Freitas | Mar 2006 | A1 |
20060075360 | Bixler | Apr 2006 | A1 |
20060080082 | Ravindra et al. | Apr 2006 | A1 |
20060080401 | Gill et al. | Apr 2006 | A1 |
20060092467 | Dumitrescu et al. | May 2006 | A1 |
20060100847 | McEntee et al. | May 2006 | A1 |
20060101386 | Gerken et al. | May 2006 | A1 |
20060101393 | Gerken et al. | May 2006 | A1 |
20060106846 | Schulz et al. | May 2006 | A1 |
20060139312 | Sinclair et al. | Jun 2006 | A1 |
20060149751 | Jade et al. | Jul 2006 | A1 |
20060167655 | Barrow et al. | Jul 2006 | A1 |
20060173724 | Trefler et al. | Aug 2006 | A1 |
20060173871 | Taniguchi et al. | Aug 2006 | A1 |
20060206303 | Kohlmeier et al. | Sep 2006 | A1 |
20060206305 | Kimura et al. | Sep 2006 | A1 |
20060209085 | Wong et al. | Sep 2006 | A1 |
20060218166 | Myers et al. | Sep 2006 | A1 |
20060271559 | Stavrakos et al. | Nov 2006 | A1 |
20060271920 | Abouelsaadat | Nov 2006 | A1 |
20060288348 | Kawamoto et al. | Dec 2006 | A1 |
20070005623 | Self et al. | Jan 2007 | A1 |
20070010991 | Lei et al. | Jan 2007 | A1 |
20070028225 | Whittaker et al. | Feb 2007 | A1 |
20070038765 | Dunn | Feb 2007 | A1 |
20070055938 | Herring et al. | Mar 2007 | A1 |
20070061789 | Kaneko et al. | Mar 2007 | A1 |
20070094199 | Deshpande et al. | Apr 2007 | A1 |
20070100782 | Reed et al. | May 2007 | A1 |
20070118497 | Katoh | May 2007 | A1 |
20070130130 | Chan et al. | Jun 2007 | A1 |
20070136068 | Horvitz | Jun 2007 | A1 |
20070143163 | Weiss et al. | Jun 2007 | A1 |
20070143851 | Nicodemus et al. | Jun 2007 | A1 |
20070203756 | Sears et al. | Aug 2007 | A1 |
20070208553 | Hastings et al. | Sep 2007 | A1 |
20070226031 | Manson et al. | Sep 2007 | A1 |
20070233902 | Trefler et al. | Oct 2007 | A1 |
20070239646 | Trefler | Oct 2007 | A1 |
20070245300 | Chan et al. | Oct 2007 | A1 |
20070260584 | Marti et al. | Nov 2007 | A1 |
20070294644 | Yost | Dec 2007 | A1 |
20080002823 | Fama et al. | Jan 2008 | A1 |
20080046462 | Kaufman et al. | Feb 2008 | A1 |
20080077384 | Agapi et al. | Mar 2008 | A1 |
20080085502 | Allen et al. | Apr 2008 | A1 |
20080109467 | Brookins et al. | May 2008 | A1 |
20080120593 | Keren et al. | May 2008 | A1 |
20080163253 | Massmann et al. | Jul 2008 | A1 |
20080184230 | Leech et al. | Jul 2008 | A1 |
20080189679 | Rodriguez et al. | Aug 2008 | A1 |
20080195377 | Kato et al. | Aug 2008 | A1 |
20080196003 | Gerken et al. | Aug 2008 | A1 |
20080208785 | Trefler et al. | Aug 2008 | A1 |
20080216055 | Khatutsky | Sep 2008 | A1 |
20080216060 | Vargas | Sep 2008 | A1 |
20080263510 | Nerome et al. | Oct 2008 | A1 |
20090007084 | Conallen et al. | Jan 2009 | A1 |
20090018998 | Patten, Jr. et al. | Jan 2009 | A1 |
20090075634 | Sinclair et al. | Mar 2009 | A1 |
20090083697 | Zhang et al. | Mar 2009 | A1 |
20090132232 | Trefler | May 2009 | A1 |
20090132996 | Eldridge et al. | May 2009 | A1 |
20090138844 | Halberstadt et al. | May 2009 | A1 |
20090150541 | Georgis | Jun 2009 | A1 |
20090158407 | Nicodemus et al. | Jun 2009 | A1 |
20090164494 | Dodin | Jun 2009 | A1 |
20090171938 | Levin et al. | Jul 2009 | A1 |
20090199123 | Albertson et al. | Aug 2009 | A1 |
20090228786 | Danton et al. | Sep 2009 | A1 |
20090271708 | Peters et al. | Oct 2009 | A1 |
20090276206 | Fitzpatrick et al. | Nov 2009 | A1 |
20090282384 | Keppler | Nov 2009 | A1 |
20090319948 | Stannard et al. | Dec 2009 | A1 |
20100011338 | Lewis | Jan 2010 | A1 |
20100083135 | Zawacki et al. | Apr 2010 | A1 |
20100088266 | Trefler et al. | Apr 2010 | A1 |
20100107137 | Trefler et al. | Apr 2010 | A1 |
20100217737 | Shama | Aug 2010 | A1 |
20110066486 | Bassin et al. | Mar 2011 | A1 |
20110148791 | Luu | Jun 2011 | A1 |
20110239113 | Hung et al. | Sep 2011 | A1 |
20110252305 | Tschani et al. | Oct 2011 | A1 |
20110264251 | Copello et al. | Oct 2011 | A1 |
20120041921 | Canaday et al. | Feb 2012 | A1 |
20120102420 | Fukahori | Apr 2012 | A1 |
20120272186 | Kraut | Oct 2012 | A1 |
20120293558 | Dilts | Nov 2012 | A1 |
20130007267 | Khatutsky | Jan 2013 | A1 |
20130031455 | Griffiths et al. | Jan 2013 | A1 |
20130047165 | Goetz et al. | Feb 2013 | A1 |
20130067392 | Leonard et al. | Mar 2013 | A1 |
20130135294 | An | May 2013 | A1 |
20130159904 | Kelappan et al. | Jun 2013 | A1 |
20130167245 | Birtwhistle et al. | Jun 2013 | A1 |
20130179816 | Seo et al. | Jul 2013 | A1 |
20130231970 | Trefler et al. | Sep 2013 | A1 |
20130254833 | Nicodemus et al. | Sep 2013 | A1 |
20130290249 | Merriman et al. | Oct 2013 | A1 |
20140019400 | Trefler et al. | Jan 2014 | A1 |
20140089819 | Andler et al. | Mar 2014 | A1 |
20140125577 | Hoang et al. | May 2014 | A1 |
20140137019 | Paulsen et al. | May 2014 | A1 |
20140277164 | Ramsay et al. | Sep 2014 | A1 |
20150058772 | Bator et al. | Feb 2015 | A1 |
20150089406 | Trefler et al. | Mar 2015 | A1 |
20150127736 | Clinton et al. | May 2015 | A1 |
20160041961 | Romney | Feb 2016 | A1 |
20160062963 | Umapathy | Mar 2016 | A1 |
20160070560 | Chase | Mar 2016 | A1 |
20160085809 | de Castro Alves | Mar 2016 | A1 |
20160098298 | Trefler et al. | Apr 2016 | A1 |
20160105370 | Mellor et al. | Apr 2016 | A1 |
20170013073 | Mendez et al. | Jan 2017 | A1 |
20170255341 | Trefler et al. | Sep 2017 | A1 |
20170351425 | D'angelo et al. | Dec 2017 | A1 |
20170357703 | Theimer | Dec 2017 | A1 |
20180011678 | Shipper et al. | Jan 2018 | A1 |
20180024901 | Tankersley et al. | Jan 2018 | A1 |
Number | Date | Country |
---|---|---|
19911098 | Dec 1999 | DE |
0 549 208 | Jun 1993 | EP |
0 669 717 | Aug 1995 | EP |
0 996 916 | May 2000 | EP |
1 015 997 | Jul 2000 | EP |
1 019 807 | Jul 2000 | EP |
1 073 955 | Feb 2001 | EP |
1 073 992 | Feb 2001 | EP |
1 135 723 | Sep 2001 | EP |
1 163 604 | Dec 2001 | EP |
1 183 636 | Mar 2002 | EP |
1 196 882 | Apr 2002 | EP |
1 203 310 | May 2002 | EP |
1 208 482 | May 2002 | EP |
1 212 668 | Jun 2002 | EP |
1 240 592 | Sep 2002 | EP |
1 277 102 | Jan 2003 | EP |
1 277 119 | Jan 2003 | EP |
1 277 120 | Jan 2003 | EP |
1 277 153 | Jan 2003 | EP |
1 277 155 | Jan 2003 | EP |
1 277 329 | Jan 2003 | EP |
1 374 083 | Jan 2004 | EP |
1 382 030 | Jan 2004 | EP |
1 386 241 | Feb 2004 | EP |
1 393 172 | Mar 2004 | EP |
1 393 188 | Mar 2004 | EP |
1 402 336 | Mar 2004 | EP |
1 407 384 | Apr 2004 | EP |
1 430 396 | Jun 2004 | EP |
1 438 649 | Jul 2004 | EP |
1 438 654 | Jul 2004 | EP |
1 438 672 | Jul 2004 | EP |
1 483 685 | Dec 2004 | EP |
1 490 747 | Dec 2004 | EP |
1 490 809 | Dec 2004 | EP |
1 492 232 | Dec 2004 | EP |
1 782 183 | May 2007 | EP |
1 830 312 | Sep 2007 | EP |
1 840 803 | Oct 2007 | EP |
2 115 581 | Nov 2009 | EP |
9838564 | Sep 1998 | WO |
9840807 | Sep 1998 | WO |
9905632 | Feb 1999 | WO |
9945465 | Sep 1999 | WO |
9950784 | Oct 1999 | WO |
0033187 | Jun 2000 | WO |
0033217 | Jun 2000 | WO |
0033226 | Jun 2000 | WO |
0033235 | Jun 2000 | WO |
0033238 | Jun 2000 | WO |
03058504 | Jul 2000 | WO |
0052553 | Sep 2000 | WO |
0052603 | Sep 2000 | WO |
0067194 | Nov 2000 | WO |
02088869 | Nov 2000 | WO |
0140958 | Jun 2001 | WO |
0175610 | Oct 2001 | WO |
0175614 | Oct 2001 | WO |
0175747 | Oct 2001 | WO |
0175748 | Oct 2001 | WO |
0176206 | Oct 2001 | WO |
0177787 | Oct 2001 | WO |
0179994 | Oct 2001 | WO |
0221254 | Mar 2002 | WO |
0244947 | Jun 2002 | WO |
02056249 | Jul 2002 | WO |
02080006 | Oct 2002 | WO |
02080015 | Oct 2002 | WO |
02082300 | Oct 2002 | WO |
02084925 | Oct 2002 | WO |
02091346 | Nov 2002 | WO |
02101517 | Dec 2002 | WO |
2103576 | Dec 2002 | WO |
0219131 | Jan 2003 | WO |
03021393 | Mar 2003 | WO |
03029923 | Apr 2003 | WO |
03029955 | Apr 2003 | WO |
03030005 | Apr 2003 | WO |
03030013 | Apr 2003 | WO |
03030014 | Apr 2003 | WO |
03069500 | Jun 2003 | WO |
03071380 | Aug 2003 | WO |
03071388 | Aug 2003 | WO |
03073319 | Sep 2003 | WO |
03077139 | Sep 2003 | WO |
03085503 | Oct 2003 | WO |
03085580 | Oct 2003 | WO |
2004001613 | Dec 2003 | WO |
2004003766 | Jan 2004 | WO |
2004003885 | Jan 2004 | WO |
20041003684 | Jan 2004 | WO |
2004046882 | Jun 2004 | WO |
2004061815 | Jul 2004 | WO |
2004086198 | Oct 2004 | WO |
20041086197 | Oct 2004 | WO |
2004095207 | Nov 2004 | WO |
2004095208 | Nov 2004 | WO |
2004114147 | Dec 2004 | WO |
2005001627 | Jan 2005 | WO |
2005003888 | Jan 2005 | WO |
2005010645 | Feb 2005 | WO |
2005117549 | Dec 2005 | WO |
2006081536 | Aug 2006 | WO |
2007033922 | Mar 2007 | WO |
2008109441 | Sep 2008 | WO |
2009097384 | Aug 2009 | WO |
Entry |
---|
Eisenstein, et al., Adaptation in Automated User-Interface Design. IUI, 2000, pp. 74-81. |
Simpson, Alan, et al., Access 97 for Windows 95/NT; 1997 SYBEX; 16 pages; USPTO STIC-EIC 2100/2400. |
Ceri, S., et al., WIDE—A distributed architecture for workflow management. Proceedings. Seventh International Workshop on Research Issues in Data Engineering, IEEE, 1997, pp. 76-79, 1997. |
Deelman, E., et al., Workflows and e-science: An overview of workflow system features and capabilities. Future Generation Computer Systems, May 2009, vol. 25, issue 5, pp. 528-540. |
Kappel, G., et al., TriGSflow active object-oriented workflow management. Proceedings of the 28th Annual Hawaii International Conference on System Sciences. 1995, pp. 727-736. |
Manolescu, D.A., et al., Dynamic object model and adaptive workflow. Proceedings of Metadata and Active Object-Model Pattern Mining Workshop co-located with OOPSLA, 1999, vol. 99, 19 pages. |
Schulze, W., Fitting the workflow management facility into the object management architecture. Business Object Design and Implementation II. Springer London, 1998, pp. 109-117. |
Surjanto, B., XML content management based on object-relational database technology. Proceedings of the First International Conference on Web Information Systems Engineering, IEEE, 2000, Jun. 19-21, 2000, vol. 1, pp. 70-79. |
[No Author Listed] About the Integrated Work Manager (IWM). Pegasystems, Inc., Apr. 30, 2009, 3 pages, <http://pdn-dev/DevNet/PRPCv5/KB/TMP9ad01zumf.asp>. |
[No Author Listed] FreeBSD Project. “EDQUOTA(8)” in Free BSD System Manager's Manual. FreeBSD 8.2 Jun. 6, 1993. pp. 1-2. Retrieved from freebsd.org on Oct. 27, 2011. |
[No Author Listed] How SmartForms for Fair Blaze Advisor works, Fair Issac White Paper, http://www.FAIRISAAC.COM/, Oct. 31, 2005, 8 pages (website no longer active). |
[No Author Listed] How to Configure and Customize the Universal Worklist. SAP Netweaver '04 and SAP Enterprise Portal 6.0. SAP AG. Version 1, May 2004, 65 pages. <http://www.erpgenie.com/sap/netweaver/ep/Configuring%20the%20UWL.pdf>. |
[No Author Listed] How to configure the IWM/IAC gateway. Pegasystems, Inc., Apr. 30, 2009, 4 paegs, <http://pdn-dev/DevNet/PRPCv5/KB/TMP9cf8fzurq4.asp>. |
[No Author Listed] How to install the Integrated Work Manager (IWM). Pegasystems, Inc., Apr. 30, 2009, 6 pages, <http://pdn-dev/DevNet/PRPCv5/KB/TMP9br1ezurp8.asp>. |
[No Author Listed] HP Integrated Lights-Out 2, User Guide, Part No. 394326-004, HP, Aug. 2006, 189 pages. |
[No Author Listed] Integrating with External Systems, PegaRULES Process Commander 5.2. Process Commander 5.2 reference. Pegasystems Inc, Cambridge, MA, 2006, 103 pages <http://pdn.pega.com/ProductSupport/Products/PegaRULESProcessCommander/documents/PRPC/V5/502/iwes/PRPC52_Integrating_with_External_Systems.pdf>. |
[No Author Listed] IP Prior Art Database, Options when returning work items in workflow management systems. IBM, IPCOM000027980D, 2004, 3 pages. |
[No Author Listed] IP Prior Art Database, Staff Queries and Assignments in Workflow Systems. IBM, IPCOM000142382D, 2006, 4 pages. |
[No Author Listed] IP Prior Art Database, Using work items to manage user interactions with adaptive business services. IBM TDB, IPCOM000015953D, 2003, 4 pages. |
[No Author Listed] Localizing an Application, PegaRULES Process Commander. Process Commander 4.2 reference. Pegasystems Inc., Cambdrige, MA, 2006, 92 pages <http://pdn.pega.com/DevNet/PRPCv4/TechnologyPapers/documents/Localization0402.pdf>. |
[No Author Listed] Oracle Universal Work Queue: Implementation Guide. Release 11i for Windows NT. Oracle Corporation. Jul. 2001, 136 pages. <http://docs.oracle.com/cd/A85964_01/acrobat/eu115ug.pdf>. |
[No Author Listed] Solaris 9 resource manager software. A technical white paper. Sun Microsystems, Inc., Palo Alto CA, 2002, 37 pages. XP-002291080. Retrieved Aug. 3, 2004 from <http://wwws.sun.com/software/whitepapers/solaris9/srm.pdf>. |
Bertino, E., and P. Foscoli, Index Organizations for Object-Oriented Database Systems, IEEE Trans. on Knowledge and Data Engineering, 7(2):193-209 (Apr. 1995). |
Bierbaum, A., et al., VR juggler: A virtual platform for virtual reality application development. Proceedings of the Virtual Reality 2001 Conference, IEEE, 2001, 8 pages, <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=8tarnumber-913774>. |
Breiman, L., Bagging predictors, Machine Learning, vol. 24, No. 2, Aug. 31, 1996, pp. 123-140, Kluwer Academic Publishers, Netherlands. |
Brusilovsky, P., and De Bra, P., Editors, “Second Workshop on Adaptive Hypertext and Hypermedia Proceedings,” Jun. 20-24, 1998. Ninth ACM Conference on Hypertext and Hypermedia, Hypertext'98. pp. 1-2. |
Burleson, D., Adding behaviors to relational databases, DBMS, 8(10): 68(5) (Sep. 1995). |
Busse, Ralph et al., Declarative and Procedural Object Oriented Views, IEEE, 1998, pp. 570-578, retrieved Mar. 22, 2007. |
Buyya et al., Economic Models for Resource Management and Scheduling in Grid Computing, Concurrency and Computation: Practice and Experience, 2002, vol. 14, pp. 1507-1542. |
Chan, T.W., and W. Hwang, Towards Integrating Logic, Object, Frame, and Production, Proc. Fourth Int'l. Conf. on Software Engineering and Knowledge Engineering, IEEE, pp. 463-469, Jun. 1992. |
Cheng, Cheng-Chung; Smith, Stephen F.; A Constraint Satisfaction Approach to Makespan Scheduling, ATPS 1996 Proceedings, pp. 45-52 (1996). |
Cheng, C.C. and Smith, Applying Constraint Satisfaction Techniques to Job Shop Scheduling, Annals of Operations Research, 70: 327-357 (1997). |
Cochrane, Roberta et al., Integrating Triggers and Declarative Constraints in SQL Database Systems, pp. 567-578, Proceedings of the 22nd VLDB Conference Mumbai (Bombay), India, 1996, retrieved Mar. 22, 2007. |
Damerau, F.J., Problems and some solutions in customization of natural language database front ends. ACM Transactions on Information Systems, vol. 3, No. 2, Apr. 1, 1985, pp. 165-184. |
Danforth, S., Integrating Object and Relational Technologies, Proc. Sixteenth Annual Int'l. Computer Software and Applications Conf., IEEE Comput. Soc. Press, pp. 225-226, Sep. 1992 (Abstract). |
Deelman, E., et al., Pegasus: A framework for mapping complex scientific workflows onto distributed systems, submitted to Scientific Programming, Jan. 2005. Pre journal publication article, 22 pages. |
Deelman, E., et al., Pegasus: A framework for mapping complex scientific workflows onto distributed systems. Scientific Programming, 13, pp. 219-237, 2005. |
DeMichiel, LG., et al., Polyglot: Extensions to Relational Databases for Sharable Types and Functions in a Multi-Language Environment, Proc. Ninth Int'l. Conf. on Data Engineering, IEEE, pp. 651-660, Apr. 1993. |
Devarakonda et al., Predictability of process resource usage: A measurement-based study on UNIX. IEEE Transactions on Software Engineering. 1989;15(12)1579-1586. |
Communication for European Patent Application No. 05755530.2, dated Sep. 6, 2007 (2 pages). |
European Search Report for Application No. 05755530.2, dated Mar. 26, 2012 (3 Pages). |
European Office Action dated Aug. 31, 2012 for Application No. 05755530.2 (4 Pages). |
Communication for European Patent Application No. 07250844.3 enclosing European Search Report, dated Jul. 11, 2007 (6 pages). |
Communication for European Patent Application No. 07250844.3, dated Mar. 28, 2008 (1 page). |
European Office Action dated Jul. 9, 2012 for Application No. 07250844.3 (8 Pages). |
Communication for European Patent Application No. 07250848.4, dated Aug. 13, 2007 (EESR enclosed) (6 pages). |
Communication for European Patent Application No. 07250848.4, dated May 29, 2008 (1 page). |
Communication for European Patent Application No. 08731127.0, dated Oct. 13, 2009 (2 pages). |
Extended European Search Report dated Oct. 29, 2012 for Application No. 08731127.0 (8 Pages). |
Extended European Search Report for Application No. 15189385.6, dated Dec. 17, 2015 (8 pages). |
Fayad, M.E., et al., Object-oriented application frameworks. Communications of the ACM, Oct. 1997, vol. 40, issue 10, pp. 32-38, <http://dl.acm.org/citation.cfm?id=262798>. |
Francisco, S., et al. Rule-Based Web Page Generation, Proceedings of the 2nd Workshop on Adaptive Hypertext and Hypermedia, Hypertext'98, Jun. 20-24, 1998, 9 pages. |
Gajos et al. SUPPLE: Automatically Generating User Interfaces. IUI 2004, 8 pages. |
Hague, Darren, Universal Worklist with SAP Netweaver Portal. Galileo Press, 2008, pp. 11-31. <http://www.sap-hefte.de/download/dateien/1461/146_leseprobe.pdf>. |
International Search Report and Written Opinion for Application No. PCT/GB2004/000677, dated Aug. 2, 2004 (15 pages). |
International Search Report for Application No. PCT/US2004/020783, dated Nov. 8, 2005 (2 pages). |
International Preliminary Report on Patentability for Application No. PCT/US2004/020783, dated Feb. 13, 2006 (6 pages). |
International Search Report for PCT/US05/018599, dated May 15, 2007 (1 page). |
[NoAuthorListed] XPath Tutorial, Mar. 18, 2005, 7 pages (Year: 2005). |
Summons to Attend Oral Proceedings pursuant to rule 115(1) EPC, issued May 2, 2018 for Application No. 08731127.0 (8 pages). |
U.S. Appl. No. 08/666,165, filed Jun. 19, 1996, Rules Bases and Methods of Access Thereof. |
U.S. Appl. No. 10/430,693, filed May 6, 2003, Methods and Apparatus for Digital Data Processing With Mutable Inheritance. |
U.S. Appl. No. 10/547,014, filed Aug. 25, 2005, Classification Using Probability Estimate Re-Sampling. |
U.S. Appl. No. 10/639,735, filed Aug. 12, 2003, Process/Viewer Interface. |
U.S. Appl. No. 10/854,017, filed May 26, 2004, Integration of Declarative Rule-Based Processing With Procedural Programming. |
U.S. Appl. No. 11/046,211, filed Jan. 28, 2005, Methods and Apparatus for Work Management and Routing. |
U.S. Appl. No. 11/203,513, filed Aug. 12, 2005, Methods and Apparatus for Digital Data Processing With Mutable Inheritance. |
U.S. Appl. No. 11/368,360, filed Mar. 3, 2006, Rules Base Systems and Methods With Circumstance Translation. |
U.S. Appl. No. 11/396,415, filed Mar. 30, 2006, User Interface Methods and Apparatus for Rules Processing. |
U.S. Appl. No. 11/681,269, filed Mar. 2, 2007, Proactive Performance Management for Multi-User Enterprise Software Systems. |
U.S. Appl. No. 12/035,682, filed Feb. 22, 2008, User Interface Methods and Apparatus for Rules Processing. |
U.S. Appl. No. 12/174,624, filed Jul. 16, 2008, Methods and Apparatus for Implementing Multilingual Software Applications. |
U.S. Appl. No. 12/381,523, filed Mar. 12, 2009, Techniques for Dynamic Data Processing. |
U.S. Appl. No. 12/386,959, filed Apr. 24, 2009, Method and Apparatus for Integrated Work Management. |
U.S. Appl. No. 12/590,454, filed Nov. 6, 2009, Techniques for Content-Based Caching in a Computer System. |
U.S. Appl. No. 12/619,215, filed Nov. 16, 2009, Rules Base Systems and Methods With Circumstance Translation. |
U.S. Appl. No. 12/649,095, filed Dec. 29, 2009, Methods and Apparatus for Integration of Declarative Rule-Based Processing With Procedural Programming in a Digital Data-Processing Evironment. |
U.S. Appl. No. 12/798,161, filed Mar. 30, 2010, System and Method for Creation and Modification of Software Applications. |
U.S. Appl. No. 13/031,097, filed Feb. 18, 2011, Systems and Methods for Distributed Rules Processing. |
U.S. Appl. No. 13/031,109, filed Feb. 18, 2011, Rule-Based User Interface Conformance Methods. |
U.S. Appl. No. 13/341,411, filed Dec. 30, 2011, System and Method for Updating or Modifying an Application Without Manual Coding. |
U.S. Appl. No. 13/536,079, filed Jun. 28, 2012, Proactive Performance Management for Multi-User Enterprise Software Systems. |
U.S. Appl. No. 13/718,255, filed Dec. 18, 2012, Methods and Apparatus for Work Management and Routing. |
U.S. Appl. No. 13/892,956, filed May 13, 2013, Content-Based Caching Using a Content Identifier at a Point in Time. |
U.S. Appl. No. 13/897,763, filed May 20, 2013, System and Software for Creation and Modification of Software. |
U.S. Appl. No. 13/907,287, filed May 31, 2013, Methods and Apparatus for Integration of Declarative Rule-Based Processing With Procedural Programming in a Digital Data-Processing Environment. |
U.S. Appl. No. 14/469,208, filed Aug. 26, 2014, Techniques for Dynamic Data Processing. |
U.S. Appl. No. 14/527,348, filed Oct. 29, 2014, Systems and Methods for Distributed Rules Processing. |
U.S. Appl. No. 14/558,084, filed Dec. 2, 2014, Methods and Apparatus for User Interface Optimization. |
U.S. Appl. No. 14/597,207, filed Jan. 14, 2015, Methods and Apparatus for Integrated Work Management. |
U.S. Appl. No. 14/928,085, filed Oct. 30, 2015, System and Method for Updating or Modifying an Application Without Manual Coding. |
International Preliminary Report on Patentability for PCT/US2005/018599, dated Jun. 5, 2007 (10 pages). |
International Search Report & Written Opinion for PCT/US06/03160, dated Jul. 21, 2008 (16 pages). |
International Preliminary Report on Patentability for PCT/US06/03160, dated Apr. 9, 2009 (14 pages). |
International Search Report for PCT/US013/55503, dated Jul. 28, 2008 (1 page). |
International Preliminary Report on Patentability for PCT/US2008/055503, dated Sep. 17, 2009 (4 pages). |
International Search Report & Written Opinion for PCT/US09/32341, dated Mar. 11, 2009 (14 pages). |
International Preliminary Report on Patentability for PCT/US2009/032341, dated Aug. 12, 2010 (8 pages). |
Johnson et al., Sharing and resuing rules-a feature comparison of five expert system shells. IEEE Expert, IEEE Services Center, New York, NY, vol. 9, No. 3, Jun. 1, 1994, pp. 3-17. |
Jones et al., A user-centered approach to functions in excel. International Conference on Functional Programming, Uppsala, Jun. 30, 2003, pp. 1-12. |
Kim, W., Object-Oriented Databases: Definition and Research Directions, IEEE Trans. on Knowledge and Data Engineering, vol. 2(3) pp. 327-341, Sep. 1990. |
Kuhn, H.W., The Hungarian Method for the Assignment Problem, Naval Research Logistics Quarterly, 2 (1955), pp. 83-97. |
Kuno, H.A., and E.A. Rundensteiner, Augmented Inherited Multi-Index Structure for Maintenance of Materialized Path Query Views, Proc. Sixth Int'l. Workshop on Research Issues in Data Engineering, pp. 128-137, Feb. 1996. |
LaRue, J., Leveraging Integration and Workflow. Integrated Solutions, Accounting Today, SourceMedia, Aug. 2006, pp. 18-19. |
Lippert, Eric, Fabulous Adventures in Coding: Metaprogramming, Toast and the Future of Development Tools, Microsoft.com Blog, MSDN Home, published Mar. 4, 2004, 6 pgs. |
Mandal, et al., Integrating existing scientific workflow systems: The kepler/pegasus example. USC Information Sciences Institute, 2007, 8 pages. |
Manghi, Paolo, et. al., Hybrid Applications Over XML: Integrating the Procedural and Declarative Approaches, 2002 ACM, pp. 1-6. Retrieved Mar. 22, 2007. |
Markiewicz, M.E., et al., Object oriented framework development ACM, 2001, 13 pages, <http://dl.acm.org/citation.cfm?id=372771>. |
Markowitz, V.M., and A. Shoshani, Object Queries over Relational Databases: Language, Implementation, and Applications, IEEE Xplore, pp. 71-80, Apr. 1993. |
Marmel, Elaine, Microsoft Office Project 2007 Bible, ISBN 0470009926, Wiley Publishing, Inc., 2007, 961 pages. |
Maryanski, F., et al., The Data Model Compiler a Tool for Generating Object-Oriented Database Systems, 1986 Int'l. Workshop on Object-Oriented Database Systems, IEEE, 73-84 (1986). |
McConnell, Steven C., Brooks' Law Repealed, IEEE Software, pp. 6-9, Nov./Dec. 1999. |
Mecca, G., et al., Cut and Paste, ACM, pp. 1-25 and Appendix I-IV (Jun. 1999). Retrieved Mar. 22, 2007. |
Mitchell, T.M., Machine Learning, Chapter 3, 1997, McGraw-Hill, pp. 52-80. |
Mitchell, T.M., Machine Learning, Chapter 6, 1997, McGraw-Hill, pp. 154-200. |
Morizet-Mahoudeaux, P., A Hierarchy of Network-Based Knowledge Systems, IEEE Trans. on Systems, Man, and Cybemetics, vol. 21(5), pp. 1184-1191, Sep./Oct. 1991. |
Pientka, B., et al., Programming with proofs and explicit contexts. International Symposium on Principles and Practice of Declarative Programming, ACM, 2008, pp. 163-173, <http://delivery.acm.org/10.1145/1390000/1389469/p163-pientka.pdf?>. |
Reinersten, Don, Is It Always a Bad Idea to Add Resources to a Late Project?, Oct. 30, 2000. Electronic Design. vol. 48, Issue 22, p. 70. |
Riccuiti, M., Oracle 8.0 on the way with objects: upgrade will also build in multidimensional engine. InfoWorld. Sep. 25, 1995;17(39):16. |
Richner, T., et al., Recovering high-level views of object-oriented applications from static and dynamic information. IEEE, 1999, 10 pages, <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=792487>. |
Salvini, S., and M.H. Williams, Knowledge Management for Expert Systems, IEEE Colloquium on ‘Knowledge Engineering’, 3 pages, May 1990. |
Schiefelbein, Mark, A Backbase Ajax Front-end for J2EE Applications, Internet Article, dev2dev <http://dev2dev.bea.com/1pt/a/433>, Aug. 29, 2005, 16 pages. |
Sellis, T., et al., Coupling Production Systems and Database Systems: A Homogeneous Approach, IEEE Trans. on Knowledge and Data Engineering, vol. 5(2), pp. 240-256, Apr. 1993. |
Shyy Y.M., and S.Y.W. Su, Refinement Preservation for Rule Selection in Active Object-Oriented Database Systems, Proc. Fourth Int'l. Workshop on Research Issues in Data Engineering, pp. 115-123, Feb. 1994. |
Singh, G., et al., Workflow task clustering for best effort systems with pegasus, Pegasus, 2008, 8 pages. |
Smedley, T.J. et al., “Expanding the Utility of Spreadsheets Through the Integration of Visual Programming and User Interface Objects,” School of Computer Science, Technical University of Nova Scotia, ACM, 1996; pp. 148-155. |
Srinivasan, V., et al., Object persistence in object-oriented applications. IBM Systems Journal, 1997, vol. 36, issue 1, pp. 66-87, <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber-5387186>. |
Stonebraker, M., The Integration of Rule Systems and Database Systems, IEEE Trans. on Knowledge and Data Engineering, vol. 4(5), pp. 415-423, Oct. 1992. |
Sun, et al., “Supporting Inheritance in Relational Database Systems,” IEEE, pp. 511-518, Jun. 1992. |
Thuraisingham, “From Rules to Frames and Frames to Rules,” Al Expert, pp. 31-39, Oct. 1989. |
Vranes, S., et al., Integrating Multiple Paradigms within the Blackboard Framework, IEEE Transactions on Software Engineering, vol. 21, No. 3, Mar. 1995, pp. 244-262. |
Yang, Bibo; Geunes, Joseph; O'Brien, William J.; Resource-Constrained Project Scheduling: Past Work and New Directions, Apr. 2001, 28 pages, Research Report 2001-2006, Department of Industrial and Systems Engineering, University of Florida. |
U.S. Appl. No. 15/613,439, filed Jun. 5, 2017, Connecting Graphical Shapes Using Gestures. |
U.S. Appl. No. 15/206,956, filed Jul. 11, 2016, Selective Sharing for Collaborative Application Usage. |
Number | Date | Country | |
---|---|---|---|
20160105370 A1 | Apr 2016 | US |
Number | Date | Country | |
---|---|---|---|
62062515 | Oct 2014 | US |