Computer data distribution architecture

Abstract
Described are methods, systems and computer readable media for computer data distribution architecture.
Description

Embodiments relate generally to computer data systems, and more particularly, to methods, systems and computer readable media for computer data distribution architecture.


Some conventional computer data systems may maintain data in one or more data sources that may include data objects such as tables. These conventional systems may include clients that independently access tables from each data source directly. In such data systems, a need may exist to provide systems and methods for an optimized composite table data service providing flexible data routing and caching across the various data sources for one or more clients, in order to reduce memory usage and to enable redundancy, high-availability, scalability, and rule-based data discovery.


Embodiments were conceived in light of the above mentioned needs, problems and/or limitations, among other things.


Some implementations can include a memory-efficient and processor-efficient computer system for reliable implementation of a table data cache proxy. The system can comprise a plurality of data server computers each being programmed with a table data service accessible via an electronic messaging protocol. The system can also comprise a table data cache proxy enabled (TDCP-enabled) server computer coupled to the one or more data server computers, the TDCP-enabled server computer having a plurality of TDCP clients, the TDCP-enabled server computer caching data from the plurality of data server computers and aggregating subscriptions of the TDCP clients to the plurality of data server computers. The TDCP-enabled server computer can comprise a cache memory device having at least a portion being a shared memory portion. The TDCP-enabled server computer can also comprise one or more hardware processors and a computer readable data storage device coupled to the one or more hardware processors, the computer readable data storage device having stored thereon software instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations. The operations can include receiving, from a first TDCP client of the plurality of TDCP clients, a first electronic message requesting table data. The operations can also include determining whether a shared memory cache stored in the shared memory portion of the cache memory device contains a cached copy of the requested table data.


The operations can further include, when the shared memory cache contains a cached copy of the requested table data, transmitting, to the first TDCP client, one or more second electronic messages comprising a reference indicating a location where the cached copy is stored in the shared memory portion of the cache memory device in response to the first electronic message, the data received from at least one of the plurality of data server computers and stored in the cache memory device being authoritative due to a data model of the plurality of data server computers. When the shared memory cache does not contain a cached copy of the requested table data, the operations can include selecting one or more data server computers from the plurality of data server computers as one or more appropriate data server computers to provide the requested table data. When the shared memory cache does not contain a cached copy of the requested table data, the operations can also include transmitting one or more third electronic messages to the one or more appropriate data server computers requesting the requested table data.


When the shared memory cache does not contain a cached copy of the requested table data, the operations can further include receiving one or more fourth electronic messages from the one or more appropriate data server computers in response to the third electronic messages. When the shared memory cache does not contain a cached copy of the requested table data, the operations can also include filtering the received fourth electronic messages and storing a result of the filtering in the shared memory portion of the cache memory device. When the shared memory cache does not contain a cached copy of the requested table data, the operations can additionally include transmitting, to the first TDCP client computer, one or more fifth electronic messages comprising a reference indicating a location where the result of the filtering is stored in the shared memory portion of the cache memory device in response to the first electronic message.


When the shared memory cache does not contain a cached copy of the requested table data, the operations can further include transmitting a request to subscribe for updates from the one or more appropriate data server computers; the operations can also include receiving a subscription update from one or more of the plurality of data server computers; and the operations can further include electronically distributing the received subscription update to one or more of the plurality of TDCP clients.


Some implementations can include a memory-efficient and processor-efficient computer system for reliable implementation of a table data cache proxy. The system can comprise at least one data server having a table data service accessible via an electronic messaging protocol. The system can also comprise a table data cache proxy enabled (TDCP-enabled) server coupled to the one or more data servers, the TDCP-enabled server caching data from the plurality of data servers and aggregating subscriptions to the plurality of data servers. The TDCP-enabled server can comprise one or more hardware processors and a computer readable data storage device coupled to the one or more hardware processors, the computer readable data storage device having stored thereon software instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations.


The operations can include receiving a first electronic message requesting table data. The operations can also include determining whether a cache of the TDCP-enabled server contains a cached copy of the requested table data. The operations can further include, when the cache of the TDCP-enabled server contains a cached copy of the requested table data, transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data received from at least one of the plurality of data servers and stored in the cache being authoritative due to a data model of the plurality of data servers.


When the cache of the TDCP-enabled server does not contain a cached copy of the requested data: the operations can include determining one or more appropriate data servers of the plurality of data servers to request the requested table data from; the operations can also include transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested table data; the operations can further include receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages; the operations can also include filtering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested table data in response to the first electronic message.


The first electronic message can be received via an inter process communication (IPC) mechanism from a client, the one or more second electronic messages can be transmitted to the client via the IPC mechanism, and the one or more fifth electronic messages can be transmitted to the client via the IPC mechanism.


The TDCP-enabled server can comprise a shared memory in which at least a portion of the cache is stored. The one or more second electronic messages can comprise an indication where the cached copy is stored in the shared memory of the TDCP-enabled server, and the one or more fifth electronic messages can comprise indications where the requested table data is stored in the shared memory of the TDCP-enabled server.


Additionally or alternatively, first electronic message can be received via a network from a remote client, the one or more second electronic messages can be transmitted to the remote client via the network, and the one or more fifth electronic messages can be transmitted to the remote client via the network.


In any of the above-mentioned implementations, the operations can also include, when the first electronic message requests a block of binary data, determining whether a future request for another block of binary data is likely based on a number (and/or pattern) of received requests, and, when the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.


Some implementations can include a memory-efficient and processor-efficient computerized method for reliable implementation of a table data cache proxy. The method can include receiving, at a table data cache proxy enabled (TDCP-enabled) server, a first electronic message requesting table data, the TDCP-enabled server being coupled to one or more data servers having a table data service accessible via an electronic messaging protocol, the TDCP-enabled server caching data from the one or more data servers and aggregating subscriptions to the one or more data servers. The method can also include determining whether a cache of the TDCP-enabled server contains a cached copy of the requested table data. The method can further include, when the cache of the TDCP-enabled server contains a cached copy of the requested table data, transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data received from at least one of the one or more data servers and stored in the cache being authoritative due to a data model of the one or more data servers.


When the cache of the TDCP-enabled server does not contain a cached copy of the requested data: the method can include determining one or more appropriate data servers of the one or more data servers to request the requested table data from; the method can also include transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested table data; the method can further include receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages; and the method can also include filtering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested table data in response to the first electronic message.


The method can include determining whether a future request for another block of binary data is likely based on a number (and/or pattern) of received requests; and when the future request is determined to be likely, the method can include prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.


When the cache of the TDCP-enabled server does not contain a cached copy of the requested data: the method can include transmitting a request to subscribe for updates from the one or more appropriate data servers; the method can also include receiving a subscription update from one or more of the one or more data servers; and the method can further include electronically distributing the received subscription update to one or more subscribers of the TDCP-enabled server.


Some implementations can include a nontransitory computer readable medium having stored thereon software instructions that, when executed by one or more processors, cause the one or more processors to perform operations. The operations can include receiving, at a table data cache proxy enabled (TDCP-enabled) server, a first electronic message requesting table data, the TDCP-enabled server being coupled to one or more data servers having a table data service accessible via an electronic messaging protocol, the TDCP-enabled server caching data from the one or more data servers and aggregating subscriptions to the one or more data servers. The operations can also include determining whether a cache of the TDCP-enabled server contains a cached copy of the requested table data. When the cache of the TDCP-enabled server contains a cached copy of the requested table data, the operations can include transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data received from at least one of the one or more data servers and stored in the cache being authoritative due to a data model of the one or more data servers. When the cache of the TDCP-enabled server does not contain a cached copy of the requested data: the operations can include determining one or more appropriate data servers of the one or more data servers to request the requested table data from; the operations can also include transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested table data; the operations can further include receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages; and the operations can also include filtering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested table data in response to the first electronic message.


The operations can include determining whether a future request for another block of binary data is likely based on a number of received requests, and when the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.


In any of the above-mentioned implementations, the first electronic message can request data from a data block and comprise a minimum response length, and when additional data above the minimum response length is available within the block, the additional data is included in response to the first electronic message to prevent redundant subsequent requests.


In any of the above-mentioned implementations, the first electronic message can be one of: a table location discovery request requesting a list of table locations for a given table key, the table key uniquely identifying a table; a table location metadata retrieval request requesting table-location-level metadata including the size of a table; a column location metadata retrieval request requesting column-location-level metadata, wherein column data is laid out in an optimized data layout to enable high performance data access, the column-location-level metadata including information indicating the optimized data layout; a column file size retrieval request requesting the size of a column file; and a column file data retrieval request requesting at least a portion of a block of binary data from a column file.


In any of the above-mentioned implementations, the at least one data server can be a plurality of data servers comprising at least two data sources providing the same data; and the TDCP-enabled server can be configured to reconnect to any of the data servers, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.


In any of the above-mentioned implementations, the one or more data servers can comprise a plurality of data servers, and, when the first electronic message requests one or more data locations, the filtering can comprise combining at least a portion of each of two or more of the one or more fourth electronic messages to create a composite of data from two or more of the plurality of data servers as the result of the filtering.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a diagram of an example computer data system showing an example data distribution configuration in accordance with some implementations.



FIG. 2 is a diagram of an example computer data system showing an example administration/process control arrangement in accordance with some implementations.



FIG. 3A is a diagram of an example query server host in accordance with some implementations.



FIG. 3B is a diagram of an example query server host in accordance with some implementations.



FIG. 4 is a diagram of an example computer data system and network 400 showing an example data distribution configuration in accordance with some implementations.



FIG. 5 is a flowchart of an example method of processing a TDCP composite table data service request in accordance with some implementations.



FIG. 6 is a flowchart of an example method of processing a table location discovery request by a TDCP server in accordance with some implementations.



FIG. 7 is a flowchart of an example method of processing a table location metadata retrieval request by a TDCP server in accordance with some implementations.



FIG. 8 is a flowchart of an example method of processing a column location metadata retrieval request by a TDCP server in accordance with some implementations.



FIG. 9 is a flowchart of an example method of processing a column file size retrieval request by a TDCP server in accordance with some implementations.



FIG. 10 is a flowchart of an example method of processing a column file data retrieval request by a TDCP server in accordance with some implementations.



FIG. 11 is a diagram of an example computing device configured for table data cache proxy (TDCP) processing in accordance with at least one implementation.





DETAILED DESCRIPTION

Reference may be made herein to the Java programming language, Java classes, Java bytecode and the Java Virtual Machine (JVM) for purposes of illustrating example implementations. It will be appreciated that implementations can include other programming languages (e.g., groovy, Scala, R, Go, etc.), other programming language structures as an alternative to or in addition to Java classes (e.g., other language classes, objects, data structures, program units, code portions, script portions, etc.), other types of bytecode, object code and/or executable code, and/or other virtual machines or hardware implemented machines configured to execute a data system query.



FIG. 1 is a diagram of an example computer data system and network 100 showing an example data distribution configuration in accordance with some implementations. In particular, the system 100 includes an application host 102, a periodic data import host 104, a query server host 106, a long-term file server 108, and a user data import host 110. While tables are used as an example data object in the description below, it will be appreciated that the data system described herein can also process other data objects such as mathematical objects (e.g., a singular value decomposition of values in a given range of one or more rows and columns of a table), TableMap objects, etc. A TableMap object provides the ability to lookup a Table by some key. This key represents a unique value (or unique tuple of values) from the columns aggregated on in a byExternal( ) statement execution, for example. A TableMap object is can be the result of a byExternal( ) statement executed as part of a query. It will also be appreciated that the configurations shown in FIGS. 1 and 2 are for illustration purposes and in a given implementation each data pool (or data store) may be directly attached or may be managed by a file server.


The application host 102 can include one or more application processes 112, one or more log files 114 (e.g., sequential, row-oriented log files), one or more data log tailers 116 and a multicast key-value publisher 118. The periodic data import host 104 can include a local table data server, direct or remote connection to a periodic table data store 122 (e.g., a column-oriented table data store) and a data import server 120. The query server host 106 can include a multicast key-value subscriber 126, a performance table logger 128, local table data store 130 and one or more remote query processors (132, 134) each accessing one or more respective tables (136, 138). The long-term file server 108 can include a long-term data store 140. The user data import host 110 can include a remote user table server 142 and a user table data store 144. Row-oriented log files and column-oriented table data stores are discussed herein for illustration purposes and are not intended to be limiting. It will be appreciated that log files and/or data stores may be configured in other ways. In general, any data stores discussed herein could be configured in a manner suitable for a contemplated implementation.


In operation, the input data application process 112 can be configured to receive input data from a source (e.g., a securities trading data source), apply schema-specified, generated code to format the logged data as it's being prepared for output to the log file 114 and store the received data in the sequential, row-oriented log file 114 via an optional data logging process. In some implementations, the data logging process can include a daemon, or background process task, that is configured to log raw input data received from the application process 112 to the sequential, row-oriented log files on disk and/or a shared memory queue (e.g., for sending data to the multicast publisher 118). Logging raw input data to log files can additionally serve to provide a backup copy of data that can be used in the event that downstream processing of the input data is halted or interrupted or otherwise becomes unreliable.


A data log tailer 116 can be configured to access the sequential, row-oriented log file(s) 114 to retrieve input data logged by the data logging process. In some implementations, the data log tailer 116 can be configured to perform strict byte reading and transmission (e.g., to the data import server 120). The data import server 120 can be configured to store the input data into one or more corresponding data stores such as the periodic table data store 122 in a column-oriented configuration. The periodic table data store 122 can be used to store data that is being received within a time period (e.g., a minute, an hour, a day, etc.) and which may be later processed and stored in a data store of the long-term file server 108. For example, the periodic table data store 122 can include a plurality of data servers configured to store periodic securities trading data according to one or more characteristics of the data (e.g., a data value such as security symbol, the data source such as a given trading exchange, etc.).


The data import server 120 can be configured to receive and store data into the periodic table data store 122 in such a way as to provide a consistent data presentation to other parts of the system. Providing/ensuring consistent data in this context can include, for example, recording logged data to a disk or memory, ensuring rows presented externally are available for consistent reading (e.g., to help ensure that if the system has part of a record, the system has all of the record without any errors), and preserving the order of records from a given data source. If data is presented to clients, such as a remote query processor (132, 134), then the data may be persisted in some fashion (e.g., written to disk).


The local table data server 124 can be configured to retrieve data stored in the periodic table data store 122 and provide the retrieved data to one or more remote query processors (132, 134) via an optional proxy (e.g., table data cache proxy (TDCP) 394 and/or 404 as shown in FIG. 3 and FIG. 4, respectively). Remote query processors (132, 134) can also receive data from DIS 120 and/or LTDS 124 via the proxy.


The remote user table server (RUTS) 142 can include a centralized consistent data writer, as well as a data server that provides processors with consistent access to the data that it is responsible for managing. For example, users can provide input to the system by writing table data that is then consumed by query processors.


The remote query processors (132, 134) can use data from the data import server 120, local table data server 124 and/or from the long-term file server 108 to perform queries. The remote query processors (132, 134) can also receive data from the multicast key-value subscriber 126, which receives data from the multicast key-value publisher 118 in the application host 102. The performance table logger 128 can log performance information about each remote query processor and its respective queries into a local table data store 130. Further, the remote query processors can also read data from the RUTS, from local table data written by the performance logger, or from user table data read over NFS, for example.


It will be appreciated that the configuration shown in FIG. 1 is a typical example configuration that may be somewhat idealized for illustration purposes. An actual configuration may include one or more of each server and/or host type. The hosts/servers shown in FIG. 1 (e.g., 102-110, 120, 124 and 142) may each be separate or two or more servers may be combined into one or more combined server systems. Data stores can include local/remote, shared/isolated and/or redundant. Any table data may flow through optional proxies indicated by an asterisk on certain connections to the remote query processors (e.g., table data cache proxy (TDCP) 392 or 404 as shown in FIG. 3B and FIG. 4, respectively). Also, it will be appreciated that the term “periodic” is being used for illustration purposes and can include, but is not limited to, data that has been received within a given time period (e.g., millisecond, second, minute, hour, day, week, month, year, etc.) and which has not yet been stored to a long-term data store (e.g., 140).



FIG. 2 is a diagram of an example computer data system 200 showing an example administration/process control arrangement in accordance with some implementations. The system 200 includes a production client host 202, a controller host 204, a GUI host or workstation 206, and query server hosts 208 and 210. It will be appreciated that there may be one or more of each of 202-210 in a given implementation.


The production client host 202 can include a batch query application 212 (e.g., a query that is executed from a command line interface or the like) and a real time query data consumer process 214 (e.g., an application that connects to and listens to tables created from the execution of a separate query). The batch query application 212 and the real time query data consumer 214 can connect to a remote query dispatcher 222 and one or more remote query processors (224, 226) within the query server host 1208.


The controller host 204 can include a persistent query controller 216 configured to connect to a remote query dispatcher 232 and one or more remote query processors 228-230. In some implementations, the persistent query controller 216 can serve as the “primary client” for persistent queries and can request remote query processors from dispatchers, and send instructions to start persistent queries. For example, a user can submit a query to 216, and 216 starts and runs the query every day. In another example, a securities trading strategy could be a persistent query. The persistent query controller can start the trading strategy query every morning before the market opened, for instance. It will be appreciated that 216 can work on times other than days. In some implementations, the controller may require its own clients to request that queries be started, stopped, etc. This can be done manually, or by scheduled (e.g., cron jobs). Some implementations can include “advanced scheduling” (e.g., auto-start/stop/restart, time-based repeat, etc.) within the controller.


The GUI/host workstation can include a user console 218 and a user query application 220. The user console 218 can be configured to connect to the persistent query controller 216. The user query application 220 can be configured to connect to one or more remote query dispatchers (e.g., 232) and one or more remote query processors (228, 230).



FIG. 3A is a diagram of an example query server host 320 (e.g., as described at 208, 210, and/or 106) in accordance with at least one embodiment. Query server host 320 can include a processor 324, a high speed memory (e.g., RAM) 326, another high speed memory (e.g., shared RAM with RQP and TDP) 336. Query server host 320 can access a medium access speed memory 346 (e.g., RAM managed by another host (actual or virtual) such as, for example, an intraday server (e.g., DIS 120 or LTDS 124)) and a slow access speed storage 355 (e.g., a file server with hard drive such as, for example, long term file server 108).


In operation, processor 324 can execute remote query processor 322 which stores/accesses data in high speed memory 326, high speed memory 336, medium access speed memory 346, and slow access speed storage 354. High speed memory 326 and high speed memory 336 can be memory on the same or different memory devices.


High speed memory 326 can contain one or more query update graphs 328, one or more table indexes 330, in memory data 332, and recent data cache 334. High speed memory 326 can request and retrieve data from one or more slow access speed storages 355 and/or from high speed memory 336.


High speed memory 336 can be memory that is shared with one or more remote query processors 322 and/or one or more table data cache proxies (e.g., TDCP 392 and 404, as shown in FIG. 3 and FIG. 4 respectively). High speed memory 336 can contain one or more data columns, for example, a symbol column data 338, a date column data 340, a time column data 342, and a quote column data 344. High speed memory 336 can exchange data with remote query processor 322, high speed memory 326, and/or medium access speed memory 346, and can request and receive data from slow access speed storage 355.


Medium access speed memory 346 can contain one or more data columns, for example, symbol column data 348, a date column data 350, a time column data 352, and a quote column data 354. Medium access speed memory 346 can exchange data with high speed memory 336 and can transmit data to a slow access speed storage 355. In some embodiments, medium access speed memory 346 is RAM that resides on a remote host, administered by a remote process (e.g., DIS 120, LTDS 124, or RUTS 142).


Slow access speed storage 355, for example, a file server with one or more hard drives, can contain persistent column data, for example, a symbol column 358, a date column 360, a time column 362, and a quote column 364. The one or more persisted column data 358-364 can be copied into medium speed solid state storage 356, for example, flash, to provide faster access for more frequently accessed data. In some embodiments, slow access speed storage 355 is used by long-term file server 108.


In some embodiments, remote query processor 322 can access a table having column data of two or more columns of the table stored in different memory devices and/or data servers. In some such embodiments, column data of a column (i.e., different rows or groups of rows) can also be stored in different memory devices and/or data servers. Processor 322 can store/access identifiers indicating where column data of the columns/rows is stored. Processor 322 can also store/access one or more table indexes associated with the table that identify for that table the valid portion(s) of the data referenced by the location identifier (e.g., the portions of the data which correspond to the rows of that table). For example, in some embodiments, a first portion of column data can be stored in shared memory with TDCP 336, a second portion of column data can be stored in medium speed memory 346, and a third portion of column data can be stored in slow speed storage 355, for the same or different columns of a table.



FIG. 3B is a diagram of an example query server host 370 (e.g., as described at 320, 208, 210, and/or 106) in accordance with at least one embodiment. Query server host 370 can contain one or more remote query processors (372, 374, 376) associated with one or more table data cache proxy clients (378, 380, 382), a shared memory 384 (e.g., as described at 336) that can exchange data (386, 388, 390) with table data cache proxy clients (378, 380, 382), and one or more table data cache proxies 392 that can exchange data with shared memory 384.



FIG. 4 is a diagram of an example computer data system and network 400 showing an example data distribution configuration in accordance with some implementations. System 400 includes local table data server (LTDS) 124, data import server (DIS) 120, remote user table server (RUTS) 142, and a host 402 (e.g., query server host 106, 208, 210, 320, and/or 370). Host 402 includes table data cache proxy (TDCP) 404, one or more remote query processors (RQP) 412/414, and one or more TDCP clients 408/410. TDCP includes a cache 406. In operation, each RQP 412/414 transmits data to and receives data from TDCP 404 via a corresponding one of TDCP clients 408/410, and TDCP 404 transmits data to and receives data from each of LTDS 124, DIS 120, and RUTS 142.


In some embodiments, TDCP 404 exports a composite table data service composed of multiple filtered remote table data services. In some embodiments, each of data sources 120, 124, and 142 exports a table data service via a messaging protocol and TDCP 404 is configured to export a composite table data service composed of the table data services of the data sources via a messaging protocol. The composite table data service of TDCP 404 can be composed of the services of data sources 120, 124, and 142 that are filtered and/or combined at a table-location level. A given “location” for a table may be provided by a single, non-composite service, but a client-visible table might be composed of locations from multiple underlying sources and the composite table data service of TDCP 404 provides data from the multiple underlying sources to a client (e.g., RQP 412/414 via TDCP clients 408/410), filtered and/or combined as appropriate.


In some embodiments, TDCP 404 is coupled to multiple of one or more of the different types of data sources 120, 124, and 142. In some such embodiments, the multiple data sources can provide the same data. The composite table data service of TDCP 404 can be configured to reconnect to any of the data sources that can provide the same data, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.


In some embodiments, LTDS 124 provides access to un-merged real-time data from previous time periods (e.g., from the same storage systems to which current-period data is persisted by DIS 120). In some such embodiments, LTDS 124 can be used as a stop-gap (i.e., alternate data source) when issues prevent timely completion of the merge operations that transform, validate, and promote periodic data.


In some embodiments, the messaging protocol is identical for data sources 120, 124, and 142. In some such embodiments, the messaging protocol for TDCP 404 is identical to that of the data sources. In some embodiments, the messaging protocol of the data sources 120, 124, and 142 and/or the messaging protocol of TDCP 404 is built on top of one or more other networking protocols (e.g., TCP/IP).


In some embodiments, TDCP 404 is configured to serve as an aggregator of subscriptions by RQP 412/414 to the same metadata updates and/or requests for individual data and/or metadata items. TDCP 404 is configured to cache data and/or metadata received from data servers 120, 124, and 142 in cache 406.


In some embodiments, cache 406 comprises data blocks indexed by {{namespace, table name, table type}, {internal partition, column partition}, {column name, column file type, offset} } or, more generally, {table key}, {table location key}, {column block key}. Data blocks in cache 406 may be evicted (e.g., based on a modified LRU policy) at any time when they are not actively in-use. If more data becomes available (e.g., as implied by table size change notifications), the missing region of an already-cached data block can be read when a client requests the block in question. Successful data block read requests provide at least as much data as requested, but may provide more if the block has grown and the source can provide an additional suffix.


In some embodiments, cache 406 is indexed per-source. Each bottom-level table data service uses either the filesystem (which may be remote or local) or at most one actively-connected server, and maintains its own index of which location data it has. All caches within a given process generally share the same pool of free/managed data block memory-space, but this is configurable. The engine code deals with table locations, which have location-level metadata, per-column location-level metadata, and buffer stores associated with each relevant column file. The buffer store handles block indexing and presents the engine with byte-oriented access, which is then translated to cell-oriented access by an intermediate level of the engine code.


In embodiments, system 400 is configured such that TDCP 404 is authoritative for data stored in cache 406 from at least one of data sources 120, 124, and 142. Cache 406 includes data requested by the clients of TDCP 404, and different TDCPs with different attached clients/workloads will have different caches. Additionally, because of the underlying model for data updates used by data sources such as 120 and 124 in some embodiments (repeatable reads by virtue of append-only data changes, in the real-time system), if cache 406 has a data block from one of those data sources, it has the correct (i.e., authoritative) data for that data block. If it has a partial data block, the range that it has is correct data for that range. TDCP 404 therefore doesn't have to worry about having its data invalidated by upstream changes.


In some embodiments, connections between data sources 120, 124, and 142 and TDCP 404 are authenticated (e.g., using ACL permissions) and/or encrypted.


In some embodiments, TDCP 404 can communicate with TDCP clients 408-410 via an inter-process communication (IPC) mechanism (e.g., sockets, shared memory, memory mapped files, message passing, pipes, and/or message queues). For example, cache 406 can be stored in shared memory (e.g., shared memory 384 as shown in FIG. 3B). In some embodiments, the shared memory is System V IPC shared memory accessed with the various “shm_system” calls. In some embodiments, mixed IPC mechanisms may be used. For example, TDCP 404 can transmit data to one or more TDCP clients on a different host (actual or virtual) via one IPC mechanism (e.g., a socket/network) and provide data to one or more different TDCP clients via a different IPC mechanism (e.g., shared memory). In another example, TDCP 404 can communicate with data sources 120/124/142 using an IPC mechanism that is the same or different than that used for communications between TDCP 404 and TDCP clients 408-410.


Although not shown, in some embodiments, TDCP clients 408-410 and RQP 412-414 can be on a separate host (actual or virtual). In some such embodiments, data can be transmitted between TDCP 404 and TDCP clients 408-410 via a network.


In some embodiments, data may be transmitted between TDCP 404, data sources 120/124/142, TDCP clients 408-410, and RQP 412-414 using single and/or multipart messages.


In some embodiments, TDCP 404 and/or TDCP clients 408-410 maintain a separate cache (or a separate portion of the cache) for each RQP 412-414. In other embodiments, TDCP 404 and/or TDCP clients 408-410 can maintain separate and/or shared caches for RQP 412-414 (e.g., sharing a cache between two or more RQPs while maintaining a separate cache for a different RQP, or sharing one cache for all RQPs).



FIG. 5 is a flowchart of an example method 500 of processing a TDCP composite table data service request in accordance with some implementations. Processing begins at 502, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests table (data and/or metadata) from a TDCP server (e.g., as shown in FIG. 3B or FIG. 4). Processing continues to 504.


At 504, the TDCP determines whether the requested table data/metadata is in the TDCP local state (or cache such as, for example, shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 510; otherwise, processing continues to 506.


At 506 the TDCP requests the latest state and subscribes to updates for the table from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4). In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 502. In some embodiments, the one or more appropriate data servers can be selected based on a table type indicated in the request received at 502 (e.g., by determining which of the data servers match the table type and selecting those that match). In some embodiments, additional user tables can come from RUTS 142 and one more “user” filesystems mounted over NFS, performance log tables can be from locally-mounted filesystems, all of which may have independent services that can be composed by the TDCP and selected by the TDCP as one of the appropriate data servers.


At 508. the appropriate data servers distribute latest table state to the TDCP. The latest table state is received in response to the request(s) made by the TDCP at 506 and is stored in the TDCP cache. Processing continues to 510.


At 510, the TDCP filters/distributes the latest table state to the RQP. The latest table state can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative). Filtering can include removing metadata changes that aren't important to certain classes of downstream consumers (e.g., some TDCPs or RQPs might only want size changes, and/or might not want modifications after a certain time of day, and/or might not need to see all metadata fields). Additionally or alternatively, the TDCP can combine data related to two or more “downstream” table locations (with the same keys), and the filtering can include coalescing changes and only advertising them when (for example) a threshold (more than one, or all) number of the data sources being combined had advertised the same thing, thereby increasing the likelihood that data can be re-read promptly in the event of a subset of the data sources crashing or becoming partitioned away.


In some embodiments, the TDCP can be configured to optimize the performance of computer data access when multiple data servers each provide access to the same data by requesting the latest table state from two or more of the multiple data servers at 506 and by filtering the received/cached table state to generate a complete, non-duplicative composite table state that includes data/metadata from two or more of the multiple data servers. In such embodiments, the table state can include table locations and the TDCP can filter table locations to be distributed to the RQP to interleave table locations from the multiple data servers, thereby distributing subsequent data access across the multiple data servers. Processing continues to 512.


At 512, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 514.


At 514, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates. In some embodiments, the updates can be filtered as described above at 510 before being distributed to subscribing RQP.


It will be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 502-514 may be repeated in whole or in part. For example, 512-514 may be repeated to continuously provide table location updates to the subscribing RQP.



FIG. 6 is a flowchart of an example method 600 of processing a table location discovery request by a TDCP server (e.g., as shown in FIG. 3B or FIG. 4) in accordance with some implementations. Processing begins at 602, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests table locations for a given table key from the TDCP. The table key can comprise a namespace, a table name, and a table type (e.g. user/system, periodic/historical). In some embodiments, table locations are keyed by path information such as, for example, internal partition and column partition. Processing continues to 604.


At 604, the TDCP determines whether the requested data is in the TDCP local state (or cache such as, e.g., shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 610; otherwise, processing continues to 606.


At 606 the TDCP requests table locations for the given table key from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4) and optionally subscribes for updates. In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 602. In some embodiments, the one or more appropriate data servers can be selected based on the table type indicated by the table key (e.g., by determining which of the data servers match the table type and selecting those that match).


At 608. the TDCP receives table locations. The table locations are received in response to the request(s) made by the TDCP at 606 and are stored in the TDCP cache. Processing continues to 610.


At 610, the TDCP filters/distributes the table locations to the RQP. The table locations can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative). In some embodiments, the TDCP can be configured to optimize the performance of computer data access when multiple data servers each provide access to the same data by requesting table locations from two or more of the multiple data servers at 606 and by filtering the received/cached table locations to generate a complete, non-duplicative set of table locations that includes locations from two or more of the multiple data servers. In such embodiments, the TDCP can, for example, filter the table locations to be distributed to the RQP to interleave table locations from the multiple data servers, thereby distributing subsequent data access across the multiple data servers.


Additionally or alternatively, the TDCP can combine data received from data servers that provide access to the same data (e.g., by combining different portions to generate a complete, non-duplicative set as discussed above, by combining all data, or by including data received from one of the data servers and excluding data received from the other data servers), and the filtering can include coalescing changes and only advertising them when (for example) a threshold (more than one, or all) number of the data sources being combined had advertised the same thing, thereby increasing the likelihood that data can be re-read promptly in the event of a subset of the data sources crashing or becoming partitioned away. Processing continues to 612.


At 612, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 614.


At 614, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates to table locations. In some embodiments, the updates can be filtered as described above at 610 before being distributed to subscribing RQP.


It will be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 602-614 may be repeated in whole or in part. For example, 612-614 may be repeated to continuously provide table location updates to the subscribing RQP.



FIG. 7 is a flowchart of an example method 700 of processing a table location metadata retrieval request by a TDCP server (e.g., as shown in FIG. 3B or FIG. 4) in accordance with some implementations. Processing begins at 702, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests table location metadata for a given table from the TDCP. The table location metadata can comprise size, modification time, validation status and validation completion time (e.g., validation being a process of ensuring that the data has passed proper quality checks), schema version used to generate the data, code version used to generate the data, user identifying information, and other metadata. Table location metadata can also include an “is valid” flag or an “is finished” flag to indicate that the data has been validated (e.g., that the data has passed proper quality checks). Processing continues to 704.


At 704, the TDCP determines whether the requested table location metadata is in the TDCP local state (or cache such as, e.g., shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 710; otherwise, processing continues to 706.


At 706 the TDCP requests table location metadata from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4) and subscribes for updates. In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 702. In some embodiments, the one or more appropriate data servers can be selected based on a table type and/or table location indicated in the request received at 702 (e.g., by determining which of the data servers match the table type and/or table location and selecting those that match).


At 708. the TDCP receives table location metadata. The table location metadata is received in response to the request(s) made by the TDCP at 706 and are stored in the TDCP cache. Processing continues to 710.


At 710, the TDCP filters/distributes the table location metadata to the RQP. The table location metadata can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative). Filtering can include removing metadata changes that aren't important to certain classes of downstream consumers (e.g., some TDCPs or RQPs might only want size changes, and/or might not want modifications after a certain time of day, and/or might not need to see all metadata fields). Additionally or alternatively, the TDCP can combine data related to two or more “downstream” table locations (with the same keys), and the filtering can include coalescing changes and only advertising them when (for example) a threshold (more than one, or all) number of the data sources being combined had advertised the same thing, thereby increasing the likelihood that data can be re-read promptly in the event of a subset of the data sources crashing or becoming partitioned away. Processing continues to 712.


At 712, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 714.


At 714, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates to table location metadata. In some embodiments, the updates can be filtered as described above at 710 before being distributed to subscribing RQP.


It will also be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 702-714 may be repeated in whole or in part. For example, 712-714 may be repeated to continuously provide table location metadata updates to the subscribing RQP (e.g. sending updates to table size as it changes).



FIG. 8 is a flowchart of an example method 800 of processing a column location metadata retrieval request by a TDCP server (e.g., as shown in FIG. 3B or FIG. 4) in accordance with some implementations. Processing begins at 802, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests column location metadata (e.g., grouping information, periodic/historical). Processing continues to 804.


At 804, the TDCP determines whether the requested column location metadata is in the TDCP local state (or cache such as, e.g., shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 810; otherwise, processing continues to 806.


At 806 the TDCP requests column location metadata from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4) and optionally subscribes for updates. For example, a subscriber could subscribe to receive updates to value indexes included in column metadata (e.g., for real-time/periodic/intraday data). In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 802. In some embodiments, the one or more appropriate data servers can be selected based on the table type indicated by a table type and/or table location indicated in the request received at 802 (e.g., by determining which of the data servers match the table type and/or table location and selecting those that match).


At 808. the TDCP receives column location metadata. The column location metadata is received in response to the request(s) made by the TDCP at 806 and are stored in the TDCP cache. Processing continues to 810.


At 810, the TDCP filters/distributes the column location metadata to the RQP. The column location metadata can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative). Column location metadata can also be filtered to eliminate updates of a nature not needed/requested by certain downstream consumers (e.g., grouping/indexing changes if the RQP/query doesn't use them). In some embodiments, filtering at 810 can include removing metadata changes that aren't important to certain classes of downstream consumers (e.g., some TDCPs or RQPs might only want size changes, and/or might not want modifications after a certain time of day, and/or might not need to see all metadata fields). Additionally or alternatively, the TDCP can combine data related to two or more “downstream” table locations (with the same keys), and filtering at 810 can include coalescing changes and only advertising them when (for example) a threshold (more than one, or all) number of the data sources being combined had advertised the same thing, thereby increasing the likelihood that data can be re-read promptly in the event of a subset of the data sources crashing or becoming partitioned away. Processing continues to 812.


At 812, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 814.


At 814, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates to column location metadata. In some embodiments, the updates can be filtered as described above at 810 before being distributed to subscribing RQP.


It will be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 802-814 may be repeated in whole or in part. For example, 812-814 may be repeated to continuously provide column location metadata updates to the subscribing RQP (e.g. sending updated column size as it changes)



FIG. 9 is a flowchart of an example method 900 of processing a column file metadata retrieval request by a TDCP server (e.g., as shown in FIG. 3B or FIG. 4) in accordance with some implementations. Processing begins at 902, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests column file metadata which can include column file size (e.g., for column files that aren't an exact multiple of table location size) from the TDCP. Processing continues to 904.


At 904, the TDCP determines whether the requested column file metadata is in the TDCP local state (or cache such as, e.g., shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 910; otherwise, processing continues to 906.


At 906 the TDCP requests column file metadata from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4) and optionally subscribes for updates. In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 902. In some embodiments, the one or more appropriate data servers can be selected based on a column and/or column location/file indicated in the request received at 902 (e.g., by determining which of the data servers match the column and/or column location/file).


At 908. the TDCP receives column file metadata. The column file metadata is received in response to the request(s) made by the TDCP at 906 and are stored in the TDCP cache. Processing continues to 910.


At 910, the TDCP filters/distributes the column file metadata to the RQP. The column file metadata can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative). Column file metadata can be filtered, for example, to throttle the rate of change notification to the same frequency as other notifications such as table metadata (e.g., size) change notifications. Column file metadata can also be filtered to eliminate updates of a nature not needed/requested by certain downstream consumers. In some embodiments, filtering at 910 can include removing metadata changes that aren't important to certain classes of downstream consumers (e.g., some TDCPs or RQPs might not want modifications after a certain time of day, and/or might not need to see all metadata fields). Additionally or alternatively, the TDCP can combine data related to two or more “downstream” table locations (with the same keys), and filtering at 910 can include coalescing changes and only advertising them when (for example) a threshold (more than one, or all) number of the data sources being combined had advertised the same thing, thereby increasing the likelihood that data can be re-read promptly in the event of a subset of the data sources crashing or becoming partitioned away. Processing continues to 912.


At 912, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 914.


At 914, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates to column file metadata. In some embodiments, the updates can be filtered as described above at 910 before being distributed to subscribing RQP.


It will be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 902-914 may be repeated in whole or in part. For example, 912-914 may be repeated to continuously provide column file size updates to the subscribing RQP (e.g. sending updated column size as it changes).



FIG. 10 is a flowchart of an example method 1000 of processing a column file data retrieval request by a TDCP server (e.g., as shown in FIG. 3B or FIG. 4) in accordance with some implementations. Processing begins at 1002, where a remote query processor (RQP) (e.g., remote query processors 132-134, 322, 372-376, or 412-414 shown in FIG. 1, FIG. 3A, FIG. 3B, and FIG. 4, respectively) requests column file data from the TDCP. The client, and each intermediating service that cannot satisfy the request out of cache, requests the data for a block of binary data. The request includes the block size (which may be standardized at the service level), starting offset of the block within the column file, starting offset desired within the block, and minimum result length. More data (e.g., up to the maximum result length=block size−starting offset within the block) may be retrieved if available to prevent redundant subsequent requests. Processing continues to 1004.


At 1004, the TDCP determines whether the column file data is in the TDCP local state (or cache such as, e.g., shared RAM 336, shared memory 384, or cache 406 shown in FIG. 3A, FIG. 3B, and FIG. 4, respectively). If so, processing continues to 1010; otherwise, processing continues to 1006.


At 1006 the TDCP requests column file data from one or more appropriate data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142 shown in FIG. 1 and FIG. 4) and optionally subscribes for updates. In some embodiments, the TDCP is coupled to multiple data servers (e.g., LTDS 124, DIS 120, and/or RUTS 142) and the TDCP selects one or more of the multiple data servers as the appropriate data servers for the request received at 1002. In some embodiments, the one or more appropriate data servers can be selected based on the column file and/or the type of table that the column is a part of (e.g., by determining which of the data servers match the table type and selecting those that match). In some embodiments, when the TDCP has only a portion of the requested data in cache the TDCP can request whatever sub-range it doesn't have, and each request might actually get more data than they asked for in order to amortize away subsequent requests.


At 1008. the TDCP receives column file data. The column file data is received in response to the request(s) made by the TDCP at 1006 and are stored in the TDCP cache. Processing continues to 1010.


At 1010, the TDCP filters responses and/or distributes the column file data to the RQP. The column file data can be filtered based on rules such as, for example, rules defining where data should come from (e.g., which data source is authoritative when the data was requested from two or more sources). In some embodiments, the TDCP can be configured to optimize the performance of computer data access when multiple data servers each provide access to the same data by requesting different portions of the requested column file data from two or more of the multiple data servers at 606 and by filtering the responses to combine the different portions of column file data received from the multiple data servers into the column file data to be distributed to the RQP. In such embodiments, the TDCP can, for example, split the request across the multiple data servers, thereby distributing data access across the multiple data servers. In some embodiments, the TDCP or RQP may determine that requests for column file data follow a pattern (e.g., sequential access, striding, etc.) and prefetch one or more additional blocks or ranges of column file data from the appropriate data sources in order to enhance system performance by decreasing perceived latency and/or amortizing request costs. Processing continues to 1012.


At 1012, the TDCP listens to table updates from the data servers. Listening to table updates can include listening to table updates from data servers to which the TDCP has subscribed for updates. Processing continues to 1014.


At 1014, the TDCP distributes updates to subscribing RQP. The TDCP can operate as an aggregator of subscriptions for multiple RQP and upon receiving an update the TDCP can distribute the update to the subscribing RQP. For example, although not shown, two different RQP can subscribe to the TDCP to receive updates. In some embodiments, the updates can be filtered as described above at 1010 before being distributed to subscribing RQP.


It will be appreciated that, although not shown, the subscribing RQP can cancel their subscription to stop receiving updates from the TDCP, that all subscriptions are cancelled for an RQP that disconnects, and that the TDCP may cancel its own data subscriptions and/or discard data it no longer needs for any RQP.


It will also be appreciated that 1002-1014 may be repeated in whole or in part. For example, 1012-1014 may be repeated to continuously provide table location updates to the subscribing RQP.



FIG. 11 is a diagram of an example computing device 300 configured for table data cache proxy (TDCP) processing in accordance with at least one implementation. The computing device 300 includes one or more processors 302, operating system 304, computer readable medium 306 and network interface 308. The memory 306 can include table data cache proxy (TDCP) application 310 and a data section 312 (e.g., for storing caches, index data structures, column source maps, etc.).


In operation, the processor 302 may execute the application 310 stored in the memory 306. The application 310 can include software instructions that, when executed by the processor, cause the processor to perform operations for table data cache proxy processing in accordance with the present disclosure (e.g., performing one or more of 502-514, 602-614, 702-714, 802-814, 902-914, and/or 1002-1014 described above).


The application program 310 can operate in conjunction with the data section 312 and the operating system 304.


Although references have been made herein to tables and table data, it will be appreciated that the disclosed systems and methods can be applied with various computer data objects to, for example, provide flexible data routing and caching for such objects in accordance with the disclosed subject matter. For example, references herein to tables can include a collection of objects generally, and tables can include column types that are not limited to scalar values and can include complex types (e.g., objects).


It will be appreciated that the modules, processes, systems, and sections described above can be implemented in hardware, hardware programmed by software, software instructions stored on a nontransitory computer readable medium or a combination of the above. A system as described above, for example, can include a processor configured to execute a sequence of programmed instructions stored on a nontransitory computer readable medium. For example, the processor can include, but not be limited to, a personal computer or workstation or other such computing system that includes a processor, microprocessor, microcontroller device, or is comprised of control logic including integrated circuits such as, for example, an Application Specific Integrated Circuit (ASIC), a field programmable gate array (FPGA), a graphics processing unit (e.g., GPGPU or GPU) or the like. The instructions can be compiled from source code instructions provided in accordance with a programming language such as Java, C, C++, C#.net, assembly or the like. The instructions can also comprise code and data objects provided in accordance with, for example, the Visual Basic™ language, a specialized database query language, or another structured or object-oriented programming language. The sequence of programmed instructions, or programmable logic device configuration software, and data associated therewith can be stored in a nontransitory computer-readable medium such as a computer memory or storage device which may be any suitable memory apparatus, such as, but not limited to ROM, PROM, EEPROM, RAM, flash memory, disk drive and the like.


Furthermore, the modules, processes systems, and sections can be implemented as a single processor or as a distributed processor. Further, it should be appreciated that the steps mentioned above may be performed on a single or distributed processor (single and/or multi-core, or cloud computing system). Also, the processes, system components, modules, and sub-modules described in the various figures of and for embodiments above may be distributed across multiple computers or systems or may be co-located in a single processor or system. Example structural embodiment alternatives suitable for implementing the modules, sections, systems, means, or processes described herein are provided below.


The modules, processors or systems described above can be implemented as a programmed general purpose computer, an electronic device programmed with microcode, a hard-wired analog logic circuit, software stored on a computer-readable medium or signal, an optical computing device, a networked system of electronic and/or optical devices, a special purpose computing device, an integrated circuit device, a semiconductor chip, and/or a software module or object stored on a computer-readable medium or signal, for example.


Embodiments of the method and system (or their sub-components or modules), may be implemented on a general-purpose computer, a special-purpose computer, a programmed microprocessor or microcontroller and peripheral integrated circuit element, an ASIC or other integrated circuit, a digital signal processor, a hardwired electronic or logic circuit such as a discrete element circuit, a programmed logic circuit such as a PLD, PLA, FPGA, PAL, GP, GPU, or the like. In general, any processor capable of implementing the functions or steps described herein can be used to implement embodiments of the method, system, or a computer program product (software program stored on a nontransitory computer readable medium).


Furthermore, embodiments of the disclosed method, system, and computer program product (or software instructions stored on a nontransitory computer readable medium) may be readily implemented, fully or partially, in software using, for example, object or object-oriented software development environments that provide portable source code that can be used on a variety of computer platforms. Alternatively, embodiments of the disclosed method, system, and computer program product can be implemented partially or fully in hardware using, for example, standard logic circuits or a VLSI design. Other hardware or software can be used to implement embodiments depending on the speed and/or efficiency requirements of the systems, the particular function, and/or particular software or hardware system, microprocessor, or microcomputer being utilized. Embodiments of the method, system, and computer program product can be implemented in hardware and/or software using any known or later developed systems or structures, devices and/or software by those of ordinary skill in the applicable art from the function description provided herein and with a general basic knowledge of the software engineering and computer networking arts.


Moreover, embodiments of the disclosed method, system, and computer readable media (or computer program product) can be implemented in software executed on a programmed general purpose computer, a special purpose computer, a microprocessor, or the like.


It is, therefore, apparent that there is provided, in accordance with the various embodiments disclosed herein, methods, systems and computer readable media for computer data distribution architecture.


Application Ser. No. 15/154,974, entitled “DATA PARTITIONING AND ORDERING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,975, entitled “COMPUTER DATA SYSTEM DATA SOURCE REFRESHING USING AN UPDATE PROPAGATION GRAPH” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,979, entitled “COMPUTER DATA SYSTEM POSITION-INDEX MAPPING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,980, entitled “SYSTEM PERFORMANCE LOGGING OF COMPLEX REMOTE QUERY PROCESSOR QUERY OPERATIONS” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,983, entitled “DISTRIBUTED AND OPTIMIZED GARBAGE COLLECTION OF REMOTE AND EXPORTED TABLE HANDLE LINKS TO UPDATE PROPAGATION GRAPH NODES” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,984, entitled “COMPUTER DATA SYSTEM CURRENT ROW POSITION QUERY LANGUAGE CONSTRUCT AND ARRAY PROCESSING QUERY LANGUAGE CONSTRUCTS” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,985, entitled “PARSING AND COMPILING DATA SYSTEM QUERIES” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,987, entitled “DYNAMIC FILTER PROCESSING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,988, entitled “DYNAMIC JOIN PROCESSING USING REAL-TIME MERGED NOTIFICATION LISTENER” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,990, entitled “DYNAMIC TABLE INDEX MAPPING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,991, entitled “QUERY TASK PROCESSING BASED ON MEMORY ALLOCATION AND PERFORMANCE CRITERIA” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,993, entitled “A MEMORY-EFFICIENT COMPUTER SYSTEM FOR DYNAMIC UPDATING OF JOIN PROCESSING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,995, entitled “QUERY DISPATCH AND EXECUTION ARCHITECTURE” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,996, entitled “COMPUTER DATA DISTRIBUTION ARCHITECTURE” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,997, entitled “DYNAMIC UPDATING OF QUERY RESULT DISPLAYS” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,998, entitled “DYNAMIC CODE LOADING” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/154,999, entitled “IMPORTATION, PRESENTATION, AND PERSISTENT STORAGE OF DATA” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,001, entitled “COMPUTER DATA DISTRIBUTION ARCHITECTURE” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,005, entitled “PERSISTENT QUERY DISPATCH AND EXECUTION ARCHITECTURE” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,006, entitled “SINGLE INPUT GRAPHICAL USER INTERFACE CONTROL ELEMENT AND METHOD” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,007, entitled “GRAPHICAL USER INTERFACE DISPLAY EFFECTS FOR A COMPUTER DISPLAY SCREEN” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,009, entitled “COMPUTER ASSISTED COMPLETION OF HYPERLINK COMMAND SEGMENTS” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,010, entitled “HISTORICAL DATA REPLAY UTILIZING A COMPUTER SYSTEM” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,011, entitled “DATA STORE ACCESS PERMISSION SYSTEM WITH INTERLEAVED APPLICATION OF DEFERRED ACCESS CONTROL FILTERS” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


Application Ser. No. 15/155,012, entitled “REMOTE DATA OBJECT PUBLISHING/SUBSCRIBING SYSTEM HAVING A MULTICAST KEY-VALUE PROTOCOL” and filed in the United States Patent and Trademark Office on May 14, 2016, is hereby incorporated by reference herein in its entirety as if fully set forth herein.


While the disclosed subject matter has been described in conjunction with a number of embodiments, it is evident that many alternatives, modifications and variations would be, or are, apparent to those of ordinary skill in the applicable arts. Accordingly, Applicants intend to embrace all such alternatives, modifications, equivalents and variations that are within the spirit and scope of the disclosed subject matter.

Claims
  • 1. A memory-efficient and processor-efficient computer system for reliable implementation of a table data cache proxy, the system comprising: a plurality of data server computers each being programmed with a table data service accessible via an electronic messaging protocol;a table data cache proxy enabled (TDCP-enabled) server computer coupled to the one or more data server computers, the TDCP-enabled server computer having a plurality of TDCP clients, the TDCP-enabled server computer caching data from the plurality of data server computers and aggregating subscriptions of the TDCP clients to the plurality of data server computers;the TDCP-enabled server computer comprising: one or more hardware processors;a cache memory device having at least a portion being a shared memory portion; anda computer readable data storage device coupled to the one or more hardware processors, the computer readable data storage device having stored thereon software instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations including: receiving, from a first TDCP client of the plurality of TDCP clients, a first electronic message requesting table data;determining whether a shared memory cache stored in the shared memory portion of the cache memory device contains a cached copy of the requested table data;when the shared memory cache contains a cached copy of the requested table data, transmitting, to the first TDCP client, one or more second electronic messages comprising a reference indicating a location where the cached copy is stored in the shared memory portion of the cache memory device in response to the first electronic message, the data received from at least one of the plurality of data server computers and stored in the cache memory device being authoritative due to a data model of the plurality of data server computers;when the shared memory cache does not contain a cached copy of the requested table data: selecting one or more data server computers from the plurality of data server computers as one or more appropriate data server computers to provide the requested table data,transmitting one or more third electronic messages to the one or more appropriate data server computers requesting the requested table data,receiving one or more fourth electronic messages from the one or more appropriate data server computers in response to the third electronic messages,filtering the received fourth electronic messages and storing a result of the filtering in the shared memory portion of the cache memory device.
  • 2. A memory-efficient and processor-efficient computer system for reliable implementation of a table data cache proxy, the system comprising: a table data cache proxy enabled (TDCP-enabled) server computer coupled to a plurality of data server computers, the plurality of data server computers each being programmed with a table data service accessible via an electronic messaging protocol, the TDCP-enabled server computer having a plurality of TDCP clients, the TDCP-enabled server computer caching data from the plurality of data server computers and aggregating subscriptions of the TDCP clients to the plurality of data server computers;the TDCP-enabled server computer comprising: one or more hardware processors;a cache memory device having at least a portion being a shared memory portion; anda computer readable data storage device coupled to the one or more hardware processors, the computer readable data storage device having stored thereon software instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations including: receiving, from a first TDCP client of the plurality of TDCP clients, a first electronic message requesting table data;determining whether a shared memory cache stored in the shared memory portion of the cache memory device contains a cached copy of the requested table data;when the shared memory cache contains a cached copy of the requested table data, transmitting, to the first TDCP client, one or more second electronic messages comprising a reference indicating a location where the cached copy is stored in the shared memory portion of the cache memory device in response to the first electronic message, the data received from at least one of the plurality of data server computers and stored in the cache memory device being authoritative due to a data model of the plurality of data server computers;when the shared memory cache does not contain a cached copy of the requested table data, requesting the requested data from a server of the plurality of data servers.
  • 3. The system of claim 2, wherein the operations further include: when the shared memory cache does not contain a cached copy of the requested table data, transmitting a request to subscribe for updates from the one or more appropriate data server computers;receiving a subscription update from one or more of the plurality of data server computers; andelectronically distributing the received subscription update to one or more of the plurality of TDCP clients.
  • 4. The system of claim 2, wherein the first electronic message requests data from a data block and comprises a minimum response length; andwherein, when additional data above the minimum response length is available within the block, the additional data is included in response to the first electronic message to prevent redundant subsequent requests.
  • 5. The system of claim 2, wherein the first electronic message includes a table location discovery request requesting a list of table locations for a given table key, the table key uniquely identifying a table.
  • 6. The system of claim 2, wherein the first electronic message is a table location metadata retrieval request requesting table-location-level metadata including the size of a table.
  • 7. The system of claim 2, wherein the first electronic message is a column location metadata retrieval request requesting column-location-level metadata.
  • 8. The system of claim 2, wherein column data is laid out in an optimized data layout to enable high performance data access, the column-location-level metadata including information indicating the optimized data layout.
  • 9. The system of claim 2, wherein the first electronic message is a column file size retrieval request requesting the size of a column file.
  • 10. The system of claim 2, wherein the first electronic message is a column file data retrieval request requesting at least a portion of a block of binary data from a column file.
  • 11. The system of claim 2, wherein the plurality of data server computers comprises at least two data server computers providing the same data; andwherein the TDCP-enabled server is configured to reconnect to any of the data server computers, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.
  • 12. The system of claim 2, wherein the operations further include: when the first electronic message requests a block of binary data, determining whether a future request for another block of binary data is likely based on a number of previously received requests; andwhen the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.
  • 13. The system of claim 2, the operations further including: when the shared memory cache does not contain a cached copy of the requested table data: selecting one or more data server computers from the plurality of data server computers as one or more appropriate data server computers to provide the requested table data,the requesting including transmitting one or more third electronic messages to the one or more appropriate data server computers requesting the requested table data,receiving one or more fourth electronic messages from the one or more appropriate data server computers in response to the third electronic messages,filtering the received fourth electronic messages and storing a result of the filtering in the shared memory portion of the cache memory device, andtransmitting, to the first TDCP client computer, one or more fifth electronic messages comprising a reference indicating a location where the result of the filtering is stored in the shared memory portion of the cache memory device in response to the first electronic message.
  • 14. A system comprising: one or more hardware processors of a table data cache proxy enabled (TDCP-enabled) server, the TDCP-enabled server being coupled to one or more data servers;a computer readable data storage device coupled to the one or more hardware processors, the computer readable data storage device having stored thereon software instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations including: receiving a first electronic message requesting table data;determining whether a cache of the TDCP-enabled server contains a cached copy of the requested table data;when the cache of the TDCP-enabled server contains a cached copy of the requested table data, transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data received from at least one of the one or more data servers and stored in the cache being authoritative due to a data model of the one or more data serverswhen the cache of the TDCP-enabled server does not contain a cached copy of the requested data, requesting the requested data from a server of the plurality of data servers.
  • 15. The system of claim 14, wherein the TDCP-enabled server further comprises a shared memory in which at least a portion of the cache is stored,wherein the one or more second electronic messages comprise an indication where the cached copy is stored in the shared memory of the TDCP-enabled server, andwherein the one or more fifth electronic messages comprise indications where the requested table data is stored in the shared memory of the TDCP-enabled server.
  • 16. The system of claim 14, wherein the first electronic message is received via a network from a remote client, the one or more second electronic messages are transmitted to the remote client via the network, and the one or more fifth electronic messages are transmitted to the remote client via the network.
  • 17. The system of claim 14, wherein the first electronic message is received via an inter process communication (IPC) mechanism from a client, the one or more second electronic messages are transmitted to the client via the IPC mechanism, and the one or more fifth electronic messages are transmitted to the client via the IPC mechanism.
  • 18. The system of claim 14, wherein the first electronic message requests data from a data block and comprises a minimum response length; andwherein, when additional data above the minimum response length is available within the block, the additional data is included in response to the first electronic message to prevent redundant subsequent requests.
  • 19. The system of claim 14, wherein the first electronic message is one of: a table location discovery request requesting a list of table locations for a given table key, the table key uniquely identifying a table;a table location metadata retrieval request requesting table-location-level metadata including the size of a table;a column location metadata retrieval request requesting column-location-level metadata, wherein column data is laid out in an optimized data layout to enable high performance data access, the column-location-level metadata including information indicating the optimized data layout;a column file size retrieval request requesting the size of a column file; anda column file data retrieval request requesting at least a portion of a block of binary data from a column file.
  • 20. The system of claim 14, wherein the one or more data servers is a plurality of data servers comprising at least two data sources providing the same data; andwherein the TDCP-enabled server is configured to reconnect to any of the data servers, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.
  • 21. The system of claim 14, wherein the operations further include: when the first electronic message requests a block of binary data, determining whether a future request for another block of binary data is likely based on a number of received requests; andwhen the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.
  • 22. The system of claim 14, the operations further including: when, based on the determining, the cache of the TDCP-enabled server does not contain a cached copy of the requested data: determining one or more appropriate data servers of the one or more data servers to request the requested table data from,the requesting including transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested table data,receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages, andfiltering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested table data in response to the first electronic message.
  • 23. A method comprising: receiving, at a table data cache proxy enabled (TDCP-enabled) server, a first electronic message requesting table data, the TDCP-enabled server caching data;determining that a cache of the TDCP-enabled server contains a cached copy of the requested table data;after the determining that the cache of the TDCP-enabled server contains a cached copy of the requested table data, transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data stored in the cache being authoritative due to a data model;receiving, at the TDCP-enabled server, another electronic message requesting second table data;determining that the cache of the TDCP-enabled server does not contain a cached copy of the requested second table data; andafter the determining that the cache of the TDCP-enabled server does not contain a cached copy of the requested second table data, requesting the requested second table data.
  • 24. The method of claim 23, wherein the one or more data servers comprises a plurality of data servers; andwherein, when the first electronic message requests one or more data locations, the filtering comprises combining at least a portion of each of two or more of the one or more fourth electronic messages to create a composite of data from two or more of the plurality of data servers as the result of the filtering.
  • 25. The method of claim 23, wherein the first electronic message is one of: a table location discovery request requesting a list of table locations for a given table key, the table key uniquely identifying a table;a table location metadata retrieval request requesting table-location-level metadata including the size of a table;a column location metadata retrieval request requesting column-location-level metadata, wherein column data is laid out in an optimized data layout to enable high performance data access, the column-location-level metadata including information indicating the optimized data layout;a column file size retrieval request requesting the size of a column file; anda column file data retrieval request requesting at least a portion of a block of binary data from a column file.
  • 26. The method of claim 23, wherein the one or more data servers comprises a plurality of data servers comprising a first and second data server providing the same data; andwherein the TDCP-enabled server is configured to reconnect to any of the data servers, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.
  • 27. The method of claim 23, further comprising: determining whether a future request for another block of binary data is likely based on a number of received requests; andwhen the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.
  • 28. The method of claim 23, further comprising: when the cache of the TDCP-enabled server does not contain a cached copy of the requested data, transmitting a request to subscribe for updates from the one or more appropriate data servers;receiving a subscription update from one or more of the one or more data servers; andelectronically distributing the received subscription update to one or more subscribers of the TDCP-enabled server.
  • 29. The method of claim 23, further comprising: after the determining that the cache of the TDCP-enabled server does not contain a cached copy of the requested second table data: determining one or more appropriate data servers of the one or more data servers to request the requested second table data from,the requesting including transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested second table data,receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages, andfiltering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested second table data in response to the another electronic message.
  • 30. A nontransitory computer readable medium having stored thereon software instructions that, when executed by one or more processors, cause the one or more processors to perform operations including: receiving, at a table data cache proxy enabled (TDCP-enabled) server, a first electronic message requesting table data, the TDCP-enabled server being coupled to one or more data servers having a table data service accessible via an electronic messaging protocol, the TDCP-enabled server caching data from the one or more data servers;determining whether a cache of the TDCP-enabled server contains a cached copy of the requested table data;when the cache of the TDCP-enabled server contains a cached copy of the requested table data, transmitting one or more second electronic messages providing the cached copy of the requested table data from the cache in response to the first electronic message, the data received from at least one of the one or more data servers and stored in the cache being authoritative due to a data model of the one or more data servers; andwhen the cache of the TDCP-enabled server does not contain a cached copy of the requested data, requesting the requested data from a server of the one or more of data servers.
  • 31. The nontransitory computer readable medium of claim 30, wherein the one or more data servers comprises a plurality of data servers; andwherein, when the first electronic message requests one or more data locations, the filtering comprises combining at least a portion of each of two or more of the one or more fourth electronic messages to create a composite of data from two or more of the plurality of data servers as the result of the filtering.
  • 32. The nontransitory computer readable medium of claim 30, wherein the first electronic message is one of: a table location discovery request requesting a list of table locations for a given table key, the table key uniquely identifying a table;a table location metadata retrieval request requesting table-location-level metadata including the size of a table;a column location metadata retrieval request requesting column-location-level metadata, wherein column data is laid out in an optimized data layout to enable high performance data access, the column-location-level metadata including information indicating the optimized data layout;a column file size retrieval request requesting the size of a column file; anda column file data retrieval request requesting at least a portion of a block of binary data from a column file.
  • 33. The nontransitory computer readable medium of claim 30, wherein the one or more data servers comprises a plurality of data servers comprising a first and second data server providing the same data; andwherein the TDCP-enabled server is configured to reconnect to any of the data servers, and upon reconnection any in-progress requests are re-sent and any subscriptions are renewed.
  • 34. The nontransitory computer readable medium of claim 30, the operations further comprising: determining whether a future request for another block of binary data is likely based on a number of received requests; andwhen the future request is determined to be likely, prefetching the another block of binary data including transmitting one or more sixth electronic messages to the one or more appropriate data servers requesting the another block of binary data.
  • 35. The nontransitory computer readable medium of claim 30, the operations further comprising: when, based on the determining, the cache of the TDCP-enabled server does not contain a cached copy of the requested data: determining one or more appropriate data servers of the one or more data servers to request the requested table data from,transmitting one or more third electronic messages to the one or more appropriate data servers requesting the requested table data,receiving one or more fourth electronic messages from the one or more appropriate servers in response to the one or more third electronic messages, andfiltering the received one or more fourth electronic messages and, based on a result of the filtering, transmitting one or more fifth electronic messages providing the requested table data in response to the first electronic message.
Parent Case Info

This application claims the benefit of U.S. Provisional Application No. 62/161,813, entitled “Computer Data System” and filed on May 14, 2015, which is incorporated herein by reference in its entirety.

US Referenced Citations (425)
Number Name Date Kind
5335202 Manning et al. Aug 1994 A
5452434 Macdonald Sep 1995 A
5469567 Okada Nov 1995 A
5504885 Alashqur Apr 1996 A
5530939 Mansfield et al. Jun 1996 A
5568632 Nelson Oct 1996 A
5673369 Kim Sep 1997 A
5701461 Dalal et al. Dec 1997 A
5701467 Freeston Dec 1997 A
5764953 Collins et al. Jun 1998 A
5787428 Hart Jul 1998 A
5806059 Tsuchida et al. Sep 1998 A
5859972 Subramaniam et al. Jan 1999 A
5875334 Chow et al. Feb 1999 A
5878415 Olds Mar 1999 A
5890167 Bridge et al. Mar 1999 A
5899990 Maritzen et al. May 1999 A
5920860 Maheshwari et al. Jul 1999 A
5943672 Yoshida Aug 1999 A
5960087 Tribble et al. Sep 1999 A
5991810 Shapiro et al. Nov 1999 A
5999918 Williams et al. Dec 1999 A
6006220 Haderle et al. Dec 1999 A
6032144 Srivastava et al. Feb 2000 A
6032148 Wilkes Feb 2000 A
6038563 Bapat et al. Mar 2000 A
6058394 Bakow et al. May 2000 A
6061684 Glasser et al. May 2000 A
6138112 Slutz Oct 2000 A
6266669 Brodersen et al. Jul 2001 B1
6289357 Parker Sep 2001 B1
6292803 Richardson et al. Sep 2001 B1
6304876 Isip Oct 2001 B1
6317728 Kane Nov 2001 B1
6327702 Sauntry et al. Dec 2001 B1
6336114 Garrison Jan 2002 B1
6353819 Edwards et al. Mar 2002 B1
6367068 Vaidyanathan et al. Apr 2002 B1
6389414 Delo et al. May 2002 B1
6389462 Cohen et al. May 2002 B1
6438537 Netz et al. Aug 2002 B1
6446069 Yaung et al. Sep 2002 B1
6460037 Weiss et al. Oct 2002 B1
6473750 Petculescu et al. Oct 2002 B1
6487552 Lei et al. Nov 2002 B1
6496833 Goldberg et al. Dec 2002 B1
6505189 Au et al. Jan 2003 B1
6505241 Pitts Jan 2003 B2
6510551 Miller Jan 2003 B1
6530075 Beadle et al. Mar 2003 B1
6538651 Hayman et al. Mar 2003 B1
6546402 Beyer et al. Apr 2003 B1
6553375 Huang et al. Apr 2003 B1
6584474 Pereira Jun 2003 B1
6604104 Smith Aug 2003 B1
6618720 Au et al. Sep 2003 B1
6631374 Klein et al. Oct 2003 B1
6640234 Coffen et al. Oct 2003 B1
6697880 Dougherty Feb 2004 B1
6701415 Hendren Mar 2004 B1
6714962 Helland et al. Mar 2004 B1
6725243 Snapp Apr 2004 B2
6732100 Brodersen et al. May 2004 B1
6745332 Wong et al. Jun 2004 B1
6748374 Madan et al. Jun 2004 B1
6748455 Hinson et al. Jun 2004 B1
6760719 Hanson et al. Jul 2004 B1
6775660 Lin et al. Aug 2004 B2
6785668 Polo et al. Aug 2004 B1
6795851 Noy Sep 2004 B1
6816855 Hartel et al. Nov 2004 B2
6820082 Cook et al. Nov 2004 B1
6829620 Michael et al. Dec 2004 B2
6832229 Reed Dec 2004 B2
6851088 Conner et al. Feb 2005 B1
6882994 Yoshimura et al. Apr 2005 B2
6925472 Kong Aug 2005 B2
6934717 James Aug 2005 B1
6947928 Dettinger et al. Sep 2005 B2
6983291 Cochrane et al. Jan 2006 B1
6985895 Witkowski et al. Jan 2006 B2
6985899 Chan et al. Jan 2006 B2
6985904 Kaluskar et al. Jan 2006 B1
7020649 Cochrane et al. Mar 2006 B2
7024414 Sah et al. Apr 2006 B2
7031962 Moses Apr 2006 B2
7058657 Berno Jun 2006 B1
7089228 Arnold et al. Aug 2006 B2
7089245 George et al. Aug 2006 B1
7096216 Anonsen Aug 2006 B2
7099927 Cudd Aug 2006 B2
7103608 Ozbutun et al. Sep 2006 B1
7110997 Turkel et al. Sep 2006 B1
7127462 Hiraga et al. Oct 2006 B2
7146357 Suzuki et al. Dec 2006 B2
7149742 Eastham et al. Dec 2006 B1
7167870 Avvari et al. Jan 2007 B2
7171469 Ackaouy et al. Jan 2007 B2
7174341 Ghukasyan et al. Feb 2007 B2
7181686 Bahrs Feb 2007 B1
7188105 Dettinger et al. Mar 2007 B2
7200620 Gupta Apr 2007 B2
7216115 Walters et al. May 2007 B1
7216116 Nilsson et al. May 2007 B1
7219302 O'Shaughnessy et al. May 2007 B1
7225189 McCormack et al. May 2007 B1
7254808 Trappen et al. Aug 2007 B2
7257689 Baird Aug 2007 B1
7272605 Hinshaw et al. Sep 2007 B1
7308580 Nelson et al. Dec 2007 B2
7316003 Dulepet et al. Jan 2008 B1
7330969 Harrison et al. Feb 2008 B2
7333941 Choi Feb 2008 B1
7343585 Lau et al. Mar 2008 B1
7350237 Vogel et al. Mar 2008 B2
7380242 Alaluf May 2008 B2
7401088 Chintakayala et al. Jul 2008 B2
7426521 Harter Sep 2008 B2
7430549 Zane et al. Sep 2008 B2
7433863 Zane et al. Oct 2008 B2
7447865 Uppala et al. Nov 2008 B2
7478094 Ho et al. Jan 2009 B2
7484096 Garg et al. Jan 2009 B1
7493311 Cutsinger et al. Feb 2009 B1
7506055 McClain Mar 2009 B2
7529734 Dirisala May 2009 B2
7529750 Bair May 2009 B2
7542958 Warren et al. Jun 2009 B1
7552223 Ackaouy Jun 2009 B1
7610351 Gollapudi et al. Oct 2009 B1
7620687 Chen et al. Nov 2009 B2
7624126 Pizzo et al. Nov 2009 B2
7627603 Rosenblum et al. Dec 2009 B2
7661141 Dutta et al. Feb 2010 B2
7664778 Yagoub et al. Feb 2010 B2
7672275 Yajnik et al. Mar 2010 B2
7680782 Chen et al. Mar 2010 B2
7711716 Stonecipher May 2010 B2
7711740 Minore et al. May 2010 B2
7761444 Zhang et al. Jul 2010 B2
7797356 Iyer et al. Sep 2010 B2
7827204 Heinzel et al. Nov 2010 B2
7827403 Wong et al. Nov 2010 B2
7827523 Ahmed et al. Nov 2010 B2
7882121 Bruno et al. Feb 2011 B2
7882132 Ghatare Feb 2011 B2
7904487 Ghatare Mar 2011 B2
7908259 Branscome et al. Mar 2011 B2
7908266 Zeringue et al. Mar 2011 B2
7930412 Yeap et al. Apr 2011 B2
7966311 Haase Jun 2011 B2
7966312 Nolan et al. Jun 2011 B2
7966343 Yang et al. Jun 2011 B2
7970777 Saxena et al. Jun 2011 B2
7979431 Qazi et al. Jul 2011 B2
7984043 Waas Jul 2011 B1
8019795 Anderson et al. Sep 2011 B2
8027293 Spaur et al. Sep 2011 B2
8032525 Bowers et al. Oct 2011 B2
8037542 Taylor et al. Oct 2011 B2
8046394 Shatdal Oct 2011 B1
8046749 Owen et al. Oct 2011 B1
8055672 Djugash et al. Nov 2011 B2
8060484 Bandera et al. Nov 2011 B2
8171018 Zane et al. May 2012 B2
8180789 Wasserman et al. May 2012 B1
8196121 Peshansky et al. Jun 2012 B2
8209356 Roesler Jun 2012 B1
8286189 Kukreja et al. Oct 2012 B2
8321833 Langworthy et al. Nov 2012 B2
8332435 Ballard et al. Dec 2012 B2
8359305 Burke et al. Jan 2013 B1
8375127 Lita Feb 2013 B1
8380757 Bailey et al. Feb 2013 B1
8418142 Ao et al. Apr 2013 B2
8433701 Sargeant et al. Apr 2013 B2
8458218 Wildermuth Jun 2013 B2
8473897 Box et al. Jun 2013 B2
8478713 Cotner et al. Jul 2013 B2
8515942 Marum et al. Aug 2013 B2
8543620 Ching Sep 2013 B2
8553028 Urbach Oct 2013 B1
8555263 Allen et al. Oct 2013 B2
8560502 Vora Oct 2013 B2
8595151 Hao et al. Nov 2013 B2
8601016 Briggs et al. Dec 2013 B2
8631034 Peloski Jan 2014 B1
8650182 Murthy Feb 2014 B2
8660869 MacIntyre et al. Feb 2014 B2
8676863 Connell et al. Mar 2014 B1
8683488 Kukreja et al. Mar 2014 B2
8713518 Pointer et al. Apr 2014 B2
8719252 Miranker et al. May 2014 B2
8725707 Chen et al. May 2014 B2
8726254 Rohde et al. May 2014 B2
8745014 Travis Jun 2014 B2
8745510 D'Alo′ et al. Jun 2014 B2
8751823 Myles et al. Jun 2014 B2
8768961 Krishnamurthy Jul 2014 B2
8788254 Peloski Jul 2014 B2
8793243 Weyerhaeuser et al. Jul 2014 B2
8805947 Kuzkin et al. Aug 2014 B1
8806133 Hay et al. Aug 2014 B2
8812625 Chitilian et al. Aug 2014 B1
8838656 Cheriton Sep 2014 B1
8855999 Elliot Oct 2014 B1
8863156 Lepanto et al. Oct 2014 B1
8874512 Jin et al. Oct 2014 B2
8880569 Draper et al. Nov 2014 B2
8880787 Kimmel et al. Nov 2014 B1
8881121 Ali Nov 2014 B2
8886631 Abadi et al. Nov 2014 B2
8903717 Elliot Dec 2014 B2
8903842 Bloesch et al. Dec 2014 B2
8922579 Mi et al. Dec 2014 B2
8924384 Driesen et al. Dec 2014 B2
8930892 Pointer et al. Jan 2015 B2
8954418 Faerber et al. Feb 2015 B2
8959495 Chafi et al. Feb 2015 B2
8996864 Maigne et al. Mar 2015 B2
9031930 Valentin May 2015 B2
9077611 Cordray et al. Jul 2015 B2
9122765 Chen Sep 2015 B1
9195712 Freedman et al. Nov 2015 B2
9298768 Varakin et al. Mar 2016 B2
9311357 Ramesh et al. Apr 2016 B2
9372671 Balan et al. Jun 2016 B2
9384184 Cervantes et al. Jul 2016 B2
9613018 Zeldis et al. Apr 2017 B2
9633060 Caudy Apr 2017 B2
9832068 McSherry et al. Nov 2017 B2
20020002576 Wollrath et al. Jan 2002 A1
20020007331 Lo et al. Jan 2002 A1
20020054587 Baker et al. May 2002 A1
20020065981 Jenne et al. May 2002 A1
20020129168 Kanai Sep 2002 A1
20020156722 Greenwood Oct 2002 A1
20030004952 Nixon et al. Jan 2003 A1
20030061216 Moses Mar 2003 A1
20030074400 Brooks et al. Apr 2003 A1
20030110416 Morrison et al. Jun 2003 A1
20030167261 Grust et al. Sep 2003 A1
20030182261 Patterson Sep 2003 A1
20030208484 Chang et al. Nov 2003 A1
20030208505 Mullins et al. Nov 2003 A1
20030233632 Aigen et al. Dec 2003 A1
20040002961 Dettinger et al. Jan 2004 A1
20040076155 Yajnik et al. Apr 2004 A1
20040111492 Nakahara et al. Jun 2004 A1
20040148630 Choi Jul 2004 A1
20040186813 Tedesco et al. Sep 2004 A1
20040216150 Scheifler et al. Oct 2004 A1
20040220923 Nica Nov 2004 A1
20040254876 Coval et al. Dec 2004 A1
20050015490 Saare et al. Jan 2005 A1
20050060693 Robison et al. Mar 2005 A1
20050097447 Serra et al. May 2005 A1
20050102284 Srinivasan et al. May 2005 A1
20050102636 McKeon et al. May 2005 A1
20050131893 Glan Jun 2005 A1
20050132384 Morrison et al. Jun 2005 A1
20050138624 Morrison et al. Jun 2005 A1
20050165866 Bohannon et al. Jul 2005 A1
20050198001 Cunningham et al. Sep 2005 A1
20060059253 Goodman et al. Mar 2006 A1
20060074901 Pirahesh et al. Apr 2006 A1
20060085490 Baron et al. Apr 2006 A1
20060100989 Chinchwadkar et al. May 2006 A1
20060101019 Nelson et al. May 2006 A1
20060116983 Dettinger et al. Jun 2006 A1
20060116999 Dettinger et al. Jun 2006 A1
20060136361 Peri et al. Jun 2006 A1
20060173693 Arazi et al. Aug 2006 A1
20060195460 Nori et al. Aug 2006 A1
20060212847 Tarditi et al. Sep 2006 A1
20060218123 Chowdhuri et al. Sep 2006 A1
20060218200 Factor et al. Sep 2006 A1
20060230016 Cunningham et al. Oct 2006 A1
20060253311 Yin et al. Nov 2006 A1
20060271510 Harward et al. Nov 2006 A1
20060277162 Smith Dec 2006 A1
20070011211 Reeves et al. Jan 2007 A1
20070027884 Heger et al. Feb 2007 A1
20070033518 Kenna et al. Feb 2007 A1
20070073765 Chen Mar 2007 A1
20070101252 Chamberlain et al. May 2007 A1
20070169003 Branda et al. Jul 2007 A1
20070256060 Ryu et al. Nov 2007 A1
20070258508 Werb et al. Nov 2007 A1
20070271280 Chandasekaran Nov 2007 A1
20070299822 Jopp et al. Dec 2007 A1
20080022136 Mattsson et al. Jan 2008 A1
20080033907 Woehler et al. Feb 2008 A1
20080046804 Rui et al. Feb 2008 A1
20080072150 Chan et al. Mar 2008 A1
20080120283 Liu et al. May 2008 A1
20080155565 Poduri Jun 2008 A1
20080168135 Redlich et al. Jul 2008 A1
20080235238 Jalobeanu et al. Sep 2008 A1
20080263179 Buttner et al. Oct 2008 A1
20080276241 Bajpai et al. Nov 2008 A1
20080319951 Ueno et al. Dec 2008 A1
20090019029 Tommaney et al. Jan 2009 A1
20090022095 Spaur et al. Jan 2009 A1
20090037391 Agrawal et al. Feb 2009 A1
20090055370 Dagum et al. Feb 2009 A1
20090083215 Burger Mar 2009 A1
20090089312 Chi et al. Apr 2009 A1
20090248902 Blue Oct 2009 A1
20090254516 Meiyyappan et al. Oct 2009 A1
20090300770 Rowney et al. Dec 2009 A1
20090319058 Rovaglio et al. Dec 2009 A1
20090319484 Golbandi et al. Dec 2009 A1
20090327242 Brown et al. Dec 2009 A1
20100036801 Pirvali et al. Feb 2010 A1
20100042587 Johnson et al. Feb 2010 A1
20100047760 Best et al. Feb 2010 A1
20100049715 Jacobsen et al. Feb 2010 A1
20100161555 Nica et al. Jun 2010 A1
20100186082 Ladki et al. Jul 2010 A1
20100199161 Aureglia et al. Aug 2010 A1
20100205017 Sichelman et al. Aug 2010 A1
20100205351 Wiener et al. Aug 2010 A1
20100281005 Carlin et al. Nov 2010 A1
20100281071 Ben-Zvi et al. Nov 2010 A1
20110126110 Vilke et al. May 2011 A1
20110126154 Boehler et al. May 2011 A1
20110153603 Adiba et al. Jun 2011 A1
20110161378 Williamson Jun 2011 A1
20110167020 Yang et al. Jul 2011 A1
20110178984 Talius et al. Jul 2011 A1
20110194563 Shen et al. Aug 2011 A1
20110219020 Oks et al. Sep 2011 A1
20110314019 Pens Dec 2011 A1
20120110030 Pomponio May 2012 A1
20120144234 Clark et al. Jun 2012 A1
20120159303 Friedrich et al. Jun 2012 A1
20120191446 Binsztok et al. Jul 2012 A1
20120192096 Bowman et al. Jul 2012 A1
20120197868 Fauser et al. Aug 2012 A1
20120209886 Henderson Aug 2012 A1
20120215741 Poole et al. Aug 2012 A1
20120221528 Renkes Aug 2012 A1
20120246052 Taylor et al. Sep 2012 A1
20120254143 Varma et al. Oct 2012 A1
20120259759 Crist et al. Oct 2012 A1
20120296846 Teeter Nov 2012 A1
20130041946 Joel et al. Feb 2013 A1
20130080514 Gupta et al. Mar 2013 A1
20130086107 Genochio et al. Apr 2013 A1
20130166556 Baeumges et al. Jun 2013 A1
20130173667 Soderberg et al. Jul 2013 A1
20130179460 Cervantes et al. Jul 2013 A1
20130185619 Ludwig Jul 2013 A1
20130191370 Chen et al. Jul 2013 A1
20130198232 Shamgunov et al. Aug 2013 A1
20130226959 Dittrich et al. Aug 2013 A1
20130246560 Feng et al. Sep 2013 A1
20130263123 Zhou et al. Oct 2013 A1
20130290243 Hazel et al. Oct 2013 A1
20130304725 Nee et al. Nov 2013 A1
20130304744 McSherry et al. Nov 2013 A1
20130311352 Kayanuma et al. Nov 2013 A1
20130311488 Erdogan et al. Nov 2013 A1
20130318129 Vingralek et al. Nov 2013 A1
20130346365 Kan et al. Dec 2013 A1
20140019494 Tang Jan 2014 A1
20140040203 Lu et al. Feb 2014 A1
20140059646 Hannel et al. Feb 2014 A1
20140082724 Pearson et al. Mar 2014 A1
20140136521 Pappas May 2014 A1
20140143123 Banke et al. May 2014 A1
20140149997 Kukreja et al. May 2014 A1
20140156618 Castellano Jun 2014 A1
20140173023 Varney et al. Jun 2014 A1
20140181036 Dhamankar et al. Jun 2014 A1
20140181081 Veldhuizen Jun 2014 A1
20140188924 Ma et al. Jul 2014 A1
20140195558 Murthy et al. Jul 2014 A1
20140201194 Reddy et al. Jul 2014 A1
20140215446 Araya et al. Jul 2014 A1
20140222768 Rambo et al. Aug 2014 A1
20140229506 Lee Aug 2014 A1
20140229874 Strauss Aug 2014 A1
20140244687 Shmueli et al. Aug 2014 A1
20140279810 Mann et al. Sep 2014 A1
20140280522 Watte Sep 2014 A1
20140282227 Nixon et al. Sep 2014 A1
20140282444 Araya et al. Sep 2014 A1
20140282540 Bonnet et al. Sep 2014 A1
20140297611 Abbour et al. Oct 2014 A1
20140317084 Chaudhry et al. Oct 2014 A1
20140324821 Meiyyappan et al. Oct 2014 A1
20140330700 Studnitzer et al. Nov 2014 A1
20140330807 Weyerhaeuser et al. Nov 2014 A1
20140344186 Nadler Nov 2014 A1
20140344391 Varney et al. Nov 2014 A1
20140359574 Beckwith et al. Dec 2014 A1
20140372482 Martin et al. Dec 2014 A1
20140380051 Edward et al. Dec 2014 A1
20150019516 Wein et al. Jan 2015 A1
20150026155 Martin Jan 2015 A1
20150067640 Booker et al. Mar 2015 A1
20150074066 Li et al. Mar 2015 A1
20150082218 Affoneh et al. Mar 2015 A1
20150088894 Czarlinska et al. Mar 2015 A1
20150095381 Chen et al. Apr 2015 A1
20150127599 Schiebeler May 2015 A1
20150154262 Yang et al. Jun 2015 A1
20150172117 Dolinsky et al. Jun 2015 A1
20150188778 Asayag et al. Jul 2015 A1
20150205588 Bates et al. Jul 2015 A1
20150254298 Bourbonnais et al. Sep 2015 A1
20150304182 Brodsky et al. Oct 2015 A1
20150317359 Tran et al. Nov 2015 A1
20160026442 Chhaparia Jan 2016 A1
20160065670 Kimmel et al. Mar 2016 A1
20160092599 Barsness et al. Mar 2016 A1
20160125018 Tomoda et al. May 2016 A1
20160171070 Hrle et al. Jun 2016 A1
20160253294 Allen et al. Sep 2016 A1
20160316038 Jolfaei Oct 2016 A1
20160335330 Teodorescu et al. Nov 2016 A1
20160335361 Teodorescu et al. Nov 2016 A1
20170161514 Dettinger et al. Jun 2017 A1
Foreign Referenced Citations (13)
Number Date Country
2309462 Dec 2000 CA
1406463 Apr 2004 EP
1198769 Jun 2008 EP
2199961 Jun 2010 EP
2423816 Feb 2012 EP
2743839 Jun 2014 EP
2421798 Jun 2011 RU
2000000879 Jan 2000 WO
2001079964 Oct 2001 WO
2011120161 Oct 2011 WO
2012136627 Oct 2012 WO
2014026220 Feb 2014 WO
2014143208 Sep 2014 WO
Non-Patent Literature Citations (141)
Entry
“About Entering Commands in the Command Window”, dated Dec. 16, 2015. Retrieved from https://knowledge.autodesk.com/support/autocad/learn-explore/caas/CloudHelp/cloudhelp/2016/ENU/AutoCAD-Core/files/GUID-BB0C3E79-66AF-4557-9140-D31B4CF3C9CF-htm.html (last accessed Jun. 16, 2016).
“Change Data Capture”, Oracle Database Online Documentation 11g Release 1 (11.1), dated Apr. 5, 2016. Retreived rom https://web.archive.org/web/20160405032625/http://docs.oracle.com/cd/B28359_01/server.111/b28313/cdc.htm.
“Chapter 24. Query access plans”, Tuning Database Performance, DB2 Version 9.5 for Linux, UNIX, and Windows, pp. 301-462, dated Dec. 2010. Retreived from http://public.dhe.ibm.com/ps/products/db2/info/vr95/pdf/en_US/DB2PerfTuneTroubleshoot-db2d3e953.pdf.
“GNU Emacs Manual”, dated Apr. 15, 2016, pp. 43-47. Retrieved from https://web.archive.org/web/20160415175915/http://www.gnu.org/software/emacs/manual/html_mono/emacs.html.
“Google Protocol RPC Library Overview”, dated Apr. 27, 2016. Retrieved from https://cloud.google.com/appengine/docs/python/tools/protorpc/ (last accessed Jun. 16, 2016).
“IBM—What is HBase?”, dated Sep. 6, 2015. Retrieved from https://web.archive.org/web/20150906022050/http://www-01.ibm.com/software/data/infosphere/hadoop/hbase/.
“IBM Informix TimeSeries data management”, dated Jan. 18, 2016. Retrieved from https://web.archive.org/web/20160118072141/http://www-01.ibm.com/software/data/informix/timeseries/.
“IBM InfoSphere BigInsights 3.0.0—Importing data from and exporting data to DB2 by using Sqoop”, dated Jan. 15, 2015. Retrieved from https://web.archive.org/web/20150115034058/http://www-01.ibm.com/support/knowledgecenter/SSPT3X_3.0.0/com.ibm.swg.im.infosphere.biginsights.import.doc/doc/data_warehouse_sqoop.html.
“Maximize Data Value with Very Large Database Management by SAP Sybase IQ”, dated 2013. Retrieved from http://www.sap.com/bin/sapcom/en_us/downloadasset.2013-06-jun-11-11.maximize-data-value-with-very-large-database-management-by-sap-sybase-iq-pdf.html.
“Microsoft Azure—Managing Access Control Lists (ACLs) for Endpoints by using PowerShell”, dated Nov. 12, 2014. Retrieved from https://web.archive.org/web/20150110170715/http://msdn.microsoft.com/en-us/library/azure/dn376543.aspx.
“Oracle Big Data Appliance—Perfect Balance Java API”, dated Sep. 20, 2015. Retrieved from https://web.archive.org/web/20131220040005/http://docs.oracle.com/cd/E41604_01/doc.22/e41667/toc.htm.
“Oracle Big Data Appliance—X5-2”, dated Sep. 6, 2015. Retrieved from https://web.archive.org/web/20150906185409/http://www.oracle.com/technetwork/database/bigdata-appliance/overview/bigdataappliance-datasheet-1883358.pdf.
“Oracle Big Data Appliance Software User's Guide”, dated Feb. 2015. Retrieved from https://docs.oracle.com/cd/E55905_01/doc.40/e55814.pdf.
“SAP HANA Administration Guide”, dated Mar. 29, 2016, pp. 290-294. Retrieved from https://web.archive.org/web/20160417053656/http://help.sap.com/hana/SAP_HANA_Administration_Guide_en.pdf.
“Sophia Database—Architecture”, dated Jan. 18, 2016. Retrieved from https://web.archive.org/web/20160118052919/http://sphia.org/architecture.html.
“Tracking Data Changes”, SQL Server 2008 R2, dated Sep. 22, 2015. Retreived from https://web.archive.org/web/20150922000614/https://technet.microsoft.com/en-us/library/bb933994(v=sql.105).aspx.
“Use Formula AutoComplete”, dated 2010. Retrieved from https://support.office.com/en-us/article/Use-Formula-AutoComplete-c7c46fa6-3a94-4150-a2f7-34140c1ee4d9 (last accessed Jun. 16, 2016).
Adelfio et al. “Schema Extraction for Tabular Data on the Web”, Proceedings of the VLDB Endowment, vol. 6, No. 6. Apr. 2013. Retrieved from http://www.cs.umd.edu/˜hjs/pubs/spreadsheets-vldb13.pdf.
Advisory Action dated Apr. 19, 2017, in U.S. Appl. No. 15/154,999.
Advisory Action dated Apr. 20, 2017, in U.S. Appl. No. 15/154,980.
Advisory Action dated Apr. 6, 2017, in U.S. Appl. No. 15/154,995.
Advisory Action dated Mar. 31, 2017, in U.S. Appl. No. 15/154,996.
Advisory Action dated May 3, 2017, in U.S. Appl. No. 15/154,993.
Borror, Jefferey A. “Q for Mortals 2.0”, dated Nov. 1, 2011. Retreived from http://code.la.com/wiki/JB:QforMortals2/contents.
Cheusheva, Svetlana. “How to change the row color based on a cell's value in Excel”, dated Oct. 29, 2013. Retrieved from https://www.ablebits.com/office-addins-blog/2013/10/29/excel-change-row-background-color/ (last accessed Jun. 16, 2016).
Corrected Notice of Allowability dated Aug. 9, 2017, in U.S. Appl. No. 15/154,980.
Corrected Notice of Allowability dated Jul. 31, 2017, in U.S. Appl. No. 15/154,999.
Corrected Notice of Allowability dated Mar. 10, 2017, in U.S. Appl. No. 15/154,979.
Corrected Notice of Allowability dated Oct. 26, 2017, in U.S. Appl. No. 15/610,162.
Decision on Pre-Appeal Conference Request dated Nov. 20, 2017, in U.S. Appl. No. 15/154,997.
Ex Parte Quayle Action dated Aug. 8, 2016, in U.S. Appl. No. 15/154,999.
Final Office Action dated Apr. 10, 2017, in U.S. Appl. No. 15/155,006.
Final Office Action dated Dec. 19, 2016, in U.S. Appl. No. 15/154,995.
Final Office Action dated Feb. 24, 2017, in U.S. Appl. No. 15/154,993.
Final Office Action dated Jan. 27, 2017, in U.S. Appl. No. 15/154,980.
Final Office Action dated Jan. 31, 2017, in U.S. Appl. No. 15/154,996.
Final Office Action dated Jul. 27, 2017, in U.S. Appl. No. 15/154,993.
Final Office Action dated Jun. 23, 2017, in U.S. Appl. No. 15/154,997.
Final Office Action dated Mar. 1, 2017, in U.S. Appl. No. 15/154,975.
Final Office Action dated Mar. 13, 2017, in U.S. Appl. No. 15/155,012.
Final Office Action dated Mar. 31, 2017, in U.S. Appl. No. 15/155,005.
Final Office Action dated May 15, 2017, in U.S. Appl. No. 15/155,010.
Final Office Action dated May 4, 2017, in U.S. Appl. No. 15/155,009.
Gai, Lei et al. “An Efficient Summary Graph Driven Method for RDF Query Processing”, dated Oct. 27, 2015. Retreived from http://arxiv.org/pdf/1510.07749.pdf.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032582 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032584 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032588 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032593 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032597 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032599 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 18, 2016, in International Appln. No. PCT/US2016/032605 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 25, 2016, in International Appln. No. PCT/US2016/032590 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 25, 2016, in International Appln. No. PCT/US2016/032592 filed May 14, 2016.
International Search Report and Written Opinion dated Aug. 4, 2016, in International Appln. No. PCT/US2016/032581 filed May 14, 2016.
International Search Report and Written Opinion dated Jul. 28, 2016, in International Appln. No. PCT/US2016/032586 filed May 14, 2016.
International Search Report and Written Opinion dated Jul. 28, 2016, in International Appln. No. PCT/US2016/032587 filed May 14, 2016.
Search Report and Written Opinion dated Jul. 28, 2016, in International Appln. No. PCT/US2016/032589 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 1, 2016, in International Appln. No. PCT/US2016/032596 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 1, 2016, in International Appln. No. PCT/US2016/032598 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 1, 2016, in International Appln. No. PCT/US2016/032601 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 1, 2016, in International Appln. No. PCT/US2016/032602 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 1, 2016, in International Appln. No. PCT/US2016/032607 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 15, 2016, in International Appln. No. PCT/US2016/032591 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 15, 2016, in International Appln. No. PCT/US2016/032594 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 15, 2016, in International Appln. No. PCT/US2016/032600 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 29, 2016, in International Appln. No. PCT/US2016/032595 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 29, 2016, in International Appln. No. PCT/US2016/032606 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 8, 2016, in International Appln. No. PCT/US2016/032603 filed May 14, 2016.
International Search Report and Written Opinion dated Sep. 8, 2016, in International Appln. No. PCT/US2016/032604 filed May 14, 2016.
Jellema, Lucas. “Implementing Cell Highlighting in JSF-based Rich Enterprise Apps (Part 1)”, dated Nov. 2008. Retrieved from http://www.oracle.com/technetwork/articles/adf/jellema-adfcellhighlighting-087850.html (last accessed Jun. 16, 2016).
Lou, Yuan. “A Multi-Agent Decision Support System for Stock Trading”, IEEE Network, Jan./Feb. 2002. Retreived from http://www.reading.ac.uk/AcaDepts/si/sisweb13/ais/papers/journal12-A%20multi-agent%20Framework.pdf.
Mallet, “Relational Database Support for Spatia-Temporal Data”, Technical Report TR 04-21, Sep. 2004, University of Alberta, Department of Computing Science.
Mariyappan, Balakrishnan. “10 Useful Linux Bash_Completion Complete Command Examples (Bash Command Line Completion on Steroids)”, dated Dec. 2, 2013. Retrieved from http://www.thegeekstuff.com/2013/12/bash-completion-complete/ (last accessed Jun. 16, 2016).
Murray, Derek G. et al. “Naiad: a timely dataflow system.” SOSP '13 Proceedings of the Twenty-Fourth ACM Symposium on Operating Systems Principles. pp. 439-455. Nov. 2013.
Non-final Office Action dated Apr. 19, 2017, in U.S. Appl. No. 15/154,974.
Non-final Office Action dated Aug. 12, 2016, in U.S. Appl. No. 15/155,001.
Non-final Office Action dated Aug. 16, 2016, in U.S. Appl. No. 15/154,993.
Non-final Office Action dated Aug. 19, 2016, in U.S. Appl. No. 15/154,991.
Non-final Office Action dated Aug. 25, 2016, in U.S. Appl. No. 15/154,980.
Non-final Office Action dated Aug. 26, 2016, in U.S. Appl. No. 15/154,995.
Non-final Office Action dated Aug. 8, 2016, in U.S. Appl. No. 15/154,981.
Non-final Office Action dated Aug. 8, 2016, in U.S. Appl. No. 15/154,985.
Non-final Office Action dated Dec. 13, 2017, in U.S. Appl. No. 15/608,963.
Non-final Office Action dated Feb. 8, 2017, in U.S. Appl. No. 15/154,997.
Non-final Office Action dated Jul. 27, 2017, in U.S. Appl. No. 15/154,995.
Non-final Office Action dated Mar. 2, 2017, in U.S. Appl. No. 15/154,984.
Non-final Office Action dated Nov. 15, 2017, in U.S. Appl. No. 15/654,461.
Non-final Office Action dated Nov. 17, 2016, in U.S. Appl. No. 15/154,999.
Non-final Office Action dated Nov. 21, 2017, in U.S. Appl. No. 15/155,005.
Non-final Office Action dated Nov. 30, 2017, in U.S. Appl. No. 15/155,012.
Non-final Office Action dated Oct. 13, 2016, in U.S. Appl. No. 15/155,009.
Non-final Office Action dated Oct. 27, 2016, in U.S. Appl. No. 15/155,006.
Non-final Office Action dated Oct. 5, 2017, in U.S. Appl. No. 15/428,145.
Non-final Office Action dated Oct. 7, 2016, in U.S. Appl. No. 15/154,998.
Non-final Office Action dated Sep. 1, 2016, in U.S. Appl. No. 15/154,979.
Non-final Office Action dated Sep. 1, 2016, in U.S. Appl. No. 15/155,011.
Non-final Office Action dated Sep. 1, 2016, in U.S. Appl. No. 15/155,012.
Non-final Office Action dated Sep. 14, 2016, in U.S. Appl. No. 15/154,984.
Non-final Office Action dated Sep. 16, 2016, in U.S. Appl. No. 15/154,988.
Non-final Office Action dated Sep. 22, 2016, in U.S. Appl. No. 15/154,987.
Non-final Office Action dated Sep. 26, 2016, in U.S. Appl. No. 15/155,005.
Non-final Office Action dated Sep. 29, 2016, in U.S. Appl. No. 15/154,990.
Non-final Office Action dated Sep. 8, 2016, in U.S. Appl. No. 15/154,975.
Non-final Office Action dated Sep. 9, 2016, in U.S. Appl. No. 15/154,996.
Non-final Office Action dated Sep. 9, 2016, in U.S. Appl. No. 15/155,010.
Notice of Allowance dated Dec. 19, 2016, in U.S. Appl. No. 15/155,001.
Notice of Allowance dated Dec. 22, 2016, in U.S. Appl. No. 15/155,011.
Notice of Allowance dated Dec. 7, 2016, in U.S. Appl. No. 15/154,985.
Notice of Allowance dated Feb. 1, 2017, in U.S. Appl. No. 15/154,988.
Notice of Allowance dated Feb. 14, 2017, in U.S. Appl. No. 15/154,979.
Notice of Allowance dated Feb. 28, 2017, in U.S. Appl. No. 15/154,990.
Notice of Allowance dated Jan. 30, 2017, in U.S. Appl. No. 15/154,987.
Notice of Allowance dated Jul. 28, 2017, in U.S. Appl. No. 15/155,009.
Notice of Allowance dated Jun. 19, 2017, in U.S. Appl. No. 15/154,980.
Notice of Allowance dated Jun. 20, 2017, in U.S. Appl. No. 15/154,975.
Notice of Allowance dated Mar. 2, 2017, in U.S. Appl. No. 15/154,998.
Notice of Allowance dated Mar. 31, 2017, in U.S. Appl. No. 15/154,998.
Notice of Allowance dated May 10, 2017, in U.S. Appl. No. 15/154,988.
Notice of Allowance dated Nov. 17, 2016, in U.S. Appl. No. 15/154,991.
Notice of Allowance dated Nov. 17, 2017, in U.S. Appl. No. 15/154,993.
Notice of Allowance dated Nov. 21, 2016, in U.S. Appl. No. 15/154,983.
Notice of Allowance dated Nov. 8, 2016, in U.S. Appl. No. 15/155,007.
Notice of Allowance dated Oct. 11, 2016, in U.S. Appl. No. 15/155,007.
Notice of Allowance dated Oct. 21, 2016, in U.S. Appl. No. 15/154,999.
Notice of Allowance dated Oct. 6, 2017, in U.S. Appl. No. 15/610,162.
Palpanas, Themistoklis et al. “Incremental Maintenance for Non-Distributive Aggregate Functions”, Proceedings of the 28th VLDB Conference, 2002. Retreived from http://www.vldb.org/conf/2002/S22P04.pdf.
PowerShell Team, Intellisense in Windows PowerShell ISE 3.0, dated Jun. 12, 2012, Windows PowerShell Blog, pp. 1-6 Retrieved: https://biogs.msdn.microsoft.com/powershell/2012/06/12/intellisense-in-windows-powershell-ise-3-0/.
Smith, Ian. “Guide to Using SQL: Computed and Automatic Columns.” Rdb Jornal, dated Sep. 2008, retrieved Aug. 15, 2016, retrieved from the Internet <URL: http://www.oracle.com/technetwork/products/rdb/automatic-columns-132042.pdf>.
Wes McKinney & PyData Development Team. “pandas: powerful Python data analysis toolkit, Release 0.16.1” Dated May 11, 2015. Retrieved from: http://pandas.pydata.org/pandas-docs/version/0.16.1/index.html.
Wes McKinney & PyData Development Team. “pandas: powerful Python data analysis toolkit, Release 0.18.1” Dated May 3, 2016. Retrieved from: http://pandas.pydata.org/pandas-docs/version/0.18.1/index.html.
Wu, Buwen et al. “Scalable SPARQL Querying using Path Partitioning”, 31st IEEE International Conference on Data Engineering (ICDE 2015), Seoul, Korea, Apr. 13-17, 2015. Retreived from http://imada.sdu.dk/˜zhou/papers/icde2015.pdf.
Advisory Action dated Dec. 21, 2017, in U.S. Appl. No. 15/154,984.
Breitbart, Update Propagation Protocols for Replicated Databases, SIGMOD '99 Philadelphia PA, 1999, pp. 97-108.
Final Office Action dated Dec. 29, 2017, in U.S. Appl. No. 15/154,974.
Kramer, The Combining DAG: A Technique for Parallel Data Flow Analysis, IEEE Transactions on Parallel and Distributed Systems, vol. 5, No. 8, Aug. 1994, pp. 805-813.
Non-final Office Action dated Dec. 28, 2017, in U.S. Appl. No. 15/154,996.
Non-final Office Action dated Dec. 28, 2017, in U.S. Appl. No. 15/796,230.
Non-final Office Action dated Feb. 12, 2018, in U.S. Appl. No. 15/466,836.
Non-final Office Action dated Jan. 4, 2018, in U.S. Appl. No. 15/583,777.
Notice of Allowance dated Feb. 12, 2018, in U.S. Appl. No. 15/813,142.
Sobell, Mark G. “A Practical Guide to Linux, Commands, Editors and Shell Programming.” Third Edition, dated Sep. 14, 2012. Retrieved from: http://techbus.safaribooksonline.com/book/operating-systems-and-server-administration/linux/9780133085129.
Related Publications (1)
Number Date Country
20170195461 A1 Jul 2017 US
Provisional Applications (1)
Number Date Country
62161813 May 2015 US
Continuations (1)
Number Date Country
Parent 15155001 May 2016 US
Child 15464314 US