Techniques and architectures for managing global installations and configurations

Information

  • Patent Grant
  • 11579940
  • Patent Number
    11,579,940
  • Date Filed
    Thursday, December 26, 2019
    5 years ago
  • Date Issued
    Tuesday, February 14, 2023
    2 years ago
Abstract
A publish and subscribe architecture can be utilized to manage records, which can be used to accomplish the various functional goals. At least one template having definitions for managing production and consumption of data within an unconfigured group of computing resources is maintained. Records organized by topic collected from multiple disparate previously configured producers are utilized to initiate configuration of the unconfigured group of computing resources. Records within a topic are organized by a corresponding topic sequence. A first portion of the computing resources are configured as consumers based on the at least one template. The consumers to consume records at a pace independent of record production. A second portion of the computing resources are configured as producers based on the at least one template. The producers to produce records at a pace independent of record consumption.
Description
TECHNICAL FIELD

Embodiments relate to techniques for managing installations and/or configurations of, for example, data centers. More specifically, embodiments relate to distributed architectures and mechanisms to provide efficient and consistent configuration functionality utilizing publish/subscribe concepts that can scale efficiently.


BACKGROUND

Configuring and/or managing data centers (or other high-volume data resources) can be a complex, expensive and/or time-consuming venture. Even currently automated functionality must be manually initiated or managed. Increased complexity and time requirements result in increased costs and increased likelihood of error. Thus, more efficient architectures and/or techniques would be beneficial.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings in which like reference numerals refer to similar elements.



FIG. 1 is a block diagram of one embodiment of an architecture that can be utilized to provide the functionality described herein.



FIG. 2 is a block diagram of one embodiment of a distributed architecture that can be utilized to provide the functionality described herein.



FIG. 3 is a flow diagram of one embodiment of a technique for managing configurations.



FIG. 4 is a block diagram of one embodiment of a global configuration agent.





DETAILED DESCRIPTION

In the following description, numerous specific details are set forth. However, embodiments of the invention may be practiced without these specific details. In other instances, well-known structures and techniques have not been shown in detail in order not to obscure the understanding of this description.


The following description provides specifics in terms of platforms (e.g., salesforce) and supporting infrastructure (e.g., Apache Kafka or a variation thereof); however, the techniques, mechanisms, architectures and functionality described and claimed herein are not limited to those specific implementation details. Various platforms and features available from, for example, salesforce.com can be used to connect data from a global network of internal and external entities to enable products that improve customer access to services and data.


However, this has resulted in increased complexity related to on-boarding of clients as well as ongoing and/or expanding operations. Managing a relatively small number of integrations and clusters can be achieved by a team utilizing various tools, but this standard approach is not sustainable at the global level. Various embodiments provide improved functionality by utilizing the concepts of producers and consumers that can operate using records that are managed by topics. Various other labels for these concepts and alternative architectures are described in greater detail below.


In various embodiments described herein one or more of the following core functionalities can be provided to enable a more efficient solution. 1) Enabling customers to self-service their producers and consumers using the concepts of Sources and Sinks. Sources and Sinks are an abstraction layer that can span clusters (e.g., Kafka clusters) which result in the automatic management of, for example, topics, access control lists (ACLs), quotas and mirroring to, for example, satisfy global on-boarding with a single command. 2) Providing global visibility on cluster health and state from a single interface (e.g., operations portal). 3) Ensuring global state marches towards desired state using event-driven mechanisms.


In general, a publish and subscribe architecture can be utilized to manage records, which can be used to accomplish the various functional goals, for example, efficient on-boarding. In some embodiments, Apache Kafka can be utilized to provide some or all of the publish and subscribe architecture. In alternate embodiments, other publish and subscribe architectures can be utilized. With these types of architectures, data consumers (data sinks) can receive data according to their own schedule in a flexible and efficient way.



FIG. 1 is a block diagram of one embodiment of an architecture that can be utilized to provide the functionality described herein. In the example of FIG. 1, one or more producer entities (e.g., 110, 112, 118) can produce records (e.g., 120, 160) that can be utilized by consumer entities (e.g., 170, 172, 178). As mentioned above, this architecture can be utilized, for example, to accomplish goals such as onboarding of clients in complex multi-client environments or configuration of devices in complex environments like data centers.


Producers can be various disparate sources that provide records to cluster 130 that can organize records into topics (e.g., 140, 142, 144, 146) that can be utilized to provide the appropriate records to consumers. While the example of FIG. 1 has relatively small numbers, any number of producers, records, clusters, topics and consumers can be supported.


The example of FIG. 1 includes a single cluster; however, any number of clusters can be supported. For example, records can be gathered by one or more local clusters (that are local to the producers of the records) and can be sent to a data center cluster (or other higher-level cluster) to be distributed to consumers and/or to lower-level clusters more local to the consumers.


As mentioned above, Apache Kafka can be utilized to provide some of the basic functionality described herein. In alternate embodiments, other similar tools can be utilized. In an example Kafka (or similar)-based embodiments records, topics, consumers, producers, brokers, logs, partitions and clusters can be utilized. In one embodiment, records (e.g., 120, 160) are immutable and can have a key, a value and a timestamp (not illustrated in FIG. 1). In one embodiment, each topic (e.g., 140, 142, 144, 146) has an associated log that functions as the storage structure of the topic. In some embodiments, a topic log can be broken into partitions and segments.


In one embodiment, a producer can produce streams of data records that can be received by cluster 130, for example, via producer application program interface (API) 150. In one embodiment, consumers can consume a stream of data records, for example, via consumer API 155. In one embodiment, a broker is a server (or similar) that runs one or more topic partitions. In various embodiments, clusters can consist of many brokers on many servers.


In one embodiment, records are appended from the producer end of the topic log. In various embodiments, a topic log can consist of many partitions spread over multiple files that can be on multiple cluster nodes.


Consumers read from topic logs at their respective cadences and can consume records (e.g., via offset) within the topic log. In various embodiments, each consumer can track offset(s) from where they left off reading. In some embodiments, topic log partitions can be distributed across multiple nodes in a cluster for performance and scalability. In some embodiments, partitions can be replicated to provide failover.


As described in greater detail below, records can be generated by various sources, managed in one or more clusters and consumed in a manner to provide self-managed onboarding to one or more environments. As one example, one or more servers within a data center (or even a new data center) can be configured by consuming records from an appropriate topic.


In various embodiments sources and sinks of data can be defined by templates that can be applied globally or to one or more data centers that can require replication that can control accesses, which is a new control on data flows. Further, orchestration of, for example, Kafka clusters and giving customers ability to define future states. Thus, centralized application command and control can be provided, which is new and useful. Further, linking to command and control can be supported.


The techniques described herein support user self-service with respect to devices or groups of devices (e.g., data centers), which eliminates or reduces the need for manual administration of these devices. Further, templated integrations can be supported that can allow new groups of devices to come online automatically configured. In various embodiments, data sources and data sinks can be conceptualized, which abstracts away from the standard “building blocks” of Kafka. Conceptually, the standard use of topics, producer ACLs, consumer ACLs and mirroring patterns are not utilized. Rather, in one embodiment, a source is defined in one group of devices (or system) and a sink is defined in another group of devices (or a different system) manages the configuration of core building blocks.


In various embodiments, a record management agent can run in one or more data centers (or other groups of computing resources) and can function to collect and manage distribution of records within the data center (or resource group). The record management agent can communicate records and/or related information through various application program interfaces (APIs). As described in greater detail below, in some embodiments, specific operations (or groups of operations) can be utilized to accomplish these tasks.


In various embodiments, the following concepts (that may have corresponding commands) can be provided. An “ENSURE_SOURCE” command (or similar concept) can be utilized to onboard or manage a producer (e.g., 110, 112, 118) including creating resources needed by the producer, and a “DECOM_SOURCE” command (or similar concept) can be utilized to remove access for the producer including decommissioning resources needed by the producer. Similarly, an “ENSURE_SINK” command (or similar concept) can be utilized to onboard or manage a consumer (e.g., 170, 172, 178) including creating resources needed by the consumer, and a “DECOM_SINK” command (or similar concept) can be utilized to remove access for the consumer including decommissioning resources needed by the consumer.


A source is a producer of data with an associated scope that defines the granularity with which topics are to be created. The scope can be, for example, per pod or per data center. A sink is a logical consumer of data and can be applied to a source. In an example Apache Kafka based embodiment, sources and sinks can be associated with a principal. In one embodiment, topics that fulfil a source have Describe and Write ACLs, and topics that fulfil a sink have Describe and Read ACLs.


In some embodiments, the ENSURE_SOURCE command can include a name (a label for the source), a principal (one or more principals that can produce data), a namespace (to be prefixed to created topics), a channel name (to be added to the channel portion of the topic), a channel format (a formatting guide for the channel, for example, #suffix# <==>${superpod}.${pod}.${channelName}), a scope (the dimensions by which topics should be created, for example, site, pod, superpod), a pod environment (that can be used with the scope is “pod,” for example, production or sandbox), a number of partitions for each topic, and/or any specific overrides. Other and/or different formats and elements can also be supported.


The following is an example onboard logbus producer:

















{



“principal”:“logbus”, // PKI cert also deployed to core app hosts



“name”:“logbus-pod-prod”,



“namespace”:“sfdc.prod.logbus”,



“scope”:“pod”,



“podEnvironment”:“production”,



“numPartitions”:15,



“channelName”:“logs.coreapp”,



“channelFormat”:“#suffix#”,



“overrides”:{









“retention.ms”:“345600000”,



“max.message.bytes”:“1048576”,



“retention.bytes”:“219902325555” // 205 GiB per partition



}









}










The following is an example onboard splunk consumer:

















{



“principal”:“splunk”,



“name”:“splunk coreapp prod logs consumer”,



“scope”:“aggregate”,



“sources”:[









{



“principal”:“logbus”,



“name”:“logbus-pod-prod”



}









]









}











These are just one example each of producers and consumers. Other configurations, types, structures, specifics can be supported.


In some embodiments, the ENSURE_SINK command can include a name (a label for the sink), a principal (one or more principals that can consume data), and/or a scope (the dimensions by which topics should be created, for example, site, aggregate). Other and/or different formats and elements can also be supported.


As described in greater detail below, the techniques and architectures described herein can be utilized to provide customer self-service in terms of simple and timely onboarding, clear visibility about resources, lifecycle controls with guardrails and/or access controls. In one embodiment, the automation techniques described herein cannot be utilized to automate the exfiltration of data. These techniques, therefore, provide improved cluster visibility, cluster self-healing and/or improved, automatic load balancing.


As used herein, the event-source design architecture described can be used for configuration, which is a new and different approach to system/data center configuration. In one embodiment, all commands going to the system are stored centrally, and each cluster controller, which are distributed globally, checks these commands to determine if there is work to be done locally. This asynchronous model enables the eventual march forward of consistency and unblocks new sites coming online. A new site, for example, proceeds through the ordered list of commands to be configured in the same way as other sites.


That is, the system uses the event sourcing design pattern to asynchronously march global actual state to desired state. Desired state can be changed when an external command is issued to the system such as create-source or create-sink. New commands result in local cluster controllers detecting a difference between actual and desired state within their purview. The local controller immediately begins the work of modifying it's clusters to remediate the deltas. Eventually the global network of controllers completes which means the global state satisfies all commands. Communication within the system is via an immutable stream of events stored centrally. The events never expire and can be replayed to satisfy unexpected future business requirements.


In one embodiment, when a new cluster comes online the local controller will detect that its actual state is empty which contradicts the desired state. The local controller can remediate the deltas by updating the new cluster with topics, ACLs and quotas to support the desired set of sources and sinks. Various embodiments can utilize the architecture as illustrated in FIG. 2.



FIG. 2 is a block diagram of one embodiment of a distributed architecture that can be utilized to provide the functionality described herein. In some embodiments, the same operations are applied to every pod. This simplifies configuration and can also be implemented using the publish-subscribe techniques described herein. In some embodiments, at least the data collection engine, local cluster and data replication engine are included in every data center.


Producer 200 can be any type of producer as described above with respect to FIG. 1. Further, any number of producers can be supported. Data collection engine 210 functions to collect records (e.g., logs) from producers (e.g., 200). Conceptually, this is the start of the pipeline and can be used as a reference point for the measurement of completeness.


In various embodiments, local cluster 220 functions as a data processor that receives records from one or more data collection engines and publishes them to data replication engine 230. Local cluster 220 can be, for example, a Kafka cluster, co-located in each data center that receives records from data collection engine 210 and manages the records in topics before the records are published to data replication engine 230.


In one embodiment, data replication engine 230 functions to replicate records from local cluster 220 to data aggregator 240. In one embodiment, data aggregator 240 is located in a secure location and can receive records from any number of local clusters and data replication engines. In one embodiment, data aggregator 240 can be a Kafka cluster that manages received records in topics.


In some embodiments, one or more consumers (e.g., 270) can consume records from one or more topics as provided by data aggregator 240. In some embodiments, records from data aggregator 240 can be stored in data storage system 250. For example, a MapReduce job (e.g., Apache Kafka Camus) can consume records from data aggregator 240 and serialize them for storage in data storage system 250, which can be, for example, a large-scale distributed Apache Hadoop ecosystem. In some embodiments, records can be stored in data storage system 250 before they are post processed for consumers.


In one embodiment, the ENSURE_SOURCE command can be utilized to create per-pod producer topics in local cluster 220. In one embodiment, the ENSURE_SOURCE command can also be utilized to set producer quotas in local cluster 220. In one embodiment, the ENSURE_SOURCE command can be utilized to set producer ACLs for local cluster 220.


In one embodiment, the ENSURE_SINK command can be utilized to create per-pod topics in data aggregator 240. In one embodiment, the ENSURE_SINK command can be utilized to set producer quotas and/or set producer ACLs in data replication engine 230. In one embodiment, the ENSURE_SINK command can be utilized to set consumer quotas and/or set consumer ACLs in data replication engine 230.


In one embodiment, the ENSURE_SINK command can be utilized to notify data replication engine 230 to mirror pod topics to data aggregator 240. In one embodiment, the ENSURE_SINK command can be utilized to create consumer quotas and/or set consumer ACLs for consumer 270. Thus, the architecture of FIGS. 1 and 2 can provide a flexible and scalable environment in which data can be managed and utilized, for example, for onboarding or configuration purposes. A few use case examples are provided herein; however, the concepts have broader applicability than the few examples included herein.


As discussed above, sources and sinks of data can be defined by templates that can be applied globally or to one or more data centers. Thus, application of the template definitions can initiate (or manage) the production and consumption of data (e.g., configuration operations) within a new data center (or other complex system) based on information from any number of sources. By managing the configuration information in topics, the appropriate ordering can be consistently provided. Further, templates that can require replication that can control accesses, which is a new control on data flows. Further, orchestration of, for example, Kafka clusters can give customers ability to define future states through centralized command and control.


The techniques described herein support user self-service with respect to devices or groups of devices (e.g., data centers), which eliminates or reduces the need for manual administration of these devices. Further, templated integrations can be supported that can allow new groups of devices to come online automatically configured. In various embodiments, data sources and data sinks can be conceptualized, which abstracts away from the standard “building blocks” of Kafka. Conceptually, the standard use of topics, producer ACLs, consumer ACLs and mirroring patterns are not utilized. Rather, in one embodiment, a source is defined in one group of devices (or system) and a sink is defined in another group of devices (or a different system) manages the configuration of core building blocks.



FIG. 3 is a flow diagram of one embodiment of a technique for managing configurations. The operations of FIG. 3 are illustrated and described in a particular order. However, this ordering can be modified, for example, configuration of producers can be performed before or in parallel with configuration of consumers.


One or more templates can be built and/or maintained that can be utilized to configure producers, consumers and/or topics 310. These templates can be generated and maintained in any appropriate format. The templates can be used to define configurations or future states for any number of producers, consumers, topics and/or clusters.


As discussed above, producers and consumers of data can be defined by templates that can be applied globally or to one or more data centers that can require replication that can control accesses, which is a new control on data flows. Further, orchestration of, for example, topics within Kafka clusters (or similar) provides the ability to define future states. Thus, centralized application command and control can be provided. An example application of the techniques described herein can be utilized to support user self-service with respect to devices or groups of devices (e.g., data centers), which eliminates or reduces the need for manual administration of these devices.


In one embodiment, records from one or more topics can be utilized to initiate configurations, 320. As discussed above, templated integrations can be supported that can allow new groups of devices to come online automatically configured. In various embodiments, data sources and data sinks can be conceptualized, which abstracts away from standard techniques. Conceptually, in one embodiment, a source is defined in one group of devices (or system) and a sink is defined in another group of devices (or a different system) manages the configuration of core building blocks.


One or more consumer devices are configured using records from one or more topics, 330. Similarly, one or more producer devices are configured from one or more topics, 340. The configured producers, consumers, topics, etc. can begin operations once sufficiently configured, 350.



FIG. 4 is a block diagram of one embodiment of a global configuration agent. In one embodiment, one or more global configuration agents may exist and/or operate within a host environment. The agent of FIG. 4 may provide functionality as described, for example, with respect to FIGS. 1-3. The agent of FIG. 4 may also provide additional functionality.


In one embodiment, global configuration agent 400 includes control logic 410, which implements logical functional control to direct operation of global configuration agent 400, and/or hardware associated with directing operation of global configuration agent 400. Logic may be hardware logic circuits and/or software routines. In one embodiment, global configuration agent 400 includes one or more applications 412, which represent a code sequence and/or programs that provide instructions to control logic 410.


Global configuration agent 400 includes memory 414, which represents a memory device and/or access to a memory resource for storing data and/or instructions. Memory 414 may include memory local to global configuration agent 400, as well as, or alternatively, including memory of the host system on which global configuration agent 400 resides. Global configuration agent 400 also includes one or more interfaces 416, which represent access interfaces to/from (an input/output interface) global configuration agent 400 with regard to entities (electronic or human) external to global configuration agent 400.


Global configuration agent 400 also includes global configuration engine 420, which represents one or more functions or modules that enable global configuration agent 400 to provide the functionality as described above. The example of FIG. 4 provides several modules that may be included in global configuration engine 420; however, different and/or additional modules may also be included. Example modules that may be involved in providing the functionality described herein include, for example, template module 430, topic module 435, cluster module 440, Producer module 445 and/or consumer module 450. Each of these modules may further include other sub-modules to provide other functions. As used herein, a module refers to routine, a subsystem, logic circuit, microcode, etc., whether implemented in hardware, software, firmware or some combination thereof.


In various embodiments, template module 430 can function to support creation and/or maintenance of templates to be utilized as described herein. For example, template module 430 can receive information via a graphical user interface (GUI, not illustrated in FIG. 4) that can provide parameters for future operations and/or desired future states to be achieved. Template module 430 can also maintain one or more templates in, for example, memory 414 which may include (or provide access to) a database.


In various embodiments topic module 435 can function to support creation and/or maintenance of one or more topics (e.g., within a Kafka cluster) that can be utilized to manage one or more records. In one embodiment, topic module 435 can function to manage created topics. In one embodiment, topic module 435 can function to configure topics based on one or more templates.


In various embodiments, cluster module 440 functions to support creation and/or maintain of one or more clusters (e.g., Kafka clusters) that can be utilized to manage one or more records. In one embodiment, cluster module 440 can function to manage and/or update created topics. In one embodiment, cluster module 440 can function to configure topics based on one or more templates.


In various embodiments, producer module 445 functions to support creation and/or operation of one or more producers that can be function to produce one or more records. In one embodiment, producer module 445 can function to create records for configuration of other devices and/or be configured to produce records for one or more topics.


In various embodiments, consumer module 450 functions to support creation and/or operation of one or more consumers that can be function to consume one or more records. In one embodiment, consumer module 450 can function to consume records for configuration of devices and/or be configured to consume records from one or more topics.


Reference in the specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.


While the invention has been described in terms of several embodiments, those skilled in the art will recognize that the invention is not limited to the embodiments described, but can be practiced with modification and alteration within the spirit and scope of the appended claims. The description is thus to be regarded as illustrative instead of limiting.

Claims
  • 1. A non-transitory computer-readable medium having stored thereon instructions that, when executed by one or more processors, cause the one or more processors to: maintain at least one template having definitions for managing production and consumption of data within a group of computing resources, the definitions defining configuration information for a set of sources, a set of sinks, and a set of topics within the group of computing resources;utilize records organized by topic collected from multiple disparate producers to initiate configuration of the group of computing resources, wherein records within a topic are organized by a corresponding topic sequence;configure a first configuration for at least a first portion of the computing resources as consumers based on a definition of the set of sinks and a definition in the set of topics in the at least one template, the consumers to consume records at a pace independent of record production, wherein the set of sinks spans the first portion;configure a second configuration for at least a second portion of the computing resources as producers based on a definition of the set of sources and a definition in the set of topics in the at least one template, the producers to produce records at a pace independent of record consumption, wherein the first portion is different from the second portion and wherein the set of sources spans the second portion; andresponsive to an addition of a new computing resource to the group of computing resources, detect at least one difference between an actual and desired state of the new computing resource, and remediate the at least one difference by updating the new computing resource with the set of topics to support the set of sources and the set of sinks.
  • 2. The non-transitory computer-readable medium of claim 1 wherein the at least one template further includes definitions for at least one topic partition to be used for the production and consumption of data.
  • 3. The non-transitory computer-readable medium of claim 1 wherein the at least one template comprises at least one parameter to cause one or more producer topics to be created in a local cluster that is local to the set of sources.
  • 4. The non-transitory computer-readable medium of claim 1 wherein the at least one template comprises at least one producer access control list (ACL) for a cluster local to the set of sources.
  • 5. The non-transitory computer-readable medium of claim 1 wherein the at least one template comprises at least one parameter to cause one or more topics to be created in an aggregator local to the consumers.
  • 6. The non-transitory computer-readable medium of claim 1 wherein the at least one template comprises at least one access control list (ACL) for a data replication engine.
  • 7. The non-transitory computer-readable medium of claim 1 wherein the at least one template comprises at least one consumer access control list (ACL) for cluster local to the set of sinks.
  • 8. A system comprising: a memory system; andat least one hardware processor coupled with the memory system, the at least one hardware processor configurable to: maintain at least one template having definitions for managing production and consumption of data within a group of computing resources, the definitions defining configuration information for a set of sources, a set of sinks, and a set of topics within the group of computing resources,utilize records organized by topic collected from multiple disparate producers to initiate configuration of the group of computing resources, wherein records within a topic are organized by a corresponding topic sequence,configure a first configuration for at least a first portion of the computing resources as consumers based on a definition of the set of sinks and a definition in the set of topics in the at least one template, the consumers to consume records at a pace independent of record production, wherein the set of sinks spans the first portion,configure a second configuration for at least a second portion of the computing resources as producers based on a definition of the set of sources and a definition in the set of topics in the at least one template, the producers to produce records at a pace independent of record consumption, wherein the first portion is different from the second portion and wherein the set of sources spans the second portion, andresponsive to an addition of a new computing resource to the group of computing resources, detect at least one difference between an actual and desired state of the new computing resource, and remediate the at least one difference by updating the new computing resource with the set of topics to support the set of sources and the set of sinks.
  • 9. The system of claim 8 wherein the at least one template further includes definitions for at least one topic partition to be used for the production and consumption of data.
  • 10. The system of claim 8 wherein the at least one template comprises at least one parameter to cause one or more producer topics to be created in a local cluster that is local to the set of sources.
  • 11. The system of claim 8 wherein the at least one template comprises at least one producer access control list (ACL) for a cluster local to the set of sources.
  • 12. The system of claim 8 wherein the at least one template comprises at least one parameter to cause one or more topics to be created in an aggregator local to the consumers.
  • 13. The system of claim 8 wherein the at least one template comprises at least one access control list (ACL) for a data replication engine.
  • 14. The system of claim 8 wherein the at least one template comprises at least one consumer access control list (ACL) for a cluster local to the set of sinks.
  • 15. A method comprising: maintaining at least one template having definitions for managing production and consumption of data within a group of computing resources, the definitions defining configuration information for a set of sources, a set of sinks, and a set of topics within the group of computing resources;utilizing records organized by topic collected from multiple disparate producers to initiate configuration of the group of computing resources, wherein records within a topic are organized by a corresponding topic sequence;configuring a first configuration for at least a first portion of the computing resources as consumers based on a definition of the set of sinks and a definition in the set of topics in the at least one template, the consumers to consume records at a pace independent of record production, wherein the set of sinks spans the first portion;configuring a second configuration for at least a second portion of the computing resources as producers based on a definition of the set of sources and a definition in the set of topics in the at least one template, the producers to produce records at a pace independent of record consumption, wherein the first portion is different from the second portion and wherein the set of sources spans the second portion; andresponsive to an addition of a new computing resource to the group of computing resources, detect at least one difference between an actual and desired state of the new computing resource, and remediate the at least one difference by updating the new computing resource with the set of topics to support the set of sources and the set of sinks.
  • 16. The method of claim 15 wherein the at least one template further includes definitions for at least one topic partition to be used for the production and consumption of data.
  • 17. The method of claim 15 wherein the at least one template comprises at least one parameter to cause one or more producer topics to be created in a local cluster that is local to the set of sources.
  • 18. The method of claim 15 wherein the at least one template comprises at least one producer access control list (ACL) for a cluster local to the set of sources.
  • 19. The method of claim 15 wherein the at least one template comprises at least one parameter to cause one or more topics to be created in an aggregator local to the consumers.
  • 20. The method of claim 15 wherein the at least one template comprises at least one access control list (ACL) for a data replication engine.
  • 21. The method of claim 15 wherein the at least one template comprises at least one consumer access control list (ACL) for a cluster local to the set of sinks.
US Referenced Citations (164)
Number Name Date Kind
5577188 Zhu Nov 1996 A
5608872 Schwartz et al. Mar 1997 A
5649104 Carleton et al. Jul 1997 A
5715450 Ambrose et al. Feb 1998 A
5761419 Schwartz et al. Jun 1998 A
5819038 Carleton et al. Oct 1998 A
5821937 Tonelii et al. Oct 1998 A
5831610 Tonelii et al. Nov 1998 A
5873096 Lim et al. Feb 1999 A
5918159 Fomukong et al. Jun 1999 A
5963953 Cram et al. Oct 1999 A
6092083 Brodersen et al. Jul 2000 A
6169534 Raffel et al. Jan 2001 B1
6178425 Brodersen et al. Jan 2001 B1
6185613 Lawson Feb 2001 B1
6189011 Lim et al. Feb 2001 B1
6216135 Brodersen et al. Apr 2001 B1
6233617 Rothwein et al. May 2001 B1
6266669 Brodersen et al. Jul 2001 B1
6295530 Ritchie et al. Sep 2001 B1
6324568 Diec Nov 2001 B1
6324693 Brodersen et al. Nov 2001 B1
6336137 Lee et al. Jan 2002 B1
D454139 Feldcamp Mar 2002 S
6367077 Brodersen et al. Apr 2002 B1
6393605 Loomans May 2002 B1
6405220 Brodersen et al. Jun 2002 B1
6434550 Warner et al. Aug 2002 B1
6446089 Brodersen et al. Sep 2002 B1
6535909 Rust Mar 2003 B1
6549908 Loomans Apr 2003 B1
6553563 Ambrose et al. Apr 2003 B2
6560461 Fomukong et al. May 2003 B1
6574635 Stauber et al. Jun 2003 B2
6577726 Huang et al. Jun 2003 B1
6601087 Zhu et al. Jul 2003 B1
6604117 Lim et al. Aug 2003 B2
6604128 Diec Aug 2003 B2
6609150 Lee et al. Aug 2003 B2
6621834 Scherpbier et al. Sep 2003 B1
6654032 Zhu et al. Nov 2003 B1
6665648 Brodersen et al. Dec 2003 B2
6665655 Warner et al. Dec 2003 B1
6684438 Brodersen et al. Feb 2004 B2
6711565 Subramaniam et al. Mar 2004 B1
6724399 Katchour et al. Apr 2004 B1
6728702 Subramaniam et al. Apr 2004 B1
6728960 Loomans Apr 2004 B1
6732095 Warshavsky et al. May 2004 B1
6732100 Brodersen et al. May 2004 B1
6732111 Brodersen et al. May 2004 B2
6754681 Brodersen et al. Jun 2004 B2
6763351 Subramaniam et al. Jul 2004 B1
6763501 Zhu et al. Jul 2004 B1
6768904 Kim Jul 2004 B2
6782383 Subramaniam et al. Aug 2004 B2
6804330 Jones et al. Oct 2004 B1
6826565 Ritchie et al. Nov 2004 B2
6826582 Chatterjee et al. Nov 2004 B1
6826745 Coker et al. Nov 2004 B2
6829655 Huang et al. Dec 2004 B1
6842748 Warner et al. Jan 2005 B1
6850895 Brodersen et al. Feb 2005 B2
6850949 Warner et al. Feb 2005 B2
7289976 Kihneman et al. Oct 2007 B2
7340411 Cook Mar 2008 B2
7620655 Larsson et al. Nov 2009 B2
7711783 Violleau May 2010 B1
9852186 Herwadkar Dec 2017 B2
10673969 Bheemarajaiah Jun 2020 B2
20010044791 Richter et al. Nov 2001 A1
20020022986 Coker et al. Feb 2002 A1
20020000355 Brodersen et al. Mar 2002 A1
20020029161 Brodersen et al. Mar 2002 A1
20020029376 Ambrose et al. Mar 2002 A1
20020042264 Kim Apr 2002 A1
20020042843 Diec Apr 2002 A1
20020072951 Lee et al. Jun 2002 A1
20020082892 Raffel et al. Jun 2002 A1
20020129352 Brodersen et al. Sep 2002 A1
20020140731 Subramaniam et al. Oct 2002 A1
20020143997 Huang et al. Oct 2002 A1
20020152102 Brodersen et al. Oct 2002 A1
20020161734 Stauber et al. Oct 2002 A1
20020162090 Parnell et al. Oct 2002 A1
20020165742 Robins Nov 2002 A1
20030004971 Gong et al. Jan 2003 A1
20030018705 Chen et al. Jan 2003 A1
20030018830 Chen et al. Jan 2003 A1
20030066031 Laane Apr 2003 A1
20030066032 Ramachadran et al. Apr 2003 A1
20030069936 Warner et al. Apr 2003 A1
20030070000 Coker et al. Apr 2003 A1
20030070004 Mukundan et al. Apr 2003 A1
20030070005 Mukundan et al. Apr 2003 A1
20030074418 Coker Apr 2003 A1
20030088545 Subramaniam et al. May 2003 A1
20030120675 Stauber et al. Jun 2003 A1
20030151633 George et al. Aug 2003 A1
20030159136 Huang et al. Aug 2003 A1
20030187921 Diec Oct 2003 A1
20030189600 Gune et al. Oct 2003 A1
20030191743 Brodersen et al. Oct 2003 A1
20030204427 Gune et al. Oct 2003 A1
20030206192 Chen et al. Nov 2003 A1
20030225730 Warner et al. Dec 2003 A1
20030225857 Flynn Dec 2003 A1
20040001092 Rothwein et al. Jan 2004 A1
20040010489 Rio Jan 2004 A1
20040015981 Coker Jan 2004 A1
20040027388 Berg et al. Feb 2004 A1
20040128001 Levin et al. Jul 2004 A1
20040186860 Lee et al. Sep 2004 A1
20040193510 Catahan, Jr. et al. Sep 2004 A1
20040199489 Barnes-Leon et al. Oct 2004 A1
20040199536 Barnes-Leon et al. Oct 2004 A1
20040199543 Braud et al. Oct 2004 A1
20040249854 Barnes-Leon et al. Dec 2004 A1
20040260534 Pak et al. Dec 2004 A1
20040260659 Chan et al. Dec 2004 A1
20040268299 Lei et al. Dec 2004 A1
20050050555 Exley et al. Mar 2005 A1
20050091098 Brodersen et al. Apr 2005 A1
20050203949 Cabrera Sep 2005 A1
20050204051 Box Sep 2005 A1
20060029054 Breh Feb 2006 A1
20080052727 Todd Feb 2008 A1
20090177744 Marlow et al. Jul 2009 A1
20120079066 Li Mar 2012 A1
20140082744 Behforooz Mar 2014 A1
20140143214 Kohl May 2014 A1
20140201135 Smid Jul 2014 A1
20150019655 Kizhakkiniyil Jan 2015 A1
20150200887 Cuomo Jul 2015 A1
20150254328 Dereszynski Sep 2015 A1
20160182652 Bevilacqua Jun 2016 A1
20160219089 Murthy Jul 2016 A1
20170041352 Martens Feb 2017 A1
20170078423 Diester Mar 2017 A1
20170126603 Chen May 2017 A1
20170132068 Parra May 2017 A1
20170300966 Dereszynski Oct 2017 A1
20170339238 Qin Nov 2017 A1
20180091586 Auradkar Mar 2018 A1
20180167476 Hoffner Jun 2018 A1
20180213018 Madani Jul 2018 A1
20180248771 Côté et al. Aug 2018 A1
20190044818 Nolan Feb 2019 A1
20190073373 Surale Mar 2019 A1
20190079818 Handa Mar 2019 A1
20190102266 Park Apr 2019 A1
20190130290 Kerber May 2019 A1
20190222640 Tock Jul 2019 A1
20190258756 Minwalla Aug 2019 A1
20190306121 Anderson Oct 2019 A1
20190377816 McCluskey Dec 2019 A1
20200042365 Tanna Feb 2020 A1
20200111033 Mahajan Apr 2020 A1
20200174854 Warshavsky Jun 2020 A1
20200183603 Takla Jun 2020 A1
20200293803 Wajs Sep 2020 A1
20200293920 Chung Sep 2020 A1
20200402058 Zhou Dec 2020 A1
20210149752 Burroughes May 2021 A1
Related Publications (1)
Number Date Country
20200326991 A1 Oct 2020 US
Provisional Applications (1)
Number Date Country
62832849 Apr 2019 US