Systems and methods for isolating applications associated with multiple tenants within a computing platform

Information

  • Patent Grant
  • 11943319
  • Patent Number
    11,943,319
  • Date Filed
    Friday, April 28, 2023
    a year ago
  • Date Issued
    Tuesday, March 26, 2024
    9 months ago
Abstract
Systems and methods for isolating applications associated with multiple tenants within a computing platform receive a request from a client associated with a tenant for running an application on a computing platform. Hosts connected to the platform are associated with a network address and configured to run applications associated with multiple tenants. A host is identified based at least in part on the request. One or more broadcast domain(s) including the identified hosts are generated. The broadcast domains are isolated in the network at a data link layer. A unique tenant identification number corresponding to the tenant is assigned to the broadcast domains. In response to launching the application on the host: the unique tenant identification number is assigned to the launched application and is added to the network address of the host; and the network address of the host is sent to the client associated with the tenant.
Description
BACKGROUND

Certain embodiments of the present invention are directed to computing platforms including hosts connected through a network. More particularly, some embodiments of the present invention provide systems and methods for isolating applications associated with multiple tenants within a computing platform.


With the number of customers of a computing platform increasing, the demands on data security provided by the computing platform are also increasing. In some examples, customers (e.g., tenants) are running multi-tenant applications on multiple hosts across the computing platform sharing resources and the network of the computing platform. Thus, managing network traffic and ensuring data security across a multi-tenant computing platform becomes increasingly more complex with the number of customers (e.g., tenants) increasing. For example, conventional computing platforms generally partition the network and hosts for running the applications on a per customer (e.g., per tenant) basis by using subnets and firewalls, which are difficult to implement and manage for an increased number of customers (e.g., tenants) of the platform.


Hence it is highly desirable to improve the techniques for isolating applications associated with multiple tenants within a computing platform.


SUMMARY

Certain embodiments of the present invention are directed to computing platforms including hosts connected through a network. More particularly, some embodiments of the present invention provide systems and methods for isolating applications associated with multiple tenants within a computing platform.


According to some embodiments, method for isolating applications associated with multiple tenants within a computing platform. For example, a method includes receiving a request from a client associated with a tenant for running an application on a computing platform. The computing platform includes a plurality of hosts connected through a network. Each host is associated with a network address and configured to run applications associated with multiple tenants. The method further includes identifying one or more hosts of the plurality of hosts based at least in part on the request. The method further includes generating one or more broadcast domains including the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The method further includes assigning to the one or more broadcast domains a unique tenant identification number corresponding to the tenant. The method further includes launching the application on at least one host of the identified one or more hosts. In response to launching the application on the at least one host: the unique tenant identification number is assigned to the launched application; the unique tenant identification number is added to the network address of the at least one host; and the network address of the at least one host is sent to the client associated with the tenant. The method is performed using one or more processors. The network address may further include a unique cluster identification number, the unique cluster identification number may be associated with a cluster of the computing platform; and the cluster may be associated with the tenant. The unique tenant identification number may include the unique cluster identification number. The network address may include a plurality of bit sequences; and one bit sequence of the plurality of bit sequences may include the unique tenant identification number. The plurality of bit sequences of the network address may include at least 128 bits. The broadcast domains may include virtual local area networks. The launched application may be included in a container; and the container may be associated with the unique tenant identification number. The container may be included in a pod running on the at least one host; the pod may include one or more containers; and each container of the one or more containers may be associated with one tenant of the computing platform. Each container of the one or more containers may be associated with a different tenant of the computing platform. The pod may map to the network address. The network address may further include a region identification number, a network device identification number, or a virtual interface of the at least one host. The one or more broadcast domain may map to the network address.


According to certain embodiments, a system for isolating applications associated with multiple tenants within a computing platform includes a plurality of hosts connected through a network and a fleet controller. Each host is associated with a network address and configured to run applications associated with multiple tenants on a computing platform. The fleet controller is configured to, in response to receiving a first request from a client associated with a tenant for running an application on the computing platform, identify one or more hosts of the plurality of hosts based at least in part on the request. The fleet controller is further configured to generate one or more broadcast domains including the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The fleet controller is further configured to assign to the one or more broadcast domains a unique tenant identification number corresponding to the tenant. The fleet controller is further configured to send a second request to a scheduler for launching the application on at least one host of the identified one or more hosts. The fleet controller is further configured to, in response to receiving confirmation from the scheduler of the application being launched on the at least one host: assign the unique tenant identification number to the launched application; add the unique tenant identification number to the network address of the at least one host; and send the network address of the at least one host to the client associated with the tenant. The network address may further include a unique cluster identification number; the unique cluster identification number may be associated with a cluster of the computing platform; and the cluster may be associated with the tenant. The network address may include a plurality of bit sequences; and one bit sequence of the plurality of bit sequences may include the unique tenant identification number. The plurality of bit sequences of the network address may include at least 128 bits. The broadcast domains may include virtual local area networks.


According to some embodiments, a system for isolating applications associated with multiple tenants within a computing platform includes a client associated with a tenant and configured to request running an application on a computing platform. The computing platform includes a plurality of hosts connected through a network. Each host is associated with a network address and is configured to run applications associated with multiple tenants. The client is further configured to send a request for running the application on the computing platform. The client is further configured to, in response to sending the request for running the application on the computing platform, receive the network address of at least one host of the plurality of hosts. One or more hosts of the plurality of hosts are identified based at least in part on the request. The identified one or more hosts include the at least one host. One or more broadcast domains are generated to include the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The one or more broadcast domains are assigned to a unique tenant identification number corresponding to the tenant. The application is launched on the at least one host of the identified one or more hosts. The launched application is assigned to the unique tenant identification number. The unique tenant identification number is added to the network address of the at least one host. The network address may further include a unique cluster identification number; the unique cluster identification number may be associated with a cluster of the computing platform; and the cluster may be associated with the tenant. The network address may include a plurality of bit sequences; and one bit sequence of the plurality of bit sequences may include the unique tenant identification number.


Depending upon embodiment, one or more benefits may be achieved. These benefits and various additional objects, features and advantages of the present invention can be fully appreciated with reference to the detailed description and accompanying drawings that follow.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a simplified diagram showing a system for isolating applications associated with multiple tenants within a computing platform according to one embodiment of the present invention.



FIG. 2 is a simplified diagram showing the system for isolating applications associated with multiple tenants within the computing platform as shown in FIG. 1 according to one embodiment of the present invention.



FIG. 3 is a simplified diagram showing the system for isolating applications associated with multiple tenants within the computing platform as shown in FIG. 1 according to one embodiment of the present invention.



FIG. 4 is a simplified diagram the system for isolating applications associated with multiple tenants within the computing platform as shown in FIG. 1 according to one embodiment of the present invention.



FIG. 5 is a simplified diagram showing a method for isolating applications associated with multiple tenants within a computing platform according to one embodiment of the present invention.



FIG. 6 is a simplified diagram showing a computing system for implementing a system for isolating applications associated with multiple tenants within a computing platform according to one embodiment of the present invention.





DETAILED DESCRIPTION

Conventional systems and methods are often not capable of efficiently isolating applications associated with multiple tenants within a multi-tenant computing platform. Conventional systems and methods typically partition the network of a computing platform and corresponding network addresses on a per customer (e.g., per tenant) basis by using subnets and firewalls. This increases the complexity and cost of managing applications in a multi-tenant environment across the platform and makes it difficult to scale the platform when the number of customers (e.g., the number of tenants) increases. For example, each customer is assigned its own subnets of network addresses and responsible for configuring and managing the assigned subnets. In turn, the use of subnets by conventional systems and methods requires the use of security groups around each subnet, including, for example, firewalls, to guarantee the security of customer data communicated across the platform.


In some embodiments, one or more solutions rooted in computer technology overcome one or more problems specifically arising in the realm of computer technology, including that of security of customer data. Some embodiments are directed to computing platforms including hosts connected through a network. More particularly, some embodiments of the present invention provide systems and methods for isolating applications associated with multiple tenants within a computing platform. In some examples, the hosts of the computing platform include virtual servers and/or virtual machines. In certain examples, the computing platforms include a virtual computing environment that provides an operating system and/or an application server for running one or more containers. For example, a container includes a containerized application. In some examples, one or more containers run on a server or host machine of the computing platform and are associated with particular resources that include CPU, memory, storage, and/or networking capacity. In certain examples, the hosts of the computing platform include physical servers and/or physical machines.


In certain embodiments, systems and methods are configured to isolate applications (e.g., containers) on a per tenant and per host basis by assigning to each application (e.g., each container) a unique tenant identification number corresponding to a particular tenant of the computing platform and embedding the unique tenant identification number in a network address of a host running the application (e.g., container). In some examples, the systems and methods are further configured to isolate applications (e.g., containers) associated with different tenants at the data link layer by generating a broadcast domain including the host, assign the broadcast domain to the unique tenant identification number, and run the applications (e.g., the containers) associated with the unique tenant identification number in the broadcast domain of the host. In certain examples, the broadcast domain associated with the unique tenant identification number is mapped to the network address including the unique tenant identification number.


According to some embodiments, benefits include significant improvements, including, for example, increased efficiency, reduced complexity, and improved scalability, in managing an increased number of tenants across a multi-tenant computing platform. In certain embodiments, other benefits include increased data security for each tenant on a multi-tenant computing platform. In some embodiments, systems and methods are configured to isolate application data from different tenants across a multi-tenant computing platform.



FIG. 1 is a simplified diagram showing a system 100 for isolating applications associated with multiple tenants within a computing platform 102 according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. The system 100 includes the computing platform 102 and a network 104. In some examples, the computing platform 102 includes a plurality of hosts. For example, the plurality of hosts includes hosts 1061-m, 1081-n. As an example, hosts 1061-m represents hosts 1061, . . . , 106m as shown in FIG. 1, and hosts 1081-n represents hosts 1081, . . . , 108n as shown in FIG. 1. As an example, each host of the hosts 1061-m, 1081-n is configured to be connected to other components of the computing platform 102 through the network 104. As an example, each host of the hosts 1061-m, 1081-n is associated with a network address. In one example, each host of the hosts 1061-m, 1081-n is configured to run applications associated with multiple tenants. In certain examples, the computing platform 102 includes one or more networking devices 1101-N. For example, networking devices 1101-N represents networking devices 1101, . . . , 110N as shown in FIG. 1. As an example, each networking device of the one or more networking devices 1101-N is configured to be connected through the network 104. In one example, each host of the hosts 1061-m, 1081-n is configured to be connected to one or more networking devices 1101-N through the network 104. In certain examples, the network 104 includes at least three networking layers (e.g., a physical layer or layer 1, a data link layer or layer 2, and a network layer or layer 3). For example, the network 104 includes an IPv4 network, an IPv6 network, or any combination thereof. In some examples, the computing platform 102 includes a plurality of racks. For example, each rack of the plurality of racks includes one or more hosts and one or more networking devices. As an example, the computing platform 102 includes N racks with the first rack including the networking device 1101 and the hosts 1061-m, . . . , and the Nth rack including the networking device 110N and the hosts 1081-n. In certain examples, the networking devices 1101-N of the racks include top-of-rack (ToR) switches.


In some embodiments, the computing platform 102 includes a cluster computing platform including clusters of one or more server or host machines (e.g. one or more hosts of the hosts 1061-m, 1081-n). In some examples, the computing platform 102 includes a distributed computing platform that allows the one or more client devices 1101-M to distribute applications and/or data over the network 104 to the cluster of servers or host machines (e.g. clusters of the hosts 1061-m, 1081-n). For example, client devices 1101-M represents client devices 1121, . . . , 112M as shown in FIG. 1. In certain examples, the computing platform 102 includes a cloud computing platform that allows the one or more client devices 1121-M access to remote servers, data storages, networks, devices, applications and/or data resources over the network 104. For example, multiple customers (e.g., multiple tenants) through the one or more client devices 1121-M store data at the data storages of the cloud computing platform. In other examples, the computing platform 102 is associated with a platform provider that provides the platform to multiple customers (e.g., multiple tenants). For example, customers (e.g., tenants) of the computing platform 102 include individuals, organizations and/or commercial companies.


In certain embodiments, the one or more servers or host machines (e.g., the one or more hosts of the hosts 1061-m, 1081-n) are divided into in one or more regions. For example, a region represents a geographic area that the one or more servers or host machines are located within. As an example, each region relates to a different geographic area. In other examples, each region of the one or more servers or host machines includes one or more separate zones. For example, each server or host machine within a region is associated with only one zone of the one or more separate zones associated with the region. As an example, each zone within a region is isolated from any other zone within the region. In one example, each zone within a region is connected with any other zone within the region through low-latency links. In some examples, the computing platform 102 is configured to not replicate applications and/or resources across different regions. For example, each region is completely independent from any other region of the computing platform 102.


According to some embodiments, the computing platform 102 includes a container-orchestration platform. In some examples, the container-orchestration platform allows for automated deployment, scaling and/or operations of containers across the platform. For example, the container-orchestration platform employs the containers across the one or more servers or host machines (e.g., one or more hosts of the hosts 1061-m, 1081-n) of the computing platform 102. In some examples, a pod of the computing platform 102 represents a basic scheduling unit of work on the computing platform 102. In certain examples, the pod includes one or more containers. In other examples, one or more pods of the computing platform 102 provide a service to the one or more client devices 1121-M.


According to certain embodiments, a container of the computing platform 102 includes one or more applications. In some examples, the container also includes data and libraries associated with the one or more applications. For example, the container allows the one and more applications and their associated data and libraries to be co-located on the same server or host machine (e.g., the same host of the hosts 1061-m, 1081-n). In one example, the container allows the one or more applications and their associated data and libraries to share resources. For example, the shared resources include CPU, memory, storage, and/or networking capacity. As an example, the container represents the lowest level of a micro-service of the computing platform 102. In one example, the micro-service includes the one or more applications, libraries and the applications' dependencies.


In some embodiments, the computing platform 102 includes a fleet controller 114. In some examples, the computing platform 102 includes a fleet catalog 116, a hardware (HW) controller 118, a hardware catalog 120, a control plane application 122, a fleet health component 124, a fleet scheduler 126, a hardware health component 128, and a hardware scheduler 130. For example, the control plane application 122 is configured to schedule and manage applications that run on the hosts 1061-m, 1081-n. As an example, the control plane application 122 is configured to manage regions, tenants and node (e.g., host) assignments of the computing platform 102. In certain examples, the control plane application 122 is configured to manage workloads and communications between applications running on the hosts 1061-m, 1081-n.


In certain embodiments, the computing platform 102 is configured to provide services to tenants based at least in part on two abstract layers including a fleet layer and a hardware layer. In some examples, the fleet layer includes logical states and entities of components of the computing platform 102. For example, logical entities include a logical entity associated with a cluster of 64 nodes (e.g., hosts). In one example, logical entities include a logical entity associated with three publicly routable IP addresses. As an example, the hardware layer includes actual physical components and resources (e.g., hardware components) of the computing platform 102. In other examples, the organization of the two abstract levels of computing platform 102 is symmetrical with respect to the services provided by the computing platform. In some examples, the fleet catalog 116 and the hardware catalog 120 store data and sources of truth relating to the state of the two abstract layers, respectively. In some examples, the fleet controller 114 is configured to actuate an actual logical state of the computing platform 102 that matches a desired logical state stored in the fleet catalog 116. In certain examples, the hardware controller 118 is configured to actuate a physical state of the computing platform 102 that matches a desired physical state stored in the hardware catalog 120. For example, the actual logical state represents a state that corresponds to the actual physical state of the computing platform 102.


According to some embodiments, the fleet controller 114 is configured to receive a request from a client (e.g., a client device) associated with a tenant for running an application on the computing platform 102. For example, each client device of the client devices 1101-M is associated with a different customer (e.g., a different tenant) of the multi-tenant computing platform 102. In some examples, the fleet controller 114 is configured to send the received request for storing to the fleet catalog 116. In certain examples, the fleet controller 114 is configured to queue requests received from the client devices 1121-M and/or other components of the computing platform 102. For example, the fleet controller 114 is configured to provide a control loop for ensuring that a declared logical state in the fleet catalog 116 is satisfied. In certain examples, the fleet catalog 116 is configured to provide a source of truth for states of resources of the computing platform 102. As an example, states of resources include logical assignment of the hosts 1061-m, 1081-n and their status. For example, the fleet catalog 116 provides information that associates a tenant with a cluster of the computing platform 102.


According to certain embodiments, the fleet controller 114 is configured to monitor changes and/or updates of the states of resources included in the fleet catalog 116. For example, the fleet controller 114 is configured to retrieve a declared state of a resource from the fleet catalog 116. In some examples, the fleet controller 114 is configured to query the hardware catalog 120 for available hardware components of the computing platform 102. For example, the hardware catalog 120 is configured to provide a source of truth for the hardware components of the computing platform 102 and its state. As an example, states of hardware components of the computing platform 102 include host serial numbers, rack locators, ports, MAC addresses, internet protocol (IP) addresses, host images, host health, and power status of hardware components of the computing platform 102.


In some embodiments, the fleet controller 114 is configured to allocate and assign hardware components (e.g., physical machines and/or hosts 1061-m, 1081-n) to tenants and/or clusters of the computing platform 102. For example, the fleet controller 114 is configured to, in response to successful allocating the hardware components of the computing platform 102, update the hardware catalog 120 to reflect the allocation of the hardware components. As an example, the fleet controller 114 is configured to send reservations for hardware components to the hardware controller 118. In one example, the fleet controller 114 is configured to map allocated compute nodes (e.g., hosts) to clusters of the computing platform 102. For example, a certificate residing on a compute node (e.g., a host) includes information that associates the compute node with a cluster of the computing platform 102.


In certain embodiments, the hardware controller 118 is configured to monitor state changes in the hardware catalog 120. In certain examples, the hardware controller 118 is configured to, in response to determining state changes in the hardware catalog 120, actuate the changes in the corresponding hardware components of the computing platform 102. For example, state changes include assignments and/or reservations added by the fleet controller 114 to the hardware catalog 120. As an example, the hardware controller 118 is configured to provide a control loop for ensuring that a declared hardware state in the hardware catalog 120 is satisfied.


According to some embodiments, the hardware controller 118 is configured to, in response to the fleet controller 114 assigning the hardware components (e.g., the physical machines and/or the hosts 1061-m, 1081-n) to tenants and/or clusters of the computing platform 102, configure and/or connect the corresponding hardware components. For example, the hardware controller 118 is configured to provide IP addresses to the hardware components and connect the ToR switches, network interface controllers (NICs) and other components of the computing platform 102 to the hardware components. In some examples, the hardware controller 118 is configured to assign IP addresses on a per cluster basis. In certain examples, the hardware controller 118 is configured to read the hardware catalog 120 for reservations of compute nodes (e.g., hosts) of the computing platform 102. For example, the hardware controller 118 is configured to communicate to a networking device connected to one or more reserved compute nodes (e.g., hosts) a request for network addresses to be assigned to the one or more reserved compute nodes. As an example, the networking device is configured to, in response to receiving the request, allocate the network addresses and assign the allocated network addresses to the one or more reserved compute nodes (e.g., hosts) connected to the networking device.


According to certain embodiments, the hardware catalog 120 is configured to store information associated with the allocated network addresses of the one or more reserved compute notes (e.g., hosts), the networking device connected to the one or more reserved compute nodes, and/or the clusters associated with the one or more reserved compute nodes. In some examples, the hardware catalog 120 is configured to provide the actual hardware configuration of the computing platform 102 and record changes in the hardware configuration relating to the employment of physical servers and association of services and servers with certain customers (e.g., tenants). For example, the hardware catalog 120 is configured to provide information associated with mapping allocated compute nodes (e.g., allocated hosts) to clusters of the computing platform 102.


In some embodiments, the fleet scheduler 126 is configured to identify one or more hosts of the hosts 1061-m, 1081-n based at least in part on the request for resources by querying the fleet catalog 116 and/or the hardware catalog 120. In some examples, the hardware scheduler 130 is configured to determine a time when hardware components (e.g., physical machines and/or resources) are made available to the hardware controller 118 for employment to satisfy a declared state in the hardware catalog 120. In some examples, the fleet health component 124 is configured to poll services running on hosts and/or other components of the computing platform 102 based at least in part on entries in the fleet catalog 116. For example, the fleet health component 124, in response to receiving results from the queried services, logs the health and state of the declared resource in the fleet catalog 116. In certain examples, the fleet health component 124 is configured to generate a custom logic for polling services regarding different types of resources. For examples, types of resources include non-public types (e.g., resources of billing and logging components). In other examples, the hardware health component 128 is configured to update the hardware catalog 120 regarding the health and the state of hardware components of one or more regions of the computing platform. For example, the health includes information about the hardware components being operational, allocated and/or ready to operate. In some examples, the hardware health component 128 is configured to poll components of the computing platform 102 regarding their health and/or state. As an example, the hardware health component 128 is configured to push hardware changes and/or updates in the hardware catalog 120 to components of the computing platform 102 based at least in part on features supported by hardware standards implemented on the computing platform.


In certain embodiments, the fleet controller 114 is configured to read the fleet catalog 116 to determine changes in the state of resources requested by tenants of the multi-tenant computing platform 102. For example, the fleet catalog 116 includes a request by a tenant for a cluster of 64 hosts in two zones. In some examples, the fleet controller 114 is configured to, in response to determining changes in the state of resources requested by tenants, request the fleet scheduler 126 for scheduling the requested resources. In certain examples, the fleet scheduler 126 is configured to query the fleet catalog 116 in response to receiving a request for resources from the fleet controller 114. For example, the fleet scheduler 126 is configured to send a reply to the fleet controller 114 in response to querying the fleet catalog 116. As an example, the reply includes a determination whether all the requested resources or a percentage of the requested resources are available for employment.



FIG. 2 is a simplified diagram showing the system 100 for isolating applications associated with multiple tenants within the computing platform 102 according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. In some examples, the fleet controller 114 is configured to receive a request from the client (e.g., the client device 1121) associated with tenant 2001 for running the application 202 on the computing platform 102. In certain examples, the fleet controller 114 is configured to identify one or more hosts 1061, . . . 106k of the plurality of hosts (e.g., the hosts 1061-m) based at least in part on the request. For example, the fleet controller 114 is configured to store information associated with the identified hosts 1061, . . . 106k in the fleet catalog 116 and/or hardware catalog 120 for recording the state changes of the identified hosts 1061, . . . 106k. In some examples, the network 104 includes one or more broadcast domains of the computing platform 102. For example, broadcast domains are isolated from other broadcast domains at the data link layer of the network 104. In one example, each broadcast domain includes different hosts of the computing platform 102. In certain examples, the broadcast domains are associated with different tenants of the multi-tenant computing platform 102. For example, each broadcast domain is associated with one tenant. As an example, the one tenant associated with a broadcast domain is different from tenants associated with the other broadcast domains. In some examples, the broadcast domains include virtual local area networks (VLANs).


In some embodiments, the fleet controller 114 is configured to generate the broadcast domain 2041 including the identified one or more hosts 1061, . . . 106k. For example, the fleet controller 114 is configured to store information associated with the broadcast domain 2041 in the fleet catalog 116 and/or hardware catalog 120. In some examples, the broadcast domain 2041 is isolated from other broadcast domains in the network 104 at a data link layer 206. In certain examples, the networking device 1101 is connected to other platform components at a network layer 208 of the computing platform 102. In other examples, the broadcast domain 2041 includes a unique domain identification number (e.g., “1”).


In certain embodiments, the fleet controller 114 is configured to assign to the broadcast domain 2041 a unique tenant identification number (e.g., “1”) corresponding to tenant 2001. For example, the fleet controller 114 is configured to store information associated with the unique tenant identification number (e.g., “1”) and the assigned broadcast domain 2041 in the fleet catalog 116. In some examples, the unique tenant identification number (e.g., “1”) corresponds to the unique domain identification number (e.g., “1”) associated with the broadcast domain 2041. For example, the unique tenant identification number includes a bit sequence with a first portion of the bit sequence representing the unique domain identification number. In other examples, the unique tenant identification number (e.g., “1”) represents a route identification number. As an example, a second portion of the bit sequence of unique tenant identification number represents the route identification number.


According to some embodiments, the fleet controller 114 is configured to launch the application 202 on at least one host of the identified hosts 1061, . . . 106k. For example, the fleet controller 114 is configured to store information associated with the at least one host in the fleet catalog 116 and/or hardware catalog 120 for recording the state change of the at least one host.


According to certain embodiments, the fleet controller 114 is configured to, in response to launching the application 202 on the at least one host, assign the unique tenant identification number (e.g., “1”) to the launched application. In some examples, the launched application is included in a container associated with tenant 2001. In certain examples, the fleet controller 114 is configured to add the unique tenant identification (e.g., “1”) to the network address of the at least one host.


In some embodiments, the network addresses of the hosts 1061-m include unique tenant identification numbers associated with one or more tenants of the multi-tenant computing platform 102. In some examples, the unique tenant identification numbers relate to unique cluster identification numbers. For example, each network address includes a unique cluster identification numbers associated with a cluster of hosts of the computing platform 102. As an example, each unique tenant identification number represents the unique cluster identification number. In certain examples, the unique cluster identification number is associated with one or more tenants of the multi-tenant computing platform 102.


In certain embodiments, each network address of the hosts 1061-m includes a region identification number, a rack identification number (e.g., a network device identification number), and/or a virtual interface of the corresponding host associated with the network address. In some examples, each network address includes an IPv6 address. In certain examples, each network address includes a predetermined network prefix. For example, the predetermined network prefix includes a /16 network prefix or a /28 network prefix. As an example, the predetermined network prefix is associated with a region of the computing platform 102. In other examples, the predetermined network prefix associated with a region of the computing platform 102 represents the region identification number. In some examples, the network addresses include a bit sequence corresponding to subnets associated with the networking devices and/or the hosts of the computing platform 102. For example, the bit sequence corresponds to 16-bit subnets associated with racks of the computing platform 102.


According to some embodiments, the network addresses of the hosts 1061-m include a bit sequence associated with the unique tenant identification number. For example, each network address includes a 20-bit sequence or a 32-bit sequence of the network address for embedding the unique tenant identification number. As an example, each pod running on the hosts 1061-m is configured to use the bit sequence to identify the tenant associated with the unique tenant identification number. In some examples, each network address includes a bit sequence used to identify the virtual interface of the corresponding host for a particular tenant and for a particular networking device (e.g., a particular rack). For example, the bit sequence includes a 64-bit sequence. As an example, the bit sequence is configured to be used for assigning IP addresses of the host's virtual interface, which are generated, for example, by stateless address autoconfiguration (SLAAC).


According to certain embodiments, the fleet controller 114 is configured to send the network address of the at least one host of the identified one or more hosts 1061, . . . 106k to the client (e.g., the client device 1121) associated with tenant 2001.



FIG. 3 is a simplified diagram showing the system 100 for isolating applications associated with multiple tenants within the computing platform 102 according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. In some examples, the fleet controller 114 is configured to generate the broadcast domains 3001 and 3021 including the identified hosts 1061, . . . , 106i, 106i+1, . . . , 106k. For example, the broadcast domain 3001 includes the hosts 1061, . . . , 106i, and the broadcast domain 3021 includes the hosts 106i+1, . . . , 106k. As an example, the fleet controller 114 is configured to store information associated with the broadcast domains 3001 and 3021 in the fleet catalog 116 and/or hardware catalog 120.


In some embodiments, the broadcast domains 3001 and 3021 are connected through the network layer 208 of the network 104. In some examples, the broadcast domains 3001 and 3021 are associated with one or more tenants of the computing platform 102. In some examples, the broadcast domains 3001 and 3021 are associated with the same tenant of the computing platform 102. In other examples, the broadcast domain 3001 is connected to the networking device 1101 at the data link layer of the network 104. In some examples, the broadcast domain 3021 is connected to the networking device 110K at the data link layer of the network 104. In certain examples, the networking device 1101 is configured to send data frames from the broadcast domains 3001 to the broadcast domain 3021 using the network layer 208 of the network 104.


In certain embodiments, the fleet controller 114 is configured to assign to the broadcast domains 3001 and 3021a unique tenant identification number corresponding to one tenant of the multi-tenant computing platform 102. For example, the fleet controller 114 is configured to store information associated with the unique tenant identification number and the assigned broadcast domains 3001 and 3021 in the fleet catalog 116. As an example, the broadcast domains 3001 and 3021 are associated with the same tenant. In some examples, the unique tenant identification number corresponds with the unique domain identification numbers (e.g., “1”) associated with the broadcast domains 3001 and 3021. For example, the unique tenant identification number includes a bit sequence with a first portion of the bit sequence representing the unique domain identification number associated with the broadcast domains 3001 and 3021. In certain examples, the broadcast domains 3001 and 3021 include a virtual local area network (VLAN) extending over the network 104 through the network layer 208.


According to some embodiments, networking devices of the computing platform 102 are configured to provide, at the data link layer 206 of the network 104, layer 2 isolation on a per-tenant basis by assigning each tenant to a separate VLAN. For example, networking devices 1101, . . . , 110K (e.g., ToR ports) of the network 104 that are connecting down to the compute nodes (e.g., the hosts) include 802.1q trunks for carrying multiple VLANs. As an example, each compute node (e.g., each host) includes at least two VLANs with one VLAN for the control plane application and the other VLAN for the container associated with the primary tenant running on the compute node (e.g., the host). In some examples, the networking devices 1101 and 110K are configured to tag received network data for sending across broadcast domains of the computing platform 102.



FIG. 4 is a simplified diagram showing the system 100 for isolating applications associated with multiple tenants within the computing platform 102 according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. In some examples, the host 1061 includes a controller runtime component 400. In certain examples, the controller runtime component 400 is configured to manage running the containers 4021 and 4022. For example, the containers 4021 and 4022 are associated with the pod 404 running on the host 1061. As an example, each of the containers 4021 and 4022 includes a container network interface (CNI) and a container runtime interface (CRI). In some examples, the container 4021 and the client device 1121 are associated with tenant 4061. In certain examples, the container 4022 and the client device 1122 are associated with tenant 4062.


In some embodiments, the container network interfaces of the containers 4021 and 4022 are configured to set up the host-level network layer of the host 1061. In some examples, the container network interfaces of the containers 4021 and 4022 are configured to generate network interfaces for launching the containers 4021 and 4022 on the host 1061 by assigning network addresses to the network interfaces. In certain examples, the container network interfaces of the containers 4021 and 4022 are configured to generate the broadcast domains (e.g., the VLANs) for each container that map to the unique tenant identification number associated with the containers, respectively. For example, the container network interface of the container 4021 is configured to generate the broadcast domain (e.g., the VLAN) for the container 4021 that maps to the unique tenant identification number (e.g., “1”) associated with the container 4021. As an example, the unique tenant identification number (e.g., “1”) associated with the container 4021 corresponds to tenant 4061. In another example, the container network interface of the container 4022 is configured to generate the broadcast domain (e.g., the VLAN) for the container 4022 that maps to the unique tenant identification number (e.g., “2”) associated with the container 4022. As an example, the unique tenant identification number (e.g., “2”) associated with the container 4022 corresponds to tenant 4062. In some examples, each controller network interface of the container 4021 and 4022 is configured to set up a virtual interface of the host 1061 for the pod 404 running on the host 1061.


In certain embodiments, the controller network interfaces of containers are configured to set up a pod-specific network for the corresponding pod at runtime of the pod. For example, the pod running on a host maps to the network address of the host. As an example, the controller network interfaces are configured to set up each pod network with the corresponding VLAN. In some examples, controller network interfaces are configured to use annotations to the network addresses associated with the corresponding pod to map a virtual ethernet interface to the corresponding VLANs. In certain examples, the controller network interface is configured to generate an interface list, IP configurations assigned to the interface, IPv6 addresses, and an internal domain name system.


According to some embodiments, the controller runtime component 400 is configured to receive the requests 4081 and 4082 from the client devices 1121 and 1122, respectively. For example, the controller runtime component 400 is configured to receive the requests 4081 and 4082 via an API server of the computing platform 102. In some examples, the API server delegates authentication and authorization of received client requests to an authentication and authorization component of the computing platform 102 for evaluating the client requests and access requirements and for granting access of the clients associated with the client requests to applications running on hosts of the computing platform 102.


According to certain embodiments, the container runtime component 400 is configured to send the requests 4081 and 4082 to the containers 4021 and 4022 based at least in part on the unique tenant identification number associated with each request, respectively. For example, the container runtime component 400 is configured to send the request 4081 to the containers 4021 based at least in part on the unique tenant identification number (e.g., “1”) associated with the request 4081. As an example, the container runtime component 400 is configured to send the request 4082 to the containers 4022 based at least in part on the unique tenant identification number (e.g., “2”) associated with the request 4082.


In some embodiments, the container runtime component 400 is configured to send the requests 4081 and 4082 to the containers 4021 and 4022 based at least in part on the network address associated with each request, respectively. For example, each network address associated with the requests 4081 and 4082 includes the corresponding unique tenant identification number associated with the client devices 1121 and 1122, respectively. As an example, the network address associated with the request 4081 includes the unique tenant identification number (e.g., “1”) that is associated with the client devices 1121 and relates to tenant 4061. In another example, the network address associated with the request 4082 includes the unique tenant identification number (e.g., “2”) that is associated with the client devices 1122 and relates to tenant 4062. In other examples, the container runtime component 400 is configured to isolate client requests from each other based at least in part on the tenants associated with each client request. For example, the container runtime component 400 is configured to extract the unique tenant identification number from the network address associated with a received client request and forward the client request to the container associated with the extracted unique tenant identification number.



FIG. 5 is a simplified diagram showing a method for isolating applications associated with multiple tenants within a computing platform according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. The method 500 includes processes 502-516 that are performed using one or more processors. Although the above has been shown using a selected group of processes for the method, there can be many alternatives, modifications, and variations. For example, some of the processes may be expanded and/or combined. Other processes may be inserted to those noted above. Depending upon the embodiment, the sequence of processes may be interchanged with others replaced.


In some embodiments, some or all processes (e.g., steps) of the method 500 are performed by the system 100. In certain examples, some or all processes (e.g., steps) of the method 500 are performed by a computer and/or a processor directed by a code. For example, a computer includes a server computer and/or a client computer (e.g., a personal computer). In some examples, some or all processes (e.g., steps) of the method 500 are performed according to instructions included by a non-transitory computer-readable medium (e.g., in a computer program product, such as a computer-readable flash drive). For example, a non-transitory computer-readable medium is readable by a computer including a server computer and/or a client computer (e.g., a personal computer, and/or a server rack). As an example, instructions included by a non-transitory computer-readable medium are executed by a processor including a processor of a server computer and/or a processor of a client computer (e.g., a personal computer, and/or server rack).


In some embodiments, at the process 502, a request is received from a client for running an application on a computing platform. The client is associated with a tenant of the computing platform. The computing platform includes a plurality of hosts connected through a network. Each host is associated with a network address and configured to run applications associated with multiple tenants. At the process 504, one or more hosts of the plurality of hosts are identified based at least in part on the request. At process 506, one or more broadcast domains including the identified one or more hosts are generated. The one or more broadcast domains are isolated in the network at the data link layer. At the process 508, the one or more broadcast domains are assigned a unique tenant identification number corresponding to the tenant. At the process 510, the application is launched on at least one host of the identified one or more hosts. At the process 512, in response to launching the application on the at least one host, the unique tenant identification number is assigned to the launched application. At the process 514, the unique tenant identification number is added to the network address of the at least one host. At the process 516, the network address of the at least one host is sent to the client associated with the tenant.



FIG. 6 is a simplified diagram showing a computing system for implementing a system for isolating applications associated with multiple tenants within a computing platform according to one embodiment of the present invention. This diagram is merely an example, which should not unduly limit the scope of the claims. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. The computing system 600 includes a bus 602 or other communication mechanism for communicating information, a processor 604, a display 606, a cursor control component 608, an input device 610, a main memory 612, a read only memory (ROM) 614, a storage unit 616, and a network interface 618. In some embodiments, some or all processes (e.g., steps) of the method 500 are performed by the computing system 600. In some examples, the bus 602 is coupled to the processor 604, the display 606, the cursor control component 606, the input device 610, the main memory 612, the read only memory (ROM) 614, the storage unit 616, and/or the network interface 618. In certain examples, the network interface is coupled to a network 620. For example, the processor 604 includes one or more general purpose microprocessors. In some examples, the main memory 612 (e.g., random access memory (RAM), cache and/or other dynamic storage devices) is configured to store information and instructions to be executed by the processor 604. In certain examples, the main memory 612 is configured to store temporary variables or other intermediate information during execution of instructions to be executed by processor 604. For examples, the instructions, when stored in the storage unit 616 accessible to processor 604, render the computing system 600 into a special-purpose machine that is customized to perform the operations specified in the instructions. In some examples, the ROM 614 is configured to store static information and instructions for the processor 604. In certain examples, the storage unit 616 (e.g., a magnetic disk, optical disk, or flash drive) is configured to store information and instructions.


In some embodiments, the display 606 (e.g., a cathode ray tube (CRT), an LCD display, or a touch screen) is configured to display information to a user of the computing system 600. In some examples, the input device 610 (e.g., alphanumeric and other keys) is configured to communicate information and commands to the processor 604. For example, the cursor control 608 (e.g., a mouse, a trackball, or cursor direction keys) is configured to communicate additional information and commands (e.g., to control cursor movements on the display 606) to the processor 604.


According to some embodiments, method for isolating applications associated with multiple tenants within a computing platform. For example, a method includes receiving a request from a client associated with a tenant for running an application on a computing platform. The computing platform includes a plurality of hosts connected through a network. Each host is associated with a network address and configured to run applications associated with multiple tenants. The method further includes identifying one or more hosts of the plurality of hosts based at least in part on the request. The method further includes generating one or more broadcast domains including the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The method further includes assigning to the one or more broadcast domains a unique tenant identification number corresponding to the tenant. The method further includes launching the application on at least one host of the identified one or more hosts. In response to launching the application on the at least one host: the unique tenant identification number is assigned to the launched application; the unique tenant identification number is added to the network address of the at least one host; and the network address of the at least one host is sent to the client associated with the tenant. The method is performed using one or more processors. For example, the method is implemented according to at least FIG. 1, FIG. 2, FIG. 3, FIG. 4 and/or FIG. 5.


In some examples, the network address further includes a unique cluster identification number. The unique cluster identification number is associated with a cluster of the computing platform. The cluster is associated with the tenant. In certain examples, the unique tenant identification number includes the unique cluster identification number. In other examples, the network address includes a plurality of bit sequences. One bit sequence of the plurality of bit sequences includes the unique tenant identification number.


In certain examples, the plurality of bit sequences of the network address includes at least 128 bits. In some examples, the broadcast domains include virtual local area networks. In certain examples, the launched application is included in a container. The container is associated with the unique tenant identification number. For example, the container is included in a pod running on the at least one host. The pod includes one or more containers. Each container of the one or more containers is associated with one tenant of the computing platform. As an example, each container of the one or more containers is associated with a different tenant of the computing platform. In one example, the pod maps to the network address. In some examples, the network address further includes a region identification number, a network device identification number, or a virtual interface of the at least one host. In certain examples, the one or more broadcast domain map to the network address.


According to certain embodiments, a system for isolating applications associated with multiple tenants within a computing platform includes a plurality of hosts connected through a network and a fleet controller. Each host is associated with a network address and configured to run applications associated with multiple tenants on a computing platform. The fleet controller is configured to, in response to receiving a first request from a client associated with a tenant for running an application on the computing platform, identify one or more hosts of the plurality of hosts based at least in part on the request. The fleet controller is further configured to generate one or more broadcast domains including the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The fleet controller is further configured to assign to the one or more broadcast domains a unique tenant identification number corresponding to the tenant. The fleet controller is further configured to send a second request to a scheduler for launching the application on at least one host of the identified one or more hosts. The fleet controller is further configured to, in response to receiving confirmation from the scheduler of the application being launched on the at least one host: assign the unique tenant identification number to the launched application; add the unique tenant identification number to the network address of the at least one host; and send the network address of the at least one host to the client associated with the tenant. For example, the system is implemented according to at least FIG. 1, FIG. 2, FIG. 3, and/or FIG. 4.


In some examples, the network address further includes a unique cluster identification number. The unique cluster identification number is associated with a cluster of the computing platform. The cluster is associated with the tenant. In certain examples, the network address includes a plurality of bit sequences. One bit sequence of the plurality of bit sequences includes the unique tenant identification number. In other examples, the plurality of bit sequences of the network address includes at least 128 bits. In some examples, the broadcast domains include virtual local area networks.


According to some embodiments, a system for isolating applications associated with multiple tenants within a computing platform includes a client associated with a tenant and configured to request running an application on a computing platform. The computing platform includes a plurality of hosts connected through a network. Each host is associated with a network address and is configured to run applications associated with multiple tenants. The client is further configured to send a request for running the application on the computing platform. The client is further configured to, in response to sending the request for running the application on the computing platform, receive the network address of at least one host of the plurality of hosts. One or more hosts of the plurality of hosts are identified based at least in part on the request. The identified one or more hosts include the at least one host. One or more broadcast domains are generated to include the identified one or more hosts. The one or more broadcast domains are isolated in the network at a data link layer. The one or more broadcast domains are assigned to a unique tenant identification number corresponding to the tenant. The application is launched on the at least one host of the identified one or more hosts. The launched application is assigned to the unique tenant identification number. The unique tenant identification number is added to the network address of the at least one host. For example, the system is implemented according to at least FIG. 1, FIG. 2, FIG. 3, and/or FIG. 4.


In some examples, the network address further includes a unique cluster identification number. The unique cluster identification number is associated with a cluster of the computing platform. The cluster is associated with the tenant. In certain examples, the network address includes a plurality of bit sequences. One bit sequence of the plurality of bit sequences includes the unique tenant identification number.


For example, some or all components of various embodiments of the present invention each are, individually and/or in combination with at least another component, implemented using one or more software components, one or more hardware components, and/or one or more combinations of software and hardware components. In another example, some or all components of various embodiments of the present invention each are, individually and/or in combination with at least another component, implemented in one or more circuits, such as one or more analog circuits and/or one or more digital circuits. In yet another example, while the embodiments described above refer to particular features, the scope of the present invention also includes embodiments having different combinations of features and embodiments that do not include all of the described features. In yet another example, various embodiments and/or examples of the present invention can be combined.


Additionally, the methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing subsystem. The software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to perform the methods and systems described herein.


The systems' and methods' data (e.g., associations, mappings, data input, data output, intermediate data results, final data results, etc.) may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, EEPROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, application programming interface, etc.). It is noted that data structures describe formats for use in organizing and storing data in databases, programs, memory, or other computer-readable media for use by a computer program.


The systems and methods may be provided on many different types of computer-readable media including computer storage mechanisms (e.g., CD-ROM, diskette, RAM, flash memory, computer's hard drive, DVD, etc.) that contain instructions (e.g., software) for use in execution by a processor to perform the methods' operations and implement the systems described herein. The computer components, software modules, functions, data stores and data structures described herein may be connected directly or indirectly to each other in order to allow the flow of data needed for their operations. It is also noted that a module or processor includes a unit of code that performs a software operation and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code. The software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.


The computing system can include client devices and servers. A client device and server are generally remote from each other and typically interact through a communication network. The relationship of client device and server arises by virtue of computer programs running on the respective computers and having a client device-server relationship to each other.


This specification contains many specifics for particular embodiments. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations, one or more features from a combination can in some cases be removed from the combination, and a combination may, for example, be directed to a subcombination or variation of a subcombination.


Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the embodiments described above should not be understood as requiring such separation in all embodiments, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.


Although specific embodiments of the present invention have been described, it will be understood by those of skill in the art that there are other embodiments that are equivalent to the described embodiments. Accordingly, it is to be understood that the invention is not to be limited by the specific illustrated embodiments, but only by the scope of the appended claims.

Claims
  • 1. A method for isolating applications associated with multiple tenants within a computing platform, the method comprising: receiving a request from a client associated with a tenant for running an application on a computing platform, the computing platform including a plurality of hosts connected through a network, each host of the plurality of hosts being associated with a network address;generating one or more broadcast domains based at least in part on the request, the one or more broadcast domains being isolated in the network at a data link layer;assigning a unique tenant identification number to the application; andadding the unique tenant identification number to the network address of at least one host of the plurality of hosts;wherein the method is performed using one or more processors.
  • 2. The method of claim 1, wherein the request is received by a fleet controller, and wherein the fleet controller sends the received request to a fleet catalog for storing.
  • 3. The method of claim 2, further comprising: receiving a state of a resource in the fleet catalog, wherein the state of the resource includes an assignment of a host of the plurality of hosts and a status of the host; andupdating the state of the resource in the fleet catalog.
  • 4. The method of claim 1 wherein: the network address includes a plurality of bit sequences; anda bit sequence of the plurality of bit sequences includes the unique tenant identification number.
  • 5. The method of claim 4, wherein the bit sequence includes a 20-bit sequence or a 32-bit sequence for embedding the unique tenant identification number in the network address.
  • 6. The method of claim 5 wherein the plurality of bit sequences of the network address includes at least 128 bits.
  • 7. The method of claim 1 wherein the one or more broadcast domains include one or more virtual local area networks.
  • 8. The method of claim 1 wherein: the application is included in a container; andthe container is associated with the unique tenant identification number.
  • 9. The method of claim 8 wherein: the container is included in a pod running on the at least one host;the pod includes one or more containers; andeach container of the one or more containers is associated with a respective tenant of the computing platform.
  • 10. The method of claim 1 wherein the network address further includes at least one in a group consisting of a region identification number, a network device identification number, and a virtual interface of the at least one host.
  • 11. The method of claim 1, further comprising mapping the one or more broadcast domains to the network address.
  • 12. The method of claim 11, wherein the one or more broadcast domain are mapped to the network address via information of a certificate.
  • 13. A system for isolating applications associated with multiple tenants within a computing platform, the system comprising: a plurality of hosts connected through a network, each host of the plurality of hosts being associated with a network address; anda fleet controller;wherein the fleet controller is configured to: receive a request from a client associated with a tenant for running an application on the computing platform;generate one or more broadcast domains based at least in part on the request, the one or more broadcast domains being isolated in the network at a data link layer;assign a unique tenant identification number to the application; andadd the unique tenant identification number to the network address of at least one host of the plurality of hosts.
  • 14. The system of claim 13, further comprising a fleet catalog, wherein the fleet controller is further configured to store information associated with the one or more broadcast domains in the fleet catalog.
  • 15. The system of claim 14, wherein the fleet controller is further configured to record a state change of the at least one host in the fleet catalog.
  • 16. The system of claim 13, further comprising a fleet scheduler, wherein the fleet scheduler is configured to identify one or more hosts of the hosts based at least in part on the request, and wherein the at least one host is at least one host of the one or more identified hosts.
  • 17. The system of claim 13 wherein the one or more broadcast domains include one or more virtual local area networks.
  • 18. A system for isolating applications associated with multiple tenants within a computing platform, the system comprising: a client associated with a tenant and configured to request running an application on a computing platform, the computing platform including a plurality of hosts connected through a network, each host of the plurality of hosts being associated with a network address;wherein the client is further configured to:send a request for running the application on the computing platform, wherein: one or more broadcast domains are generated based at least in part on the request, the one or more broadcast domains being isolated in the network at a data link layer;the application is assigned to a unique tenant identification number; andthe unique tenant identification number is added to a network address of at least one host of the plurality of hosts.
  • 19. The system of claim 18 wherein: the network address further includes a unique cluster identification number;the unique cluster identification number is associated with a cluster of the computing platform; andthe cluster is associated with the tenant.
  • 20. The system of claim 19 wherein: the network address includes a plurality of bit sequences; andone bit sequence of the plurality of bit sequences includes the unique tenant identification number.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 17/714,722, filed Apr. 6, 2022, which is a continuation of U.S. application Ser. No. 17/097,389 (now U.S. Pat. No. 11,330,076), filed Nov. 13, 2020, which is a continuation of U.S. application Ser. No. 16/545,073 (now U.S. Pat. No. 10,868,887), filed Aug. 20, 2019, which claims priority to U.S. Provisional Application No. 62/803,257, filed Feb. 8, 2019, all incorporated by reference herein for all purposes.

US Referenced Citations (214)
Number Name Date Kind
5978475 Schneier et al. Nov 1999 A
6253203 O'Flaherty et al. Jun 2001 B1
6725240 Asad et al. Apr 2004 B1
6807569 Bhimani et al. Oct 2004 B1
7017046 Doyle et al. Mar 2006 B2
7069586 Winneg et al. Jun 2006 B1
7225468 Waisman et al. May 2007 B2
7596285 Brown et al. Sep 2009 B2
7770032 Nesta et al. Aug 2010 B2
7801871 Gosnell Sep 2010 B2
7962495 Jain et al. Jun 2011 B2
8141036 Wagner et al. Mar 2012 B2
8181253 Zaitsev et al. May 2012 B1
8190893 Benson et al. May 2012 B2
8196184 Amirov et al. Jun 2012 B2
8239668 Chen et al. Aug 2012 B1
8301904 Gryaznov Oct 2012 B1
8312546 Alme Nov 2012 B2
8392902 Reinz Mar 2013 B2
8646080 Williamson et al. Feb 2014 B2
8667456 Czymontek Mar 2014 B1
8683322 Cooper Mar 2014 B1
8726379 Stiansen et al. May 2014 B1
8769412 Gill et al. Jul 2014 B2
8782794 Ramcharran Jul 2014 B2
8931043 Cooper et al. Jan 2015 B2
9021260 Falk et al. Apr 2015 B1
9049117 Nucci et al. Jun 2015 B1
9100428 Visbal Aug 2015 B1
9122562 Stickle Sep 2015 B1
9288117 Angrish et al. Mar 2016 B1
9304798 Mudigonda et al. Apr 2016 B2
9335897 Goldenberg May 2016 B2
9338013 Castellucci et al. May 2016 B2
9350657 Gross, IV et al. May 2016 B2
9444785 Jungck et al. Sep 2016 B2
9497039 Greenberg et al. Nov 2016 B2
9509662 Mudigonda et al. Nov 2016 B2
9514327 Ford Dec 2016 B2
9721117 Pleau et al. Aug 2017 B2
9754122 Christopher et al. Sep 2017 B1
9813303 Guilli Nov 2017 B1
9825822 Holland Nov 2017 B1
9825905 Thakkar et al. Nov 2017 B2
9900313 Wiest et al. Feb 2018 B2
10009443 Guigli Jun 2018 B1
10068089 Shavell et al. Sep 2018 B1
10079832 DeArment et al. Sep 2018 B1
10333889 Thakkar et al. Jun 2019 B2
10348767 Lee et al. Jul 2019 B1
10491516 Ram et al. Nov 2019 B2
10691493 Barry et al. Jun 2020 B1
10805216 Sinha et al. Oct 2020 B2
10868887 DeArment et al. Dec 2020 B2
11159366 Gawade et al. Oct 2021 B1
11330076 DeArment et al. May 2022 B2
20020112157 Doyle et al. Aug 2002 A1
20040073814 Miyazaki et al. Apr 2004 A1
20040123139 Aiello et al. Jun 2004 A1
20040153418 Hanweck Aug 2004 A1
20040250124 Chesla et al. Dec 2004 A1
20050157662 Bingham et al. Jul 2005 A1
20050229256 Banzhof Oct 2005 A2
20050262556 Waisman et al. Nov 2005 A1
20050275638 Kalmykov-Zotov et al. Dec 2005 A1
20060031928 Conley et al. Feb 2006 A1
20060069912 Zheng et al. Mar 2006 A1
20060179003 Steele et al. Aug 2006 A1
20060121131 Shull et al. Sep 2006 A1
20060218637 Thomas et al. Sep 2006 A1
20060265747 Judge Nov 2006 A1
20070074204 Curtis et al. Mar 2007 A1
20070143851 Nicodemus Jun 2007 A1
20070294766 Mir et al. Dec 2007 A1
20080104407 Horne et al. May 2008 A1
20080201580 Savitzky et al. Aug 2008 A1
20080222706 Renaud et al. Sep 2008 A1
20080229422 Hudis et al. Sep 2008 A1
20090103442 Douville Apr 2009 A1
20090228701 Lin Sep 2009 A1
20090328222 Helman et al. Dec 2009 A1
20100077481 Polyakov et al. Mar 2010 A1
20100100963 Maheffey Apr 2010 A1
20100103837 Jungck et al. Apr 2010 A1
20100179831 Brown et al. Jul 2010 A1
20100235915 Memon et al. Sep 2010 A1
20100262688 Hussain et al. Oct 2010 A1
20100309510 Hansen Dec 2010 A1
20100325159 Wright et al. Dec 2010 A1
20100330801 Rouh Dec 2010 A1
20110060910 Gormish et al. Mar 2011 A1
20110126127 Mariotti et al. May 2011 A1
20110202555 Cordover et al. Aug 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110277027 Hayton et al. Nov 2011 A1
20120079592 Pandrangi Mar 2012 A1
20120084866 Stolfo Apr 2012 A1
20120110055 Van Biljon et al. May 2012 A1
20120110633 An et al. May 2012 A1
20120110674 Belani et al. May 2012 A1
20120120964 Koponen et al. May 2012 A1
20120144023 Guest et al. Jun 2012 A1
20120169593 Mak et al. Jul 2012 A1
20120180039 Bravery et al. Jul 2012 A1
20120218305 Patterson et al. Aug 2012 A1
20120254129 Wheeler et al. Oct 2012 A1
20120281706 Agarwal et al. Nov 2012 A1
20120284791 Miller et al. Nov 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20130019306 Lagar-Cavilla et al. Jan 2013 A1
20130086322 Pelletier et al. Apr 2013 A1
20130097709 Basavapatna et al. Apr 2013 A1
20130110876 Meijer et al. May 2013 A1
20130136123 Ge et al. May 2013 A1
20130139268 An et al. May 2013 A1
20130211911 Krietzman Aug 2013 A1
20130239217 Kindler et al. Sep 2013 A1
20130247142 Nishizawa et al. Sep 2013 A1
20140013451 Kulka et al. Jan 2014 A1
20140043972 Li et al. Feb 2014 A1
20140059683 Ashley Feb 2014 A1
20140098815 Mishra et al. Apr 2014 A1
20140123279 Bishop et al. May 2014 A1
20140129550 Weatherhead et al. May 2014 A1
20140143009 Brice et al. May 2014 A1
20140165038 Kaler et al. Jun 2014 A1
20140173712 Ferdinand Jun 2014 A1
20140173738 Condry et al. Jun 2014 A1
20140188895 Wang et al. Jul 2014 A1
20140204937 Matsuoka Jul 2014 A1
20140229422 Jain et al. Aug 2014 A1
20140280595 Mani et al. Sep 2014 A1
20140283107 Walton et al. Sep 2014 A1
20140366132 Stiansen et al. Dec 2014 A1
20140373007 Donnellan et al. Dec 2014 A1
20150006730 Helfman et al. Jan 2015 A1
20150030024 Venkataswami et al. Jan 2015 A1
20150039565 Lucas Feb 2015 A1
20150063353 Kapadia et al. Mar 2015 A1
20150082301 Garg et al. Mar 2015 A1
20150106489 Duggirala Apr 2015 A1
20150128274 Giokas May 2015 A1
20150134600 Eisner et al. May 2015 A1
20150163158 Ryland Jun 2015 A1
20150188715 Castelluci et al. Jul 2015 A1
20150195178 Bhattacharya et al. Jul 2015 A1
20150200948 Cairns et al. Jul 2015 A1
20150229664 Hawthorn et al. Aug 2015 A1
20150248563 Alfarano et al. Sep 2015 A1
20150261847 Ducott et al. Sep 2015 A1
20150263970 Macchiano et al. Sep 2015 A1
20150278311 Isherwood et al. Oct 2015 A1
20150286665 Nyman et al. Oct 2015 A1
20150312188 White et al. Oct 2015 A1
20150326467 Fullbright et al. Nov 2015 A1
20150326601 Grandin et al. Nov 2015 A1
20150370549 Zhang et al. Dec 2015 A1
20160004864 Falk et al. Jan 2016 A1
20160013973 Onoue Jan 2016 A1
20160020921 Tsai et al. Jan 2016 A1
20160028759 Visbal Jan 2016 A1
20160080502 Yadav et al. Mar 2016 A1
20160087941 Mudigonda et al. Mar 2016 A1
20160105392 Thakkar et al. Apr 2016 A1
20160105393 Thakkar et al. Apr 2016 A1
20160182309 Maturana et al. Jun 2016 A1
20160182473 Cignetti et al. Jun 2016 A1
20160188527 Cherian et al. Jun 2016 A1
20160239343 Holt Aug 2016 A1
20160253198 Gallant et al. Sep 2016 A1
20160292236 Joshi et al. Oct 2016 A1
20160344687 Rong et al. Nov 2016 A1
20160352632 Nedeltchev et al. Dec 2016 A1
20160371108 Madtha et al. Dec 2016 A1
20170005923 Babakian Jan 2017 A1
20170026245 Rao et al. Jan 2017 A1
20170063542 Vilella et al. Mar 2017 A1
20170091466 Meyer et al. Mar 2017 A1
20170099160 Mithyantha et al. Apr 2017 A1
20170185715 Smith Jun 2017 A9
20170192986 Isherwood et al. Jul 2017 A1
20170317972 Bansal et al. Nov 2017 A1
20170323117 Pleau et al. Nov 2017 A1
20170324660 Sinha et al. Nov 2017 A1
20170331742 He et al. Nov 2017 A1
20170331791 Wardell et al. Nov 2017 A1
20170346686 Mudigonda Nov 2017 A1
20170346736 Chander et al. Nov 2017 A1
20170353444 Karangutkar et al. Dec 2017 A1
20170373860 Kshirsagar et al. Dec 2017 A1
20180024975 Ramakrishnan et al. Jan 2018 A1
20180025152 Ben Ali et al. Jan 2018 A1
20180069809 Oswal et al. Mar 2018 A1
20180095774 Atta et al. Apr 2018 A1
20180095898 Khosravi et al. Apr 2018 A1
20180115470 Huang et al. Apr 2018 A1
20180123903 Holla et al. May 2018 A1
20180139150 Jain et al. May 2018 A1
20180176176 Kapur et al. Jun 2018 A1
20180219743 Garcia et al. Aug 2018 A1
20180295036 Krishnamurthy et al. Oct 2018 A1
20180332001 Ferrero et al. Nov 2018 A1
20180351791 Nagarajan et al. Dec 2018 A1
20180367393 Harneja et al. Dec 2018 A1
20190052532 Chen et al. Feb 2019 A1
20190068495 Jeuk et al. Feb 2019 A1
20190097918 Ponnuswamy et al. Mar 2019 A1
20190132229 McCormack et al. May 2019 A1
20190141010 Chander et al. May 2019 A1
20190334868 Tewari et al. Oct 2019 A1
20200007583 Dixit et al. Jan 2020 A1
20200067763 Vytla Feb 2020 A1
20200136966 Mishra Apr 2020 A1
20220232098 DeArment et al. Jul 2022 A1
Foreign Referenced Citations (12)
Number Date Country
101729531 Jun 2010 CN
103281301 Sep 2013 CN
1962222 Aug 2008 EP
2892197 Jul 2015 EP
2897051 Jul 2015 EP
2963578 Jan 2016 EP
2985974 Feb 2016 EP
3188069 Jul 2017 EP
3694173 Sep 2022 EP
2011642 Aug 2015 NL
WO 2005010685 Feb 2005 WO
WO 2017177401 Oct 2017 WO
Non-Patent Literature Citations (16)
Entry
European Patent Office; Official Communication for Application No. 19192652.6; dated Feb. 11, 2020.
European Patent Office; Official Communication for Application No. 17197000.7; dated Mar. 8, 2018.
VirusTotal—About, <http://www.virustotal.com/en/about/> Printed Jun. 30, 2014 in 8 pages.
Schneier et al., “Automatic Event Stream Notarization Using Digital Signatures,” Security Protocols, International Workshop Apr. 1996 Proceedings, Springer-Veriag, 1997.
Crosby et al., “Efficient Data Structures for Tamper-Evident Logging,” Department of Computer Science, Rice University, 2009, pp. 17.
Ma et al., “A New Approach to Secure Logging,” ACM Transactions on Storage, vol. 5, No. 1, Article 2, Published Mar. 2009, 21 pages.
Glaab et al., “EnrichNet: Network-Based Gene Set Enrichment Analysis,” Bioinformatics 28.18 (2012): pp. i451-i457.
FireEye, <http://www.fireeye.com/> Printed Jun. 30, 2014 in 2 pages.
Lee et al., “A Data Mining and CIDF Based Approach for Detecting Novel and Distributed Intrusions,” Lecture Notes in Computer Science, vol. 1907 Nov. 11, 2000, pp. 49-65.
Schneier et al., “Cryptographic Support for Secure Logs on Untrusted Machines,” The Seventh USENIX Security Symposium Proceedings, USENIX Press, Jan. 1998, pp. 53-62, https://www.schneier.com/paper-secure-logs.pdf.
FireEye—Products and Solutions Overview, <http://www.fireeye.com/products-and-solutions> Printed Jun. 30, 2014 in 3 pages.
Bhuyan et al., “Network Anomaly Detection: Methods, Systems and Tools,” First Quarter 2014, IEEE.
Baker et al., “The Development of a Common Enumeration of Vulnerabilities and Exposures,” Presented at the Second International Workshop on Recent Advances in Intrusion Detection, Sep. 7-9, 1999, pp. 35.
Hur et al., “SciMiner: web-based literature mining tool for target identification and functional enrichment analysis,” Bioinformatics 25.6 (2009): pp. 838-840.
Waters et al., “Building an Encrypted and Searchable Audit Log,” Published Jan. 9, 2004, 11 pages, http://www.parc.com/content/attachments/building_encrypted_searchable_5059_parc.pdf.
Zheng et al., “GOEAST: a web-based software toolkit for Gene Ontology enrichment analysis,” Nucleic acids research 36.suppl 2 (2008): pp. W385-W363.
Related Publications (1)
Number Date Country
20230269307 A1 Aug 2023 US
Provisional Applications (1)
Number Date Country
62803257 Feb 2019 US
Continuations (3)
Number Date Country
Parent 17714722 Apr 2022 US
Child 18140918 US
Parent 17097389 Nov 2020 US
Child 17714722 US
Parent 16545073 Aug 2019 US
Child 17097389 US