The present disclosure relates generally to maintaining a security environment for a network environment and service provider cloud infrastructure client instance.
This section is intended to introduce the reader to various aspects of art that may be related to various aspects of the present disclosure, which are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present disclosure. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
Organizations, regardless of size, rely upon access to information technology (IT) and data and services for their continued operation and success. A respective organization's IT infrastructure may have associated hardware resources (e.g., computing devices, load balancers, firewalls, switches, etc.) and software resources (e.g., productivity software, database applications, custom applications, and so forth). Over time, more and more organizations have turned to cloud computing approaches to supplement or enhance their IT infrastructure solutions.
Cloud computing relates to the sharing of computing resources that are generally accessed via the Internet. In particular, a cloud computing infrastructure allows users, such as individuals and/or enterprises, to access a shared pool of computing resources, such as servers, storage devices, networks, applications, and/or other computing based services. By doing so, users are able to access computing resources on demand that are located at remote locations, which resources may be used to perform a variety of computing functions (e.g., storing and/or processing large quantities of computing data). For enterprise and other organization users, cloud computing provides flexibility in accessing cloud computing resources without accruing large up-front costs, such as purchasing expensive network equipment or investing large amounts of time in establishing a private network infrastructure. Instead, by utilizing cloud computing resources, users are able redirect their resources to focus on their enterprise's core functions.
A summary of certain embodiments disclosed herein is set forth below. It should be understood that these aspects are presented merely to provide the reader with a brief summary of these certain embodiments and that these aspects are not intended to limit the scope of this disclosure. Indeed, this disclosure may encompass a variety of aspects that may not be set forth below.
A cloud computing system is disclosed that includes a hub client instance and at least one spoke client instance that is generated based on copying the hub client instance. The hub client instance may include multiple objects (“hub objects”), and the spoke client instance may include multiple objects (“spoke objects”) that are copied from the hub objects. The hub objects may be maintained using hub object tables in the hub client instance, while the spoke objects may be maintained using spoke object tables in the spoke client instance. To synchronize the spoke client instance with the hub client instance, the spoke client instance may request data indicating changes made to the hub object tables. The request may be one-way, such that the spoke client may block or prevent requests from the hub client instance. In response to receiving the request, the hub client instance may send the data indicating changes made to the hub object tables to the spoke client instance. The spoke client instance may update the spoke object tables based on the data.
In this manner, the cloud computing system may maintain synchronization between the hub client instance and a spoke client instance, such that the cloud computing system may only allow the spoke client instance to synchronize by sending a request to the hub client instance (e.g., one way requests).
Various refinements of the features noted above may exist in relation to various aspects of the present disclosure. Further features may also be incorporated in these various aspects as well. These refinements and additional features may exist individually or in any combination. For instance, various features discussed below in relation to one or more of the illustrated embodiments may be incorporated into any of the above-described aspects of the present disclosure alone or in any combination. The brief summary presented above is intended only to familiarize the reader with certain aspects and contexts of embodiments of the present disclosure without limitation to the claimed subject matter.
Various aspects of this disclosure may be better understood upon reading the following detailed description and upon reference to the drawings in which:
One or more specific embodiments will be described below. In an effort to provide a concise description of these embodiments, not all features of an actual implementation are described in the specification. It should be appreciated that in the development of any such actual implementation, as in any engineering or design project, numerous implementation-specific decisions must be made to achieve the developers' specific goals, such as compliance with system-related and enterprise-related constraints, which may vary from one implementation to another. Moreover, it should be appreciated that such a development effort might be complex and time consuming, but would nevertheless be a routine undertaking of design, fabrication, and manufacture for those of ordinary skill having the benefit of this disclosure.
As used herein, the term “computing system” refers to an electronic computing device such as, but not limited to, a single computer, virtual machine, virtual container, host, server, laptop, and/or mobile device, or to a plurality of electronic computing devices working together to perform the function described as being performed on or by the computing system. As used herein, the term “medium” refers to one or more non-transitory, computer-readable physical media that together store the contents described as being stored thereon. Embodiments may include non-volatile secondary storage, read-only memory (ROM), and/or random-access memory (RAM). As used herein, the term “application” refers to one or more computing modules, programs, processes, workloads, threads and/or a set of computing instructions executed by a computing system. Example embodiments of an application include software modules, software objects, software instances and/or other types of executable code.
In certain situations, a client instance implemented on a cloud-based infrastructure that may be managed by a service provider and operating as a “hub” (a hub client instance) may be copied or cloned, and each copy or clone (e.g., a spoke client instance) may operate separately to enable access to different departments, offices, or employees of a client. However, some of the spoke client instances may operate in more restrictive environments (e.g., different jurisdictional or regulatory entities (such as different countries, states, counties, cities, and to forth)) than others. The more restrictive environments may have different or higher security requirements that enforce certain rules not enforced in other environments in which the client operates. For example, in a less restrictive environment, a spoke client instance may synchronize with the hub client instance by receiving data from the hub client instance in response to sending requests to and receiving requests from the hub client instance. However, in the more restrictive environment, the spoke client instance may not be permitted to receive any requests (e.g., from the hub client instance), and may only send requests to the hub client instance for data associated with synchronization purposes.
With the preceding in mind, a cloud computing system is disclosed that includes a hub client instance and at least one spoke client instance that is generated based on copying the hub client instance. The hub client instance may include multiple objects (“hub objects”), and the spoke client instance may include multiple objects (“spoke objects”) that are copied from the hub objects. The hub objects may be maintained using hub object tables in the hub client instance, while the spoke objects may be maintained using spoke object tables in the spoke client instance. To synchronize the spoke client instance with the hub client instance, the spoke client instance may request data indicating changes made to the hub object tables. The request may be one-way, such that the spoke client may block or prevent requests from the hub client instance. In response to receiving the request, the hub client instance may send the data indicating changes made to the hub object tables to the spoke client instance. The spoke client instance may update the spoke object tables based on the data.
If the spoke object tables were updated, the spoke client instance may send a confirmation to the hub client instance. In some cases, as the spoke client instance updates the spoke object tables, the spoke client instance may generate an error with respect to updating a certain spoke object table. As such, the spoke client instance may stop updating any other spoke object tables, and send an indication to the hub client instance of the spoke object tables that were successfully updated, the spoke object table that generated the error, and the spoke object tables that were deferred or not updated due to the update process being stopped. The spoke client instance may subsequently request data for updating the spoke object table that generated the error and the spoke object tables that were deferred. In this manner, the cloud computing system may maintain synchronization between the hub client instance and a spoke client instances that operates in a more restrictive environment.
With the preceding in mind, the following figures relate to various types of generalized system architectures or configurations that may be employed to provide services to an organization in a multi-instance framework and on which the present approaches may be employed. Correspondingly, these system and platform examples may also relate to systems and platforms on which the techniques discussed herein may be implemented or otherwise utilized. Turning now to
For the illustrated embodiment,
In
To utilize computing resources within the platform 16, network operators may choose to configure the data centers 18 using a variety of computing infrastructures. In one embodiment, one or more of the data centers 18 are configured using a multi-tenant cloud architecture, such that one of the server instances 26 handles requests from and serves multiple customers. Data centers 18 with multi-tenant cloud architecture commingle and store data from multiple customers, where multiple customer instances are assigned to one of the virtual servers 26. In a multi-tenant cloud architecture, the particular virtual server 26 distinguishes between and segregates data and other information of the various customers. For example, a multi-tenant cloud architecture could assign a particular identifier for each customer in order to identify and segregate the data from each customer. Generally, implementing a multi-tenant cloud architecture may suffer from various drawbacks, such as a failure of a particular one of the server instances 26 causing outages for all customers allocated to the particular server instance.
In another embodiment, one or more of the data centers 18 are configured using a multi-instance cloud architecture to provide every customer its own unique customer instance or instances. For example, a multi-instance cloud architecture could provide each customer instance with its own dedicated application server(s) and dedicated database server(s). In other examples, the multi-instance cloud architecture could deploy a single physical or virtual server 26 and/or other combinations of physical and/or virtual servers 26, such as one or more dedicated web servers, one or more dedicated application servers, and one or more database servers, for each customer instance. In a multi-instance cloud architecture, multiple customer instances could be installed on one or more respective hardware servers, where each customer instance is allocated certain portions of the physical server resources, such as computing memory, storage, and processing power. By doing so, each customer instance has its own unique software stack that provides the benefit of data isolation, relatively less downtime for customers to access the platform 16, and customer-driven upgrade schedules. An example of implementing a customer instance within a multi-instance cloud architecture will be discussed in more detail below with reference to
Although
As may be appreciated, the respective architectures and frameworks discussed with respect to
By way of background, it may be appreciated that the present approach may be implemented using one or more processor-based systems such as shown in
With this in mind, an example computer system may include some or all of the computer components depicted in
The one or more processors 202 may include one or more microprocessors capable of performing instructions stored in the memory 206. In some embodiments, the instructions may be pipelined from execution stacks of each process in the memory 206 and stored in an instruction cache of the one or more processors 202 to be processed more quickly and efficiently. Additionally or alternatively, the one or more processors 202 may include application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), and/or other devices designed to perform some or all of the functions discussed herein without calling instructions from the memory 206.
With respect to other components, the one or more busses 204 include suitable electrical channels to provide data and/or power between the various components of the computing system 200. The memory 206 may include any tangible, non-transitory, and computer-readable storage media. Although shown as a single block in
With the preceding technical and background discussion in mind, in certain situations a client instance 102 may be copied or cloned, and each copy or clone may operate separately to enable access to different departments, offices, or employees of a client, such as at different geographic sites or locations. The original client instance 102 may be referred to as a “hub instance”, while the copied client instance may be referred to as a “spoke client instance”. However, some of the spoke client instances may operate in more restrictive environments than others.
In some cases, the more regulated environment may be associated with regulations that may be enforced by a governmental agency, a different country, a different region, a different department, and so on. As such, while a first set of client devices 20 may access the hub client instance 302 in a less restrictive environment (e.g., at a first country or department) via a first communication network (e.g., a first client network 12), a second set of client devices 20 may access the spoke client instance 303 in a more restrictive environment (e.g., at a second country or department) via a second communication network (e.g., a second client network 12).
Data in the spoke client instances 303 may be updated or synchronized (e.g., periodically or based on certain triggering events) to continue mirroring the hub client instance 302. For example, a spoke client instance 303 may be updated every minute, every thirty minutes, every hour, every two hours, every six hours, every twelve hours, every day, every week, every month, and so on. As another example, a spoke client instance 303 may be update each time a threshold number of changes have been made to the hub client instance 302, each time certain data in the hub client instance 302 (e.g., in a certain object table) has been updated, and so on.
In a normal environment (e.g., an environment that is less restricted), a spoke client instance 303 may synchronize with the hub client instance 302 by receiving data from the hub client instance 302 in response to sending requests to and receiving requests from the hub client instance 302. However, in a more restrictive environment, the spoke client instance 303 may not be permitted to receive any requests (e.g., from the hub client instance 302). That is, the spoke client instance 303 may block or prevent requests from being received, or ignore any received requests for data. Instead, the spoke client instance 303 may send requests to the hub client instance 302 for data associated with synchronization purposes.
The more restrictive environment may operate under additional or alternative rules, such that the data requested by the spoke client instance 303 may be requested and processed by the spoke client instance 303 on a first in first out basis. Additionally or alternatively, there may be no parallel processing of records from the same table of the hub client instance 302. Controls may be implemented to limit the number of records being retrieved by a single client process.
A single hub client instance 302 may accommodate multiple spoke client instance 303. In some embodiments, credentials and configuration storage and management may comply with service provider standards (e.g., storing passwords in an encrypted password field). As another example, actions may be restricted to those with minimal impact. That is, producing a single record for consumption by multiple spoke client instances 303 may result in a convoluted cleanup approach once all spoke client instances 303 have consumed the record, compared to producing multiple copies of the same record, because it may be more efficient and less complex to purge individual records once they are consumed. In addition to synchronization of data, due care may be taken to maintain a simple to understand record numbering approach (e.g., avoid using a duplicate record numbers). Further, all tables which are synchronized may be included in a number management script to prevent duplicate numbering, and changes to the number management script may be propagated from the hub client instance 302 and across all spoke client instance 303 as soon as possible to avoid collisions.
In particular, the hub client instance 302 may include one or more objects 304 (“hub objects”). An object 304 may include a data structure, such as a Javascript construct, programmed or written to hold and/or manipulate data. Each hub object 304 may be defined or have attributes configured by an associated table 305 (“hub object table”). Similarly, each spoke client instance 303 may include one or more objects 306 (“spoke objects”) that may be defined or have attributes configured by an associated table 307 (“spoke object table”). At least some of the spoke object tables 307 may mirror the hub object tables 305 by storing the same data. As such, to synchronize a spoke client instance 303 with the hub client instance 302, changes to at least some of the hub object tables 305 may be made to the spoke object tables 307.
The spoke client instance 303 may then update spoke object tables 307 that correspond to (e.g., that should mirror) the hub object tables 305. In particular, the spoke client instance 303 may update the corresponding spoke object tables 307 with the staging records sent in the communication 312, such that the corresponding spoke object tables 307 store the same records as the hub object tables 305. In this manner, the spoke client instance 303 may be synchronized with the hub client instance 302.
As the spoke client instance 303 updates each record of a spoke object table 307, the spoke client instance 303 may indicate or otherwise flag the staging record as successfully updated. However, in some cases, the spoke client instance 303 may not update the spoke object tables 307 successfully. For instance, the staging records may have been corrupted during transmission, there may be an invalid entry sent from the hub client instance 302, the spoke object table 307 may have been inadvertently corrupted, and so on. In such cases, the spoke client instance 303 may generate an error when a record of a spoke object table 307 is not updated successfully. In such cases, the spoke client instance 303 may indicate or flag the corresponding staging record as generating the error. Moreover, in some embodiments, the spoke client instance 303 may stop updating the remaining records, and indicate or flag the remaining staging records as deferred. In alternative or additional embodiments, the spoke client instance 303 may continue updating the remaining records of the spoke object table 307, and indicate or flag the remaining staging records as successfully updated or generating errors based on the success or failure of updating the records.
With the foregoing in mind,
The spoke client instance 303 may send the statuses of the staging records to the hub client instance 302 in the communication 314. In response, the hub client instance 302 may update corresponding staging records stored by the hub client instance 302.
When the hub client instance 302 receives a request for updated (e.g., staging) records, such as in the communication 310, the hub client instance 302 sends at least some of the staging records to the spoke client instance 303 in the communication 312. The hub client instance 302 may send any suitable number of staging records in any suitable data structure to the spoke client instance 303 in the communication 312. For example, the hub client instance 302 may send one, two, ten, 25, 50, 100, 200, 500, 1000, and so on, staging records to the spoke client instance 303 in the communication 312. The hub client instance 302 may then change the status of the staging records sent to the spoke client instance 303 to “in progress” 330.
After attempting to update records of one or more corresponding spoke object tables 307 based on the staging records, the spoke client instance 303 may send the statuses of the staging records to the hub client instance 302 in the communication 314. For the staging records that correspond to a successful status as indicated by the spoke client instance 303, the hub client instance 302 may change the status of the staging records to “synced” 332. For the staging records that correspond to an error status as indicated by the spoke client instance 303, the hub client instance 302 may change the status of the staging records to “error” 334. For the staging records that correspond to a deferred status as indicated by the spoke client instance 303, the hub client instance 302 may change the status of the staging records to “deferred” 336. At a later time, the spoke client instance 303 may reattempt to update the records and/or send a subsequent communication (such as the communication 310) requesting updates for the records that were not successfully updated (e.g., corresponding to the staging records with error statuses and/or deferred statuses).
As illustrated, the hub client instance 302 may include a synchronized objects table 346, a hub staging records table 348, and a spokes table 350. The spoke client instance 303 may include an object configuration table 358 and a spoke staging records table 358. The synchronized objects table 346 may tabulate the hub objects 304 in the hub client instance 302 for synchronization with one or more respective spoke client instances 303. The hub staging records table 348 may stage records that were inserted, updated, and/or deleted in the hub object tables 305, and thus should be synchronized with one or more spoke client instances 303. The spokes table 350 may tabulate the spoke client instances 303 which should be synchronized with the hub client interface 302. The object configuration table 358 may tabulate configurations of the hub objects 304 that are synchronized with the one or more spoke client instances 303 (e.g., fields that should be synchronized, fields that should be excluded during synchronization, fields that should be set to default values, and the like). The spoke staging records table 358 may cache the staging records received from hub client instance 302 and track the type of synchronization operation (e.g., insertion, update, and/or deletion) to be performed. Table 1 below lists the tables for managing the hub client instance 302 and the spoke client instances 303, and their respective purposes.
In some embodiments, the hub staging records table 348 (stored in the hub client interface 302) may include the following fields shown in Table 2 below: “Instance ID”, “Table Name”, “Operation”, “Record Payload”, “Response Message”, “Sync Record”, and/or “State”. The Instance ID field may identify which specific spoke client instance 303, and which spoke object table 307 in the specific spoke client instance 303, that a record is being staged for. The Table Name may identify which hub object table 305 and which spoke object table 307 the record getting staged belongs to. The Operation field may detail what operation (e.g., insertion, update, and/or deletion) that is to be performed for the staging record. The Record Payload field may store a representation of the staging record to be synchronized with the spoke client instance 303. The representation may be populated while staging the record, and may be an open-standard file format or data interchange format that uses human-readable text to transmit data objects having attribute-value pairs and/or array data types. For example, the representation may be in JavaScript Object Notation (JSON) format. The Response Message field may store a received message from a spoke client instance 303 regarding a result of an operation on a spoke client instance 303. The State function may track the state of the synchronization process for the staging record, as described in
Additionally or alternatively, the spokes table 350 (stored in the hub client interface 302) may include the following fields shown in Table 3: “Active”, “Spoke Instance ID”, and/or “Sync Retrieval Limit”. In particular, the Active field may allow or disallow the spoke client instance 303 to synchronize with the hub client instance 302. The Spoke Instance ID field may identify the spoke client instance 303. The Sync Retrieval Limit field may control the number of staging records to be returned per synchronization request.
The synchronized objects table 346 (stored in the hub client interface 302) may include the following fields showing in Table 4: “Active”, “Spoke Instance ID”, “Get Everything”, and/or “Payload Fields”. In particular, the Active field may allow or disallow a hub object table's 305 data to be staged for a specific spoke client instance 303. The Spoke Instance ID field may identify the spoke client instance 303 for which the hub object table 305 record is staged. The Get Everything field, if true, may return fields (e.g., all fields) of a hub object table 305 as part of the payload, and, if false, stage the fields present in the Payload Fields field. The Payload Fields field may list the fields to be staged for synchronization, where the fields may be separated by commas.
The specific embodiments described above have been shown by way of example, and it should be understood that these embodiments may be susceptible to various modifications and alternative forms. It should be further understood that the claims are not intended to be limited to the particular forms disclosed, but rather to cover all modifications, equivalents, and alternatives falling within the spirit and scope of this disclosure.
The techniques presented and claimed herein are referenced and applied to material objects and concrete examples of a practical nature that demonstrably improve the present technical field and, as such, are not abstract, intangible or purely theoretical. Further, if any claims appended to the end of this specification contain one or more elements designated as “means for [perform]ing [a function] . . . ” or “step for [perform]ing [a function] . . . ”, it is intended that such elements are to be interpreted under 35 U.S.C. 112(f). However, for any claims containing elements designated in any other manner, it is intended that such elements are not to be interpreted under 35 U.S.C. 112(f).
Number | Name | Date | Kind |
---|---|---|---|
6321229 | Goldman | Nov 2001 | B1 |
6609122 | Ensor | Aug 2003 | B1 |
6799189 | Huxoll | Sep 2004 | B2 |
6816898 | Scarpelli | Nov 2004 | B1 |
6895586 | Brasher | May 2005 | B1 |
7020706 | Cates | Mar 2006 | B2 |
7027411 | Pulsipher | Apr 2006 | B1 |
7028301 | Ding | Apr 2006 | B2 |
7062683 | Warpenburg | Jun 2006 | B2 |
7131037 | LeFaive | Oct 2006 | B1 |
7170864 | Matharu | Jan 2007 | B2 |
7216142 | Whipple | May 2007 | B2 |
7350209 | Shum | Mar 2008 | B2 |
7392300 | Anantharangachar | Jun 2008 | B2 |
7468956 | Leelanivas | Dec 2008 | B1 |
7610512 | Gerber | Oct 2009 | B2 |
7617073 | Trinon | Nov 2009 | B2 |
7685167 | Mueller | Mar 2010 | B2 |
7689628 | Garg | Mar 2010 | B2 |
7698398 | Lai | Apr 2010 | B1 |
7716353 | Golovinsky | May 2010 | B2 |
7769718 | Murley | Aug 2010 | B2 |
7783744 | Garg | Aug 2010 | B2 |
7801030 | Aggarwal | Sep 2010 | B1 |
7890802 | Gerber | Feb 2011 | B2 |
7925981 | Pourheidari | Apr 2011 | B2 |
7930396 | Trinon | Apr 2011 | B2 |
7933927 | Dee | Apr 2011 | B2 |
7945860 | Vambenepe | May 2011 | B2 |
7962743 | Fluhrer | Jun 2011 | B2 |
7966398 | Wiles | Jun 2011 | B2 |
8051164 | Peuter | Nov 2011 | B2 |
8082222 | Rangarajan | Dec 2011 | B2 |
8085690 | Leelanivas | Dec 2011 | B1 |
8151261 | Sirota | Apr 2012 | B2 |
8224683 | Manos | Jul 2012 | B2 |
8266096 | Navarrete | Sep 2012 | B2 |
8380645 | Kowalski | Feb 2013 | B2 |
8400945 | Barnes | Mar 2013 | B2 |
8402127 | Solin | Mar 2013 | B2 |
8457928 | Dang | Jun 2013 | B2 |
8478569 | Scarpelli | Jul 2013 | B2 |
8554750 | Rangaranjan | Oct 2013 | B2 |
8612408 | Trinon | Dec 2013 | B2 |
8646093 | Myers | Feb 2014 | B2 |
8674992 | Poston | Mar 2014 | B2 |
8743121 | De Peuter | Jun 2014 | B2 |
8745040 | Kowalski | Jun 2014 | B2 |
8812539 | Milousheff | Aug 2014 | B2 |
8818994 | Kowalski | Aug 2014 | B2 |
8832652 | Mueller | Sep 2014 | B2 |
8887133 | Behnia | Nov 2014 | B2 |
8984598 | Branch | Mar 2015 | B2 |
9015188 | Behne | Apr 2015 | B2 |
9037536 | Vos | May 2015 | B2 |
9065783 | Ding | Jun 2015 | B2 |
9098322 | Apte | Aug 2015 | B2 |
9122552 | Whitney | Sep 2015 | B2 |
9137115 | Mayfield | Sep 2015 | B2 |
9185078 | Wong | Nov 2015 | B2 |
9239857 | Trinon | Jan 2016 | B2 |
9317327 | Apte | Apr 2016 | B2 |
9323801 | Morozov | Apr 2016 | B2 |
9363252 | Mueeler | Jun 2016 | B2 |
9419857 | Ryan | Aug 2016 | B1 |
9412084 | Kowalski | Sep 2016 | B2 |
9467344 | Gere | Oct 2016 | B2 |
9535737 | Joy | Jan 2017 | B2 |
9613070 | Kumar | Apr 2017 | B2 |
9645833 | Mueller | May 2017 | B2 |
9646169 | Wong | May 2017 | B2 |
9654473 | Miller | May 2017 | B2 |
9659051 | Hutchins | May 2017 | B2 |
9674108 | Ryan | Jun 2017 | B1 |
9674285 | Detienne | Jun 2017 | B2 |
9766935 | Kelkar | Sep 2017 | B2 |
9792387 | George | Oct 2017 | B2 |
9805322 | Kelkar | Oct 2017 | B2 |
9852165 | Morozov | Dec 2017 | B2 |
10002203 | George | Jun 2018 | B2 |
10592829 | Mueller | Mar 2020 | B2 |
20030126137 | McFadden | Jul 2003 | A1 |
20050055382 | Ferrat | Mar 2005 | A1 |
20050086285 | Balasubramanian | Apr 2005 | A1 |
20070255812 | Mirtorabi | Nov 2007 | A1 |
20080082989 | Walker | Apr 2008 | A1 |
20090097417 | Asati | Apr 2009 | A1 |
20090129383 | Maalouf | May 2009 | A1 |
20120185290 | Mueller | Jul 2012 | A1 |
20130086280 | James | Apr 2013 | A1 |
20140173702 | Wong | Jun 2014 | A1 |
20150304174 | Zhao | Oct 2015 | A1 |
20160098327 | Detienne | Apr 2016 | A1 |
20160261506 | Hegde | Sep 2016 | A1 |
20170244810 | Ryan | Aug 2017 | A1 |
Number | Date | Country | |
---|---|---|---|
20210218802 A1 | Jul 2021 | US |