The present technology pertains to communications networks and, more specifically, to systems and methods for understanding identity and organizational access to applications within an enterprise environment.
The approaches described in this section could be pursued but are not necessarily approaches that have previously been conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.
Due to the extensive use of computer networks by enterprises, there has been a dramatic rise in network attacks, a proliferation of computer viruses, and a constant distribution of other types of malicious content that attempts to attack, infect, or otherwise infiltrate the computer networks. Attackers breach internal networks and public clouds to steal critical data. For example, attackers target low-risk assets to enter the internal network. Inside the internal network and public clouds, and behind the hardware firewall, attackers move laterally across the internal network, exploiting East-West traffic flows, to critical enterprise assets. Once there, attackers siphon off valuable company and customer data.
Tracking possible network attacks on enterprise assets can be difficult because an enterprise may include multiple organizational units. The multiple organizational units may not be rigid because users change their roles and business functions within organizational units and can belong to different organizational units.
This summary is provided to introduce a selection of concepts in a simplified form that are further described in the Detailed Description below. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
The present disclosure is related to various methods and systems for understanding identity and organizational access to applications within an enterprise environment.
According to one example embodiment, a method for understanding identity and organizational access to applications within an enterprise environment is provided. The method may include collecting data concerning relationships between applications and metadata associated with the applications in a computing environment of an enterprise. The metadata may include information concerning a plurality of users accessing the applications. The method may also include updating a graph database. The graph database may include nodes representing the applications of the computing environment of the enterprise and edges representing relationships between the applications. The method may also include enriching the graph database by associating the nodes with metadata associated with the applications. The method may include analyzing the graph database to identify a subset of nodes being accessed by a user of the plurality of users. The method may also include accessing an identity store to classify the user behavior into organizational units and roles in order to represent organizational behavior. The method may also include displaying, via a graphical user interface, a graphical representation of the subset of nodes and relationships between the nodes in the subset of the nodes and the relationship between the users and the nodes.
The metadata may include network logs of the user access events into the applications or other user activity conducted on client devices. The metadata may include telemetry data concerning access operations, time of day, a client device used, or an amount of data written to or read from the applications. The metadata may also include data from identity stores which the system can utilize to organize information into a consumable form. The metadata may also include Role Based Access Control (RBAC) rules and permissions associated with users and groups of users to understand the access currently allowed from the users to the applications.
The method may include analyzing the graph database to detect a violation by the user of an access right to at least one application of the applications. The method may include, in response to the violation, generating a security policy disallowing at least one relationship between the at least one user and at least one further application in the graph database. The method may also include permitting a subset of communications in the form of a whitelist in order to restrict the attack surface exposed to a user or group of users.
The method may include graphically visualizing usage behavior orientated around an application, or set of applications by demonstrating the organizational structure of users accessing those applications, facilitating intuitive evaluation of access patterns and business dependencies. Alternatively, those visualizations can be oriented around the behaviors of members of an organizational unit, facilitating the assessment of the dependencies of that organizational unit. Throughout the visualizations, it may be possible for the user to drill up and down in order to understand group behavior and the behavior of individuals within the group in order to evaluate outliers and anomalous behavior.
The method may include graphically visualizing the permissions provided to organizational units of users or individual users and overlaying that information with actual user behavior. This visualization may include visual identification of permissions that are not being utilized and user behavior (both attempted and successful) that violates the permissions. The method may include the identification of individual permissions that are recommended for removal and also the production of a score reflecting the accuracy of permissions which may be used by users of the system to prioritize remediation.
The method may include the scoring of risk of permissions per application based upon the understanding by the system of an application criticality and also the scope of permissions assigned, including a context relation to a degree of privilege.
The method may include the ability of users to select access relationships or permissions to be used to monitor future access behavior in order to produce alerts by the system.
The method may include automatic conversion of observed access behaviors into permissions, which can be delivered by the system in a configuration file or program via application programming interfaces (APIs) of the identity and access management system.
The method may include reporting on overall user access risk based on the scoring of permission accuracy and based on degree of privilege and application criticality.
According to another embodiment, a system for understanding identity and organizational access to applications within an enterprise environment is provided. The system may include at least one processor and a memory storing processor-executable codes, wherein the processor can be configured to implement the operations of the above-mentioned method for understanding identity and organizational access to applications within an enterprise environment.
According to yet another aspect of the disclosure, there is provided a non-transitory processor-readable medium, which stores processor-readable instructions. When the processor-readable instructions are executed by a processor, they cause the processor to implement the above-mentioned method for understanding identity and organizational access to applications within an enterprise environment.
Additional objects, advantages, and novel features will be set forth in part in the detailed description section of this disclosure, which follows, and in part will become apparent to those skilled in the art upon examination of this specification and the accompanying drawings or may be learned by production or operation of the example embodiments. The objects and advantages of the concepts may be realized and attained by means of the methodologies, instrumentalities, and combinations particularly pointed out in the appended claims.
Embodiments are illustrated by way of example, and not by limitation, in the figures of the accompanying drawings, in which like references indicate similar elements and in which:
While this technology is susceptible of embodiment in many different forms, there is shown in the drawings and will herein be described in detail several specific embodiments with the understanding that the present disclosure is to be considered as an exemplification of the principles of the technology and is not intended to limit the technology to the embodiments illustrated. The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the technology. As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises,” “comprising,” “includes,” and/or “including,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will be understood that like or analogous elements and/or components, referred to herein, may be identified throughout the drawings with like reference characters. It will be further understood that several of the figures are merely schematic representations of the present technology. As such, some of the components may have been distorted from their actual scale for pictorial clarity.
Each of workloads 1101,1-110X,Y can be a unit of computing resource, such as a physical computing system (also referred to as a bare metal server), virtual machine, container, pod, and combinations thereof. A physical computing system is computer hardware and not a virtual computing system, such as a virtual machine and container. In addition to running operating systems and applications, physical computing systems can be the hardware that virtual computing systems run on.
A virtual machine provides a substitute for a physical computing system, including functionality to execute entire operating systems. Virtual machines are created and run by a hypervisor or virtual machine monitor (VMM). A hypervisor is computer software or firmware which can run on workloads 1101,1-110X,Y. A hypervisor uses native execution to share and manage hardware, allowing for multiple environments which are isolated from one another, yet exist on the same physical computing system.
Containers are an operating system-level virtualization method for deploying and running distributed applications without launching an entire virtual machine for each application. Containers can look like physical computing systems from the point of view of programs running in them. Generally, a computer program running on an operating system can see all resources (e.g., connected devices, files and folders, network shares, central processing unit (CPU) power, etc.) of that physical computing system. However, programs running inside a container can only see the container's contents and devices assigned to the container. A pod is a group of containers with shared storage and/or network resources, and a shared specification for how to run the containers.
A container is an instance of an image. An image can be a file, comprised of multiple layers, with information to create a complete and executable version of an application. Containers can be arranged, coordinated, and managed—including means of discovery and communications between containers—by container orchestration (e.g., Docker Swarm®, Kubernetes®, Amazon EC2 Container Service (ECS), Diego, Red Hat OpenShift, and Apache® Mesos™). In contrast to hypervisor-based virtualization, containers may be an abstraction performed at the operating system (OS) level, whereas virtual machines are an abstraction of physical hardware.
Typically, workloads 1101,1-110X,Y of computing environment 100 individually and/or collectively run applications and/or services. Applications and/or services are programs designed to carry out operations for a specific purpose. By way of non-limiting example, applications can be a database (e.g., Microsoft® SQL Server®, MongoDB, Hadoop Distributed File System (HDFS), etc.), email server (e.g., Sendmail®, Postfix, qmail, Microsoft® Exchange Server, etc.), message queue (e.g., Apache® Qpid™, RabbitMQ®, etc.), web server (e.g., Apache® HTTP Server™, Microsoft® Internet Information Services (IIS), Nginx, etc.), Session Initiation Protocol (SIP) server (e.g., Kamailio® SIP Server, Avaya® Aura® Application Server 5300, etc.), other media server (e.g., video and/or audio streaming, live broadcast, etc.), file server (e.g., Linux server, Microsoft® Windows Server®, etc.), service-oriented architecture (SOA) and/or microservices process, object-based storage (e.g., Lustre®, EMC® Centera , Scality® RING®, etc.), directory service (e.g., Microsoft® Active Directory®, Domain Name System (DNS) hosting service, etc.), and the like.
Physical computing systems and computing environments are described further in relation to
Network logs 270 and middleware system logs can be data sources such as flow logs from cloud services 2601-260Z (e.g., Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP)), vArmour DSS Distributed Security System, Software Defined Networking (SDN) (e.g., VMware NSX and Cisco Application Centric Infrastructure (ACI)), monitoring agents (e.g., Tanium Asset and Falco), and the like. Generally, streaming telemetry 275 can be low-level data about relationships between applications. Streaming telemetry 275 can include 5-tuple, layer 7 (application layer) process information, management plane logs, and the like. 5-tuple refers to a set of five different values that comprise a Transmission Control Protocol/Internet Protocol (TCP/IP) connection: a source IP address/port number, destination IP address/port number and the protocol in use. Streaming telemetry can alternatively or additionally include a volume of data (i.e., how much data is or how many data packets are) exchanged between workloads (e.g., workloads 1101,1-110X,Y in
Cloud control plane 280 establishes and controls the network and computing resources within a computing environment (e.g., computing environment 100 in
Events 285 can include information about a container (and/or a pod) being created, having a state change, having an error, and the like. For example, when a container is created, information about the workload such as a service name, image deployed, and the like can be received in events 285. By way of further example, additional information from an image registry corresponding to the deployed image can be gathered by controller 210.
Configuration management database (CMDB) 290 can be a database of information about the hardware and software components (also known as assets) used in a computing environment (e.g., computing environment 100 in
For the purposes of identity, an identity store (directory) 272 (such as Lightweight Directory Access Protocol (LDAP)) is utilized to provide metadata associated with the organizational membership of the user, including organizational unit membership, roles, groups, permissions, and administrative status.
In an example embodiment, roles of users can be stored in a database. The role may include description of actions that the user is entitled to perform according to the role. Even though the user may be allowed to perform a number of actions according to their role, the user may not necessarily perform all of the allowed actions. Furthermore, the user may perform some actions only once (e.g., when accessing the application for the first time).
Streaming identity 277, streaming telemetry 275, events 285, and inventory 295 can be ingested by graph 220. Graph 220 normalizes information received in streaming telemetry 275, events 285, and inventory 295 into a standard data format and/or model, graph database 225. Graph database 225 uses a graph data model comprised of nodes (also referred to as vertices), which is an entity such as a workload (e.g., of workloads 1101,1-110X,Y in
Conventionally, security management systems stored raw logs of each and every individual communication between workloads. The amount of data scaled linearly and consumed massive amounts of storage. In contrast, streaming telemetry 275, events 285, and inventory 295, graph 220 (
Using streaming telemetry 275, events 285, and inventory 295, graph 220 (
Referring back to
On the other hand, a web server connected to the Internet will have high entropy, because the number of relationships (connections) to clients on the Internet (nodes) is huge and continues to grow. To protect the size of graph database 225, graph 220 can recognize when there is high entropy and summarize the nodes. For example, the vast (and growing) number of clients on the Internet is represented by a single “Internet” object with one edge to the web server node.
According to some embodiments, a new relationship can be created around a particular node in graph database 225, as streaming telemetry 275, events 285, and inventory 295 are processed by graph 220. Graph 220 (
Visualize 230 can visually present information from graph database 225 to users according to various criteria, such as by application, application type, organization, and the like.
Visualize 230 can visually organize information from graph database 225. In some embodiments, nodes that behave similarly can be clustered together (i.e., be put in a cluster). For example, when two nodes have similar edges (relationships) and behave in a similar fashion (e.g., run the same application, are associated with the same organization, and the like), the two nodes can be clustered together. Nodes that are clustered together can be visually presented as a shape (e.g., circle, rectangle, and the like) which denotes that there are a certain number of workloads fulfilling the same function, instead of presenting a shape for each workload in the cluster.
In various embodiments, visualize 230 can detect and present communities. Communities are workloads (e.g., of workloads 1101,1-110X,Y in
Protect 240 can use information in the graph database 225 to design security policies. Security policies can implement security controls, for example, to protect an application wherever it is in a computing environment (e.g., computing environment 100 in
Nouns and verbs can be described in a security template. A security template can include logic about how to process information in graph database 225 relating to workloads having a particular label/selection (nouns). Labels can be provided by logs 270 (e.g., layer 7 information), cloud control planes 280 (e.g., container orchestration), and CMDB 290. Protect 240 uses a security template to extract workloads to be protected (nouns) from graph database 225. Protect 240 further applies logic in the security template about how to protect the workloads (verbs) to produce a security policy. In various embodiments, security templates are JavaScript Object Notation (JSON) documents, documents in Jinja (or Jinja2), YAML Ain't Markup Language (YAML) documents, Open Policy Agent (OPA) rules, and the like. Jinja and Jinja2are a web template engine for the Python programming language. YAML is a human-readable data-serialization language. OPA is an open source, general-purpose policy engine that enables unified, context-aware policy enforcement. Security templates are described further in U.S. patent application Ser. No. 16/428,838 (Attorney Docket No. PA9274US) filed May 31, 2019, titled “Template-Driven Intent-Based Security,” which is hereby incorporated by reference for disclosure of generating a security policy using security templates.
Protect 240 can produce multiple security policies, each reflecting independent pieces of security logic that can be implemented by protect 240. In various embodiments, security policies are JavaScript Object Notation (JSON) documents which are described to a user (responsible for system administration, security, and the like) in natural language. A natural language is any language that has evolved naturally in humans through use and repetition without conscious planning or premeditation. Natural language can broadly be defined in contrast to artificial or constructed languages such as computer programming languages. The multiple security policies can be placed in an order of precedence to resolve potential conflicts. Visualize 230 can be used to visualize the security policy (or security policies), showing the workloads protected, permitted relationships, and prohibited relationships. Protect 240 can then be used to edit the security policy. For example, there can be a primary and backup server (e.g., of workloads 1101,1-110X,Y in
Protect 240 can validate a security policy. The security policy can be simulated using graph database 225. For example, a simulation can report which applications are broken (e.g., communications among nodes needed by the application to operate are prohibited) by the security policy, are unnecessarily exposed by weak policy, and the like. Security policy validation is described further in U.S. patent application Ser. No. 16/428,849 filed May 31, 2019, titled “Validation of Cloud Security Policies,” which is incorporated by reference herein for disclosure of security policy validation.
Protect 240 can test a security policy. Protect can use historical data in graph database 225 to determine entropy in the computing environment (e.g., computing environment 100 in
Protect 240 can deploy a security policy (or security policies). The security policy is deployed as needed in one or more computing environments of cloud services 2601-260Z (e.g., Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP)), vArmour DSS Distributed Security System, VMware NSX, and the like). Protect 240 can provide the security policy to one or more of cloud drivers 2501-250Z. Cloud drivers 2501-250Z maintain an inventory and topology (i.e., current state) of the workloads in the computing environments hosted by cloud services 2601-260Z, respectively. Cloud drivers 2501-250Z can use their respective inventory and topology to apply the security policy to the appropriate workloads, and respond immediately to changes in workload topology and workload placement.
Cloud drivers 2501-250Z can serve as an interface between protect 240 (having a centralized security policy) and cloud services 2601-260Z. In other words, cloud drivers 2501-250Z implement the security policy using the different facilities (e.g., application programming interfaces (APIs)) and capabilities available from cloud services 2601-260Z. For example, each of cloud services 2601-260Z can have different syntax and semantics for implementing security controls. Moreover, each of cloud services 2601-260Z can have different security capabilities (e.g., communications/connections between workloads can only be expressly permitted and not expressly prohibited), rule capacity (limit on the number of rules), optimization methods, and the like.
Cloud drivers 2501-250Z can maintain the integrity of the security policy in the computing environments hosted by cloud services 2601-260Z (referred to as the “cloud”). Cloud drivers 2501-250Z can check that the security policy actually deployed in the cloud is as it should be, using the security policy's JSON source. When the security policy deployed in the cloud does not comport with the centralized security policy—such as when a bad actor logs into one of the cloud services and removes all the security rules—the responsible cloud driver (of cloud drivers 2501-250Z) can re-deploy the security policy and/or raise an operational alert. Where supported, cloud services 2601-260Z can notify the respective cloud driver (of cloud drivers 2501-250Z) of changes to the topology and/or configuration. Otherwise, the respective cloud driver (of cloud drivers 2501-250Z) can poll the cloud service (cloud services 2601-260Z) to ensure the security rules are in place.
As described above, a security policy can be pushed down to the computing environments hosted by cloud services 2601-260Z using cloud drivers 2501-250Z, respectively. Additionally, or alternatively, as new data comes into graph 220 as network logs 270, events 285 from cloud control plane 280, and inventory 295, protect 240 can check the new data against the security policy to detect violations and or drift (e.g., change in the environment and/or configuration).
Protect 240 can dynamically update a security policy as changes occur in the computing environments hosted by cloud services 2601-260Z. For example, when a container (or pod) is deployed by container orchestration, it can be given a label, and cloud control plane 280 reports a container is deployed (as events 285). Labels can be predefined to specify identifying attributes of containers (and pods), such the container's application function. When the label corresponds to an attribute covered by an active (deployed) security policy, protect 240 can dynamically add the new container to the active security policy (as a target). For example, when a pod is deployed for a particular organization and there is an active policy for that organization, the new workload is added to the security policy. Similarly, when a container is killed, the workload is removed from the security policy. Dynamically updating security policy is described further in U.S. patent application Ser. No. 9,521,115 issued Dec. 13, 2016, titled “Security Policy Generation Using Container Metadata,” which is hereby incorporated by reference for disclosure of dynamically updating security policy.
At step 520, a graph database can be created or updated using the cloud data. For example, streaming telemetry 275, events 285, and inventory 295 (
At step 530, a visual representation of the computing environment as modeled by the graph database can be provided. For example, visualize 230 (
At step 540, a security template can be received. A security template can include logic about how to extract information from graph database 225 to identify workloads to be targets of a security policy. In addition, a security template can specify how the workloads are to be protected (e.g., security intent).
At step 550, a security policy can be created. For example, protect 240 can use the security template to extract information from graph database 225 (
At step 560, the security policy can be validated. For example, protect 240 (
At step 570, the security policy can be tested. For example, protect 240 (
At step 580, the security policy can be deployed to the computing environment (e.g., computing environment 100 in
Optionally at step 580, the security policy can be maintained. For example, cloud drivers 2501-250Z can make sure the security policy remains in force at the computing environment hosted by a respective one of cloud services 2601-260Z. Optionally at step 580, the security policy can be dynamically updated as workloads subject to the deployed security policy are deployed and/or killed.
Although steps 510-580 are shown in a particular sequential order, various embodiments can perform steps 510-580 in different orders, perform some of steps 510-580 concurrently, and/or omit some of steps 510-580.
The components shown in
Mass data storage 630, which can be implemented with a magnetic disk drive, solid state drive, or an optical disk drive, is a non-volatile storage device for storing data and instructions for use by processor unit(s) 610. Mass data storage 630 stores the system software for implementing embodiments of the present disclosure for purposes of loading that software into main memory 620.
Portable storage device 640 operates in conjunction with a portable non-volatile storage medium, such as a flash drive, floppy disk, compact disk, digital video disc, or Universal Serial Bus (USB) storage device, to input and output data and code to and from the computer system 600 in
User input devices 660 can provide a portion of a user interface. User input devices 660 may include one or more microphones, an alphanumeric keypad, such as a keyboard, for inputting alphanumeric and other information, or a pointing device, such as a mouse, a trackball, stylus, or cursor direction keys. User input devices 660 can also include a touchscreen. Additionally, the computer system 600 as shown in
Graphics display system 670 include a liquid crystal display (LCD) or other suitable display device. Graphics display system 670 is configurable to receive textual and graphical information and processes the information for output to the display device.
Peripheral device(s) 680 may include any type of computer support device to add additional functionality to the computer system.
Some of the components provided in the computer system 600 in
Some of the above-described functions may be composed of instructions that are stored on storage media (e.g., computer-readable medium). The instructions may be retrieved and executed by the processor. Some examples of storage media are memory devices, tapes, disks, and the like. The instructions are operational when executed by the processor to direct the processor to operate in accord with the technology. Those skilled in the art are familiar with instructions, processor(s), and storage media.
In some embodiments, the computer system 600 may be implemented as a cloud-based computing environment, such as a virtual machine operating within a computing cloud. In other embodiments, the computer system 600 may itself include a cloud-based computing environment, where the functionalities of the computer system 600 are executed in a distributed fashion. Thus, the computer system 600, when configured as a computing cloud, may include pluralities of computing devices in various forms, as will be described in greater detail below.
In general, a cloud-based computing environment is a resource that typically combines the computational power of a large grouping of processors (such as within web servers) and/or that combines the storage capacity of a large grouping of computer memories or storage devices. Systems that provide cloud-based resources may be utilized exclusively by their owners or such systems may be accessible to outside users who deploy applications within the computing infrastructure to obtain the benefit of large computational or storage resources.
The cloud is formed, for example, by a network of web servers that comprise a plurality of computing devices, such as the computer system 600, with each server (or at least a plurality thereof) providing processor and/or storage resources. These servers manage workloads provided by multiple users (e.g., cloud resource customers or other users). Typically, each user places workload demands upon the cloud that vary in real-time, sometimes dramatically. The nature and extent of these variations typically depends on the type of business associated with the user.
It is noteworthy that any hardware platform suitable for performing the processing described herein is suitable for use with the technology. The terms “computer-readable storage medium” and “computer-readable storage media” as used herein refer to any medium or media that participate in providing instructions to a CPU for execution. Such media can take many forms, including, but not limited to, non-volatile media, volatile media and transmission media. Non-volatile media include, for example, optical, magnetic, and solid-state disks, such as a fixed disk. Volatile media include dynamic memory, such as system random-access memory (RAM). Transmission media include coaxial cables, copper wire and fiber optics, among others, including the wires that comprise one embodiment of a bus. Transmission media can also take the form of acoustic or light waves, such as those generated during radio frequency (RF) and infrared (IR) data communications. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, a hard disk, magnetic tape, any other magnetic medium, a CD-ROM disk, digital video disk (DVD), any other optical medium, any other physical medium with patterns of marks or holes, a RAM, a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), a Flash memory, any other memory chip or data exchange adapter, a carrier wave, or any other medium from which a computer can read.
Various forms of computer-readable media may be involved in carrying one or more sequences of one or more instructions to a CPU for execution. A bus carries the data to system RAM, from which a CPU retrieves and executes the instructions. The instructions received by system RAM can optionally be stored on a fixed disk either before or after execution by a CPU.
Computer program code for carrying out operations for aspects of the present technology may be written in any combination of one or more programming languages, including an object oriented programming language such as JAVA, SMALLTALK, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Using streaming telemetry 275, events 285, and inventory 295, graph 220 (
According to various exemplary embodiments, within the graph 220 (
By interfacing with inventory systems such as a configuration management database (“CMDB”), which could be a ServiceNow® CMDB (for example, Node 710), an enterprise governance, risk and compliance (“GRC”) tool such as RSA Archer® (for example, Node 730), or a “home built” tool such as the CMDB tools that many banks have developed (for example, Node 750), there is what is known as “common metadata.” Common metadata may comprise compliance information.
For example, a system may be considered to be a Payment Card Industry Data Security Standard (“PCI DSS”) Category 1 system, or the system may have certain operational service level objectives.
In one example, the system (for example, a Node) has a recovery time objective (“RTO”). This is the number of hours that a system needs to recover within in order to meet its service level agreement, its service level requirements, or its recovery point objective (“RPO”), which is the amount of time that a system can lose data.
When there is a failure, if the RPO is zero, it means the system is not allowed to lose any data. For example, by mapping into the graph relationships between systems that are publishing data and systems that are consuming data, one may query whether or not there are risks or issues relating to the ability of a system to recover.
For example, Node 710 may send out information consumed by Node 730. Node 710 may have a recovery time objective of four hours. Node 730 may have a recovery time objective of one hour. Generally, organizations know what the recovery time objective is for individual components of their systems. But they have difficulty mapping it across the dynamic dependencies that have been created. This is a problem. In the case of Node 730, it needs to recover its data within an hour. Node 710, however, might not recover the data for four hours. This is known as a “toxic combination.”
According to the exemplary systems and methods herein, it has a standard set of policies to recognize where these mismatches occur. It can identify in advance problematic dependencies such as described herein.
As another example, if a system is labeled as being PCI DSS Level 1, there is a rule within the PCI DSS specification that it cannot be accessed directly from a system on the Internet. The graph will show if a system on the Internet is connecting to a system that is labelled as PCI DSS Level 1 and that it is a toxic combination. This can be prevented by writing a policy, a set of permissions, or triggering an alarm as soon as it occurs.
Another example of the ability to use this dependency mapping to find toxic combinations is if Node 730 has a recovery point objective for data loss of zero and Node 710 has a recovery point objective of one hour. Node 710 could lose data for an hour, however, Node 730 cannot afford to lose any data at all. Connecting together these dependencies based on metadata about operational requirements or compliance obligations means that the exemplary systems and methods described herein can immediately find these problems. In most cases, the application controller performs this function.
In further exemplary embodiments, multiple dependencies may be analyzed within a dependency tree. It is not constrained to just an analysis of just two assets. As shown in dependency risk report 800, 90 internal workloads, 26 dependency workloads, and 2 unknown workloads comprising 2,374 relationships, 26 relationships per workload and 226 active rules were analyzed. Additionally, the exemplary systems and methods herein may be used with audits. Auditors require organizations to prove that they have the capability to ensure that their critical functions meet their business risk requirements. The output of this system can show when a mismatch occurs or when a violation occurs . For example, dependency risk report 800 shows four violations between two nodes (source and destination, respectively) where each violation is recovery time objective (“RTO”) mismatch of 1 to 4. For example, the source needs to recover its data in 1 hour, however, the destination might take up to four hours to recover its data. A violation can occur for a number of reasons. It could be because the needs of an application have changed, and/or it could be because there is a new dependency between applications.
In some exemplary embodiments, organizations can produce an audit trail to prove that they have continuous and immediate capability to understand when risk profiles change. They can also demonstrate that they have a means of preventing risk profiles from changing. For example, the exemplary systems and methods herein can prevent a relationship between an out of scope PCI DSS system and a properly functioning PCI DSS Level 1 system that stores and processes credit card information.
Many regulatory agencies feel that business continuity should not be focused only on the planning process to recover operations after an event, but it should include the continued maintenance of systems and controls for the resilience of operations. Resilience incorporates, prior to measures to mitigate disruptive events, the ability to evaluate an entity's recovery capabilities. The exemplary systems and methods herein can automatically mitigate any toxic combinations in a complex network. Rules can be generated to find where there are combinations of conditions that will not allow an organization to meet their resilience and operational risk requirements. In some cases, policies may be provided with the system upon acquisition. But equally, users can label the system with their own operational conditions or their own metadata and write a rule, which allows them to evaluate when a problematic combination of things exists.
The nodes 910-960 can be enriched with metadata 910-M, 920-M, 930-M, 940-M, 950-M, and 960-M. The metadata 910-M, 920-M, 930-M, 940-M, 950-M may include information concerning users accessing the applications corresponding to the nodes 910, 920, 930, 940, 950, and 960. The metadata 910-M, 920-M, 930-M, 940-M, 950-M can be collected by controller 210 from network logs 270 (shown in
The metadata may also include information concerning connections between the applications. The information concerning the connections can be used to determine relationships between the nodes. For example, if two applications communicate with each other, via an Application Programming Interface (API) gateway, these two applications need to identify, authenticate, and authorize each other. These applications may pass a set of credentials and establish a set of tokens, Jason web tokens to allow these applications to transact with each other. The relationships between the applications can be mapped based on flow telemetry. The relationships can be attributed meanings based of semantics of identity, such as a person identity associated with a user, machine identity, and API identity.
The metadata discovered about user access behavior may include network logs of the user access events into the applications or other user activity conducted on client devices. The metadata may include telemetry data concerning access operations, time of day, a client device used or an amount of data written to or read from the applications. The metadata may also include data from an identity store utilized by the system to organize information into a consumable form around organization units, groups, and roles. The metadata may also include the RBAC rules and permissions associated with the users and groups of users to understand the access to the applications currently allowed for the users.
The controller 210 can analyze metadata to determine the access behavior of users within different organizational units defined within the identity store. This access behavior can be compared against policies that are defined around organizational access permissions to ensure that corporate policies are not breached.
For example, the first organizational unit may be a trading department in a bank and the second organizational unit may be an advisory department of the same bank. Advisors in the advisory department may possess insider information (e.g., acquisition negotiations). The traders in the trader department are not allowed to have access to the insider information because it is against the law.
In other embodiments, the subset 915 may include applications used by users performing a first role in the enterprise and the subset 925 may include applications used by users performing a second role in the enterprise. The subset 935 may represent potentially vulnerable nodes if security rules of the enterprise prohibit using the same applications or access the same data by users with different roles or business functions within the enterprise.
The roles of users, groups of users within the enterprise, and applications used by the enterprise are not static. The users may also be members of overlapping groups. Therefore, the graph database needs to be constantly updated and analyzed to determine inappropriate combinations of roles of users and overlapping groups. The automated updating may occur continuously via APIs. This approach may allow detecting, for example, when the movement of a user to another organizational unit violates a policy for access to resources across organizational units.
Some embodiments may allow assessing the relationships between the applications and groups of users. The groups may include groups of employees within an organizational unit, employees with similar roles, groups of customers of the enterprise, and so forth. Certain embodiments may allow detecting attempts or potential exploits to gain access between different environments. The detection of potential exploits may be implemented with monitoring policies, so if anybody in the future attempts to gain access between different environments, an authorized person can be notified immediately. The detection of potential exploits may also allow creating a security policy. The security policy may allow only those relationships that have been established and validated, and disallow creation of any other relationships in the future.
Alternatively, the prevention of potential exploits may be implemented using RBAC-based permissions which are deployed via API onto the Identity Access Management (IAM) system within the environment (for example, LDAP, Okta, Sailpoint). This approach facilitates the prevention of access which violates corporate policy. These RBAC policies can be defined via a whitelist (commonly known as a zero trust rule set) or as a backlist defining the access that in not permitted.
Some embodiments may allow, upon identifying current relationships between different environments, review of the relationships between applications and creating a baseline. The baseline may include currently allowed relationships. The graph database can be constantly updated based on metadata collected from networks logs and analyzed to detect deviations from the baseline.
The graph database can be analyzed to determine connections between the applications associated with an individual user, users of a specific role, or users belonging to specific groups. The result of this analysis may allow providing a view of functions of a department, division, and members of the certain roles within the enterprise. This analysis may also be used to represent the groups of users accessing an application to enable an application owner to improve the access policies for their service.
The method 1000 may commence in block 1002 with collecting data concerning relationships between applications and metadata associated with the applications in a computing environment of an enterprise. The metadata may include information concerning a plurality of users accessing the applications. The metadata may include network logs of the logins of the users into the applications and network communications between systems to which users are connected and the applications. The metadata may include telemetry data concerning an amount of data written to or read from the applications and operations performed during the access session where available.
In block 1004, the method 1000 may include updating a graph database including nodes representing the applications of the computing environment of the enterprise and edges representing relationships between the applications.
In block 1006, the method 1000 may include enriching the graph database by associating the nodes with metadata associated with the applications.
In block 1008, the method 1000 may include enriching the graph database by associating user accounts associated with the plurality of users with metadata associated with roles, organizational membership, privileges, and permissions provided to the users through an IAM system.
In block 1010, the method 1000 may include analyzing the graph database to identify a subset of nodes being accessed by a user of the plurality of users.
In block 1012, the method 1000 may include displaying, via a graphical user interface, a graphical representation of the subset of nodes and relationships and permissions between the nodes in the subset of the nodes.
In block 1014, the method 1000 may display, via the graphical user interface, a graphical representation of a subset of users defined by at least one of a group, a role, and an organizational membership, and relationships between the nodes associated with the subset of users.
In block 1016, the method 1000 may display, via the graphical user interface, a graphical representation of the nodes representing the applications and groups of users accessing the applications.
In block 1018, the method 1000 may display, via the graphical user interface, a graphical representation of the permissions provided to the subset of users defined by at least one of the group, the role, and organizational unit in relation to the nodes representing the applications.
In block 1020, the method 1000 may compare the permissions with relationships related to accessing the applications by the subset of users. The relationships related to accessing the applications may be recorded to the graph database.
The method 1000 may further include analyzing the graph database to detect a violation by the user of an access right to at least one application of the applications.
The method 1000 may further include generating, in response to the violation, a security policy disallowing at least one relationship between the at least one application and at least one further application in the graph database.
In an example embodiment, the method 1000 may include accessing an identity store to classify behavior of the plurality of users into organizational units and roles associated with the plurality of users to represent organizational behavior associated with the plurality of users.
The method 1000 may further include permitting a subset of communications between the nodes by generating a whitelist identifying at least one user of the plurality of users permitted to access at least one application.
In an example embodiment, the method 1000 may include identifying one or more of the permissions unutilized by at least one of the plurality of users and generate a score reflecting an accuracy of the permissions provided to the plurality of users. The one or more of the permissions may be recommended for removal. Additionally, a risk associated with permissions for each application may be scored to determine a criticality associated with the applications, and a degree of privilege associated with the plurality of users. An overall user access risk may be determined based on the accuracy of the permissions, the criticality associated with the applications, and the degree of privilege associated with the plurality of users.
The method 1000 may further include generating further permissions for the plurality of users based on the metadata associated with accessing the applications by the plurality of users.
The method 1000 may include providing a graphical representation of actions of the user with respect to the actions the user is entitled to perform and actions that the user actually performs with respect to the application. Thus, the system can analyze the activity and behavior of users, and analyze the information related to the access of the users to the application and actions performed by the users with respect to the application. Permissions provided to the user based on the role may be compared with access of the user to the application and actions performed by the user with respect to the application. If it is determined that a part of permissions are not used by the user, the authorized person may decide to remove one or more actions from the user permissions. Changing permission for user access by disallowing the one or more actions that have been never performed by the user of a particular role or users of a particular organizational unit may help reduce security risk associated with the application. Therefore, the system can enable comparing permissions given to a user and permissions used by the user, determining that the permissions provided to the user are broader than necessary (e.g., some permissions are never used by the user), and allow limiting the permissions.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present technology has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. Exemplary embodiments were chosen and described in order to best explain the principles of the present technology and its practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
Aspects of the present technology are described above with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present technology. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The description of the present technology has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. Exemplary embodiments were chosen and described in order to best explain the principles of the present technology and its practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
The present application is a continuation-in-part of U.S. application Ser. No. 17/133,466, filed on Dec. 23, 2020 and titled “Modeling Application Dependencies to identify Operational Risk”, which is a continuation-in-part and claims the priority benefit of U.S. patent application Ser. No. 16/428,828 filed on May 31, 2019 and titled “Cloud Security Management,” which are hereby incorporated by reference. The U.S. application Ser. No. 16/428,828 is related to U.S. patent application Ser. No. 17/133,451 filed on Dec. 23, 2020 and titled “Modeling Topic-Based Message-Oriented Middleware Within a Security System” and U.S. patent application Ser. No. 17/133,458 filed on Dec. 23, 2020 and titled “Modeling Queue-Based Message-Oriented Middleware Relationships in a Security System.” The subject matter of afore-mentioned applications is incorporated herein by reference for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
6158007 | Moreh et al. | Dec 2000 | A |
6253321 | Nikander et al. | Jun 2001 | B1 |
6484261 | Wiegel | Nov 2002 | B1 |
6578076 | Putzolu | Jun 2003 | B1 |
6765864 | Natarajan et al. | Jul 2004 | B1 |
6832243 | Mikaisen et al. | Dec 2004 | B1 |
6970459 | Meier | Nov 2005 | B1 |
6981155 | Lyle et al. | Dec 2005 | B1 |
7058712 | Vasko et al. | Jun 2006 | B1 |
7062566 | Amara et al. | Jun 2006 | B2 |
7096260 | Zavalkovsky et al. | Aug 2006 | B1 |
7373524 | Motsinger et al. | May 2008 | B2 |
7397794 | Lacroute et al. | Jul 2008 | B1 |
7467408 | O'Toole, Jr. | Dec 2008 | B1 |
7475424 | Lingafelt et al. | Jan 2009 | B2 |
7516476 | Kraemer et al. | Apr 2009 | B1 |
7519062 | Kloth et al. | Apr 2009 | B1 |
7533128 | Sanchez et al. | May 2009 | B1 |
7627671 | Palma | Dec 2009 | B1 |
7694181 | Noller et al. | Apr 2010 | B2 |
7725937 | Levy | May 2010 | B1 |
7742414 | Iannaccone et al. | Jun 2010 | B1 |
7774837 | McAlister | Aug 2010 | B2 |
7797306 | Pather et al. | Sep 2010 | B1 |
7849495 | Huang et al. | Dec 2010 | B1 |
7900240 | Terzis et al. | Mar 2011 | B2 |
7904454 | Raab | Mar 2011 | B2 |
7996255 | Shenoy et al. | Aug 2011 | B1 |
8051460 | Lum et al. | Nov 2011 | B2 |
8112304 | Scates | Feb 2012 | B2 |
8254381 | Allen et al. | Aug 2012 | B2 |
8259571 | Raphel et al. | Sep 2012 | B1 |
8291495 | Burns et al. | Oct 2012 | B1 |
8296459 | Brandwine et al. | Oct 2012 | B1 |
8307422 | Varadhan et al. | Nov 2012 | B2 |
8321862 | Swamy et al. | Nov 2012 | B2 |
8353021 | Satish et al. | Jan 2013 | B1 |
8369333 | Hao et al. | Feb 2013 | B2 |
8396986 | Kanada et al. | Mar 2013 | B2 |
8429647 | Zhou | Apr 2013 | B2 |
8468113 | Harrison et al. | Jun 2013 | B2 |
8490153 | Bassett et al. | Jul 2013 | B2 |
8494000 | Nadkarni et al. | Jul 2013 | B1 |
8499330 | Albisu et al. | Jul 2013 | B1 |
8528091 | Bowen et al. | Sep 2013 | B2 |
8539548 | Overby, Jr. et al. | Sep 2013 | B1 |
8565118 | Shukla et al. | Oct 2013 | B2 |
8612744 | Shieh | Dec 2013 | B2 |
8661434 | Liang et al. | Feb 2014 | B1 |
8677496 | Wool | Mar 2014 | B2 |
8688491 | Shenoy et al. | Apr 2014 | B1 |
8726343 | Borzycki et al. | May 2014 | B1 |
8730963 | Grosser, Jr. et al. | May 2014 | B1 |
8798055 | An | Aug 2014 | B1 |
8813169 | Shieh | Aug 2014 | B2 |
8813236 | Saha et al. | Aug 2014 | B1 |
8819762 | Harrison et al. | Aug 2014 | B2 |
8898788 | Aziz et al. | Nov 2014 | B1 |
8935457 | Feng et al. | Jan 2015 | B2 |
8938782 | Sawhney et al. | Jan 2015 | B2 |
8990371 | Kalyanaraman et al. | Mar 2015 | B2 |
9009829 | Stolfo et al. | Apr 2015 | B2 |
9015299 | Shah | Apr 2015 | B1 |
9021546 | Banerjee | Apr 2015 | B1 |
9027077 | Bharali et al. | May 2015 | B1 |
9036639 | Zhang | May 2015 | B2 |
9060025 | Xu | Jun 2015 | B2 |
9141625 | Thornewell et al. | Sep 2015 | B1 |
9191327 | Shieh et al. | Nov 2015 | B2 |
9258275 | Sun et al. | Feb 2016 | B2 |
9294302 | Sun et al. | Mar 2016 | B2 |
9294442 | Lian et al. | Mar 2016 | B1 |
9361089 | Bradfield et al. | Jun 2016 | B2 |
9380027 | Lian et al. | Jun 2016 | B1 |
9405665 | Shashi et al. | Aug 2016 | B1 |
9407602 | Feghali et al. | Aug 2016 | B2 |
9521115 | Woolward | Dec 2016 | B1 |
9609083 | Shieh | Mar 2017 | B2 |
9621595 | Lian et al. | Apr 2017 | B2 |
9680852 | Wager et al. | Jun 2017 | B1 |
9762599 | Wager et al. | Sep 2017 | B2 |
9794289 | Banerjee et al. | Oct 2017 | B1 |
9973472 | Woolward et al. | May 2018 | B2 |
10009317 | Woolward | Jun 2018 | B2 |
10009381 | Lian et al. | Jun 2018 | B2 |
10091238 | Shieh et al. | Oct 2018 | B2 |
10116441 | Rubin et al. | Oct 2018 | B1 |
10191758 | Ross et al. | Jan 2019 | B2 |
10193929 | Shieh et al. | Jan 2019 | B2 |
10264025 | Woolward | Apr 2019 | B2 |
10333827 | Xu et al. | Jun 2019 | B2 |
10333986 | Lian et al. | Jun 2019 | B2 |
10382467 | Wager et al. | Aug 2019 | B2 |
10528897 | Labat et al. | Jan 2020 | B2 |
10554604 | Burcham et al. | Feb 2020 | B1 |
10630703 | Ghosh et al. | Apr 2020 | B1 |
10755334 | Eades et al. | Aug 2020 | B2 |
10862748 | Deruijter | Dec 2020 | B1 |
11290493 | Woolward et al. | Mar 2022 | B2 |
11290494 | Li et al. | Mar 2022 | B2 |
11310284 | Woolward et al. | Apr 2022 | B2 |
11470684 | Hu et al. | Oct 2022 | B2 |
11575563 | Woolward et al. | Feb 2023 | B2 |
20020031103 | Wiedeman et al. | Mar 2002 | A1 |
20020066034 | Schlossberg et al. | May 2002 | A1 |
20020178273 | Pardo-Bastellote et al. | Nov 2002 | A1 |
20030055950 | Cranor et al. | Mar 2003 | A1 |
20030177389 | Albert et al. | Sep 2003 | A1 |
20030227392 | Ebert et al. | Dec 2003 | A1 |
20040062204 | Bearden et al. | Apr 2004 | A1 |
20040095897 | Vafaei | May 2004 | A1 |
20040143631 | Banerjee et al. | Jul 2004 | A1 |
20040172557 | Nakae et al. | Sep 2004 | A1 |
20040240458 | Tv et al. | Dec 2004 | A1 |
20050021943 | Ikudome et al. | Jan 2005 | A1 |
20050033989 | Poletto et al. | Feb 2005 | A1 |
20050114829 | Robin et al. | May 2005 | A1 |
20050119905 | Wong et al. | Jun 2005 | A1 |
20050125768 | Wong et al. | Jun 2005 | A1 |
20050154576 | Tarui | Jul 2005 | A1 |
20050190758 | Gai et al. | Sep 2005 | A1 |
20050201343 | Sivalingham et al. | Sep 2005 | A1 |
20050246241 | Irizarry, Jr. et al. | Nov 2005 | A1 |
20050283823 | Okajo et al. | Dec 2005 | A1 |
20060005228 | Matsuda | Jan 2006 | A1 |
20060037077 | Gadde et al. | Feb 2006 | A1 |
20060050696 | Shah et al. | Mar 2006 | A1 |
20060085412 | Johnson et al. | Apr 2006 | A1 |
20060168070 | Thompson et al. | Jul 2006 | A1 |
20070016945 | Bassett et al. | Jan 2007 | A1 |
20070019621 | Perry et al. | Jan 2007 | A1 |
20070022090 | Graham | Jan 2007 | A1 |
20070027801 | Botzer et al. | Feb 2007 | A1 |
20070064617 | Reves | Mar 2007 | A1 |
20070079308 | Chiaramonte et al. | Apr 2007 | A1 |
20070130566 | Van Rietschote | Jun 2007 | A1 |
20070157286 | Singh et al. | Jul 2007 | A1 |
20070162400 | Brew et al. | Jul 2007 | A1 |
20070168971 | Royzen et al. | Jul 2007 | A1 |
20070192861 | Varghese et al. | Aug 2007 | A1 |
20070192863 | Kapoor et al. | Aug 2007 | A1 |
20070198656 | Mazzaferri et al. | Aug 2007 | A1 |
20070234369 | Paramisivam et al. | Oct 2007 | A1 |
20070239987 | Hoole et al. | Oct 2007 | A1 |
20070271612 | Fang et al. | Nov 2007 | A1 |
20070277222 | Pouliot | Nov 2007 | A1 |
20080016339 | Shukla | Jan 2008 | A1 |
20080016550 | McAlister | Jan 2008 | A1 |
20080083011 | McAlister et al. | Apr 2008 | A1 |
20080155239 | Chowdhury et al. | Jun 2008 | A1 |
20080163207 | Reumann et al. | Jul 2008 | A1 |
20080195670 | Boydstun | Aug 2008 | A1 |
20080229382 | Vitalos | Sep 2008 | A1 |
20080239961 | Hilerio et al. | Oct 2008 | A1 |
20080263179 | Buttner et al. | Oct 2008 | A1 |
20080301770 | Kinder | Dec 2008 | A1 |
20080307110 | Wainner et al. | Dec 2008 | A1 |
20090077621 | Lang et al. | Mar 2009 | A1 |
20090077666 | Chen et al. | Mar 2009 | A1 |
20090083445 | Ganga | Mar 2009 | A1 |
20090138316 | Weller et al. | May 2009 | A1 |
20090165078 | Samudrala et al. | Jun 2009 | A1 |
20090190585 | Allen et al. | Jul 2009 | A1 |
20090249470 | Litvin et al. | Oct 2009 | A1 |
20090260051 | Igakura | Oct 2009 | A1 |
20090268667 | Gandham et al. | Oct 2009 | A1 |
20090328187 | Meisel | Dec 2009 | A1 |
20100043068 | Varadhan et al. | Feb 2010 | A1 |
20100064341 | Aldera | Mar 2010 | A1 |
20100071025 | Devine et al. | Mar 2010 | A1 |
20100088738 | Birnbach | Apr 2010 | A1 |
20100095367 | Narayanaswamy | Apr 2010 | A1 |
20100125476 | Yeom et al. | May 2010 | A1 |
20100191863 | Wing | Jul 2010 | A1 |
20100192223 | Ismael et al. | Jul 2010 | A1 |
20100192225 | Ma et al. | Jul 2010 | A1 |
20100199349 | Ellis | Aug 2010 | A1 |
20100208699 | Lee et al. | Aug 2010 | A1 |
20100228962 | Simon et al. | Sep 2010 | A1 |
20100235880 | Chen et al. | Sep 2010 | A1 |
20100274970 | Treuhaft et al. | Oct 2010 | A1 |
20100281539 | Burns et al. | Nov 2010 | A1 |
20100287544 | Bradfield et al. | Nov 2010 | A1 |
20100333165 | Basak et al. | Dec 2010 | A1 |
20110003580 | Belrose et al. | Jan 2011 | A1 |
20110022812 | van der Linden et al. | Jan 2011 | A1 |
20110069710 | Naven et al. | Mar 2011 | A1 |
20110072486 | Hadar et al. | Mar 2011 | A1 |
20110090915 | Droux et al. | Apr 2011 | A1 |
20110113472 | Fung et al. | May 2011 | A1 |
20110138384 | Bozek et al. | Jun 2011 | A1 |
20110138441 | Neystadt et al. | Jun 2011 | A1 |
20110184993 | Chawla et al. | Jul 2011 | A1 |
20110225624 | Sawhney et al. | Sep 2011 | A1 |
20110249679 | Lin et al. | Oct 2011 | A1 |
20110263238 | Riley et al. | Oct 2011 | A1 |
20120017258 | Suzuki | Jan 2012 | A1 |
20120113989 | Akiyoshi | May 2012 | A1 |
20120130936 | Brown et al. | May 2012 | A1 |
20120131685 | Broch et al. | May 2012 | A1 |
20120185913 | Martinez et al. | Jul 2012 | A1 |
20120207174 | Shieh | Aug 2012 | A1 |
20120210428 | Blackwell | Aug 2012 | A1 |
20120216273 | Rolette et al. | Aug 2012 | A1 |
20120224057 | Gill et al. | Sep 2012 | A1 |
20120278903 | Pugh | Nov 2012 | A1 |
20120284792 | Liem | Nov 2012 | A1 |
20120297383 | Meisner et al. | Nov 2012 | A1 |
20120311144 | Akelbein et al. | Dec 2012 | A1 |
20120311575 | Song | Dec 2012 | A1 |
20120324069 | Nori et al. | Dec 2012 | A1 |
20120324567 | Couto et al. | Dec 2012 | A1 |
20130019277 | Chang et al. | Jan 2013 | A1 |
20130054536 | Sengupta | Feb 2013 | A1 |
20130081142 | McDougal et al. | Mar 2013 | A1 |
20130086399 | Tychon et al. | Apr 2013 | A1 |
20130097138 | Barkol et al. | Apr 2013 | A1 |
20130097692 | Cooper et al. | Apr 2013 | A1 |
20130145465 | Wang et al. | Jun 2013 | A1 |
20130151680 | Salinas et al. | Jun 2013 | A1 |
20130166490 | Elkins et al. | Jun 2013 | A1 |
20130166681 | Thompson et al. | Jun 2013 | A1 |
20130166720 | Takashima et al. | Jun 2013 | A1 |
20130198799 | Staggs | Aug 2013 | A1 |
20130198840 | Drissi et al. | Aug 2013 | A1 |
20130219384 | Srinivasan et al. | Aug 2013 | A1 |
20130223226 | Narayanan et al. | Aug 2013 | A1 |
20130250956 | Sun et al. | Sep 2013 | A1 |
20130263125 | Shamsee et al. | Oct 2013 | A1 |
20130275592 | Xu et al. | Oct 2013 | A1 |
20130276092 | Sun et al. | Oct 2013 | A1 |
20130283336 | Macy et al. | Oct 2013 | A1 |
20130291088 | Shieh et al. | Oct 2013 | A1 |
20130298181 | Smith et al. | Nov 2013 | A1 |
20130298184 | Ermagan et al. | Nov 2013 | A1 |
20130298243 | Kumar et al. | Nov 2013 | A1 |
20130318617 | Chaturvedi et al. | Nov 2013 | A1 |
20130343396 | Yamashita et al. | Dec 2013 | A1 |
20140007181 | Sarin et al. | Jan 2014 | A1 |
20140022894 | Oikawa et al. | Jan 2014 | A1 |
20140033267 | Aciicmez | Jan 2014 | A1 |
20140096229 | Burns et al. | Apr 2014 | A1 |
20140137240 | Smith et al. | May 2014 | A1 |
20140153577 | Janakiraman et al. | Jun 2014 | A1 |
20140157352 | Paek et al. | Jun 2014 | A1 |
20140250524 | Meyers et al. | Sep 2014 | A1 |
20140282027 | Gao et al. | Sep 2014 | A1 |
20140282518 | Banerjee | Sep 2014 | A1 |
20140283030 | Moore et al. | Sep 2014 | A1 |
20140310765 | Stuntebeck et al. | Oct 2014 | A1 |
20140337743 | Branton | Nov 2014 | A1 |
20140344435 | Mortimore, Jr. et al. | Nov 2014 | A1 |
20150047046 | Pavlyushchik | Feb 2015 | A1 |
20150058983 | Zeitlin et al. | Feb 2015 | A1 |
20150082417 | Bhagwat et al. | Mar 2015 | A1 |
20150124606 | Alvarez et al. | May 2015 | A1 |
20150163088 | Anschutz | Jun 2015 | A1 |
20150180894 | Sadovsky et al. | Jun 2015 | A1 |
20150180949 | Maes | Jun 2015 | A1 |
20150229641 | Sun et al. | Aug 2015 | A1 |
20150235229 | Pryor | Aug 2015 | A1 |
20150249676 | Koyanagi et al. | Sep 2015 | A1 |
20150269383 | Lang et al. | Sep 2015 | A1 |
20150295943 | Malachi | Oct 2015 | A1 |
20160028851 | Shieh | Jan 2016 | A1 |
20160072899 | Tung et al. | Mar 2016 | A1 |
20160105370 | Mellor et al. | Apr 2016 | A1 |
20160162179 | Annett | Jun 2016 | A1 |
20160171682 | Abedini et al. | Jun 2016 | A1 |
20160173521 | Yampolskiy et al. | Jun 2016 | A1 |
20160191466 | Pernicha | Jun 2016 | A1 |
20160191545 | Nanda et al. | Jun 2016 | A1 |
20160203331 | Khan et al. | Jul 2016 | A1 |
20160234250 | Ashley et al. | Aug 2016 | A1 |
20160269442 | Shieh | Sep 2016 | A1 |
20160294774 | Woolward et al. | Oct 2016 | A1 |
20160294875 | Lian et al. | Oct 2016 | A1 |
20160301704 | Hassanzadeh et al. | Oct 2016 | A1 |
20160323245 | Shieh et al. | Nov 2016 | A1 |
20160337390 | Sridhara et al. | Nov 2016 | A1 |
20160350105 | Kumar et al. | Dec 2016 | A1 |
20160357424 | Pang et al. | Dec 2016 | A1 |
20160357774 | Gauchi et al. | Dec 2016 | A1 |
20170005986 | Bansal et al. | Jan 2017 | A1 |
20170006135 | Siebel et al. | Jan 2017 | A1 |
20170063795 | Lian et al. | Mar 2017 | A1 |
20170075744 | Deshpande et al. | Mar 2017 | A1 |
20170085654 | Mikhailov et al. | Mar 2017 | A1 |
20170118218 | Koottayi | Apr 2017 | A1 |
20170134422 | Shieh et al. | May 2017 | A1 |
20170168864 | Ross et al. | Jun 2017 | A1 |
20170180421 | Shieh et al. | Jun 2017 | A1 |
20170195454 | Shieh | Jul 2017 | A1 |
20170208100 | Lian et al. | Jul 2017 | A1 |
20170223033 | Wager et al. | Aug 2017 | A1 |
20170223038 | Wager et al. | Aug 2017 | A1 |
20170251013 | Kirti | Aug 2017 | A1 |
20170264619 | Narayanaswamy et al. | Sep 2017 | A1 |
20170264640 | Narayanaswamy et al. | Sep 2017 | A1 |
20170279770 | Woolward | Sep 2017 | A1 |
20170295188 | David et al. | Oct 2017 | A1 |
20170302685 | Ladnai et al. | Oct 2017 | A1 |
20170339188 | Jain et al. | Nov 2017 | A1 |
20170374032 | Woolward et al. | Dec 2017 | A1 |
20170374101 | Woolward | Dec 2017 | A1 |
20180005296 | Eades et al. | Jan 2018 | A1 |
20180083977 | Murugesan | Mar 2018 | A1 |
20180095976 | Shelksohn | Apr 2018 | A1 |
20180113773 | Kirshnan et al. | Apr 2018 | A1 |
20180137506 | Kel et al. | May 2018 | A1 |
20180191779 | Shieh et al. | Jul 2018 | A1 |
20180191813 | Harpaz et al. | Jul 2018 | A1 |
20180225795 | Napoli | Aug 2018 | A1 |
20180232262 | Chowdhury | Aug 2018 | A1 |
20180254892 | Egorov et al. | Sep 2018 | A1 |
20180293701 | Appu et al. | Oct 2018 | A1 |
20180329940 | Tiku et al. | Nov 2018 | A1 |
20180375877 | Jakbsson et al. | Dec 2018 | A1 |
20190014153 | Lang et al. | Jan 2019 | A1 |
20190043534 | Sievert | Feb 2019 | A1 |
20190052549 | Duggal et al. | Feb 2019 | A1 |
20190081963 | Waghorn | Mar 2019 | A1 |
20190109820 | Clark et al. | Apr 2019 | A1 |
20190141075 | Gay | May 2019 | A1 |
20190273746 | Coffing | Sep 2019 | A1 |
20190278760 | Smart | Sep 2019 | A1 |
20190317728 | Chen et al. | Oct 2019 | A1 |
20190342307 | Gamble et al. | Nov 2019 | A1 |
20190394225 | Vajipayajula et al. | Dec 2019 | A1 |
20200043008 | Hrabik | Feb 2020 | A1 |
20200065343 | Morkovine | Feb 2020 | A1 |
20200074078 | Saxe et al. | Mar 2020 | A1 |
20200076826 | Ford | Mar 2020 | A1 |
20200145441 | Patterson et al. | May 2020 | A1 |
20200169565 | Badawy et al. | May 2020 | A1 |
20200259852 | Wolff | Aug 2020 | A1 |
20200329118 | Wang et al. | Oct 2020 | A1 |
20200382363 | Woolward et al. | Dec 2020 | A1 |
20200382556 | Woolward et al. | Dec 2020 | A1 |
20200382557 | Woolward et al. | Dec 2020 | A1 |
20200382560 | Woolward et al. | Dec 2020 | A1 |
20200382586 | Badawy et al. | Dec 2020 | A1 |
20210120029 | Ross et al. | Apr 2021 | A1 |
20210126837 | Dinh et al. | Apr 2021 | A1 |
20220036302 | Cella et al. | Feb 2022 | A1 |
20220201024 | Ross et al. | Jun 2022 | A1 |
20220201025 | Ross et al. | Jun 2022 | A1 |
Number | Date | Country |
---|---|---|
201642616 | Dec 2016 | TW |
201642617 | Dec 2016 | TW |
201642618 | Dec 2016 | TW |
201703483 | Jan 2017 | TW |
201703485 | Jan 2017 | TW |
WO2002098100 | Dec 2002 | WO |
WO2016148865 | Sep 2016 | WO |
WO2016160523 | Oct 2016 | WO |
WO2016160533 | Oct 2016 | WO |
WO2016160595 | Oct 2016 | WO |
WO2016160599 | Oct 2016 | WO |
WO2017100365 | Jun 2017 | WO |
Entry |
---|
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/024116, dated May 3, 2016, 12 pages. |
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/024300, dated May 3, 2016, 9 pages. |
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/024053, dated May 3, 2016, 12 pages. |
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/019643, dated May 6, 2016, 27 pages. |
Dubrawsky, Ido, “Firewall Evolution—Deep Packet Inspection,” Symantec, Created Jul. 28, 2003; Updated Nov. 2, 2010, symantec.com/connect/articles/firewall-evolution-deep-packet-inspection, 3 pages. |
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/024310, dated Jun. 20, 2016, 9 pages. |
“Feature Handbook: NetBrain® Enterprise Edition 6.1” NetBrain Technologies, Inc., Feb. 25, 2016, 48 pages. |
Arendt, Dustin L. et al., “Ocelot: User-Centered Design of a Decision Support Visualization for Network Quarantine”, IEEE Symposium on Visualization for Cyber Security (VIZSEC), Oct. 25, 2015, 8 pages. |
“International Search Report” and “Written Opinion of the International Searching Authority,” Patent Cooperation Treaty Application No. PCT/US2016/065451, dated Jan. 12, 2017, 20 pages. |
Maniar, Neeta, “Centralized Tracking and Risk Analysis of 3rd Party Firewall Connections,” SANS Institute InfoSec Reading Room, Mar. 11, 2005, 20 pages. |
Hu, Hongxin et al., “Detecting and Resolving Firewall Policy Anomalies,” IEEE Transactions on Dependable and Secure Computing, vol. 9, No. 3, May/Jun. 2012, pp. 318-331. |
Woolward et al., “Template-Driven Intent-Based Security,” U.S. Appl. No. 16/428,838, filed May 31, 2019, Specification, Claims, Abstract, and Drawings, 60pages. |
Woolward et al., “Validation of Cloud Security Policies,” U.S. Appl. No. 16/428,849, filed May 31, 2019, Specification, Claims, Abstract, and Drawings, 54 pages. |
Woolward et al., “Reliability Prediction for Cloud Security Policies,” U.S. Appl. No. 16/428,858, filed May 31, 2019, Specification, Claims, Abstract, and Drawings, 59 pages. |
Bates, Adam Macneil, “Designing and Leveraging Trustworthy Provenance-Aware Architectures”, ProQuest Dissertations and Theses ProQuest Dissertations Publishing, 2017, 147 pages. |
Wang et al., “System and Method for Attributing User Behavior from Multiple Technical Telemetry Sources,” U.S. Appl. No. 17/162,761, filed Jan. 29, 2021; Specification, Claims, Abstract, and Drawings, 31 pages. |
Badidi et al., “A Cloud-based Approach for Context Information Provisioning,” World of Computer Science and Information Technology Journal (WCSIT), vol. 1, No. 3, 2011, pp. 63-70. |
Fazio et al., “Message Oriented Middleware for Cloud Computing to Improve Efficiency in Risk Management Systems,” Scalable Computing: Practice and Expirence, vol. 14, No. 4, 2013, pp. 201-213. |
Souto et al., “A Message-Oriented Middleware for Sensor Networks,” 2nd International Workshop on Middleware for Pervasive and Ad-Hoc Computing, Oct. 18-22, 2004, Toronto, Ontario, Canada, pp. 127-134. |
Wun et al., “A Policy Management Framework for Content-Based Publish/Subscribe Middleware,” ACM/IFIP/USENIX International Conference on Distributed Systems Platforms and Open Distributed Processing Middleware 2007, pp. 368-388. |
Wang et al., “Anomaly Detection in the Case of Message Oriented Middleware,” MidSec '08: Proceedings of the 2008 workshop on Middleware security, Dec. 2008, https://doi.org/10.1145/1463342.1463350, pp. 40-42. |
Carranza et al., “Brokering Policies and Execution Monitors for ioT Middleware,” SACMAT '19, Jun. 3-6, 2019, Toronto, ON, Canada, pp. 49-60. |
Albano et al., “Message-oriented Middleware for Smart Grids,” Computer Standards & Interfaces, vol. 38, Feb. 2015, pp. 133-143. |
“Middleware for Communications,” Qusay Mahmoud (Ed.), John Wiiey & Sons Ltd, The Atrium, Southern Gate, Chichester, England, Jun. 29, 2004, Online ISBN:9780470862087 |DOI:10.1002/0470862084, abstract, 1 page. |
Chew, Zhen Bob, “Modelling Message-Oriented-Middleware Brokers Using Autoregressive Models for Bottleneck Prediction,” PhD diss, Queen Mary, University of London, 2013, 179 pages. |
Vinoski, S. “An overview of Middleware,” In: Llamosi, A, Strohmeier, A. (eds), Reliable Software Technologies—Ada-Europe 2004, Ada-Europe 2004. Lecture Notes in Computer Science, vol. 3063, Springer, Berlin, Heidelberg, https://doi.org/10.1007/978-3-540-24841-5_3, pp. 35-51. |
Casola et al., “A Security Monitoring System for Internet of Things,” Internet of Things, vol. 7, Sep. 2019, 20 pages. |
Pierleoni et al., “Amazon, Google and Microsoft Solutions for IoT: Architectures and a Performance Comparison,” in IEEE Access, vol. 8, 2020, doi: 10.1109/ACCESS.2019.2961511, pp. 5455-5470. |
Rusello et al., “A Policy-Based Pulish/Subscribe Middleware for Sense-and-React Applications,” Journal of Systems and Software vol. 84, No. 4, 2011 pp. 638-654. |
Number | Date | Country | |
---|---|---|---|
20210168150 A1 | Jun 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17133466 | Dec 2020 | US |
Child | 17170320 | US | |
Parent | 16428828 | May 2019 | US |
Child | 17133466 | US |