Fine-grained security policy enforcement for applications

Information

  • Patent Grant
  • 12242599
  • Patent Number
    12,242,599
  • Date Filed
    Friday, September 27, 2024
    5 months ago
  • Date Issued
    Tuesday, March 4, 2025
    15 hours ago
  • Inventors
    • Hassey; Philip David (Rye, CO, US)
    • Zeng; Yanran (Pittsburgh, PA, US)
    • Rowe; Brian David (Portland, OR, US)
    • Jamieson; Kevin David
    • Stephen; Patrick David (Minneapolis, MN, US)
    • Fanous; Karim (Seattle, WA, US)
  • Original Assignees
    • strongDM, Inc. (Palo Alto, CA, US)
  • Examiners
    • Keller; Michael A
    Agents
    • Branch; John W.
    • Branch Partners PLLC
Abstract
Embodiments generate state elements based on application requests from a client. The state elements may be enqueued in a state queue associated with an application session for an application requests and the application requests may be forwarded to the application. Application responses from the application may be employed to perform further actions, including: generating message elements based on the application responses such that the message elements may be enqueued in a message queue associated in the application session; determining a portion of the state elements in the state queue that may be associated the message elements; updating the portion of the state elements to advance a protocol state based on the message elements such that the application responses may be communicated to the client.
Description
TECHNICAL FIELD

The present invention relates generally to network security, and more particularly, but not exclusively, to fine-grained security policy enforcement for applications.


BACKGROUND

As organizations become increasingly dependent on networked environments, remote services, distributed services, or the like, managing and monitoring infrastructure access in networked environments can become both critically important and more difficult. Difficulties in managing network environments may not be new, however, interconnections among remote offices, data centers, remote employees, remote customers, and so on, have resulted in organizations relying more broadly on heterogeneous distributed networked services, or the like. Also, in some cases, the regulatory environment has been adapting to the increase in computer-based services. Accordingly, organizations may be required to comply with regulatory regimes from multiple jurisdictions related to various critical subjects, such as finance, privacy, employee rights, cross-jurisdiction taxation, and so on. The combination of the increase in reliance on distributed and networked services and ongoing changes in regulatory environments has tended to elevate the importance of managing and monitoring infrastructure access in networked environments both for operations as well as compliance with various regulatory regimes. Thus, it is with respect to these considerations and others that the present invention has been made.





BRIEF DESCRIPTION OF THE DRAWINGS

Non-limiting and non-exhaustive embodiments of the present innovations are described with reference to the following drawings. In the drawings, like reference numerals refer to like parts throughout the various figures unless otherwise specified. For a better understanding of the described innovations, reference will be made to the following Detailed Description of Various Embodiments, which is to be read in association with the accompanying drawings, wherein:



FIG. 1 illustrates a system environment in which various embodiments may be implemented;



FIG. 2 illustrates a schematic embodiment of a client computer;



FIG. 3 illustrates a schematic embodiment of a network computer;



FIG. 4 illustrates a logical architecture of a system for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 5 illustrates a logical schematic of a system for fine-grained security policy enforcement for applications in secured networks in accordance with one or more of the various embodiments;



FIG. 6 illustrates a logical schematic of a system for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 7 illustrates a logical schematic of a terminal and shortcut 706 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 8 illustrates a logical schematic of a system for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 9 illustrates an overview flowchart of a process for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 10 illustrates a flowchart of a process for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 11 illustrates a flowchart of a process for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments;



FIG. 12 illustrates a flowchart of a process for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments; and



FIG. 13 illustrates a flowchart of a process for employing protocol maps for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments.





DETAILED DESCRIPTION OF VARIOUS EMBODIMENTS

Various embodiments now will be described more fully hereinafter with reference to the accompanying drawings, which form a part hereof, and which show, by way of illustration, specific exemplary embodiments by which the invention may be practiced. The embodiments may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the embodiments to those skilled in the art. Among other things, the various embodiments may be methods, systems, media or devices. Accordingly, the various embodiments may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. The following detailed description is, therefore, not to be taken in a limiting sense.


Throughout the specification and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrase “in one embodiment” as used herein does not necessarily refer to the same embodiment, though it may. Furthermore, the phrase “in another embodiment” as used herein does not necessarily refer to a different embodiment, although it may. Thus, as described below, various embodiments may be readily combined, without departing from the scope or spirit of the invention.


In addition, as used herein, the term “or” is an inclusive “or” operator, and is equivalent to the term “and/or,” unless the context clearly dictates otherwise. The term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of “a,” “an,” and “the” include plural references. The meaning of “in” includes “in” and “on.”


For example, embodiments, the following terms are also used herein according to the corresponding meaning, unless the context clearly dictates otherwise.


As used herein the term, “engine” refers to logic embodied in hardware or software instructions, which can be written in a programming language, such as C, C++, Objective-C, COBOL, Java™, PHP, Perl, Python, R, Julia, JavaScript, Ruby, VBScript, Microsoft.NET™ languages such as C #, or the like. An engine may be compiled into executable programs or written in interpreted programming languages. Software engines may be callable from other engines or from themselves. Engines described herein refer to one or more logical modules that can be merged with other engines or applications or can be divided into sub-engines. The engines can be stored in non-transitory computer-readable medium or computer storage devices and be stored on and executed by one or more general purpose computers, thus creating a special purpose computer configured to provide the engine.


As used herein, the term “session” refers to a semi-permanent interactive packet interchange between two or more communicating endpoints in a networked environment. A session is set up or established at a certain point in time and torn down at a later point in time. An established communication session may involve more than one message in each direction. A session may have stateful communication where at least one of the communicating endpoints saves information about the session history to be able to communicate. A session may also provide stateless communication, where the communication consists of independent requests with responses between the endpoints. An established session is the basic requirement to perform a connection-oriented communication. A session also is the basic step to transmit in connectionless communication modes.


As used herein, the terms “network connection,” and “connection” refer to an interactive packet interchange between two or more communicating endpoints, such as network devices. Connections may be established before application data is transferred, and where a stream of data is delivered in the same or different order than it was sent. The alternative to connection-oriented transmission is connectionless communication. For example, the datagram mode of communication used by the Internet Protocol (IP) and the Universal Datagram Protocol (UDP) may deliver packets out of order, since different packets may be routed independently and could be delivered over different paths. Packets associated with a TCP protocol connection may also be routed independently and could be delivered over different paths. However, for TCP connections the network communication system may provide the packets to application endpoints in the correct order.


As used herein, the term “tuple” refers to a set of values that identify a source and destination of a network communication or network traffic packet, which may, under some circumstances, be a part of a network connection. In one embodiment, a tuple may include a source Internet Protocol (IP) address, a destination IP address, a source port number, a destination port number, virtual LAN segment identifier (VLAN ID), tunnel identifier, routing interface identifier, physical interface identifier, or a protocol identifier. Tuples may be used to identify network flows (e.g., connection flows).


As used herein, the term, “protocol” refers generally to network protocols that may be employed in a network, including data-link layer protocols, transport layer protocols, application layer protocols, or the like. Thus, unless otherwise indicated, innovations described as working with or being associated with a protocol may be applicable to protocols of various OSI layers, or the like, or combination thereof.


As used herein, the term, “application protocol” refers generally to communication protocols that may be employed in a network that enable one or more applications or services and their client applications to communicate in a networked environment. Application protocols may be considered distinct from transport protocols that may be used to ferry application protocol traffic in networks or among processes/services.


As used herein, the term “mesh agent” refers to programs, process, or services that provide a node, link, or hop in a software defined network. Multiple mesh agents may provide secure tunnels between each other to provide a secure overlay network that may be provided in a conventional underlay network. In some cases, one or more mesh agents may be hosted on network computers in a networked environment.


As used herein, the term “ingress agent” refers to a mesh agent that a client application or user gains access to an overlay network. Ingress agents may be considered mesh agents that are on a logical edge of an overlay network. For example, if a client application requires access to an overlay network to access a protected resource, the first mesh agent that the client application communicates to join or access the overlay network may be considered an ingress agent.


As used herein, the term “egress agent” refers to a mesh agent that may directly communicate with a protected resource. Egress agents may be considered mesh agents that are on a logical edge of an overlay network. For example, client requests provided by a client to an ingress agent may be forwarded through one or more mesh agents in an overlay network until they reach an egress agent associated with the target resource.


As used herein, the term “credential information” refers to data or data structures that include credential secrets that enable access to protected resource servers. For example, credential information may include usernames, passwords, pass-phrases, security certificates, or the like.


As used herein, the “activity” refers to one or more application protocols, applications, or the like. For brevity, policy containers may be described as being associated with activities. For example, a policy container may be associated with HTTP traffic while another policy container may be associated with a database application. Accordingly, in some embodiments, policies in a given policy container may be directed to enforcing policies on for a particular activity. Note, one of ordinary skill in the art will appreciate that the definition or declaration of an activity may vary depending on local requirements or local circumstances. Accordingly, in some embodiments, infrastructure security computers may provide user interfaces or configuration information that enable administrators of overlay networks to have broad authority to define or declare activities that may be relevant.


As used herein, the term “policy” refers to data or data structures that include particular instructions or declarations that may be evaluated to determine if a resource may be accessed by a client in an overlay network.


As used herein, the term “request” refers to data that may be directed from a client endpoint in an overlay network to a target resource in the overlay network. Requests may be considered to include data or messages that may initiate or establish transactions, sessions, actions, or the like, with target resources. In the course of a session, the different endpoints of a session may take on the role of clients or servers depending on the direction of communication or behavior of the application. Likewise, in some cases, requests may be considered part of streaming sessions and are not limited to being single requests that synchronously wait for a single response.


As used herein, the term “response” refers to data that may be directed from a target resource to a client endpoint in an overlay network. Responses may be considered to include data or messages that may initiate, establish, or complete transactions, sessions, actions, or the like, based on requests from clients. In the course of a session, the different endpoints of a session may take on the role of clients or servers depending on the direction of communication or behavior of the application. Likewise, in some cases, responses may be considered part of streaming sessions and are not limited to being single responses to a single response.


As used herein, the term, “configuration information” refers to information that may include rule based policies, pattern matching, scripts (e.g., computer readable instructions), or the like, that may be provided from various sources, including, configuration files, databases, user input, built-in defaults, or the like, or combination thereof. In some cases, configuration information may include or reference information stored in other systems or services, such as, configuration management databases, Lightweight Directory Access Protocol (LDAP) servers, name services, public key infrastructure services, or the like.


The following briefly describes embodiments of the invention to provide a basic understanding of some aspects of the invention. This brief description is not intended as an extensive overview. It is not intended to identify key or critical elements, or to delineate or otherwise narrow the scope. Its purpose is merely to present some concepts in a simplified form as a prelude to the more detailed description that is presented later.


Briefly stated, various embodiments are directed to managing interactions with applications in a computing environment. In one or more of the various embodiments, one or more state elements may be generated based on one or more application requests from a client such that the one or more state elements each include information associated with the one or more application requests or an application server that may be a target of the one or more application requests.


In one or more of the various embodiments, the one or more state elements may be enqueued in a state queue to reduce memory consumption that may be associated with an application session for the one or more application requests such that the one or more application requests may be forwarded to the application server.


In one or more of the various embodiments, one or more application responses from the application server may be employed to perform further actions, including: generating one or more message elements based on the one or more application responses such that the one or more message elements may be enqueued in a message queue to reduce memory consumption associated in the application session; determining a portion of the one or more state elements in the state queue that may be associated the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements; updating the portion of the one or more state elements to advance a protocol state based on the one or more message elements that may be associated with the portion of the one or more state elements such that the one or more application responses may be communicated to the client; dequeing the portion of the one or more state elements from the state queue; dequeing the one or more message elements from the message queues; or the like.


In one or more of the various embodiments, a first application request may be employed to perform additional actions, including: initializing the application session based on the client such that the application session may be associated with the client and the application server; generating the state queue and associating it with the application session; generating the message queue and associating it with the application session; or the like. In some embodiments, another application request from another client may be employed to perform other actions, including: initializing another application session that may be associated with the other client and the application server; generating another state queue and another message queue such that the other state queue and the other message queue may be associated with the other application session; or the like.


In one or more of the various embodiments, a protocol map may be determined based on a protocol associated with the protocol state. In some embodiments, a sequence of the one or more application requests and another sequence of the one or more application responses may be validated as conforming to the protocol based on the protocol map. In some embodiments, the contents of the one or more application requests and other contents of the one or more application responses may be validated as conforming to the protocol based on the protocol map. In some embodiments, a failure to validate the one or more application requests or another failure to validate the one or more application responses may be employed to terminate the application session.


In one or more of the various embodiments, generating the one or more state elements may include: determining one or more policy instructions associated with the application session based on an overlay network policy; including the one or more policy instructions in the one or more state elements; modifying the one or more application requests based on the one or more policy instructions; or the like.


In one or more of the various embodiments, generating the one or more message elements may include: determining one or more policy instructions associated with the application session based on an overlay network policy; including the one or more policy instructions in the one or more message elements; modifying the one or more application responses based on the one or more policy instructions; or the like.


In one or more of the various embodiments, determining the portion of the one or more state elements in the state queue that may be associated with the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements may include: determining one or more features of a protocol corresponding to the protocol state that associates an application response with an application request based the protocol such that the one or more features include one or more of a sequence number, a timestamp, the client identity, a network address, or the like; determining the association between the portion of the one or more state elements and the one or more message elements based on the one or more features of the protocol; or the like.


In one or more of the various embodiments, one or more of the state queue or the message queue that exceeds its memory capacity may be determined based on one or more of the application requests or the application responses. In some embodiments, terminating the application session.


Illustrated Operating Environment



FIG. 1 shows components of one embodiment of an environment in which embodiments of the innovations disclosed herein may be practiced. Not all of the components may be required to practice these innovations, and variations in the arrangement and type of the components may be made without departing from the spirit or scope of these innovations. As shown, system 100 of FIG. 1 includes local area networks (LANs)/wide area networks (WANs)-(network) 110, wireless network 108, client computers 102-105, application server computer 116, infrastructure security computer 118, one or more mesh agent computers 120, or the like.


At least one embodiment of client computers 102-105 is described in more detail below in conjunction with FIG. 2. In one embodiment, at least some of client computers 102-105 may operate over one or more wired or wireless networks, such as networks 108, or 110. Generally, client computers 102-105 may include virtually any computer capable of communicating over a network to send and receive information, perform various online activities, offline actions, or the like. In one embodiment, one or more of client computers 102-105 may be configured to operate within a business or other entity to perform a variety of services for the business or other entity. For example, client computers 102-105 may be configured to operate as a web server, firewall, client application, media player, mobile telephone, game console, desktop computer, or the like. However, client computers 102-105 are not constrained to these services and may also be employed, for example, as for end-user computing in other embodiments. It should be recognized that more or less client computers (as shown in FIG. 1) may be included within a system such as described herein, and embodiments are therefore not constrained by the number or type of client computers employed.


Computers that may operate as client computer 102 may include computers that typically connect using a wired or wireless communications medium such as personal computers, multiprocessor systems, microprocessor-based or programmable electronic devices, network PCs, or the like. In some embodiments, client computers 102-105 may include virtually any portable computer capable of connecting to another computer and receiving information such as, laptop computer 103, mobile computer 104, tablet computers 105, or the like. However, portable computers are not so limited and may also include other portable computers such as cellular telephones, display pagers, radio frequency (RF) devices, infrared (IR) devices, Personal Digital Assistants (PDAs), handheld computers, wearable computers, integrated devices combining one or more of the preceding computers, or the like. As such, client computers 102-105 typically range widely in terms of capabilities and features. Moreover, client computers 102-105 may access various computing applications, including a browser, or other web-based application.


A web-enabled client computer may include a browser application that is configured to send requests and receive responses over the web. The browser application may be configured to receive and display graphics, text, multimedia, and the like, employing virtually any web-based language. In one embodiment, the browser application is enabled to employ JavaScript, HyperText Markup Language (HTML), extensible Markup Language (XML), JavaScript Object Notation (JSON), Cascading Style Sheets (CSS), or the like, or combination thereof, to display and send a message. In one embodiment, a user of the client computer may employ the browser application to perform various activities over a network (online). However, another application may also be used to perform various online activities.


Client computers 102-105 also may include at least one other client application that is configured to receive or send content between another computer. The client application may include a capability to send or receive content, or the like. The client application may further provide information that identifies itself, including a type, capability, name, and the like. In one embodiment, client computers 102-105 may uniquely identify themselves through any of a variety of mechanisms, including an Internet Protocol (IP) address, a phone number, Mobile Identification Number (MIN), an electronic serial number (ESN), a client certificate, or other device identifier. Such information may be provided in one or more network packets, or the like, sent between other client computers, application server computer 116, infrastructure security computer 118, mesh agent computers 120, or other computers.


Client computers 102-105 may further be configured to include a client application that enables an end-user to log into an end-user account that may be managed by another computer, such as application server computer 116, infrastructure security computer 118, mesh agent computers 120, or the like. Such an end-user account, in one non-limiting example, may be configured to enable the end-user to manage one or more online activities, including in one non-limiting example, project management, software development, system administration, configuration management, search activities, social networking activities, browse various websites, communicate with other users, or the like. Further, client computers may be arranged to enable users to provide configuration information, policy information, or the like, to infrastructure security computer 118. Also, client computers may be arranged to enable users to display reports, interactive user-interfaces, results provided by infrastructure security computer 118, mesh agent computers 120, or the like. Wireless network 108 is configured to couple client computers 103-105 and its components with network 110. Wireless network 108 may include any of a variety of wireless sub-networks that may further overlay stand-alone ad-hoc networks, and the like, to provide an infrastructure-oriented connection for client computers 103-105. Such sub-networks may include mesh networks, Wireless LAN (WLAN) networks, cellular networks, and the like. In one embodiment, the system may include more than one wireless network.


Wireless network 108 may further include an autonomous system of terminals, gateways, routers, and the like connected by wireless radio links, and the like. These connectors may be configured to move freely and randomly and organize themselves arbitrarily, such that the topology of wireless network 108 may change rapidly.


Wireless network 108 may further employ a plurality of access technologies including 2nd (2G), 3rd (3G), 4th (4G) 5th (5G) generation radio access for cellular systems, WLAN, Wireless Router (WR) mesh, and the like. Access technologies such as 2G, 3G, 4G, 5G, and future access networks may enable wide area coverage for mobile computers, such as client computers 103-105 with various degrees of mobility. In one non-limiting example, wireless network 108 may enable a radio connection through a radio network access such as Global System for Mobile communication (GSM), General Packet Radio Services (GPRS), Enhanced Data GSM Environment (EDGE), code division multiple access (CDMA), time division multiple access (TDMA), Wideband Code Division Multiple Access (WCDMA), High Speed Downlink Packet Access (HSDPA), Long Term Evolution (LTE), and the like. In essence, wireless network 108 may include virtually any wireless communication mechanism by which information may travel between client computers 103-105 and another computer, network, a cloud-based network, a cloud instance, or the like.


Network 110 is configured to couple network computers with other computers, including, application server computer 116, infrastructure security computer 118, mesh agent computers 120, client computers 102-105 through wireless network 108, or the like. Network 110 is enabled to employ any form of computer readable media for communicating information from one electronic device to another. Also, network 110 can include the Internet in addition to local area networks (LANs), wide area networks (WANs), direct connections, such as through a universal serial bus (USB) port, Ethernet port, other forms of computer-readable media, or any combination thereof. On an interconnected set of LANs, including those based on differing architectures and protocols, a router acts as a link between LANs, enabling messages to be sent from one to another. In addition, communication links within LANs typically include twisted wire pair or coaxial cable, while communication links between networks may utilize analog telephone lines, full or fractional dedicated digital lines including T1, T2, T3, and T4, or other carrier mechanisms including, for example, E-carriers, Integrated Services Digital Networks (ISDNs), Digital Subscriber Lines (DSLs), wireless links including satellite links, or other communications links known to those skilled in the art. Moreover, communication links may further employ any of a variety of digital signaling technologies, including without limit, for example, DS-0, DS-1, DS-2, DS-3, DS-4, OC-3, OC-12, OC-48, or the like. Furthermore, remote computers and other related electronic devices could be remotely connected to either LANs or WANs via a modem and temporary telephone link. In one embodiment, network 110 may be configured to transport information using one or more network protocols, such Internet Protocol (IP).


Additionally, communication media typically embodies computer readable instructions, data structures, program modules, or other transport mechanism and includes any information non-transitory delivery media or transitory delivery media. By way of example, communication media includes wired media such as twisted pair, coaxial cable, fiber optics, wave guides, and other wired media and wireless media such as acoustic, RF, infrared, and other wireless media.


One embodiment of application server computer 116, infrastructure security computer 118, and mesh agent computers 120 are described in more detail below in conjunction with FIG. 3. Although FIG. 1 illustrates application server computer 116, infrastructure security computer 118, and mesh agent computers 120 each as a single computer, the innovations or embodiments are not so limited. For example, one or more functions of application server computer 116, infrastructure security computer 118, and mesh agent computers 120, or the like, may be distributed across one or more distinct network computers. Moreover, in one or more embodiments, infrastructure security computer 118 may be implemented using a plurality of network computers. Further, in one or more of the various embodiments, application server computer 116, infrastructure security computer 118, or mesh agents 120 may be implemented using one or more cloud instances in one or more cloud networks. Accordingly, these innovations and embodiments are not to be construed as being limited to a single environment, and other configurations, and other architectures are also envisaged.


Illustrative Client Computer



FIG. 2 shows one embodiment of client computer 200 that may include many more or less components than those shown. Client computer 200 may represent, for example, at least one embodiment of mobile computers or client computers shown in FIG. 1.


Client computer 200 may include processor 202 in communication with memory 204 via bus 228. Client computer 200 may also include power supply 230, network interface 232, audio interface 256, display 250, keypad 252, illuminator 254, video interface 242, input/output interface 238, haptic interface 264, global positioning systems (GPS) receiver 258, open air gesture interface 260, temperature interface 262, camera(s) 240, projector 246, pointing device interface 266, processor-readable stationary storage device 234, and processor-readable removable storage device 236. Client computer 200 may optionally communicate with a base station (not shown), or directly with another computer. And in one embodiment, although not shown, a gyroscope may be employed within client computer 200 for measuring or maintaining an orientation of client computer 200.


Power supply 230 may provide power to client computer 200. A rechargeable or non-rechargeable battery may be used to provide power. The power may also be provided by an external power source, such as an AC adapter or a powered docking cradle that supplements or recharges the battery.


Network interface 232 includes circuitry for coupling client computer 200 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the OSI model for mobile communication (GSM), CDMA, time division multiple access (TDMA), UDP, TCP/IP, SMS, MMS, GPRS, WAP, UWB, WiMax, SIP/RTP, GPRS, EDGE, WCDMA, LTE, UMTS, OFDM, CDMA2000, EV-DO, HSDPA, or any of a variety of other wireless communication protocols. Network interface 232 is sometimes known as a transceiver, transceiving device, or network interface card (NIC).


Audio interface 256 may be arranged to produce and receive audio signals such as the sound of a human voice. For example, audio interface 256 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others or generate an audio acknowledgement for some action. A microphone in audio interface 256 can also be used for input to or control of client computer 200, e.g., using voice recognition, detecting touch based on sound, and the like.


Display 250 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computer. Display 250 may also include a touch interface 244 arranged to receive input from an object such as a stylus or a digit from a human hand, and may use resistive, capacitive, surface acoustic wave (SAW), infrared, radar, or other technologies to sense touch or gestures.


Projector 246 may be a remote handheld projector or an integrated projector that is capable of projecting an image on a remote wall or any other reflective object such as a remote screen.


Video interface 242 may be arranged to capture video images, such as a still photo, a video segment, an infrared video, or the like. For example, video interface 242 may be coupled to a digital video camera, a web-camera, or the like. Video interface 242 may comprise a lens, an image sensor, and other electronics. Image sensors may include a complementary metal-oxide-semiconductor (CMOS) integrated circuit, charge-coupled device (CCD), or any other integrated circuit for sensing light.


Keypad 252 may comprise any input device arranged to receive input from a user. For example, keypad 252 may include a push button numeric dial, or a keyboard. Keypad 252 may also include command buttons that are associated with selecting and sending images.


Illuminator 254 may provide a status indication or provide light. Illuminator 254 may remain active for specific periods of time or in response to event messages. For example, when illuminator 254 is active, it may backlight the buttons on keypad 252 and stay on while the client computer is powered. Also, illuminator 254 may backlight these buttons in various patterns when particular actions are performed, such as dialing another client computer. Illuminator 254 may also cause light sources positioned within a transparent or translucent case of the client computer to illuminate in response to actions.


Further, client computer 200 may also comprise hardware security module (HSM) 268 for providing additional tamper resistant safeguards for generating, storing or using security/cryptographic information such as, keys, digital certificates, passwords, passphrases, two-factor authentication information, or the like. In some embodiments, hardware security module may be employed to support one or more standard public key infrastructures (PKI), and may be employed to generate, manage, or store keys pairs, or the like. In some embodiments, HSM 268 may be a stand-alone computer, in other cases, HSM 268 may be arranged as a hardware card that may be added to a client computer.


Client computer 200 may also comprise input/output interface 238 for communicating with external peripheral devices or other computers such as other client computers and network computers. The peripheral devices may include an audio headset, virtual reality headsets, display screen glasses, remote speaker system, remote speaker and microphone system, and the like. Input/output interface 238 can utilize one or more technologies, such as Universal Serial Bus (USB), Infrared, WiFi, WiMax, Bluetooth™, and the like.


Input/output interface 238 may also include one or more sensors for determining geolocation information (e.g., GPS), monitoring electrical power conditions (e.g., voltage sensors, current sensors, frequency sensors, and so on), monitoring weather (e.g., thermostats, barometers, anemometers, humidity detectors, precipitation scales, or the like), or the like. Sensors may be one or more hardware sensors that collect or measure data that is external to client computer 200.


Haptic interface 264 may be arranged to provide tactile feedback to a user of the client computer. For example, the haptic interface 264 may be employed to vibrate client computer 200 in a particular way when another user of a computer is calling. Temperature interface 262 may be used to provide a temperature measurement input or a temperature changing output to a user of client computer 200. Open air gesture interface 260 may sense physical gestures of a user of client computer 200, for example, by using single or stereo video cameras, radar, a gyroscopic sensor inside a computer held or worn by the user, or the like. Camera 240 may be used to track physical eye movements of a user of client computer 200.


GPS transceiver 258 can determine the physical coordinates of client computer 200 on the surface of the Earth, which typically outputs a location as latitude and longitude values. GPS transceiver 258 can also employ other geo-positioning mechanisms, including, but not limited to, triangulation, assisted GPS (AGPS), Enhanced Observed Time Difference (E-OTD), Cell Identifier (CI), Service Area Identifier (SAI), Enhanced Timing Advance (ETA), Base Station Subsystem (BSS), or the like, to further determine the physical location of client computer 200 on the surface of the Earth. It is understood that under different conditions, GPS transceiver 258 can determine a physical location for client computer 200. In one or more embodiments, however, client computer 200 may, through other components, provide other information that may be employed to determine a physical location of the client computer, including for example, a Media Access Control (MAC) address, IP address, and the like.


Human interface components can be peripheral devices that are physically separate from client computer 200, allowing for remote input or output to client computer 200. For example, information routed as described here through human interface components such as display 250 or keyboard 252 can instead be routed through network interface 232 to appropriate human interface components located remotely. Examples of human interface peripheral components that may be remote include, but are not limited to, audio devices, pointing devices, keypads, displays, cameras, projectors, and the like. These peripheral components may communicate over a Pico Network such as Bluetooth™, Zigbee™ and the like. One non-limiting example of a client computer with such peripheral human interface components is a wearable computer, which might include a remote pico projector along with one or more cameras that remotely communicate with a separately located client computer to sense a user's gestures toward portions of an image projected by the pico projector onto a reflected surface such as a wall or the user's hand.


A client computer may include web browser application 226 that is configured to receive and to send web pages, web-based messages, graphics, text, multimedia, and the like. The client computer's browser application may employ virtually any programming language, including a wireless application protocol messages (WAP), and the like. In one or more embodiment, the browser application is enabled to employ Handheld Device Markup Language (HDML), Wireless Markup Language (WML), WMLScript, JavaScript, Standard Generalized Markup Language (SGML), HyperText Markup Language (HTML), extensible Markup Language (XML), HTML5, and the like.


Memory 204 may include RAM, ROM, or other types of memory. Memory 204 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data. Memory 204 may store BIOS 208 for controlling low-level operation of client computer 200. The memory may also store operating system 206 for controlling the operation of client computer 200. It will be appreciated that this component may include a general-purpose operating system such as a version of UNIX®, or Linux®, or a specialized client computer communication operating system such as Windows Phone™, or the Symbian® operating system. The operating system may include, or interface with a Java virtual machine module that enables control of hardware components or operating system operations via Java application programs.


Memory 204 may further include one or more data storage 210, which can be utilized by client computer 200 to store, among other things, applications 220 or other data. For example, data storage 210 may also be employed to store information that describes various capabilities of client computer 200. The information may then be provided to another device or computer based on any of a variety of methods, including being sent as part of a header during a communication, sent upon request, or the like. Data storage 210 may also be employed to store social networking information including address books, buddy lists, aliases, user profile information, or the like. Data storage 210 may further include program code, data, algorithms, and the like, for use by processors, such as processor 202 to execute and perform actions. In one embodiment, at least some of data storage 210 might also be stored on another component of client computer 200, including, but not limited to, non-transitory processor-readable removable storage device 236, processor-readable stationary storage device 234, or even external to the client computer.


Applications 220 may include computer executable instructions which, when executed by client computer 200, transmit, receive, or otherwise process instructions and data. Applications 220 may include, for example, mesh agent 222, flow engine 224, other client applications 226, web browser 227, or the like. Client computers may be arranged to exchange communications, such as, queries, searches, messages, notification messages, event messages, alerts, log data, API calls, or the like, combination thereof, with application servers. Other examples of application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth.


Additionally, in one or more embodiments (not shown in the figures), client computer 200 may include one or more embedded logic hardware devices instead of CPUs, such as, an Application Specific Integrated Circuit (ASIC), Field Programmable Gate Array (FPGA), Programmable Array Logic (PAL), or the like, or combination thereof. The embedded logic hardware devices may directly execute embedded logic to perform actions. Also, in one or more embodiments (not shown in the figures), client computer 200 may include one or more hardware microcontrollers instead of CPUs. In one or more embodiments, the microcontrollers may directly execute their own embedded logic to perform actions and access their own internal memory and their own external Input and Output Interfaces (e.g., hardware pins or wireless transceivers) to perform actions, such as System On a Chip (SOC), or the like.


Illustrative Network Computer



FIG. 3 shows one embodiment of network computer 300 that may be included in a system implementing at least one of the various embodiments. Network computer 300 may include many more or less components than those shown in FIG. 3. However, the components shown are sufficient to disclose an illustrative embodiment for practicing these innovations. Network computer 300 may represent, for example, one embodiment of at least one of application server computer 116, infrastructure security computer 118, or mesh agent computers 120 in FIG. 1.


As shown in the figure, network computer 300 includes a processor 302 that may be in communication with a memory 304 via a bus 328. In some embodiments, processor 302 may be comprised of one or more hardware processors, or one or more processor cores. In some cases, one or more of the one or more processors may be specialized processors designed to perform one or more specialized actions, such as, those described herein. Network computer 300 also includes a power supply 330, network interface 332, audio interface 356, display 350, keyboard 352, input/output interface 338, processor-readable stationary storage device 334, and processor-readable removable storage device 336. Power supply 330 provides power to network computer 300.


Network interface 332 includes circuitry for coupling network computer 300 to one or more networks, and is constructed for use with one or more communication protocols and technologies including, but not limited to, protocols and technologies that implement any portion of the Open Systems Interconnection model (OSI model), global system for mobile communication (GSM), code division multiple access (CDMA), time division multiple access (TDMA), user datagram protocol (UDP), transmission control protocol/Internet protocol (TCP/IP), Short Message Service (SMS), Multimedia Messaging Service (MMS), general packet radio service (GPRS), WAP, ultra-wide band (UWB), IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMax), Session Initiation Protocol/Real-time Transport Protocol (SIP/RTP), or any of a variety of other wired and wireless communication protocols. Network interface 332 is sometimes known as a transceiver, transceiving device, or network interface card (NIC). Network computer 300 may optionally communicate with a base station (not shown), or directly with another computer.


Audio interface 356 is arranged to produce and receive audio signals such as the sound of a human voice. For example, audio interface 356 may be coupled to a speaker and microphone (not shown) to enable telecommunication with others or generate an audio acknowledgement for some action. A microphone in audio interface 356 can also be used for input to or control of network computer 300, for example, using voice recognition.


Display 350 may be a liquid crystal display (LCD), gas plasma, electronic ink, light emitting diode (LED), Organic LED (OLED) or any other type of light reflective or light transmissive display that can be used with a computer. In some embodiments, display 350 may be a handheld projector or pico projector capable of projecting an image on a wall or other object.


Network computer 300 may also comprise input/output interface 338 for communicating with external devices or computers not shown in FIG. 3. Input/output interface 338 can utilize one or more wired or wireless communication technologies, such as USB™, Firewire™, WiFi, WiMax, Thunderbolt™, Infrared, Bluetooth™, Zigbee™, serial port, parallel port, and the like.


Also, input/output interface 338 may also include one or more sensors for determining geolocation information (e.g., GPS), monitoring electrical power conditions (e.g., voltage sensors, current sensors, frequency sensors, and so on), monitoring weather (e.g., thermostats, barometers, anemometers, humidity detectors, precipitation scales, or the like), or the like. Sensors may be one or more hardware sensors that collect or measure data that is external to network computer 300. Human interface components can be physically separate from network computer 300, allowing for remote input or output to network computer 300. For example, information routed as described here through human interface components such as display 350 or keyboard 352 can instead be routed through the network interface 332 to appropriate human interface components located elsewhere on the network. Human interface components include any component that allows the computer to take input from, or send output to, a human user of a computer. Accordingly, pointing devices such as mice, styluses, track balls, or the like, may communicate through pointing device interface 358 to receive user input.


GPS transceiver 340 can determine the physical coordinates of network computer 300 on the surface of the Earth, which typically outputs a location as latitude and longitude values. GPS transceiver 340 can also employ other geo-positioning mechanisms, including, but not limited to, triangulation, assisted GPS (AGPS), Enhanced Observed Time Difference (E-OTD), Cell Identifier (CI), Service Area Identifier (SAI), Enhanced Timing Advance (ETA), Base Station Subsystem (BSS), or the like, to further determine the physical location of network computer 300 on the surface of the Earth. It is understood that under different conditions, GPS transceiver 340 can determine a physical location for network computer 300. In one or more embodiment, however, network computer 300 may, through other components, provide other information that may be employed to determine a physical location of the client computer, including for example, a Media Access Control (MAC) address, IP address, and the like.


In at least one of the various embodiments, applications, such as, operating system 306, infrastructure security engine 322, mesh agent 324, flow engine 326, web services 329, or the like, may be arranged to employ geo-location information to select one or more localization features, such as, time zones, languages, currencies, calendar formatting, or the like. Also, localization features may be used when interpreting network traffic, application protocols, modifying/localizing client requests, modifying/localizing server responses, user-interfaces, generating reports, monitoring infrastructure access in different regions, or the like. Localization may be employed by one or more internal processes or databases. In at least one of the various embodiments, geo-location information used for selecting localization information may be provided by GPS 340. Also, in some embodiments, geolocation information may include information provided using one or more geolocation protocols over the networks, such as, wireless network 108 or network 111.


Memory 304 may include Random Access Memory (RAM), Read-Only Memory (ROM), or other types of memory. Memory 304 illustrates an example of computer-readable storage media (devices) for storage of information such as computer-readable instructions, data structures, program modules or other data. Memory 304 stores a basic input/output system (BIOS) 308 for controlling low-level operation of network computer 300. The memory also stores an operating system 306 for controlling the operation of network computer 300. It will be appreciated that this component may include a general-purpose operating system such as a version of UNIX®, or Linux®, or a specialized operating system such as Microsoft Corporation's Windows® operating system, or the Apple Corporation's IOS operating system. Operating systems may include, or interface with a Java virtual machine module that enables control of hardware components or operating system operations via Java application programs. Likewise, other runtime environments may be included.


Memory 304 may further include one or more data storage 310, which can be utilized by network computer 300 to store, among other things, applications 320 or other data. For example, data storage 310 may also be employed to store information that describes various capabilities of network computer 300. The information may then be provided to another device or computer based on any of a variety of methods, including being sent as part of a header during a communication, sent upon request, or the like. Data storage 310 may also be employed to store social networking information including address books, buddy lists, aliases, user profile information, or the like. Data storage 310 may further include program code, data, algorithms, and the like, for use by a processor, such as processor 302 to execute and perform actions such as those actions described below. In one embodiment, at least some of data storage 310 might also be stored on another component of network computer 300, including, but not limited to, non-transitory media inside processor-readable removable storage device 336, processor-readable stationary storage device 334, or any other computer-readable storage device within network computer 300, or even external to network computer 300. Data storage 310 may include, for example, overlay network information 314, application protocol information 316, activity logs 318, frame capture logs 319, or the like.


Applications 320 may include computer executable instructions which, when executed by network computer 300, transmit, receive, or otherwise process messages (e.g., SMS, Multimedia Messaging Service (MMS), Instant Message (IM), email, or other messages), audio, video, and enable telecommunication with another user of another mobile computer. Other examples of application programs include calendars, search programs, email client applications, IM applications, SMS applications, Voice Over Internet Protocol (VOIP) applications, contact managers, task managers, transcoders, database programs, word processing programs, security applications, spreadsheet programs, games, search programs, and so forth. Applications 320 may include infrastructure security engine 322, mesh agent 324, flow engine 326, web services 329, or the like, that may be arranged to perform actions for embodiments described below. In one or more of the various embodiments, one or more of the applications may be implemented as modules or components of another application. Further, in one or more of the various embodiments, applications may be implemented as operating system extensions, modules, plugins, or the like.


Furthermore, in one or more of the various embodiments, infrastructure security engine 322, mesh agent 324, flow engine 326, web services 329, or the like, may be operative in a cloud-based computing environment. In one or more of the various embodiments, these applications, and others may be executing within virtual machines or virtual servers that may be managed in a cloud-based computing environment. In one or more of the various embodiments, in this context the applications may flow from one physical network computer within the cloud-based environment to another depending on performance and scaling considerations automatically managed by the cloud computing environment. Likewise, in one or more of the various embodiments, virtual machines or virtual servers dedicated to infrastructure security engine 322, mesh agent 324, flow engine 326, web services 329, or the like, may be provisioned and de-commissioned automatically.


Also, in one or more of the various embodiments, infrastructure security 322, mesh agent 324, flow engine 326, web services 329, or the like, may be located in virtual servers running in a cloud-based computing environment rather than being tied to one or more specific physical network computers. Likewise, in some embodiments, one or more of infrastructure security engine 322, mesh agent 324, flow engine 326, web services 329, or the like, may be configured to execute in a container-based environment.


Further, network computer 300 may also comprise hardware security module (HSM) 360 for providing additional tamper resistant safeguards for generating, storing or using security/cryptographic information such as, keys, digital certificates, passwords, passphrases, two-factor authentication information, or the like. In some embodiments, hardware security modules may be employed to support one or more standard public key infrastructures (PKI), and may be employed to generate, manage, or store keys pairs, or the like. In some embodiments, HSM 360 may be a stand-alone network computer, in other cases, HSM 360 may be arranged as a hardware card that may be installed in a network computer.


Additionally, in one or more embodiments (not shown in the figures), network computer 300 may include one or more embedded logic hardware devices instead of CPUs, such as, an Application Specific Integrated Circuit (ASIC), Field Programmable Gate Array (FPGA), Programmable Array Logic (PAL), or the like, or combination thereof. The embedded logic hardware device may directly execute its embedded logic to perform actions. Also, in one or more embodiments (not shown in the figures), the network computer may include one or more hardware microcontrollers instead of CPUs. In one or more embodiments, the one or more microcontrollers may directly execute their own embedded logic to perform actions and access their own internal memory and their own external Input and Output Interfaces (e.g., hardware pins or wireless transceivers) to perform actions, such as System On a Chip (SOC), or the like.


Illustrative Logical System Architecture



FIG. 4 illustrates a logical architecture of system 400 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. In some embodiments, system 400 may comprise: one or more infrastructure security computers, such as, infrastructure security computer 402; one or more networks, such as, network 404; one or more mesh agents, such as, mesh agents 406-410 or mesh agents 412-416; one or more edge agents, such as, ingress agent 418 or egress agent 420; one or more clients, such as, client 422 or client 424; one or more activity log data stores, such as, activity log 426; or the like.


Note, one or more portions of system 400 illustrates a non-limiting example of a software defined network system or overlay network system. One of ordinary skill in the art will appreciate that other mesh network architectures or overlay network management schemes may be employed without departing from the scope of the innovations disclosed herein. However, the description of system 400 is sufficient for enabling one of ordinary skill in the art to understand the innovations disclosed herein.


In one or more of the various embodiments, mesh agents may be arranged to provide a software-defined overlay network that runs on one or more conventionally configured/provisioned physical (underlay) networks. In some embodiments, if a client application attempts to access resources managed by an infrastructure security service, the mesh agent used by clients to gain access to resources in a managed environment may be considered an ingress agent. Similarly, mesh agents that enable resource servers to respond to client request via the overlay network may be considered to be egress agents. Accordingly, in some cases mesh agents may be considered to be ingress agents in some contexts and they may be considered to be egress agents in other contexts. The different nomenclature is used herein as a convenience to describe features of embodiments in the different contexts.


In one or more of the various embodiments, infrastructure security computers, such as, infrastructure security computer 402, may host one or more infrastructure security engines that enforce policies for a secure overlay network that provides managed access to one or more resources (or endpoints) in a networked environment.


In one or more of the various embodiments, if a client employs a mesh agent, such as, ingress agent 418, the ingress agent may be arranged communicate with infrastructure security computer 402 to authenticate the client request and determine a network path from the client to the target resource server via the overlay network.


In one or more of the various embodiments, mesh agents may be arranged to determine a next ‘hop’ to a next mesh agent. Accordingly, the next mesh agent, as well as, other mesh agents, may determine the next hop to a next mesh agent. Eventually, in some embodiments, the communication or request may reach an egress agent for the target resource server. And, in some embodiments, if each intervening mesh agent validates or authenticates the client communication, the client communication may reach the intended target resource server.


In one or more of the various embodiments, if an ingress agent receives a client request, the ingress agent may send a communication to an infrastructure security computer, such as, infrastructure security computer 402 to determine one or more authorized routes through the overlay network to reach the target resource server. In some embodiments, routes through the overlay network may be considered to be overlay paths or secure tunnels that go from mesh agent-to-mesh agent until the target server may be reached.


In one or more of the various embodiments, mesh agents may establish a cryptographically secure virtual network tunnel between clients and servers such that the client/server traffic may be opaque to observers or other network devices that may be unaffiliated with the infrastructure security computer.


Note, one of ordinary skill in the art will appreciate that system 400 may comprise one or more network devices, network computers, routers, switches, or the like, that comprise the underlay network. For brevity and clarity, the underlay network components are omitted from FIG. 4.


In one or more of the various embodiments, mesh agents may be considered to be hosted on physical or virtual computers that have access to the underlay networks. Also, in some embodiments, the number of overlay network hops (between/via mesh agents) between endpoints may be different than the actual network hops required by the underlay network. For example, for some embodiments, system 400 shows one overlay hop to get from ingress agent 418 to mesh agent 408. However, in some cases, there may be multiple hops in the underlay network (not shown) to exchange network traffic between ingress agent 418 and mesh agent 408. For example, in some embodiments, one hop in the overlay network may traverse one or more sub-networks that may require multiple hops through multiple underlay network routers.


In one or more of the various embodiments, if a client provides a request to communicate with a managed endpoint (e.g., resource server 424), the corresponding ingress agent (e.g., ingress agent 418) forwards information about the request to an infrastructure security computer (e.g., infrastructure security computer 402). Accordingly, in some embodiments, an infrastructure security engine (hosted on the infrastructure security computer) may be arranged to determine if the client user has permission to communicate with the target endpoint. Also, in some embodiments, the infrastructure security engine may be arranged to determine one or more next mesh agents where the ingress agent may forward the client request. Accordingly, in some embodiments, infrastructure security engines may be arranged to generate an overlay route table that includes one or more available mesh agents that may be candidates that may be suitable and authorized for handling the communication.


In some embodiments, the communication may be forwarded to subsequent mesh agents, each intervening mesh agent may be arranged to validate and authenticate the client communication using the infrastructure security engine. In some embodiments, if the client communication may be authorized, the infrastructure security engine may provide an overlay route table that identifies one or more mesh agents for the next hop through the overlay network.


In this example, the overlay path determined for client 422 to communicate with server 424 is ingress client 418 to mesh agent 408 to mesh agent 416 to egress agent 420 and ultimately to server 424. At the final mesh agent (e.g., egress agent 420), the egress agent may determine/obtain the credentials that enable access to the server. In some embodiments, egress agents may be arranged to communicate with an infrastructure security computer to obtain credentials for a server. In this example, the connections (double-arrow lines) illustrated with solid lines represent the determined route through the overlay network. In contrast, the connections (double-arrow lines) illustrated using dashed lines represent mesh agents that may be part of the overlay network that were not selected for a particular communication between client 422 and server 424.


In one or more of the various embodiments, ingress agents, egress agents, or mesh agents may be configured to capture or record activity that may be associated with the communication through the secure tunnel. In this example, for some embodiments, activity log data store 426 represents a data store for storing logged or recorded activity for managed infrastructure. In some embodiments, infrastructure security engines may be arranged to enable different types of activity logging. In some embodiments, infrastructure security engines may be configured to record one or more of the user information associated with an action, occurrence of actions, the accompanying application payload (if any), response from servers, or the like. Further, in some embodiments, infrastructure security engines may enable log information to be forwarded to another data store for storage or archival.


In one or more of the various embodiments, infrastructure security engines may be arranged to generate authenticity tokens that may act as a fingerprint for activity that may occur during the secure tunnel session. In one or more of the various embodiments, authenticity tokens may be generated based on the payload content, user identities, client identities, or the like, that may be associated with an overlay session. For example, if the secure tunnel session includes a response to a query, the authenticity token may be generated based on a hash of the response to the query. Among other things, in some embodiments, authenticity tokens may be employed as part of a scheme to determine the authenticity activity log information that may be stored elsewhere.



FIG. 5 illustrates a logical schematic of system 500 for fine-grained security policy enforcement for applications in secured networks in accordance with one or more of the various embodiments. in this example, for some embodiments, system 500 includes client computer 502, ingress agent 504, client application 506, mesh agent(s) 510, target resource computer 514, egress agent 516, resource server 518, infrastructure security server 520, log data store 522. Note, in some embodiments, ingress agents may be considered mesh agents that a client application employs to communicate via an overlay network. Similarly, in some embodiments, egress agents may be considered mesh agents that communicate with server applications. Note, as mentioned above, ingress agents and egress agents may be considered mesh agents that may provide ingress or egress of traffic in the overlay network. Thus, for brevity and clarity mesh agents may be referred to as ingress agents, egress agents, or mesh agents depending on their immediate role in an overlay communication session.


In this example, for some embodiments, network path 508 may represent communication over a single overlay network hop of a secure tunnel that may comprise one or more underlay network hops to reach the computer that is hosting mesh agent 510. Likewise, network path 512 may represent one or more overlay network hops (intervening mesh agents not shown) that reach target computer 514 and egress agent 516.


As described above, in some embodiments, mesh agents may be arranged to communicate with infrastructure security engines that may be hosted on infrastructure security computers, such as, infrastructure security computer 520. Also, in some embodiments, infrastructure security engines may be arranged to log overlay session activity into activity log data store 522.


In one or more of the various embodiments, client applications, such as, client application 506 may employ one or more application protocols to communicate with resource servers, such as, resource server 518. In some embodiments, application protocols may define particular data in particular arrangements or sequences that enable client applications to communicate with associated/related server applications. In some embodiments, two or more applications or services may share the same application protocol. Also, in some embodiments, one or more applications may employ unique or custom protocols to enable communication between clients or servers. In some embodiments, servers may provide one or more APIs or interfaces that enable applications to access the servers rather than requiring dedicated/proprietary client applications. In such circumstances, a program or service configured to employ such APIs or interfaces may be considered a client application.


In some embodiments, application protocol packets originating from client applications may be provided to ingress agents and subsequently forwarded through the overlay network until they reach the egress agent. Accordingly, in some embodiments, egress agents may be arranged to act as an endpoint to the connection from the client application.


In this example, for some embodiments, application protocol packet 524A represents one or more packets of network traffic that may be sent from a client application, such as client application 506. Accordingly, the application protocol packet may be routed through the overlay network until it reaches its egress agent. In this example, application protocol packet 524B represents the application protocol packet at the egress agent for the client request. And, in this example, for some embodiments, application protocol packet 524 (A, B, C) represents application protocol packets sent from an egress agent to the target resource server.


In one or more of the various embodiments, mesh agents including ingress agents or egress agents may be arranged to re-write application protocol packets to enforce one or more policies in the overlay network. For example, if an application protocol defines one or more fields for holding credential information, a policy may be provided such that egress agents may be arranged to insert the appropriate credential information into the application protocol packets before communicating them to the target resource server. Thus, in some embodiments, credential information provided by the client (if any) may be automatically replaced by other credential information determined by the egress agent as directed by its infrastructure security computer.


Accordingly, in some embodiments, egress agents may be arranged to communicate with infrastructure security computers to obtain credential instructions that declare one or more actions the egress agent may perform to obtain or activate credential information to access the target resource server. In some embodiments, egress agents may be arranged to replace the credential information included in application protocol packet 524B (if any) with credential information determined based on credential instructions provided by its associated infrastructure security computer. Further, in some embodiments, egress agents may be arranged to perform one or more ‘fix-up’ operations on application protocol packet 524B, such as, computing hash signatures based on packet content, adjusting packet size field values, adjusting sequence numbers, adjusting packet offset/pointer values (e.g., values that point to the location of particular values or sections in a given packet), or the like. Accordingly, in some embodiments, application protocol packet 524C represents a modified packet that includes credential information as well as other packet modifications that may be required for a particular application protocol or credential mechanism.



FIG. 6 illustrates a logical schematic of system 600 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. In some embodiments, systems, such as, system 600 may include client computer 602, client application 604, ingress agent 606, policy container 608, policies 610, network computer 612, egress agent 614, policy container 616, policies 618, target resource server 620, or the like. Further, in this example, path 622 represents a network connection between client computer 602 and network computer 612 as facilitated by ingress agent 606 and egress agent 614.


As described above, in one or more of the various embodiments, ingress agents and egress agents may be arranged to establish a secure network path between a client application, such as, client application 604 and target resources, such as, target resource 620.


As described above, mesh agents, including, ingress agents or egress agents may be enabled to inspect or modify network traffic exchanged between client applications and protected target resources. Accordingly, in some embodiments, mesh agents may be arranged to employ one or more policies included in policy containers that may include rules, instructions, parsers, grammars, or the like, directed to one or more application protocols that particular client applications and target resources may employ to communicate.


Herein, for brevity and clarity, network traffic, such as, traffic over connection 622 may be referred to as requests or responses such that clients may send/provide requests and target resources may send/provide responses. One of ordinary skill in the art will appreciate that the requests or responses may be part of ongoing/continuous sessions rather than being limited to strict one-for-one exchanges of requests and responses. For example, policies may be applied to data packets that comprise streaming video, streaming data, streaming audio, or the like.


In one or more of the various embodiments, particular policies for particular application protocols may be included in policy containers, such as, policy container 608 and policy container 616. In some embodiments, policy containers may be arranged to include one or more policies that declare access rules for one or more resources or network segments.


In some embodiments, policy containers may be associated with one or more application protocols, applications, or the like. For brevity, policy containers may be described as being associated with activities. For example, a policy container may be associated with HTTP traffic while another policy container may be associated with a database application. Accordingly, in some embodiments, policies in a given policy container may be directed to enforcing policies on for a particular activity. Note, one of ordinary skill in the art will appreciate that the definition or declaration of an activity may vary depending on local requirements or local circumstances. Accordingly, in some embodiments, infrastructure security computers may provide user interfaces or configuration information that enable administrators of overlay networks to have broad authority to define or declare activities that may be relevant.



FIG. 7 illustrates a logical schematic of terminal 700 and shortcut 706 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. In some embodiments, terminal 700 may represent a computer terminal application that enables users access to command-line interfaces for various applications. In this example, for some embodiments, command line 702 represents a command line that may be provided by a user. In this example, the user is attempting to retrieve data from a remote database using the program dbclient. In the example, the command ‘dbclient’ is shown as prefixed by the command ‘edge’. Accordingly, in this example, the prefix represents a command to employ an ingress agent to execute the dbclient. Accordingly, in this example, rather than routing the dbclient command via the conventional underlay network, the ingress agent may capture the execution of dbclient and employ the overlay network to establish a secure tunnel for the requested operation.


Similarly, in some embodiments, shortcut 706 illustrates how a shortcut for a GUI based operating system or window manager may be configured to route commands initiated from a desktop (e.g., mouse-clicks) through secure tunnels in the overlay network. In this example, shortcut 706 includes an application display name, such as, display name 708 and a launch/execution command represented by target 710. Thus, in this example, shortcuts may be configured to enable launched applications to access the overlay network.


Note, often client applications may support users providing server identifiers (e.g., URIs, IP addresses, hostnames, or the like) that declare the server that for the client application. For example, a conventional command to launch a dbclient application may include the hostname where the database server of interest is located. However, for resources in the overlay network, the infrastructure security engines may determine the location of the server.



FIG. 8 illustrates a logical schematic of system 800 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. In some embodiments, mesh agents or other computer may include systems (or portions thereof), such as system 800. Accordingly, in some embodiments, system 800 may be arranged to include one or more dual queues, such as dual queue 802. In some embodiments, dual queues may be established for each pending application session. Dual queues may include message queues, such as message queue 804 and state queues, such as state queue 806. In some embodiments, flow engines may provide one or more dual queue processors, such as dual queue processor 808.


In one or more of the various embodiments, clients, such as client 810 may be communicatively coupled with system 800 via network connections, such as network connection 812. Accordingly, in some embodiments, message from clients (e.g., client requests) may be received at a protocol handler, such as protocol handler 814. In some embodiments, protocol handlers, such as protocol handler 814 may be arranged to interpret communication/transport protocols to determine the application protocol traffic that may be included in the client messages.


In some embodiments, client messages processors, such as client message processor 816 may be arranged to perform various actions, including: ensuring that client message may be authorized; determining if the client message should be modified before forwarding to the target server; included and stored in a state element that may be stored in the state queue; forwarded to the target server; determining if additional/associated state elements may be added to the state queue; or the like.


In some embodiments, messages received from servers, such as server 824 may be provided to system 800 via network connections such as network connection 822. Also, similar to client messages, in some embodiments, incoming server messages may be provided to a protocol handler, such as protocol handler 820 which may be arranged to interpret the communication/transport protocol to determine the application protocol traffic that may be carried by the transport protocol traffic such that the application protocol traffic may be provided to server message processor 818.


In some embodiments, server message process 818 may be arranged to generate a message queue element that includes the server messages which may be enqueued in message queue 804.


Further, in some embodiments, dual queue processors, such as dual queue processor 808 may be arranged to monitor state queue 806 and message queue 804. Accordingly, in some embodiments, dual queue processor 808 may be arranged to check if state queue 806 may be empty which may enable the server message to be forwarded directly to client 810. Also, in some embodiments, if there may be one or more state elements in state queue 806, dual queue processor 808 may be arranged to read one or more state elements to determine the current state of the pending application session. Accordingly, in some embodiments, dual queue processors may be arranged to take various actions depending on the current state. For example, if the application session may be associated with security policies instructions that require messages from the server to be modified before sending to the client, flow engines may perform such modifications. Also, for example, such security policy instructions may be executed to perform actions, including omitting/discarding messages, sending additional messages to the client, or the like. In some cases, for some embodiments, modification instructions may be provided for by infrastructure security computers, or the like, during the authorization or validation of the client messages that initiate that application session.


Also, in some cases, more than one transport protocol packet may be required for a single application protocol message if the size of an application protocol packet exceeds the payload capacity of the transportation protocol packets. Accordingly, in some embodiments, protocol handlers may be arranged to include parsers, or the like, that may enable application protocol messages to be determined from one or more transport protocol packets or to enable application protocol messages to be included in transport protocol packets for transmission over the network.


In some embodiments, application sessions may be defined based on the target application. Accordingly, in some embodiments, one or more network connections with a client may be associated with the same state queue message queue pair. Likewise, in some embodiments, one or more network connections with the target application server may be associated with the same state queue message queue pair. Likewise, in some cases, for some embodiments, a single network connection from a client and a single connection to the target application server may be associated with a state queue message queue pair.


In some embodiments, flow engines may be arranged to include meta-data associated with the application session in state queue elements or message queue elements. In some embodiments, this meta-data may include one or more instructions or rules that may be associated with the security policies that may be associated with the application session. Also, in some embodiments, meta-data may include timestamps, check-sums, sequence numbers, or the like, that may be employed to improve the management of the communication flow of the application session.


In some embodiments, if other application sessions may be initialized for the same or different application servers, flow engines may be arranged to generate additional state queue and message queue pairs to service the other application sessions (not shown). Also, in some embodiments, depending on the associated application or application protocol, one or more network connections may be associated with the same state queue message queue pair.


As described above, if a client sends a request to an application server in the managed overlay network, mesh agents may be arranged to communicate with its infrastructure security computer to authenticate the user and validate the request. In some embodiments, this may include determining any policies that may be associated with the request, the user, the application, the network, or the like. In some cases, for some embodiments, there may be particular policy rules that may be applied based on the details of the client request or the application server. In some embodiments, if the request is disallowed by the infrastructure security computer outright, the request may be discarded before providing it to the flow engine. Also, in some embodiments, infrastructure security computers may determine the routing information, credential information, logging rules, redaction rules, or the like, based on the policies that may be applied to the request. Accordingly, in some embodiments, it may generally be assumed that requests, users, or the like, may be vetted by an infrastructure security computer, or the like, before the application protocol messages may be provided to flow engines.


As described above, in some embodiments, flow engines may be arranged to generate dual queues each with a state queue and a message queue for each application session. Accordingly, in some embodiments, if a new application session may be initiated, flow engines may be arranged to generate another state queue and another message queue.


In some embodiments, one or more message elements may be associated with application server messages that may be sent in response to application requests from clients that may be associated with state elements. Also, in some embodiments, one or more message elements may be associated with other application protocol messages provided by application servers that may be unassociated with client application requests or state elements, including, warning messages, error messages, status messages, or the like, that may not be sent in response to client requests.


In some embodiments, flow engines may employ data structures that include rules, instructions, state maps, or the like for managing or validating application protocols. For brevity and clarity these data structures may be referred to as protocol maps. In some embodiments, protocol maps may include protocol information used by flow engines to evaluate the sequence or contents of application protocol messages or application protocol requests. Accordingly, in some embodiments, flow engines may be enabled to monitor communication of clients or applications to determine if the communication conforms to the application protocol being used for the application session. Thus, in some embodiments, flow engines may be arranged to track communication that occurs during the application session to identify out of order messages, erroneous messages, or otherwise anomalous messages. For example, in some embodiments, if a comprised client application sends out-of-order or otherwise incorrect or unexpected messages, the flow engines may employ protocol maps to detect them and take appropriate actions. For example, for some embodiments, flow engines may be configured terminate application sessions if malformed application protocol messages may be encountered or observed.


In some embodiments, flow engines may be arranged to be adaptable to various application protocols. Accordingly, in some embodiments, flow engines or infrastructure security computers may include one or more data stores of protocol maps to support a variety of application protocols. Also, in some embodiments, flow engines may be arranged to dynamically load protocol maps from configuration information to support additional application protocols as needed to account for local circumstances or local requirements.


Further, in some embodiments, state queue or message queues may be configured to have memory capacity limits. Accordingly, in some embodiments, state queue memory capacity restrictions may prevent incoming application messages from overloading flow engines (or the computers hosting the flow engines). For example, if a client application may be maliciously (e.g., denial of service attacks) or inadvertently generating an excessive number of application request, the state queue capacity limit may protect the flow engine or the computer hosting the flow engine from being starved of memory. Likewise, in some embodiments, messages queues may be configured to have memory capacity limits. Accordingly, in some embodiments, message queues may protect against application servers that may maliciously or inadvertently flood the network with application responses. Accordingly, in some embodiments, in addition to managing the flow order of application requests, state queues may be arranged to reduce memory consumption associated with the one or more application requests or application sessions. And, in some embodiments, in addition to managing the flow order of application responses, message queues may be arranged to reduce memory consumption associated with the one or more application responses or application sessions. In some embodiments, flow engines may be arrange terminate application sessions associated with state queues or messages queues that may be overloaded by requests or responses that may result in their memory capacity limits being exceeded. In some embodiments, flow engines may generate one or more notifications, log entries, report records, or the like, to indicate that an application session has been terminated because of overloaded state queues or message queues.


Generalized Operations



FIGS. 9-13 represent generalized operations for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. In one or more of the various embodiments, processes 900, 1000, 1100, 1200, and 1300 described in conjunction with FIGS. 9-13 may be implemented by or executed by one or more processors on a single network computer such as network computer 300 of FIG. 3. In other embodiments, these processes, or portions thereof, may be implemented by or executed on a plurality of network computers, such as network computer 300 of FIG. 3. In yet other embodiments, these processes, or portions thereof, may be implemented by or executed on one or more virtualized computers, such as, those in a cloud-based or containerized environment. However, embodiments are not so limited and various combinations of network computers, client computers, or the like may be utilized. Further, in one or more of the various embodiments, the processes described in conjunction with FIGS. 9-13 may be used for fine-grained security policy enforcement for applications in accordance with at least one of the various embodiments or architectures such as those described in conjunction with FIGS. 4-8 Further, in one or more of the various embodiments, some or all of the actions performed by processes 900, 1000, 1100, 1200, and 1300 may be executed in part by infrastructure security engine 322, flow engine 326, one or more mesh agents, such as, mesh agent 324, or the like, running on one or more processors of one or more network computers.



FIG. 9 illustrates an overview flowchart of process 900 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. After a start block, at block 902, in one or more of the various embodiments, one or more mesh agents and one or more infrastructure security engines may be provided. As described above, mesh agents may be deployed in an underlay network where an infrastructure security computer may be arranged to dynamically configure the mesh agents to establish an overlay network.


At block 904, in one or more of the various embodiments, clients may request access to a target resource (e.g., application servers). In some embodiments, client applications may be configured to send requests that may be routed via mesh agents to the target resource. In some embodiments, the mesh agent that may be directly communicated to by the client may be considered to be an ingress agent for the application session because it enables the client to access the overlay network.


At block 906, in one or more of the various embodiments, mesh agents may be arranged to apply one or more policies before the request may be forwarded towards the target resource if allowed by the relevant policies. As described above, infrastructure security computers may be arranged to promulgate policy information by pushing policy information to mesh agents to enforce the policies of the overlay network. In some embodiments, these policies may be enforced irrespective of how the target resource may be configured. For example, if a target resource is configured to support request-type A, a policy enforced in the overlay network may block or restrict request-type A requests. For example, if a database is enabled to allow remote users to drop tables, the overlay network may enforce a policy that disables remote users from dropping tables.


At block 908, in one or more of the various embodiments, mesh agents may be arranged to receive responses from the target resource. Similar to how clients access the overlay network, resources responding to requests may also be configured to respond via the overlay network. Accordingly, in some embodiments, target resources may communicate responses to mesh agents that enable access to a secure tunnel to the client provided by the overlay network. Note, the mesh agent that enables resources to respond to requests may be referred to as egress agents since from the point-of-view of the client, the egress agent is the mesh agent at the end of the secure tunnel that directly communicates with the target resource.


At block 910, in one or more of the various embodiments, mesh agents may be arranged to apply one or more policies and forward responses towards the clients, if allowed by the relevant policies. Responses may be forwarded to a sequence of mesh agents that form a secure tunnel between the ingress agent and the egress agent associated with the tunnel. Each intervening mesh agent as well as the ingress agent and egress agent may enforce one or more policies in the overlay network.


Next, in one or more of the various embodiments, control may be returned to a calling process.



FIG. 10 illustrates a flowchart of process 1000 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. After a start block, at block 1002, in one or more of the various embodiments, mesh agents may be arranged to receive a request from a client or a response from a resource. In some embodiments, infrastructure security computers may provide route information for a secure tunnel in the overlay network between a client and a target resource. In some embodiments, the secure tunnel may be comprised of two or more mesh agents.


At block 1004, in one or more of the various embodiments, mesh agents may be arranged to determine one or more policies that may be associated with the request or response. In some embodiments, infrastructure security computers may be arranged to associate particular activities that may correspond to applications, application protocols, originating endpoints, resources, or the like, with one or more policies. For example, in some embodiments, a particular policy may be associated with outbound HTTP traffic. Likewise, for example, another policy may be associated with network traffic/connections to a particular database server. In some cases, a request or response may be associated with more than one policy container. For example, a request from a using a particular application protocol may be associated with a policy for the application protocol and another policy for the server type (e.g., web-server, database, or the like). For example, an HTTP associated policy may include policies that block or restrict HTTP traffic from particular servers/web-sites. While a policy associated with the underlying application may include polices that may block or restrict requests or responses that include certain SQL expressions.


At block 1006, in one or more of the various embodiments, mesh agents may be arranged to apply the one or more policies to the request/response.


In one or more of the various embodiments, policies may include one or more rules, instructions, or the like, for implementing a particular policy. In some cases, policy rules or instructions may conform to one or more standard or customized policy description languages.


In one or more of the various embodiments, policies may be associated with one or more actions or one or more identities. In some embodiments, actions of interest for a policy may be declared within the policy itself using a policy description language. In some cases, policies may include a map, dictionary, or the like, that may explicitly associate one or more policies with one or more actions or identities.


In some embodiments, actions may be associated with applications. For example, particular SQL expressions may be considered actions that may be directed towards SQL databases.


In some embodiments, identities may represent users, services, application servers, machines, or the like, that the actions may be directed by or performed on behalf of. As described above, infrastructure security computers may be arranged to assign identities to the entities (e.g., users, services, machines, application servers, or the like) that may be enabled to access overlay networks. Accordingly, in some embodiments, identities may be assigned or determined by an infrastructure security computer if a client is requesting access to a protected resource.


At decision block 1008, in one or more of the various embodiments, if the request or response may be determined to be valid, control may flow to block 1012; otherwise, control may flow to block 1010. In some embodiments, mesh agents may execute the one or more policies that may apply to the activity and identity associated with the request/response. In some embodiments, a policy may be configured to block or restrict requests/responses that do not conform to a policy. In some embodiments, if there may be two or more applicable policies, policies that block or restrict traffic may be enforced over other less restrictive policies. For example, in some cases, a first policy may be configured to mask sensitive fields in a response while a second policy may be configured to block the response. Accordingly, in this example, the mesh agent that may be applying the policies may block or restrict the response.


At block 1010, in one or more of the various embodiments, mesh agents may be arranged to block or discard the request or response. As described, in some embodiments, policies may be defined to perform a variety of actions, including validating or rejecting requests or responses. Accordingly, for brevity and clarity the innovations herein are generally described in terms of the policies validating requests/responses. However, one of ordinary skill in the art will appreciate that policies are not limited to validating requests/responses and may perform other actions, such as, data substitution, data masking, sending notifications, raising alerts, collecting metrics, or the like.


At block 1012, in one or more of the various embodiments, mesh agents may be arranged to forward the request or response to the next mesh agent. As described above, in some embodiments, mesh agents may establish a secure tunnel that “hops” from one mesh agent to another mesh agent to connect the ingress agent and the egress agent. Accordingly, in some embodiments, some or all mesh agents that comprise the secure tunnel may similarly enforce one or more policies in the overlay network.


Next, in one or more of the various embodiments, control may be returned to a calling process.



FIG. 11 illustrates a flowchart of process 1100 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. After a start block, at block 1102, in one or more of the various embodiments, application protocol requests from clients may be provided to flow engines. As described above, mesh agents may be arranged to capture one or more communication protocol packets (e.g., transport protocol packets) that have been routed from an ingress agent associated with the client application to the egress agent associated with an application server. In some cases, there may be one or more application protocol messages in each communication protocol packet depending on the application protocol and the communication protocol. Also, in some cases, it may require more than one communication protocol packet to obtain one application protocol message depending on the application protocol and the communication protocol.


In some embodiments, flow engines may be arranged to depend on the overlay network routing services or security services for enforcing overlay network policies that enable the client to communicate with the target resource (e.g., the application server). Accordingly, in some embodiments, communications that violate overlay network policy may be prevented from being provided to flow engines.


At decision block 1104, in some embodiments, if the request is associated with a new application session, control may flow to block 1106; otherwise, control may flow to block 1108.


In some embodiments, some applications or application protocols may support on-going or long-lived sessions. Accordingly, in some cases, for some embodiments, an application session may have been initiated by previously provided application protocol messages.


At block 1106, in some embodiments, flow engines may be arranged to initialize a state queue message queue pair for the new application session.


In some embodiments, flow engines may be arranged to manage a state queue message queue pair for each application session. Accordingly, in some embodiments, since this may be a new applications session, flow engines may be arranged to generate a new state queue and a new message queue for the application session.


At block 1108, in some embodiments, flow engines may be arranged to generate a state element based on the application protocol request.


As described above, in some embodiments, flow engines may be arranged to generate state elements for client provided application protocol messages. In some embodiments, flow engines may be arranged to employ protocol maps to determine the appropriate state names, or other protocol related meta-data for the state element. Also, in some embodiments, flow engines may be arranged to include one or more policy instructions or policy rules determined by the infrastructure security computer in the state element meta-data.


In some cases, for some embodiments, state elements may include some or all of the application protocol message that was sent by the client. In some cases, for some embodiments, flow engines may be arranged to employ the client application protocol message if evaluating responses from the application server. In some embodiments, one or more fields in the application protocol message may be employed to associate state elements with messages from the application servers, or the like. Accordingly, in some embodiments, flow engines may be arranged to selectively store one or more portions of the application message sent by the client in the state element. For example, in some embodiments, flow engines may be configured store one or more key fields of the application protocol message in the state element while discarding the payload after the application message has been sent to the application server.


At block 1110, in some embodiments, flow engines may be arranged to enqueue the state element in the state queue and forward the request to the application server. As described above, in some embodiments, flow engines may be arranged to add the state element to the message queue. Also, in some embodiments, the state element may be activated or ‘opened’ if it is ready to be immediately processed. Accordingly, in some embodiments, if the new state element may be only element in the state queue, the state element may be opened and the associated application protocol message may be forwarded to the application server.


In some cases, in some embodiments, another state element for the same application session may be active which may require flow engines to wait until that state may be completed before processing messages for the new state elements. However, in some cases, for some application protocols or application servers, multiple state elements may be active at the same time.


Next, in one or more of the various embodiments, control may be returned to a calling process.



FIG. 12 illustrates a flowchart of process 1200 for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. After a start block, at block 1202, in one or more of the various embodiments, application protocol messages from application servers may be provided to flow engines. As described above, in some embodiments, flow engines may be arranged to send application protocol messages associated with client requests to application servers. Similarly, in some embodiments, application servers may provide application protocol messages to flow engines.


In most cases, for some embodiments, application protocol messages provided by application server may be in response to client requests. For example, in some embodiments, a client may employ an application protocol message to send a query to an application server, while the application server may employ one or more application protocol messages to respond to the query from the client.


Also, in some cases, for some embodiments, application servers may send one or more application protocol messages that may or may not be responses to client requests. For example, in some cases, an application server may send status messages, error messages, warning messages, or the like, that may not be directly associated with a client request.


At block 1204, in some embodiments, flow engines may be arranged to generate message element.


As described above, message elements may be data structures that may be somewhat similar to state elements. Accordingly, in some embodiments, flow engines may be arranged to collect meta-data associated with the message, such as time stamps, source network address information, or the like. In some embodiments, flow engines may be arranged to determine the particular meta-data based on protocol maps or information/instructions provided by the infrastructure security computer.


Also, in some embodiments, flow engines may be arranged to include the application protocol message in message elements.


At block 1206, in some embodiments, flow engines may be arranged to enqueue message elements in the message queue associated with the application session.


At block 1208, in some embodiments, flow engines may be arranged to evaluate state elements in view of new message elements. In some embodiments, as message elements may be enqueued in message queues, flow engines may check if there may be relevant state elements in the state queue. In some cases, a state element may indicate that a state is open or active. In other cases, none of the enqueued state elements may be active or open. Also, in some cases, for some embodiments, the state queue may be empty of state elements.


Accordingly, in some embodiments, flow engines may be arranged to compare the message element or the message from the application server with state elements associated with active or open states to determine if the incoming message may be associated with the open states.


Likewise, in some embodiments, flow engines may be arranged to compare the message element/message from the application service with state elements that may be inactive.


Additionally, in some embodiments, flow engines may be arranged to validate that the message from the application server conforms to the current application protocol. In general, flow engines may employ protocol maps to anticipate messages (or message types) from the application server. Likewise, in some embodiments, protocol maps may include instructions or rules that indicate particular fields in the state elements or message elements that may be compared, including fields included in the application protocol request/response messages that may be compared. For example, if an application protocol uses sequence numbers, message digest hashes, security keys, user identifiers, or the like, the protocol maps may include rules to ensure that these values may be appropriate for a given session. For example, for some embodiments, if a compromised server application sends an out-of-order or non-sequitur message, the flow engine may detect it based on a comparison with the protocol map.


At decision block 1210, in some embodiments, if there may be a state element in the state queue that is associated with the new message element, control may transfer to block 1212; otherwise, control may transfer to block 1216.


At block 1212, in some embodiments, flow engines may be arranged to update the one or more state elements associated with the message element.


In some cases, a state may have one or more conditions, such as NEW, OPEN, ACTIVE, DELAYED, CANCELED, or the like. Also, in some embodiments, states may be labeled based on the application protocol being used. For example, for some embodiments, a database application protocol may include various response messages labeled ‘Command Complete’, ‘Ready For Query’, ‘Data Row’, ‘Warning’, ‘Error’, or the like. Accordingly, for this example, in some embodiments, a protocol map may declare state element labels or conditions that correspond to message types used by the database application.


In some embodiments, if a state element may be matched to a application server message, the protocol map may dictate that the flow engine may remove that state element from the state queue. Also, in some embodiments, for some protocols, the protocol map may dictate that the flow engine may advance a state label or state indicator of matched state element rather than removing it from the state queue. For example, for some embodiments, if the protocol map indicates that two or more messages from the application server may be required or expected for a particular client request message, the flow engine may wait for the require application messages to be received before advancing a state or removing a state element from the state queue.


At block 1214, in some embodiments, flow engines may be arranged to apply relevant policy instructions to the application message content.


In some embodiments, overlay network policies may impose requirements or restrictions on the application protocol message. For example, in some cases, an overlay network policy may restrict particular values from being provided to the client. Accordingly, in some embodiments, the matched state element may include or reference additional rules or instructions that may be executed by flow engines to modify the application protocol message before it may be sent to the client. For example, a policy may restrict fields that include personally identifiable information (PII) from being sent to some users/clients. Accordingly, in this example, the flow engine may mask one or more fields as indicated in the additional instructions included in the matched state element.


At block 1216, in one or more of the various embodiments, optionally, flow engines may be arranged to send the application message to the client.


In some embodiments, if there may be a matched state element, the application protocol message from the application server may be sent to the client as per the controlling overlay network policy.


Also, in some embodiments, if the message may be unmatched to a state element, the application message may be sent to the client. Accordingly, in some embodiments, spontaneous messages from the server may be forwarded absent an associated state or state elements. As described above, typically, these may be status messages, client pings/keep-alive-checks, error conditions, reports of exceptions or crashes, or the like.


However, in some embodiments, protocol map rules or overlay network policy rules may indicate that one or more of these types of the message be discarded or otherwise not sent to the client. For example, in some embodiments, some spontaneous server messages may be blocked or discarded for performance/convenience reasons rather than being automatically forwarded to clients.


Also, in some embodiments, one or more spontaneous server message may include information that may be deemed to be a security risk or security leak. For example, if the spontaneous server message includes network infrastructure information (e.g., network address), sensitive geographic information, sensitive timing/timestamp information, application version information, or the like, protocol maps or overlay network policies may be configured to discard those messages or mask the sensitive information.


In some embodiments, if an application protocol message may be sent to the client, the corresponding message element (if any) may be removed from the message queue.


Note, this block is indicated as being optional, because in some cases application protocol messages from application servers may be suppressed or discarded rather than being sent to clients.


Next, in one or more of the various embodiments, control may be returned to a calling process.



FIG. 13 illustrates a flowchart of process 1300 for employing protocol maps for fine-grained security policy enforcement for applications in accordance with one or more of the various embodiments. After a start block, at decision block 1302, in one or more of the various embodiments, if an application protocol may be provided to a flow engine, control may transfer to block 1304; otherwise, control may loop back to decision block 1302. As described above, flow engines may be provided application protocol messages that may originate from client applications or application servers.


At block 1304, in one or more of the various embodiments, optionally, flow engines may be arranged to determine a protocol map associated with the application protocol message. As described above, in some embodiments, protocol maps may be configured for various application protocols that may be supported by flow engines.


In some embodiments, flow engines or infrastructure security computers may be arranged to associate one or more particular protocol maps with one or more application protocols. In some embodiments, the authorization process of the infrastructure security computer may include providing the protocol map (or reference thereto) to the flow engine.


Also, in some embodiments, flow engines may be arranged to determine the protocol map based on the application session information which may include application name, application version, or the like. In some cases, for some embodiments, overlay network policies may pin a particular application protocol version to a particular session to restrict dynamic application protocol version downgrades, or the like, that may be requested by clients or application servers.


Also, in some embodiments, overlay network policies may associate particular protocol maps to particular users, applications, networks, or the like. Accordingly, in some embodiments, customized protocol maps may be assigned to particular users, user roles, locations, applications, or the like.


Note, this block is indicated as being optional because in some cases, flow engines may have previously determined the protocol map based on previously application protocol messages. Also, in some embodiments, infrastructure security computer may indicate the appropriate protocol map as part of authorizing the communication.


At block 1306, in one or more of the various embodiments, flow engines may be arranged to evaluate the application protocol message in view of the determined protocol map. In some embodiments, protocol maps may be logically considered to be a state machine that may be used to monitor the state of the application session. Accordingly, in some embodiments, out of order messages may be identified.


Further, in some embodiments, protocol maps may include rules, grammars, or the like, that may be used to validate the provided application protocol messages. Accordingly, in some embodiments, various characteristics of the application protocol messages may be verified to include the correct fields, formats, or the like, for the application protocol being used for the application session.


Accordingly, in some embodiments, flow engines may be arranged to employ protocol maps to evaluate various characteristics of the application protocol message, including, order received, fields, field values, length, session identifiers, sequence numbers, or the like.


In some embodiments, flow engines may be arranged to be adaptable to various standard or custom application protocols. Accordingly, in some embodiments, flow engines may be arranged to load protocol maps from configuration information to account for local requirements or local circumstances. In some embodiments, that absence of a suitable protocol maps may prevent the initiation of an application session.


At decision block 1308, in one or more of the various embodiments, if the application protocol message may be validated, control may transfer to block 1312; otherwise, control may transfer to block 1310.


At block 1310, in one or more of the various embodiments, flow engines may be arranged to execute one or more incident responses actions. In some embodiments, flow engines may be arranged to take various actions in response to invalid application protocol messages depending on the overlay network policy. In some cases, for some embodiments, various actions may include one or more of canceling the application session, dropping the invalid messages, raising notifications, generating log entries, alerting security teams, or the like, or combination thereof.


Next, in one or more of the various embodiments, control may be returned to a calling process.


At block 1312, in one or more of the various embodiments, flow engines may be arranged to generate a state element or message element based on the application protocol message.


As described above, in some embodiments, flow engines may be arranged to generate state elements for application protocol messages coming from clients or generate message elements for application protocol messages coming from application servers. Accordingly, in some embodiments, flow engines may be arranged to enqueue the state element in the state queue for the application session or enqueue the message element in the message queue for the application session.


Next, in one or more of the various embodiments, control may be returned to a calling process.


It will be understood that each block of the flowchart illustrations or combinations of blocks in these flowchart illustrations, may be implemented by computer program instructions. These program instructions may be provided to a processor to produce a machine, such that the instructions, which execute on the processor, create means for implementing the actions specified in the flowchart block or blocks. The computer program instructions may be executed by a processor to cause a series of operational steps to be performed by the processor to produce a computer-implemented process such that the instructions, which execute on the processor to provide steps for implementing the actions specified in the flowchart block or blocks. The computer program instructions may also cause at least some of the operational steps shown in the blocks of the flowchart to be performed in parallel. Moreover, some of the steps may also be performed across more than one processor, such as might arise in a multi-processor computer system. In addition, one or more blocks or combinations of blocks in the flowchart illustration may also be performed concurrently with other blocks or combinations of blocks, or even in a different sequence than illustrated without departing from the scope or spirit of the invention.


Accordingly, blocks of the flowchart illustration support combinations of means for performing the specified actions, combinations of steps for performing the specified actions and program instruction means for performing the specified actions. It will also be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by special purpose hardware based systems, which perform the specified actions or steps, or combinations of special purpose hardware and computer instructions. The foregoing example should not be construed as limiting or exhaustive, but rather, an illustrative use case to show an implementation of at least one of the various embodiments of the invention.


Further, in one or more embodiments (not shown in the figures), the logic in the illustrative flowcharts may be executed using an embedded logic hardware device instead of a CPU, such as, an Application Specific Integrated Circuit (ASIC), Field Programmable Gate Array (FPGA), Programmable Array Logic (PAL), or the like, or combination thereof. The embedded logic hardware device may directly execute its embedded logic to perform actions. In one or more embodiments, a microcontroller may be arranged to directly execute its own embedded logic to perform actions and access its own internal memory and its own external Input and Output Interfaces (e.g., hardware pins or wireless transceivers) to perform actions, such as System On a Chip (SOC), or the like.

Claims
  • 1. A method for managing interactions with applications in a computing environment using one or more processors that are configured to execute instructions that cause performance of actions, comprising: generating one or more state elements based on one or more application requests from a client, wherein the one or more state elements each include information associated with the one or more application requests or an application server that is a target of the one or more application requests;enqueuing the one or more state elements in a state queue to reduce memory consumption that is associated with an application session for the one or more application requests, wherein the one or more application requests are forwarded to the application server; andemploying one or more application responses from the application server to perform further actions, including: generating one or more message elements based on the one or more application responses, wherein the one or more message elements are enqueued in a message queue to reduce memory consumption associated in the application session;determining a portion of the one or more state elements in the state queue that are associated the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements;updating the portion of the one or more state elements to advance a protocol state based on the one or more message elements that are associated with the portion of the one or more state elements, wherein the one or more application responses are communicated to the client;dequeing the portion of the one or more state elements from the state queue; anddequeing the one or more message elements from the message queues.
  • 2. The method of claim 1, further comprising: employing a first application request to perform additional further actions, including: initializing the application session based on the client, wherein the application session is associated with the client and the application server;generating the state queue and associating it with the application session;generating the message queue and associating it with the application session; andemploying another application request from another client to perform other actions, including: initializing another application session that is associated with the other client and the application server; andgenerating another state queue and another message queue, wherein the other state queue and the other message queue are associated with the other application session.
  • 3. The method of claim 1, further comprising: determining a protocol map based on a protocol associated with the protocol state;validating a sequence of the one or more application requests and another sequence of the one or more application responses conform to the protocol based on the protocol map;validating that contents of the one or more application requests and other contents of the one or more application responses conforms to the protocol based on the protocol map; andemploying a failure to validate the one or more application requests or another failure to validate the one or more application responses to terminate the application session.
  • 4. The method of claim 1, wherein generating the one or more state elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more state elements; andmodifying the one or more application requests based on the one or more policy instructions.
  • 5. The method of claim 1, wherein generating the one or more message elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more message elements; andmodifying the one or more application responses based on the one or more policy instructions.
  • 6. The method of claim 1, wherein determining the portion of the one or more state elements in the state queue that are associated with the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements, further comprises: determining one or more features of a protocol for the protocol state that associates an application response with an application request based the protocol, wherein the one or more features include one or more of a sequence number, a timestamp, the client identity, or a network address; anddetermining the association between the portion of the one or more state elements and the one or more message elements based on the one or more features of the protocol.
  • 7. The method of claim 1, further comprising: determining one or more of the state queue or the message queue that exceeds its memory capacity based on one or more of the application requests or the application responses; andterminating the application session.
  • 8. A network computer for managing interactions with applications, comprising: a memory that stores at least instructions; andone or more processors that execute instructions that are configured to cause actions, including: generating one or more state elements based on one or more application requests from a client, wherein the one or more state elements each include information associated with the one or more application requests or an application server that is a target of the one or more application requests;enqueuing the one or more state elements in a state queue to reduce memory consumption that is associated with an application session for the one or more application requests, wherein the one or more application requests are forwarded to the application server; andemploying one or more application responses from the application server to perform further actions, including:generating one or more message elements based on the one or more application responses, wherein the one or more message elements are enqueued in a message queue to reduce memory consumption associated in the application session;determining a portion of the one or more state elements in the state queue that are associated the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements;updating the portion of the one or more state elements to advance a protocol state based on the one or more message elements that are associated with the portion of the one or more state elements, wherein the one or more application responses are communicated to the client;dequeing the portion of the one or more state elements from the state queue; anddequeing the one or more message elements from the message queues.
  • 9. The network computer of claim 8, wherein the one or more processors execute instructions that are configured to cause actions, further comprising: employing a first application request to perform additional further actions, including: initializing the application session based on the client, wherein the application session is associated with the client and the application server;generating the state queue and associating it with the application session;generating the message queue and associating it with the application session; andemploying another application request from another client to perform other actions, including: initializing another application session that is associated with the other client and the application server; andgenerating another state queue and another message queue, wherein the other state queue and the other message queue are associated with the other application session.
  • 10. The network computer of claim 8 wherein the one or more processors execute instructions that are configured to cause actions, further comprising: determining a protocol map based on a protocol associated with the protocol state;validating a sequence of the one or more application requests and another sequence of the one or more application responses conform to the protocol based on the protocol map;validating that contents of the one or more application requests and other contents of the one or more application responses conforms to the protocol based on the protocol map; andemploying a failure to validate the one or more application requests or another failure to validate the one or more application responses to terminate the application session.
  • 11. The network computer of claim 8, wherein generating the one or more state elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more state elements; andmodifying the one or more application requests based on the one or more policy instructions.
  • 12. The network computer of claim 8, wherein generating the one or more message elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more message elements; andmodifying the one or more application responses based on the one or more policy instructions.
  • 13. The network computer of claim 8, wherein determining the portion of the one or more state elements in the state queue that are associated with the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements, further comprises: determining one or more features of a protocol for the protocol state that associates an application response with an application request based the protocol, wherein the one or more features include one or more of a sequence number, a timestamp, the client identity, or a network address; anddetermining the association between the portion of the one or more state elements and the one or more message elements based on the one or more features of the protocol.
  • 14. The network computer of claim 8, wherein the one or more processors execute instructions that are configured to cause actions, further comprising: determining one or more of the state queue or the message queue that exceeds its memory capacity based on one or more of the application requests or the application responses; andterminating the application session.
  • 15. A processor readable non-transitory storage media that includes instructions configured for managing interactions with applications in a computing environment, wherein execution of the instructions by one or more processors on one or more network computers performs actions, comprising: generating one or more state elements based on one or more application requests from a client, wherein the one or more state elements each include information associated with the one or more application requests or an application server that is a target of the one or more application requests;enqueuing the one or more state elements in a state queue to reduce memory consumption that is associated with an application session for the one or more application requests, wherein the one or more application requests are forwarded to the application server; andemploying one or more application responses from the application server to perform further actions, including: generating one or more message elements based on the one or more application responses, wherein the one or more message elements are enqueued in a message queue to reduce memory consumption associated in the application session;determining a portion of the one or more state elements in the state queue that are associated the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements;updating the portion of the one or more state elements to advance a protocol state based on the one or more message elements that are associated with the portion of the one or more state elements, wherein the one or more application responses are communicated to the client;dequeing the portion of the one or more state elements from the state queue; anddequeing the one or more message elements from the message queues.
  • 16. The media of claim 15, further comprising: employing a first application request to perform additional further actions, including: initializing the application session based on the client, wherein the application session is associated with the client and the application server;generating the state queue and associating it with the application session;generating the message queue and associating it with the application session; andemploying another application request from another client to perform other actions, including: initializing another application session that is associated with the other client and the application server; andgenerating another state queue and another message queue, wherein the other state queue and the other message queue are associated with the other application session.
  • 17. The media of claim 15, further comprising: determining a protocol map based on a protocol associated with the protocol state;validating a sequence of the one or more application requests and another sequence of the one or more application responses conform to the protocol based on the protocol map;validating that contents of the one or more application requests and other contents of the one or more application responses conforms to the protocol based on the protocol map; andemploying a failure to validate the one or more application requests or another failure to validate the one or more application responses to terminate the application session.
  • 18. The media of claim 15, wherein generating the one or more state elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more state elements; andmodifying the one or more application requests based on the one or more policy instructions.
  • 19. The media of claim 15, wherein generating the one or more message elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more message elements; andmodifying the one or more application responses based on the one or more policy instructions.
  • 20. The media of claim 15, wherein determining the portion of the one or more state elements in the state queue that are associated with the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements, further comprises: determining one or more features of a protocol for the protocol state that associates an application response with an application request based the protocol, wherein the one or more features include one or more of a sequence number, a timestamp, the client identity, or a network address; anddetermining the association between the portion of the one or more state elements and the one or more message elements based on the one or more features of the protocol.
  • 21. The media of claim 15, further comprising: determining one or more of the state queue or the message queue that exceeds its memory capacity based on one or more of the application requests or the application responses; andterminating the application session.
  • 22. A system for method for managing interactions with applications, comprising: a network computer, comprising: a memory that stores at least instructions; andone or more processors that execute instructions that are configured to cause actions, including: generating one or more state elements based on one or more application requests from a client, wherein the one or more state elements each include information associated with the one or more application requests or an application server that is a target of the one or more application requests;enqueuing the one or more state elements in a state queue to reduce memory consumption that is associated with an application session for the one or more application requests, wherein the one or more application requests are forwarded to the application server; andemploying one or more application responses from the application server to perform further actions, including: generating one or more message elements based on the one or more application responses, wherein the one or more message elements are enqueued in a message queue to reduce memory consumption associated in the application session;determining a portion of the one or more state elements in the state queue that are associated the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements;updating the portion of the one or more state elements to advance a protocol state based on the one or more message elements that are associated with the portion of the one or more state elements, wherein the one or more application responses are communicated to the client;dequeing the portion of the one or more state elements from the state queue; anddequeing the one or more message elements from the message queues; anda client computer, comprising: another memory that stores at least instructions; andone or more other processors that execute other instructions that are configured to cause actions, including: generating the one or more application requests.
  • 23. The system of claim 22, wherein the one or more processors of the network computer execute instructions that are configured to cause actions, further comprising: employing a first application request to perform additional further actions, including: initializing the application session based on the client, wherein the application session is associated with the client and the application server;generating the state queue and associating it with the application session;generating the message queue and associating it with the application session; andemploying another application request from another client to perform other actions, including: initializing another application session that is associated with the other client and the application server; andgenerating another state queue and another message queue, wherein the other state queue and the other message queue are associated with the other application session.
  • 24. The system of claim 22, wherein the one or more processors of the network computer execute instructions that are configured to cause actions, further comprising: determining a protocol map based on a protocol associated with the protocol state;validating a sequence of the one or more application requests and another sequence of the one or more application responses conform to the protocol based on the protocol map;validating that contents of the one or more application requests and other contents of the one or more application responses conforms to the protocol based on the protocol map; andemploying a failure to validate the one or more application requests or another failure to validate the one or more application responses to terminate the application session.
  • 25. The system of claim 22, wherein generating the one or more state elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more state elements; andmodifying the one or more application requests based on the one or more policy instructions.
  • 26. The system of claim 22, wherein generating the one or more message elements further comprises: determining one or more policy instructions associated with the application session based on an overlay network policy;including the one or more policy instructions in the one or more message elements; andmodifying the one or more application responses based on the one or more policy instructions.
  • 27. The system of claim 22, wherein determining the portion of the one or more state elements in the state queue that are associated with the one or more message elements based on an association between the portion of the one or more state elements and the one or more message elements, further comprises: determining one or more features of a protocol for the protocol state that associates an application response with an application request based the protocol, wherein the one or more features include one or more of a sequence number, a timestamp, the client identity, or a network address; anddetermining the association between the portion of the one or more state elements and the one or more message elements based on the one or more features of the protocol.
  • 28. The system of claim 22, wherein the one or more processors of the network computer execute instructions that are configured to cause actions, further comprising: determining one or more of the state queue or the message queue that exceeds its memory capacity based on one or more of the application requests or the application responses; andterminating the application session.
US Referenced Citations (610)
Number Name Date Kind
5884794 Calhoun et al. Mar 1999 A
5987247 Lau Nov 1999 A
6345386 Delo et al. Feb 2002 B1
6418447 Frey et al. Jul 2002 B1
6418554 Delo et al. Jul 2002 B1
6442564 Frey et al. Aug 2002 B1
6463470 Mohaban et al. Oct 2002 B1
6466932 Dennis et al. Oct 2002 B1
6502103 Frey et al. Dec 2002 B1
6505210 Frey et al. Jan 2003 B1
6523166 Mishra et al. Feb 2003 B1
6553384 Frey et al. Apr 2003 B1
6560609 Frey et al. May 2003 B1
6567818 Frey et al. May 2003 B1
6594671 Aman et al. Jul 2003 B1
6836794 Lucovsky et al. Dec 2004 B1
7174361 Paas Feb 2007 B1
7233569 Swallow Jun 2007 B1
7752466 Ginter et al. Jul 2010 B2
7752487 Feeser et al. Jul 2010 B1
7797406 Patel Sep 2010 B2
7886339 Keohane et al. Feb 2011 B2
8102814 Rahman et al. Jan 2012 B2
8135815 Mayer Mar 2012 B2
8165905 Yamamoto Apr 2012 B2
9002805 Barber et al. Apr 2015 B1
9052942 Barber et al. Jun 2015 B1
9063946 Barber et al. Jun 2015 B1
9082402 Yadgar et al. Jul 2015 B2
9176758 Swaminathan Nov 2015 B2
9178793 Marlow Nov 2015 B1
9185082 Dashora Nov 2015 B2
9239834 Donabedian et al. Jan 2016 B2
9300635 Gilde et al. Mar 2016 B1
9355060 Barber et al. May 2016 B1
9380456 Lee et al. Jun 2016 B1
9380523 Mijar et al. Jun 2016 B1
9380562 Vetter et al. Jun 2016 B1
9417917 Barber et al. Aug 2016 B1
9471775 Wagner et al. Oct 2016 B1
9521115 Woolward Dec 2016 B1
9560081 Woolward Jan 2017 B1
9584523 Santhiveeran Feb 2017 B2
9619673 Vetter et al. Apr 2017 B1
9632828 Mehta et al. Apr 2017 B1
9658983 Barber et al. May 2017 B1
9667703 Vetter et al. May 2017 B1
9697629 Vetter et al. Jul 2017 B1
9727522 Barber et al. Aug 2017 B1
9787639 Sun et al. Oct 2017 B1
9800517 Anderson Oct 2017 B1
9819593 Vetter et al. Nov 2017 B1
10063595 Qureshi et al. Aug 2018 B1
10075334 Kozura et al. Sep 2018 B1
10089476 Roth et al. Oct 2018 B1
10104185 Sharifi et al. Oct 2018 B1
10110593 Karroumi et al. Oct 2018 B2
10117098 Naguthanawala et al. Oct 2018 B1
10154065 Buchler et al. Dec 2018 B1
10292033 Beyer et al. May 2019 B2
10360010 Maehler et al. Jul 2019 B1
10395042 Agarwal et al. Aug 2019 B2
10409582 Maehler et al. Sep 2019 B1
10476745 McCormick et al. Nov 2019 B1
10516667 Roth et al. Dec 2019 B1
10579362 Maehler et al. Mar 2020 B1
10579403 Antony et al. Mar 2020 B2
10609041 Wilczynski et al. Mar 2020 B1
10645562 Beyer et al. May 2020 B2
10659533 Zhao et al. May 2020 B1
10735263 Mcalary et al. Aug 2020 B1
10827020 Cao et al. Nov 2020 B1
10944691 Raut et al. Mar 2021 B1
10999326 Pollitt et al. May 2021 B1
11038861 Agarwal et al. Jun 2021 B2
11075747 Holsman Jul 2021 B1
11102076 Pieczul et al. Aug 2021 B1
11102147 Mehta et al. Aug 2021 B2
11163614 Francisco Nov 2021 B1
11228945 Yang et al. Jan 2022 B2
11249809 Tang et al. Feb 2022 B1
11290143 Sternowski Mar 2022 B1
11310650 Zhou Apr 2022 B2
11316822 Gawade et al. Apr 2022 B1
11323919 Parulkar et al. May 2022 B1
11375300 Sagie et al. Jun 2022 B2
11412051 Chiganmi et al. Aug 2022 B1
11431497 Liguori et al. Aug 2022 B1
11521444 Badik et al. Dec 2022 B1
11528147 Madisetti et al. Dec 2022 B2
11538287 Fang et al. Dec 2022 B2
11546323 Jones et al. Jan 2023 B1
11546763 Filho et al. Jan 2023 B1
11599714 Munro et al. Mar 2023 B2
11599841 Anisingaraju et al. Mar 2023 B2
11620103 Graham et al. Apr 2023 B2
11632669 Xu et al. Apr 2023 B2
11657145 Cristina et al. May 2023 B2
11729620 Filho et al. Aug 2023 B1
11736531 Filho et al. Aug 2023 B1
11765159 Crawford et al. Sep 2023 B1
11765207 McCarthy et al. Sep 2023 B1
11784999 Jones et al. Oct 2023 B1
11831511 Zhou et al. Nov 2023 B1
11861221 Richardson et al. Jan 2024 B1
11916885 Cirello Filho et al. Feb 2024 B1
11916968 Cirello Filho et al. Feb 2024 B1
11954219 Makmal et al. Apr 2024 B1
11973752 Crawford et al. Apr 2024 B2
12028321 Cirello Filho et al. Jul 2024 B1
12063148 Dabell et al. Aug 2024 B2
12063218 Wilczynski et al. Aug 2024 B2
12063550 Qiao et al. Aug 2024 B2
20020099952 Lambert et al. Jul 2002 A1
20020140738 West et al. Oct 2002 A1
20020149623 West et al. Oct 2002 A1
20030058286 Dando Mar 2003 A1
20030093465 Banerjee et al. May 2003 A1
20030145317 Chamberlain Jul 2003 A1
20030177182 Clark et al. Sep 2003 A1
20040019898 Frey et al. Jan 2004 A1
20040184070 Kiraly et al. Sep 2004 A1
20050022185 Romero Jan 2005 A1
20050132227 Reasor et al. Jun 2005 A1
20050164650 Johnson Jul 2005 A1
20050209876 Kennis et al. Sep 2005 A1
20060074876 Kakivaya et al. Apr 2006 A1
20060136928 Crawford et al. Jun 2006 A1
20060177005 Shaffer et al. Aug 2006 A1
20060177024 Frifeldt et al. Aug 2006 A1
20060177025 Frifeldt et al. Aug 2006 A1
20060190991 Iyer Aug 2006 A1
20060200856 Salowey et al. Sep 2006 A1
20060212487 Kennis et al. Sep 2006 A1
20060240824 Henderson et al. Oct 2006 A1
20060265708 Blanding et al. Nov 2006 A1
20060265758 Khandelwal et al. Nov 2006 A1
20060288204 Sood et al. Dec 2006 A1
20070009104 Renkis Jan 2007 A1
20070014413 Oliveira et al. Jan 2007 A1
20070033273 White et al. Feb 2007 A1
20070124797 Gupta et al. May 2007 A1
20070162359 Gokhale et al. Jul 2007 A1
20070220009 Morris et al. Sep 2007 A1
20080144502 Jackowski et al. Jun 2008 A1
20080184336 Sarukkai et al. Jul 2008 A1
20080201454 Soffer Aug 2008 A1
20080229383 Buss et al. Sep 2008 A1
20080313699 Starostin et al. Dec 2008 A1
20090037607 Farinacci et al. Feb 2009 A1
20090049509 Chan et al. Feb 2009 A1
20090063381 Chan et al. Mar 2009 A1
20090083336 Srinivasan Mar 2009 A1
20090164663 Ransom et al. Jun 2009 A1
20090222559 Anipko et al. Sep 2009 A1
20100131650 Pok et al. May 2010 A1
20100132013 Van Riel et al. May 2010 A1
20100154025 Esteve Balducci et al. Jun 2010 A1
20100161632 Rosen Jun 2010 A1
20100217853 Alexander et al. Aug 2010 A1
20100262717 Critchley et al. Oct 2010 A1
20110002333 Karuppiah et al. Jan 2011 A1
20110106757 Pickney et al. May 2011 A1
20110106770 McDonald et al. May 2011 A1
20110106771 McDonald et al. May 2011 A1
20110106802 Pinkney et al. May 2011 A1
20110225311 Liu et al. Sep 2011 A1
20120102050 Button et al. Apr 2012 A1
20120216240 Gottumukkala et al. Aug 2012 A1
20120278293 Bulkowski et al. Nov 2012 A1
20120304265 Richter et al. Nov 2012 A1
20130031157 Mckee et al. Jan 2013 A1
20130125112 Mittal et al. May 2013 A1
20130173794 Agerbak et al. Jul 2013 A1
20130198558 Rao et al. Aug 2013 A1
20130239192 Linga et al. Sep 2013 A1
20130268260 Lundberg et al. Oct 2013 A1
20130268740 Holt Oct 2013 A1
20130298183 McGrath et al. Nov 2013 A1
20140044265 Kocher et al. Feb 2014 A1
20140057676 Lord et al. Feb 2014 A1
20140136970 Xiao May 2014 A1
20140195818 Neumann et al. Jul 2014 A1
20140282900 Wang et al. Sep 2014 A1
20150082374 Dobson et al. Mar 2015 A1
20150089575 Vepa et al. Mar 2015 A1
20150127949 Patil et al. May 2015 A1
20150135300 Ford May 2015 A1
20150169871 Achutha et al. Jun 2015 A1
20150208273 Raleigh et al. Jul 2015 A1
20150281079 Fan et al. Oct 2015 A1
20150281952 Patil et al. Oct 2015 A1
20150282058 Forssell Oct 2015 A1
20150301824 Patton et al. Oct 2015 A1
20150309849 Lau et al. Oct 2015 A1
20150310025 Rathgeber et al. Oct 2015 A1
20150370793 Chen et al. Dec 2015 A1
20150370846 Zhou Dec 2015 A1
20150382198 Kashef et al. Dec 2015 A1
20160014669 Patil et al. Jan 2016 A1
20160014818 Reitsma et al. Jan 2016 A1
20160065618 Banerjee Mar 2016 A1
20160080128 Hebron Mar 2016 A1
20160085841 Dorfman et al. Mar 2016 A1
20160147529 Coleman et al. May 2016 A1
20160173501 Brown Jun 2016 A1
20160180102 Kim et al. Jun 2016 A1
20160191545 Nanda et al. Jun 2016 A1
20160212237 Nishijima Jul 2016 A1
20160224360 Wagner et al. Aug 2016 A1
20160224785 Wagner et al. Aug 2016 A1
20160262021 Lee et al. Sep 2016 A1
20160294826 Han et al. Oct 2016 A1
20160314355 Laska et al. Oct 2016 A1
20160337474 Rao Nov 2016 A1
20160378846 Luse et al. Dec 2016 A1
20160380909 Antony et al. Dec 2016 A1
20170005790 Brockmann et al. Jan 2017 A1
20170010826 Basham et al. Jan 2017 A1
20170011078 Gerrard et al. Jan 2017 A1
20170061006 Hildebrand et al. Mar 2017 A1
20170061956 Sarikaya et al. Mar 2017 A1
20170078721 Brockmann et al. Mar 2017 A1
20170093923 Duan Mar 2017 A1
20170103440 Xing et al. Apr 2017 A1
20170111368 Hibbert et al. Apr 2017 A1
20170126734 Harney May 2017 A1
20170134422 Shieh et al. May 2017 A1
20170142810 Cho May 2017 A1
20170149843 Amulothu et al. May 2017 A1
20170171245 Lee et al. Jun 2017 A1
20170177222 Singh et al. Jun 2017 A1
20170177892 Tingstrom et al. Jun 2017 A1
20170206207 Bondurant et al. Jul 2017 A1
20170212830 Thomas et al. Jul 2017 A1
20170237747 Quinn et al. Aug 2017 A1
20170244606 Htay Aug 2017 A1
20170257357 Wang et al. Sep 2017 A1
20170279971 Raleigh et al. Sep 2017 A1
20170339561 Wennemyr et al. Nov 2017 A1
20170353433 Antony et al. Dec 2017 A1
20170353496 Pai et al. Dec 2017 A1
20170364505 Sarikaya et al. Dec 2017 A1
20170364748 Maji et al. Dec 2017 A1
20170372087 Lee Dec 2017 A1
20170374032 Woolward et al. Dec 2017 A1
20170374101 Woolward Dec 2017 A1
20180027009 Santos et al. Jan 2018 A1
20180032258 Edwards et al. Feb 2018 A1
20180035126 Lee et al. Feb 2018 A1
20180041467 Vats et al. Feb 2018 A1
20180041598 Vats et al. Feb 2018 A1
20180061158 Greene Mar 2018 A1
20180083915 Medam et al. Mar 2018 A1
20180083944 Vats et al. Mar 2018 A1
20180101422 Flanigan et al. Apr 2018 A1
20180115585 Rubakha Apr 2018 A1
20180121110 Sawhney May 2018 A1
20180121129 Sawhney et al. May 2018 A1
20180123957 Chen et al. May 2018 A1
20180159701 Krause et al. Jun 2018 A1
20180167373 Anderson et al. Jun 2018 A1
20180167415 Khan et al. Jun 2018 A1
20180176262 Kavi Jun 2018 A1
20180196680 Wang et al. Jul 2018 A1
20180210801 Wu et al. Jul 2018 A1
20180218148 D'errico et al. Aug 2018 A1
20180218149 Jacobs et al. Aug 2018 A1
20180220472 Schopp Aug 2018 A1
20180233141 Solomon et al. Aug 2018 A1
20180255591 Valicherla et al. Sep 2018 A1
20180270732 Garcia Martin et al. Sep 2018 A1
20180288026 Callaghan Oct 2018 A1
20180293024 Baptist et al. Oct 2018 A1
20180314846 Schultz et al. Nov 2018 A1
20180316725 Mani et al. Nov 2018 A1
20180341519 Vyas et al. Nov 2018 A1
20180349629 Chenchev et al. Dec 2018 A1
20180359369 Golshenas et al. Dec 2018 A1
20180367308 Kacin et al. Dec 2018 A1
20190020665 Surcouf et al. Jan 2019 A1
20190039569 Reed et al. Feb 2019 A1
20190065278 Jeuk et al. Feb 2019 A1
20190068449 Sikand et al. Feb 2019 A1
20190073373 Surale et al. Mar 2019 A1
20190075095 Venable et al. Mar 2019 A1
20190081955 Chugtu et al. Mar 2019 A1
20190104411 Hotchkiss et al. Apr 2019 A1
20190116132 Suzuki Apr 2019 A1
20190124507 Dotchkoff et al. Apr 2019 A1
20190147154 Das May 2019 A1
20190156023 Gerebe et al. May 2019 A1
20190163406 Kazi et al. May 2019 A1
20190174449 Shan et al. Jun 2019 A1
20190213319 Gerebe et al. Jul 2019 A1
20190235605 Chenchev et al. Aug 2019 A1
20190254094 Babu et al. Aug 2019 A1
20190258813 Lewis et al. Aug 2019 A1
20190260753 Lewis et al. Aug 2019 A1
20190268167 White et al. Aug 2019 A1
20190281127 Schmidt et al. Sep 2019 A1
20190289459 Shan Sep 2019 A1
20190303345 Zhu et al. Oct 2019 A1
20190306754 Shan et al. Oct 2019 A1
20190313359 Lee et al. Oct 2019 A1
20190317821 O'neal et al. Oct 2019 A1
20190318240 Kulkarni et al. Oct 2019 A1
20190332777 Edwards et al. Oct 2019 A1
20190349742 Kim et al. Nov 2019 A1
20190349758 Zhu et al. Nov 2019 A1
20200004451 Prohofsky Jan 2020 A1
20200007540 Kawaguchi et al. Jan 2020 A1
20200021615 Wainner et al. Jan 2020 A1
20200026850 Levin et al. Jan 2020 A1
20200034248 Nara et al. Jan 2020 A1
20200067801 McCormick et al. Feb 2020 A1
20200089516 Vijayvargiya et al. Mar 2020 A1
20200097195 Fritz et al. Mar 2020 A1
20200112850 Lee et al. Apr 2020 A1
20200177333 Liu Jun 2020 A1
20200193016 Zeng et al. Jun 2020 A1
20200205209 Pan et al. Jun 2020 A1
20200214054 Qiao et al. Jul 2020 A1
20200218798 Kosaka et al. Jul 2020 A1
20200219023 Duchastel Jul 2020 A1
20200220848 Patwardhan Jul 2020 A1
20200228936 Talebi Fard et al. Jul 2020 A1
20200229069 Chun Jul 2020 A1
20200257179 Barnum et al. Aug 2020 A1
20200257810 Vrabec et al. Aug 2020 A1
20200260401 So Aug 2020 A1
20200267552 Lee et al. Aug 2020 A1
20200267554 Faccin et al. Aug 2020 A1
20200267785 Talebi Fard et al. Aug 2020 A1
20200272911 Quiros Araya et al. Aug 2020 A1
20200296619 Pan et al. Sep 2020 A1
20200296779 Moghe et al. Sep 2020 A1
20200314613 Lee et al. Oct 2020 A1
20200314701 Talebi Fard et al. Oct 2020 A1
20200320189 Zhang et al. Oct 2020 A1
20200322286 Mehta et al. Oct 2020 A1
20200323030 Mehta et al. Oct 2020 A1
20200334084 Jacobson Oct 2020 A1
20200344637 Kim et al. Oct 2020 A1
20200351699 Pan et al. Nov 2020 A1
20200351980 Talebi Fard et al. Nov 2020 A1
20200351984 Talebi Fard et al. Nov 2020 A1
20200367141 Cakulev et al. Nov 2020 A1
20200382488 Liu et al. Dec 2020 A1
20200396678 Lee et al. Dec 2020 A1
20200401452 Piercey et al. Dec 2020 A1
20200404107 Cheng et al. Dec 2020 A1
20210011984 Renke et al. Jan 2021 A1
20210026819 Beard et al. Jan 2021 A1
20210029132 Wilczynski et al. Jan 2021 A1
20210029168 Jeong Jan 2021 A1
20210037380 Lee et al. Feb 2021 A1
20210056524 Isgar Feb 2021 A1
20210056536 Carter et al. Feb 2021 A1
20210075631 Liao Mar 2021 A1
20210076444 Shu et al. Mar 2021 A1
20210091941 Pancras et al. Mar 2021 A1
20210099490 Crabtree et al. Apr 2021 A1
20210105596 Prabhakar et al. Apr 2021 A1
20210109900 Mcilroy et al. Apr 2021 A1
20210117241 Xia Apr 2021 A1
20210120473 Yao et al. Apr 2021 A1
20210120596 Youn et al. Apr 2021 A1
20210122261 Qiao et al. Apr 2021 A1
20210133313 Sakib et al. May 2021 A1
20210141655 Gamage et al. May 2021 A1
20210157655 Foreman May 2021 A1
20210168613 Suh et al. Jun 2021 A1
20210168661 Wong et al. Jun 2021 A1
20210168751 Stojanovski et al. Jun 2021 A1
20210211408 Porras et al. Jul 2021 A1
20210211423 Tan et al. Jul 2021 A1
20210211870 Perras et al. Jul 2021 A1
20210218652 Raut et al. Jul 2021 A1
20210218801 Zhang et al. Jul 2021 A1
20210223128 Kirch Jul 2021 A1
20210224091 Hayatnagarkar et al. Jul 2021 A1
20210226850 Xu et al. Jul 2021 A1
20210226910 Ranpise et al. Jul 2021 A1
20210227438 Xu Jul 2021 A1
20210243678 Drevon et al. Aug 2021 A1
20210250890 Won Aug 2021 A1
20210258859 Watfa Aug 2021 A1
20210274340 Sun et al. Sep 2021 A1
20210281572 Fernandez-spadaro et al. Sep 2021 A1
20210294970 Bender et al. Sep 2021 A1
20210306310 Tan Sep 2021 A1
20210307101 Kim et al. Sep 2021 A1
20210312400 Irimie et al. Oct 2021 A1
20210314399 Hyun et al. Oct 2021 A1
20210320897 Stojanovski et al. Oct 2021 A1
20210336934 Deshmukh et al. Oct 2021 A1
20210344492 Goodsitt et al. Nov 2021 A1
20210360294 Liao et al. Nov 2021 A1
20210360371 Qiao et al. Nov 2021 A1
20210368339 Watfa et al. Nov 2021 A1
20210368581 Shan Nov 2021 A1
20210382739 Guo et al. Dec 2021 A1
20210385642 Di Girolamo et al. Dec 2021 A1
20210385742 Liao Dec 2021 A1
20210397313 Desai et al. Dec 2021 A1
20210397463 Schimon et al. Dec 2021 A1
20210399954 Dabell et al. Dec 2021 A1
20210406127 Knierim et al. Dec 2021 A1
20210406902 Bernert et al. Dec 2021 A1
20210409449 Crabtree et al. Dec 2021 A1
20210410059 Talebi Fard et al. Dec 2021 A1
20220007161 Shan Jan 2022 A1
20220007180 Liao et al. Jan 2022 A1
20220007437 Goenka et al. Jan 2022 A1
20220030495 Qiao et al. Jan 2022 A1
20220038544 Grinstein et al. Feb 2022 A1
20220038986 Soliman et al. Feb 2022 A1
20220039004 Soliman et al. Feb 2022 A1
20220043665 Vidyadhara et al. Feb 2022 A1
20220052850 Fagan et al. Feb 2022 A1
20220052961 Chauhan et al. Feb 2022 A1
20220053449 Shan Feb 2022 A1
20220060398 Shishir et al. Feb 2022 A1
20220060507 Crabtree et al. Feb 2022 A1
20220060509 Crabtree et al. Feb 2022 A1
20220070967 Li et al. Mar 2022 A1
20220086639 Lu et al. Mar 2022 A1
20220092024 Kavaipatti Anantharamakrishnan et al. Mar 2022 A1
20220095260 Shan Mar 2022 A1
20220103987 Shan Mar 2022 A1
20220104164 Kedalagudde et al. Mar 2022 A1
20220104296 Mary et al. Mar 2022 A1
20220110023 Wu et al. Apr 2022 A1
20220116814 Di Girolamo et al. Apr 2022 A1
20220124521 Xu et al. Apr 2022 A1
20220124595 Xu et al. Apr 2022 A1
20220129583 Balasubramanian et al. Apr 2022 A1
20220141662 Liao May 2022 A1
20220150312 Ranpise et al. May 2022 A1
20220159029 Bendersky et al. May 2022 A1
20220159527 Lee et al. May 2022 A1
20220159605 Li et al. May 2022 A1
20220164186 Pamidala et al. May 2022 A1
20220167299 Xu et al. May 2022 A1
20220174580 You et al. Jun 2022 A1
20220191052 Garcia Azorero et al. Jun 2022 A1
20220191100 Kim et al. Jun 2022 A1
20220191736 Barton et al. Jun 2022 A1
20220191765 Ding Jun 2022 A1
20220200924 Lu et al. Jun 2022 A1
20220201638 Arrobo Vidal et al. Jun 2022 A1
20220210698 Ly et al. Jun 2022 A1
20220217084 Arora et al. Jul 2022 A1
20220225448 Li et al. Jul 2022 A1
20220232363 Watfa et al. Jul 2022 A1
20220255966 Sienicki et al. Aug 2022 A1
20220263835 Pieczul et al. Aug 2022 A1
20220263913 Zhang et al. Aug 2022 A1
20220264370 Qiao et al. Aug 2022 A1
20220264403 Watfa et al. Aug 2022 A1
20220264444 Ryu et al. Aug 2022 A1
20220264503 Starsinic et al. Aug 2022 A1
20220272614 Lu et al. Aug 2022 A1
20220272620 Ninglekhu et al. Aug 2022 A1
20220278900 Pieczul et al. Sep 2022 A1
20220279075 Fan et al. Sep 2022 A1
20220286480 Jadhav et al. Sep 2022 A1
20220294540 Black et al. Sep 2022 A1
20220295440 Kumar et al. Sep 2022 A1
20220309152 Araujo et al. Sep 2022 A1
20220311767 Ouellet Sep 2022 A1
20220312176 Matolia et al. Sep 2022 A1
20220312299 Mochizuki et al. Sep 2022 A1
20220312517 Xu et al. Sep 2022 A1
20220329477 Chiganmi et al. Oct 2022 A1
20220329495 Xie et al. Oct 2022 A1
20220330138 Shan Oct 2022 A1
20220330193 Shan Oct 2022 A1
20220334864 Kn et al. Oct 2022 A1
20220338000 Lee et al. Oct 2022 A1
20220345875 Kumar Oct 2022 A1
20220353799 Talebi Fard et al. Nov 2022 A1
20220360670 Singh et al. Nov 2022 A1
20220361136 Watfa et al. Nov 2022 A1
20220368726 Balasubramaniam et al. Nov 2022 A1
20220369408 Shariat et al. Nov 2022 A1
20220377529 Kim et al. Nov 2022 A1
20220377617 Deng et al. Nov 2022 A1
20220377654 Soliman et al. Nov 2022 A1
20220386100 Lee et al. Dec 2022 A1
20220394566 Liu et al. Dec 2022 A1
20220400378 Wang et al. Dec 2022 A1
20220408396 Youn et al. Dec 2022 A1
20220413883 Clebsch et al. Dec 2022 A1
20230007439 Williams et al. Jan 2023 A1
20230007464 Lee et al. Jan 2023 A1
20230007611 Karampatsis et al. Jan 2023 A1
20230011468 Giri et al. Jan 2023 A1
20230019448 Deshmukh et al. Jan 2023 A1
20230022134 Talwar et al. Jan 2023 A1
20230023571 Xing et al. Jan 2023 A1
20230024999 Wu et al. Jan 2023 A1
20230027290 Chen et al. Jan 2023 A1
20230027507 He et al. Jan 2023 A1
20230029714 Xu et al. Feb 2023 A1
20230032790 Mahajan et al. Feb 2023 A1
20230040747 Watfa et al. Feb 2023 A1
20230042646 Mahadevan et al. Feb 2023 A1
20230044346 Nuggehalli et al. Feb 2023 A1
20230052699 Ninglekhu et al. Feb 2023 A1
20230052827 Araujo et al. Feb 2023 A1
20230056442 Ly et al. Feb 2023 A1
20230058336 Kim et al. Feb 2023 A1
20230059173 Moon Feb 2023 A1
20230067223 Freed et al. Mar 2023 A1
20230073757 Chandramouli et al. Mar 2023 A1
20230078317 Xing et al. Mar 2023 A1
20230083175 Xiong Mar 2023 A1
20230094062 Kim et al. Mar 2023 A1
20230098558 Savir et al. Mar 2023 A1
20230109272 Ryu et al. Apr 2023 A1
20230113180 Srinivasan et al. Apr 2023 A1
20230113519 Fernandez Alonso et al. Apr 2023 A1
20230116463 Rath et al. Apr 2023 A1
20230118271 Punathil Apr 2023 A1
20230129117 Ding et al. Apr 2023 A1
20230130746 Binder et al. Apr 2023 A1
20230133444 Dimitrovski et al. May 2023 A1
20230135699 Liao et al. May 2023 A1
20230136984 Lee et al. May 2023 A1
20230138033 Rajadurai et al. May 2023 A1
20230147538 Lee et al. May 2023 A1
20230153447 Kapadia May 2023 A1
20230156513 Xing et al. May 2023 A1
20230156514 Eriksson et al. May 2023 A1
20230163984 Shan May 2023 A1
20230164523 Wu et al. May 2023 A1
20230171824 Purkayastha et al. Jun 2023 A1
20230188525 Singh et al. Jun 2023 A1
20230189192 Talebi Fard et al. Jun 2023 A1
20230195884 Belair et al. Jun 2023 A1
20230199632 Fard et al. Jun 2023 A1
20230205505 Chen et al. Jun 2023 A1
20230209329 Guo et al. Jun 2023 A1
20230224304 Lukanov et al. Jul 2023 A1
20230247003 Chanak et al. Aug 2023 A1
20230262818 Kim et al. Aug 2023 A1
20230284077 Pateromichelakis Sep 2023 A1
20230300651 Kim et al. Sep 2023 A1
20230300702 You et al. Sep 2023 A1
20230309158 Qiao et al. Sep 2023 A1
20230319112 Kaimal et al. Oct 2023 A1
20230319679 Sung et al. Oct 2023 A1
20230319685 Talebi Fard et al. Oct 2023 A1
20230319915 Paladugu et al. Oct 2023 A1
20230328821 Talebi Fard et al. Oct 2023 A1
20230336471 Dhody Oct 2023 A1
20230336977 Henry et al. Oct 2023 A1
20230362198 Jung et al. Nov 2023 A1
20230362623 Lee et al. Nov 2023 A1
20230362632 Hu et al. Nov 2023 A1
20230362640 Edge Nov 2023 A1
20230362704 Edge et al. Nov 2023 A1
20230368193 Russinovich et al. Nov 2023 A1
20230370992 You et al. Nov 2023 A1
20230371111 Xu et al. Nov 2023 A1
20230388785 Ferdi et al. Nov 2023 A1
20230397086 Kim et al. Dec 2023 A1
20230422149 Kim et al. Dec 2023 A1
20240007983 Liu et al. Jan 2024 A1
20240008130 Kim et al. Jan 2024 A1
20240012674 Xie et al. Jan 2024 A1
20240015567 Mladin et al. Jan 2024 A1
20240022469 Hu et al. Jan 2024 A1
20240048229 Kumar et al. Feb 2024 A1
20240048966 Suh Feb 2024 A1
20240049171 Khirallah et al. Feb 2024 A1
20240056904 Watfa et al. Feb 2024 A1
20240056947 Watfa et al. Feb 2024 A1
20240056957 Wang Feb 2024 A1
20240073249 Cirello Filho et al. Feb 2024 A1
20240073772 Gupta et al. Feb 2024 A1
20240080340 Ben Henda et al. Mar 2024 A1
20240080342 Mishra Gupta et al. Mar 2024 A1
20240080730 Zhang et al. Mar 2024 A1
20240080791 Aghili et al. Mar 2024 A1
20240086558 Jadhav et al. Mar 2024 A1
20240089291 Shpilyuck et al. Mar 2024 A1
20240104553 Thorpe Mar 2024 A1
20240106821 Crawford et al. Mar 2024 A1
20240107481 Tiwari et al. Mar 2024 A1
20240114057 Bjerrum et al. Apr 2024 A1
20240147207 Watfa et al. May 2024 A1
20240152602 Belair et al. May 2024 A1
20240168793 Pyka et al. May 2024 A1
20240169062 Lee et al. May 2024 A1
20240171969 Kumar et al. May 2024 A1
20240179066 Shen et al. May 2024 A1
20240179070 Zhou et al. May 2024 A1
20240179071 Zhou et al. May 2024 A1
20240179509 Fu et al. May 2024 A1
20240187968 Tiwari et al. Jun 2024 A1
20240187990 Chen et al. Jun 2024 A1
20240205863 Nassar et al. Jun 2024 A1
20240236047 Cirello Filho et al. Jul 2024 A1
20240241978 Chopra et al. Jul 2024 A1
20240244014 De Foy et al. Jul 2024 A1
20240273211 Singh Aug 2024 A1
20240275680 Palnati et al. Aug 2024 A1
Foreign Referenced Citations (1)
Number Date Country
2847507 Nov 2014 CA
Non-Patent Literature Citations (27)
Entry
Office Communication for U.S. Appl. No. 17/900,121 mailed Apr. 4, 2023, 13 Pages.
Office Communication for U.S. Appl. No. 18/091,006 mailed Apr. 19, 2023, 9 Pages.
Office Communication for U.S. Appl. No. 17/954,697 mailed May 12, 2023, 5 Pages.
Office Communication for U.S. Appl. No. 18/131,151 mailed Jun. 8, 2023, 9 Pages.
Office Communication for U.S. Appl. No. 17/954,697 mailed Jan. 20, 2023, 13 Pages.
Office Communication for U.S. Appl. No. 18/094,858 mailed Mar. 16, 2023, 10 Pages.
Office Communication for U.S. Appl. No. 18/091,006 mailed Mar. 31, 2023, 12 Pages.
Office Communication for U.S. Appl. No. 17/733,735 mailed Jul. 12, 2022, 38 Pages.
Office Communication for U.S. Appl. No. 17/733,735 mailed Nov. 1, 2022, 7 Pages.
Office Communication for U.S. Appl. No. 17/889,788 mailed Nov. 7, 2022, 12 Pages.
Office Communication for U.S. Appl. No. 18/091,895 mailed Apr. 5, 2023, 12 Pages.
Office Communication for U.S. Appl. No. 18/094,858 mailed Jul. 7, 2023, 12 Pages.
Office Communication for U.S. Appl. No. 18/131,151 mailed Jul. 10, 2023, 7 Pages.
Office Communication for U.S. Appl. No. 18/094,858 mailed Sep. 19, 2023, 4 Pages.
Office Communication for U.S. Appl. No. 18/091,895 mailed Jul. 11, 2023, 9 pages.
Office Communication for U.S. Appl. No. 18/236,360 mailed Oct. 25, 2023, 12 pages.
Office Communication for U.S. Appl. No. 18/236,360 mailed Nov. 13, 2023, 2 pages.
Office Communication for U.S. Appl. No. 18/238,649 mailed Nov. 1, 2023, 13 pages.
Office Communication for U.S. Appl. No. 18/094,858 mailed Oct. 25, 2023, 7 pages.
Office Communication for U.S. Appl. No. 18/238,649 mailed Jan. 26, 2024, 11 pages.
Office Communication for U.S. Appl. No. 18/238,649 mailed Feb. 26, 2024, 4 pages.
Office Communication for U.S. Appl. No. 18/238,649 mailed Mar. 14, 2024, 5 pages.
Office Communication for U.S. Appl. No. 18/587,821 mailed May 3, 2024, 12 pages.
Office Communication for U.S. Appl. No. 18/740,622 mailed Sep. 5, 2024, 54 pages.
Jennings et al REsource Location and Discovery (RELOAD) Base Protocol, Internet Engineering Task Force (IETF), Request for Comments:RFC 6940, pp. 1-176 (Year: 2014).
Uriarte et al “Expressive Policy-Based Access Control for Resource-Constrained Devices,” Special Section on Security and Privacy in Applications and Services for Future Internet of Things, IEEE Access, pp. 15-46 (Year: 2017).
Office Communication for U.S. Appl. No. 18/740,667 mailed Sep. 19, 2024, 8 pages.