Deceiving attackers accessing network data

Information

  • Patent Grant
  • 11695800
  • Patent Number
    11,695,800
  • Date Filed
    Wednesday, April 15, 2020
    4 years ago
  • Date Issued
    Tuesday, July 4, 2023
    a year ago
Abstract
Endpoints in a network execute a sensor module that intercepts commands. The sensor module compares a source of commands to a sanctioned list of applications received from a management server. If the source does not match a sanctioned application the command is ignored and a simulated acknowledgment is sent or, deception data is returned instead. In some embodiments, certain data is protected such that commands will be ignored or modified to refer to deception data where the source is not a sanctioned application. The source may be verified to be a sanctioned application by evaluating a certificate, hash, or path of the source. Responses from an active directory server may be intercepted and modified to reference a decoy server when not addressed to a sanctioned application. Requests to view network resources may be responded to with references to a decoy server.
Description
BACKGROUND

Once an end point system is compromised, attackers try to move laterally in the network. Attackers harvest data from end point systems and then use that information to move laterally. The systems and methods disclosed herein provide an improved approach for preventing unauthorized access to application data on endpoint systems.





BRIEF DESCRIPTION OF THE FIGURES

In order that the advantages of the invention will be readily understood, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered limiting of its scope, the invention will be described and explained with additional specificity and detail through use of the accompanying drawings, in which:



FIG. 1 is a schematic block diagram of a network environment for performing methods in accordance with an embodiment of the present invention;



FIG. 2 is a diagram illustrating components for preventing unauthorized access to application data in accordance with an embodiment of the present invention;



FIGS. 3A and 3B are process flow diagrams of methods for preventing unauthorized access in accordance with an embodiment of the present invention;



FIG. 4 is a schematic block diagram of an alternative network environment for performing methods in accordance with an embodiment of the present invention;



FIG. 5 is a diagram illustrating components for preventing unauthorized access to active directory data in accordance with an embodiment of the present invention;



FIG. 6 is a process flow diagram of a method for preventing unauthorized access to active directory data in accordance with an embodiment of the present invention;



FIG. 7 is a diagram illustrating components for preventing unauthorized access to domain data in accordance with an embodiment of the present invention;



FIG. 8 is a process flow diagram of a method for preventing unauthorized access to domain data in accordance with an embodiment of the present invention;



FIG. 9A is a schematic block diagram illustrating exploitation of an active directory server using an infected endpoint;



FIG. 9B is a schematic block diagram illustrating prevention of exploitation of an active directory server using an infected endpoint in accordance with an embodiment of the present invention;



FIG. 10 is a process flow diagram illustrating the use of decoy session data in accordance with an embodiment of the present invention;



FIG. 11 is a schematic block diagram of configuration of dynamic link libraries in accordance with an embodiment of the present invention;



FIG. 12 is a process flow diagram of a method for intercepting system calls in accordance with an embodiment of the present invention;



FIG. 13 is a process flow diagram of a method for intercepting calls to list network shares in accordance with an embodiment of the present invention;



FIG. 14 is a process flow diagram of a method for intercepting calls to list remote network resources in accordance with an embodiment of the present invention;



FIG. 15 is a process flow diagram of a method for intercepting calls to a database for a privileged group in accordance with an embodiment of the present invention;



FIG. 16 is a process flow diagram of a method for intercepting access requests with DNS names in accordance with an embodiment of the present invention; and



FIG. 17 is a schematic block diagram of a computer system suitable for implementing methods in accordance with embodiments of the present invention.





DETAILED DESCRIPTION

It will be readily understood that the components of the invention, as generally described and illustrated in the Figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of the embodiments of the invention, as represented in the Figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of certain examples of presently contemplated embodiments in accordance with the invention. The presently described embodiments will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout.


Embodiments in accordance with the invention may be embodied as an apparatus, method, or computer program product. Accordingly, the invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.), or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “module” or “system.” Furthermore, the invention may take the form of a computer program product embodied in any tangible medium of expression having computer-usable program code embodied in the medium.


Any combination of one or more computer-usable or computer-readable media may be utilized. For example, a computer-readable medium may include one or more of a portable computer diskette, a hard disk, a random access memory (RAM) device, a read-only memory (ROM) device, an erasable programmable read-only memory (EPROM or Flash memory) device, a portable compact disc read-only memory (CDROM), an optical storage device, and a magnetic storage device. In selected embodiments, a computer-readable medium may comprise any non-transitory medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.


Computer program code for carrying out operations of the invention may be written in any combination of one or more programming languages, including an object-oriented programming language such as Java, Smalltalk, C++, or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages, and may also use descriptive or markup languages such as HTML, XML, JSON, and the like. The program code may execute entirely on a computer system as a stand-alone software package, on a stand-alone hardware unit, partly on a remote computer spaced some distance from the computer, or entirely on a remote computer or server. In the latter scenario, the remote computer may be connected to the computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).


The invention is described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions or code. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.


These computer program instructions may also be stored in a non-transitory computer-readable medium that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable medium produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.


The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.


Referring to FIG. 1, the methods disclosed herein may be practiced in a network environment 100 including a plurality of domains 102a-102c. The domains 102a-102c may be any network division, such as a subnet, local area network (LAN), virtual local area network (VLAN), or the like. The domains 102a-102c may be distributed within a same building or over a large geographic area with interconnecting links including the Internet 104. The illustrated domains 102a-102c may represent one or more network components, such as routers, switches, servers, and the like that implement routing of traffic within the domains 102a-102c and control traffic flowing into and out of the domains 102a-102c


Each domain may include one or more endpoints 106a-106g. The endpoints 106a-106g are production computing devices that operate as personal computers for users or servers providing production services to other endpoints or to external computers accessing the network environment by way of the internet 104. The endpoints 106a-106g may be desktop or laptop computers, mobile phones, tablet computers, server computers, and any other type of computing device. Some endpoints 106a-106g may include internet-enabled devices, i.e. so-called internet of things (IoT) devices that are often a vulnerability.


The endpoints 106a-106g are not dedicated honeypots, but rather perform non-decoy functions and process legitimate production data and legitimate production tasks of an enterprise, such as functioning as user computers executing applications such as word processors, browsers, graphics programs etc. The endpoints 106a-106g may also function as web servers, database servers, remote login servers, application servers, and the like.


Some or all of the endpoints 106a-106g execute a sensor module 108. The sensor module 108 stores or accesses a list of sanctioned applications 110a and may also store or access a listing or description of protected data 110b. The endpoints 106a-106g may execute one or more instances of one or more of the sanctioned applications 110a and store data generated or used by these applications that corresponds to the protected data. The sanctioned applications 110a are production applications for performing any of the production tasks and functionality mentioned above. Accordingly, the protected data 110b may also be production data for use in performing the production tasks and functionality mentioned above, as opposed to decoy or deceptive data.


The sensor module 108 may store or access deception data 110c stored locally on the endpoint 106a-106g or accessed from another location. The deception data 110c may mimic the format of production data in the form of web pages, word processor documents, spreadsheets, databases, etc. The deception data 110c may also mimic other files used by applications such as credentials for authenticating the application with a remote server, configuration files, browser histories, a listing of recently accessed files, configuration files, and the like.


In the case of credentials or other files that are used to access a remote server or provide a record of accessing a remote server, the deception data 110c may reference a BotSink 120. The BotSink 120 may function as a honey pot programmed to engage an attacker while preventing access to production data or computer systems. For example, the BotSink 120 may execute one or more virtual machines implementing network services that engage and monitor malicious code while preventing access to other endpoints 106a-106g of the network. The BotSink 120 may implement any of the method methods for detecting and engaging malicious code disclosed in the following applications (herein after “the incorporated applications”), which are hereby incorporated herein by reference in their entirety:


U.S. application Ser. No. 14/458,026, filed Aug. 12, 2014, and entitled DISTRIBUTED SYSTEM FOR BOT DETECTION;


U.S. application Ser. No. 14/466,646, filed Aug. 22, 2014, and entitled EVALUATING URLS FOR MALICIOUS CONTENT;


U.S. application Ser. No. 14/549,112, filed Nov. 20, 2014, and entitled METHOD FOR DIRECTING MALICIOUS ACTIVITY TO A MONITORING SYSTEM;


U.S. application Ser. No. 15/157,082, filed May 17, 2016, and entitled EMULATING SUCCESSFUL SHELLCODE ATTACKS;


U.S. application Ser. No. 14/805,202, filed Jul. 21, 2015, and entitled MONITORING ACCESS OF NETWORK DARKSPACE;


U.S. application Ser. No. 14/965,574, filed Dec. 10, 2015, and entitled DATABASE DECEPTION IN DIRECTORY SERVICES;


U.S. application Ser. No. 15/142,860, filed Apr. 29, 2016, and entitled AUTHENTICATION INCIDENT DETECTION AND MANAGEMENT;


U.S. application Ser. No. 15/153,471, filed May 12, 2016, and entitled LURING ATTACKERS TOWARDS DECEPTION SERVERS;


U.S. application Ser. No. 15/204,779, filed Jul. 7, 2016, and entitled DETECTING MAN-IN-THE-MIDDLE ATTACKS; and


U.S. application Ser. No. 15/360,117, filed Nov. 23, 2016, and entitled IMPLEMENTING DECOYS IN NETWORK ENDPOINTS.


In some embodiments, the data 110a-110c is provided to the endpoints 106a-106g by a management server 112. The management server 112 may implement an application policy module 114. The application policy module 114 stores or accesses a listing 116a of sanctioned applications and may provide an interface for an administrator to specific what applications are included in the listing 116a. The listing 116a may indicate which applications are sanctioned for particular endpoints 106a-106g or for endpoints in a particular domain 102a-102c. The listing 116a may be automatically updated to include applications as they are installed on endpoints 106a-106g by an administrator.


In a similar manner, the management server 112 may store or access protected data 116b that lists data files, folders, or other descriptors of data that are protected. The protected data 116b may be obtained automatically from configuration files for applications. For example, an application on installation will often create directories for files used by the application. Accordingly, these directories and files may be observed by analyzing the configuration files for instructions to create these directories and files or by observing how the file system changes following installation of the application on a sample endpoint.


The management server 112 may also store or access deception data 118. As noted above, this deception data may mimic production data for one or more applications and may reference the BotSink 120 in order to lure attackers into engagement with the BotSink 120.


The management server 112 distributes some or all of the data 116a-116b, 118 to the endpoints 106a-106g. For example, deception data 118 provided to an endpoint may mimic files generated or used by one or more production applications actually installed on that endpoint. Likewise, the listing of sanctioned applications 116a provide to the endpoint may include only those applications that were installed by an administrator or authorized user on that endpoint. The deception data 118 provided to the endpoint may then include deception data mimicking the production data of those applications. The management server 112 may periodically update the data 116a-116b, 118 and distribute updated data to the endpoints 106a-106g.


The methods disclosed herein are invoked in response to an attacker system 122 attempting to access production application data on an endpoint 106a-106g. This may include the attacker system 122 issuing commands to the endpoint 106a-106g or uploading malicious code to the endpoint 106a-106g, which then attempts to access the production application data. However, unauthorized access may be prevented using the methods disclosed herein in either case.


Referring to FIG. 2, the sensor module 108 may be incorporated into the operating system 200, such as by modifying the functions of one or more libraries for performing access to a file system. The sensor module 108 intercepts file system command from instances of applications 202 installed on the endpoint as well as file system commands from an attacker, such as an attacker tool 204 executing on the endpoint. The file system commands that are intercepted may include write commands, read commands, delete comments, or instructions to list the contents of a directory or other commands to navigate through a directory.


The sensor module 108 evaluates a source of each command, e.g. the binary instance executing on the endpoint that issued the command to the operating system. If the source of the command is an application listed in the sanctioned applications 110a, the command is passed to the file system I/O (input output) functions 206 of the operating system 200, which then executes the command with respect to the production application data 208.


If the source of the command is not found to be in the sanctioned applications 110a, the command may be modified such that it refers to deception data 110c. The modified command may then be input to the file system I/O functions 206. In the case of a write or delete command, the sensor module 108 may suppress execution of the command and return an acknowledgment to the source of the command indicating that the command was executed successfully.


In the embodiment of FIG. 2, the modified commands are executed by the same file system I/O functions 206 as other commands. Accordingly, the deception data 110c may be stored in the file system and accessible to such functions 206. However, the deception data 110c may be hidden such that it is not viewable by users or applications executing on the endpoint.


In other embodiments, the sensor module 108 itself may access and return the deception data 110c, in which case the deception data 110c may be stored anywhere, including remotely from the endpoint and referenced by the sensor module 108. The deception data 110c could, for example, be encrypted such that only the sensor module 108 can decrypt the deception data 110c in order to return it in response to a read command from the attacker tool 204 or other unauthorized source.


In still other embodiments, the sensor module 108 may automatically generate deception data 110c in response to file system commands, such as based on templates that are populated with random data in order to mimic a type of file requested by the file system command from the attacker tool 204.


Referring to FIG. 3A, the sensor module 108 may include any executable code programmed to execute the illustrated method 300a. The method 300a may include receiving 302 a file system command, such as by intercepting a command made to the operating system 200 of the endpoint in which the sensor module 108 is embedded.


The method 300a may include evaluating 304 whether the data (file, directory, type of file, etc.) is protected data, such as might be indicated by the listing 110b of protected data on the endpoint. If not, the method 300a may include executing 306 the file system command without modification, i.e. passing it to the operating system 200 for execution. In some embodiments, only the sanctioned applications 110a are permitted to issue file system commands, which may include operating system utilities. Accordingly, in some embodiments, step 304 may be omitted.


The method 300a may further include evaluating a source of the file system command according to some or all of steps 308-312. For example, if the source of the file system command is found 308 to have a certificate matching that of a sanctioned application 110a, then file system command may be executed 306 with respect to the data referenced in the command.


If a hash, e.g. the SHA-1 (secure hash algorithm) hash of the binary code that issued the file system command is found 310 to match a hash of the binary executable for one of the sanctioned applications 110a, then the file system command may be executed 306 with respect to the data referenced in the command.


If a path to the binary code that issued the file system command is found 312 to match the path to the binary executable of one of the sanctioned applications 110a, then the file system command may be executed 306 with respect to the data referenced in the command.


Steps 308-312 are just examples of checks that may be used to verify whether binary code issuing a command is in fact an instance of a sanctioned application. In some embodiments, all of steps 308-312 must be satisfied before step 306 will be executed with respect to the data referenced in the command. In other embodiments, other checks may be used as alternatives or as additional requirements before step 306 will be executed with respect to the data referenced in the command.


In some embodiments, certain protected data may be bound to a particular sanctioned application 110a. Accordingly steps 308-312 may be evaluated only for those applications that are bound to the protected data 110b referenced in the file system command, i.e. the command will be executed with respect to the data referenced in the command only if one of 308-312 (or each and every one of 308-312 in some embodiments) is satisfied for at least one sanctioned application 110a that is also bound to the protected data 110b referenced in the file system command.


If the tests of steps 308-312 are not sufficient to invoke execution of step 306 for the data referenced in the file system command according to any of the embodiments mentioned above, then the source of the file system command may be determined not to be one of the sanctioned applications 110a and processing continues at steps 314, 318.


If the file system command is found 314 to be a read comment, then file system command may be changed 318 to refer to the deception data 110c of the endpoint. Step 306 may be executed with respect to the modified file system command, which will include returning the deception data referenced by the modified file to the source of the file system command, such as to the attacker tool 204.


If the file system command is found 316 not to be a read command, such as in the case of a write or a delete command, then the file system command is ignored and a simulated acknowledgment of the command is returned 316 to the source of the file system command.


Referring to FIG. 3B, in some embodiments, the sensor module 108, or a different sensor module 108 may execute the illustrated method 300b with respect to commands other than file system commands. In some embodiments, a plurality of sensor modules 108 execute on the endpoint and each sensor module 108 may intercept a different types of command. Each sensor module will then execute the method 300a or 300b upon intercepting that type of command. For example, requests to modify, delete, or read information regarding processes executing on the endpoint, the registry of the endpoint, or an application programming interface (API) available to applications executing on the endpoint. The method 300b may include receiving 320 a command and evaluating 322 whether the command references protected data or is a protected command. If not, then the command may be executed 324. For example, the protected data 110b may indicate which commands are restricted to sanctioned applications 110a in addition to data or types of data that are protected. In some embodiments, only sanctioned applications are permitted to access any data or invoke execution of any command. Accordingly, step 322 may be omitted as a path to execution 324 of a command.


The method 300b may further include evaluating the source of the command of step 320 according to steps 308-312 in the same manner as for the method 300a. If the conditions of steps 308-312 are met, then the command may be executed in the same form it was received. As for the method 300a, all of steps 308-312 must have a positive result before the command is executed 324 and one or more additional tests may be required to be satisfied before the command is executed 324 without modification.


If the steps 308-312 indicate that the source is a sanctioned application, then the method 300b may include evaluating 326 whether the command is a request for information, such as a request for information regarding executing processes, the registry, available APIs, or other system information. If so, then deception data is returned 328 to the source of the command. As for other embodiments disclosed herein, the deception data that is returned may mimic the legitimate data that is requested but not correspond to actual system data.


If the command is not a request for information, the method 300b may include returning 330 a result that simulates successful execution of the command. As for the method 300a, a delete command may indicate that data was successfully deleted, a modify command may indicate that the system data or process or operating parameter referenced in the command was modified as requested.


An example of the use of the method 300a may include the browser history for a browser. For example, the FIREFOX browser history may be stored at C:\Users\<user name>\AppData\Roaming\Mozilla\Firefox\Profiles\<some profile number>.default\formhistory.sqlite. Ordinarily, only the FIREFOX browser should access this file inasmuch as it will include functions for displaying the browser history on request. Accordingly, any other application attempting to access this file or its directory may be blocked by the sensor module 108 and instead receive a simulated browser history file including decoy data.


In another example, in Windows, the “shawdowcopy delete” command deletes the volume shadow copies. Malware often deletes this to accessing of backup volumes. Accordingly, this file may be listed as protected data 110b and sensor module 108 will detect attempts to delete the volume shadow copies by non-sanctioned applications and prevent their execution. As noted above, acknowledgments of such commands may be returned indicating that the volume shadow copies were in fact deleted.


In an example of the use of the method 30b, an attacker, e.g. attacker tool 122, tries to access the registry, which may be in a listing of protected data 110b. For example, the attacker may attempt to read the registry to determine a registry key for one or more antivirus application (e.g., MCAFEE, SYMANTEC, or the like) is present in the registry. No legitimate application would generally need to access the registry to determine whether an antivirus application is installed. In some embodiments, the sensor module 108 may intercept such attempts and return a result including a registry key for the antivirus tool, regardless of whether the antivirus application is installed. In many cases, this will cause the attacker tool to refrain from installing itself or otherwise attempting to perform malicious activities.


In another example, the attacker, e.g. attacker tool 122, seeks to determine whether the endpoint is a virtual machine (VM), such as by evaluating the registry to determine whether a hypervisor is installed and/or executing on the endpoint. The sensor module 108 intercepts these requests and returns an output indicating that the endpoint is executing a VM, e.g. indicate that the current operating environment in which the command was received is a VM or that a hypervisor is installed and/or executing on the endpoint. In another example, the attacker, e.g. attacker tool 122, seeks to view a list of processes executing on the endpoint. In response, the sensor module 108 will return a “correct list of processes list” to the attacker.


Referring to FIG. 4, the network environment 100 may further include an active directory server 402 in one or more of the domains 102a-102c. The active directory server 402 may implement a directory service. A directory service functions as databases that map and store the names of network resources to their respective network addresses. Users referencing network objects need not remember the physical address of the object. The directory may store network resources and those resources may or may not have a name. The directory can also store user identifiers (very common), departments of user identifiers, access level of user identifiers, hosts accessible by user identifiers, the access level associated with each user identifier. The directory may further store organizational details, a network topology, an IT policy, and the like. Although Active Directory, the most popular directory server, depends on DNS very heavily, it is distinct therefrom and provides much more than a simple mapping between domain names and IP addresses. In particular, directory services may perform authentication for access to resources as well. LDAP (Lightweight Directory Access Protocol) is one of the popular methods available to access the data in directory services. LDAP also provides authentication and authorization to let user access resources in directory services.


The directory service implemented by the active directory server 402 may provide authorization and access to key assets in corporate networks. Attackers may use various methods of exploitation to get unauthorized access to directory services. Once an attacker obtains access to a directory service, the attacker can easily log into key servers, databases etc. by impersonating credentials stored in directory services. The attacker may then exfiltrate data. For example, confidential data may be disclosed. In some instances, unauthorized modifications could be made (such as new account creation, access control modifications, document forgery, backup corruption, unauthorized financial transactions etc.) or data may be made unavailable (such as crypto malware, bank account takeover, bringing down or defacement of corporate web servers).


The systems and methods disclosed herein incorporate database deception into directory services so that attackers cannot differentiate between real production assets with respect to fake assets. Once an attacker access fake credentials pointing to a BotSink 120, the system prevents outbound access and simulates access to actual enterprise assets. This enables the identification and analysis of attackers for use in preventing data breaches.


“Active Directory” (AD) is one of the directory services supported by WINDOWS operating systems. The AD domain controller is a server that provides authentication services within a domain whereby it provides access to computer resources. AD provides a database storing information about objects. Each object can be a user, computer, or a group of users. Although the systems and methods disclosed herein are described for use with an Active Directory system, they may be applied with equal effectiveness to any directory service.


The systems and methods described herein returns simulated responses to requests to an active directory server 402. These responses reference the BotSink 120 and may include credentials for authenticating with respect to the BotSink 120, such as a user account defined on the BotSink 120 or a service executing on the BotSink 120. These responses lure an attacker to use the service or to attempt to use the user account. However, use of such services or accounts results in engagement of the attacker with the BotSink 120 and other resources described herein. This enables early detection of malicious intent and account misuse and diverts attention from real targets and resources.


Referring to FIG. 5, in some embodiments, the sensor module 108 on an endpoint 106a-106g may detect responses from an active directory server 402. These responses may be the result of requests issued to the active directory server 402 by an application 202 or attacker tool 204 executing on that endpoint 106a-106g. For example, the sensor module 108 may operate in conjunction with substituted or modified functions in an active directory API (application programming interface) in the operating system 200 of an endpoint 106a-106g. The substituted or modified functions may route responses from the active directory server 402 to the sensor module 108 prior to return of the response to the application that requested it. The sensor module 108 may then determine whether to return the response to an application 202 to which the response is addressed without modification or with modification. In particular, if the response is addressed to an attacker tool 204 that is not a sanctioned application, the response may be modified such that any system, user account, or credential included in the response is replaced with corresponding references to the BotSink 120, decoy computer on the BotSink 120, a credential or user account for authenticating with the BotSink 120 (e.g., logging in, accessing a particular service, etc.).


The manner in which the application to which the response is addressed is determined to be sanctioned or not may be according to the approach described in either of FIGS. 3A and 3B. In particular, the response from the active directory server 402 may be received on a port with respect to which an application is registered by the operating system 200. Accordingly, the sensor module 108 may evaluate that application to determine whether it is sanctioned according to the approach of either of FIGS. 3A and 3B. If not, the response is modified at step 606 to replace references to addresses, actual user accounts, services, computer systems, or other production resources with corresponding references to the BotSink 120, i.e. an address of the BotSink 120, a decoy user account, decoy service, decoy virtual machine, or other decoy resource executing on the BotSink 120.


Referring to FIG. 6, the sensor module 108 may execute the illustrated method 600 with respect to responses from the active directory server 402. The method 600 includes receiving 602 a response from the active directory server 402 by the sensor module 108. The sensor module 108 evaluates 604 whether the application to which the response is addressed is sanctioned. For example, this may include evaluating an application listening to a port to which the response was addressed. Alternatively, the response may include some other identifier that is sufficient to identify the application (i.e. instance of an application executing on operating system 200 to which the response is addressed. As noted above, receiving the response may be performed by interfacing with an API programmed to interface with the active directory server 402 such that response are intercepted before they are returned to the requesting application. For example, the sensor module 108 may interface with substituted or modified versions of the Power Shell such that commands input to the Power Shell are intercepted and possibly modified by the sensor module 108. Commands that may be intercepted may include commands such as ‘net group /domain “domain users”’ will return list of domain users present in Active Directory. Other example commands include ‘Get-ADComputer-Filter *’ will return all computer accounts in Active Directory, “nltest /dclist:domain name” will return domains in the network.


The modified response is then returned 608 to the application to which it is addressed, i.e. the application that requested the information included in the response from the active directory server 402. Where the application is found 604 to be sanctioned, the original response, or data from the response, is returned 608 to the source of the request without substituting references to the BotSink 120.


Note that FIG. 6 illustrates the case where a response from the active directory service is received and modified. In other embodiments, a request from an application that is not sanctioned that is addressed to the active directory service is intercepted and simulated response is generated that references the BotSink 120, i.e. a response that has an expected format corresponding to the request but with names of systems, addresses, credentials, services, etc. referencing such entities as implemented by the BotSink 120.


In some embodiments, the active directory server 402 may host decoy information describing a configuration of the BotSink 120, such as decoy information describing decoy users accounts, addresses of decoy systems (virtual machines), decoy services, and decoy credentials for accessing any of these accounts, systems, and services. Accordingly, modifying a request by the sensor module may include modifying the request such that it is a request for decoy information on the active directory server (e.g., references the BotSink 120) such that the response from the active directory server 402 includes only the decoy information rather than the information actually requested.


In some embodiments, authenticated users of an endpoint 106a-106g are added to a whitelist accessed by the sensor module 108 such that requests to the active directory server 402 and responses received are not substituted according to the method 600.


Referring to FIG. 7, in some embodiments, the sensor module 108 intercepts responses to requests to a domain controller 700. In some embodiments, the function of the domain controller 700 is performed by a domain joined computer. For example, the sensor module 108 may monitor one or both of requests to a network API 702 of the operating system 200 to the domain controller 700 and responses to such requests.


For example, referring to the method 800 of FIG. 8, the responses may be responses to requests for session data issued received 802 from a network API 702. A response determined 804 to be to a non-sanctioned application may be modified at step 806 to refer to the BotSink 120 and returned 808 to the non-sanctioned application. In one scenario, an attacker tool 204 executing on an infected endpoint (e.g., endpoint 106a) requests session data from the domain controller 700 or domain joined computer 700 using information learned via earlier active directory queries.


Responses determined 804 to be in response to requests from sanctioned applications 202 may be returned 808 to them without replacing a reference to a computer system in the response with a reference to the BotSink 120. The manner in which an application is determined 804 to be sanctioned may be performed in the same manner described above with respect to either of FIGS. 3A and 3B.



FIGS. 9A and 9B illustrate how the methods 600 and 800 may be used to redirect and occupy an attacker. In an unprotected case shown in FIG. 9A, an attacker tool 204 on an infected endpoint 106a uses the active directory API on the infected system to access an active directory database 900 through the active directory server 402. Using this information, the attacker tool 204 uses the network API to request session data in order to identity of an endpoint 106b logged in to the domain account for a domain to which infected endpoint 106a belong. The attacker tool may infect the administrator endpoint 106b and access the active directory database 900 and discover that the admin user account is referenced in domain data 904 defining another group, e.g. an administrator group. The endpoint 106b may further store credentials enabling the endpoint 106b to access a service hosted by server 106f.


Accordingly, the attacker tool 204 on the infected system 106a may attempt to move laterally by requesting information from the active directory server 402 in order to discover the endpoint 106b of the admin user account, infect it, and use it to access the service on server 106f as an administrator.



FIG. 9B illustrates an alternative view seen by an attacker tool 204 on the infected endpoint 106a when the methods according to FIGS. 4 through 6 are implemented. The attacker tool 204 requests information from the active directory server 402. The responses to these requests are modified to reference decoy information, such as decoy active directory data 906 having a same format as the active directory database 900 but with references to computer systems and services being replaced with references to the BotSink 120 (e.g. virtual machines executing on the BotSink 120) and services executing on the BotSink 120.


A response to a request to obtain the identity of an administrator of a domain to which the endpoint 106a belongs using the network API is intercepted and modified to reference to a decoy domain 908. A request for the identity of the administrator of the decoy domain 908 may include a reference to a user account 912 or virtual machine logged in to a user account 912 on the BotSink 120. That user account 912 may be defined as part of an admin domain 910 and that user account 912 may include credentials (e.g., VM for which the user account 912 is logged in may store or reference credentials) that are sufficient to authenticate a user with respect to a service 914 executing on the BotSink 120.


Accordingly, the attacker tool 204 may attempt to move laterally as in the case of FIG. 9A but only engage the BotSink 120. The activities of the attacker tool 204 may be monitored and logged by the BotSink 120 in order to characterize the attacker tool 204 and alerts may be generated to alert and administrator to remove the attacker tool 204 from the infected endpoint 106a. Engaging, monitoring, logging, characterizing, and generation of alerts may be performed according to any of the approaches described in the incorporated applications.



FIG. 10 illustrates a method 1000 that may be implemented using the system shown in FIG. 9B. An attacker tool 204 executing on an infected endpoint (e.g., endpoint 106a) requests 1002 session data using a network API for the domain of the infected endpoint, the session data indicating computers (e.g., computers in the same domain) that are connected to the active directory server 402. For example, a network API command for requesting such session data may be used. The sensor module 108 intercepts 1004 a response to this request and determines that the attacker tool 204 is a non-sanctioned application attempting to access restricted data (see FIG. 6). In response, the sensor module 108 returns 1006 the session data with references to one or more computers in the session data replaced with one or more addresses assigned to the BotSink 120. In the alternative, the sensor module 108 intercepts the request for session data and creates a decoy response referencing the BotSink 120 rather than modifying a response from the active directory server 402.


In particular, the IP address of the computer logged in to the administrator account of the domain may be replaced with an IP address assigned to the BotSink 120. For example, the BotSink 120 may acquire IP addresses in various domains as described in the incorporated applications. Each address added to the session data may be assigned to a virtual machine (VM) executing on the BotSink 120.


The attacker tool 204 may then attempt to access 1008 the computer logged in to the administrator account, which is in fact the BotSink 120. The BotSink 120 (e.g., the VM assigned the IP address represented as that of the administrator's computer at step 1006 (“the admin VM”)) may then engage 1010 the attacker tool. The activities of the attacker tool 204 with respect to the BotSink 120 may be monitored and stored in order to characterize the attacker tool 204 as described in the incorporated applications.


In particular, the attacker tool 204 may attempt to obtain credentials or identifiers of services cached or stored by the admin VM. In response, these credentials may be returned 1012 to the attacker tool. However, these credentials or identifiers of services may reference services implemented by the BotSink 120, such as by other VMs being executed by the BotSink 120. The attacker tool 204 receives the credentials or identifiers and attempts to access 1014 the services represented by them.


In response, the BotSink 120 authenticates the attacker tool 204 and engages 1016 with the attacker tool 204 using the service, i.e. executes commands from the attacker tool in accordance with the service. Actions of the attacker tool 204 may be monitored and used to characterize the attacker tool 204 as described in the incorporated applications. In addition or as an alternative, interaction with the BotSink 120 at steps 1010, 1012, and 1016 by the attacker tool may trigger an alert or remedial action such as removing the attacker tool 204 from the infected endpoint 106a or isolating (disconnecting) the infected endpoint 106a from a network.


Note that in some embodiments, one of the services implemented on the BotSink 120 may be a decoy active directory service that implements an interface for responding to commands in the active directory API and is programmed to response to these commands with decoy data referencing decoy computers, user accounts, and services implemented on the BotSink 120.


Referring to FIG. 11, in some implementations, intercepting of system calls according to the foregoing and subsequent embodiments may be implemented by substituting a reference to a detour function 1100 in the place of a target function 1102 in source function 1104 such that when the source function 1104 executes code that would have called the target function 1102, the detour function 1100 is called instead. For example, this reference replacement may be made in dynamic link libraries (DLL) of the operating system on the endpoint 106a-106g or other files of the operating system.


The detour function 1100 may perform evaluations described herein to determine whether the call should be handled in cooperation with the BotSink 120 or should be passed on to the target function. The detour function 1100 may also perform actions described herein relating to substituting decoy information in a response to a call to the target function 1102. Where it is determined that the call should be routed to the target function 110 this may be performed by way of a pass through function 1106 or “trampoline” function, which invokes the target function 1102 using arguments passed to the detour function 1100 by the source function 1104. The target function 1102 may return its result to the detour function 1100, which then returns the result to the source function 1104. Alternatively, the result may be routed to the detour function 1100 by way of the pass through function 1106. Where the function call is routed to the BotSink 120, a result form the BotSink 120 may be routed to the source function 1104 by the detour function 1100.


The processing of system calls according to the methods disclosed herein may be preceded by modifying source functions making the system calls on the endpoint 106a-106g to refer to one or more detour functions as described above.



FIG. 12 illustrates a method 1200 for identifying whitelisted applications. For a period, e.g. 1 day, 2 days, 1 week, etc. following processing of source functions 1104 on the operating system of one or more endpoints 106a-106g to reference the detour function(s) 1100, a learning mode may be invoked 1202. During this period the detour functions 1100 may operate in a learning mode in which all calls to a target function 1102 are intercepted 1204 and the detour function 1100 invokes 1210 the target function 1102. I.e., a call to a target function 1102 will be passed to the target function 1102. The detour function 1100 may report data regarding calls during the learning period, particularly identifiers of applications or processes that issue the call, e.g., a registry key, executable name, file system path to the executable, hash of the executable, or other identifier of the calling application or process. The detour function 1100 may make this report to the BotSink 120 for each system call or make a report of each unique identifier that made a call during the learning period. Note also that in some embodiments, not all system calls are intercepted, i.e. have references thereto replaced with detour function(s) 1100. For example, either (a) only references to certain target functions 1102 are replaced with references to detour function(s) 1100 or (b) some functions 1100 simply pass a function call to a corresponding target function 1100 while others perform interception, inspection, and substitution functions as described herein.


If the application or process that issued the intercepted call is not found 1206 to be in a list of sanctioned applications, it is added 1208 to the list, i.e., the unique identifier of the application or process is added to the list. If the application or process that issued the intercepted call is found 1206 to already be in a list of sanctioned applications, then step 1210 is performed without repeating step 1208.


In some embodiments, each endpoint 106a-106g creates its own list of sanctioned applications during the learning period and then sends these to the BotSink 120, which then compiles a list of sanctioned applications that are then distributed back to the endpoints 106a-106g. Alternatively, the endpoints 106a-106g may coordinate with one another to share identifiers of sanctioned applications such that each endpoint 106a-106g has the same list. In another approach, each endpoint 106a-106g compiles its own list of sanctioned applications and uses it according to the methods described herein without obtaining identifiers of other applications identified by other endpoints 106a-106g.


Referring to FIG. 13, the illustrated method 1300 may be executed on an endpoint 106a-106g. For example, the method 1300 may be executed by a detour function 1104, sensor module 108, or other component executing on the endpoint 106a-106g (hereinafter “the endpoint”).


The method 1300 may include receiving 1302 a call to display network shares mounted to the endpoint. Network shares are file systems on other computers remote from the endpoint that are mounted to the file system of the endpoint and accessed through the file system of the endpoint. The call may be the “net use” command in the MICROSOFT operating system. Other function calls may include GetLogicalDrives( ) GetVolumeInformation( ). This call may be intercepted 1304 by virtue of a reference to it being a call to a detour function 1100 substituted for a target function corresponding to the call. Interception may also be performed by a sensor module 108 or according to any interception approach described herein.


The detour function 1100 or other component performing the method 1300, may evaluate 1306 whether the source of the call is a sanctioned application, e.g. a unique identifier of the source of the call is in the list of sanctioned applications. If so, the detour function 1100 invokes 1308 the target function 1102, i.e. the “net use” function and returns a result of the function to the source of the call. If not, then the detour function 1100 returns 1310 a simulated result that includes a reference to a network share on the BotSink 120. For example, the actual response to the “net use” function call. The simulated result may be obtained by the detour function 1100 issuing a call to the corresponding target function 1102, i.e. the net use call, receiving the response, and replacing references to network shares with references to the BotSink 120. In other approaches, the simulated result is generated without generating a call to the target function 1102.


For example, the actual result from the “net use” function may be:















Status
Local
Remote
Network







OK
Z:
\\network426.example.com
Microsoft Windows





Network









The result returned to the source of the call may be:















Status
Local
Remote
Network







OK
Z:
\\network521.example.com
Microsoft Windows





Network









In this example, “network426.example.com” may be an actual server providing a network share. “network521.example.com” may be the domain name assigned to the BotSink 120.


Attempts by the source of the call to access the network shares may result in the source of the call engaging 1312 the BotSink 120. For example, the BotSink 120 may create a network share referenced by the result returned at step 1310. Any attempt to access the share may be responded to by providing simulated files. Attempts to access the network share may be monitored by connection attempts on BotSink 120 to host network426.example.com or by monitoring the DNS queries to the host network426.example.com. Attempts to access the BotSink 120 through the references in the modified response may invoke generation of an alert or other action to reduce risk to the endpoint or other network resources, such as blocking access to the endpoint or other network resources by the source address or executable that is the source of the call.


Referring to FIG. 14, the illustrated method 1400 may be executed on an endpoint 106a-106g. For example, the method 1400 may be executed by a detour function 1100, sensor module 108, or other component executing on the endpoint.


The method 1400 may include receiving 1402 a call to list a remote network resource or request to list credentials managed by a credential manager, or a request to list of domain controllers accessible from the endpoint, or a request to enumerate remote network shares. This call may be intercepted 1304 by virtue of a reference to it being a reference to a detour function 1100 when the interception DLL was loaded into the application or process making the call. Interception may also be performed by a sensor module 108 or according to any interception approach described herein.


The detour function 1100 or other component performing the method 1400, may execute 1404 the call on the operating system of the endpoint, such as by invoking the corresponding target function 1102. The detour function 1100 may intercept 1406 a response to the call from the operating system. If the source of the call is found 1408 to be a sanctioned application, the result from the operating system is returned 1410 to the source of the call. The manner in which the source is determined to be sanctioned may be according to the approach of any of the methods described hereinabove.


If not, then the detour function 1100 returns 1412 a simulated result that includes a reference to a network share on the BotSink 120. For example, response from step 1406 may be modified to include references to the BotSink 120 in the place of references to network resources.


For example, the system call may be to an API (application programming interface) of a credential manager, such as that provided by the WINDOWS CREDENTIAL MANAGER API. For example, such a call may include “CredEnumerateA( ), CredReadA( ), or other function call defined in Advapi32.dll. The result from the operating system received at step 1406 may be:


TERMSRV/network426.internalcorp.com, UserName rdpuser@internalcorp.com, Password: HASH1


The modified result returned to the source of the call may be


TERMSRV/network521.internalcorp.com, UserName decoyuser@internalcorp.com, Password: HASH2


where network426.internalcorp.com is a server for which the credential manager has a credential and network521.internalcorp.com is a domain name assigned to the BotSink 120. Note that the username and hash of the password are also altered in this example.


In another example, the call intercepted at step 1402 is a call to list domain controllers. Such a call may include “WNetOpenEnumA( )”, “WNetEnumResourceA( )”, or other function call defined in Mpr.dll of the Win32 API. In another example, the call is “NetShareEnum( . . . )” in Netapi32.dll of the Win32 API.


For example, the source of the call may issue the call “net view bantestadoO5.” The operating system may return the result:


Shared resources at bantestad05

















Share name
Type
Used as
Comment



















NETLOGON
Disk
Logon server share



SYSVOL
Disk
Login server share









In particular, the domain controllers may return NETLOGON and SYSVOL shares type. A result referencing these shares would then be modified to either (a) remove them to hide them from malware or (b) replace them with references to a decoy domain controller implemented on the BotSink 120. This hinders an attacker from fingerprinting or finding the domain controllers of a network.


In another example, the call received at step 1402 is to find network shares on a remote machine using a call such as WNetOpenEnumA( ), WNetEnumResourceA( ) in the Mpr.dll of the Win32 API or NetShareEnum( ) in the Netapi32.dll of the Win32 API.


For example, the result of a call such as “net view sx1.testsystem/all” may include:


Shared resources at sx1.testsystem

















Share name
Type
Used as
Comment



















ADMIN$
Disk
Remote Admin



SHARE_CIFS
Disk




C$
Disk
Default share



IPC$
IPC
Remote IPC









The entries of this result indicate network shares from a remote system. These network shares may be hidden from a non-sanctioned application according to the methods described herein. For example, for a non-sanctioned application, the entries may simply be deleted, returning a response at step 1410 that does not indicate the presence of any network shares. This enables the hiding of these network shares from malware so that it cannot find these shares to exploit them or spread a ransomware-type attack. Alternatively, network shares on the BotSink 120 may be substituted in the result.


In another example, the call received at step 1402 is a request to get information about network servers and/or computers. For example, the “NetServerEnum( )” call in the Netapi32.dll of the Win32 API.


For example, the input call may return “NETWORK426” when executed by the operations, where NETWORK426 is an identifier of an actual server on the network. The detour function 1100 may return NETWORK521, which refers to the BotSink 120.


In another example, the call at step 1402 is a call to enumerate a registry of recently accessed servers or other computers. Such a call may include RegEnumValue( ) in the Advapi32.dll of the Win32 API, which enumerates a list of remote servers or computers recently accessed using a remote desktop connection.


The result at step 1406 may read as follow:

  • HKEY_CURRENT_USER\Software\Microsoft\Terminal server Client\Default MRU0 REG_SZ NETWORK426


The modified result returned at step 1410 may read as follows:

  • HKEY_CURRENT_USER\Software\Microsoft\Terminal server Client\Default MRU0 REG_SZ NETWORK521


where NETWORK426 refers to a server in the network other than the BotSink 120 and NETWORK521 refers to the BotSink 120.


As for other embodiments, attempts to access the network resources in the modified response will be received by the BotSink 120, which may then engage 1414 the source of the call in order to detect and characterize it. Attempts to access the BotSink 120 through the references in the modified response may invoke generation of an alert or other action to reduce risk to the endpoint or other network resources, such as blocking access to the endpoint or other network resources by the source address or executable that is the source of the call.


Referring to FIG. 15, the illustrated method 1500 relates to a circumstance in which malware obtains a local administrative credential for a security account manager (SAM) database, such as from a local security authority subsystem service (LSAS) process. In this case, the endpoint receives 1502 a query to obtain information regarding a privileged group database. The privileged group may be referenced by the data obtained by the malware. The query may be a call to SamEnumerateGroupsInDomain( ), SamGetMembersIngroup( ) in the Win32 API.


If the query is found 1504 to be from a sanctioned application according to any of the approaches described herein, then the end point executes 1506 the query and returns 1508 a result of the query to the application, process, or remote computer system from which the query was received. If the query is not found 1504 to be from a sanctioned application, the endpoint provides 1510 decoy data to the source of the call, i.e. data including references to users, groups, or both users and groups that do not correspond to actual users and groups on the endpoint or in the production SAM database of the endpoint and network 100. The decoy users and/or groups may reference decoy users and groups defined on the BotSink 120, e.g., in a decoy SAM database on the BotSink 120.


If a request is found 1512 to have been received that references a decoy user or group returned at step 1510, the method 1500 may include providing decoy information from decoy accounts corresponding to the decoy user or group referenced in the request. For example, this may be performed in response to receiving a SamGetMembersInGroup( ) call that references a decoy group provided at step 1510. Steps 1512 and 1514 may be performed any number of times. In particular, decoy data may be generated in response to each request such that an attacker will be fed decoy data to occupy and deceive the attacker.


Further attempts to access or exploit the user accounts may be directed to the BotSink 120 that may engage 1516 the attacker, such as by implementing the user accounts on the BotSink 120 and providing access to services in the context of the user accounts on the BotSink 120. These activities may be monitored and characterized. Engaging 1516 may invoke generation of an alert or other action to reduce risk to the endpoint or other network resources, such as blocking access to the endpoint or other network resources by the source of the call


Referring to FIG. 16, in some embodiments an endpoint 106a-106g may execute the illustrated method 1600. A source may issue a call that references a DNS (domain name service) address that must be resolved to an IP (interne protocol address). Such a call may include DNSQuery_A( ) from the Dnsapi.dll of the Win32 API. This call is received 1602 by the endpoint, such as by way of a detour function 1100, which evaluates 1604 whether the call was received from a sanctioned application according to any of the approaches described herein. If so, the endpoint invokes 1606 performing of the access request. In some embodiments, prior to performing step 1606, the endpoint resolves the DNS address to an IP address, such as by issuing a call with the DNS address to a DNS server and receiving a response including the IP address. The endpoint may then invoke performing of the access request with the IP address by the operating system of the endpoint, i.e. invoke the target function 1102 corresponding to the call. Alternatively, the arguments of the call, including the DNS address may be passed to the target function 1102 without first resolving the DNS address to an IP address. The target function 1102 may then execute with respect to the IP address. For example, if the functions 1104, 1106 are a “ping” command, the target function 1102 will ping the BotSink 120 and return a result of that function, i.e. when and whether replies to ping messages were received, a delay between a ping message and a reply, and the IP address of the destination of the ping message.


If the source of the call is not found 1604 to be a sanctioned application, the DNS address may be resolved 1608 to an address that does not actually correspond to the DNS address. Instead, the DNS address is resolved 1608 to one of one or more IP addresses assigned to the BotSink 120. The access request may be performed 1610, e.g. target function 1102 invoked, with the arguments from the call received at step 1602 except that the DNS address is replaced with the IP address from step 1608. The result of the access request is then returned to the source of the access request. For example, where the access request is a ping command, the result of pining the BotSink 120 may be returned to the source. As for other embodiments disclosed herein, attempts to communicate using the information returned may result in engaging 1612 the attacker as described above.



FIG. 17 is a block diagram illustrating an example computing device 1700 which can be used to implement the system and methods disclosed herein. The endpoints 106a-106g, management server 112, BotSink 120, attacker system 122, and active directory server 402 may also have some or all of the attributes of the computing device 1700. In some embodiments, a cluster of computing devices interconnected by a network may be used to implement any one or more components of the invention.


Computing device 1700 may be used to perform various procedures, such as those discussed herein. Computing device 1700 can function as a server, a client, or any other computing entity. Computing device can perform various monitoring functions as discussed herein, and can execute one or more application programs, such as the application programs described herein. Computing device 1700 can be any of a wide variety of computing devices, such as a desktop computer, a notebook computer, a server computer, a handheld computer, tablet computer and the like.


Computing device 1700 includes one or more processor(s) 1702, one or more memory device(s) 1704, one or more interface(s) 1706, one or more mass storage device(s) 1708, one or more Input/Output (I/O) device(s) 1710, and a display device 1730 all of which are coupled to a bus 1712. Processor(s) 1702 include one or more processors or controllers that execute instructions stored in memory device(s) 1704 and/or mass storage device(s) 1708. Processor(s) 1702 may also include various types of computer-readable media, such as cache memory.


Memory device(s) 1704 include various computer-readable media, such as volatile memory (e.g., random access memory (RAM) 1714) and/or nonvolatile memory (e.g., read-only memory (ROM) 1716). Memory device(s) 1704 may also include rewritable ROM, such as Flash memory.


Mass storage device(s) 1708 include various computer readable media, such as magnetic tapes, magnetic disks, optical disks, solid-state memory (e.g., Flash memory), and so forth. As shown in FIG. 17, a particular mass storage device is a hard disk drive 1724. Various drives may also be included in mass storage device(s) 1708 to enable reading from and/or writing to the various computer readable media. Mass storage device(s) 1708 include removable media 1726 and/or non-removable media.


I/O device(s) 1710 include various devices that allow data and/or other information to be input to or retrieved from computing device 1700. Example I/O device(s) 1710 include cursor control devices, keyboards, keypads, microphones, monitors or other display devices, speakers, printers, network interface cards, modems, lenses, CCDs or other image capture devices, and the like.


Display device 1730 includes any type of device capable of displaying information to one or more users of computing device 1700. Examples of display device 1730 include a monitor, display terminal, video projection device, and the like.


Interface(s) 1706 include various interfaces that allow computing device 1700 to interact with other systems, devices, or computing environments. Example interface(s) 1706 include any number of different network interfaces 1720, such as interfaces to local area networks (LANs), wide area networks (WANs), wireless networks, and the Internet. Other interface(s) include user interface 1718 and peripheral device interface 1722. The interface(s) 1706 may also include one or more user interface elements 1718. The interface(s) 1706 may also include one or more peripheral interfaces such as interfaces for printers, pointing devices (mice, track pad, etc.), keyboards, and the like.


Bus 1712 allows processor(s) 1702, memory device(s) 1704, interface(s) 1706, mass storage device(s) 1708, and I/O device(s) 1710 to communicate with one another, as well as other devices or components coupled to bus 1712. Bus 1712 represents one or more of several types of bus structures, such as a system bus, PCI bus, IEEE 1394 bus, USB bus, and so forth.


For purposes of illustration, programs and other executable program components are shown herein as discrete blocks, although it is understood that such programs and components may reside at various times in different storage components of computing device 1700, and are executed by processor(s) 1702. Alternatively, the systems and procedures described herein can be implemented in hardware, or a combination of hardware, software, and/or firmware. For example, one or more application specific integrated circuits (ASICs) can be programmed to carry out one or more of the systems and procedures described herein.

Claims
  • 1. A method comprising: replacing in a first system call, by a computer system, a reference to a target function with a reference to a detour function, such that the first system call calls the detour function instead of the target function;receiving, by the computer system, the first system call from a first source function to obtain information regarding a remote network resource;evaluating, by the detour function, whether the first source function is sanctioned;determining, by the computer system, that the first source function is not sanctioned;in response to determining that the first source function is not sanctioned, returning, by the detour function, a response to the first system call having the information regarding the remote network resource replaced with information regarding a decoy server;replacing in a second system call, by a computer system, a reference to the target function with a reference to the detour function, such that the second system call calls the detour function instead of the target function;receiving, by the computer system, the second system call from a second source function;evaluating, by the detour function, whether the second source function is sanctioned;determining, by the computer system, that the second source function is sanctioned; andin response to determining that the second source function is sanctioned, invoking, by the detour function, the target function,wherein the target function is in a dynamic link library.
  • 2. The method of claim 1, further comprising: invoking, by the detour function, the target function in response to the first system call;receiving, by the detour function, a result from the target function;modifying, by the detour function, the result to obtain a modified result referencing the decoy server;replacing, by the decoy function, a first reference in the response to the first system call with a second reference referencing a decoy server to obtain a modified response;returning, by the detour function, the modified response to the first system call.
  • 3. The method of claim 1, further comprising: modifying, by the decoy function, an argument of the first system call to replace a domain name service (DNS) address with an internet protocol (IP) address of the decoy server to obtain a modified argument;invoking, by the decoy function, the target function in response to the first system call with the modified argument;receiving, by the decoy function, a result from the target function; andreturning, by the decoy function, the result to the first system call.
  • 4. The method of claim 1, wherein the first system call is an instruction to list network shares mounted to the computer system.
  • 5. The method of claim 1, wherein the first system call is an instruction to list credentials for network services stored on the computer system.
  • 6. The method of claim 1, wherein the first system call is an instruction to list domain controllers.
  • 7. The method of claim 1, wherein the first system call is an instruction to enumerate network computers.
  • 8. The method of claim 1, wherein the first system call is an instruction to list users and groups, the method further comprising: generating, by the computer system, a first response including references to a decoy group defined on the decoy server;returning, by the computer system, the first response to the first source function;receiving, by the computer system from the first source function, a request for information regarding the decoy group;in response to the request for information regarding the decoy group, returning a second response to the first system call, the second response including decoy account information for a decoy user in the decoy group.
  • 9. A system comprising: a computer system including one or more processing devices and one or more memory devices operably coupled to the one or more processing devices, the one or more memory devices storing executable code that, when executed by the one or more processing devices, causes the one or more processing devices to: replace in a first system call a reference to a target function with a reference to a detour function, such that the system call calls the detour function instead of the target function;receive the system call from a first source function to obtain information regarding a remote network resource;evaluate whether the first source function is sanctioned;determine that the first source function is not sanctioned;in response to determining that the first source function is not sanctioned, return a response to the first system call having the information regarding the remote network resource replaced with information regarding a decoy server;replace in a second system call a reference to the target function with a reference to the detour function, such that the second system call calls the detour function instead of the target function;receive the second system call from the second source function;determine that the second source function is sanctioned; andin response to determining that the second source function is sanctioned, invoke, by the detour function, the target function,wherein the target function is in a dynamic link library.
  • 10. The system of claim 9, wherein the executable code, when executed by the one or more processing devices, further causes the one or more processing devices to: invoke, by the detour function, the target function in response to the first system call;receive, by the detour function, a result from the target function;modify, by the detour function, the result to obtain a modified result referencing the decoy server;replace, by the detour function, a first reference in the response with a second reference referencing a decoy server to obtain a modified response;return, by the decoy function, the modified response to the first system call.
  • 11. The system of claim 9, wherein the executable code, when executed by the one or more processing devices, further causes the one or more processing devices to: modify an argument of the first system call to replace a domain name service (DNS) address with an internet protocol (IP) address of the decoy server to obtain a modified argument;invoke, by the decoy function, the target function in response to the first system call with the modified argument;receive, by the decoy function, a result from the target function; andreturn, by the decoy function, the result to the first system call.
  • 12. The system of claim 9, wherein the first system call is an instruction to list network shares mounted to the computer system.
  • 13. The system of claim 9, wherein the first system call is an instruction to list credentials for network services stored on the computer system.
  • 14. The system of claim 9, wherein the first system call is an instruction to list domain controllers.
  • 15. The system of claim 9, wherein the first system call is an instruction to enumerate network computers.
  • 16. The system of claim 9, wherein the first system call is an instruction to list users and groups; wherein the executable code, when executed by the one or more processing devices, further causes the one or more processing devices to:generate a first response including references to a decoy group defined on the decoy server;return the first response to the first system call;receive a request for information regarding the decoy group;in response to the request for information regarding the decoy group, return a second response to the first system call, the second response including decoy account information for a decoy user in the decoy group.
RELATED APPLICATIONS

This application is a continuation-in-part of U.S. application Ser. No. 16/543,189, filed Aug. 16, 2019, and entitled “Deceiving Attackers Accessing Active Directory Data”, which is hereby incorporated herein by reference in its entirety. U.S. patent application Ser. No. 16/543,189 is a continuation-in-part of U.S. application Ser. No. 15/383,522, filed Dec. 19, 2016, and entitled “Deceiving Attackers in Endpoint Systems”, which is hereby incorporated herein by reference in its entirety for all purposes

US Referenced Citations (470)
Number Name Date Kind
4979118 Kheradpir Dec 1990 A
5311593 Carmi May 1994 A
6154844 Touboul et al. Nov 2000 A
6157953 Chang et al. Dec 2000 A
6728716 Bhattacharya et al. Apr 2004 B1
6804780 Touboul Oct 2004 B1
6836888 Basu et al. Dec 2004 B1
7076696 Stringer Jul 2006 B1
7093239 Van der Made Aug 2006 B1
7181769 Keanini et al. Feb 2007 B1
7225468 Waisman et al. May 2007 B2
7299294 Bruck et al. Nov 2007 B1
7305546 Miller Dec 2007 B1
7322044 Hrastar Jan 2008 B2
7464407 Nakae et al. Dec 2008 B2
7530106 Zaitsev et al. May 2009 B1
7543269 Krueger Jun 2009 B2
7546587 Marr et al. Jun 2009 B2
7574740 Kennis Aug 2009 B1
7596807 Ptacek et al. Sep 2009 B2
7596808 Wilkinson et al. Sep 2009 B1
7710933 Sundaralingam et al. May 2010 B1
7739516 Bender et al. Jun 2010 B2
7832012 Huddleston Nov 2010 B2
7882538 Palmer Feb 2011 B1
7890612 Todd et al. Feb 2011 B2
7937755 Guruswamy May 2011 B1
7958549 Nakae et al. Jun 2011 B2
7984129 Vaught Jul 2011 B2
8015605 Yegneswaran Sep 2011 B2
8024795 Newton Sep 2011 B2
8042186 Polyakov et al. Oct 2011 B1
8065722 Barford et al. Nov 2011 B2
8078556 Adi et al. Dec 2011 B2
8082471 Khan Dec 2011 B2
8131281 Hildner et al. Mar 2012 B1
8141154 Gruzman et al. Mar 2012 B2
8156556 Krishnamurthy Apr 2012 B2
8171545 Cooley et al. May 2012 B1
8181033 Paul May 2012 B1
8181250 Rafalovich et al. May 2012 B2
8204984 Aziz Jun 2012 B1
8205035 Reddy et al. Jun 2012 B2
8230505 Ahrens et al. Jul 2012 B1
8296842 Singh et al. Oct 2012 B2
8327442 Herz et al. Dec 2012 B2
8353033 Chen et al. Jan 2013 B1
8370931 Chien et al. Feb 2013 B1
8375444 Aziz et al. Feb 2013 B2
8375447 Amoroso et al. Feb 2013 B2
8413238 Sutton Apr 2013 B1
8413241 Weeks et al. Apr 2013 B2
8418250 Morris et al. Apr 2013 B2
8438386 Hegli et al. May 2013 B2
8438626 Anderson et al. May 2013 B2
8443442 Wang et al. May 2013 B2
8474044 Zawadowskiy et al. Jun 2013 B2
8488466 Breslin et al. Jul 2013 B2
8528057 Garrett Sep 2013 B1
8528087 Hsu et al. Sep 2013 B2
8538578 Battles et al. Sep 2013 B2
8539582 Aziz et al. Sep 2013 B1
8549643 Shou Oct 2013 B1
8555385 Bhatkar et al. Oct 2013 B1
8561177 Aziz et al. Oct 2013 B1
8566946 Aziz et al. Oct 2013 B1
8607054 Ramarathinam et al. Dec 2013 B2
8607340 Wright Dec 2013 B2
8627475 Loveland et al. Jan 2014 B2
8677494 Edery et al. Mar 2014 B2
8713306 Bennett Apr 2014 B1
8719937 Sundaram et al. May 2014 B2
8725898 Vincent May 2014 B1
8726389 Morris et al. May 2014 B2
8732296 Thomas et al. May 2014 B1
8752173 Yadav Jun 2014 B2
8789135 Pani Jul 2014 B1
8793151 Delzoppo et al. Jul 2014 B2
8821242 Hinman et al. Sep 2014 B2
8839369 Dai et al. Sep 2014 B1
8849880 Thelen Sep 2014 B2
8850582 Endoh et al. Sep 2014 B2
8880435 Catlett et al. Nov 2014 B1
8881282 Aziz et al. Nov 2014 B1
8893278 Chechik Nov 2014 B1
8898788 Aziz et al. Nov 2014 B1
8904527 Dawson et al. Dec 2014 B2
8943594 Arrowood Jan 2015 B1
8949986 Ben-Shalom Feb 2015 B2
8959338 Snow et al. Feb 2015 B2
8973142 Shulman et al. Mar 2015 B2
8984637 Karecha et al. Mar 2015 B2
9009829 Stolfo et al. Apr 2015 B2
9027135 Aziz May 2015 B1
9043920 Gula et al. May 2015 B2
9117078 Chien et al. Aug 2015 B1
9141792 Baluda et al. Sep 2015 B2
9166993 Liu Oct 2015 B1
9185136 Dulkin et al. Nov 2015 B2
9197601 Pasdar Nov 2015 B2
9213838 Lu Dec 2015 B2
9225734 Hastings Dec 2015 B1
9240976 Murchison Jan 2016 B1
9246774 Mataitis et al. Jan 2016 B2
9270690 Kraitsman et al. Feb 2016 B2
9305165 Snow et al. Apr 2016 B2
9329973 Bhuyan May 2016 B2
9330259 Klein et al. May 2016 B2
9356942 Joffe May 2016 B1
9356950 Vissamsetty et al. May 2016 B2
9369476 Chekina et al. Jun 2016 B2
9386034 Cochenour Jul 2016 B2
9398001 Tidd Jul 2016 B1
9407602 Feghali et al. Jul 2016 B2
9413721 Morris et al. Aug 2016 B2
9430646 Mushtaq et al. Aug 2016 B1
9432360 Triandopoulos et al. Aug 2016 B1
9438614 Herz Sep 2016 B2
9495188 Ma et al. Nov 2016 B1
9503470 Gertner et al. Nov 2016 B2
9547516 Thakkar et al. Jan 2017 B2
9578045 Jaroch et al. Feb 2017 B2
9591006 Siva et al. Mar 2017 B2
9601000 Gruss et al. Mar 2017 B1
9602531 Wallace et al. Mar 2017 B1
9606893 Gupta et al. Mar 2017 B2
9607146 Sridhara et al. Mar 2017 B2
9609019 Vissamsetty et al. Mar 2017 B2
9628498 Aziz et al. Apr 2017 B1
9641544 Treat et al. May 2017 B1
9641550 Kraitsman et al. May 2017 B2
9710648 Weingarten Jul 2017 B2
9712547 Touboul et al. Jul 2017 B2
9769204 Vissamsetty et al. Aug 2017 B2
9781148 Mahaffey et al. Oct 2017 B2
9807092 Gutzmann Oct 2017 B1
9807115 Kolton et al. Oct 2017 B2
9813451 Honda et al. Nov 2017 B2
9871766 Syed et al. Jan 2018 B2
9877210 Hildner et al. Jan 2018 B1
9888032 Dekel et al. Feb 2018 B2
9898763 Vaynblat et al. Feb 2018 B1
9942270 Vissamsetty et al. Mar 2018 B2
10025928 Jaroch et al. Jul 2018 B2
10044675 Ettema et al. Aug 2018 B1
10102374 Cohen et al. Oct 2018 B1
10169586 Maciejak et al. Jan 2019 B2
10237282 Nelson et al. Mar 2019 B2
10250636 Vissamsetty et al. Apr 2019 B2
10257224 Jaroch et al. Apr 2019 B2
10284591 Giuliani et al. May 2019 B2
10375110 Vissamsetty et al. Jul 2019 B2
10382484 Shayevitz et al. Aug 2019 B2
10476891 Vissamsetty et al. Nov 2019 B2
10509905 Vissamsetty et al. Dec 2019 B2
10542044 Vissamsetty et al. Jan 2020 B2
10567431 Vissamsetty et al. Jan 2020 B2
10574698 Sharifi Feb 2020 B1
10599842 Vissamsetty et al. Mar 2020 B2
10599844 Schmidtler et al. Mar 2020 B2
10609074 Vissamsetty et al. Mar 2020 B2
10757090 Kahol et al. Aug 2020 B2
10826941 Jain et al. Nov 2020 B2
10855671 Kahol et al. Dec 2020 B2
10938854 Strogov et al. Mar 2021 B2
11032301 Mandrychenko et al. Jun 2021 B2
11038658 Vissamsetty et al. Jun 2021 B2
11171974 Gertner et al. Nov 2021 B2
11470115 Vissamsetty et al. Oct 2022 B2
20020010800 Riley et al. Jan 2002 A1
20020016826 Johansson et al. Feb 2002 A1
20020078382 Sheikh et al. Jun 2002 A1
20020093917 Knobbe et al. Jul 2002 A1
20020095607 Lin-Hendel Jul 2002 A1
20020178374 Swimmer et al. Nov 2002 A1
20020194489 Almogy et al. Dec 2002 A1
20030065950 Yarborough Apr 2003 A1
20030145225 Bruton, III et al. Jul 2003 A1
20030145226 Bruton, III et al. Jul 2003 A1
20030152034 Zhang et al. Aug 2003 A1
20030188189 Desai et al. Oct 2003 A1
20030223367 Shay et al. Dec 2003 A1
20040083369 Erlingsson Apr 2004 A1
20040172557 Nakae et al. Sep 2004 A1
20040243699 Koclanes et al. Dec 2004 A1
20040255157 Ghanea-Hercock Dec 2004 A1
20050050353 Thiele et al. Mar 2005 A1
20050076235 Ormazabal et al. Apr 2005 A1
20050076238 Ormazabal et al. Apr 2005 A1
20050108562 Khazan et al. May 2005 A1
20050138402 Yoon et al. Jun 2005 A1
20050204157 Johnson Sep 2005 A1
20050223239 Dotan Oct 2005 A1
20050240989 Kim et al. Oct 2005 A1
20060085543 Hrastar Apr 2006 A1
20060101515 Amoroso et al. May 2006 A1
20060126522 Oh Jun 2006 A1
20060161989 Reshef et al. Jul 2006 A1
20060203774 Carrion-Rodrigo Sep 2006 A1
20060209701 Zhang et al. Sep 2006 A1
20060230129 Swami et al. Oct 2006 A1
20060236401 Fosdick Oct 2006 A1
20070022090 Graham Jan 2007 A1
20070025374 Stefan et al. Feb 2007 A1
20070067623 Ward Mar 2007 A1
20070097976 Wood et al. May 2007 A1
20070100905 Masters et al. May 2007 A1
20070101431 Clift et al. May 2007 A1
20070115993 Cohen May 2007 A1
20070143827 Nicodemus et al. Jun 2007 A1
20070143851 Nicodemus et al. Jun 2007 A1
20070177499 Gavrilescu et al. Aug 2007 A1
20070208822 Wang et al. Sep 2007 A1
20070208936 Ramos Sep 2007 A1
20070209070 Yadav Sep 2007 A1
20070226320 Hager Sep 2007 A1
20070240215 Flores et al. Oct 2007 A1
20070250930 Aziz et al. Oct 2007 A1
20070282782 Carey et al. Dec 2007 A1
20080005782 Aziz Jan 2008 A1
20080018927 Martin et al. Jan 2008 A1
20080022000 Furuya et al. Jan 2008 A1
20080034429 Schneider Feb 2008 A1
20080046989 Wahl Feb 2008 A1
20080060074 Okuyama Mar 2008 A1
20080071728 Lim Mar 2008 A1
20080082722 Savagaonkar et al. Apr 2008 A1
20080083034 Kim et al. Apr 2008 A1
20080098476 Syversen Apr 2008 A1
20080104046 Singla et al. May 2008 A1
20080127346 Oh et al. May 2008 A1
20080162397 Zaltzman Jul 2008 A1
20080168559 Touitou et al. Jul 2008 A1
20080170566 Akimoto Jul 2008 A1
20080229415 Kapoor et al. Sep 2008 A1
20090077664 Hsu et al. Mar 2009 A1
20090089040 Monastyrsky et al. Apr 2009 A1
20090104046 Martin et al. Apr 2009 A1
20090158407 Nicodemus et al. Jun 2009 A1
20090158418 Rao et al. Jun 2009 A1
20090170566 Kwon et al. Jul 2009 A1
20090199296 Xie et al. Aug 2009 A1
20090241173 Troyansky Sep 2009 A1
20090249466 Motil et al. Oct 2009 A1
20090249470 Litvin et al. Oct 2009 A1
20090254973 Kwan Oct 2009 A1
20090288158 Izatt et al. Nov 2009 A1
20090296641 Bienas et al. Dec 2009 A1
20090327688 Li et al. Dec 2009 A1
20090328196 Bovee Dec 2009 A1
20100005339 Hooks Jan 2010 A1
20100077483 Stolfo et al. Mar 2010 A1
20100122317 Yadav May 2010 A1
20100122343 Ghosh et al. May 2010 A1
20100169973 Kim et al. Jul 2010 A1
20100269175 Stolfo et al. Oct 2010 A1
20100293615 Ye Nov 2010 A1
20100299430 Powers et al. Nov 2010 A1
20110023118 Wright et al. Jan 2011 A1
20110067107 Weeks et al. Mar 2011 A1
20110078309 Bloch Mar 2011 A1
20110099633 Aziz Apr 2011 A1
20110113427 Dotan May 2011 A1
20110138456 Ormazabal et al. Jun 2011 A1
20110141937 Breslin et al. Jun 2011 A1
20110145920 Mahaffey et al. Jun 2011 A1
20110167494 Bowen Jul 2011 A1
20110178930 Scheldt et al. Jul 2011 A1
20110185430 Sallam Jul 2011 A1
20110214176 Burch et al. Sep 2011 A1
20110214182 Adams et al. Sep 2011 A1
20110219443 Hampel Sep 2011 A1
20110219449 St. Neitzel Sep 2011 A1
20110247071 Hooks et al. Oct 2011 A1
20110271341 Satish et al. Nov 2011 A1
20110288940 Hordan et al. Nov 2011 A1
20120023572 William, Jr. et al. Jan 2012 A1
20120030745 Bauer Feb 2012 A1
20120030750 Bhargava et al. Feb 2012 A1
20120079596 Thomas et al. Mar 2012 A1
20120084866 Stolfo Apr 2012 A1
20120106377 Sommers et al. May 2012 A1
20120124363 Dietrich et al. May 2012 A1
20120137342 Hartrell et al. May 2012 A1
20120137367 Dupont et al. May 2012 A1
20120144488 Sankruthi Jun 2012 A1
20120151565 Fiterman Jun 2012 A1
20120185563 Sugiyama et al. Jul 2012 A1
20120240182 Marayanaswamy et al. Sep 2012 A1
20120255003 Sallam Oct 2012 A1
20120255004 Sallam Oct 2012 A1
20120291090 Srinivasan et al. Nov 2012 A1
20120297488 Kapoor et al. Nov 2012 A1
20120324094 Wyatt et al. Dec 2012 A1
20120331553 Aziz et al. Dec 2012 A1
20130052992 Lee et al. Feb 2013 A1
20130061097 Mendel et al. Mar 2013 A1
20130080641 Lui et al. Mar 2013 A1
20130086684 Mohler Apr 2013 A1
20130091573 Herz et al. Apr 2013 A1
20130097706 Titonis et al. Apr 2013 A1
20130111547 Kraemer May 2013 A1
20130133072 Kraitsman et al. May 2013 A1
20130152200 Alme et al. Jun 2013 A1
20130167236 Sick Jun 2013 A1
20130191924 Tedesco et al. Jul 2013 A1
20130212658 Amaya et al. Aug 2013 A1
20130219217 Seren et al. Aug 2013 A1
20130231084 Raleigh Sep 2013 A1
20130242743 Thomas et al. Sep 2013 A1
20130247190 Spurlock Sep 2013 A1
20130290662 Teal Oct 2013 A1
20130290729 Pettigrew et al. Oct 2013 A1
20130291111 Zhou et al. Oct 2013 A1
20130298192 Kumar et al. Nov 2013 A1
20130298244 Kumar et al. Nov 2013 A1
20130329732 Vyas et al. Dec 2013 A1
20130333040 Diehl et al. Dec 2013 A1
20130340033 Jones Dec 2013 A1
20130346472 Wheeldon Dec 2013 A1
20130347052 Choudrie Dec 2013 A1
20140046645 White et al. Feb 2014 A1
20140053267 Klein et al. Feb 2014 A1
20140068326 Quinn Mar 2014 A1
20140068779 Tan Mar 2014 A1
20140090061 Avasarala et al. Mar 2014 A1
20140096229 Burns et al. Apr 2014 A1
20140108794 Barton et al. Apr 2014 A1
20140123280 Kedma May 2014 A1
20140137246 Baluda May 2014 A1
20140150094 Rao et al. May 2014 A1
20140157366 Ko et al. Jun 2014 A1
20140165203 Friedrichs et al. Jun 2014 A1
20140215617 Smith et al. Jul 2014 A1
20140215621 Xaypanya et al. Jul 2014 A1
20140215625 Paul et al. Jul 2014 A1
20140237562 Nandakumar Aug 2014 A1
20140237595 Sridhara et al. Aug 2014 A1
20140237599 Gertner et al. Aug 2014 A1
20140245376 Hibbert et al. Aug 2014 A1
20140250524 Meyers et al. Sep 2014 A1
20140259092 Boucher et al. Sep 2014 A1
20140282816 Xie Sep 2014 A1
20140283076 Muttik Sep 2014 A1
20140298419 Boubez et al. Oct 2014 A1
20140349611 Kant Nov 2014 A1
20150006384 Shaikh Jan 2015 A1
20150007312 Pidathala et al. Jan 2015 A1
20150013006 Shulman et al. Jan 2015 A1
20150039513 Adjaoute Feb 2015 A1
20150074810 Saher Mar 2015 A1
20150082430 Sridhara et al. Mar 2015 A1
20150089655 Choi et al. Mar 2015 A1
20150096048 Zhang et al. Apr 2015 A1
20150113616 Sampas Apr 2015 A1
20150121524 Fawaz et al. Apr 2015 A1
20150121529 Quinlan et al. Apr 2015 A1
20150128206 Haim et al. May 2015 A1
20150128246 Feghali et al. May 2015 A1
20150143496 Thomas et al. May 2015 A1
20150150125 Dulkin et al. May 2015 A1
20150150130 Fiala May 2015 A1
20150156214 Kaminsky Jun 2015 A1
20150163121 Mahaffey et al. Jun 2015 A1
20150172300 Cochenour Jun 2015 A1
20150195291 Zuk et al. Jul 2015 A1
20150199512 Kim et al. Jul 2015 A1
20150200928 Burch et al. Jul 2015 A1
20150205962 Swidowski et al. Jul 2015 A1
20150220735 Paithane et al. Aug 2015 A1
20150254161 Baril et al. Sep 2015 A1
20150257194 Cho Sep 2015 A1
20150264068 Beauchesne Sep 2015 A1
20150264077 Berger et al. Sep 2015 A1
20150268947 Ionescu Sep 2015 A1
20150271200 Brady et al. Sep 2015 A1
20150281267 Danahy et al. Oct 2015 A1
20150286820 Sridhara et al. Oct 2015 A1
20150288706 Marshall Oct 2015 A1
20150310196 Turgeman et al. Oct 2015 A1
20150326587 Vissamsetty et al. Nov 2015 A1
20150326588 Vissamsetty Nov 2015 A1
20150326592 Vissamsetty et al. Nov 2015 A1
20150326599 Vissamsetty et al. Nov 2015 A1
20150350236 Klinghofer et al. Dec 2015 A1
20150358345 Clark et al. Dec 2015 A1
20150373039 Wang Dec 2015 A1
20150381376 Wardman et al. Dec 2015 A1
20160028750 Di Pietro et al. Jan 2016 A1
20160042179 Weingarten Feb 2016 A1
20160042180 Sayre et al. Feb 2016 A1
20160055334 Herwono et al. Feb 2016 A1
20160055337 El-Moussa Feb 2016 A1
20160072838 Kolton et al. Mar 2016 A1
20160078365 Baumard Mar 2016 A1
20160080414 Kolton Mar 2016 A1
20160127352 Xu et al. May 2016 A1
20160127413 Kraitsman et al. May 2016 A1
20160142399 Pace et al. May 2016 A1
20160191554 Kaminsky Jun 2016 A1
20160212225 Smith et al. Jul 2016 A1
20160261631 Vissamsetty et al. Sep 2016 A1
20160285914 Singh et al. Sep 2016 A1
20160315909 Von Gravrock et al. Oct 2016 A1
20160323300 Boss et al. Nov 2016 A1
20160323316 Kolton et al. Nov 2016 A1
20160381023 Dulce et al. Dec 2016 A1
20170019425 Ma et al. Jan 2017 A1
20170026387 Vissamsetty et al. Jan 2017 A1
20170032122 Thakar Feb 2017 A1
20170054754 Saher et al. Feb 2017 A1
20170093910 Gukal et al. Mar 2017 A1
20170126718 Baradaran et al. May 2017 A1
20170134405 Ahmadzadeh et al. May 2017 A1
20170141980 Palanciuc et al. May 2017 A1
20170147796 Sardesai May 2017 A1
20170149787 Niemela May 2017 A1
20170149795 Day, II May 2017 A1
20170149832 Touboul et al. May 2017 A1
20170171244 Vissamsetty et al. Jun 2017 A1
20170180421 Shieh et al. Jun 2017 A1
20170206357 Gorelik et al. Jul 2017 A1
20170230384 Touboul et al. Aug 2017 A1
20170230402 Greenspan et al. Aug 2017 A1
20170235967 Ray Aug 2017 A1
20170244729 Fahmy et al. Aug 2017 A1
20170244749 Shulman et al. Aug 2017 A1
20170244755 Tsao Aug 2017 A1
20170264639 Sama et al. Sep 2017 A1
20170279846 Osterweil et al. Sep 2017 A1
20170286676 Weingarten et al. Oct 2017 A1
20170302458 Berger Oct 2017 A1
20170302653 Ortner Oct 2017 A1
20170302665 Zou et al. Oct 2017 A1
20170302696 Schutz Oct 2017 A1
20170318054 Vissamsetty et al. Nov 2017 A1
20170322959 Tidwell et al. Nov 2017 A1
20170324774 Ohayon et al. Nov 2017 A1
20170324777 Ohayon et al. Nov 2017 A1
20170331849 Yu Nov 2017 A1
20170331856 Vissamsetty et al. Nov 2017 A1
20170346802 Gruskin Nov 2017 A1
20170346853 Wyatt et al. Nov 2017 A1
20170359370 Humphries Dec 2017 A1
20180013788 Vissamsetty et al. Jan 2018 A1
20180020005 Belter et al. Jan 2018 A1
20180027006 Zimmermann Jan 2018 A1
20180027017 Touboul et al. Jan 2018 A1
20180039776 Loman et al. Feb 2018 A1
20180048665 Shulman et al. Feb 2018 A1
20180063187 St. Pierre Mar 2018 A1
20180089430 Mayo Mar 2018 A1
20180143915 Gonzalez May 2018 A1
20180146008 Vissamsetty et al. May 2018 A1
20180173876 Vissamsetty et al. Jun 2018 A1
20180183815 Enfinger Jun 2018 A1
20180248896 Challita et al. Aug 2018 A1
20190042745 Chen et al. Feb 2019 A1
20190052659 Weingarten et al. Feb 2019 A1
20190068642 Araujo et al. Feb 2019 A1
20190073475 Vissamsetty et al. Mar 2019 A1
20190114426 Cohen et al. Apr 2019 A1
20190253453 Vissamsetty et al. Aug 2019 A1
20190354355 Jacobson et al. Nov 2019 A1
20190379697 Vissamsetty et al. Dec 2019 A1
20200143054 Cohen et al. May 2020 A1
20200218806 Cho Jul 2020 A1
20200252429 Vissamsetty et al. Aug 2020 A1
20200374087 Vissamsetty et al. Nov 2020 A1
20210397710 Cohen et al. Dec 2021 A1
Foreign Referenced Citations (17)
Number Date Country
103607399 Feb 2014 CN
3 171 568 May 2017 EP
2016-512631 Apr 2016 JP
2017-504102 Feb 2017 JP
10-2015-0101811 Sep 2015 KR
10-1969572 Apr 2019 KR
WO 0227440 Apr 2002 WO
WO 2010030169 Mar 2010 WO
WO 2012027669 Mar 2012 WO
WO 2013014672 Jan 2013 WO
WO 2015171780 Nov 2015 WO
WO 2015171789 Nov 2015 WO
WO 2016024268 Feb 2016 WO
WO 2016081561 May 2016 WO
WO 2017064710 Apr 2017 WO
WO 2019092530 May 2019 WO
WO 2019032728 Mar 2020 WO
Non-Patent Literature Citations (46)
Entry
Barbhuiya et al., “Detection of neighbor solicitation and advertisement spoofing in IPv6 neighbor discovery protocol.” Proceedings of the 4th international conference on Security of information and networks. (2011).
Chen et al., “MitM attack by name collision: Cause analysis and vulnerability assessment in the new gTLD era.” Security and Privacy (SP), 2016 IEEE Symposium on. IEEE (2016).
Dini et al., “Probabilistic Contract Compliance for Mobile Applications”, Eighth International Conference on Availability, Reliability and Security (ARES) IEEE, Sep. 2-6, 2013, pp. 599-606.
Gu et al., “Malicious Shellcode Detection with Virtual Memory Snapshots,” 2010 Proceedings IEEE INFOCOM, San Diego, CA, 2010, pp. 1-9, accessed Mar. 6, 2019.
Installing a Tanium Client—distributed also using user/password (not encrypted) from a hosted Module server, downloaded from https://docs.tanium.eom/client/client/deployment.html#client_management on Apr. 1, 2021.
IBM Software, “Stepping up the battle against advanced threats”, IBM Trusteer Apex Thought Leadership White Paper, Dec. 2013, WGW03043-USEN-00.
“IBM Security Qradar SIEM Installation Guide” downloaded from http://www.siem.su/docs/IBM/Installation_and_updating/IBM_Security_QRadar_installation_Guide.pdf, printed 2013, vol. 7.2 in 54 pages.
IBM Security Guardium Insights for IBM Cloud Park for Security, printed from https://www.IBM.com/downloads/cas/QY1RYRLP, printed May 26, 2021 in 12 pages.
IBM Security Qradar Solutuion Brief “Visibility, detection, investigation and response” printed from https://www.ibm.com/downloads/cas/OP62GKAR, printed on May 26, 2021 in 11 pgs.
“IBM Security Trusteer Apex Advanced malware Protection (SaaS) delivers an enterprise endpoint security solution to help protect organizations from advanced malware and targeted attacks” downloaded from https://www-01 .ibm.com/common/ssi/rep_ca/0/877/ENUSZP14-0390/ENUSZP14-0390.PDF, printed Aug. 5, 2014, in 9 pgs.
IBM Guardium Documentation “Components and Topology”, downloaded from https://www.ibm.com/docs/en/qsip/7.4?topic=deployment-qradar-architecture-overview, printed on Jun. 8, 2021 in 3 pages.
IBM Software “Stepping up the battle against advanced threats”, downloaded from https://citrixready.citrix.com/content/dam/ready/partners/ib/ibm-global-services/ibm-security-trusteer-apex-advanced-malware-protection/wgw03043usen.pdf, Printed Dec. 2013 in 12 pages.
Laureano et al., M., “Intrusion detection in virtual machine environments. In Euromicro Conference, 2004. Proceedings.” 30th (pp. 520-525). IEEE Sep. 30, 2004.
Liu et al., “Detecting Trojan Horses Based on System Behavior Using Machine Learning Method”, International Conference on Machine Learning and Cybernetics (ICMLC), IEEE, Jul. 11-14, 2010, pp. 855-860.
IBM, “Qradar Architecture overview”, downloaded from https://www.ibm.com/docs/en/qsip/7.4?topic=deployment-qradar-architecture-overview printed May 28, 2021 in 6 pgs.
Moussaileb et al., “Ransomware's Early Mitigation Mechanisms,” Proceedings ARES 2018 Proceedings of the 13th International Conference on Availability, Reliability and Security Article No. 2, pp. 1-10 (2018).
Ornaghi et al., “Man in the middle attacks.” Blackhat Conference Europe (2003).
Ramachandran et al., “Detecting ARP spoofing: An active technique.” International Conference on Information Systems Security, Springer, Berlin, Heidelberg (2005).
Rüdiger Schollmeier, A Definition of Peer-to-Peer Networking for the Classification of Peer-to-Peer Architectures and Applications, Proceedings of the First International Conference on Peer-to-Peer Computing, IEEE (2002).
Shosha et al., A.F., “Evasion-resistant malware signature based on profiling kernel data structure objects.” In 2012 7th International Conference on Risks and Security of Internet and Systems (CRISIS) (pp. 1-8). IEEE., Oct. 31, 2012.
Tanium™ Client Management 1.6 User Guide, downloaded from https://docs.tanium.com/client_management/client_management/index.html on Apr. 1, 2021.
Ullrich et al., “IPv6 Security: Attacks and Countermeasures in a Nutshell.” 8th USENIX Workshop on Offensive Technologies (2015).
Xu, J-Y; Sung, A.H.; Chavez, P.; Mukkamala, S.; “Polymorphic Malicious Executable Scanner by API Sequence Analysis”, Fourth International Conference on Hybrid Intelligent Systems, IEEE Dec. 5-8, 2004, pp. 378-383.
Communication Pursuant to Article 94(3) EPC dated Dec. 11, 2018 for European Application 15 760 520.5, in 7 pages.
Extended European Search Report dated May 9, 2019 for Application No. 16855062.2, in 7 pages.
European Search Report dated Apr. 29, 2021 in European Patent Application No. 18844671 in 38 pages.
Extended European Search Report dated Jan. 25, 2021 for European Patent Application No. 20181537.0, in 10 pages.
Reaqta Hive, A.I. Based Endpoint Threat Response, Whitepaper, 27 pages (Apr. 6, 2017).
International Search Report and Written Opinion dated Feb. 18, 2016 for International Application No. PCT/IL2015/050802, in 10 pages.
International Preliminary Report on Patentability dated Feb. 14, 2017 for International Application No. PCT/IL2015/050802, in 7 pages.
International Search Report and Written Opinion dated Apr. 20, 2017 for International Application No. PCT/IL2016/051110, in 10 pages.
International Search Report and Written Opinion dated Dec. 11, 2018 for International Application No. PCT/US2018/045850, in 12 pages.
International Search Report and Written Opinion dated Aug. 24, 2020 for International Application No. PCT/US2020/033872, in 8 pages.
International Search Report issued in application No. PCT/US2021/050129 dated Dec. 21, 2021.
International Search Report in corresponding International Patent Application No. PCT/US2015/29490, dated Aug. 7, 2015, in 2 pages.
Written Opinion in corresponding International Patent Application No. PCT/US2015/29490, dated Aug. 7, 2015, in 6 pages.
International Preliminary Reporton Patentability in corresponding International Patent Application No. PCT/US2015/29490, dated Nov. 8, 2016, in 7 pages.
International Search Report in corresponding International Patent Application No. PCT/US2015/29501, dated Aug. 7, 2015, in 2 pages.
Written Opinion in corresponding International Patent Application No. PCT/US2015/29501, dated Aug. 7, 2015, in 6 pages.
International Preliminary Reporton Patentability in corresponding International Patent Application No. PCT/US2015/29501, dated Nov. 8, 2016, in 7 pages.
International Search Report in corresponding International Patent Application No. PCT/US2015/61271, dated Feb. 2, 2016, in 2 pages.
Written Opinion in corresponding International Patent Application No. PCT/US2015/61271, dated Feb. 2, 2016, in 6 pages.
International Preliminary Reporton Patentability in corresponding International Patent Application No. PCT/US2015/61271, dated May 23, 2017, in 7 pages.
Office Action dated Jul. 24, 2019 in European Patent Application No. 15760520.5, in 8 pages.
Office Action dated May 31, 2022 in Japanese Patent Application No. 2020-503272, in 7 pages.
Extended European Search Report dated Aug. 25, 2021 for European Patent Application No. 21162973.8, in 5 pages.
Related Publications (1)
Number Date Country
20200252429 A1 Aug 2020 US
Continuation in Parts (2)
Number Date Country
Parent 16543189 Aug 2019 US
Child 16849813 US
Parent 15383522 Dec 2016 US
Child 16543189 US