The present invention relates to systems, computer-implemented methods and computer program products for performing adaptive autonomic threat detection and quarantine of detected threats.
In a network computing environment, connected devices that are infected by worms, Trojans, spyware and other forms of malware impose burdens on network resources and create additional challenges for network administrators. For example, certain malware may pose a nuisance to the computing environment, e.g., by causing system slowdowns. Other forms of malware may lead to the loss or corruption of information stored within the computing environment. Still further, devices that are logged into or otherwise connected to the computing environment may create problems for network administrators when controlled by malicious users with ill intent, e.g., to misappropriate information, which may be used for nefarious purposes such as identity theft.
In addition to the potential for infection of other devices connected to the computing environment, data corruption and data theft, infected devices can often consume enormous amounts of network bandwidth through port-scanning or sending mass emails. The problems created by a malware infected computing environment are magnified in the context of mobile computing since mobile computing devices often connect into an intranet through unsecured public networks and are thus potentially exposed to malicious traffic from untrusted devices on these unsecured public networks.
There are several techniques that may be used to detect infected devices and to curtail activity caused by malware. For example, devices may be required to successfully pass a validation test prior to logging into a corresponding computing environment. However, conventional validation tests typically require a lengthy validation process, e.g., by running virus scans on every connection. Validation tests are also not optimal because typical validation tests cannot respond in real-time if a device becomes compromised by malware after it has been connected to the network. Another exemplary approach to curtail activity caused by malware is to define static heuristics in the form of statistical roles and attempt to use these heuristics to detect and disconnect compromised devices in real-time. However, conventional statistical heuristics approaches rely on information determined from an analysis of previous traffic and attack patterns, thus the heuristics may be invalid for new patterns of attacks.
According to various aspects of the present invention, methods, computer program products and systems are provided for performing autonomic threat detection. Autonomic threat detection is performed by collecting traffic samples of traffic patterns associated with a device having a device resident validation module, which is connected to a network. Autonomic threat detection is further performed by using a threat analysis system to recognize a pattern of traffic determined to be indicative of a compromised device, based at least in part upon the traffic samples. If the samples indicate a compromised device, the device is quarantined and a security check is performed on the device.
If the device has been quarantined, the security check is performed by requesting data from the corresponding device resident validation module to determine if the device is compromised, analyzing data from the device resident validation module of the quarantined device and taking an action based upon analysis of the data. Autonomic threat detection is further performed by utilizing at least one of the data from the device resident validation module of the quarantined device or the traffic samples to autonomically train the threat analysis system to identify compromised devices.
According to various aspects of the present invention, autonomic adaptive threat detection and quarantine of detected threats is implemented in a network computing environment. A threat detection system analyzes network connections to detect connected devices that may be compromised, e.g., devices that are believed to be infected by malware or devices that are believed to be used for malicious or otherwise inappropriate purposes. The threat detection system quarantines devices believed to be compromised, then resolves whether the quarantined device is actually a threat through interaction with a device resident security validation component provided on each quarantined device. Based upon the interaction with quarantined devices, the threat detection system autonomically trains itself to adapt to changing threat conditions, as will be described in greater detail herein.
The threat detection system according to various aspects of the present invention provides a layer of protection for the network environment against compromised devices and may be used in conjunction with, or as an alternative to conventional threat detection methods.
Referring now to the drawings and particularly to
The network 104 provides communications links between the various processing devices 102, and may be supported by networking components 106 that interconnect the processing devices 102, including for example, routers, hubs, firewalls, network interfaces wired or wireless communications links and corresponding interconnections. Moreover, the network 104 may comprise connections using one or more intranets, extranets, local area networks (LAN), wide area networks (WAN), wireless networks (WIFI), the internet, including the world wide web, and/or other arrangements for enabling communication between the processing devices 102, in either real time or otherwise, e.g., via time shifting, batch processing, etc. The network 104 is shown by way of illustration, and not by way of limitation, as a computing environment in which various aspects of the present invention may be practiced.
It is possible that malicious attacks on a network processing system may occur. For example, a network server or other processing device 102 may become infected by malware, which can potentially spread throughout the network, causing a nuisance such as slow downs, spurious email generation and other undesired network activity and/or loss of data. Further, a malicious attack may take the form of a user that is accessing a networked processing device 102 for malicious or otherwise inappropriate purposes, e.g., to commit nefarious acts such as identity theft. As such, threat detection and quarantine of processing devices 102 that are believed to be threats to the network computing environment 100 may be performed.
Referring to
A decision is made at 114 if a monitored device is believed to pose a potential threat to the network computing environment or components thereof. If no potential threat is suspected at 114, monitoring of the connected device(s) continues. If a potential threat is suspected in a connected device at 114, then that connected device that is believed to be a possible threat is placed into quarantine at 116.
A security check is performed at 118 on the quarantined device and an appropriate action is taken as a result of the security check at 120. For example, if the security check results in a conclusion that the quarantined device is not a security risk, then the device may be removed from quarantine. If the quarantined device is determined to be a risk to the network or components thereof, appropriate security measures are taken, examples of which will be described in greater detail herein. The results of the security check are also utilized to adaptively improve the analysis capabilities of monitoring for threats at 122 as will be described in greater detail herein.
Referring to
The authentication/access control component 132 may comprise, for example, a VPN server or other access point to the network computing environment that provides the necessary log-on, user authentication and/or access control as required by the particular implementation of the network computing environment. The authentication/access control component 132 further includes a restriction module 136 to restrict traffic from a specific device, such as by using network filters or other exclusionary techniques as will be described in greater detail herein. The threat analysis component 134 monitors each processing device 102, further designated by the reference numeral 102B, which is connected to the network through the network server side authentication/access control component 132, in real-time. As used herein, the term “real-time” should be interpreted in a practical manner to include substantially near real time, e.g., to account for reasonable propagation and communication delays, threat analysis processing delays, etc.
If a connected processing device 102B is believed to be compromised, the authentication/access control component 132 and the threat analysis component 134 work together to interact with a device resident security validation component 138 provided with the corresponding connected processing device 102B, to implement autonomic threat detection and quarantine of detected threats. In an exemplary implementation, upon suspecting a compromised device, the threat analysis component 134 instructs the authentication/access control component 132 to quarantine the suspected device.
As shown, connected and monitored processing devices that are quarantined are further designated by the reference 102C to differentiate these devices from non-quarantined processing devices 102B. In a quarantined state, traffic to and from a given quarantined processing device 102C is restricted. For example, upon receiving a request from the threat analysis component 134 to quarantine a connected processing device, the authentication/access control component 132 uses its restriction module 136 to prevent the quarantined processing device 102C from communicating with devices 102 and system components 139, which may comprise, devices, processes, servers, databases, storage retrieval systems and other networked components, features, services, etc., which are deployed or otherwise connected to the network computing environment 100.
If a processing device has been quarantined, the associated device resident security validation component 138 performs a security check on its corresponding quarantined device 102C and the results of the security check are fed back into the threat analysis component 134. According to an aspect of the present invention, the device resident security validation component 138 comprises a tamper-resistant, device resident program that has the ability to run a suite of validation tests on the quarantined processing device 102C, which might include such things as a virus scan, verification of firewall rules, and analysis of patch levels.
Based upon the results of the security audit, an action is taken. For example, if the quarantined processing device 102C is found not to be a threat, the device can be released from quarantine, e.g., by turning off the filter or other control component of the restriction module 136. If the quarantined processing device 102C is found to be a threat, the authentication/access control component 132 may disconnect the corresponding quarantined processing device 102C from the network. Another exemplary action may be to attempt to correct, repair, clean or otherwise remedy the quarantined processing device 102C that is believed to be compromised, e.g., by using a suitable virus/repair technique to clean the device. Any other number of reasonable actions may further be taken, in addition to or in lieu of the examples given above. Still further, the results of the security check are used to adaptively improve the analysis capabilities of the threat analysis component 134.
According to various aspects of the present invention, the system can dynamically detect whether a connected processing device 102B has become compromised and block it from accessing other network devices 102 and/or system components 139 in real-time, thus providing enhanced capabilities over the “scan at connect” types of conventional validation. Moreover, according to various aspects of the present invention, the system 130 does not rely solely upon filters making use of heuristic rules because the system can adapt to different types of threats without the intervention of the operator as will be described in greater detail below.
Referring to
Referring to
Referring to
The device resident security validation component 138 sends the audit results to the threat detection system at 178. Based on the results from the device resident security validation component 138, the threat detection system 130 sends a message to the authentication/access control component 132 telling it to either lift the restrictions on the corresponding quarantined processing device 102C, or to take some other action on the quarantined processing device 102C, such as by disconnecting the quarantined processing device 102C from the network, etc. When the quarantined processing device 102C disconnects, a stop event may be sent from the authentication/access control component 132 to the threat analysis component 134. Other exemplary actions may comprise cleaning or otherwise removing a virus from the infected device or taking other appropriate action to remove detected malware, etc., as described in greater detail herein.
The threat detection system 130 now has empirical evidence from the device resident security validation component 138 as to whether its classification of the quarantined processing device 102 as being “compromised” was correct. This feedback may be processed to improve a classification algorithm used to identify connected processing devices 102B that may be suspected of being compromised.
According to an aspect of the present invention, the threat analysis component 134 may make use of the traffic patterns observed during the “uninfected” portions of device connections in combination with the audit data from device resident security validation components 138 of corresponding quarantined processing devices 102C to improve its analysis capabilities. For example, the threat analysis component 134 may collect historical information about the traffic patterns of connected devices 102B. Historical information may be collected by observing each connection and maintaining a sliding window that contains data for a discrete unit of time. At a fixed interval, each window is updated and then analyzed, e.g., using neural network techniques such as a Hopfield network or self-organizing map, or by using Bayesian and/or other statistical techniques.
Referring to
Further, for devices which are placed into the compromised group, i.e., quarantined processing devices 102C, the threat analysis system may enter a training mode at 188 in which the threat detection system 130 combines the feedback from a corresponding device resident security validation component 138, e.g., as obtained during the quarantine process, with the data in one or more traffic samples as input to adjust the connection weights in its neural network at 190. This enables the threat detection system 130 to “learn” from the traffic patterns of the quarantined device. The threat detection system 130 can also use other parameters, such as traffic patterns of other non quarantined devices to adjust the weights in its neural network. Still further, the neural network can learn in other additional ways.
The autonomic learning of the threat detection system 130 may be augmented by operator assisted learning. For example, a network administrator may manually flag devices or patterns which should (or should not) be quarantined. Thus, for example, the autonomic learning capability of the threat detection system does not need to start off without a base of knowledge, but rather may be seeded with log data that has been analyzed by a development lab or is otherwise known to a network administrator.
According to aspects of the present invention, the threat detection system has the ability to dynamically adjust the threat detection criteria based on observed changes in the behavior of connected devices. That is, the baseline behaviors that are being compared against by the threat detection system 130 can be changed over time by “training” the artificial neural network model using data about the behaviors and threat classification of connected devices. Thus, static heuristic rules or other static, operator determined rules, which are based on information about previous suspicious behavior, which may be used to “pre-program” or otherwise augment the threat detection system knowledge can be autonomically changed, revised, expanded, discarded, etc.
In this way, training data is collected so that the behavior of the classification system, e.g., the neural network, used by the threat detection system 130 can be updated without requiring a human operator to actually analyze the historical information by hand and try to develop heuristics. The overall system allows the neural network to do this itself, although the networks own learning could also be augmented with human analyzed training data as noted above.
In practice, there may be multiple access points to a particular implementation of a network computing environment. As such, the threat detection system 130 may monitor multiple access points, e.g., by implementing multiple authentication/access control components 132 or multiple instances of the threat detection system 130 may be deployed across the network 104 to provide threat mitigation at desired points of access to the network.
Referring to
Also connected to the I/O bus may be devices such as a graphics adapter 216, storage 218 and a computer usable storage medium 220 having computer usable program code embodied thereon. The computer usable program code may be executed to execute any aspect of the present invention, for example, to implement any aspect of any of the methods illustrated in FIGS. 2 and 4-7. In the case of a network side server, the computer usable program code may be executed to implement the authentication/access control component 132, including the restriction module 136. In the case of a network side server, the computer usable program code may also be executed to implement the threat analysis component 134 including any required threat analysis engine features, such as the artificial neural network and other processes performed thereby. In the case of a connected device, the computer usable program code may be executed to implement the device resident security validation component 138, etc.
The data processing system depicted in
The various aspects of the present invention may be embodied as systems, computer-implemented methods and computer program products. Also, various aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including software, firmware, micro-code, etc.) or an embodiment combining software and hardware, wherein the embodiment or aspects thereof may be generally referred to as a “circuit,” “component” or “system.” Furthermore, the various aspects of the present invention may take the form of a computer program product on a computer-usable storage medium having computer-usable program code embodied in the medium or a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system.
The software aspects of the present invention may be stored, implemented and/or distributed on any suitable computer usable or computer readable medium(s). For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer program product aspects of the present invention may have computer usable or computer readable program code portions thereof which are stored together or distributed, either spatially or temporally across one or more devices. A computer-usable or computer-readable medium may comprise, for example, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. As yet further examples, a computer usable or computer readable medium may comprise cache or other memory in a network processing device or group of networked processing devices such that one or more processing devices stores at least a portion of the computer program product. The computer-usable or computer-readable medium may also comprise a computer network itself as the computer program product moves from buffer to buffer propagating through the network. As such, any physical memory associated with part of a network or network component can constitute a computer readable medium.
More specific examples of the computer usable or computer readable medium comprise for example, a semiconductor or solid state memory, magnetic tape, an electrical connection having one or more wires, a swappable intermediate storage medium such as floppy drive or other removable computer diskette, tape drive, external hard drive, a portable computer diskette, a hard disk, a rigid magnetic disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), a read/write (CD-R/W) or digital video disk (DVD), an optical fiber, disk or storage device, or a transmission media such as those supporting the Internet or an intranet. The computer-usable or computer-readable medium may also comprise paper or another suitable medium upon which the program is printed or otherwise encoded, as the program can be captured, for example, via optical scanning of the program on the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory. The computer-usable medium may include a propagated data signal with the computer-usable program code embodied therewith, either in baseband or as part of a carrier wave or a carrier signal. The computer usable program code may also be transmitted using any appropriate medium, including but not limited to the Internet, wire line, wireless, optical fiber cable, RF, etc.
Computer program code for carrying out operations of the present invention may be written in any suitable language, including for example, an object oriented programming language such as Java, Smalltalk, C++ or the like. The computer program code for carrying out operations of the present invention may also be written in conventional procedural programming languages, such as the “C” programming language, or in higher or lower level programming languages. The program code may execute entirely on a single processing device, partly on one or more different processing devices, as a stand-alone software package or as part of a larger system, partly on a local processing device and partly on a remote processing device or entirely on the remote processing device. In the latter scenario, the remote processing device may be connected to the local processing device through a network such as a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external processing device, for example, through the Internet using an Internet Service Provider.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus systems and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams may be implemented by system components or computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
The present invention may be practiced on any form of computer system, including a stand alone computer or one or more processors participating on a distributed network of computers. Thus, computer systems programmed with instructions embodying the methods and/or systems disclosed herein, or computer systems programmed to perform various aspects of the present invention and storage or storing media that store computer readable instructions for converting a general purpose computer into a system based upon the various aspects of the present invention disclosed herein, are also considered to be within the scope of the present invention. Once a computer is programmed to implement the various aspects of the present invention, including the methods of use as set out herein, such computer in effect, becomes a special purpose computer particular to the methods and program structures of this invention. The techniques necessary for this are well known to those skilled in the art of computer systems.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, one or more blocks in the flowchart or block diagrams may represent a component, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently or in the reverse order.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention.
Having thus described the invention of the present application in detail and by reference to embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the invention defined in the appended claims.