1. Field of the Invention
The present invention relates generally to network security and more particularly to detecting malicious software operating in computers and other digital devices.
2. Related Art
Malicious software, or malware for short, may include any program or file that is harmful by design to a computer. Malware includes computer viruses, worms, Trojan horses, adware, spyware, and any programming that gathers information about a computer or its user or otherwise operates without permission. The owners of the computers are often unaware that these programs have been added to their computers and are often similarly unaware of their function.
Malicious network content is a type of malware distributed over a network via websites, e.g., servers operating on a network according to an HTTP standard or other well-known standard. Malicious network content distributed in this manner may be actively downloaded and installed on a computer, without the approval or knowledge of its user, simply by the computer accessing the web site hosting the malicious network content (the “malicious web site”). Malicious network content may be embedded within objects associated with web pages hosted by the malicious web site. Malicious network content may also enter a computer on receipt or opening of email. For example, email may contain an attachment, such as a PDF document, with embedded malicious executable programs. Furthermore, malicious content may exist in files contained in a computer memory or storage device, having infected those files through any of a variety of attack vectors.
Various processes and devices have been employed to prevent the problems associated with malicious content. For example, computers often run antivirus scanning software that scans a particular computer for viruses and other forms of malware. The scanning typically involves automatic detection of a match between content stored on the computer (or attached media) and a library or database of signatures of known malware. The scanning may be initiated manually or based on a schedule specified by a user or system administrator associated with the particular computer. Unfortunately, by the time malware is detected by the scanning software, some damage on the computer or loss of privacy may have already occurred, and the malware may have propagated from the infected computer to other computers. Additionally, it may take days or weeks for new signatures to be manually created, the scanning signature library updated and received for use by the scanning software, and the new signatures employed in new scans.
Moreover, anti-virus scanning utilities may have limited effectiveness to protect against all exploits by polymorphic malware. Polymorphic malware has the capability to mutate to defeat the signature match process while keeping its original malicious capabilities intact. Signatures generated to identify one form of a polymorphic virus may not match against a mutated form. Thus polymorphic malware is often referred to as a family of virus rather than a single virus, and improved anti-virus techniques to identify such malware families is desirable.
Another type of malware detection solution employs virtual environments to replay content within a sandbox established by virtual machines (VMs). Such solutions monitor the behavior of content during execution to detect anomalies that may signal the presence of malware. One such system offered by FireEye, Inc., the assignee of the present patent application, employs a two-phase malware detection approach to detect malware contained in network traffic monitored in real-time. In a first or “static” phase, a heuristic is applied to network traffic to identify and filter packets that appear suspicious in that they exhibit characteristics associated with malware. In a second or “dynamic” phase, the suspicious packets (and typically only the suspicious packets) are replayed within one or more virtual machines. For example, if a user is trying to download a file over a network, the file is extracted from the network traffic and analyzed in the virtual machine. The results of the analysis aids in determining whether the file is malicious. The two-phase malware detection solution may detect numerous types of malware and, even malware missed by other commercially available approaches. Through verification, the two-phase malware detection solution may also achieve a significant reduction of false positives relative to such other commercially available approaches. Dealing with false positives in malware detection may needlessly slow or interfere with download of network content or receipt of email, for example. This two-phase approach has even proven successful against many types of polymorphic malware and other forms of advanced persistent threats.
Further enhancement to malware detection effectiveness while avoiding false positives is desirable of course, particularly as malware developers continue to create new forms of exploits.
Embodiments of the invention may automatically detect a type of malicious software running in a computer or other digital device called by the coined term “instruction backdoor.” An instruction backdoor is malware that opens a communications channel, preferably, a communication port for listening for instructions or other communications from an external malicious computer system or program. Characteristically, the communications are effected in a clandestine fashion. For example, such instructions may trigger activation of or otherwise control a Trojan Horse or other “sleeper” virus, or otherwise direct activity of malicious code that has previously gained entry into the digital device through other means.
In some embodiments, detection of an instruction backdoor may be carried out within a virtual machine that mimics a real digital device and, in a controlled and monitored environment, permits content (whether network downloads, email, files or other objects), when loaded and executed therein, to open one or more local communication ports to listen for instructions from a supposed (decoy) external server or program. When those communication ports are not the authorized ones used for communications by programs installed on the digital device by its user or administrator, an instruction backdoor may be discovered. The inventors recognized that key to detecting instruction backdoors are “listened to” ports rather than ports involved in communication connections. Embodiments may monitor “listened to” ports in digital devices to detect instruction backdoors during normal operation or during diagnostic or forensic testing of the digital device. Upon detecting an instruction backdoor, embodiments of the invention may issue an alert, report the incident as malware, isolate the affected digital device from network communication, or take other steps to remediate.
Using backdoors of the type described herein as a keystone of malicious attack and exploit analysis may permit embodiments of the invention to detect disparate forms of malware, and even families of polymorphic virus that use the same communication mechanisms to obtain instructions and other communications in furtherance of their nefarious purposes.
The invention will be more fully understood with reference to the following detailed description in conjunction with the drawings, of which:
Generally speaking, a backdoor is a type of (or part of) an active infiltration attack that uses an entry point into a digital device, whether an undocumented entry point or subverted use of a documented entry point. In so doing, the backdoor bypasses normal security and authentication mechanisms, and thereby achieves unauthorized, often remote access to the computer. A hallmark of backdoors is the manner by which they provide access is designed to remain undetected to the digital device's user and system/network administrators.
Generally herein, “instruction backdoor” will be used in the foregoing sense of being malware itself; it may also be viewed as denoting the instrumentality or mechanism associated with an attack, such as a communications channel or port, in which case detection of the instruction backdoor is tantamount to detecting an attack, leading to the conclusion that malware is present. This later sense should be regarded as merely a different perspective of the same concept, as will be apparent to those of skill in the art in light of this specification. Also, while this specification often states that content comprising a malware candidate or test sample opens an instruction backdoor, it will be understood that this phrasing is intended to encompass the typical situation where a program processing content may open an instruction backdoor as a result of malicious code within the content. After all, it may be the program that sends system calls to open a port for listening, even a black port.
As used herein, a “digital device” may include any form or type of computer (such as a client, server, desk top, or notebook), mobile device (such as a smart phone or tablet) or other device, apparatus or system equipped for communications over a network (such as the Internet or an intranet). The terms “digital device” and “computer” or “computer system” may be used interchangeably herein, unless the context requires otherwise.
An instruction backdoor may be designed to use an otherwise innocuous local entry point on an affected digital device and subvert it for its own malicious purposes, often targeting a personal computer running Microsoft® Windows and Microsoft® Outlook. The entry point may allow instructions to be received from the network by malicious code executing on the digital device, which directs the malicious code for example to take surreptitious control of the digital device or a program installed thereon, or may install a new, unauthorized program. For example, the malware may be directed by a malicious external server to take control of electronic mail to spam junk emails from the infected computer, or to install and use data-gathering agents to steal sensitive personal, proprietary or even classified military or other governmental data.
Network content may include any data transmitted over a network. Network data may include text, software, images, audio, or other digital data. An example of network content includes web content, or any network data that may be transmitted using a Hypertext Transfer Protocol (HTTP), HyperText Markup Language (HTML) protocol, other data transfer protocols, such as File Transfer Protocol (FTP), or be transmitted in a manner suitable for display on a web browser software application. Other examples of network content include email messages and Instant Messages. Network content may include one or more objects encapsulated in packets for communication. The packets may include a network header carrying information used for addressing, routing, and otherwise delivering the network content, and a data payload carrying the network content.
Discovery of an instruction backdoor in accordance with embodiments of the invention may involve execution of content constituting a malware candidate within a virtual environment. An illustrative technique may inspect a malware candidate and, in response, use a virtual machine configured to mimic a target computer or other digital device having a pre-determined software profile to detect the instruction backdoor. The software profile specifies at least one program and an operating system to be executed by the digital device. The malware candidate is then executed or replayed in the virtual machine, which may execute the specified program(s) and operating system in a virtualized environment and monitor the behavior of the program(s) during execution. Important to embodiments of the invention is that the monitored behavior may include attempts by any of the program(s) to open and listen to one or more ports provided by the virtual machine. As used herein, the terms computer program and application may be used interchangeably unless the context requires otherwise.
More specifically, embodiments of the invention employ a backdoor detection analyzer to identify a port number associated with each computer port at which the malware candidate is listening, if any, during its execution in the virtual machine and, in some embodiments, for purposes hereof, associated with only the “listened to” ports. Embodiments of the invention may detect attempts to open and listen to ports by intercepting system calls issued by the program to bind and listen to specific ports, or, alternatively, by reading data providing port number-process identifier (ID) associations stored by the operating system's network service module. It should be recalled that a process in this sense is an instance of a computer program being executed. The backdoor analyzer may then determine whether the identified port number of the port at which the malware candidate is listening is a “legal” or “white” port. This may be achieved through a look-up in a blacklist repository of ports that are not legal (i.e., “black ports”) or, alternatively, in both a blacklist repository of black ports and a whitelist repository of white ports. The whitelist repository may store port numbers of white ports, which, in alternative embodiments, may include (a) ports that program(s) specified by the software profile may use for listening in accordance with their respective, associated protocols and applicable standards and specifications, or (b) ports that programs that generally run on digital devices of any relevant type or of a specific type (e.g., tablet computer or desktop) are authorized to use for listening. The blacklist repository may store port numbers of black ports, which, in alternative embodiments, may include (a) ports other than white ports, (b) ports that are opened for listening by known malware or a specific type of malware that the system is to detect, or (c) ports that satisfy both (a) and (b). “Known” malware is intended to encompass content that is verified to contain malware, such as through a verification process provided by embodiments of the invention described herein or by other reliable methodologies.
In some embodiments, if the malware candidate listens at a port whose port number is on the blacklist, the malware candidate is associated with a high probability of containing malware. In other embodiments, if the content listens at a port whose port number is on the blacklist and not on the whitelist, the malware candidate is associated with a high probability of containing malware. The latter embodiments provide a level of additional protection, since it may be deemed an error to have the port number on both the blacklist and whitelist, and its inclusion on both lists may require further review and testing. When a malware candidate is considered to have a high probability of being malware, it may be recommended that the content be treated as being malware for practical purposes to contain or eliminate the threat that it poses to the network, devices and/or data.
Embodiments of the invention may monitor communications channels in digital devices to detect instruction backdoors during normal operation or during diagnostic or testing. Utilities running in the background may monitor all ports during execution of programs, and perform look-ups in the appropriate repositories either in real time or off-line to detect instruction backdoors. The embodiments described herein may also be put into practice at malware forensic labs, whose purpose is to detect, verify and study malware candidates for any reason. Occasionally, the reason may involve prior, non-conclusive testing of the content by any of various known malware detection systems, including the commercially available systems and utilities described hereinabove. Another reason may involve analysis of malware candidates to identify a root cause of seemingly malicious harm suffered by one or more digital devices.
Accordingly, sources of content comprising malware candidates may be of various types, such as web pages, emails and files, as typically received by or stored in a digital device. The term, “sources,” should be broadly construed. The content may be delivered to the backdoor detection system on a storage medium or may constitute portions of network traffic that underwent malware detection previously and require further analysis.
Embodiments of the backdoor detection system may be deployed as a stand-alone system or integrated into malware detection systems that examine and analyze other additional characteristics or attributes of the content that may indicate the presence of malware, such as, for example, malware signature scanning utilities or the afore-mentioned two-phase malware detection solutions. For example, the instruction backdoor detection mechanism described herein can be added to appliances and other systems for detecting malware, either by adding a virtual machine execution capability to such systems or by adding the appropriate aspects regarding port detection described herein to two-phase malware detection systems that already employ virtual execution environments.
Upon detecting an instruction backdoor, embodiments of the invention may issue an alert, report the incident as malware, isolate the affected digital device from network communication, or take other steps to remediate.
Ports, Listening and Connections
Since the constructs of computer ports, listening and connections are important to the understanding of embodiments of the invention, those constructs will now be described in greater detail.
As used herein, a port serves as a communications endpoint in a computer's operating system. A port is associated with an IP address of the computer, as well as the type of protocol used for communication. A port is a program-specific software construct, in that (i) each port provided by a computer's operating system may be uniquely associated at any point of time with a specific program running on the computer (or, in other words, the port is bound to the program), and (ii) plural ports may share a single physical connection to a packet switched network, such as, for example, the Internet.
For network communication, a computer forms packets for transmission by adding a destination port number to a destination computer's IP address to form the destination address for a communications session, and the destination address is then placed in the packets' headers. The data packets are placed on the network and routed across the network to the specified IP address, and then, upon reaching the destination computer having that IP address, are further routed to the specific process identified by the destination port number.
The TCP/IP protocol, for example, uses port numbers to identify both the sending and receiving end-points, called “internal sockets.” Each side of a TCP connection has an associated, e.g., 16-bit, port number reserved by the sending or receiving program. Arriving TCP/IP data packets are identified as belonging to a specific TCP/IP connection by its sockets, that is, the combination of source address, source port, destination address, and destination port. The specifics of communications depend on the protocol, however, continuing the discussion with TCP/IP as the example, the steps involved in establishing a socket for communication between a server and client will next be described. A program running on a server creates a socket by issuing a SOCKET( ) system call to its operating system. Next, it binds the socket to an IP address using a BIND( ) system call specifying a port number on the host machine. Next, the server listens for connections with a LISTEN( ) system call. Afterwards, a program on another computer serving as a client may accept a connection with an ACCEPT( ) system call to its operating system. By doing so, the client connects with the server or, in other words, a connection is made. Thereupon, data may be sent and received between the server and client over the connection. The connection continues until closed at the end of the data transfer by exchanging control and acknowledgment packets between the server and client.
Accordingly, when a program running on a computer is waiting for another computer to connect with it, the program is said to be “listening” for connections or in a “listening state”. In other words, the program (or at least an instance of the program) attaches or binds to a port of the computer on which it is executing and waits for a connection. Once a program binds itself to a port, that port cannot be used by any other program until communication has ended and the program has disconnected. When a remote computer connects to that particular port it is said to be “establishing” a connection and initiating a communication session between the two computers. The way it does this is set forth in applicable standards and specifications governing communications in the relevant protocol.
Thousands of ports are specified by communications-related “Requests for Comment” (“RFC”), Protocol Standards, and technical specification documents of software vendors, such as Microsoft, Inc. and Oracle Corp., of which far fewer are well known and widely used to identify specific types of protocols and services on a computer. The port numbers are assigned by the Internet Assigned Numbers Authority (IANA). Examples of well-known ports include: Ports 20 & 21 for File Transfer Protocol (FTP), Port 22 for Secure Shell (SSH), Port 23 for Telnet remote login service, Port 25 for Simple Mail Transfer Protocol (SMTP), Port 40 for TCP, Port 53 for Domain Name System (DNS) service, Port 80 for Hypertext Transfer Protocol (HTTP) used in the World Wide Web, Port 110 for Post Office Protocol (POP3), Port 161 for Simple Network Management Protocol (SNMP), and Port 443 for HTTP Secure (HTTPS).
A couple examples may prove useful. Web servers typically bind to the TCP/IP authorized port 80, which is what the HTTP protocol uses by default, and then will wait and “listen” for connections from remote devices. Once connected, the web server will send the requested web pages to the remote device and, when done, disconnect the connection. On the other hand, the remote device connecting to a web server may do this in reverse. The remote web browser may pick an available port in a seemingly random fashion from a range of port numbers authorized for the TCP/IP protocol, and attempt to connect to port 80 at the IP address of the web server. When the connection is established, the web browser may send a request for a particular web page and receive the requested web page from the web server. Afterwards, the two computers may disconnect the connection. Concurrent communications from both a web server and an FTP server may be established. An FTP server is a server that allows transfer to and receipt of files from remote computers under the FTP protocol. FTP servers typically use TCP ports 20 and 21 to send and receive information; therefore, an FTP server program may bind itself to TCP ports 20 or 21 to listen and wait for connections in order to send and receive data. It follows that such communications will not normally conflict with a web server running on TCP/IP port 80.
Backdoor Detection System
The virtual execution engine 120 includes a scheduler 160, a replayer 162, a virtual switch 164, and at least one virtual machine 166. The scheduler 160 receives content (e.g., packets) from the source(s) 102, assigns an execution priority to the packets, and schedules the packets for execution by the virtual execution engine 120. The execution priority may be set automatically by default as “first in, first out” or may reflect a different order of execution indicative of a perceived seriousness of a malware issue. The priority may be set or modified by a user through a user interface such as a user interface 260 (
The scheduler 160 retrieves the packets from the repository in accordance with its schedule and provides them to the replayer 162, which sends them via the virtual switch 164 to the virtual machine 166 for execution (also known as “replay”). In some embodiments, the replayer 162 mimics the behavior of a server in transmitting them to the virtual machine 120, which mimics a client in a server-client model. There may be any number of replayers 162, each associated with and acting as a virtual server to one or more virtual machines 166. In a further embodiment, the replayer 162 may dynamically modify session variables, as appropriate, to emulate a “real” server of the content being replayed. In other embodiments, the replayer 162 furnishes the packets to a virtual machine, which simulates a stand-alone digital device, such as a server or client. The virtual switch 164 may route data packets to select ports of the virtual machine 166 that would normally be used for communication between a server and client for the applicable communication protocol, such as TCP/IP.
The virtual machine 166 includes a simulation of the digital device. In one embodiment, the scheduler 140 retrieves an instance of the virtual machine 166 from a pool of preferably concurrently existing virtual machine 166 which may be pre-configured to mimic a digital device in accordance with a software profile. The software profile specifies an operating system and at least one computer program, may be provided by the source 102 along with the content to be tested, and establishes the runtime environment for virtual execution of program(s) included in the software profile. The configured virtual machine 166 is then ready to provide virtual execution of packets.
As the virtual machine executes the packets, the backdoor detection analyzer 140 monitors its behavior closely for unauthorized activity, and in particular, for instance, for any and all black ports within the virtual machine opened by objects for listening. In one example, the backdoor analyzer 140 may, upon detecting opening of a black port for listening, transmit a command to the virtualized digital device to stop execution. However, it is preferable to allow execution to proceed so that the backdoor detection analyzer 140 can detect additional illegal or unauthorized activity, such as the opening of additional black ports, which may further evidence the characteristic behavior of malware contained in the content under test.
In some embodiments, the backdoor detection analyzer 140 monitors and analyzes the behavior of the virtual machine 166 in order to determine and identify a specific type of malware, such as malware of a specific family of polymorphic virus. The backdoor detection analyzer 140 may also pursue remediation, for example, through generating computer code configured to eliminate the malware or malicious network content or limit its propagation. In various embodiments, the backdoor detection analyzer 140 generates computer code configured to repair damage performed by malware.
The controller 200 may also have a communication network interface 240, an input/output (I/O) interface 250, and a user interface 260. The communication network interface 240 may be coupled with a communication network 272 via a communication medium 270. The communications network interface 240 may communicate with other digital devices (not shown) via the communications medium 270. The communication interface 240 may include a network tap 840 (
The I/O interface 250 may include any device that can receive input from or provide output to a user. The I/O interface 250 may include, but is not limited to, a flash drive, a compact disc (CD) drive, a digital versatile disc (DVD) drive, or other type of I/O peripheral (not separately shown). The user interface 260 may include, but is not limited to a keyboard, mouse, touchscreen, keypad, biosensor, display monitor or other human-machine interface (not separately shown) to allow a user to control the controller 200. The display monitor may include a screen on which is provided a command line interface or graphical user interface.
As noted above, the controller 200 may implement the backdoor detection system 100. In other embodiments of the invention, a number of different controllers (for example, each as illustrated and described for controller 200) may be used to implement the backdoor detection system. For example, a plurality of controllers (not separately shown) may be used to implement the virtual execution engine 120, the backdoor analyzer 140, and the event generator and reporting module 150.
Whitelist and Blacklist Generation
In alternative embodiments, the whitelist may include (a) ports that program(s) specified by the software profile are authorized to use for listening in accordance with their respective, associated protocols and applicable standards, or (b) ports that programs that generally run on digital devices of any type or of a specific type (e.g., tablet computer or desktop) are authorized to use for listening in accordance with their respective, associated protocols and applicable standards. For example, RFC 959 (File Transfer Protocol (FTP)) as promulgated by the Network Working Group, Internet Society (1985), specifies that ports 20 and 21 may be used for listening for FTP complying communications with an FTP server, and so the whitelist may include both ports 20 and 21. RFC 2476 (“Message Submission”) as promulgated by the xxx (1995), specifies that ports 25 and 587 may be used SMTP-complying email communications with an email server, and so the whitelist may include both ports 25 and 587 for programs that use SMTP for communications. RFC 2616 specifies that port 80 is used for HTTP complying accesses to web content.
As an alternative to referencing separately all those documents, a compilation of authorized ports may be had with reference to certain websites, which have already collected information about ports. For example, speedguide.net offers a website providing information about ports at http://www.speedguide.net/port.php?port=1111. An advantage of this approach is that the resulting stored whitelist may be suitable for use across a plurality of different potential malware candidates' specific software profiles.
Accordingly, in some embodiments, the generation and storage of the whitelist may be regarded as infrastructural, and malware detection may then be carried out expeditiously on more than one malware candidate without interruption to develop a new whitelist for each profile. Embodiments of the invention may alternatively employ a whitelist specific to the software profile associated with each malware candidates being evaluated. Accordingly, the whitelist may need to be changed and the repository updated for each malware candidate before testing on that candidate may be conducted. This latter approach may increase the accuracy of malware detection in embodiments that rely on the whitelist since the authorized ports in the whitelist so generated will likely be fewer in number. Accordingly, if the software profile includes a web server application, the whitelist for that software may list ports 80. Similarly, if the software profile includes an SMTP server, Microsoft® Exchange®, the whitelist for that software may list port 25. If the software profile included both a web server and an email server, then the whitelist would include 80, 25.
Finally, in step 330, the resulting whitelist port numbers are stored in data structures in a whitelist repository 340.
In step 362, the method 300 detects any ports at which the sample of malware is listening, for example, in accordance with the approaches described above in conjunction with
In step 368, if the port number is not on the whitelist, the port number of the port at which the sample of malware was listening is stored to a blacklist repository and is available thereafter for use in testing unknown (though possibly suspicious) samples, which are also herein called malware candidates. Such use may be on the any backdoor detection system, including the one performing the method 300. Where it is desired to have other backdoor detection systems avail themselves of the updated backlist, in step 362, the revised backlist (or just the updates to the blacklist) may be distributed to other backdoor detection systems, perhaps on a paid subscription basis. In that way, subscribers to the updates can remain current in recognizing non-legal ports opened for listening by verified malware for digital devices with the same or similar software profiles.
Backdoor Detection System Using Blacklist
The RSQ 430 receives or otherwise obtains a software profile with the content to be tested. The software profile may specify the operating system and one or more specific programs, including version identifiers thereof, for software intended for execution in the digital device (e.g., client device 830 of
The virtual execution engine 420 includes a VM configurator 446, a replayer 448, and one or more virtual machines (VMs) 450, all coupled for communication with one another (either directly or indirectly). The VM configurator 446 is adapted to configure one or more VMs 450 in accordance with a software profile provided by the RSQ 430, that is, at least one of the VMs will be furnished with copies of the same software programs, preferably at the same release/version level, for execution in a virtual environment. The replayer 448 may serve to provide the content to the one or more VMs as though it were a source device transmitting content over a network. Alternatively, and depending on the nature of the content, the replayer 448 may provide the content to the VMs 450 simply by loading the content as a file or other object. The latter approach is particularly useful where the content constitutes data at rest rather than data in flight.
The backdoor detection analyzer 430 includes a port detector 452, blacklist comparator 454, and, in some embodiments, a signature generator 462, all coupled for communication with one another (either directly or indirectly). The port detector 452 monitors the one or more VMs executing the content to detect any and all ports that the content has opened for listening, for example, using any of the approaches described above in conjunction with
The report module 440 includes an event/alert generator 472, an intelligence sharing module 474, a user interface 476, and an I/O interface 478, all coupled for communication with one another (either directly or indirectly). The alert/event generator 472 receives metadata regarding the newly determined malware-containing content, such as the port numbers that it attempted to open for listening, as well as its signature. The alert/event generator 462 may generate an alert or a report, to be displayed or otherwise presented on a management console displayed via the GUI 476 and/or emailed or otherwise sent to a user or administrator via the I/O interface 478. The metadata and signature can also be shared with other, often remote malware detection systems (see
The backdoor detection system 400 may be implemented by one or more controllers, each, in one embodiment, in the form of controller 200 (
In step 506, the content is executed in the virtual execution environment established by the configured virtual machine. In step 508, any ports opened for listening by the content is detected during execution of the content in the virtual machine. For this purpose, the virtual machine may be equipped with one or more virtual agents, each configured to monitor behavior and/or state of the program(s) and operating system executing within the virtual machine during processing of the malware candidate. The virtual machine may implement any of the approaches to detect “listened to” ports, as described in conjunction with
If the port number is on the blacklist, in step 514, an alert is generated to indicate that the malware candidate is associated with a high probability that it contains malware. Next, in some embodiments, in step 516, the method 500 may generate a signature for the detected malicious content, as described elsewhere herein. Also, the signature and test results may be uploaded to a malware management system or otherwise shared with other malware detection systems that can use the signatures for expedited evaluation of other content that may have signatures that match the current signature.
Backdoor Detection System Using Whitelist and Blacklist
The virtual execution engine 620 includes a VM configurator 646, a replayer 648, and one or more virtual machines (VMs) 650, all coupled for communication with one another (either directly or indirectly). The VM configurator 666 is adapted to configure one or more VMs 650 in accordance with a software profile provided by the RSQ 630. The replayer 648 serves to provide the content to the VMs 650 as though it were a source device transmitting content over a network or by loading the content as a file or other object.
The backdoor detection analyzer 630 includes a port detector 652, blacklist comparator 654, whitelist comparator 658, backdoor detection logic engine 664 and, in some embodiments, a signature generator 662, all coupled for communication with one another (either directly or indirectly). The port detector 652 monitors the one or more VMs 650 executing the content to detect any and all ports that the content opens for listening (and preferably on those “listened to” ports), and extracts the port numbers for those ports. To that end, the port detector 652 may advantageously implement any of the approaches for detecting “listened to” ports described above in conjunction with
The signature generator 656 receives the content having high probability scores of containing malware, and forms a signature from that content for use in identifying that content. The signature may be formed by applying any of numerous known hash functions to the content, for example, as will be understood to those of ordinary skill in the art. The signature may be used in anti-virus scanning utilities and other malware detection systems, for example.
In various embodiments, the report module 660 may include an event/alert generator 668, a blacklist updater 672, an intelligence sharing module 676, a user interface 676, and an I/O interface 678, all coupled for communication with one another (either directly or indirectly). The alert/event generator 668 receives metadata regarding the newly determined malware-containing content, such as the port numbers that it attempted to open for listening, as well as its signature, at least in some embodiments. The alert/event generator 668 may generate an alert or a report, to be displayed or otherwise presented on a management console (see
The backdoor detection system 600 may be implemented by one or more controllers, each, in one embodiment, in the form of controller 200 (
In step 705, the content is executed in the virtual execution environment established by the configured virtual machine. In step 708, any port opened for listening by the content (and preferably only those ports) is detected during execution of the content in the virtual machine in accordance with any of the approaches described above in conjunction with
Next, in step 715, a backdoor logic engine may combine the results from steps 710 and 712, by applying the following logic: If the port number has a match in the blacklist but not in the whitelist, in step 718, an alert is generated signaling a high probability that the content contains malware. If the port number matches neither any entry in the blacklist nor any entry in the whitelist, the port number is added to the blacklist in step 726 and the method 700 proceeds to step 718, as described above. If the part number matches entries in both the whitelist and the blacklist or matches only an entry in the whitelist but not in the blacklist, a report is generated in step 725 to the effect that additional testing is required. After step 718, in step 720, in some embodiments, a signature is generated for the content and, in step 722, the signature and test results are uploaded, both as described previously.
In some embodiments, a backdoor detection system may utilize only a whitelist of “listened to” ports generated as described above, and stored in a whitelist repository rather than a blacklist as in the foregoing embodiments. Accordingly, if content opens a port for listening that has a port number not on the whitelist, the content is deemed to have a probability score indicating a high probability of containing malware, or at least a probability score indicating that the content is suspicious and may require further evaluation.
Computer Network System
The malicious network content detection system 850 may monitor exchanges of network content (e.g., web content) in real-time rather than intercepting and holding the network content until such time as it can determine whether the network content includes malicious network content. The malicious network content detection system 825 may be configured to inspect exchanges of network content over the communication network 820, identify suspicious network content, and analyze the suspicious network content using a virtual machine to detect malicious network content. In this way, the malicious network content detection system 850 may be computationally efficient and scalable as data traffic volume and the number of computing devices communicating over the communication network 820 increase. Therefore, the malicious network content detection system 825 may not become a bottleneck in the computer network system 800.
The communication network 820 may include a public computer network such as the Internet, in which case a firewall 825 may be interposed between the communication network 820 and the client device 830. Alternatively, the communication network may be a private computer network such as a wireless telecommunication network, wide area network, or local area network, or a combination of networks. Though the communication network 820 may include any type of network and be used to communicate different types of data, communications of web data may be discussed below for purposes of example.
The malicious network content detection system 825 is shown as coupled with the network 820 by a network tap 840. The network tap 840 may include a digital network tap configured to monitor network data and provide a copy of the network data to the malicious network content detection system 825. Network data may comprise signals and data that are transmitted over the communication network 820 including data flows from the server device 810 to the client device 830. In one example, the network tap 840 monitors and copies the network data without an appreciable decline in performance of the server device 810, the client device 830, or the communication network 820. The network tap 840 may copy any portion of the network data, for example, any number of data packets from the network data. In embodiments where the malicious content detection system 850 is implemented as an dedicated appliance or a dedicated computer system, the network tap 840 may include an assembly integrated into the appliance or computer system that includes network ports, network interface card and related logic (not shown) for connecting to the communication network 860 to non-disruptively “tap” traffic thereon and provide a copy of the traffic to the heuristic module 850. In other embodiments, the network tap 840 can be integrated into a firewall, router, switch or other network device (not shown) or can be a standalone component, such as an appropriate commercially available network tap. In virtual environments, a virtual tap (vTAP) can be used to copy traffic from virtual networks.
The network tap 840 may also capture metadata from the network data. The metadata may be associated with the server device 810 and/or the client device 110. For example, the metadata may identify the server device 810 and/or the client device 830. In some embodiments, the server device 810 transmits metadata which is captured by the tap 815. In other embodiments, a heuristic module 860 (described herein) may determine the server device 810 and the client device 830 by analyzing data packets within the network data in order to generate the metadata. The term, “content,” as used herein may be construed to include the intercepted network data and the metadata unless the context requires otherwise.
The malicious network content detection system 825 may include a heuristic module 860, a heuristics database 862, a scheduler 870, a virtual machine pool 880, an analysis engine 882 and a reporting module 884. In some embodiments, the network tap 840 may be contained within the malicious network content detection system 125.
The heuristic module 860 receives the copy of the network data from the network tap 840 and applies heuristics to the data to determine if the network data might contain suspicious network content. The heuristics applied by the heuristic module 860 may be based on data and/or rules stored in the heuristics database 862. In one example, the heuristic module 860 flags network data as suspicious after applying a heuristic analysis. The suspicious network data may then be provided to the scheduler 870. In some embodiments, the suspicious network data is provided directly to the scheduler 870 with or without buffering or organizing one or more data flows.
When a characteristic of the packet, such as a sequence of characters or keyword, is identified that meets the conditions of a heuristic, a suspicious characteristic of the network content is identified. The identified characteristic may be stored for reference and analysis. In some embodiments, the entire packet may be inspected and multiple characteristics may be identified before proceeding to the next step. In some embodiments, the characteristic may be determined as a result of an analysis across multiple packets comprising the network content. A score related to a probability that the suspicious characteristic identified indicates malicious network content is determined.
The heuristic module 860 may also provide a priority level for the packet and/or the features present in the packet. The scheduler 160 may then load and configure a virtual machine from the virtual machine pool 880 in an order related to the priority level, and dispatch the virtual machine to the analysis engine 882 to process the suspicious network content.
The heuristic module 860 may provide the packet containing the suspicious network content to the scheduler 870, along with a list of the features present in the packet and the malicious probability scores associated with each of those features. Alternatively, the heuristic module 860 may provide a pointer to the packet containing the suspicious network content to the scheduler 870 such that the scheduler 870 may access the packet via a memory shared with the heuristic module 860. In another embodiment, the heuristic module 860 may provide identification information regarding the packet to the scheduler 870 such that the scheduler 870, replayer 805, or virtual machine may query the heuristic module 860 for data regarding the packet as needed.
The scheduler 160 may identify the client device 110 and retrieve a virtual machine associated with the client device 110. A virtual machine may itself be executable software that is configured to mimic the performance of a device (e.g., the client device 110). The virtual machine may be retrieved from the virtual machine pool 165. Furthermore, the scheduler 160 may identify, for example, a web browser running on the client device 110, and retrieve a virtual machine associated with the web browser.
In some embodiments, the heuristic module 860 transmits the metadata identifying the client device 830 to the scheduler 870. In other embodiments, the scheduler 870 receives one or more data packets of the network data from the heuristic module 860 and analyzes the one or more data packets to identify the client device 830. In yet other embodiments, the metadata may be received from the network tap 840.
The scheduler 870 may retrieve and configure the virtual machine to mimic the pertinent performance characteristics of the client device 830. In one example, the scheduler 870 configures the characteristics of the virtual machine to mimic only those features of the client device 830 that are affected by the network data copied by the network tap 840. The scheduler 870 may determine the features of the client device 830 that are affected by the network data by receiving and analyzing the network data from the network tap 840. Such features of the client device 830 may include ports that are to receive the network data, select device drivers that are to respond to the network data, and any other devices coupled to or contained within the client device 830 that can respond to the network data. In other embodiments, the heuristic module 860 may determine the features of the client device 830 that are affected by the network data by receiving and analyzing the network data from the network tap 840. The heuristic module 850 may then transmit the features of the client device to the scheduler 870.
The virtual machine pool 880 may be configured to store one or more virtual machines. The virtual machine pool 880 may include software and/or a storage medium capable of storing software. In one example, the virtual machine pool 880 stores a single virtual machine that can be configured by the scheduler 870 to mimic the performance of any client device 830 on the communication network 820. The virtual machine pool 880 may store any number of distinct virtual machines that can be configured to simulate the performance of a wide variety of client devices 810.
The analysis engine 882 simulates the receipt and/or display of the network content from the server device 810 after the network content is received by the client device 110 to analyze the effects of the network content upon the client device 830. The analysis engine 882 may identify the effects of malware or malicious network content by analyzing the simulation of the effects of the network content upon the client device 830 that is carried out on the virtual machine. There may be multiple analysis engines 850 to simulate multiple streams of network content. The analysis engine 882 may be configured to monitor the virtual machine for indications that the suspicious network content is in fact malicious network content. Such indications may include unusual memory accesses, unusual network transmissions, unusual changes in performance, and the like. Importantly, such indications may also include the detection of an instruction backdoor. The detection of the instruction backdoor may be used to indicate that that the network content contains malware. Indeed, in some embodiments, the conclusion that malware is present with a high probability may be predicated solely on the detection of an instruction backdoor. In other embodiments, the conclusion that malware is present with a high probability may be predicated on detecting plural indications that the suspicious network content is in fact malicious network content, including the detection of an instruction backdoor. The analysis engine 882 may flag the suspicious network content as malicious network content according to the observed behavior of the virtual machine. The reporting module 884 may issue alerts indicating the presence of malware, and using pointers and other reference information, identify the packets of the network content containing the malware. Additionally, the server device 810 may be added to a list of malicious network content providers, and future network transmissions originating from the server device 810 may be blocked from reaching their intended destinations, e.g., by firewall 825.
The computer network system may also include a further communication network 890, which couples the malicious content detection system (MCDS) 850 with one or more other MCDS, of which MCDS 892 and MCDS 894 are shown, and a management console 896. The communication network 890 may, in some embodiments, be coupled for communication with or part of network 820. The management console 896 is responsible for managing the MCDS 850, 892, 894 and providing updates to their operation systems and software programs. Also, the management console may cause malware signatures generated by any of the MCDS 850, 892, 894 to be shared with one or more of the other MCDS 850, 892, 894, for example, on a subscription basis. Moreover, the backdoor detection system as described in the foregoing embodiments may be incorporated into one or more of the MCDS 850, 892, 894, or into all of them, depending on the deployment. Also, the management console 896 itself or another dedicated computer station may incorporate the backdoor detection system in deployments where such detection is to be conducted at a centralized resource.
Further information regarding an embodiment of a malicious content detection system can be had with reference to U.S. Pat. No. 8,181,553, the disclosure of which being incorporated herein by reference in its entirety.
The embodiments discussed herein are illustrative. As these embodiments are described with reference to illustrations, various modifications or adaptations of the methods and/or specific structures described may become apparent to those skilled in the art. For example, aspects of the invention may be performed by executable software, such as a program or operating system. For example, embodiments of the invention may be implemented in an operating system instance, which monitors any and all ports that an executing program (i.e., a process) has requested be opened for listening as by, for example, tracking system calls received by the operating system instance from the program instance. If the ports are included on a blacklist available to the operating system instance, then it may report an instruction backdoor. For example, where the program instance is a browser, a web page being downloaded for display may be associated with a high probability of containing malicious code; or, where the program instance is an email application, an email or attachment to an email being processed for display may be associated with a high probability of containing malicious code. Of course, the operating system may incorporate other aspects of the invention instead of or in addition to that just described, as will be appreciated in light of the description contained in this specification. Similarly, a utility or other program may also implement the invention. Noteworthy, these embodiments need not employ a virtual environment, but rather test for an instruction backdoor during normal execution of a program within a computer system.
It should be understood that the operations performed by the above-described illustrative embodiments are purely exemplary and imply no particular order. Further, the operations can be used in any sequence when appropriate and can be partially used. With the above embodiments in mind, it should be understood that the present invention may employ various computer-implemented operations involving data stored in computer systems. These operations include physical manipulation of physical quantities. Usually, though not necessarily, these quantities take the form of electrical, magnetic, or optical signals capable of being stored, transferred, combined, compared, and otherwise manipulated.
Any of the operations described herein are useful machine operations. The present invention also relates to a device or an apparatus for performing these operations. The apparatus can be specially constructed for the required purpose, or the apparatus can be a general-purpose computer selectively activated or configured by a computer program stored in the computer. In particular, various general-purpose machines can be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations, or multiple apparatus each performing a portion of the operations. Where apparatus or components of apparatus are described herein as being coupled or connected to other apparatus or other components, the connection may be direct or indirect, unless the context requires otherwise.
The present invention can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data, which can be thereafter be read by a computer system. Examples of the computer readable medium include hard drives, flash drives, read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes, and other optical and non-optical data storage devices. The computer readable medium can also be distributed over a network-coupled computer system so that the computer readable code is stored and executed in a distributed fashion. The computer readable medium can also be distributed using a switching fabric, such as used in computer farms.
The above-described components, modules and engines may be comprised of instructions that are stored on storage media (e.g., non-transitory computer readable media). The instructions may be retrieved and executed by a processor. Some examples of instructions include software, program code, and firmware. Some examples of storage media comprise memory devices and integrated circuits. The instructions are operational when executed by the processor to direct the processor to operate in accordance with embodiments of the present invention. Those skilled in the art are familiar with instructions, processor(s), and storage media.
It will be appreciated by those of ordinary skill in the art that modifications to and variations of the above-described embodiments of a system and method of detecting backdoors and malware may be made without departing from the inventive concepts disclosed herein. Accordingly, the specification and drawings are to be regarded as illustrative rather than restrictive, and the invention should not be viewed as limited except as by the scope and spirit of the appended claims. It will be recognized that the terms “comprising,” “including,” and “having,” as used herein, are specifically intended to be read as open-ended terms of art.
Number | Name | Date | Kind |
---|---|---|---|
4292580 | Ott et al. | Sep 1981 | A |
5175732 | Hendel et al. | Dec 1992 | A |
5440723 | Arnold et al. | Aug 1995 | A |
5490249 | Miller | Feb 1996 | A |
5537540 | Miller et al. | Jul 1996 | A |
5603027 | Ohkami | Feb 1997 | A |
5657473 | Killean et al. | Aug 1997 | A |
5842002 | Schnurer et al. | Nov 1998 | A |
5978917 | Chi | Nov 1999 | A |
6088803 | Tso et al. | Jul 2000 | A |
6094677 | Capek et al. | Jul 2000 | A |
6108799 | Boulay et al. | Aug 2000 | A |
6269330 | Cidon et al. | Jul 2001 | B1 |
6272641 | Ji | Aug 2001 | B1 |
6279113 | Vaidya | Aug 2001 | B1 |
6298445 | Shostack | Oct 2001 | B1 |
6357008 | Nachenberg | Mar 2002 | B1 |
6424627 | Sorhaug et al. | Jul 2002 | B1 |
6442696 | Wray et al. | Aug 2002 | B1 |
6484315 | Ziese | Nov 2002 | B1 |
6487666 | Shanklin et al. | Nov 2002 | B1 |
6493756 | O'Brien et al. | Dec 2002 | B1 |
6550012 | Villa et al. | Apr 2003 | B1 |
6775657 | Baker | Aug 2004 | B1 |
6832367 | Choi et al. | Dec 2004 | B1 |
6895550 | Kanchirayappa et al. | May 2005 | B2 |
6898632 | Gordy et al. | May 2005 | B2 |
6907396 | Muttik et al. | Jun 2005 | B1 |
6971097 | Wallman | Nov 2005 | B1 |
6981279 | Arnold et al. | Dec 2005 | B1 |
7007107 | Ivchenko et al. | Feb 2006 | B1 |
7028179 | Anderson et al. | Apr 2006 | B2 |
7043757 | Hoefelmeyer et al. | May 2006 | B2 |
7069316 | Gryaznov | Jun 2006 | B1 |
7080407 | Zhao et al. | Jul 2006 | B1 |
7080408 | Pak et al. | Jul 2006 | B1 |
7093002 | Wolff et al. | Aug 2006 | B2 |
7093239 | van der Made | Aug 2006 | B1 |
7100201 | Izatt | Aug 2006 | B2 |
7107617 | Hursey et al. | Sep 2006 | B2 |
7159149 | Spiegel et al. | Jan 2007 | B2 |
7231667 | Jordan | Jun 2007 | B2 |
7240364 | Branscomb et al. | Jul 2007 | B1 |
7240368 | Roesch et al. | Jul 2007 | B1 |
7243371 | Kasper et al. | Jul 2007 | B1 |
7251215 | Turner et al. | Jul 2007 | B1 |
7287278 | Liang | Oct 2007 | B2 |
7308716 | Danford et al. | Dec 2007 | B2 |
7328453 | Merkle, Jr. et al. | Feb 2008 | B2 |
7346486 | Ivancic et al. | Mar 2008 | B2 |
7356736 | Natvig | Apr 2008 | B2 |
7386888 | Liang et al. | Jun 2008 | B2 |
7392542 | Bucher | Jun 2008 | B2 |
7418729 | Szor | Aug 2008 | B2 |
7428300 | Drew et al. | Sep 2008 | B1 |
7441272 | Durham et al. | Oct 2008 | B2 |
7448084 | Apap et al. | Nov 2008 | B1 |
7458098 | Judge et al. | Nov 2008 | B2 |
7464404 | Carpenter et al. | Dec 2008 | B2 |
7464407 | Nakae et al. | Dec 2008 | B2 |
7467408 | O'Toole, Jr. | Dec 2008 | B1 |
7478428 | Thomlinson et al. | Jan 2009 | B1 |
7480773 | Reed | Jan 2009 | B1 |
7487543 | Arnold et al. | Feb 2009 | B2 |
7496960 | Chen et al. | Feb 2009 | B1 |
7496961 | Zimmer et al. | Feb 2009 | B2 |
7516488 | Kienzle et al. | Apr 2009 | B1 |
7519990 | Xie | Apr 2009 | B1 |
7523493 | Liang et al. | Apr 2009 | B2 |
7530104 | Thrower et al. | May 2009 | B1 |
7540025 | Tzadikario | May 2009 | B2 |
7565550 | Liang et al. | Jul 2009 | B2 |
7568233 | Szor et al. | Jul 2009 | B1 |
7584455 | Ball | Sep 2009 | B2 |
7603715 | Costa et al. | Oct 2009 | B2 |
7607171 | Marsden et al. | Oct 2009 | B1 |
7639714 | Stolfo et al. | Dec 2009 | B2 |
7644441 | Schmid et al. | Jan 2010 | B2 |
7657419 | van der Made | Feb 2010 | B2 |
7676841 | Sobchuk et al. | Mar 2010 | B2 |
7698548 | Shelest et al. | Apr 2010 | B2 |
7707633 | Danford et al. | Apr 2010 | B2 |
7712136 | Sprosts et al. | May 2010 | B2 |
7730011 | Deninger et al. | Jun 2010 | B1 |
7739740 | Nachenberg et al. | Jun 2010 | B1 |
7779463 | Stolfo et al. | Aug 2010 | B2 |
7784097 | Stolfo et al. | Aug 2010 | B1 |
7832008 | Kraemer | Nov 2010 | B1 |
7836502 | Zhao et al. | Nov 2010 | B1 |
7849506 | Dansey et al. | Dec 2010 | B1 |
7854007 | Sprosts et al. | Dec 2010 | B2 |
7869073 | Oshima | Jan 2011 | B2 |
7877803 | Enstone et al. | Jan 2011 | B2 |
7904959 | Sidiroglou et al. | Mar 2011 | B2 |
7908660 | Bahl | Mar 2011 | B2 |
7930738 | Petersen | Apr 2011 | B1 |
7937761 | Benett | May 2011 | B1 |
7949849 | Lowe et al. | May 2011 | B2 |
7996556 | Raghavan et al. | Aug 2011 | B2 |
7996836 | McCorkendale et al. | Aug 2011 | B1 |
7996904 | Chiueh et al. | Aug 2011 | B1 |
7996905 | Arnold et al. | Aug 2011 | B2 |
8006305 | Aziz | Aug 2011 | B2 |
8010667 | Zhang et al. | Aug 2011 | B2 |
8020206 | Hubbard et al. | Sep 2011 | B2 |
8028338 | Schneider et al. | Sep 2011 | B1 |
8042184 | Batenin | Oct 2011 | B1 |
8045094 | Teragawa | Oct 2011 | B2 |
8045458 | Alperovitch et al. | Oct 2011 | B2 |
8069484 | McMillan et al. | Nov 2011 | B2 |
8087086 | Lai et al. | Dec 2011 | B1 |
8171553 | Aziz et al. | May 2012 | B2 |
8176049 | Deninger et al. | May 2012 | B2 |
8176480 | Spertus | May 2012 | B1 |
8201246 | Wu et al. | Jun 2012 | B1 |
8204984 | Aziz et al. | Jun 2012 | B1 |
8214905 | Doukhvalov et al. | Jul 2012 | B1 |
8220055 | Kennedy | Jul 2012 | B1 |
8225288 | Miller et al. | Jul 2012 | B2 |
8225373 | Kraemer | Jul 2012 | B2 |
8233882 | Rogel | Jul 2012 | B2 |
8234640 | Fitzgerald et al. | Jul 2012 | B1 |
8234709 | Viljoen et al. | Jul 2012 | B2 |
8239944 | Nachenberg et al. | Aug 2012 | B1 |
8260914 | Ranjan | Sep 2012 | B1 |
8266091 | Gubin et al. | Sep 2012 | B1 |
8286251 | Eker et al. | Oct 2012 | B2 |
8291499 | Aziz et al. | Oct 2012 | B2 |
8307435 | Mann et al. | Nov 2012 | B1 |
8307443 | Wang et al. | Nov 2012 | B2 |
8312545 | Tuvell et al. | Nov 2012 | B2 |
8321936 | Green et al. | Nov 2012 | B1 |
8321941 | Tuvell et al. | Nov 2012 | B2 |
8332571 | Edwards et al. | Dec 2012 | B1 |
8365286 | Poston | Jan 2013 | B2 |
8365297 | Parshin et al. | Jan 2013 | B1 |
8370938 | Daswani et al. | Feb 2013 | B1 |
8370939 | Zaitsev et al. | Feb 2013 | B2 |
8375444 | Aziz et al. | Feb 2013 | B2 |
8381299 | Stolfo et al. | Feb 2013 | B2 |
8402529 | Green et al. | Mar 2013 | B1 |
8464340 | Ahn et al. | Jun 2013 | B2 |
8479174 | Chiriac | Jul 2013 | B2 |
8479276 | Vaystikh et al. | Jul 2013 | B1 |
8479291 | Bodke | Jul 2013 | B1 |
8510827 | Leake et al. | Aug 2013 | B1 |
8510828 | Guo et al. | Aug 2013 | B1 |
8510842 | Amit et al. | Aug 2013 | B2 |
8516478 | Edwards et al. | Aug 2013 | B1 |
8516590 | Ranadive et al. | Aug 2013 | B1 |
8516593 | Aziz | Aug 2013 | B2 |
8522348 | Chen et al. | Aug 2013 | B2 |
8528086 | Aziz | Sep 2013 | B1 |
8533824 | Hutton et al. | Sep 2013 | B2 |
8539582 | Aziz et al. | Sep 2013 | B1 |
8549638 | Aziz | Oct 2013 | B2 |
8555391 | Demir et al. | Oct 2013 | B1 |
8561177 | Aziz et al. | Oct 2013 | B1 |
8566946 | Aziz et al. | Oct 2013 | B1 |
8584094 | Dadhia et al. | Nov 2013 | B2 |
8584234 | Sobel et al. | Nov 2013 | B1 |
8584239 | Aziz et al. | Nov 2013 | B2 |
8595834 | Xie et al. | Nov 2013 | B2 |
8627476 | Satish et al. | Jan 2014 | B1 |
8635696 | Aziz | Jan 2014 | B1 |
8682812 | Ranjan | Mar 2014 | B1 |
8695096 | Zhang | Apr 2014 | B1 |
8713631 | Pavlyushchik | Apr 2014 | B1 |
8713681 | Silberman et al. | Apr 2014 | B2 |
8726392 | McCorkendale et al. | May 2014 | B1 |
8782792 | Bodke | Jul 2014 | B1 |
8789172 | Stolfo et al. | Jul 2014 | B2 |
8789178 | Kejriwal et al. | Jul 2014 | B2 |
8805947 | Kuzkin et al. | Aug 2014 | B1 |
8806647 | Daswani et al. | Aug 2014 | B1 |
8850570 | Ramzan | Sep 2014 | B1 |
8850571 | Staniford et al. | Sep 2014 | B2 |
20010005889 | Albrecht | Jun 2001 | A1 |
20010047326 | Broadbent et al. | Nov 2001 | A1 |
20020018903 | Kokubo et al. | Feb 2002 | A1 |
20020038430 | Edwards et al. | Mar 2002 | A1 |
20020091819 | Melchione et al. | Jul 2002 | A1 |
20020095607 | Lin-Hendel | Jul 2002 | A1 |
20020144156 | Copeland, III | Oct 2002 | A1 |
20020162015 | Tang | Oct 2002 | A1 |
20020166063 | Lachman et al. | Nov 2002 | A1 |
20020169952 | DiSanto et al. | Nov 2002 | A1 |
20020184528 | Shevenell et al. | Dec 2002 | A1 |
20020188887 | Largman et al. | Dec 2002 | A1 |
20020194490 | Halperin et al. | Dec 2002 | A1 |
20030074578 | Ford et al. | Apr 2003 | A1 |
20030084318 | Schertz | May 2003 | A1 |
20030101381 | Mateev et al. | May 2003 | A1 |
20030115483 | Liang | Jun 2003 | A1 |
20030188190 | Aaron et al. | Oct 2003 | A1 |
20030191957 | Hypponen et al. | Oct 2003 | A1 |
20030200460 | Morota et al. | Oct 2003 | A1 |
20030212902 | Van Der Made | Nov 2003 | A1 |
20030229801 | Kouznetsov et al. | Dec 2003 | A1 |
20030237000 | Denton et al. | Dec 2003 | A1 |
20040003323 | Bennett et al. | Jan 2004 | A1 |
20040015712 | Szor | Jan 2004 | A1 |
20040019832 | Arnold et al. | Jan 2004 | A1 |
20040047356 | Bauer | Mar 2004 | A1 |
20040083408 | Spiegel et al. | Apr 2004 | A1 |
20040093513 | Cantrell et al. | May 2004 | A1 |
20040111531 | Staniford et al. | Jun 2004 | A1 |
20040128355 | Chao et al. | Jul 2004 | A1 |
20040165588 | Pandya | Aug 2004 | A1 |
20040236963 | Danford et al. | Nov 2004 | A1 |
20040243349 | Greifeneder et al. | Dec 2004 | A1 |
20040249911 | Alkhatib et al. | Dec 2004 | A1 |
20040255161 | Cavanaugh | Dec 2004 | A1 |
20040268147 | Wiederin et al. | Dec 2004 | A1 |
20050021740 | Bar et al. | Jan 2005 | A1 |
20050033960 | Vialen et al. | Feb 2005 | A1 |
20050033989 | Poletto et al. | Feb 2005 | A1 |
20050050148 | Mohammadioun et al. | Mar 2005 | A1 |
20050086523 | Zimmer et al. | Apr 2005 | A1 |
20050091513 | Mitomo et al. | Apr 2005 | A1 |
20050091533 | Omote et al. | Apr 2005 | A1 |
20050091652 | Ross et al. | Apr 2005 | A1 |
20050108562 | Khazan et al. | May 2005 | A1 |
20050114663 | Cornell et al. | May 2005 | A1 |
20050125195 | Brendel | Jun 2005 | A1 |
20050149726 | Joshi et al. | Jul 2005 | A1 |
20050157662 | Bingham et al. | Jul 2005 | A1 |
20050183143 | Anderholm et al. | Aug 2005 | A1 |
20050201297 | Peikari | Sep 2005 | A1 |
20050210533 | Copeland et al. | Sep 2005 | A1 |
20050238005 | Chen et al. | Oct 2005 | A1 |
20050240781 | Gassoway | Oct 2005 | A1 |
20050265331 | Stolfo | Dec 2005 | A1 |
20050283839 | Cowburn | Dec 2005 | A1 |
20060010495 | Cohen et al. | Jan 2006 | A1 |
20060015715 | Anderson | Jan 2006 | A1 |
20060015747 | Van de Ven | Jan 2006 | A1 |
20060021029 | Brickell et al. | Jan 2006 | A1 |
20060021054 | Costa et al. | Jan 2006 | A1 |
20060031476 | Mathes et al. | Feb 2006 | A1 |
20060047665 | Neil | Mar 2006 | A1 |
20060070130 | Costea et al. | Mar 2006 | A1 |
20060075496 | Carpenter et al. | Apr 2006 | A1 |
20060095968 | Portolani et al. | May 2006 | A1 |
20060101516 | Sudaharan et al. | May 2006 | A1 |
20060101517 | Banzhof et al. | May 2006 | A1 |
20060117385 | Mester et al. | Jun 2006 | A1 |
20060123477 | Raghavan et al. | Jun 2006 | A1 |
20060143709 | Brooks et al. | Jun 2006 | A1 |
20060150249 | Gassen et al. | Jul 2006 | A1 |
20060161983 | Cothrell et al. | Jul 2006 | A1 |
20060161987 | Levy-Yurista | Jul 2006 | A1 |
20060161989 | Reshef et al. | Jul 2006 | A1 |
20060164199 | Gilde et al. | Jul 2006 | A1 |
20060173992 | Weber et al. | Aug 2006 | A1 |
20060179147 | Tran et al. | Aug 2006 | A1 |
20060184632 | Marino et al. | Aug 2006 | A1 |
20060191010 | Benjamin | Aug 2006 | A1 |
20060221956 | Narayan et al. | Oct 2006 | A1 |
20060236393 | Kramer et al. | Oct 2006 | A1 |
20060242709 | Seinfeld et al. | Oct 2006 | A1 |
20060248519 | Jaeger et al. | Nov 2006 | A1 |
20060248582 | Panjwani et al. | Nov 2006 | A1 |
20060251104 | Koga | Nov 2006 | A1 |
20060288417 | Bookbinder et al. | Dec 2006 | A1 |
20070006288 | Mayfield et al. | Jan 2007 | A1 |
20070006313 | Porras et al. | Jan 2007 | A1 |
20070011174 | Takaragi et al. | Jan 2007 | A1 |
20070016951 | Piccard et al. | Jan 2007 | A1 |
20070033645 | Jones | Feb 2007 | A1 |
20070038943 | FitzGerald et al. | Feb 2007 | A1 |
20070064689 | Shin et al. | Mar 2007 | A1 |
20070074169 | Chess et al. | Mar 2007 | A1 |
20070094730 | Bhikkaji et al. | Apr 2007 | A1 |
20070101435 | Konanka et al. | May 2007 | A1 |
20070143827 | Nicodemus et al. | Jun 2007 | A1 |
20070156895 | Vuong | Jul 2007 | A1 |
20070157180 | Tillmann et al. | Jul 2007 | A1 |
20070157306 | Elrod et al. | Jul 2007 | A1 |
20070168988 | Eisner et al. | Jul 2007 | A1 |
20070171824 | Ruello et al. | Jul 2007 | A1 |
20070174915 | Gribble et al. | Jul 2007 | A1 |
20070192500 | Lum | Aug 2007 | A1 |
20070192858 | Lum | Aug 2007 | A1 |
20070198275 | Malden et al. | Aug 2007 | A1 |
20070208822 | Wang et al. | Sep 2007 | A1 |
20070220607 | Sprosts et al. | Sep 2007 | A1 |
20070240218 | Tuvell et al. | Oct 2007 | A1 |
20070240219 | Tuvell et al. | Oct 2007 | A1 |
20070240220 | Tuvell et al. | Oct 2007 | A1 |
20070240222 | Tuvell et al. | Oct 2007 | A1 |
20070250930 | Aziz et al. | Oct 2007 | A1 |
20070271446 | Nakamura | Nov 2007 | A1 |
20080005782 | Aziz | Jan 2008 | A1 |
20080028463 | Dagon et al. | Jan 2008 | A1 |
20080040710 | Chiriac | Feb 2008 | A1 |
20080046781 | Childs et al. | Feb 2008 | A1 |
20080072326 | Danford et al. | Mar 2008 | A1 |
20080077793 | Tan et al. | Mar 2008 | A1 |
20080080518 | Hoeflin et al. | Apr 2008 | A1 |
20080086720 | Lekel | Apr 2008 | A1 |
20080098476 | Syversen | Apr 2008 | A1 |
20080120722 | Sima et al. | May 2008 | A1 |
20080134178 | Fitzgerald et al. | Jun 2008 | A1 |
20080134334 | Kim et al. | Jun 2008 | A1 |
20080141376 | Clausen et al. | Jun 2008 | A1 |
20080184373 | Traut et al. | Jul 2008 | A1 |
20080189787 | Arnold et al. | Aug 2008 | A1 |
20080201778 | Guo et al. | Aug 2008 | A1 |
20080209557 | Herley et al. | Aug 2008 | A1 |
20080215742 | Goldszmidt et al. | Sep 2008 | A1 |
20080222728 | Chavez et al. | Sep 2008 | A1 |
20080222729 | Chen et al. | Sep 2008 | A1 |
20080263665 | Ma et al. | Oct 2008 | A1 |
20080295172 | Bohacek | Nov 2008 | A1 |
20080301810 | Lehane et al. | Dec 2008 | A1 |
20080307524 | Singh et al. | Dec 2008 | A1 |
20080313738 | Enderby | Dec 2008 | A1 |
20080320594 | Jiang | Dec 2008 | A1 |
20090007100 | Field et al. | Jan 2009 | A1 |
20090013408 | Schipka | Jan 2009 | A1 |
20090031423 | Liu et al. | Jan 2009 | A1 |
20090036111 | Danford et al. | Feb 2009 | A1 |
20090037835 | Goldman | Feb 2009 | A1 |
20090044024 | Oberheide et al. | Feb 2009 | A1 |
20090044274 | Budko et al. | Feb 2009 | A1 |
20090064332 | Porras et al. | Mar 2009 | A1 |
20090077666 | Chen et al. | Mar 2009 | A1 |
20090083369 | Marmor | Mar 2009 | A1 |
20090083855 | Apap et al. | Mar 2009 | A1 |
20090089879 | Wang et al. | Apr 2009 | A1 |
20090094697 | Provos et al. | Apr 2009 | A1 |
20090113425 | Ports et al. | Apr 2009 | A1 |
20090125976 | Wassermann et al. | May 2009 | A1 |
20090126015 | Monastyrsky et al. | May 2009 | A1 |
20090126016 | Sobko et al. | May 2009 | A1 |
20090133125 | Choi et al. | May 2009 | A1 |
20090144823 | Lamastra et al. | Jun 2009 | A1 |
20090158430 | Borders | Jun 2009 | A1 |
20090172815 | Gu et al. | Jul 2009 | A1 |
20090187992 | Poston | Jul 2009 | A1 |
20090193293 | Stolfo et al. | Jul 2009 | A1 |
20090199296 | Xie et al. | Aug 2009 | A1 |
20090228233 | Anderson et al. | Sep 2009 | A1 |
20090241187 | Troyansky | Sep 2009 | A1 |
20090241190 | Todd et al. | Sep 2009 | A1 |
20090265692 | Godefroid et al. | Oct 2009 | A1 |
20090271867 | Zhang | Oct 2009 | A1 |
20090300415 | Zhang et al. | Dec 2009 | A1 |
20090300761 | Park et al. | Dec 2009 | A1 |
20090328185 | Berg et al. | Dec 2009 | A1 |
20090328221 | Blumfield et al. | Dec 2009 | A1 |
20100005146 | Drako et al. | Jan 2010 | A1 |
20100011205 | McKenna | Jan 2010 | A1 |
20100017546 | Poo et al. | Jan 2010 | A1 |
20100031353 | Thomas et al. | Feb 2010 | A1 |
20100037314 | Perdisci et al. | Feb 2010 | A1 |
20100043073 | Kuwamura | Feb 2010 | A1 |
20100054278 | Stolfo et al. | Mar 2010 | A1 |
20100058474 | Hicks | Mar 2010 | A1 |
20100064044 | Nonoyama | Mar 2010 | A1 |
20100077481 | Polyakov et al. | Mar 2010 | A1 |
20100083376 | Pereira et al. | Apr 2010 | A1 |
20100100718 | Srinivasan | Apr 2010 | A1 |
20100115621 | Staniford et al. | May 2010 | A1 |
20100132038 | Zaitsev | May 2010 | A1 |
20100154056 | Smith et al. | Jun 2010 | A1 |
20100192223 | Ismael et al. | Jul 2010 | A1 |
20100220863 | Dupaquis et al. | Sep 2010 | A1 |
20100235831 | Dittmer | Sep 2010 | A1 |
20100251104 | Massand | Sep 2010 | A1 |
20100281102 | Chinta et al. | Nov 2010 | A1 |
20100281541 | Stolfo et al. | Nov 2010 | A1 |
20100281542 | Stolfo et al. | Nov 2010 | A1 |
20100287260 | Peterson et al. | Nov 2010 | A1 |
20100306173 | Frank | Dec 2010 | A1 |
20110004737 | Greenebaume | Jan 2011 | A1 |
20110025504 | Lyon et al. | Feb 2011 | A1 |
20110041179 | Stahlberg | Feb 2011 | A1 |
20110047594 | Mahaffey et al. | Feb 2011 | A1 |
20110047620 | Mahaffey et al. | Feb 2011 | A1 |
20110055907 | Narasimhan et al. | Mar 2011 | A1 |
20110078794 | Manni et al. | Mar 2011 | A1 |
20110093951 | Aziz | Apr 2011 | A1 |
20110099620 | Stavrou et al. | Apr 2011 | A1 |
20110099633 | Aziz | Apr 2011 | A1 |
20110113231 | Kaminsky | May 2011 | A1 |
20110145918 | Jung et al. | Jun 2011 | A1 |
20110145920 | Mahaffey et al. | Jun 2011 | A1 |
20110145934 | Abramovici et al. | Jun 2011 | A1 |
20110167493 | Song et al. | Jul 2011 | A1 |
20110167494 | Bowen et al. | Jul 2011 | A1 |
20110173460 | Ito et al. | Jul 2011 | A1 |
20110219450 | McDougal et al. | Sep 2011 | A1 |
20110225624 | Sawhney et al. | Sep 2011 | A1 |
20110225655 | Niemela et al. | Sep 2011 | A1 |
20110247072 | Staniford et al. | Oct 2011 | A1 |
20110265182 | Peinado et al. | Oct 2011 | A1 |
20110289582 | Kejriwal et al. | Nov 2011 | A1 |
20110302587 | Nishikawa et al. | Dec 2011 | A1 |
20110307954 | Melnik et al. | Dec 2011 | A1 |
20110307955 | Kaplan et al. | Dec 2011 | A1 |
20110307956 | Yermakov et al. | Dec 2011 | A1 |
20110314546 | Aziz et al. | Dec 2011 | A1 |
20120023593 | Puder et al. | Jan 2012 | A1 |
20120054869 | Yen et al. | Mar 2012 | A1 |
20120066698 | Yanoo | Mar 2012 | A1 |
20120079596 | Thomas et al. | Mar 2012 | A1 |
20120084859 | Radinsky et al. | Apr 2012 | A1 |
20120110667 | Zubrilin et al. | May 2012 | A1 |
20120117652 | Manni et al. | May 2012 | A1 |
20120124426 | Maybee et al. | May 2012 | A1 |
20120174186 | Aziz et al. | Jul 2012 | A1 |
20120174196 | Bhogavilli et al. | Jul 2012 | A1 |
20120174218 | McCoy et al. | Jul 2012 | A1 |
20120198279 | Schroeder | Aug 2012 | A1 |
20120210423 | Friedrichs et al. | Aug 2012 | A1 |
20120222121 | Staniford et al. | Aug 2012 | A1 |
20120255015 | Sahita et al. | Oct 2012 | A1 |
20120255017 | Sallam | Oct 2012 | A1 |
20120260342 | Dube et al. | Oct 2012 | A1 |
20120266244 | Green et al. | Oct 2012 | A1 |
20120278886 | Luna | Nov 2012 | A1 |
20120297489 | Dequevy | Nov 2012 | A1 |
20120330801 | McDougal et al. | Dec 2012 | A1 |
20130014259 | Gribble et al. | Jan 2013 | A1 |
20130036472 | Aziz | Feb 2013 | A1 |
20130047257 | Aziz | Feb 2013 | A1 |
20130074185 | McDougal et al. | Mar 2013 | A1 |
20130086684 | Mohler | Apr 2013 | A1 |
20130097699 | Balupari et al. | Apr 2013 | A1 |
20130097706 | Titonis et al. | Apr 2013 | A1 |
20130111587 | Goel et al. | May 2013 | A1 |
20130117852 | Stute | May 2013 | A1 |
20130117855 | Kim et al. | May 2013 | A1 |
20130139264 | Brinkley et al. | May 2013 | A1 |
20130160125 | Likhachev et al. | Jun 2013 | A1 |
20130160127 | Jeong et al. | Jun 2013 | A1 |
20130160130 | Mendelev et al. | Jun 2013 | A1 |
20130160131 | Madou et al. | Jun 2013 | A1 |
20130167236 | Sick | Jun 2013 | A1 |
20130174214 | Duncan | Jul 2013 | A1 |
20130185795 | Winn et al. | Jul 2013 | A1 |
20130185798 | Saunders et al. | Jul 2013 | A1 |
20130191915 | Antonakakis et al. | Jul 2013 | A1 |
20130196649 | Paddon et al. | Aug 2013 | A1 |
20130227691 | Aziz et al. | Aug 2013 | A1 |
20130246370 | Bartram et al. | Sep 2013 | A1 |
20130263260 | Mahaffey at al. | Oct 2013 | A1 |
20130291109 | Staniford et al. | Oct 2013 | A1 |
20130298243 | Kumar et al. | Nov 2013 | A1 |
20140053260 | Gupta et al. | Feb 2014 | A1 |
20140053261 | Gupta et al. | Feb 2014 | A1 |
20140130158 | Wang et al. | May 2014 | A1 |
20140137180 | Lukacs et al. | May 2014 | A1 |
20140169762 | Ryu | Jun 2014 | A1 |
20140179360 | Jackson et al. | Jun 2014 | A1 |
20140328204 | Klotsche et al. | Nov 2014 | A1 |
20140351935 | Shao et al. | Nov 2014 | A1 |
Number | Date | Country |
---|---|---|
2439806 | Jan 2008 | GB |
2490431 | Oct 2012 | GB |
WO-0206928 | Jan 2002 | WO |
WO-0223805 | Mar 2002 | WO |
WO 2007022454 | Feb 2007 | WO |
WO-2007-117636 | Oct 2007 | WO |
WO-2008041950 | Apr 2008 | WO |
WO 2008084259 | Jul 2008 | WO |
WO-2011084431 | Jul 2011 | WO |
WO 2011112348 | Sep 2011 | WO |
WO 2012075336 | Jun 2012 | WO |
WO-2012145066 | Oct 2012 | WO |
WO 2013067505 | May 2013 | WO |
Entry |
---|
Zhang et al., The Effects of Threading, Infection Time, and Multiple-Attacker Collaboration on Malware Propagation, Sep. 2009, IEEE 28th International Symposium on Reliable Distributed Systems, pp. 73-82. |
Wahid et al., Characterising the Evolution in Scanning Activity of Suspicious Hosts, Oct. 2009, Third International Conference on Network and System Security, pp. 344-350. |
IEEE Xplore Digital Library Sear Results for “detection of unknown computer worms”. Http//ieeexplore.ieee.org/searchresult.jsp?SortField=Score&SortOrder=desc&ResultC . . . , (Accessed on Aug. 28, 2009). |
AltaVista Advanced Search Results. “Event Orchestrator”. Http://www.altavista.com/web/results?Itag=ody&pg=aq&aqmode=aqa=Event+Orchesrator . . . , (Accessed on Sep. 3, 2009). |
AltaVista Advanced Search Results. “attack vector identifier”. Http://www.altavista.com/web/results?Itag=ody&pg=aq&aqmode=aqa=Event+Orchestrator . . . , (Accessed on Sep. 15, 2009). |
Cisco, Configuring the Catalyst Switched Port Analyzer (SPAN) (“Cisco”), (1992-2003). |
Reiner Sailer, Enriquillo Valdez, Trent Jaeger, Roonald Perez, Leendert van Doorn, John Linwood Griffin, Stefan Berger., sHype: Secure Hypervisor Appraoch to Trusted Virtualized Systems (Feb. 2, 2005) (“Sailer”). |
Excerpt regarding First Printing Date for Merike Kaeo, Designing Network Security(“Kaeo”), (2005). |
The Sniffers's Guide to Raw Traffic available at: yuba.stanford.edu/˜casado/pcap/section1.html, Jan. 6, 2014). |
“Network Security: NetDetector—Network Intrusion Forensic System (NIFS) Whitepaper”, “NetDetector Whitepaper”), (2003). |
“Packet”, Microsoft Computer Dictionary, Microsoft Press, (Mar. 2002), 1 page. |
“When Virtual is Better Than Real”, IEEEXplore Digital Library, available at, http://ieeexplore.ieee.org/xpl/articleDetails.jsp?reload=true&arnumber=990073, (Dec. 7, 2013). |
Abdullah, et al., Visualizing Network Data for Intrusion Detection, 2005 IEEE Workshop on Information Assurance and Security, pp. 100-108. |
Adetoye, Adedayo , et al., “Network Intrusion Detection & Response System”, (“Adetoye”), (Sep. 2003). |
Aura, Tuomas, “Scanning electronic documents for personally identifiable information”, Proceedings of the 5th ACM workshop on Privacy in electronic society. ACM, 2006. |
Baecher, “The Nepenthes Platform: An Efficient Approach to collect Malware”, Springer-verlag Berlin Heidelberg, (2006), pp. 165-184. |
Bayer, et al., “Dynamic Analysis of Malicious Code”, J Comput Virol, Springer-Verlag, France., (2006), pp. 67-77. |
Boubalos, Chris , “extracting syslog data out of raw pcap dumps, seclists.org, Honeypots mailing list archives”, available at http://seclists.org/honeypots/2003/q2/319 (“Boubalos”), (Jun. 5, 2003). |
Chaudet, C. , et al., “Optimal Positioning of Active and Passive Monitoring Devices”, International Conference on Emerging Networking Experiments and Technologies, Proceedings of the 2005 ACM Conference on Emerging Network Experiment and Technology, CoNEXT '05, Toulousse, France, (Oct. 2005), pp. 71-82. |
Cohen, M.I. , “PyFlag—An advanced network forensic framework”, Digital investigation 5, Elsevier, (2008), pp. S112-S120. |
Costa, M. , et al., “Vigilante: End-to-End Containment of Internet Worms”, SOSP '05, Association for Computing Machinery, Inc., Brighton U.K., (Oct. 23-26, 2005). |
Crandall, J.R. , et al., “Minos:Control Data Attack Prevention Orthogonal to Memory Model”, 37th International Symposium on Microarchitecture, Portland, Oregon, (Dec. 2004). |
Deutsch, P. , “Zlib compressed data format specification version 3.3” RFC 1950, (1996). |
Distler, “Malware Analysis: An Introduction”, SANS Institute InfoSec Reading Room, SANS Institute, (2007). |
Dunlap, George W. et al., “ReVirt: Enabling Intrusion Analysis through Virtual-Machine Logging and Replay”, Proceeding of the 5th Symposium on Operating Systems Design and Implementation, USENIX Association, (“Dunlap”), (Dec. 9, 2002). |
Filiol, Eric , et al., “Combinatorial Optimisation of Worm Propagation on an Unknown Network”, International Journal of Computer Science 2.2 (2007). |
Goel, et al., Reconstructing System State for Intrusion Analysis, Apr. 2008 SIGOPS Operating Systems Review, vol. 42 Issue 3, pp. 21-28. |
Hjelmvik, Erik , “Passive Network Security Analysis with NetworkMiner”, (IN)Secure, Issue 18, (Oct. 2008), pp. 1-100. |
Kaeo, Merike , “Designing Network Security”, (“Kaeo”), (Nov. 2003). |
Kim, H. , et al., “Autograph: Toward Automated, Distributed Worm Signature Detection”, Proceedings of the 13th Usenix Security Symposium (Security 2004), San Diego, (Aug. 2004), pp. 271-286. |
Krasnyansky, Max , et al., Universal TUN/TAP driver, available at https://www.kernel.org/doc/Documentation/networking/tuntap.txt (2002) (“Krasnyansky”). |
Kreibich, C. , et al., “Honeycomb-Creating Intrusion Detection Signatures Using Honeypots”, 2nd Workshop on Hot Topics in Networks (HotNets-11), Boston, USA, (2003). |
Kristoff, J. , “Botnets, Detection and Mitigation: DNS-Based Techniques”, NU Security Day, (2005), 23 pages. |
Liljenstam, Michael et al., “Simulating Realistic Network Traffic for Worm Warning System Design and Testing”, Institute for Security Technology studies, Dartmouth College, (“Liljenstam”) (Oct. 27, 2003). |
Marchette, David J., “Computer Intrusion Detection and Network Monitoring: A Statistical Viewpoint”, “Marchette”, (2001). |
Margolis, P.E. , “Random House Webster's ‘Computer & Internet Dictionary 3rd Edition’”, ISBN 0375703519, (Dec. 1998). |
Moore, D. , et al., “Internet Quarantine: Requirements for Containing Self-Propagating Code”, INFOCOM, vol. 3, (Mar. 30-Apr. 3, 2003), pp. 1901-1910. |
Morales, Jose A., et al., ““Analyzing and exploiting network behaviors of malware.””, Security and Privacy in Communication Networks. Springer Berlin Heidelberg, 2010. 20-34. |
Natvig, Kurt , “SANDBOXII: Internet”, Virus Bulletin Conference, (“Natvig”), (Sep. 2002). |
NetBIOS Working Group. Protocol Standard for a NetBIOS Service on a TCP/UDP transport: Concepts and Methods. STD 19, RFC 1001, Mar. 1987. |
Newsome, J. , et al., “Dynamic Taint Analysis for Automatic Detection, Analysis, and Signature Generation of Exploits on Commodity Software”, In Proceedings of the 12th Annual Network and Distributed System Security, Symposium (NDSS '05), (Feb. 2005). |
Newsome, J. , et al., “Polygraph: Automatically Generating Signatures for Polymorphic Worms”, In Proceedings of the IEEE Symposium on Security and Privacy, (May 2005). |
Nojiri, D. , et al., “Cooperation Response Strategies for Large Scale Attack Mitigation”, DARPA Information Survivability Conference and Exposition, vol. 1, (Apr. 22-24, 2003), pp. 293-302. |
Silicon Defense, “Worm Containment in the Internal Network”, (Mar. 2003), pp. 1-25. |
Singh, S. , et al., “Automated Worm Fingerprinting”, Proceedings of the ACM/USENIX Symposium on Operating System Design and Implementation, San Francisco, California, (Dec. 2004). |
Spitzner, Lance , “Honeypots: Tracking Hackers”, (“Spizner”), (Sep. 17, 2002). |
Thomas H. Ptacek, and Timothy N. Newsham , “Insertion, Evasion, and Denial of Service: Eluding Network Intrusion Detection”, Secure Networks, (“Ptacek”), (Jan. 1998). |
Venezia, Paul , “NetDetector Captures Intrusions”, InfoWorld Issue 27, (“Venezia”), (Jul. 14, 2003). |
Whyte, et al., “DNS-Based Detection of Scanning Works in an Enterprise Network”, Proceedings of the 12th Annual Network and Distributed System Security Symposium, (Feb. 2005), 15 pages. |
Williamson, Matthew M., “Throttling Viruses: Restricting Propagation to Defeat Malicious Mobile Code”, ACSAC Conference, Las Vegas, NV, USA, (Dec. 2002), pp. 1-9. |
Rautiainen et al: “A look at Portable Document Format vulnerabilities”, Information Security Technical Report Elsevier Advanced Technology, Amsterdam, NL, vol. 14 No. 1, Feb. 1, 2009, pp. 30-33, XP026144094, ISSN: 1363-4127, DOI: 10.1016/J.ISTR.2009.04.001 [retrieved on May 8, 2009]. |
Didier Stevens: “Malicious PDF Documents Explained”, Security & Privacy, IEEE, IEEE Service Center, Los Alamitos, CA, US, vol. 9, No. 1, Jan. 1, 2011, pp. 80-82, XP011329453, ISSN: 1540-7993, DOI: 10.1109/MSP.2011.14. |
Anonymous: “Inside Adobe Reader Protected Mode—Part 1—Design”, Oct. 5, 2010, XP055116424, Retrieved from the Internet: URL:http://blogs.adobe.com/security/2010/10/inside-adobe-reader-protected-mode-part-1-design.html [retrieved on May 6, 2014]. |
Adobe Systems Incorporated, “PDF 32000-1:2008, Document management—Portable document format—Part1:PDF 1.7”, First Edition, Jul. 1, 2008, 756 pages. |
Apostolopoulos, George; Hassapis, Constantinos; “V-eM: A Cluster of Virtual Machines for Robust, Detailed, and High-Performance Network Emulation”, 14th IEEE International Symposium on Modeling, Analysis, and Simulation of Computer and Telecommunication Systems, Sep. 11-14, 2006, pp. 117-126. |
Baldi, Mario; Risso, Fulvio; “A Framework for Rapid Development and Portable Execution of Packet-Handling Applications”, 5th IEEE International Symposium on Signal Processing and Information Technology, Dec. 21, 2005, pp. 233-238. |
Gregg Keizer: “Microsoft's HoneyMonkeys Show Patching Windows Works”, Aug. 8, 2005, XP055143386, Retrieved from the Internet: URL:https://web.archive.org/web/20121022220617/http://www.informationweek.com/microsofts-honeymonkeys-show-patching-wi/167600716 [retrieved on Sep. 29, 2014]. |
Idika et al., A-Survey-of-Malware-Detection-Techniques, Feb. 2, 2007, Department of Computer Science, Purdue University. |
Bayer et al., Dynamic analysis of malicious code, 2006, Springer-Verlag France. |
Mori, Detecting Unknown Computer Viruses, 2004, Springer-Verlag Berlin Heidelberg. |
Oberheide et al., CloudAV—N-Version Antivirus in the Network Cloud, 17th USENIX Security Symposium USENIX Security '08 Jul. 28-Aug. 1, 2008 San Jose, CA. |
Li et al., A VMM-Based System Call Interposition Framework for Program Monitoring, Dec. 2010, IEEE 16th International Conference on Parallel and Distributed Systems, pp. 706-711. |
Clark, John, Sylvian Leblanc, and Scott Knight. “Risks associated with usb hardware trojan devices used by insiders.” Systems Conference (SysCon), 2011 IEEE International. IEEE, 2011. |
Gibler, Clint, et al. AndroidLeaks: automatically detecting potential privacy leaks in android applications on a large scale. Springer Berlin Heidelberg, 2012. |
Isohara, Takamasa, Keisuke Takemori, and Ayumu Kubota. “Kernel-based behavior analysis for android malware detection.” Computational intelligence and Security (CIS), 2011 Seventh International Conference on. IEEE, 2011. |
Lindorter, Martina, Clemens Kolbitsch, and Paolo Milani Comparetti. “Detecting environment-sensitive malware.” Recent Advances in Intrusion Detection. Springer Berlin Heidelberg, 2011. |
Kevin A Roundy et al: “Hybrid Analysis and Control of Malware”, Sep. 15, 2010, Recent Advances in Intrusion Detection, Springer Berlin Heidelberg, Berlin, Heidelberg, pp. 317-338, XP019150454 ISBN:978-3-642-15511-6. |
Lok Kwong et al: “DroidScope: Seamlessly Reconstructing the OS and Dalvik Semantic Views for Dynamic Android Malware Analysis”, Aug. 10, 2012, XP055158513, Retrieved from the Internet: URL:https://www.usenix.org/system/files/conference/usenixsecurity12/sec12-final107.pdf [retrieved on Dec. 15, 2014]. |
Yuhei Kawakoya et al: “Memory behavior-based automatic malware unpacking in stealth debugging environment”, Malicious and Unwanted Software (Malware), 2010 5TH International Conference on, IEEE, Piscataway, NJ, USA, Oct. 19, 2010, pp. 39-46, XP031833827, ISBN:978-1-4244-8-9353-1. |