Method to detect application execution hijacking using memory protection

Information

  • Patent Grant
  • 11244044
  • Patent Number
    11,244,044
  • Date Filed
    Friday, February 15, 2019
    5 years ago
  • Date Issued
    Tuesday, February 8, 2022
    2 years ago
Abstract
According to one embodiment, a malware detection software being loaded into non-transitory computer readable medium for execution by a processor. The malware detection software comprises exploit detection logic, rule-matching logic, reporting logic and user interface logic. The exploit detection logic is configured to execute certain event logic with respect to a loaded module. The rule-matching logic includes detection logic that is configured to determine whether an access source is attempting to access a protected region and determine whether the access source is from a dynamically allocated memory. The reporting logic includes alert generating logic that is configured to generate an alert while the user interface logic is configured to notify a user or a network administrator of a potential cybersecurity attack.
Description
FIELD

Embodiments of the disclosure relate to the field of cyber-security. More specifically, one embodiment of the disclosure relates to a system, apparatus and method for identifying potential application-execution hijacking attacks using memory protection techniques.


GENERAL BACKGROUND

Over the last decade, malicious software has become a pervasive problem for Internet users as many networked resources include vulnerabilities that are subject to attack. For instance, over the past few years, more and more vulnerabilities are being discovered in software that is loaded onto network devices, such as vulnerabilities within operating systems for example. While some vulnerabilities continue to be addressed through software patches, prior to the release of such software patches, network devices will continue to be targeted for attack by malware, namely information such as computer code that attempts during execution to take advantage of a vulnerability in computer software by acquiring sensitive information or adversely influencing or attacking normal operations of the network device or the entire enterprise network.


For example, one commonly exploited vulnerability is known as a buffer overflow. In general, programs write data to a buffer. However, during a buffer overflow, the written data overruns the buffer's allocated boundary and overwrites adjacent memory locations. As a result, buffer overflows are the basis of many software vulnerabilities and can be maliciously exploited to cause erratic program behavior, memory access errors, incorrect results, and/or the misappropriation of sensitive data such as intellectual property.


Various techniques have been attempted to detect and prevent software exploits, however each technique has various tradeoffs One of the most generic techniques include Data. Execution Prevention (DEP), which is generally provided for by a processor. Using. DEP, memory spaces are automatically marked as non-executable unless they are explicitly told they are being allocated for executable code Specifically, a flag is set on a per-page basis and is set via a bit in the page table entry (PTE) for that page. If an attempt is made to execute code from a memory region that is marked as non-executable, the hardware feature passes an exception to DEP within the operating system and provides a corresponding indication. Consequently, DEP causes an exception within the code stack that is executing, thereby causing a failure coupled with an access violation DEP may be made stronger by CPU support with the No-Execute (NX) bit, also known as the XD bit, EVP bit, or XN bit, which allows the CPU to enforce execution rights at the hardware level.


Unfortunately, in short order, bypasses were developed by hackers to overcome DEP schemes. Specifically, a technique known as Return-Oriented Programming (ROP) was developed to circumvent DEP schemes. ROP techniques search for portions of code known as ROP gadgets in legitimate modules within a particular process. ROP gadgets generally comprise of one or more instructions, followed by a return. Combining a plurality of ROP gadgets along with appropriate values in the stack allows for the malicious shell code to be executed. Typically, the hacker's goal is to locate the address of a memory protection API, such as Virtual Protect, and mark the relevant memory region as executable (as compared to non-executable) Thereafter, the hacker may introduce a final ROP gadget to transfer the execution to the relevant memory region to execute the shellcode. As a result, the DEP scheme may be bypassed.


In an effort to make potential DEP bypasses more difficult, Address Space Layout Randomization (ASLR) was developed. ASLR involves randomly offsetting memory structures and module base addresses such that merely “guessing” the location of ROP gadgets and APIs becomes exceedingly difficult. On certain operating systems, such as the Microsoft® Windows® operating system, ASLR may be configured to randomize the location of executables and Dynamic Link Libraries (DLLs) in memory, stacks and heaps. For example, when an executable is loaded into memory, the operating system may receive a processor's timestamp counter (TSC), shift the TSC by a nominal amount, perform a division (e.g., a modulo operation), and then add a constant. The result of this operation may then be multiplied by yet another constant, at which point an executable image is loaded at the calculated offset.


However, some DLLs (including for example, ntdll, kernel32, etc.) are shared in memory across processes, their offsets are determined by a system-wide bias value that is computed at boot. Notably, the offset value is computed only once per boot. When DLLs are loaded, they are disposed into a shared memory region. The order in which modules are loaded is randomized too. Furthermore, when threads are created, their stack base address is randomized. Once the base address has been calculated, another value is derived from the TSC to compute the final stack base address. By using this method, ASLR was intended to provide a high theoretical degree of randomness.


When all of these ALSR mechanisms were combined with DEP, it was understood that shellcode would be prevented from executing because the memory region could not be executed. Moreover, it was expected that potential hackers also would not know the location of any ROP instructions in memory because the ROP gadget's address would be unreliable due to the randomization.


Nonetheless, bypasses were developed by hackers to overcome ASLR mechanisms so that DEP schemes could be easily exploited. For example, NOP sleds may be utilized to create a probabilistic exploit. Furthermore, using a pointer leak, a hacker can make an educated guess regarding a value on the stack at a reliable location to locate a usable function pointer or ROP gadget.


In other words, using these and other techniques, it may be possible to create a payload that reliably bypasses both DEP and ASLR. Moreover, once ASLR and DEP are compromised, it is a straight forward matter to control the execution of shellcode in the context of the application. Therefore, there exists a need for a system, apparatus and method for identifying potential application-execution hijacking attacks using memory protection techniques so as to prevent the execution of malicious shellcode.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the invention are illustrated by way of example and not by way of limitation in the figures of the accompanying drawings, in which like references indicate similar elements and in which:



FIG. 1A is an exemplary logical representation of a communication system deploying an application-execution hijacking detection system communicatively coupled to a network.



FIG. 1B an exemplary embodiment of a representation of the application-execution hijacking detection system of FIG. 1A.



FIG. 2 is an exemplary representation of a plurality of mobile devices deploying an application-execution hijacking detection application communicatively coupled to a network.



FIG. 3 is an exemplary embodiment of a mobile device deloying an application-execution hijacking detection application according to the present disclosure.



FIG. 4 is an exemplary block diagram of an operational flow of an application-execution hijacking detection system.



FIG. 5 is an exemplary block diagram of an alternative operational flow of an application-execution hijacking detection system.



FIG. 6 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a guard page to a base address of a loaded module.



FIG. 7 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a guard page to an import table of a loaded module.



FIG. 8 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a hardware breakpoint to a base address of a loaded module.



FIG. 9 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a guard page to a process environment block of a loaded module.





DETAILED DESCRIPTION

Various embodiments of the disclosure relate to a network appliance, such as an application-execution hijacking detection system (AEH) system for example, where the network appliance comprises a dynamic analysis server. According to one embodiment of the disclosure, information from received network traffic is analyzed to determine whether at least a portion of the received network traffic is likely to be associated with malware. A portion of the received network traffic, (hereinafter “object(s)”), that is determined to likely be associated with malware is deemed “suspicious.” The dynamic analysis server comprises virtual execution logic to automatically analyze one or more objects while the object(s) executes within a virtual machine (VM). In particular, the dynamic analysis server comprises event logic to analyze whether attempts have been made to access protected (guarded) pages of loaded modules so as to detect a potential AEH attack. In one embodiment, instead of implementing virtual execution logic, an AEH detection application may be configured so as to communicate the output of DLL/kernel logic directly into an application running on a mobile device, for example. It is envisioned that deployments other than VM-based deployments may also be used, including but not limited to runtime system deployments and the like.


Herein, according to one embodiment of the disclosure, the dynamic analysis server is configured to monitor and store access events of guarded page areas of any of various loaded modules. Any time a read, write or execute operation (“access event”) is performed on such guarded page areas, the access event is analyzed to determine whether it is malicious (and therefore associated with a malware attack), or non-malicious. For example, if the access source is from the heap, then there is a high likelihood that the access event is malicious. As used herein, the heap refers to any portion of memory where dynamically allocated memory resides. More specifically, in the context of loaded modules, the memory type is a “MEM_IMAGE” type. Conversely, if the memory type at any address is not the “MEM_IMAGE” type, then the address is from the heap, and therefore malicious. It is envisioned that upon a finding of maliciousness, an alert may be generated to communicate details of the access event, including for example the access source and corresponding memory type. In some embodiments, determining that an access event is malicious may result in the termination of the respective application, process and/or operation.


I. Terminology

In the following description, certain terminology is used to describe features of the invention. For example, in certain situations, term “logic” is representative of hardware, firmware and/or software that is configured to perform one or more functions. As hardware, logic may include circuitry having data processing or storage functionality. Examples of such circuitry may include, but are not limited or restricted to a microprocessor, one or more processor cores, a programmable gate array, a microcontroller, an application specific integrated circuit, wireless receiver, transmitter and/or transceiver circuitry, semiconductor memory, or combinatorial logic.


Logic may be software in the form of one or more software modules, such as executable code in the form of an executable application, an application programming interface (API), a subroutine, a function, a procedure, an applet, a servlet, a routine, source code, object code, a shared library/dynamic load library, or one or more instructions. These software modules may be stored in any type of a suitable non-transitory storage medium, or transitory storage medium (e.g., electrical, optical, acoustical or other form of propagated signals such as carrier waves, infrared signals, or digital signals). Examples of non-transitory storage medium may include, but are not limited or restricted to a programmable circuit; a semiconductor memory; non-persistent storage such as volatile memory (e.g., any type of random access memory “RAM”); persistent storage such as non-volatile memory (e.g., read-only memory “ROM”, power-backed RAM, flash memory, phase-change memory, etc.), a solid-state drive, hard disk drive, an optical disc drive, or a portable memory device. As firmware, the executable code is stored in persistent storage.


The term “object” generally refers to a collection of data, whether in transit (e.g., over a network) or at rest (e.g., stored), often having a logical structure or organization that enables it to be classified for purposes of analysis. During analysis, for example, the object may exhibit a set of expected characteristics and, during processing, a set of expected behaviors. The object may also exhibit a set of unexpected characteristics and a set of unexpected behaviors that may evidence an exploit and potentially allow the object to be classified as an exploit.


Examples of objects may include one or more flows or a self-contained element within a flow itself. A “flow” generally refers to related packets that are received, transmitted, or exchanged within a communication session. For convenience, a packet is broadly referred to as a series of bits or bytes having a prescribed format, which may include packets, frames, or cells.


As an illustrative example, an object may include a set of flows such as (1) a sequence of transmissions in accordance with a particular communication protocol (e.g., User Datagram Protocol (UDP); Transmission Control Protocol (TCP); or Hypertext Transfer Protocol (HTTP); etc.), or (2) inter-process communications (e.g., Remote Procedure Call “RPC” or analogous processes, etc.). Similar, as another illustrative example, the object may be a self-contained element, where different types of such objects may include an executable file, non-executable file (such as a document or a dynamically link library), a Portable Document Format (PDF) file, a JavaScript file, Zip file, a Flash file, a document (for example, a Microsoft Office® document), an electronic mail (email), downloaded web page, an instant messaging element in accordance with Session Initiation Protocol (SIP) or another messaging protocol, or the like.


According to one embodiment, the term “malware” may be construed broadly as any code or activity that initiates a malicious attack and/or operations associated with anomalous or unwanted behavior. For instance, malware may correspond to a type of malicious computer code that executes an exploit to take advantage of a vulnerability, for example, to harm or co-opt operation of a network device or misappropriate, modify or delete data. Malware may also correspond to an exploit, namely information (e.g., executable code, data, command(s), etc.) that attempts to take advantage of a vulnerability in software and/or an action by a person gaining unauthorized access to one or more areas of a network device to cause the network device to experience undesirable or anomalous behaviors. The undesirable or anomalous behaviors may include a communication-based anomaly or an execution-based anomaly, which, for example, could (1) alter the functionality of an network device executing application software in an atypical manner (a file is opened by a first process where the file is configured to be opened by a second process and not the first process); (2) alter the functionality of the network device executing that application software without any malicious intent; and/or (3) provide unwanted functionality which may be generally acceptable in another context. Additionally, malware may be code that initiates unwanted behavior which may be, as one example, uploading a contact list from an endpoint device to cloud storage without receiving permission from the user.


The term “shellcode” refers to a small piece of executable code that resides in data (e.g., is injected into data), is used as a payload of malware, or, in some cases, contains a shell command to execute an exploit.


The term “transmission medium” is a physical or logical communication path between two or more network devices (e.g., any devices with data processing and network connectivity such as, for example, a security appliance, a server, a mainframe, a computer such as a desktop or laptop, netbook, tablet, firewall, smart phone, router, switch, bridge, etc.). For instance, the communication path may include wired and/or wireless segments. Examples of wired and/or wireless segments include electrical wiring, optical fiber, cable, bus trace, or a wireless channel using infrared, radio frequency (RF), or any other wired/wireless signaling mechanism.


In certain instances, the term “detected” is used herein to represent that there is a prescribed level of confidence (or probability) on the presence of an exploit within an object under analysis.


The term “pattern” should be construed as a predefined grouping of instructions. A pattern may appear in memory, such as memory allocated in a virtual execution environment for use by an application being executed by a virtual machine. In some embodiments, the length of the pattern may correspond to the operating system of the network device which is undergoing analysis. For example, a pattern may consist of four bytes when the network device is running a 32-bit operating system (this may be referred to as a double word, or “DWORD”). Therefore, the DWORD may contain up to four (4) instructions, which may be four (4) NOP instructions, for example. Alternatively, a pattern may consist of eight bytes when the network device is running a 64-bit operating system (this may be referred to as a quad word, or “QWORD”). Therefore, the QWORD may contain up to eight (8) instructions, which may be eight (8) NOP instructions, for example.


The terms “network device” or “network appliance” should be construed as any electronic device with the capability of connecting to a network. Such a network may be a public network such as the Internet or a private network such as a wireless data telecommunication network, wide area network, a type of local area network (LAN), or a combination of networks. Examples of a network device may also include, but are not limited or restricted to mobile devices, such as a laptop, a mobile phone, a tablet, a computer, etc. or any other relatively portable device.


The term “computerized” generally represents that any corresponding operations are conducted by hardware in combination with software and/or firmware. Also, the terms “compare” or “comparison” generally mean determining if a match (e.g., a certain level of correlation) is achieved between two items where one of the items may include a particular signature pattern.


Lastly, the terms “or” and “and/or” as used herein are to be interpreted as inclusive or meaning any one or any combination. Therefore, “A, B or C” or “A, B and/or C” mean “any of the following: A; B; C; A and B; A and C; B and C; A, B and C.” An exception to this definition will occur only when a combination of elements, functions, steps or acts are in some way inherently mutually exclusive.


As this invention is susceptible to embodiments of many different forms, it is intended that the present disclosure is to be considered as an example of the principles of the invention and not intended to limit the invention to the specific embodiments shown and described.


II. Application-Execution Hijacking Detection System

A. General Architecture of Network Appliance Deploying Application-Execution Hijacking Detection Logic


Referring to FIG. 1A, an exemplary logical representation of an AEH detection environment 102 communicatively coupled to a network 110 via a network interface 120 is shown. The AEH detection environment 102 comprises a server device 105, an optional firewall 115, a client device 125 and an AEH detection system 100 communicatively coupled to the network 110 via a network interface 120. The AEH system 100 is further communicatively coupled to an analysis network 195. It is envisioned that the analysis network 195 may be configured to store and access data regarding malware attacks across a number of objects, including for example, web-based, email-based, and file-based threats. Moreover, it is envisioned that the analysis network 195 may be configured to store historical information regarding previously analyzed and/or known malware attacks. The analysis network 195 may also be periodically or aperiodically updated so as to store information regarding new malware attacks, reports, alerts, and/or corresponding features, for example.


Herein, according to the embodiment illustrated in FIG. 1A, the AEH detection system 100 comprises a network appliance that is adapted to analyze information associated with network traffic routed over a communication network 110 between at least one server device 105 and at least one client device 125. The communication network 110 may include a public network such as the Internet, in which case an optional firewall 115 (represented by dashed lines) may be interposed in the communication path between the public network and the client device 125. Alternatively, the communication network 110 may be a private network such as a wireless data telecommunication network, wide area network, a type of local area network (LAN), or a combination of networks.


As shown, the AEH detection system 100 may be communicatively coupled with the communication network 110 via a network interface 120. In general, the network interface 120 operates as a data-capturing device (sometimes referred to as a “tap” or “network tap”) that is configured to receive data propagating to/from the client device 125 and provide at least some of this data to the AEH detection system 100. Alternatively, it should be understood that the AEH detection system 100 may be positioned behind the firewall 115 and in-line with client device 125.


According to one embodiment of the disclosure, the network interface 120 is capable of receiving and routing objects associated with network traffic to the AEH detection system 100. The network interface 120 may provide the entire object or certain content within the object, for example, one or more files or packets that are part of a set of flows, packet payloads, or the like. In some embodiments, although not shown, network interface 120 may be contained within the AEH detection system 100.


It is contemplated that, for any embodiments where the AEH detection system 100 is implemented as a dedicated appliance or a dedicated computer system, the network interface 120 may include an assembly integrated into the appliance or computer system that includes a network interface card and related logic (not shown) for connecting to the communication network 110 to non-disruptively “tap” network traffic propagating through firewall 115 and provide either a duplicate copy of at least a portion of the network traffic or at least a portion the network traffic itself to the dynamic analysis server 130 and an optional static analysis server, if included within the AEH detection system 100. In other embodiments, the network interface 120 can be integrated into an intermediary device in the communication path (e.g., firewall 115, router, switch or other networked network device, which in some embodiments may be equipped with Switched Port Analyzer “SPAN” ports) or can be a standalone component, such as an appropriate commercially available network tap. In virtual environments, a virtual tap (vTAP) can be used to duplicate files from virtual networks.


As further shown in FIG. 1A, the AEH detection system 100 comprises the dynamic analysis server 130, rule-matching logic 150, and reporting logic 160. In some embodiments, an optional static analysis server may be provided within the AEH detection system 100 so as to perform static scanning on a particular object, namely heuristics, exploit signature checks and/or vulnerability signature checks for example. The optional static analysis server and the dynamic analysis server 130 may be one or more software modules executed by the same processor or different processors, where these different processors may be located within the same processor package (e.g., different processor cores) and/or located at remote or even geographically remote locations that are communicatively coupled (e.g., by a dedicated communication link) or a network.


Herein, the dynamic analysis server 130 comprises a virtual machine manager 140, a database 145 and one or more virtual machines (VMs) VM1-VMN (N≥1) that may be configured to perform in-depth dynamic analysis with respect to one or more suspicious objects. For instance, the dynamic analysis server 130 may simulate transmission and/or receipt of packets or other objects by a destination device comprising the virtual machines to determine whether certain guarded areas of loaded modules are being accessed in an effort to hijack a particular application.


According to one embodiment, each of the VMs (e.g., VM1-VMN) within the dynamic analysis server 130 may be configured with a software profile corresponding to a software image stored within the database 145 that is communicatively coupled with the virtual machine manager 140. Alternatively, the VMs (e.g., VM1-VMN) may be configured according to a prevalent software configuration, software configuration used by a network device within a particular enterprise network (e.g., client device 125), or an environment that is associated with the object to be processed, including software such as a web browser application, PDF™ reader application, or the like.


However, for a known vulnerability, the VMs (e.g., VM1-VMN) may be more narrowly configured to profiles associated with vulnerable modules. For example, if the access source comprises a certain memory type, VM1-VMN may be configured for faster processing and corresponding log file generation. Similarly, when relevant, if the access source is attempting to access a guarded page of its own module, then VM1-VMN may be configured accordingly.


In general, the dynamic analysis server 130 is adapted to execute one or more VMs (e.g., VM1-VMN) to detect an application-execution hijacking attempt by simulating the execution of an object under analysis within a run-time environment as expected by the type of object. For example, the dynamic analysis server 130 analyzes the received network traffic and determines which application is suitable for executing an object of the received network traffic within one or more VMs, namely VM1, and/or VMN.


Each of the VMs (VM1-VMN) comprise event logic 135 that is configured so as to detect and/or store all access events, however the event logic 135 may be more narrowly tailored to only focus on access events with respect to guarded areas. The event logic 135 comprises Dynamic Link Library (DLL)/kernel logic 138 that is configured to respond to the occurrence, during computation for example, of exceptions. As used herein, “exceptions” generally refer to anomalous or exceptional conditions requiring special processing that may change the normal flow of a program's execution. It is envisioned that when an attempt is made to access a guarded area, an exception is generated and appropriately handled by the DLL/kernel logic 138. Preferably, the DLL/kernel logic 138 is configured to handle all exceptions that may occur in the event that a guarded area is accessed. In general, the DLL/kernel logic 138 may be considered a library that contains code and data that may be used by more than one programs simultaneously to promote code reuse and efficient memory usage. By using the DLL/kernel logic 138, a program may be modularized into separate components, known as modules. Each module may be loaded into a main program at run time, if that module is loaded. Consequently, as used herein, the DLL/kernel logic 138 utilizes this modular nature to detect access events in conjunction with the AEH detection system 100 to ultimately determine whether the access events are malicious or not.


In one embodiment, the event logic 135 may be configured to generate a log file corresponding to the access events, with special focus being paid to guarded page areas. It is envisioned that the log file may comprise any suitable file format. Once generated, the log file may be communicated to the rule-matching logic 150 so that certain detection logic 155 may be applied thereon.


In one embodiment, the detection logic 155 is configured to apply certain rules on the generated log file. Once the rules are applied, the result is communicated to the reporting logic 160. If maliciousness is found, an alert 190 is generated. The alert 190 may comprise details with respect to the object, such as, by way of non-limiting example, the source of a particular access event, and/or its memory type. In one embodiment, the DLL/kernel logic 138 may be configured to directly block one or more malicious objects by terminating the application and/or process that has been attacked, rather than generating a log file, using the optional process-handling logic 180, for example. It is envisioned that the rules may also be configured so as to perform a probabilistic analysis with respect to some or all of the data associated with the generated log file. For example, the occurrence of an access event may indicate to some level of probability, often well less than 100%, that the access event comprises a certain exploit or exhibits certain elements associated with malware. In one embodiment, the rule-matching logic 150 may be configured to take certain action, including for example, generating an alert 190 if the probability exceeds a prescribed value, for example.


Referring now to FIG. 1B, an exemplary embodiment of a representation of the AEH detection system of FIG. 1A is shown. In one embodiment, a network appliance 106 comprises a housing 103, which is made entirely or partially of a rigid material (e.g., hardened plastic, metal, glass, composite or any combination thereof) that protect circuitry within the housing 103, namely one or more processors 109 that are coupled to communication interface logic 112 via a first transmission medium 116. Communication interface logic 112 enables communications with other AEH detection systems 100 and/or the analysis network 195 of FIG. 1A, for example. According to one embodiment of the disclosure, communication interface logic 112 may be implemented as a physical interface including one or more ports for wired connectors. Additionally, or in the alternative, communication interface logic 112 may be implemented with one or more radio units for supporting wireless communications with other electronic devices.


Processor(s) 109 may further be coupled to persistent storage 122 via a second transmission medium 119. According to one embodiment of the disclosure, persistent storage 122 may include the AEH detection system 100, which in one embodiment comprises (a) dynamic analysis server 130; (b) rule-matching logic 150; (c) optional process handling logic 180; and reporting logic 160. It is envisioned that one or more of these systems (or logic units) could be implemented externally from the AEH detection system 100 without extending beyond the spirit and scope of the present disclosure.


B. General Architecture of a Mobile Device Deploying an Application-Execution Hijacking Detection Application


Referring to FIG. 2, a plurality of mobile devices 230A-230C deploying an AEH detection application communicatively coupled to an analysis network 205 is shown. In general, a network environment 200 is shown, wherein a router 215 is communicatively coupled to analysis network 205 and Internet 210. The router is also communicatively coupled to an optional firewall 220, which itself may be communicatively coupled to a network switch 225. As shown, the plurality of mobile devices 230A-230C may also be communicatively coupled to the analysis network 205 and Internet 210 using any transmission medium, including without limitation, wireless and hardwired connection schemes. It is envisioned that an exemplary AEH detection application 231A-231C corresponding to each of the mobile devices may be installed to detect application-execution hijacking malware. Of course, although only three mobile devices 230A-230C are shown in FIG. 2, any number of devices may have the exemplary AEH detection application 231A-231C loaded thereon. In one embodiment, the AEH detection application may be a downloaded from an app store, website, etc., and thus stored and made locally available with respect to a mobile device.


Referring to FIG. 3, a mobile device may be configured to deploy the AEH detection application 231 of FIG. 2. As shown in FIG. 3, for illustrative purposes, the network device 300 is represented as a mobile device (e.g., smartphone, tablet, laptop computer, netbook, etc.). The mobile device 300 includes a display screen 302, one or more processors 310, a receiver and/or transmitter (e.g. transceiver) such as an antenna 355, and communication interface logic 305. In one embodiment, the AEH detection application 231 comprises exploit detection logic 315, rule-matching logic 330, reporting logic 345, and user interface logic 350.


As shown, the AEH detection application 231 is substantially similar to the AEH detection environment 102 of FIG. 1, except that instead of implementing VMs, the AEH detection application 231 communicates the output of the DLL/kernel logic 315 directly into the AEH detection application 231 running on the mobile device 300. After the output of the DLL/kernel logic 315 is loaded, a log file is generated comprising access events, again with special attention being paid to access events corresponding to guarded page areas. It is envisioned that the log file may be customized so as to only focus on certain criteria. For example, the log file may be filtered according to a particular access source, the memory type being accessed, and in some instances, whether the access source is from the heap. Once the log file is generated, it is communicated to the reporting logic 340, comprising the alert generation logic 345, which may generate an alert and/or route the alert to the analysis network 205 via the communication interface logic 305, for further analysis. In addition, the alert may be routed to the Internet 210 using communication interface logic 305 for further analysis by a network administrator, for example. The reporting logic 340 may issue an alert or report (e.g., an email message, text message, display screen image, etc.) to security administrators or the user, for example, communicating the urgency in handling one or more predicted attacks using the user interface logic 350. The AEH detection application 231 may trigger a further analysis of the object to verify the behavior of the object as an exploit. It is envisioned that the generated alert or report may also comprise instructions so as to prevent one or more predicted malware attacks.


II. Application-Execution Hijacking Detection Methods


FIG. 4 is a flowchart of an exemplary method for detecting AEH malware by applying any of various protection mechanisms, including by way of non-limiting example, applying a page guard to an element of a loaded module so as to establish a protected region. Specifically, as shown at block 405, a page guard is applied to an element of a loaded module, such as by way of non-limiting example, a base address, import table (IT), and/or a process environment block. In general, page guards (also referred to as a “guard page” or “PAGE_GUARD”) provide an alarm for memory page access. Page guards may be used to monitor the growth of large dynamic data structures. For example, certain operating systems use page guards to implement automatic stack checking.


It is to be understood that to create a page guard, the PAGE_GUARD page protection modifier may be set with respect to a particular page. This value may be specified, along with other page protection modifiers, by way of non-limiting example, in various functions such as VirtualAlloc. The PAGE_GUARD modifier may also be used with any other page protection modifiers. In one embodiment, if a code attempts to access an address within a guard page, the system may raise an exception or any other type of violation. Although the principles described herein are especially relevant to the Microsoft® Windows® operating system, many other operating systems and platforms, including without limitation, Apple OS®, iOS®, Android®, Linux®, etc. may also be utilized. At block 410, access events comprising read, write and/or execute operations are monitored and stored, and particular attention is paid with respect to the source of such access attempts. In the event that a guarded area is accessed, an exception is generated at block 415 by the event logic as discussed herein. At block 420, the dynamic analysis server as discussed herein generate a log file comprising access events. At block 440, detection logic is applied to the log file to determine whether an access event is malicious or not. At block 460, after applying the detection logic as discussed in more detail below, a report/alert may be generated. The report/alert may be communicated to the user or a network administrator, for example, and/or stored for further analysis by the analysis network.



FIG. 5 is an exemplary block diagram of an alternative operational flow of an AEH detection system that is substantially similar to the flow described in FIG. 4, except that instead of generating an output file, the potentially malicious object and/or corresponding operation is terminated. For example, at block 505, a page guard is applied to an element of a loaded module, such as by way of non-limiting example, a base address, import table (IT), and/or a process environment block. At block 510, access events are monitored and stored, and particular attention is paid to the source of such access attempts. At block 520, if the access source is determined to be non-malicious, no alert is generated. At block 530, if the access source is indeed malicious, an alert may be generated. Furthermore, if the access source is malicious, the corresponding object and/or process may be terminated, and therefore prevented from executing, as shown at block 535. As stated above, it is envisioned that a probabilistic analysis may also be used in determining whether or not an alert should be generated. For example, the occurrence of an access event may indicate to some level of probability, often well less than 100%, that the access event comprises a certain exploit or exhibits certain elements associated with malware. In one embodiment, the rule-matching logic 150 may be configured to take certain action, including for example, generating an alert if the probability exceeds a prescribed value, for example.



FIGS. 6-9 are flowcharts of exemplary methods for detecting AEH malware, that may be implemented using the rule-matching logic as discussed herein so as to establish a protected region.


Specifically, FIG. 6 shows an exemplary method for detecting AEH malware by applying any of various protection mechanisms such as, by way of non-limiting example, applying a page guard to the base address of a loaded module so as to establish a protected region. At block 605, a loaded module is identified. At block 610, a page guard is applied to the base address of the loaded module as discussed herein, although any number of methods may be used. At block 615, rule-matching logic is configured to determine whether a guarded page is being accessed. If the result of this query is “no”, then there is a high likelihood that the access attempt is not malicious. If the result at block 615 is “yes”, then a second query is presented at block 625.


Specifically, at block 625, the rule-matching logic determines whether the access source is from the heap. In one embodiment, exceptions may comprise context information including for example, values in certain registers at the time the exception was generated that correspond to the address of a source instruction that attempted to access a page guarded area. Once the address of the source instruction is known, a plurality of methods may be used to determine if the address is from the heap. For example, the operating system may provide an API that is configured so as to retrieve information corresponding to the memory at any address. It is envisioned that the API may return a structure that includes information such as memory type. If the memory type is “MEM_IMAGE” then the address is from a loaded module. If the memory type is not “MEM_IMAGE” type, then it is from the heap, and likely malicious. If the result at block 625 is “no”, then there is a high likelihood that the access attempt is not malicious. On the other hand, if the result at block 625 is “yes”, then there is a high likelihood that the access attempt is indeed malicious. In some embodiments, if it is determined that an access attempt is malicious, then an alert may be generated and communicated as discussed herein.



FIG. 7 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a guard page to an import table (IT) of a loaded module. In general, the IT is used as a lookup table when an application is calling a function in a different module. Accordingly, at block 705, a loaded module is identified, and at block 710, a protection mechanism such as a page guard is applied to the IAT of the loaded module as discussed herein, although any number of methods may be used so as to establish a protected region. As indicated at block 710, however, the instant method is only applicable with respect to scriptable DLLs that provide a scripting environment to applications, such as, by way of non-limiting example “jscript”, “vbscript”, and “flash.ocx”. At block 715, the rule-matching logic determines whether a guarded page is being accessed. If the result of this query is “no”, then there is a high likelihood that the access attempt is not malicious. If the result at block 715 is “yes”, then a second query is analyzed at block 725. Specifically, the rule-matching logic determines whether the access source is from the heap. If the result of this query is “yes”, then there is a high likelihood that the access attempt is malicious. If the result of this query is “no”, then a third query is presented at block 740. Specifically, at block 740, the rule-matching logic is configured to determine whether the access source is from the same loaded module, and accessing its own import table. If the result of this query is “no”, then there is a high likelihood that the access attempt is not malicious. On the other hand, if the result of this query is “yes”, then there is a high likelihood than the access attempt is malicious, as indicated at block 735.



FIG. 8 is a flowchart of an exemplary method for detecting AEH malware by applying a hardware breakpoint to a base address of a loaded module. At block 805, a loaded module is identified. At block 810, a hardware breakpoint is applied to the base address of the loaded module. Unlike software breakpoints, hardware breakpoints may be configured to establish breakpoints that are initiated when any instruction attempts to read, write, and/or execute a specific memory address. It should be appreciated, however, that hardware breakpoints have certain limitations, including for example the limited number of hardware breakpoints that may be active simultaneously. In some embodiments, for example, only four hardware may be active at the same time.


At block 815, the rule-matching logic determines whether a hardware breakpoint is being accessed. If the result of this query is “no”, then there is a high likelihood that the access attempt is not malicious. If the result at block 815 is “yes”, then a second query is analyzed at block 825. Specifically, the rule-matching logic determines whether the access source is from the heap. If the result at block 825 is “no”, then there is a high likelihood that the access attempt is not malicious. On the other hand, if the result at block 825 is “yes”, then there is a high likelihood that the access attempt is indeed malicious. In some embodiments, if it is determined that an access attempt is malicious, then an alert may be generated as discussed herein.



FIG. 9 is a flowchart of an exemplary method for detecting application-execution hijacking malware by applying a protection mechanism such as a page guard to a process environment block of a loaded module. It should be understood that the process environment block (PEB) is a data structure that is internally utilized by an operating system, most of whose fields are not intended for use by anything other than the operating system. Conventionally, the PEB comprises data structures that apply across a whole process, including global context, startup parameters, data structures for the program image loader, the program image base address, and synchronization objects used to provide mutual exclusion for process-wide data structures, for example.


Consequently, at block 905, a loaded module is identified. At block 910, a protection mechanism, such as a page guard is applied to the process environment block of the loaded module as discussed herein, although any number of methods may be used so as to establish a protected region. At block 915, the rule-matching logic determines whether a guarded page is being accessed. If the result of this query is “no”, then there is a high likelihood that the access attempt is not malicious. If the result at block 915 is “yes”, then a second query is analyzed at block 925. Specifically, the rule-matching logic determines whether the access source is from the heap. In one embodiment, the operating system may provide an API that is configured so as to retrieve information corresponding to the memory at any address. For example, the API may return an output structure that includes information such as the memory address. Similarly, the PEB may be configured so as to provide a list of all of the loaded modules, including, for example, DLLs. The list of all of the loaded modules from the PEB may be analyzed and compared with the output structure to determine if a particular address is from the heap, and therefore malicious. If the result at block 925 is “no”, then there is a high likelihood that the access attempt is not malicious. On the other hand, if the result at block 925 is “yes”, then there is a high likelihood that the access attempt is malicious. In some embodiments, if it is determined that an AEH attack is present, then an alert may be generated as discussed herein.


It should be understood that unless otherwise indicated, the principles described herein are not exclusive to any particular operating system, and thus, systems and methods may be implemented in and/or executed using many operating systems and platforms, including without limitation, Windows®, Apple OS®, iOS®, Android®, etc.


In the foregoing description, the invention is described with reference to specific exemplary embodiments thereof. However, it will be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention as set forth in the appended claims.

Claims
  • 1. An electronic device, comprising: one or more processors; anda non-transitory storage medium communicatively coupled to the one or more processors, the non-transitory storage medium includes one or more software modules and logic that, upon execution by the one or more processors, performs operations comprising: applying a page guard to an element of a stored software module of the one or more software modules, the page guard being code that detects a potential application-execution hijacking attack,determining whether an access source is attempting to access the stored software module,storing access events associated with the attempted access to the stored software module within a log file,analyzing information associated with each of the stored access events corresponding to the attempted access within the log file by at least (i) conducting, in accordance with prescribed rules, an analysis of the information associated with the stored access events to produce a result, the result including a non-zero probability value of the stored access event is malicious, (ii) comparing the probability value to a prescribed threshold value, the attempted access is determined to be malicious based on, at least in part, a determination that the probability value exceeds a prescribed threshold value, and (iii) determining whether the access source is from a prescribed region of dynamically allocated memory, andgenerating an alert when the result signifies that the attempted access is determined to be malicious, the alert comprises information associated with the access source being an object that, upon execution, is attempting to access the stored software module, the information includes a source of the access events and information associated with the dynamically allocated memory.
  • 2. The electronic device of claim 1, wherein the element of the stored software module comprises a base address of a loaded module.
  • 3. The electronic device of claim 1, wherein the element of the stored software module comprises the import table of the loaded module.
  • 4. The electronic device of claim 1, wherein the element of the stored software module comprises a process environment block.
  • 5. The electronic device of claim 1, wherein the alert to notify a user or a network administrator of the potential application-execution hijacking attack being a malicious access.
  • 6. The electronic device of claim 1, wherein the logic that, upon execution by the one or more processors, further performs the operations comprising: terminating the loaded module so as to prevent an application-execution hijacking attack being a malicious access.
  • 7. The electronic device of claim 1, wherein the prescribed memory region associated with the dynamically allocated memory includes a heap.
  • 8. The electronic device of claim 7, wherein at least one of the access events associated with the attempted access includes an exception, the exception comprises context information including one or more register values at the time the exception was generated and the one or more register values correspond to an address of a source instruction that attempted to access the stored software module guarded by the page guard.
  • 9. The electronic device of claim 8, wherein the attempted access is malicious when the address of the source instruction correspond to an address of a heap.
  • 10. A non-transitory storage medium including logic that, upon execution by the one or more processors, detects an application-execution hijacking, the non-transitory storage medium comprising: logic to identify a loaded module stored in a memory;logic to determine whether an access source is attempting to access the loaded module based on use a page guard applied to an element of the loaded module, the page guard being code that detects an attempted access to an address associated with the loaded module to detect a potential application-execution hijacking attack;logic to raise an exception or memory access violation when a code attempts to access an address within the page guard, the attempted access corresponding to one or more access events;logic to store the one or more access events within a log file,logic to analyze information associated with each of the one or more access events within the log file by at least (i) conducting an analysis, in accordance with prescribed rules, on the information associated with the one or more access events to produce a result, the result including a non-zero probability value of the stored access event is malicious, (ii) comparing the probability value to a prescribed threshold value, the access being determined to be malicious based, at least in part, a determination that the probability value exceeds a prescribed threshold value, and (iii) determining on whether the access source is from a prescribed region of the memory associated with dynamically allocated memory, andlogic to generate an alert when the result signifies that the access is determined to be malicious, the alert comprises information associated with the access source being an object that, upon execution, is attempting to access the prescribed region of the memory, the information includes a source of the access events and a type of the memory.
  • 11. The non-transitory storage medium of claim 10, wherein the element of the loaded module comprises a base address of the loaded module.
  • 12. The non-transitory storage medium of claim 10, wherein the element of the loaded module comprises an import table of the loaded module.
  • 13. The non-transitory storage medium of claim 10, wherein the element of the loaded module comprises a process environment block being a data structure that is internally utilized by an operating system.
  • 14. The non-transitory storage medium of claim 10, wherein the alert to notify a user or a network administrator of the potential application-execution hijacking attack.
  • 15. The non-transitory storage medium of claim 10, wherein the logic that, upon execution by the one or more processors, further comprising: logic to terminate the loaded module so as to prevent an application-execution hijacking attack.
  • 16. The non-transitory storage medium of claim 10, wherein the dynamically allocated memory corresponds to a heap.
  • 17. The non-transitory storage medium of claim 16, wherein the logic to generate the alert comprises logic that, upon execution by the one or more processors, generates the alert to communicate details of an access event.
  • 18. A method comprising: applying a page guard to an element of a stored software module of the one or more software modules, the page guard being code that detects a potential application-execution hijacking attack;determining whether an access source is attempting to access the stored software module; storing access events associated with the attempted access to the stored software module within a log file;analyzing information associated with each of the stored access events corresponding to the attempted access within the log file by at least (i) conducting, in accordance with prescribed rules, an analysis of the information associated with the stored access events to produce a result, the result including a non-zero probability value of the stored access event is malicious, (ii) comparing the probability value to a prescribed threshold value, the attempted access is determined to be malicious based on, at least in part, a determination that the probability value exceeds a prescribed threshold value, and (iii) determining whether the access source is from a prescribed region of dynamically allocated memory; andgenerating an alert when the result signifies that the attempted access is determined to be malicious, the alert comprises information associated with the access source being an object that, upon execution, is attempting to access the stored software module, the information includes a source of the access events and information associated with the dynamically allocated memory.
  • 19. The method of claim 18, wherein the element of the stored software module comprises a base address of the stored software module.
  • 20. The method of claim 18, wherein the element of the stored software module comprises an import table of the stored software module.
  • 21. The method of claim 18, wherein the element of the stored software module comprises a process environment block being a data structure that is internally utilized by an operating system.
  • 22. The method of claim 18, wherein the alert to notify a user or a network administrator of the potential application-execution hijacking attack.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/871,987 filed Sep. 30, 2015, now U.S. Pat. No. 10,210,329, issued Feb. 19, 2019, the entire contents of which is incorporated herein by reference.

US Referenced Citations (717)
Number Name Date Kind
4292580 Ott et al. Sep 1981 A
5175732 Hendel et al. Dec 1992 A
5319776 Hile et al. Jun 1994 A
5398196 Chambers Mar 1995 A
5440723 Arnold et al. Aug 1995 A
5490249 Miller Feb 1996 A
5657473 Killean et al. Aug 1997 A
5802277 Cowlard Sep 1998 A
5842002 Schnurer et al. Nov 1998 A
5960170 Chen et al. Sep 1999 A
5978917 Chi Nov 1999 A
5983348 Ji Nov 1999 A
6088803 Tso et al. Jul 2000 A
6092194 Touboul Jul 2000 A
6094677 Capek et al. Jul 2000 A
6108799 Boulay et al. Aug 2000 A
6154844 Touboul et al. Nov 2000 A
6269330 Cidon et al. Jul 2001 B1
6272641 Ji Aug 2001 B1
6279113 Vaidya Aug 2001 B1
6298445 Shostack et al. Oct 2001 B1
6357008 Nachenberg Mar 2002 B1
6424627 Sørhaug et al. Jul 2002 B1
6442696 Wray et al. Aug 2002 B1
6484315 Ziese Nov 2002 B1
6487666 Shanklin et al. Nov 2002 B1
6493756 O'Brien et al. Dec 2002 B1
6550012 Villa et al. Apr 2003 B1
6775657 Baker Aug 2004 B1
6831893 Ben Nun et al. Dec 2004 B1
6832367 Choi et al. Dec 2004 B1
6895550 Kanchirayappa et al. May 2005 B2
6898632 Gordy et al. May 2005 B2
6907396 Muttik et al. Jun 2005 B1
6941348 Petry et al. Sep 2005 B2
6971097 Wallman Nov 2005 B1
6981279 Arnold et al. Dec 2005 B1
7007107 Ivchenko et al. Feb 2006 B1
7028179 Anderson et al. Apr 2006 B2
7043757 Hoefelmeyer et al. May 2006 B2
7058822 Edery et al. Jun 2006 B2
7069316 Gryaznov Jun 2006 B1
7080407 Zhao et al. Jul 2006 B1
7080408 Pak et al. Jul 2006 B1
7093002 Wolff et al. Aug 2006 B2
7093239 van der Made Aug 2006 B1
7096498 Judge Aug 2006 B2
7100201 Izatt Aug 2006 B2
7107617 Hursey et al. Sep 2006 B2
7159149 Spiegel et al. Jan 2007 B2
7213260 Judge May 2007 B2
7231667 Jordan Jun 2007 B2
7240364 Branscomb et al. Jul 2007 B1
7240368 Roesch et al. Jul 2007 B1
7243371 Kasper et al. Jul 2007 B1
7249175 Donaldson Jul 2007 B1
7287278 Liang Oct 2007 B2
7308716 Danford et al. Dec 2007 B2
7328453 Merkle, Jr. et al. Feb 2008 B2
7346486 Ivancic et al. Mar 2008 B2
7356736 Natvig Apr 2008 B2
7386888 Liang et al. Jun 2008 B2
7392542 Bucher Jun 2008 B2
7418729 Szor Aug 2008 B2
7428300 Drew et al. Sep 2008 B1
7441272 Durham et al. Oct 2008 B2
7448084 Apap et al. Nov 2008 B1
7458098 Judge et al. Nov 2008 B2
7464404 Carpenter et al. Dec 2008 B2
7464407 Nakae et al. Dec 2008 B2
7467408 O'Toole, Jr. Dec 2008 B1
7478428 Thomlinson Jan 2009 B1
7480773 Reed Jan 2009 B1
7484239 Tester Jan 2009 B1
7487543 Arnold et al. Feb 2009 B2
7496960 Chen et al. Feb 2009 B1
7496961 Zimmer et al. Feb 2009 B2
7519990 Xie Apr 2009 B1
7523493 Liang et al. Apr 2009 B2
7530104 Thrower et al. May 2009 B1
7540025 Tzadikario May 2009 B2
7546638 Anderson et al. Jun 2009 B2
7565550 Liang et al. Jul 2009 B2
7568233 Szor et al. Jul 2009 B1
7584455 Ball Sep 2009 B2
7603715 Costa et al. Oct 2009 B2
7607171 Marsden et al. Oct 2009 B1
7639714 Stolfo et al. Dec 2009 B2
7644441 Schmid et al. Jan 2010 B2
7657419 van der Made Feb 2010 B2
7676841 Sobchuk et al. Mar 2010 B2
7698548 Shelest et al. Apr 2010 B2
7707633 Danford et al. Apr 2010 B2
7712136 Sprosts et al. May 2010 B2
7730011 Deninger et al. Jun 2010 B1
7739740 Nachenberg et al. Jun 2010 B1
7779463 Stolfo et al. Aug 2010 B2
7784097 Stolfo et al. Aug 2010 B1
7832008 Kraemer Nov 2010 B1
7836502 Zhao et al. Nov 2010 B1
7849506 Dansey et al. Dec 2010 B1
7854007 Sprosts et al. Dec 2010 B2
7869073 Oshima Jan 2011 B2
7877803 Enstone et al. Jan 2011 B2
7904959 Sidiroglou et al. Mar 2011 B2
7908660 Bahl Mar 2011 B2
7930738 Petersen Apr 2011 B1
7937387 Frazier et al. May 2011 B2
7937761 Bennett May 2011 B1
7949849 Lowe et al. May 2011 B2
7996556 Raghavan et al. Aug 2011 B2
7996836 McCorkendale et al. Aug 2011 B1
7996904 Chiueh et al. Aug 2011 B1
7996905 Arnold et al. Aug 2011 B2
8006305 Aziz Aug 2011 B2
8010667 Zhang et al. Aug 2011 B2
8020206 Hubbard et al. Sep 2011 B2
8028338 Schneider et al. Sep 2011 B1
8042184 Batenin Oct 2011 B1
8045094 Teragawa Oct 2011 B2
8045458 Alperovitch et al. Oct 2011 B2
8069484 McMillan et al. Nov 2011 B2
8087086 Lai et al. Dec 2011 B1
8171553 Aziz et al. May 2012 B2
8176049 Deninger et al. May 2012 B2
8176480 Spertus May 2012 B1
8201246 Wu et al. Jun 2012 B1
8204984 Aziz et al. Jun 2012 B1
8214905 Doukhvalov et al. Jul 2012 B1
8220055 Kennedy Jul 2012 B1
8225288 Miller et al. Jul 2012 B2
8225373 Kraemer Jul 2012 B2
8233882 Rogel Jul 2012 B2
8234640 Fitzgerald et al. Jul 2012 B1
8234709 Viljoen et al. Jul 2012 B2
8239944 Nachenberg et al. Aug 2012 B1
8260914 Ranjan Sep 2012 B1
8266091 Gubin et al. Sep 2012 B1
8286251 Eker et al. Oct 2012 B2
8291499 Aziz et al. Oct 2012 B2
8307435 Mann et al. Nov 2012 B1
8307443 Wang et al. Nov 2012 B2
8312545 Tuvell et al. Nov 2012 B2
8321936 Green et al. Nov 2012 B1
8321941 Tuvell et al. Nov 2012 B2
8332571 Edwards, Sr. Dec 2012 B1
8365286 Poston Jan 2013 B2
8365297 Parshin et al. Jan 2013 B1
8370938 Daswani et al. Feb 2013 B1
8370939 Zaitsev et al. Feb 2013 B2
8375444 Aziz et al. Feb 2013 B2
8381299 Stolfo et al. Feb 2013 B2
8402529 Green et al. Mar 2013 B1
8464340 Ahn et al. Jun 2013 B2
8479174 Chiriac Jul 2013 B2
8479276 Vaystikh et al. Jul 2013 B1
8479291 Bodke Jul 2013 B1
8510827 Leake et al. Aug 2013 B1
8510828 Guo et al. Aug 2013 B1
8510842 Amit et al. Aug 2013 B2
8515075 Saraf Aug 2013 B1
8516478 Edwards et al. Aug 2013 B1
8516590 Ranadive et al. Aug 2013 B1
8516593 Aziz Aug 2013 B2
8522348 Chen et al. Aug 2013 B2
8528086 Aziz Sep 2013 B1
8533824 Hutton et al. Sep 2013 B2
8539582 Aziz et al. Sep 2013 B1
8549638 Aziz Oct 2013 B2
8555391 Demir et al. Oct 2013 B1
8561177 Aziz et al. Oct 2013 B1
8566476 Shiffer et al. Oct 2013 B2
8566946 Aziz et al. Oct 2013 B1
8584094 Dadhia et al. Nov 2013 B2
8584234 Sobel et al. Nov 2013 B1
8584239 Aziz et al. Nov 2013 B2
8595834 Xie et al. Nov 2013 B2
8627476 Satish et al. Jan 2014 B1
8635696 Aziz Jan 2014 B1
8682054 Xue et al. Mar 2014 B2
8682812 Ranjan Mar 2014 B1
8689333 Aziz Apr 2014 B2
8695096 Zhang Apr 2014 B1
8713294 Gooding et al. Apr 2014 B2
8713631 Pavlyushchik Apr 2014 B1
8713681 Silberman et al. Apr 2014 B2
8726392 McCorkendale et al. May 2014 B1
8739280 Chess et al. May 2014 B2
8776229 Aziz Jul 2014 B1
8782792 Bodke Jul 2014 B1
8789172 Stolfo et al. Jul 2014 B2
8789178 Kejriwal et al. Jul 2014 B2
8793278 Frazier et al. Jul 2014 B2
8793787 Ismael et al. Jul 2014 B2
8805947 Kuzkin et al. Aug 2014 B1
8806647 Daswani et al. Aug 2014 B1
8832829 Manni et al. Sep 2014 B2
8850570 Ramzan Sep 2014 B1
8850571 Staniford et al. Sep 2014 B2
8881234 Narasimhan et al. Nov 2014 B2
8881271 Butler, II Nov 2014 B2
8881282 Aziz et al. Nov 2014 B1
8898788 Aziz et al. Nov 2014 B1
8935779 Manni et al. Jan 2015 B2
8949257 Shiffer et al. Feb 2015 B2
8984638 Aziz et al. Mar 2015 B1
8990939 Staniford et al. Mar 2015 B2
8990944 Singh et al. Mar 2015 B1
8997219 Staniford et al. Mar 2015 B2
9009822 Ismael et al. Apr 2015 B1
9009823 Ismael et al. Apr 2015 B1
9027135 Aziz May 2015 B1
9071638 Aziz et al. Jun 2015 B1
9104867 Thioux et al. Aug 2015 B1
9106630 Frazier et al. Aug 2015 B2
9106694 Aziz et al. Aug 2015 B2
9118715 Staniford et al. Aug 2015 B2
9159035 Ismael et al. Oct 2015 B1
9171160 Vincent et al. Oct 2015 B2
9176843 Ismael et al. Nov 2015 B1
9189627 Islam Nov 2015 B1
9195829 Goradia et al. Nov 2015 B1
9197664 Aziz et al. Nov 2015 B1
9223972 Vincent et al. Dec 2015 B1
9225740 Ismael et al. Dec 2015 B1
9241010 Bennett et al. Jan 2016 B1
9251343 Vincent et al. Feb 2016 B1
9262635 Paithane et al. Feb 2016 B2
9268936 Butler Feb 2016 B2
9275229 LeMasters Mar 2016 B2
9282109 Aziz et al. Mar 2016 B1
9292686 Ismael et al. Mar 2016 B2
9294501 Mesdaq et al. Mar 2016 B2
9300686 Pidathala et al. Mar 2016 B2
9306960 Aziz Apr 2016 B1
9306974 Aziz et al. Apr 2016 B1
9311479 Manni et al. Apr 2016 B1
9355247 Thioux et al. May 2016 B1
9356944 Aziz May 2016 B1
9363280 Rivlin et al. Jun 2016 B1
9367681 Ismael et al. Jun 2016 B1
9398028 Karandikar et al. Jul 2016 B1
9413781 Cunningham et al. Aug 2016 B2
9426071 Caldejon et al. Aug 2016 B1
9430646 Mushtaq et al. Aug 2016 B1
9432389 Khalid et al. Aug 2016 B1
9438613 Paithane et al. Sep 2016 B1
9438622 Staniford et al. Sep 2016 B1
9438623 Thioux et al. Sep 2016 B1
9459901 Jung et al. Oct 2016 B2
9467460 Otvagin et al. Oct 2016 B1
9483644 Paithane et al. Nov 2016 B1
9495180 Ismael Nov 2016 B2
9497213 Thompson et al. Nov 2016 B2
9507935 Ismael et al. Nov 2016 B2
9516057 Aziz Dec 2016 B2
9519782 Aziz et al. Dec 2016 B2
9536091 Paithane et al. Jan 2017 B2
9537972 Edwards et al. Jan 2017 B1
9560059 Islam Jan 2017 B1
9565202 Kindlund et al. Feb 2017 B1
9591015 Amin et al. Mar 2017 B1
9591020 Aziz Mar 2017 B1
9594904 Jain et al. Mar 2017 B1
9594905 Ismael et al. Mar 2017 B1
9594912 Thioux et al. Mar 2017 B1
9609007 Rivlin et al. Mar 2017 B1
9626509 Khalid et al. Apr 2017 B1
9628498 Aziz et al. Apr 2017 B1
9628507 Haq et al. Apr 2017 B2
9633134 Ross Apr 2017 B2
9635039 Islam et al. Apr 2017 B1
9641546 Manni et al. May 2017 B1
9654485 Neumann May 2017 B1
9661009 Karandikar et al. May 2017 B1
9661018 Aziz May 2017 B1
9674298 Edwards et al. Jun 2017 B1
9680862 Ismael et al. Jun 2017 B2
9690606 Ha et al. Jun 2017 B1
9690933 Singh et al. Jun 2017 B1
9690935 Shiffer et al. Jun 2017 B2
9690936 Malik et al. Jun 2017 B1
9736179 Ismael Aug 2017 B2
9740857 Ismael et al. Aug 2017 B2
9747446 Pidathala et al. Aug 2017 B1
9756074 Aziz et al. Sep 2017 B2
9773112 Rathor et al. Sep 2017 B1
9781144 Otvagin et al. Oct 2017 B1
9787700 Amin et al. Oct 2017 B1
9787706 Otvagin et al. Oct 2017 B1
9792196 Ismael et al. Oct 2017 B1
9824209 Ismael et al. Nov 2017 B1
9824211 Wilson Nov 2017 B2
9824216 Khalid et al. Nov 2017 B1
9825976 Gomez et al. Nov 2017 B1
9825989 Mehra et al. Nov 2017 B1
9838408 Karandikar et al. Dec 2017 B1
9838411 Aziz Dec 2017 B1
9838416 Aziz Dec 2017 B1
9838417 Khalid et al. Dec 2017 B1
9846776 Paithane et al. Dec 2017 B1
9876701 Caldejon et al. Jan 2018 B1
9888016 Amin et al. Feb 2018 B1
9888019 Pidathala et al. Feb 2018 B1
9910988 Vincent et al. Mar 2018 B1
9912644 Cunningham Mar 2018 B2
9912681 Ismael et al. Mar 2018 B1
9912684 Aziz et al. Mar 2018 B1
9912691 Mesdaq et al. Mar 2018 B2
9912698 Thioux et al. Mar 2018 B1
9916440 Paithane et al. Mar 2018 B1
9921978 Chan et al. Mar 2018 B1
9934376 Ismael Apr 2018 B1
9934381 Kindlund et al. Apr 2018 B1
9946568 Ismael et al. Apr 2018 B1
9954890 Staniford et al. Apr 2018 B1
9973531 Thioux May 2018 B1
10002252 Ismael et al. Jun 2018 B2
10019338 Goradia et al. Jul 2018 B1
10019573 Silberman et al. Jul 2018 B2
10025691 Ismael et al. Jul 2018 B1
10025927 Khalid et al. Jul 2018 B1
10027689 Rathor et al. Jul 2018 B1
10027690 Aziz et al. Jul 2018 B2
10027696 Rivlin et al. Jul 2018 B1
10033747 Paithane et al. Jul 2018 B1
10033748 Cunningham et al. Jul 2018 B1
10033753 Islam et al. Jul 2018 B1
10033759 Kabra et al. Jul 2018 B1
10050998 Singh Aug 2018 B1
10068091 Aziz et al. Sep 2018 B1
10075455 Zafar et al. Sep 2018 B2
10083302 Paithane et al. Sep 2018 B1
10084813 Eyada Sep 2018 B2
10089461 Ha et al. Oct 2018 B1
10097573 Aziz Oct 2018 B1
10104102 Neumann Oct 2018 B1
10108446 Steinberg et al. Oct 2018 B1
10121000 Rivlin et al. Nov 2018 B1
10122746 Manni et al. Nov 2018 B1
10133863 Bu et al. Nov 2018 B2
10133866 Kumar et al. Nov 2018 B1
10146810 Shiffer et al. Dec 2018 B2
10148693 Singh et al. Dec 2018 B2
10165000 Aziz et al. Dec 2018 B1
10169585 Pilipenko et al. Jan 2019 B1
10176321 Abbasi et al. Jan 2019 B2
10181029 Ismael et al. Jan 2019 B1
10191861 Steinberg et al. Jan 2019 B1
10192052 Singh et al. Jan 2019 B1
10198574 Thioux et al. Feb 2019 B1
10200384 Mushtaq et al. Feb 2019 B1
10210329 Malik et al. Feb 2019 B1
10216927 Steinberg Feb 2019 B1
10218740 Mesdaq et al. Feb 2019 B1
10242185 Goradia Mar 2019 B1
20010005889 Albrecht Jun 2001 A1
20010047326 Broadbent et al. Nov 2001 A1
20020018903 Kokubo et al. Feb 2002 A1
20020038430 Edwards et al. Mar 2002 A1
20020091819 Melchione et al. Jul 2002 A1
20020095607 Lin-Hendel Jul 2002 A1
20020116627 Tarbotton et al. Aug 2002 A1
20020144156 Copeland Oct 2002 A1
20020162015 Tang Oct 2002 A1
20020166063 Lachman et al. Nov 2002 A1
20020169952 DiSanto et al. Nov 2002 A1
20020184528 Shevenell et al. Dec 2002 A1
20020188887 Largman et al. Dec 2002 A1
20020194490 Halperin et al. Dec 2002 A1
20030021728 Sharpe et al. Jan 2003 A1
20030074578 Ford et al. Apr 2003 A1
20030084318 Schertz May 2003 A1
20030101381 Mateev et al. May 2003 A1
20030115483 Liang Jun 2003 A1
20030188190 Aaron et al. Oct 2003 A1
20030191957 Hypponen et al. Oct 2003 A1
20030200460 Morota et al. Oct 2003 A1
20030212902 van der Made Nov 2003 A1
20030229801 Kouznetsov et al. Dec 2003 A1
20030237000 Denton et al. Dec 2003 A1
20040003323 Bennett et al. Jan 2004 A1
20040006473 Mills et al. Jan 2004 A1
20040015712 Szor Jan 2004 A1
20040019832 Arnold et al. Jan 2004 A1
20040047356 Bauer Mar 2004 A1
20040083408 Spiegel et al. Apr 2004 A1
20040088581 Brawn et al. May 2004 A1
20040093513 Cantrell et al. May 2004 A1
20040111531 Staniford et al. Jun 2004 A1
20040117478 Triulzi et al. Jun 2004 A1
20040117624 Brandt et al. Jun 2004 A1
20040128355 Chao et al. Jul 2004 A1
20040133777 Kiriansky et al. Jul 2004 A1
20040165588 Pandya Aug 2004 A1
20040236963 Danford et al. Nov 2004 A1
20040243349 Greifeneder et al. Dec 2004 A1
20040249911 Alkhatib et al. Dec 2004 A1
20040255161 Cavanaugh Dec 2004 A1
20040268147 Wiederin et al. Dec 2004 A1
20050005159 Oliphant Jan 2005 A1
20050021740 Bar et al. Jan 2005 A1
20050033960 Vialen et al. Feb 2005 A1
20050033989 Poletto et al. Feb 2005 A1
20050050148 Mohammadioun et al. Mar 2005 A1
20050086523 Zimmer et al. Apr 2005 A1
20050091513 Mitomo et al. Apr 2005 A1
20050091533 Omote et al. Apr 2005 A1
20050091652 Ross et al. Apr 2005 A1
20050108562 Khazan et al. May 2005 A1
20050114663 Cornell et al. May 2005 A1
20050125195 Brendel Jun 2005 A1
20050149726 Joshi et al. Jul 2005 A1
20050157662 Bingham et al. Jul 2005 A1
20050183143 Anderholm et al. Aug 2005 A1
20050201297 Peikari Sep 2005 A1
20050210533 Copeland et al. Sep 2005 A1
20050238005 Chen et al. Oct 2005 A1
20050240781 Gassoway Oct 2005 A1
20050262562 Gassoway Nov 2005 A1
20050265331 Stolfo Dec 2005 A1
20050283839 Cowburn Dec 2005 A1
20060010495 Cohen et al. Jan 2006 A1
20060015416 Hoffman et al. Jan 2006 A1
20060015715 Anderson Jan 2006 A1
20060015747 Van de Ven Jan 2006 A1
20060021029 Brickell et al. Jan 2006 A1
20060021054 Costa et al. Jan 2006 A1
20060031476 Mathes et al. Feb 2006 A1
20060047665 Neil Mar 2006 A1
20060070130 Costea et al. Mar 2006 A1
20060075496 Carpenter et al. Apr 2006 A1
20060095968 Portolani et al. May 2006 A1
20060101516 Sudaharan et al. May 2006 A1
20060101517 Banzhof et al. May 2006 A1
20060117385 Mester et al. Jun 2006 A1
20060123477 Raghavan et al. Jun 2006 A1
20060143709 Brooks et al. Jun 2006 A1
20060150249 Gassen et al. Jul 2006 A1
20060161983 Cothrell et al. Jul 2006 A1
20060161987 Levy-Yurista Jul 2006 A1
20060161989 Reshef et al. Jul 2006 A1
20060164199 Gilde et al. Jul 2006 A1
20060173992 Weber et al. Aug 2006 A1
20060179147 Tran et al. Aug 2006 A1
20060184632 Marino et al. Aug 2006 A1
20060191010 Benjamin Aug 2006 A1
20060221956 Narayan et al. Oct 2006 A1
20060236393 Kramer et al. Oct 2006 A1
20060242709 Seinfeld et al. Oct 2006 A1
20060248519 Jaeger et al. Nov 2006 A1
20060248582 Panjwani et al. Nov 2006 A1
20060251104 Koga Nov 2006 A1
20060288417 Bookbinder et al. Dec 2006 A1
20070006288 Mayfield et al. Jan 2007 A1
20070006313 Porras et al. Jan 2007 A1
20070011174 Takaragi et al. Jan 2007 A1
20070016951 Piccard et al. Jan 2007 A1
20070019286 Kikuchi Jan 2007 A1
20070033645 Jones Feb 2007 A1
20070038943 FitzGerald et al. Feb 2007 A1
20070064689 Shin et al. Mar 2007 A1
20070074169 Chess et al. Mar 2007 A1
20070094730 Bhikkaji et al. Apr 2007 A1
20070101435 Konanka et al. May 2007 A1
20070128855 Cho et al. Jun 2007 A1
20070142030 Sinha et al. Jun 2007 A1
20070143827 Nicodemus et al. Jun 2007 A1
20070156895 Vuong Jul 2007 A1
20070157180 Tillmann et al. Jul 2007 A1
20070157306 Elrod et al. Jul 2007 A1
20070168988 Eisner et al. Jul 2007 A1
20070171824 Ruello et al. Jul 2007 A1
20070174915 Gribble et al. Jul 2007 A1
20070192500 Lum Aug 2007 A1
20070192858 Lum Aug 2007 A1
20070198275 Malden et al. Aug 2007 A1
20070208822 Wang et al. Sep 2007 A1
20070220607 Sprosts et al. Sep 2007 A1
20070240218 Tuvell et al. Oct 2007 A1
20070240219 Tuvell et al. Oct 2007 A1
20070240220 Tuvell et al. Oct 2007 A1
20070240222 Tuvell et al. Oct 2007 A1
20070250930 Aziz et al. Oct 2007 A1
20070256132 Oliphant Nov 2007 A2
20070271446 Nakamura Nov 2007 A1
20080005782 Aziz Jan 2008 A1
20080018122 Zierler et al. Jan 2008 A1
20080028463 Dagon et al. Jan 2008 A1
20080040710 Chiriac Feb 2008 A1
20080046781 Childs et al. Feb 2008 A1
20080066179 Liu Mar 2008 A1
20080072326 Danford et al. Mar 2008 A1
20080077793 Tan et al. Mar 2008 A1
20080080518 Hoeflin et al. Apr 2008 A1
20080086720 Lekel Apr 2008 A1
20080098476 Syversen Apr 2008 A1
20080120722 Sima et al. May 2008 A1
20080134178 Fitzgerald et al. Jun 2008 A1
20080134334 Kim et al. Jun 2008 A1
20080141376 Clausen et al. Jun 2008 A1
20080184367 McMillan et al. Jul 2008 A1
20080184373 Traut et al. Jul 2008 A1
20080189787 Arnold et al. Aug 2008 A1
20080201778 Guo et al. Aug 2008 A1
20080209557 Herley et al. Aug 2008 A1
20080215742 Goldszmidt et al. Sep 2008 A1
20080222729 Chen et al. Sep 2008 A1
20080263665 Ma et al. Oct 2008 A1
20080295172 Bohacek Nov 2008 A1
20080301810 Lehane et al. Dec 2008 A1
20080307524 Singh et al. Dec 2008 A1
20080313738 Enderby Dec 2008 A1
20080320594 Jiang Dec 2008 A1
20090003317 Kasralikar et al. Jan 2009 A1
20090007100 Field et al. Jan 2009 A1
20090013408 Schipka Jan 2009 A1
20090031423 Liu et al. Jan 2009 A1
20090036111 Danford et al. Feb 2009 A1
20090037835 Goldman Feb 2009 A1
20090044024 Oberheide et al. Feb 2009 A1
20090044274 Budko et al. Feb 2009 A1
20090064332 Porras et al. Mar 2009 A1
20090077666 Chen et al. Mar 2009 A1
20090083369 Marmor Mar 2009 A1
20090083855 Apap et al. Mar 2009 A1
20090089879 Wang et al. Apr 2009 A1
20090094697 Provos et al. Apr 2009 A1
20090113425 Ports et al. Apr 2009 A1
20090125976 Wassermann et al. May 2009 A1
20090126015 Monastyrsky et al. May 2009 A1
20090126016 Sobko et al. May 2009 A1
20090133125 Choi et al. May 2009 A1
20090144823 Lamastra et al. Jun 2009 A1
20090158430 Borders Jun 2009 A1
20090172815 Gu et al. Jul 2009 A1
20090187992 Poston Jul 2009 A1
20090193293 Stolfo et al. Jul 2009 A1
20090198651 Shiffer et al. Aug 2009 A1
20090198670 Shiffer et al. Aug 2009 A1
20090198689 Frazier et al. Aug 2009 A1
20090199274 Frazier et al. Aug 2009 A1
20090199296 Xie et al. Aug 2009 A1
20090228233 Anderson et al. Sep 2009 A1
20090241187 Troyansky Sep 2009 A1
20090241190 Todd et al. Sep 2009 A1
20090265692 Godefroid et al. Oct 2009 A1
20090271867 Zhang Oct 2009 A1
20090300415 Zhang et al. Dec 2009 A1
20090300761 Park et al. Dec 2009 A1
20090328185 Berg et al. Dec 2009 A1
20090328221 Blumfield et al. Dec 2009 A1
20100005146 Drako et al. Jan 2010 A1
20100011205 McKenna Jan 2010 A1
20100017546 Poo et al. Jan 2010 A1
20100030996 Butler, II Feb 2010 A1
20100031353 Thomas et al. Feb 2010 A1
20100037314 Perdisci et al. Feb 2010 A1
20100043073 Kuwamura Feb 2010 A1
20100054278 Stolfo et al. Mar 2010 A1
20100058474 Hicks Mar 2010 A1
20100064044 Nonoyama Mar 2010 A1
20100077481 Polyakov et al. Mar 2010 A1
20100083376 Pereira et al. Apr 2010 A1
20100115621 Staniford et al. May 2010 A1
20100132038 Zaitsev May 2010 A1
20100154056 Smith et al. Jun 2010 A1
20100180344 Malyshev et al. Jul 2010 A1
20100192223 Ismael et al. Jul 2010 A1
20100220863 Dupaquis et al. Sep 2010 A1
20100235831 Dittmer Sep 2010 A1
20100251104 Massand Sep 2010 A1
20100281102 Chinta et al. Nov 2010 A1
20100281541 Stolfo et al. Nov 2010 A1
20100281542 Stolfo et al. Nov 2010 A1
20100287260 Peterson et al. Nov 2010 A1
20100299754 Amit et al. Nov 2010 A1
20100306173 Frank Dec 2010 A1
20110004737 Greenebaum Jan 2011 A1
20110025504 Lyon et al. Feb 2011 A1
20110041179 Ståhlberg Feb 2011 A1
20110047594 Mahaffey et al. Feb 2011 A1
20110047620 Mahaffey et al. Feb 2011 A1
20110055907 Narasimhan et al. Mar 2011 A1
20110078794 Manni et al. Mar 2011 A1
20110093951 Aziz Apr 2011 A1
20110099620 Stavrou et al. Apr 2011 A1
20110099633 Aziz Apr 2011 A1
20110099635 Silberman et al. Apr 2011 A1
20110113231 Kaminsky May 2011 A1
20110119445 Gooding et al. May 2011 A1
20110145918 Jung et al. Jun 2011 A1
20110145920 Mahaffey et al. Jun 2011 A1
20110145934 Abramovici et al. Jun 2011 A1
20110167493 Song et al. Jul 2011 A1
20110167494 Bowen et al. Jul 2011 A1
20110173213 Frazier et al. Jul 2011 A1
20110173460 Ito et al. Jul 2011 A1
20110219449 St. Neitzel et al. Sep 2011 A1
20110219450 McDougal et al. Sep 2011 A1
20110225624 Sawhney et al. Sep 2011 A1
20110225655 Niemela et al. Sep 2011 A1
20110247072 Staniford et al. Oct 2011 A1
20110265182 Peinado et al. Oct 2011 A1
20110289582 Kejriwal et al. Nov 2011 A1
20110302587 Nishikawa et al. Dec 2011 A1
20110307954 Melnik et al. Dec 2011 A1
20110307955 Kaplan et al. Dec 2011 A1
20110307956 Yermakov et al. Dec 2011 A1
20110314546 Aziz et al. Dec 2011 A1
20120023593 Puder et al. Jan 2012 A1
20120054869 Yen et al. Mar 2012 A1
20120066698 Yanoo Mar 2012 A1
20120079596 Thomas et al. Mar 2012 A1
20120084859 Radinsky et al. Apr 2012 A1
20120096553 Srivastava et al. Apr 2012 A1
20120110667 Zubrilin et al. May 2012 A1
20120117652 Manni et al. May 2012 A1
20120121154 Xue et al. May 2012 A1
20120124426 Maybee et al. May 2012 A1
20120174186 Aziz et al. Jul 2012 A1
20120174196 Bhogavilli et al. Jul 2012 A1
20120174218 McCoy et al. Jul 2012 A1
20120198279 Schroeder Aug 2012 A1
20120210423 Friedrichs et al. Aug 2012 A1
20120222121 Staniford et al. Aug 2012 A1
20120255015 Sahita et al. Oct 2012 A1
20120255017 Sallam Oct 2012 A1
20120260342 Dube et al. Oct 2012 A1
20120266244 Green et al. Oct 2012 A1
20120278886 Luna Nov 2012 A1
20120297489 Dequevy Nov 2012 A1
20120330801 McDougal et al. Dec 2012 A1
20120331553 Aziz et al. Dec 2012 A1
20130014259 Gribble et al. Jan 2013 A1
20130036472 Aziz Feb 2013 A1
20130047257 Aziz Feb 2013 A1
20130074185 McDougal et al. Mar 2013 A1
20130086684 Mohler Apr 2013 A1
20130097699 Balupari et al. Apr 2013 A1
20130097706 Titonis et al. Apr 2013 A1
20130111587 Goel et al. May 2013 A1
20130117852 Stute May 2013 A1
20130117855 Kim et al. May 2013 A1
20130139264 Brinkley et al. May 2013 A1
20130160125 Likhachev et al. Jun 2013 A1
20130160127 Jeong et al. Jun 2013 A1
20130160130 Mendelev et al. Jun 2013 A1
20130160131 Madou et al. Jun 2013 A1
20130167236 Sick Jun 2013 A1
20130174214 Duncan Jul 2013 A1
20130185789 Hagiwara et al. Jul 2013 A1
20130185795 Winn et al. Jul 2013 A1
20130185798 Saunders et al. Jul 2013 A1
20130191915 Antonakakis et al. Jul 2013 A1
20130196649 Paddon et al. Aug 2013 A1
20130227691 Aziz et al. Aug 2013 A1
20130246370 Bartram et al. Sep 2013 A1
20130247186 LeMasters Sep 2013 A1
20130263260 Mahaffey et al. Oct 2013 A1
20130291109 Staniford et al. Oct 2013 A1
20130298243 Kumar et al. Nov 2013 A1
20130318038 Shiffer et al. Nov 2013 A1
20130318073 Shiffer et al. Nov 2013 A1
20130325791 Shiffer et al. Dec 2013 A1
20130325792 Shiffer et al. Dec 2013 A1
20130325871 Shiffer et al. Dec 2013 A1
20130325872 Shiffer et al. Dec 2013 A1
20140032875 Butler Jan 2014 A1
20140053260 Gupta et al. Feb 2014 A1
20140053261 Gupta et al. Feb 2014 A1
20140130158 Wang et al. May 2014 A1
20140137180 Lukacs et al. May 2014 A1
20140169762 Ryu Jun 2014 A1
20140179360 Jackson et al. Jun 2014 A1
20140181131 Ross Jun 2014 A1
20140189687 Jung et al. Jul 2014 A1
20140189866 Shiffer et al. Jul 2014 A1
20140189882 Jung et al. Jul 2014 A1
20140237600 Silberman et al. Aug 2014 A1
20140280245 Wilson Sep 2014 A1
20140283037 Sikorski et al. Sep 2014 A1
20140283063 Thompson et al. Sep 2014 A1
20140328204 Klotsche et al. Nov 2014 A1
20140337836 Ismael Nov 2014 A1
20140344926 Cunningham et al. Nov 2014 A1
20140351935 Shao et al. Nov 2014 A1
20140351941 Teller et al. Nov 2014 A1
20140380473 Bu et al. Dec 2014 A1
20140380474 Paithane et al. Dec 2014 A1
20150007312 Pidathala et al. Jan 2015 A1
20150096022 Vincent et al. Apr 2015 A1
20150096023 Mesdaq et al. Apr 2015 A1
20150096024 Haq et al. Apr 2015 A1
20150096025 Ismael Apr 2015 A1
20150180886 Staniford et al. Jun 2015 A1
20150186645 Aziz et al. Jul 2015 A1
20150199513 Ismael et al. Jul 2015 A1
20150199531 Ismael et al. Jul 2015 A1
20150199532 Ismael et al. Jul 2015 A1
20150220735 Paithane et al. Aug 2015 A1
20150372980 Eyada Dec 2015 A1
20160004869 Ismael et al. Jan 2016 A1
20160006756 Ismael et al. Jan 2016 A1
20160044000 Cunningham Feb 2016 A1
20160127393 Aziz et al. May 2016 A1
20160191547 Zafar et al. Jun 2016 A1
20160191550 Ismael et al. Jun 2016 A1
20160261612 Mesdaq et al. Sep 2016 A1
20160285914 Singh et al. Sep 2016 A1
20160301703 Aziz Oct 2016 A1
20160335110 Paithane et al. Nov 2016 A1
20170083703 Abbasi et al. Mar 2017 A1
20180013770 Ismael Jan 2018 A1
20180048660 Paithane et al. Feb 2018 A1
20180121316 Ismael et al. May 2018 A1
20180288077 Siddiqui et al. Oct 2018 A1
Foreign Referenced Citations (11)
Number Date Country
2439806 Jan 2008 GB
2490431 Oct 2012 GB
0206928 Jan 2002 WO
0223805 Mar 2002 WO
2007117636 Oct 2007 WO
2008041950 Apr 2008 WO
2011084431 Jul 2011 WO
2011112348 Sep 2011 WO
2012075336 Jun 2012 WO
2012145066 Oct 2012 WO
2013067505 May 2013 WO
Non-Patent Literature Citations (82)
Entry
Li et al., A VMM-Based System Call Interposition Framework for Program Monitoring, Dec. 2010, IEEE 16th International Conference on Parallel and Distributed Systems, pp. 706-711.
Liljenstam, Michael, et al., “Simulating Realistic Network Traffic for Worm Warning System Design and Testing”, Institute for Security Technology studies, Dartmouth College (“Liljenstam”), (Oct. 27, 2003).
Lindorfer, Martina, Clemens Kolbitsch, and Paolo Milani Comparetti “Detecting environment-sensitive malware.” Recent Advances in Intrusion Detection. Springer Berlin Heidelberg, 2011.
Lok Kwong et al: “DroidScope: Seamlessly Reconstructing the OS and Dalvik Semantic Views for Dynamic Android Malware Analysis”, Aug. 10, 2012, XP055158513, Retrieved from the Internet: URL:https://www.usenix.org/system/files/conference/usenixsecurity12/sec12--final107.pdf [retrieved on Dec. 15, 2014].
Marchette, David J., “Computer Intrusion Detection and Network Monitoring: A Statistical Viewpoint”, (“Marchette”), (2001).
Margolis, P.E., “Random House Webster's ‘Computer & Internet Dictionary 3rd Edition’”, ISBN 0375703519, (Dec. 1998).
Moore, D., et al., “Internet Quarantine: Requirements for Containing Self-Propagating Code”, INFOCOM, vol. 3, (Mar. 30-Apr. 3, 2003), pp. 1901-1910.
Morales, Jose A., et al., ““Analyzing and exploiting network behaviors of malware.””, Security and Privacy in Communication Networks. Springer Berlin Heidelberg, 2010. 20-34.
Mori, Detecting Unknown Computer Viruses, 2004, Springer-Verlag Berlin Heidelberg.
Natvig, Kurt, “SANDBOXII: Internet”, Virus Bulletin Conference, (“Natvig”), (Sep. 2002).
NetBIOS Working Group. Protocol Standard for a NetBIOS Service on a TCP/UDP transport: Concepts and Methods. STD 19, RFC 1001, Mar. 1987.
Newsome, J., et al., “Dynamic Taint Analysis for Automatic Detection, Analysis, and Signature Generation of Exploits an Commodity Software”, In Proceedings of the 12th Annual Network and Distributed System Security, Symposium (NDSS '05), (Feb. 2005).
Newsome, J., et al., “Polygraph: Automatically Generating Signatures for Polymorphic Worms”, In Proceedings of the IEEE Symposium on Security and Privacy, (May 2005).
Nojiri, D., et al., “Cooperation Response Strategies for Large Scale Attack Mitigation”, DARPA Information Survivability Conference and Exposition, vol. 1, (Apr. 22-24, 2003), pp. 293-302.
Oberheide et al., CloudAV.sub.--N-Version Antivirus in the Network Cloud, 17th USENIX Security Symposium USENIX Security '08 Jul. 28-Aug. 1, 2008 San Jose, CA.
Reiner Sailer, Enriquillo Valdez, Trent Jaeger, Roonald Perez, Leendert van Doorn, John Linwood Griffin, Stefan Berger., sHype: Secure Hypervisor Appraoch to Trusted Virtualized Systems (Feb. 2, 2005) (“Sailer”).
Silicon Defense, “Worm Containment in the Internal Network”, (Mar. 2003), pp. 1-25.
Singh, S., et al., “Automated Worm Fingerprinting”, Proceedings of the ACM/USENIX Symposium on Operating System Design and Implementation, San Francisco, California, (Dec. 2004).
Spitzner, Lance , “Honeypots: Tracking Hackers”, (“Spizner”), (Sep. 17, 2002).
The Sniffers's Guide to Raw Traffic available at: yuba.stanford.edu/.about.casado/pcap/section1.html, (Jan. 6, 2014).
Thomas H. Ptacek, and Timothy N. Newsham , “Insertion, Evasion, and Denial of Service: Eluding Network Intrusion Detection”, Secure Networks, (“Ptacek”), (Jan. 1998).
U.S. Appl. No. 14/871,987, filed Sep. 30, 2015 Non-Final Office Action dated Feb. 23, 2018.
U.S. Appl. No. 14/871,987, filed Sep. 30, 2015 Notice of Allowance dated Sep. 21, 2018.
U.S. Pat. No. 8,171,553 filed Apr. 20, 2006, Inter Parties Review Decision dated Jul. 10, 2015.
U.S. Pat. No. 8,291,499 filed Mar. 16, 2012, Inter Parties Review Decision dated Jul. 10, 2015.
Venezia, Paul, “NetDetector Captures Intrusions”, InfoWorld Issue 27, (“Venezia”), (Jul. 14, 2003).
Vladimir Getov: “Security as a Service in Smart Clouds—Opportunities and Concerns”, Computer Software and Applications Conference (COMPSAC), 2012 IEEE 36th Annual, IEEE, Jul. 16, 2012 (Jul. 16, 2012).
Wahid et al., Characterising the Evolution in Scanning Activity of Suspicious Hosts, Oct. 2009, Third International Conference on Network and System Security, pp. 344-350.
Whyte, et al., “DNS-Based Detection of Scanning Works in an Enterprise Network”, Proceedings of the 12th Annual Network and Distributed System Security Symposium, (Feb. 2005), 15 pages.
Williamson, Matthew M., “Throttling Viruses: Restricting Propagation to Defeat Malicious Mobile Code”, ACSAC Conference, Las Vegas, NV, USA, (Dec. 2002), pp. 1-9.
Yuhei Kawakoya et al: “Memory behavior-based automatic malware unpacking in stealth debugging environment”, Malicious and Unwanted Software (Malware), 2010 5th International Conference on, IEEE, Piscataway, NJ, USA, Oct. 19, 2010, pp. 39-46, XP031833827, ISBN:978-1-4244-8-9353-1.
Zhang et al., The Effects of Threading, Infection Time, and Multiple-Attacker Collaboration on Malware Propagation, Sep. 2009, IEEE 28th International Symposium on Reliable Distributed Systems, pp. 73-82.
“Mining Specification of Malicious Behavior”—Jha et al, UCSB, Sep. 2007 https://www.cs.ucsb.edu/.about.chris/research/doc/esec07.sub.--mining.pdf-.
“Network Security: NetDetector—Network Intrusion Forensic System (NIFS) Whitepaper”, (“NetDetector Whitepaper”), (2003).
“Packet”, Microsoft Computer Dictionary, Microsoft Press, (Mar. 2002), 1 page.
“When Virtual is Better Than Real”, IEEEXplore Digital Library, available at, http://ieeexplore.ieee.org/xpl/articleDetails.iso?reload=true&arnumber=990073, (Dec. 7, 2013).
Abdullah, et al., Visualizing Network Data for Intrusion Detection, 2005 IEEE Workshop on Information Assurance and Security, pp. 100-108.
Adetoye, Adedayo, et al., “Network Intrusion Detection & Response System”, (“Adetoye”) (Sep. 2003).
AltaVista Advanced Search Results, “attack vector identifier”. Http://www.altavista.com/web/results?ltag=ody&pg=aq&aqmode=aqa=Event+Orch- estrator . . . , (Accessed on Sep. 15, 2009).
AltaVista Advanced Search Results. “Event Orchestrator”. Http://www.altavista.com/web/results?ltag=ody&pg=aq&aqmode=aqa=Event+Orch- esrator . . . , (Accessed on Sep. 3, 2009).
Apostolopoulos, George; hassapis, Constantinos; “V-eM: A cluster of Virtual Machines for Robust, Detailed, and High-Performance Network Emulation”, 14th IEEE International Symposium on Modeling, Analysis, and Simulation of Computer and Telecommunication Systems, Sep. 11-14, 2006, pp. 117-126.
Aura, Tuomas, “Scanning electronic documents for personally identifiable information”, Proceedings of the 5th ACM workshop on Privacy in electronic society. ACM, 2006.
Baecher, “The Nepenthes Platform: An Efficient Approach to collect Malware”, Springer-verlaq Berlin Heidelberg, (2006), pp. 165-184.
Baldi, Mario; Risso, Fulvio; “A Framework for Rapid Development and Portable Execution of Packet-Handling Applications”, 5th IEEE International Symposium Processing and Information Technology, Dec. 21, 2005, pp. 233-238.
Bayer, et al., “Dynamic Analysis of Malicious Code”, J Comput Virol, Springer-Verlag, France., (2006), pp. 67-77.
Boubalos, Chris , “extracting syslog data out of raw pcap dumps, seclists.org, Honeypots mailing list archives”, available at http://seclists.org/honeypots/2003/q2/319 (“Boubalos”), (Jun. 5, 2003).
Chaudet, C., et al., “Optimal Positioning of Active and Passive Monitoring Devices”, International Conference on Emerging Networking Experiments and Technologies, Proceedings of the 2005 ACM Conference on Emerging Network Experiment and Technology, CoNEXT '05, Toulousse, France, (Oct. 2005), pp. 71-82.
Chen, P. M. and Noble, B. D., “When Virtual is Better Than Real, Department of Electrical Engineering and Computer Science”, University of Michigan (“Chen”) (2001).
Cisco “Intrusion Prevention for the Cisco ASA 5500-x Series” Data Sheet (2012).
Cisco, Configuring the Catalyst Switched Port Analyzer (SPAN) (“Cisco”), (1992).
Clark, John, Sylvian Leblanc,and Scott Knight. “Risks associated with usb hardware trojan devices used by insiders.” Systems Conference (SysCon), 2011 IEEE International. IEEE, 2011.
Cohen, M.I. , “PyFlag—An advanced network forensic framework”, Digital investigation 5, Elsevier, (2008), pp. S112-S120.
Costa, M., et al., “Vigilante: End-to-End Containment of Internet Worms”, SOSP '05, Association for Computing Machinery, Inc., Brighton U.K., (Oct. 23-26, 2005).
Crandall, J.R. , et al., “Minos:Control Data Attack Prevention Orthogonal to Memory Model”, 37th International Symposium on Microarchitecture, Portland, Oregon, (Dec. 2004).
Deutsch, P. , “Zlib compressed data format specification version 3.3” RFC 1950, (1996).
Didier Stevens, “Malicious PDF Documents Explained”, Security & Privacy, IEEE, IEEE Service Center, Los Alamitos, CA, US, vol. 9, No. 1, Jan. 1, 2011, pp. 80-82, XP011329453, ISSN: 1540-7993, DOI: 10.1109/MSP.2011.14.
Distler, “Malware Analysis: An Introduction”, SANS Institute InfoSec Reading Room, SANS Institute, (2007).
Dunlap, George W., et al., “ReVirt: Enabling Intrusion Analysis through Virtual-Machine Logging and Replay”, Proceeding of the 5th Symposium on Operating Systems Design and Implementation, USENIX Association, (“Dunlap”), (Dec. 9, 2002).
Excerpt regarding First Printing Date for Merike Kaeo, Designing Network Security (“Kaeo”), (2005).
Filiol, Eric, et al., “Combinatorial Optimisation of Worm Propagation on an Unknown Network”, International Journal of Computer Science 2.2 (2007).
FireEye Malware Analysis & Exchange Network, Malware Protection System, FireEye Inc., 2010.
FireEye Malware Analysis, Modern Malware Forensics, FireEye Inc., 2010.
FireEye v.6.0 Security Target, pp. 1-35, Version 1.1, FireEye Inc., May 2011.
Gibler, Clint, et al. AndroidLeaks: automatically detecting potential privacy leaks in android applications on a large scale. Springer Berlin Heidelberg, 2012.
Goel, et al., Reconstructing System State for Intrusion Analysis, Apr. 2008 SIGOPS Operating Systems Review, vol. 42 Issue 3, pp. 21-28.
Gregg Keizer: “Microsoft's HoneyMonkeys Show Patching Windows Works”, Aug. 8, 2005, XP055143386, Retrieved from the Internet: URL:http://www.informationweek.com/microsofts-honeymonkeys-show-patching-windows-works/d/d-id/1035069? [retrieved on Jun. 1, 2016].
Heng Yin et al, Panorama: Capturing System-Wide Information Flow for Malware Detection and Analysis, Research Showcase @ CMU, Carnegie Mellon University, 2007.
Hiroshi Shinotsuka, Malware Authors Using New Techniques to Evade Automated Threat Analysis Systems, Oct. 26, 2012, http://www.symantec.com/connect/blogs/, pp. 1-4.
Hjelmvik, Erik , “Passive Network Security Analysis with NetworkMiner”, (IN)Secure, Issue 18, (Oct. 2008), pp. 1-100.
Idika et al., A-Survey-of-Malware-Detection-Techniques, Feb. 2, 2007, Department of Computer Science, Purdue University.
IEEE Xplore Digital Library Sear Results for “detection of unknown computer worms”. Http//ieeexplore.ieee.org/searchresult.jsp?SortField=Score&SortOrder=desc- &ResultC . . . , (Accessed on Aug. 28, 2009).
Isohara, Takamasa, Keisuke Takemori, and Ayumu Kubota. “Kernel-based behavior analysis for android malware detection.” Computational intelligence and Security (CIS), 2011 Seventh International Conference on. IEEE, 2011.
Kaeo, Merike , “Designing Network Security”, (“Kaeo”), (Nov. 2003).
Kevin A Roundy et al: “Hybrid Analysis and Control of Malware”, Sep. 15, 2010, Recent Advances in Intrusion Detection, Springer Berlin Heidelberg, Berlin, Heidelberg, pp. 317-338, XP019150454 ISBN:978-3-642-15511-6.
Khaled Salah et al: “Using Cloud Computing to Implement a Security Overlay Network”, Security & Privacy, IEEE, IEEE Service Center, Los Alamitos, CA, US, vol. 11, No. 1, Jan. 1, 2013 (Jan. 1, 2013).
Kim, H., et al., “Autograph: Toward Automated, Distributed Worm Signature Detection”, Proceedings of the 13th Usenix Security Symposium (Security 2004), San Diego, (Aug. 2004), pp. 271-286.
King, Samuel T., et al., “Operating System Support for Virtual Machines”, (“King”) (2003).
Krasnyansky, Max , et al., Universal TUN/TAP driver, available at https://www.kernel.org/doc/Documentation/networking/tuntap.txt (2002) (“Krasnyansky”).
Kreibich, C., et al., “Honeycomb-Creating Intrusion Detection Signatures Using Honeypots”, 2nd Workshop on Hot Topics in Networks (HotNets-11), Boston, USA, (2003).
Kristoff, J. , “Botnets, Detection and Mitigation: DNS-Based Techniques”, NU Security Day, (2005), 23 pages.
Lastline Labs, The Threat of Evasive Malware, Feb. 25, 2013, Lastline Labs, pp. 1-8.
Leading Colleges Select FireEye to Stop Malware-Related Data Breaches, FireEye Inc., 2009.
Continuations (1)
Number Date Country
Parent 14871987 Sep 2015 US
Child 16277907 US