The present invention relates to a computer-based system for capturing security events from heterogeneous and homogenous sources, and specifically to correlating a number of security events.
Computer networks and systems have become indispensable tools for modern business. Today terabits of information on virtually every subject imaginable are stored in and accessed across such networks by users throughout the world. Much of this information is, to some degree, confidential and its protection is required. Not surprisingly then, intrusion detection systems (IDS) have been developed to help uncover attempts by unauthorized persons and/or devices to gain access to computer networks and the information stored therein. In addition, network devices such as routers and firewalls maintain activity logs that can be used to examine such attempts.
Intrusion detection may be regarded as the art of detecting inappropriate, incorrect or anomalous activity within or concerning a computer network or system. The most common approaches to intrusion detection are statistical anomaly detection and pattern-matching detection. IDS that operate on a host to detect malicious activity on that host are called host-based IDS (HIDS), which may exist in the form of host wrappers/personal firewalls or agent-based software, and those that operate on network data flows are called network-based IDS (NIDS). Host-based intrusion detection involves loading software on the system (the host) to be monitored and using log files and/or the host's auditing agents as sources of data. In contrast, a network-based intrusion detection system monitors the traffic on its network segment and uses that traffic as a data source. Packets captured by the network interface cards are considered to be of interest if they match a signature.
Regardless of the data source, there are two complementary approaches to detecting intrusions: knowledge-based approaches and behavior-based approaches. Almost all IDS tools in use today are knowledge-based. Knowledge-based intrusion detection techniques involve comparing the captured data to information regarding known techniques to exploit vulnerabilities. When a match is detected, an alarm is triggered. Behavior-based intrusion detection techniques, on the other hand, attempt to spot intrusions by observing deviations from normal or expected behaviors of the system or the users (models of which are extracted from reference information collected by various means). When a suspected deviation is observed, an alarm is generated.
Advantages of the knowledge-based approaches are that they have the potential for very low false alarm rates, and the contextual analysis proposed by the intrusion detection system is detailed, making it easier for a security officer using such an intrusion detection system to take preventive or corrective action. Drawbacks include the difficulty in gathering the required information on the known attacks and keeping it up to date with new vulnerabilities and environments.
Advantages of behavior-based approaches are that they can detect attempts to exploit new and unforeseen vulnerabilities. They are also less dependent on system specifics. However, the high false alarm rate is generally cited as a significant drawback of these techniques and because behaviors can change over time, the incidence of such false alarms can increase.
Regardless of whether a host-based or a network-based implementation is adopted and whether that implementation is knowledge-based or behavior-based, an intrusion detection system is only as useful as its ability to discriminate between normal system usage and true intrusions (accompanied by appropriate alerts). If intrusions can be detected and the appropriate personnel notified in a prompt fashion, measures can be taken to avoid compromises to the protected system. Otherwise such safeguarding cannot be provided. Accordingly, what is needed is a system that can provide accurate and timely intrusion detection and alert generation so as to effectively combat attempts to compromise a computer network or system.
Different network segments can have overlapping address spaces. In one embodiment, the present invention includes a distributed agent of a security system receiving a security event from a network device monitored by the agent. In one embodiment, the agent normalizes the security event into an event schema including one or more zone fields. In one embodiment, the agent also determines one or more zones associated with the received security event, the one or more zones each describing a part of a network, and populates the one or more zone fields using the determined one or more zones.
The present invention is illustrated by way of example, and not limitation, in the figures of the accompanying drawings in which:
Although the present system will be discussed with reference to various illustrated examples, these examples should not be read to limit the broader spirit and scope of the present invention. For example, the examples presented herein describe distributed agents, managers and consoles, which are but one embodiment of the present invention. The general concepts and reach of the present invention are much broader and may extend to any computer-based or network-based security system. Also, examples of the messages that may be passed to and from the components of the system and the data schemas that may be used by components of the system are given in an attempt to further describe the present invention, but are not meant to be all-inclusive examples and should not be regarded as such.
Some portions of the detailed description that follows are presented in terms of algorithms and symbolic representations of operations on data within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the computer science arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers or the like. It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise, it will be appreciated that throughout the description of the present invention, use of terms such as “processing”, “computing”, “calculating”, “determining”, “displaying” or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
As indicated above, one embodiment of the present invention is instantiated in computer software, that is, computer readable instructions, which, when executed by one or more computer processors/systems, instruct the processors/systems to perform the designated actions. Such computer software may be resident in one or more computer readable media, such as hard drives, CD-ROMs, DVD-ROMs, read-only memory, read-write memory and so on. Such software may be distributed on one or more of these media, or may be made available for download across one or more computer networks (e.g., the Internet). Regardless of the format, the computer programming, rendering and processing techniques discussed herein are simply examples of the types of programming, rendering and processing techniques that may be used to implement aspects of the present invention. These examples should in no way limit the present invention, which is best understood with reference to the claims that follow this description.
Referring now to
Agents 12 are software programs that provide efficient, real-time (or near real-time) local event data capture and filtering from a variety of network security devices and/or applications. The primary sources of security events are common network security devices, such as firewalls, intrusion detection systems and operating system logs. Agents 12 can collect events from any source that produces event logs or messages and can operate at the native device, at consolidation points within the network, and/or through simple network management protocol (SNMP) traps.
Agents 12 are configurable through both manual and automated processes and via associated configuration files. Each agent 12 may include one or more software modules including a normalizing component, a time correction component, an aggregation component, a batching component, a resolver component, a transport component, and/or additional components. These components may be activated and/or deactivated through appropriate commands in the configuration file.
Managers 14 may be server-based components that further consolidate, filter and cross-correlate events received from the agents, employing a rules engine 18 and a centralized event database 20. One role of manager 14 is to capture and store all of the real-time and historic event data to construct (via database manager 22) a complete, enterprise-wide picture of security activity. The manager 14 also provides centralized administration, notification (through one or more notifiers 24), and reporting, as well as a knowledge base 28 and case management workflow. The manager 14 may be deployed on any computer hardware platform and one embodiment utilizes a relational database management system such as an Oracle™ database to implement the event data store component. Communications between manager 14 and agents 12 may be bi-directional (e.g., to allow manager 14 to transmit commands to the platforms hosting agents 12) and encrypted. In some installations, managers 14 may act as concentrators for multiple agents 12 and can forward information to other managers (e.g., deployed at a corporate headquarters).
Consoles 16 are computer-(e.g., workstation-) based applications that allow security professionals to perform day-to-day administrative and operation tasks such as event monitoring, rules authoring, incident investigation and reporting. Access control lists allow multiple security professionals to use the same system and event database, with each having their own views, correlation rules, alerts, reports and knowledge base appropriate to their responsibilities. A single manager 14 can support multiple consoles 16.
In some embodiments, a browser-based version of the console 16 may be used to provide access to security events, knowledge base articles, reports, notifications and cases. That is, the manager 14 may include a web server component accessible via a web browser hosted on a personal or handheld computer (which takes the place of console 16) to provide some or all of the functionality of a console 16. Browser access is particularly useful for security professionals that are away from the consoles 16 and for part-time users. Communication between consoles 16 and manager 14 is bi-directional and may be encrypted.
Through the above-described architecture the present invention can support a centralized or decentralized environment. This is useful because an organization may want to implement a single instance of system 10 and use an access control list to partition users. Alternatively, the organization may choose to deploy separate systems 10 for each of a number of groups and consolidate the results at a “master” level. Such a deployment can also achieve a “follow-the-sun” arrangement where geographically dispersed peer groups collaborate with each other by passing primary oversight responsibility to the group currently working standard business hours. Systems 10 can also be deployed in a corporate hierarchy where business divisions work separately and support a rollup to a centralized management function.
The exemplary network security system illustrated in
The agents 12 described above are configured, in one embodiment, to perform various pre-correlation processing on the security events they observe at their respective monitor devices. An agent 12, for example, can normalize observed events (i.e., map events to some universal schema used by the network security system 10), aggregate events to save memory and bandwidth, and batch events for efficient transmission. Such agent 12 functionalities, and others, are described in further detail in U.S. application Ser. No. 10/308,584, entitled “Method for Aggregating Events to be Reported by software agent”, filed Dec. 2, 2002, which is hereby incorporated fully by reference.
Another configuration of the network security system 10 is illustrated by a simplified diagram in
In one embodiment, the Denver LAN 30 shares an address space with the Austin LAN 32. Since IP addresses are scarce and/or expensive, many companies reuse the same address range in two or more network segments. Using network address translation (“NAT” also referred to as “natting”) implemented for example in routers 34 and 46, packets can be routed off the local network segments without confusion. However, the Manager 14 may have difficulty distinguishing IP addresses contained in security event fields.
For example, agent 12(d) may collect an attack by machine 38 targeted at fax 36, while agent 12(e) may collect an attack by machine 52 targeted at machine 50. If machine 52 has the same IP address as machine 38, then the source IP of both security events representing the attacks will be the same. This may cause confusion and possible faulty correlation at the manager 14.
Various issues related to address translation are overcome in one embodiment of the present invention using zone labeling. In one embodiment, a zone describes a part of the network, such as “Denver LAN.” Zones may be on a smaller scale as well, or sub-zones can be further defined, such as “Denver: Engineering.” Any range of IP addresses, or any collection of non-consecutive IP addresses can be designated as a zone.
In one embodiment, zone labeling is performed by the agents 12. In one embodiment, zone labeling can be a part of the normalization process, but it may be performed at any time during event processing. In one embodiment, each security event has one zone field to be populated by a label of the zone that the monitor device and the agent 12 monitor. For example, agent 12(e) would label each event as “Austin Zone.”
In another embodiment, multiple zone fields can identify various zones associated with the security event. In one embodiment, a security event includes the zones of the source of the event, the destination of the event, the monitor device that is responsible for the original event, and the agent 12 that generated the normalized event. These zones can be used to populate event fields having some descriptive name, such as “Device Zone,” “Source Zone,” “Destination Zone,” “Agent Zone,” and other similar names.
In one embodiment, the zone field contains a zone reference identifier that can be used to address into a table containing additional zone attributes, such as zone name, the zone's external identifier, and various other values or identifiers associated with the zone. In another embodiment, the zone field may contain any of these attributes directly. In yet another embodiment, each event can have several zone fields for each zone identified, such as “Agent Zone ID,” “Agent Zone Name,” and so on.
Such labeling is even more useful when zones are on a smaller scale than entire facility networks. For example, the attacker machine 38 may be in zone “Denver: Engineering,” the target router 34 may be in zone “Denver: DMZ,” the monitor device may be in zone “Denver: IT,” and the agent 12(d) may be in yet another zone, or also in the “Denver: IT” zone. Various other entities and their zones may be included in other embodiments of security events.
One embodiment of an agent 12 configured to perform zone identification is now described with reference to
In one embodiment, these fields are populated by a zone mapper 66. The zone mapper accesses a zone table 64. In one embodiment, the zone table associates ranges of IP addresses with zones. An example zone table is shown in Table 1 below:
Table 1 above is only a simplified example. A real world zone table 64 may specify one hundred or more zones, and cover the entire range of possible IP addresses. The zone mapper 66 thus uses the zone table 64 to map certain IP addresses to zones according to the associations provided by the zone table 64. For example, if the source IP of an event is 192.168.0.55, then the zone mapper 66 would populate the “Source Zone” field with “Denver: Engineering.”
In one embodiment, the zone table 64 shown in Table 1 would be resident on agent 12(d) on the Denver LAN 30. In one embodiment, the zone table 64 of agent 12(e) describes the zones of the Austin LAN 32 instead of the Denver LAN 30 in the same IP address range.
One embodiment of zone identification is now described with reference to
In block 106, the agent determines the zone to which the destination of the security event belongs. If the destination is not on the local network monitored by the device associated with the agent, then the zone of the destination may not be accurately determined, since the destination IP address may be translated before delivery at a remote site.
In block 108, the agent determines the zone to which the monitor device that generated the security event belongs. Since the monitor device will generally not shift zones on a regular basis, the device zone may be fixed at agent configuration. In one embodiment, the device IP address is mapped to a zone for each security event.
In block 110, the agent determines the zone to which the agent itself belongs. Since the agent will generally not shift zones on a regular basis, the agent zone may be fixed at agent configuration. In one embodiment, the agent's IP address is mapped to a zone for each security event.
In block 112, the agent generates a normalized security event. In one embodiment, this includes populating the various zone fields with the appropriate zones determined in blocks 104, 106, 108, and 110. The normalized event may undergo additional processing before being sent on to a manager.
In other embodiments, zones other than the four zones discussed above can also be determined and used to further identify the security event, such as a target and attacker zones, where there are different from source and destination. In yet other embodiments, less than four zones may be used. In one embodiment, only one of the four zones discussed above is used, e.g., the monitor device zone. Other embodiments can use any two or any three of the zones discussed above.
The manager 14 can use the zone fields of the normalized security events in any number of ways. In one embodiment, it can use it to keep track of events from various network segments with overlapping address spaces. Correlation rules can also be created that respond to the observation of certain zones, such as prohibited zones. Furthermore, by distributing the zone identification to the agents, the manager 14 is spared this computational task.
Thus, a network security system has been described. In the forgoing description, various specific values and data structures were given names, such as “security event” and “zone table,” and various specific modules, such as “agents” and “agent normalize module” have been described. However, these names are merely to describe and illustrate various aspects of the present invention, and in no way limit the scope of the present invention. Furthermore, various modules, such as the manager 14, and the agents 12 in
Number | Name | Date | Kind |
---|---|---|---|
5557742 | Smaha et al. | Sep 1996 | A |
5717919 | Kodavalla et al. | Feb 1998 | A |
5850516 | Schneier | Dec 1998 | A |
5956404 | Schneier et al. | Sep 1999 | A |
5978475 | Schneier et al. | Nov 1999 | A |
6070244 | Orchier et al. | May 2000 | A |
6134664 | Walker | Oct 2000 | A |
6192034 | Hsieh et al. | Feb 2001 | B1 |
6275942 | Bernhard et al. | Aug 2001 | B1 |
6321338 | Porras et al. | Nov 2001 | B1 |
6408391 | Huff et al. | Jun 2002 | B1 |
6408404 | Ladwig | Jun 2002 | B1 |
6484203 | Porras et al. | Nov 2002 | B1 |
6542075 | Barker et al. | Apr 2003 | B2 |
6694362 | Secor et al. | Feb 2004 | B1 |
6704874 | Porras et al. | Mar 2004 | B1 |
6708212 | Porras et al. | Mar 2004 | B2 |
6711615 | Porras et al. | Mar 2004 | B2 |
6839850 | Campbell et al. | Jan 2005 | B1 |
6928556 | Black et al. | Aug 2005 | B2 |
6966015 | Steinberg et al. | Nov 2005 | B2 |
6981146 | Sheymov | Dec 2005 | B1 |
6985920 | Bhattacharya et al. | Jan 2006 | B2 |
6988208 | Hrabik et al. | Jan 2006 | B2 |
7039953 | Black et al. | May 2006 | B2 |
7043727 | Bennett et al. | May 2006 | B2 |
7089428 | Farley et al. | Aug 2006 | B2 |
7127743 | Khanolkar et al. | Oct 2006 | B1 |
7159237 | Schneier et al. | Jan 2007 | B2 |
7171689 | Beavers | Jan 2007 | B2 |
7219239 | Njemanze et al. | May 2007 | B1 |
7260844 | Tidwell et al. | Aug 2007 | B1 |
7278160 | Black et al. | Oct 2007 | B2 |
7308689 | Black et al. | Dec 2007 | B2 |
7333999 | Njemanze | Feb 2008 | B1 |
7376969 | Njemanze et al. | May 2008 | B1 |
7483972 | Bhattacharya et al. | Jan 2009 | B2 |
7644365 | Bhattacharya et al. | Jan 2010 | B2 |
20020019945 | Houston et al. | Feb 2002 | A1 |
20020053033 | Cooper et al. | May 2002 | A1 |
20020069356 | Kim | Jun 2002 | A1 |
20020099958 | Hrabik et al. | Jul 2002 | A1 |
20020104014 | Zobel et al. | Aug 2002 | A1 |
20020147803 | Dodd et al. | Oct 2002 | A1 |
20020162026 | Neuman et al. | Oct 2002 | A1 |
20020184532 | Hackenberger et al. | Dec 2002 | A1 |
20030004688 | Gupta et al. | Jan 2003 | A1 |
20030084349 | Friedrichs et al. | May 2003 | A1 |
20030093514 | Valdes et al. | May 2003 | A1 |
20030093692 | Porras | May 2003 | A1 |
20030097590 | Syvanne | May 2003 | A1 |
20030101358 | Porras et al. | May 2003 | A1 |
20030188189 | Desai et al. | Oct 2003 | A1 |
20030221123 | Beavers | Nov 2003 | A1 |
20040010718 | Porras et al. | Jan 2004 | A1 |
20040024864 | Porras et al. | Feb 2004 | A1 |
20040044912 | Connary et al. | Mar 2004 | A1 |
20040088571 | Jerrim et al. | May 2004 | A1 |
20040221191 | Porras et al. | Nov 2004 | A1 |
20040255167 | Knight | Dec 2004 | A1 |
20040260763 | Bhattacharya et al. | Dec 2004 | A1 |
20040260945 | Raikar et al. | Dec 2004 | A1 |
20050027845 | Secor et al. | Feb 2005 | A1 |
20050060562 | Bhattacharya et al. | Mar 2005 | A1 |
20050204404 | Hrabik et al. | Sep 2005 | A1 |
20050251860 | Saurabh et al. | Nov 2005 | A1 |
20060069956 | Steinberg et al. | Mar 2006 | A1 |
20060095587 | Bhattacharya et al. | May 2006 | A1 |
20060212932 | Patrick et al. | Sep 2006 | A1 |
20070118905 | Morin et al. | May 2007 | A1 |
20070136437 | Shankar et al. | Jun 2007 | A1 |
20070150579 | Morin et al. | Jun 2007 | A1 |
20070162973 | Schneier et al. | Jul 2007 | A1 |
20070169038 | Shankar et al. | Jul 2007 | A1 |
20070234426 | Khanolkar et al. | Oct 2007 | A1 |
20070260931 | Aguilar-Macias et al. | Nov 2007 | A1 |
20080104046 | Singla et al. | May 2008 | A1 |
20080104276 | Lahoti et al. | May 2008 | A1 |
20080162592 | Huang et al. | Jul 2008 | A1 |
20080165000 | Morin et al. | Jul 2008 | A1 |
20100058165 | Bhattacharya et al. | Mar 2010 | A1 |
Number | Date | Country |
---|---|---|
WO 0245315 | Jun 2002 | WO |
WO 02060117 | Aug 2002 | WO |
WO 02078262 | Oct 2002 | WO |
WO 02101988 | Dec 2002 | WO |
WO 03009531 | Jan 2003 | WO |
WO 2004019186 | Mar 2004 | WO |
WO 2005001655 | Jan 2005 | WO |
WO 2005026900 | Mar 2005 | WO |
Entry |
---|
Labib, Khaled, and Rao Vemuri. “NSOM: A real-time network-based intrusion detection system using self-organizing maps.” Networks and Security (2002): 1-6. |
Heberlein, L. T., et al., “A Method to Detect Intrusive Activity in a Networked Environment,” Proceedings of the Fourteenth National Computer Security Conference, NIST/NCSC, Oct. 1-4, 1991, Washington, D.C., pp. 362-371. |
Javitz, H. S., et al., “The NIDES Statistical Component Description and Justification,” SRI Project 3131, Contract N00039-92-C-0015, Annual Report, A010, Mar. 7, 1994. |
Jou, Y. F., et al., “Architecture Design of a Scalable Intrusion Detection System for the Emerging Network Infrastructure,” MCNC, Technical Report CDRL A005, Apr. 1997. |
Porras, P. A., et al., “Live Traffic Analysis of TCP/IP Gateways,” Symposium on Networks and Distributed Systems Security, Internet Society, Mar. 1998. |
Robinson, S. L., “Memorandum Opinion” in SRI International, Inc. v. Internet Security Systems, Inc. and Symantec Corporation (D. Del., Civ. No. 04-1199-SLR), Oct. 17, 2006. |
Valdes, A., et al., “Statistical Methods for Computer Usage Anomaly Detection Using NIDES (Next-Generation Intrusion Detection Expert System),” Proceedings of the Third International Workship on Rough Sets and Soft Computing (RSSC 94), Jan. 27, 1995, San Jose, CA, pp. 306-311. |
ArcSight, “About ArcSight Team,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/about—team.htm>. |
ArcSight, “About Overview,” Oct. 14, 2002, [online] [Retrieved on Apr. 21, 2006] Retrieved from the Internet <URL: http://web.archive.org/web/20021014041614/http://www.arcsight.com/about.htm>. |
ArcSight, “Contact Info,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/contact.htm>. |
ArcSight, “Enterprise Coverage: Technology Architecture,” date unknown, [online] Retrieved from the Internet <URL: http://www.snaiso.com/Documentation/Arcsight/arcsight—archdta.pdf>. |
ArcSight, “Managed Process: ArcSight Reporting System,” date unknown, [online] Retrieved from the Internet <URL: http://www.snaiso.com/Documentation/Arcsight/arcsight—reportsys.pdf>. |
ArcSight, “Managed Process: Console-Based Management,” date unknown, [online] Retrieved from the Internet <URL: http://www.snaiso.com/Documentation/Arcsight/arcsight—console.pdf >. |
ArcSight, “Precision Intelligence: SmartRules™ and Cross-Correlation,” date unknown, [online] Retrieved from the Internet <URL: http://www.snaiso.com/Documentation/Arcsight/arcsight—correlation.pdf>. |
ArcSight, “Precision Intelligence: SmartAgent™,” date unknown, [online] Retrieved from the Internet <URL: http://www.ossmanagement.com/SmartAgent.pdf>. |
ArcSight, “Product Info: Product Overview and Architecture,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/product.htm>. |
ArcSight, “Product Info: 360° Intelligence Yields Precision Risk Management,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/product—info01.htm>. |
ArcSight, “Product Info: ArcSight SmartAgents,” Oct. 10, 2002, [online] [Retrieved on Apr. 21, 2006] Retrieved from the Internet <URL:http://web.archive.org/web/20021010135236/http://www.arcsight.com/product—info02.htm>. |
ArcSight, “Product Info: ArcSight Cross-Device Correlation,” date unknown, [online] [Retrieved on Oct. 25, 2005] Retrieved from the Internet <URL: http://www.arcsight.com/product—info03.htm>. |
ArcSight, “Product Info: ArcSight Manager,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/product—info04.htm>. |
ArcSight, “Product Info: ArcSight Console,” date unknown, [online] [Retrieved on Nov. 15, 2002] Retrieved from the Internet <URL: http:www.arcsight.com/product—info05.htm>. |
ArcSight, “Product Info: ArcSight Reporting System,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http:www.arcsight.com/product—info06.htm>. |
ArcSight, “Product Info: Enterprise Scaling,” date unknown, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http://www.arcsight.com/product—info07.htm>. |
ArcSight, “Security Management for the Enterprise,” 2002, [online] [Retrieved on Oct. 25, 2002] Retrieved from the Internet <URL: http:www.arcsight.com/>. |
ArcSight, “Technical Brief: How Correlation Eliminates False Positives,” date unknown, source unknown. |
Burleson, D., “Taking Advantage of Object Partitioning in Oracle8i,” Nov. 8, 2000; [online] [Retrieved on Apr. 20, 2004] Retrieved from the Internet <URL: http://www.dba-oracle.com/art—partit.htm>. |
Derodeff, C. “Got Correlation? Not Without Normalization,” 2002, [online] Retrieved from the Internet <URL: http://www.svic.com/papers/pdf/Got-Correlation—rmalization.pdf>. |
Cheung, S. et al., “EMERALD Intrusion Incident Report: 601 Message Specification,” Aug. 10, 2000, System Design Laboratory, SRI International. |
National Institute of Standards and Technology (NIST), “Federal Information Processing Standards Publication (FIPS PUB) 199: Standards for Security Categorization of Federal Information and Information Systems”, Feb. 2004. |
Haley Enterprise, “Production Systems,” 2002, [online] [Retrieved on Oct. 29, 2002] Retrieved from the Internet <URL: http://www.haley.com/0072567836705810/ProductionSystems.html>. |
Haley Enterprise, “The Rete Algorithm,” 2002, [online] [Retrieved on Oct. 29, 2002] Retrieved from the Internet <URL: http://www.haley.com/0072567836705810/ReteAlgorithm.html>. |
Haley Enterprise, “A Rules Engine for Java Based on the Rete Algorithm,” 2002, [online] [Retrieved on Oct. 29, 2002] Retrieved from the Internet <URL: http://www.haley.com/0072567836705810/ReteAlgorithmForRules.html>. |
Halme, L.R. et al., “AINT Misbehaving: A Taxonomy of Anti-Intrusion Techniques,” 2000, [online] [Retrieved on Nov. 1, 2002] Retrieved from the Internet <URL: http://www.sans.org/newlook/resources/IDFAQ/aint.htm>. |
Lindqvist, U. et al., “Detecting Computer and Network Misuse Through the Production-Based Expert System Toolset (P-BEST),” Proceedings of the IEEE Symposium on Security and Privacy, Oakland, California, May 9-12, 1999. |
CERT Coordination Center, “Overview of Attack Trends,” 2002, [online] Retrieved from the Internet <URL: http://www.cert.org/archive/pdf/attack—trends.pdf>. |
Porras, P.A. et al., “EMERALD: Event Monitoring Enabling Responses to Anomalous Live Disturbances,” Oct. 1997, Proceedings of the 20th NIST-NCSC National Information Systems Security (NISS) Conference. |
Porras, P.A. et al., “A Mission-Impact-Based Approach to INFOSEC Alarm Correlation,” Oct. 2002, Lecture Notes in Computer Science, Proceedings: Recent Advances in Intrusion Detection, pp. 95-114, Zurich, Switzerland. |
Ingargiola, G., “The Rete Algorithm,” date unknown, [online] [Retrieved on Oct. 29, 2002] Retrieved from the Internet <URL: http://yoda.cis.temple.edu:8080/UGAIWWW/lectures/rete.html>. |
Bruneau, G., “What Difficulties are Associated on Matching Events with Attacks. Why is Event/Data Correlation Important?,” 2001, [online] [Retrieved on Nov. 1, 2002] Retrieved from the Internet <URL: http://www.sans.org/newlook/resources/IDFAQ/matching.htm>. |
National Institutes of Health (NIH), “Table 1: Security Categorization of Federal Information and Information Systems,” revised Jul. 8, 2005, [online] [retrieved on Apr. 6, 2006] Retrieved from the Internet <URL: http://irm.cit.nih.gov/security/table1.htm>. |
Wood, M., et al., “Internet-Draft: Intrusion Detection Message Exchange Requirements,” Jun. 23, 2002, [online] [Retrieved on Nov. 1, 2002] Retrieved from the Internet <URL: http://www.silicondefense.com/idwg/draft-ietf-idwg-requirements-07.txt>. |
U.S. Appl. No. 60/405,921, filed Aug. 26, 2002, Gisby et al. |
U.S. Appl. No. 10/308,767, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,548, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,941, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,416, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,418, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,417, filed Dec. 2, 2002. |
U.S. Appl. No. 10/308,584, filed Dec. 2, 2002. |
U.S. Appl. No. 10/733,073, filed Dec. 10, 2003. |
U.S. Appl. No. 10/713,471, filed Nov. 14, 2003. |
U.S. Appl. No. 10/683,221, filed Oct. 10, 2003. |
U.S. Appl. No. 10/683,191, filed Oct. 10, 2003. |
U.S. Appl. No. 10/821,459, filed Apr. 9, 2004. |
U.S. Appl. No. 10/839,563, filed May 4, 2004. |
U.S. Appl. No. 10/976,075, filed Oct. 27, 2004. |
U.S. Appl. No. 10/975,962, filed Oct. 27, 2004. |
U.S. Appl. No. 11/029,920, filed Jan. 4, 2005. |
U.S. Appl. No. 11/021,601, filed Dec. 23, 2004. |
U.S. Appl. No. 11/070,024, filed Mar. 1, 2005. |
U.S. Appl. No. 11/740,203, filed Apr. 25, 2007. |
U.S. Appl. No. 11/836,251, filed Aug. 9, 2007. |
U.S. Appl. No. 12/098,322, filed Apr. 4, 2008. |
U.S. Appl. No. 11/023,942, filed Dec. 24, 2004, pp. 1-26. |