The present disclosure relates generally to information handling systems (IHSs), and more particularly to IHS snoop filter optimization.
As the value and use of information continues to increase, individuals and businesses seek additional ways to process and store information. One option is an information handling system (IHS). An IHS generally processes, compiles, stores, and/or communicates information or data for business, personal, or other purposes. Because technology and information handling needs and requirements may vary between different applications, IHSs may also vary regarding what information is handled, how the information is handled, how much information is processed, stored, or communicated, and how quickly and efficiently the information may be processed, stored, or communicated. The variations in IHSs allow for IHSs to be general or configured for a specific user or specific use such as financial transaction processing, airline reservations, enterprise data storage, or global communications. In addition, IHSs may include a variety of hardware and software components that may be configured to process, store, and communicate information and may include one or more computer systems, data storage systems, and networking systems.
IHS server and workstation chipsets use snoop-filter caches (SF Caches) to reduce the percentage of cache line snoops on a remote bus, to improve performance. The snoop filter cache stores a directory of all processor cache lines to minimize snoop traffic on the dual front-side buses during a cache miss.
In theory, a snoop filter ensures that snoop requests for cache lines go to the appropriate processor bus (e.g., on a system with multiple front side busses (FSBs)) and not all of the available busses, thereby improving performance. Therefore, applications will benefit from a reduced snoop activity that the snoop filter cache provides.
Experiments have shown that a snoop filter does not improve performance for all applications, and moreover its performance impact is sensitive to the system configuration. In many cases, the snoop filter can cause performance degradation for certain workloads.
Accordingly, it would be desirable to provide a static and dynamic optimization of a snoop filter to optimize performance of systems with a snoop filter cache, absent the deficiencies described above.
According to one embodiment, a snoop filter optimization system includes one or more subsystems to operate a snoop filter, determine information that that affects operation of the snoop filter, and adjust operation of the snoop filter relative to the information that affects operation of the snoop filter.
For purposes of this disclosure, an IHS 100 includes any instrumentality or aggregate of instrumentalities operable to compute, classify, process, transmit, receive, retrieve, originate, switch, store, display, manifest, detect, record, reproduce, handle, or utilize any form of information, intelligence, or data for business, scientific, control, or other purposes. For example, an IHS 100 may be a personal computer, a network storage device, or any other suitable device and may vary in size, shape, performance, functionality, and price. The IHS 100 may include random access memory (RAM), one or more processing resources such as a central processing unit (CPU) or hardware or software control logic, read only memory (ROM), and/or other types of nonvolatile memory. Additional components of the IHS 100 may include one or more disk drives, and one or more network ports for communicating with external devices as well as various input and output (I/O) devices, such as a keyboard, a mouse, and a video display. The IHS 100 may also include one or more buses operable to transmit communications between the various hardware components.
Other resources can also be coupled to the system through the memory I/O hub 104 using a data bus, including an optical drive 114 or other removable-media drive, one or more hard disk drives 116, one or more network interfaces 118, one or more Universal Serial Bus (USB) ports 120, and a super I/O controller 122 to provide access to user input devices 124, etc. The IHS 100 may also include a solid state drive (SSDs) 126 in place of, or in addition to main memory 108, the optical drive 114, and/or a hard disk drive 116. It is understood that any or all of the drive devices 114, 116, and 126 may be located locally with the IHS 100, located remotely from the IHS 100, and/or they may be virtual with respect to the IHS 100.
Not all IHSs 100 include each of the components shown in
The snoop filter system 130 includes a front side bus 106 to communicatively couple the processor 102 to the memory I/O hub chipset 104. In an embodiment, communication information/data passes through processor cache memory 134 to the memory I/O hub 140 via the front side bus or system bus 106. In an embodiment, a front side bus 106 is the primary pathway between a processor 102 and memory 108. Speed of a front side bus 106 is generally derived from the number of parallel channels (e.g., 16 bit, 32 bit, and etc.) and clock speed and is generally faster than a peripheral bus such as, PCI, ISA, and etc. As the information/data passes through the memory I/O hub 104 a snoop filter 140 determines and stores the status of the processor cache 134 lines, filters unnecessary snoops on the processor 102 and via the I/O controller 122 to any remote bus, and lowers front side bus 106 utilization. When cache memory 134 has been changed, the snoop filter 140 allows other processors 102 to check which cache memory 134 has been changed.
In an embodiment, The snoop filter system 130 also includes a plurality of memory files 134 (e.g., fully buffered dynamic random access memory (FBD)), as all or part of the main memory 108. One or more memory busses 136 couple the FBD 134 with the memory I/O hub 104 to allow communication between the FBD 134 and the memory I/O hub 104.
Because the impact of a snoop filter 140 is sensitive to many factors such as, workloads, memory configurations, processor 102 architecture, and a variety of other factors, the present disclosure contemplates that the snoop filter 140 operation is enabled for those scenarios in which it will be beneficial to the IHS 100 performance. Otherwise, the snoop filter 140 may be disabled for IHS 100 configurations and/or applications that may not benefit from the snoop filter 140.
In an embodiment, the method 170 analyzes the IHS 100 system configuration during a POST and makes a decision to enable/disable the snoop filter 140 based on a table lookup. The table may be populated with any configuration information that impacts the snoop filter 140 performance (e.g., see
In an embodiment, a decision at POST may be made based on the snoop filter 140 configuration (e.g., coverage and policy) and its relationship with the processor 102's and memory configuration in the IHS 100. This helps the IHS 100 get the maximum performance from their IHS 100. For example, if the snoop filter 140 size is less than the sum of processor 102 caches 134, then the snoop filter 140 cannot provide 1× coverage. In such instances the snoop filter 140 should be turned off or otherwise disabled to reduce performance degradation due to back-invalidate operations that cause cache misses to increase. Similarly, the table lookup in the BIOS should be populated by such data when running standard benchmarks for different processor 102 and memory configurations to determine if the snoop filter 140 should be enabled or disabled if sufficient coverage is not provided.
In an embodiment, another variable that determines the impact of the snoop filter 140 is the application or workload characteristics, as shown in
In an embodiment, if a memory I/O hub chipset 104 supports the option to toggle the snoop filter 140 operation without requiring a system reboot (e.g., Hyper Threading), then an adaptive process may be used to optimize performance based on workload characteristics. In this adaptive process, the snoop filter 140 may be either used or disabled based on both system configuration and workload characteristics. Depending on the snoop activity that is measured over time, the snoop filter 140 may be enabled or disabled without rebooting the IHS 100 to ensure optimal system performance. Thus, it should be apparent to one having ordinary skill in the art that many combinations of methods 170 and 188 may be used within the scope of the present disclosure.
Although illustrative embodiments have been shown and described, a wide range of modification, change and substitution is contemplated in the foregoing disclosure and in some instances, some features of the embodiments may be employed without a corresponding use of other features. Accordingly, it is appropriate that the appended claims be construed broadly and in a manner consistent with the scope of the embodiments disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
5426741 | Butts et al. | Jun 1995 | A |
5966729 | Phelps | Oct 1999 | A |
6304945 | Koenen | Oct 2001 | B1 |
6401156 | Mergard et al. | Jun 2002 | B1 |
6519685 | Chang | Feb 2003 | B1 |
6598123 | Anderson et al. | Jul 2003 | B1 |
6615319 | Khare et al. | Sep 2003 | B2 |
6694409 | Chang | Feb 2004 | B2 |
6704845 | Anderson et al. | Mar 2004 | B2 |
6725337 | Tan et al. | Apr 2004 | B1 |
6810467 | Khare et al. | Oct 2004 | B1 |
6829665 | Jones et al. | Dec 2004 | B2 |
6842827 | Blankenship et al. | Jan 2005 | B2 |
6842830 | Khare et al. | Jan 2005 | B2 |
6857048 | Rankin et al. | Feb 2005 | B2 |
6859864 | Khare et al. | Feb 2005 | B2 |
6959364 | Safranek et al. | Oct 2005 | B2 |
6961800 | Lester et al. | Nov 2005 | B2 |
7076539 | Reese et al. | Jul 2006 | B2 |
7093079 | Quach et al. | Aug 2006 | B2 |
7117311 | Rankin et al. | Oct 2006 | B1 |
7117312 | Cypher | Oct 2006 | B1 |
7173450 | Atkinson | Feb 2007 | B2 |
7386683 | Blumrich et al. | Jun 2008 | B2 |
7392351 | Blumrich et al. | Jun 2008 | B2 |
7581068 | Gilbert et al. | Aug 2009 | B2 |
20070079074 | Collier et al. | Apr 2007 | A1 |
Number | Date | Country | |
---|---|---|---|
20090158022 A1 | Jun 2009 | US |