a depicts a graphical illustration of the H.264 Intra—4×4_Horizontal_Up prediction mode.
b depicts a graphical illustration of the H.264 Intra—4×4_Horizontal prediction mode.
c depicts a graphical illustration of the H.264 Intra—4×4_Horizontal_Down prediction mode.
d depicts a graphical illustration of the H.264 Intra—4×4_Diagonal_Down_Right prediction mode.
e depicts a graphical illustration of the H.264 Intra—4×4_Vertical_Right prediction mode.
f depicts a graphical illustration of the H.264 Intra—4×4_Vertical prediction mode.
g depicts a graphical illustration of the H.264 Intra—4×4_Vertical_Left prediction mode.
h depicts a graphical illustration of the H.264 Intra—4×4 Diagonal_Down_Left prediction mode.
Video decoding system 900 comprises: processor 901, primary memory 911, secondary memory 912, tertiary memory 913, and memory management unit 902, interconnected as shown.
Processor 901 is a general-purpose processor that can read and write to primary memory 911 and that perform the functionality described herein.
Primary memory 911 is the fastest addressable memory and processor 901 can access data within primary memory 911 in one clock cycle. Primary memory 911 is not a content-addressable memory with a hardwired cache-retention discipline—or a cache that is invisible to the system programmer. In accordance with the illustrative embodiment, processor 901 and primary 902 are on the same monolithic die. In accordance with the illustrative embodiment, primary memory 911 has its own address space, which is distinct from the address space of secondary memory 912. It will be clear to those skilled in the art, however, how to make and use alternative embodiments of the present invention in which primary memory 911 and secondary memory 912 are in the same address space.
Secondary memory 912 is the second-fastest addressable memory in the system and processor 901 can access data within secondary memory 912 in approximately 100 clock cycles. In accordance with the illustrative embodiment, secondary memory 912 is semiconductor memory but it is not on the same die as processor 901. This accounts for the substantial difference in speed between it and secondary memory 912.
Tertiary memory 913 is the slowest addressable memory and processor 901 can access data within tertiary memory 913 in approximately 10,000 clock cycles. In accordance with the illustrative embodiment, tertiary memory 913 is a mass storage device, such as a hard drive, and this accounts for the substantial difference in speed between it and secondary memory 912.
Primary memory 911 costs substantially more, per bit, than does secondary memory 912, and secondary memory 912 costs substantially more, per bit, than does tertiary memory 913. For this reason, primary memory 911 comprises substantially fewer bytes than secondary memory 912, and secondary memory 912 comprises substantially fewer bytes than tertiary memory 913.
When processor 901 seeks a word of data and the word is in primary memory 911, processor 901 can continue processing very quickly. In contrast, when processor 901 seeks a word of data and the word in not in primary memory 911, processor 901 waits until the word can be retrieved. Given that secondary memory 912 is 1/100th of the speed of primary memory 911, processing can become very slow if processor 901 must regularly wait for data to be retrieved from secondary memory 912.
One solution is to ensure that the size of primary memory 911 is large because this reduces, probabilistically, the frequency with which a desired word is in primary memory 911. This approach is problematic, however, because it is expensive and because some applications, such as video decoding, use such large quantities of data that any reasonably-sized primary memory would be ineffective.
To overcome this problem, the illustrative embodiment employs memory management unit 902 which controls what data is in primary memory 911 and what is not. In other words, memory management unit 902 retains in primary memory 911 data that will be needed by processor 901 soon and expunges from primary memory 911 data that will not be needed by processor 901 again soon. By retaining in primary memory 911 data that will be needed soon, the illustrative embodiment reduces the frequency and likelihood that processor 901 must wait until data can be retrieved from secondary memory 912, and by expunging from primary memory 911 data that will not be needed by processor 901 again soon, the illustrative embodiment frees up space in primary memory 911 for data that will be needed by processor 901 soon.
In many cases, a memory management unit cannot predict what data the processor will need again soon and what data it will not, but there are applications, such as video decoding, when reasonable predictions can be made.
As can be seen in
It is to be understood that the above-described embodiments are merely illustrative of the present invention and that many variations of the above-described embodiments can be devised by those skilled in the art without departing from the scope of the invention. It is therefore intended that such variations be included within the scope of the following claims and their equivalents.