Not applicable.
Not applicable.
Not applicable.
1. Technical Field
The present invention relates generally to electrical computers and digital processing systems having processing architectures and performing instruction processing, and more particularly to such for processing instruction data that specifically supports or performs a data transfer operation.
2. Background Art
In the art of computing, processing speed is a much desired quality, and the quest to create faster computers and processors is ongoing. However, it is generally acknowledged in the industry that the limits for increasing the speed in microprocessors are rapidly being approached, at least using presently known technology. Therefore, there is an increasing interest in the use of multiple processors to increase overall computer speed by sharing computer tasks among the processors. But it is also generally acknowledged that there will, almost inevitably, be some decrease in overall efficiency involved in the sharing of the workload. That is, the old adage will apply that just because one person can dig a post hole in 60 minutes, it does necessarily follow that 60 people could dig a post hole in 1 minute. The same principle applies to almost any division of tasks, and the division of tasks among processors is no exception.
Of course, efforts are being made to make the sharing of tasks among computer processors more efficient. The question of exactly how the tasks are to be allocated is being examined and processes improved. In the course of work in this area it has been the present inventors' observation that it may be very cumbersome under some circumstances to transfer data from one CPU to another in a multi-CPU environment. For example, if data must be transferred from one CPU to another, and the target CPU is separated from the source CPU by one CPU between them, the source CPU must write the data to the CPU directly in line, which must then in turn read the data and then write it to the target CPU, which must then read the data. Such a process requires many read and write operations, and if a large quantity of data is being transferred, so many read and write commands may clog system operations.
To satisfy the need to allow multiple read and write operations in various different directions—that is, between any of various other CPUs in the same system—all at the same time, systems and methods for multi-port read and write operations have been developed. These address most of the concerns discussed above but, as with any major advancement, these systems and methods have raised new challenges. For example, in multi-CPU environments were the CPUs are arraigned in a pipeline or a multidimensional array, inversing can occur where a CPU writes to a prior rather than a subsequent CPU. Mechanisms can be crafted to prevent this, but these entail hardware modifications or substantial programming and inter-CPU communications. As another example, many applications today require real time processing or it is simply desirable to increase processing speed and efficiency. It follows that optimization of multi-port read and write operations would be beneficial. In a similar vein, now that multi-port operations are available, it would also be beneficial to make the set-up and the performance of these operations more flexible.
Accordingly, it is an object of the present invention to provide improved systems and methods to process data in pipelines and arrays of computers.
Briefly, one preferred embodiment of the present invention is a method for a series of computers to process data. The series of computers includes a first and a last computer, and wherein each of the computers except the first is preceded by a prior computer and each except the last is followed by a subsequent computer. The process can be viewed as each of the computers being considered as a current computer. New data is read with the current computer. Then old data is written with the current computer. And then the new data is processed in the current computer to produce the next old data. After this, if the current computer is not the last computer, the old data is held in the current computer.
Briefly, another preferred embodiment of the present invention is a series of computers to process data. The series includes a first and a last computer, wherein each of the computers except the first is preceded by a prior computer and each except the last is followed by a subsequent computer. The computers each have a logic to read new data via a first data path, a logic to write old data via a second data path, and a logic to process the new data to produce the next old data. Except for the last computer, a storage element stores the old data. The logic to write operates after the logic to read and the logic to write operates before the logic to process.
An advantage of the present invention is that it avoids inversing, wherein data is written from a higher order to a lower order computer occurs.
Another advantage of the invention is that it improves the initial delivery of data through a pipeline or array of the computers so that respective processing can begin sooner.
Another advantage of the invention is that it is particularly suitable for use where a same initial data value needs to be provided to all of a series of computers.
And another advantage of the invention is that it is particularly suitable for use with pipelines or arrays of computers capable of asynchronous multi-port read and multi-port communications.
These and other objects and advantages of the present invention will become clear to those skilled in the art in view of the description of the best presently known mode of carrying out the invention and the industrial applicability of the preferred embodiment as described herein and as illustrated in the figures of the drawings.
The purposes and advantages of the present invention will be apparent from the following detailed description in conjunction with the appended figures of drawings in which:
a-f are table diagrams showing an overview of port address decoding that is usable in the computers in the section in
a-b are schematic diagrams stylistically showing the initial flow of data in the pipeline of
In the various figures of the drawings, like references are used to denote like or similar elements or steps.
While this invention is described in terms of modes for achieving this invention's objectives, it will be appreciated by those skilled in the art that variations may be accomplished in view of these teachings without deviating from the spirit or scope of the present invention.
The embodiments and variations of the invention described herein, and/or shown in the drawings, are presented by way of example only and are not limiting as to the scope of the invention. Unless otherwise specifically stated, individual aspects and components of the invention may be omitted or modified, or may have substituted therefore known equivalents, or as yet unknown substitutes such as may be developed in the future or such as may be found to be acceptable substitutes in the future. The invention may also be modified for a variety of applications while remaining within the spirit and scope of the claimed invention, since the range of potential applications is great, and since it is intended that the present invention be adaptable to many such variations.
Preferred embodiments of the present invention are improved systems and methods to process data in pipelines and arrays of computers. As illustrated in the various drawings herein, and particularly in the view of
As context and a foundation to the present invention, a detailed background example of asynchronous computer communication is first presented and then a detailed background example of multi-port read and multi-port write operations in such an asynchronous computer communication is further presented.
For the first background example, a computer array is depicted in a diagrammatic view in
One skilled in the art will recognize that there will be additional components on the die 14 that are omitted from the view of
Computer 12e is an example of one of the computers 12 that is not on the periphery of the array 10. That is, computer 12e has four orthogonally adjacent computers 12a, 12b, 12c and 12d. This grouping of computers 12a through 12e will be used hereinafter in relation to a more detailed discussion of the communications between the computers 12 of the array 10. As can be seen in the view of
A computer 12, such as the computer 12e can set one, two, three or all four of its read lines 18 such that it is prepared to receive data from the respective one, two, three or all four adjacent computers 12. Similarly, it is also possible for a computer 12 to set one, two, three or all four of its write lines 20 high. (Both cases are discussed in more detail hereinafter.)
When one of the adjacent computers 12a, 12b, 12c or 12d sets a write line 20 between itself and the computer 12e high, if the computer 12e has already set the corresponding read line 18 high, then a word is transferred from that computer 12a, 12b, 12c or 12d to the computer 12e on the associated data lines 22. Then the sending computer 12 will release the write line 20 and the receiving computer 12e (in this example) pulls both the write line 20 and the read line 18 low. The latter action will acknowledge to the sending computer 12 that the data has been received. Note that the above description is not intended necessarily to denote the sequence of events in order. In actual practice, the receiving computer may try to set the write line 20 low slightly before the sending computer 12 releases (stops pulling high) its write line 20. In such an instance, as soon as the sending computer 12 releases its write line 20 the write line 20 will be pulled low by the receiving computer 12e.
In the present example, only a programming error would cause both computers 12 on the opposite ends of one of the buses 16 to try to set either both of the read lines 18 there-between high or to set both of the write lines 20 there-between high at the same time. However, it is presently anticipated that there will be occasions wherein it is desirable to set different combinations of the read lines 18 high such that one of the computers 12 can be in a wait state awaiting data from the first one of the chosen computers 12 to set its corresponding write line 20 high.
In the example discussed above, computer 12e was described as setting one or more of its read lines 18 high before an adjacent computer (selected from one or more of the computers 12a, 12b, 12c or 12d) has set its write line 20 high. However, this process can certainly occur in the opposite order. For example, if the computer 12e were attempting to write to the computer 12a, then computer 12e would set the write line 20 between computer 12e and computer 12a to high. If the read line 18 between computer 12e and computer 12a has then not already been set to high by computer 12a, then computer 12e will simply wait until computer 12a does set that read line 18 high. Then, as discussed above, when both of a corresponding pair of read line 18 and write line 20 are high the data awaiting to be transferred on the data lines 22 is transferred. Thereafter, the receiving computer 12a (in this example) sets both the read line 18 and the write line 20 between the two computers 12e and 12a (in this example) to low as soon as the sending computer 12e releases it.
Whenever a computer 12 such as the computer 12e has set one of its write lines 20 high in anticipation of writing it will simply wait, using essentially no power, until the data is “requested,” as described above, from the appropriate adjacent computer 12, unless the computer 12 to which the data is to be sent has already set its read line 18 high, in which case the data is transmitted immediately. Similarly, whenever a computer 12 has set one or more of its read lines 18 to high in anticipation of reading it will simply wait, using essentially no power, until the write line 20 connected to a selected computer 12 goes high to transfer an instruction word between the two computers 12.
There may be several potential means and/or methods to cause the computers 12 to function as described above. However, in this present example, the computers 12 so behave simply because they are operating generally asynchronously internally (in addition to transferring data there-between in the asynchronous manner described). That is, instructions are completed sequentially. When either a write or read instruction occurs, there can be no further action until that instruction is completed (or, perhaps alternatively, until it is aborted, as by a “reset” or the like). There is no regular clock pulse, in the prior art sense. Rather, a pulse is generated to accomplish a next instruction only when the instruction being executed either is not a read or write type instruction (given that a read or write type instruction would require completion by another entity) or else when the read or write type operation is, in fact, completed.
Other basic components of the computer 12 are a return stack 28, an instruction area 30, an arithmetic logic unit (ALU 32), a data stack 34, and a decode logic section 36 for decoding instructions. One skilled in the art will be generally familiar with the operation of stack based computers such as the computers 12 of this present example. The computers 12 are dual stack computers having the data stack 34 and separate return stack 28.
In this embodiment, the computer 12 has four communication ports 38 for communicating with adjacent computers 12. The communication ports 38 are tri-state drivers, having an off status, a receive status (for driving signals into the computer 12) and a send status (for driving signals out of the computer 12). Of course, if the particular computer 12 is not on the interior of the array (
Although the technology is not limited by this example, the present computer 12 is implemented to execute native Forth language instructions. As one familiar with the Forth computer language will appreciate, complicated Forth instructions, known as Forth “words” are constructed from the native processor instructions designed into the computer. The collection of Forth words is known as a “dictionary.” In other languages, this might be known as a “library.” As will be described in greater detail hereinafter, the computer 12 reads eighteen bits at a time from RAM 24, ROM 26, or directly from one of the data buses 16 (
When the slot sequencer 42 is triggered, either by the first OR gate input 62 going high or by the second OR gate input 64 going high (as will be discussed hereinafter), then a signal will travel around the slot sequencer 42 twice, producing an output at a slot sequencer output 68 each time. The first time the signal passes the slot sequencer output 68 it will be low, and the second time the output at the slot sequencer output 68 will be high. The relatively wide output from the slot sequencer output 68 is provided to a pulse generator 70 (shown in block diagrammatic form) that produces a narrow timing pulse as an output. One skilled in the art will recognize that the narrow timing pulse is desirable to accurately initiate the operations of the computer 12.
When the particular instruction 52 being executed is a read or a write instruction, or any other instruction wherein it is not desired that the instruction 52 being executed triggers immediate execution of the next instruction 52 in sequence, then the i4 bit 66 is ‘0’ (low) and the first OR gate input 62 is, therefore, also low. One skilled in the art will recognize that the timing of events in a device such as the computers 12 is generally quite critical, and this is no exception. Upon examination of the slot sequencer 42 one skilled in the art will recognize that the output from the OR gate 60 must remain high until after the signal has circulated past the NAND gate 58 in order to initiate the second “lap” of the ring. Thereafter, the output from the OR gate 60 will go low during that second “lap” in order to prevent unwanted continued oscillation of the circuit.
As can be appreciated in light of the above discussion, when the i4 bit 66 is ‘0,’ then the slot sequencer 42 will not be triggered—assuming that the second OR gate input 64, which will be discussed hereinafter, is not high.
As discussed, above, the i4 bit 66 of each instruction 52 is set according to whether or not that instruction is a read or write type of instruction. The remaining bits 50 in the instruction 52 provide the remainder of the particular opcode for that instruction. In the case of a read or write type instruction, one or more of the bits may be used to indicate where data is to be read from or written to in that particular computer 12. In the present example, data to be written always comes from the T register 44 (the top of the data stack 34), however data can be selectively read into either the T register 44 or else the instruction area 30 from where it can be executed. That is because, in this particular embodiment, either data or instructions can be communicated in the manner described herein and instructions can, therefore, be executed directly from the data bus 16, although this is not necessary. Furthermore, one or more of the bits 50 will be used to indicate which of the ports 38, if any, is to be set to read or write. This later operation is optionally accomplished by using one or more bits to designate a register 40, such as the A register 40a, the B register 40b, or the like. In such an example, the designated register 40 will be preloaded with data having a bit corresponding to each of the ports 38 (and, also, any other potential entity with which the computer 12 may be attempting to communicate, such as memory, an external communications port, or the like.) For example, each of four bits in the particular register 40 can correspond to each of the up port 38a, the right port 38b, the left port 38c, or the down port 38d. In such case, where there is a ‘1’ at any of those bit locations, communication will be set to proceed through the corresponding port 38.
The immediately following example will assume a communication wherein computer 12e is attempting to write to computer 12c, although the example is applicable to communication between any adjacent computers 12. When a write instruction is executed in a writing computer 12e, the selected write line 20 is set high (in this example, the write line 20 between computers 12e and 12c). If the corresponding read line 18 is already high, then data is immediately sent from the selected location through the selected communications port 38. Alternatively, if the corresponding read line 18 is not already high, then computer 12e will simply stop operation until the corresponding read line 18 does go high. The mechanism for stopping (or, more accurately, not enabling further operations of) the computer 12a when there is a read or write type instruction has been discussed previously herein. In short, the opcode of the instruction 52 will have a ‘0’ at the i4 bit 66 position, and so the first OR gate input 62 of the OR gate 60 is low, and so the slot sequencer 42 is not triggered to generate an enabling pulse.
As for how the operation of the computer 12e is resumed when a read or write type instruction is completed, the mechanism for that is as follows: When both the read line 18 and the corresponding write line 20 between computers 12e and 12c are high, then both lines 18 and 20 will be released by each of the respective computers 12 that is holding it high. (In this example, the sending computer 12e will be holding the write line 20 high while the receiving computer 12c will be holding the read line 18 high). Then the receiving computer 12c will pull both lines 18 and 20 low. In actual practice, the receiving computer 12c may attempt to pull the lines 18 and 20 low before the sending computer 12e has released the write line 20. However, since the lines 18 and 20 are pulled high and only weakly held (latched) low, any attempt to pull a line 18 or 20 low will not actually succeed until that line 18 or 20 is released by the computer 12 that is latching it high.
When both lines 18 and 20 in a data bus 16 are pulled low, this is an “acknowledge” condition. Each of the computers 12e and 12c will, upon the acknowledge condition, set its own internal acknowledge line 72 high. As can be seen in the view of
In any case when the instruction 52 being executed is in the slot three position of the instruction word 48, the computer 12 will fetch the next awaiting eighteen-bit instruction word 48 unless, of course, the i4 bit 66 is a ‘0.’ In actual practice, a method and apparatus for “prefetching” instructions can be included such that the fetch can begin before the end of the execution of all instructions 52 in the instruction word 48. However, this also is not necessary for asynchronous data communications.
The above example wherein computer 12e is writing to computer 12c has been described in detail. As can be appreciated in light of the above discussion, the operations are essentially the same whether computer 12e attempts to write to computer 12c first, or whether computer 12c first attempts to read from computer 12e. The operation cannot be completed until both computers 12e and 12c are ready and, whichever computer 12e or 12c is ready first, that first computer 12 simply “goes to sleep” until the other computer 12e or 12c completes the transfer. Another way of looking at the above described process is that, actually, both the writing computer 12e and the receiving computer 12c go to sleep when they execute the write and read instructions, respectively, but the last one to enter into the transaction reawakens nearly instantaneously when both the read line 18 and the write line 20 are high, whereas the first computer 12 to initiate the transaction can stay asleep nearly indefinitely until the second computer 12 is ready to complete the process.
It is believed that a key feature for enabling efficient asynchronous communications between devices is some sort of acknowledge signal or condition. In the prior art, most communication between devices has been clocked and there is no direct way for a sending device to know that the receiving device has properly received the data. Methods such as checksum operations may have been used to attempt to insure that data is correctly received, but the sending device has no direct indication that the operation is completed. The present method, as described herein, provides the necessary acknowledge condition that allows, or at least makes practical, asynchronous communications between the devices. Furthermore, the acknowledge condition also makes it possible for one or more of the devices to “go to sleep” until the acknowledge condition occurs. Of course, an acknowledge condition could be communicated between the computers 12 by a separate signal being sent between the computers 12 (either over the interconnecting data bus 16 or over a separate signal line). However, it can be appreciated that there is even more economy involved here, in that the method for acknowledgement does not require any additional signal, clock cycle, timing pulse, or any such resource beyond that described, to actually affect the communication.
In light of the above discussion of the procedures and means for accomplishing them, the following brief description of an example of the background method can now be understood.
For the second background example,
As can be seen in
The CPU 12e has its own memory 102 (e.g., the RAM 24 and the ROM 26 shown in
a-f are table diagrams showing an overview of port address decoding that is usable in the CPUs 12 of the section 100 in
Note, for consistency and to minimize confusion we stick to the general convention here that a high value or “1” denotes a true condition and a low value or “0” denotes a false condition. This is not a requirement, however, and alternate conventions can be used. For example, some presently preferred embodiments of the of the CPUs 12 use “0” for true in the RR bit locations and use “1” for true in the WR bit locations.
In passing, it should be noted that this port address decoding approach also permits the high address bit 108 to be set to “1” and none of the select bits 110 to be set. This can beneficially be used to address another element in the CPU 12. For example, the IOCS register 40d can be addressed in this manner.
In present embodiments of the CPUs 12, the IOCS register 40d uses the same port address arrangement to report the current status of the read lines 18 and write lines 20 of the ports 38. This makes these respective bits in the IOCS register 40d useful to permit programmatically testing the status of I/O operations. For example, rather than have CPU 12e commit to an asynchronous read from CPU 12b, wherein CPU 12e will go to sleep if CPU 12b has not yet set the shared write line 20 high, CPU 12e can test the state of bit 13 (Down/WR) in the IOCS register 40d (reflecting the state of the write line 20 that connects CPU 12b to CPU 12e) and either branch to and immediately read the ready data from CPU 12b or branch to and immediately execute another instruction.
b shows a simple first example. Here the select bit 110 for Right/RR is set, indicating that port 38b is to be read from.
Conventionally, only one select bit 110 would be enabled to specify a single port 38 and a single action (read or write) at any given time. Multiple high bits would then be decoded as an error condition. The novel approach disclosed herein, however, does not follow this convention. Rather, more than one of the select bits 110 for the ports 38 may be beneficially enabled at the same time, thus requesting, multiple read and/or write operations. In such cases, the data is presented on all of the respective ports 38, including a signal that the new data is present.
d-f show some examples of multiple read and/or write operations.
In practice during a multiple write, the CPU 12e will present the data and set the write lines 20 high on the buses 16 that it shares with one or more of the target CPUs 12a, 12b, 12c, or 12d. The source CPU 12e then will wait until it receives an indication that the data has been read. At some eventual point, presumably, one or more of the target CPUs 12a, 12b, 12c, or 12d sets its respective read line 18 high on the bus 16 shared with CPU 12e. A target CPU 12 then formally reads the data and pulls both the respective read line 18 and write line 20 low on the bus 16 shared with CPU 12e, thus acknowledging receipt of the data from CPU 12e.
In the following, @=fetch, !=store, and p refers to the “program counter” or P register 40c. The “+” in @p+ and !p+refer to incrementing a memory address in the register after execution, except that the register content is not incremented if it addresses another register or a port. Thus, the “+” in these latter cases differentiates these instructions as “special” rather than as normal @p and !p instructions.
For this particular example the P register 40c can be loaded with 101100000b and the top of the return stack 28 can contain 101b (5 decimal). Since the P register 40c contains 101100000b (see e.g.,
In summary, the P register 40c in the example here is loaded with one address value that specifies both a source and destination (ports 38b and 38a and thus CPUs 12b and 12a), the return stack 28 has been loaded with an iteration count (5). Then five data words 120 are efficiently transferred (“pipelined”) through CPU 12e, which then continues at the instruction in slot zero 54a of a sixth data word 120 also provided by CPU 12b.
Various other advantages flow from the use of this simple but elegant approach. For instance, the A register 40a and the B register 40b need not be used and thus can be employed by CPU 12e for other data purposes. Following from this, pointer swapping (trashing) can also be eliminated when performing data transfers.
For example, a conventional software routine for data pipelining would at some point read data from an input port and at another point write data to an output port. For this at least one pointer into memory would be needed, in addition to pointers to the respective input and output ports that are being used. Since the ports would have different addresses, the most direct way to proceed here would be to load the input port address onto a stack with a literal instruction, put that address into an addressing register, perform a read from the input port, then load the address of the output port onto the stack with a literal instruction, put that address into an addressing register, and perform a write to the output port.
The two literal loads in this approach would take 4 cycles each, and the two register set instructions will take 1 cycles each. That is a total of 10 cycles spent inside of the loop just on setting the input and output pointers. Furthermore, there is an additional penalty when such pointer swapping is needed because three words of memory are required inside of the loop, thus not allowing the use of a loop contained inside a single 18-bit word. Accordingly, an instruction loop in this example will require a branch with a memory access, which adds 4 cycles of further overhead and makes the total pointer swap and loop overhead at least 14 cycle.
In contrast, however, since multi-port addressing is possible in the CPU 12, the address that selects both the input port 38 and the output port 38 can be loaded outside of an I/O loop and used for both input and output. This approach works because data from only one neighbor is read during a multi-port read and only one neighbor reads during a multi-port write. Thus the 14-cycle overhead inside of a loop that would traditionally be spent setting the input and output pointers is not needed. The loop still has a read instruction and a write instruction, but these can now both use the same pointer, so it does not have to be changed.
This means that the use of the multi-port write technique can reduce the overhead of some types of I/O loops by 14 cycles (or more). It has been the inventors' observation that, in the best case, this permits a reduction from 23 cycles to 6 cycles in the processing loop of a CPU 12. In a situation where one cycle takes approximately one nanosecond, this represents an increase from 43 MHz to 167 MHz in effective processor speed, which represents a considerable improvement.
Briefly continuing now with
Summarizing, the CPUs 12 have to deal with both reading and jumping to ports 38. In reading from, or jumping to, a multi-port address, WHICH port 38 that data or instruction is gotten from is unknown without explicit code being executed to find out. (The fastest way relies on the ports 38 being the same for both CPUs 12.) Traditionally this would be seen as a problem to avoid, because different data or code could come from different ports. However, in the cooperative environment postulated, the inventors have been figuring out how to turn everything into a benefit. And this has been such a case.
If a CPU 12 executes from a multiport address, and all of the addressed neighbor CPUs 12 are writing cooperatively (i.e., synchronized), one neighbor CPU 12 can be supplying the instruction stream while different CPUs 12 provide the literal data. The literal fetch opcode (@p+) causes a read from the multi-port address in the P register 40c that selectively (not all literals need to do this) can be satisfied by different neighboring CPUs 12. This merely requires extensive “cooperation” between the neighboring CPUs 12.
In the pipeline multi-port usage, however, where one neighbor CPU 12 is reading and one CPU 12 is writing, reads and writes to the same multi-port address do not cause problems. The idea is that jumping to such a multi-port address and executing the literal store opcode (!p+) allows the P register 40c to address two ports 38 with complete safety. This frees up BOTH the A register 40a and the B register 40b for local use.
The CPUs 12 can also be subject to other optimizations when data (actual data or instructions being transferred as data) is propagated.
Rule 2 avoids the pipeline of CPUs becoming a “bottleneck.” Obviously, if the pipeline of CPUs cannot keep up with the data being supplied to it, it is not going to be able to operate in real time. It follows that each CPU should optimally be ready to read before or at the very instant that a prior CPU becomes ready to write. Of course, this is not always possible (as
Rule 3 avoids the pipeline of CPUs “braking” (the analogy to a liquid carrying pipeline becomes somewhat strained here).
a shows the data flow through the pipeline 1100 if the conventional read (R), process (P), and write (W) order of operations is employed. All of the operations have a minimum time to execute (shown here as the same for simplicity), but the read (R) and write (W) operations can require additional time beyond the minimum while waiting for a corresponding write (W) or read (R) to occur. Depending on the tasks at hand, the time for the process (P) operations will vary considerably, especially in asynchronous CPUs. Thus, in actual applications, the process (P) operations would typically take longer than depicted here and problems like those shown with
In
a also shows how the inverse 1112 adds substantially to the time that CPU 1110 spends reading (i.e., waiting) for data to start work on. For that matter, however, the timing throughout the pipeline 1100 in
b shows the data flow through the pipeline 1100 if a read (R), write (W), and process (P) order of operations is employed. As can be seen here, there is no inverse and the CPUs 1102, 1104, 1106, 1108, 1110 all receive data to start work on as soon as possible.
The junctions 1120 shown in
Alternately, each of CPUs 1102, 1104, 1106, 1108, 1110 can be provided with different first data values by using initial read (R), write (W), and a single nop instruction as the process (P) until all CPUs in the pipeline have data, with which they all then perform actual processing in parallel.
Various additional modifications may be made to the present invention without altering its value or scope. For example, while this invention has been described herein in terms of read instructions and write instructions, in actual practice there may be more than one read type instruction and/or more than one write type instruction. As just one example, in one embodiment of the computers 12 there is a write instruction that increments the register and other write instructions that do not. Similarly, write instructions can vary according to which register 40 is used to select communications ports 38, or the like, as discussed previously herein. There can also be a number of different read instructions, depending only upon which variations the designer of the computers 12 deems to be a useful choice of alternative read behaviors.
Similarly, while the present invention has been described herein in relation to communications between computers 12 in an array 10 on a single die 14, the same principles and method can be used, or modified for use, to accomplish other inter-device communications, such as communications between a computer 12 and its dedicated memory or between a computer 12 in an array 10 and an external device (through an input/output port, or the like). Indeed, it is anticipated that some applications may require arrays of arrays—with the presently described inter device communication method being potentially applied to communication among the arrays of arrays.
While specific examples of the computer array 10 and computer 12 and of the rules 1000 have been discussed therein, it is expected that there will be a great many applications for these which have not yet been envisioned. Indeed, it is one of the advantages of the present invention that the inventive method and apparatus may be adapted to a great variety of uses.
All of the above are only some of the examples of available embodiments of the present invention. Those skilled in the art will readily observe that numerous other modifications and alterations may be made without departing from the spirit and scope of the invention. Accordingly, the disclosure herein is not intended as limiting and the appended claims are to be interpreted as encompassing the entire scope of the invention.
This is a continuation-in-part of application Ser. No. 11/741,649, filed Apr. 27, 2007, hereby incorporated by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | 11741649 | Apr 2007 | US |
Child | 11741659 | US |