Modified Harvard architecture processor having program memory space mapped to data memory space

Information

  • Patent Grant
  • 6728856
  • Patent Number
    6,728,856
  • Date Filed
    Friday, June 1, 2001
    23 years ago
  • Date Issued
    Tuesday, April 27, 2004
    20 years ago
Abstract
A processor has an architecture that provides the processing speed advantages of the Harvard architecture, but does not require special purpose instructions or two separate external memories in order to expand both data memory and program instruction memory. The processor has separate program memory space and data memory space, but provides the capability to map at least a portion of the program memory space to the data memory space. This allows most program instructions that are processed to obtain the speed advantages of simultaneous program instruction and data access, yet provides a means to access program memory resident data without special purpose instructions. It also allows program memory space and data memory space to be expanded externally to the processor using only one external memory device that includes both program instructions and data. The processor includes a program memory space operable to store program instructions and data, a data memory space operable to store data, and mapping circuitry operable to map at least a portion of the program memory space to the data memory space. The program memory space may be internal to the processor. The processor may further comprise a page register operable to specify a location of the program memory space that is mapped to the data memory space.
Description




FIELD OF THE INVENTION




The present invention relates to a modified Harvard architecture processor having data memory space mapped to program memory space.




BACKGROUND OF THE INVENTION




Processors, including microprocessors, digital signal processors and microcontrollers, operate by running software programs that are embodied in one or more series of program instructions stored in a memory. The processors run the software by fetching the program instructions from the series of program instructions, decoding the program instructions and executing them. In addition to program instructions, data is also stored in memory that is accessible by the processor. Generally, the program instructions process data by accessing data in memory, modifying the data and storing the modified data into memory.




One well-known architecture for processors is known as the Harvard architecture. In this architecture, data and program instructions are stored in separate memories that can be accessed simultaneously. Because of this simultaneous access, the Harvard architecture provides significant processing speed advantages over other architectures. A typical Harvard architecture processor that includes internal memory includes two separate memories, one for data, and one for program instructions. In order to expand the memory capacity of such a processor, memory external to the processor must be added. However, since a Harvard architecture processor has two separate memories, in order to expand both data memory and program instruction memory, two separate external memories must be added. This is a significant disadvantage when low-cost systems are being built. Some conventional processors provide the capability to access data that is stored in program instruction memory using special instructions, but this is a limited solution.




A need arises for a processor having an architecture that provides the processing speed advantages of the Harvard architecture, but does not require special purpose instructions or two separate external memories in order to expand both data memory and program instruction memory.




SUMMARY OF THE INVENTION




The present invention is a processor that has separate program memory space and data memory space, but provides the capability to map at least a portion of the program memory space to the data memory space. This allows data that is stored in the program memory space to be accessed as though it were actually stored in data memory space, using standard data memory access instructions, including arithmetic and logical operation instructions. The processor has an architecture that provides the processing speed advantages of the Harvard architecture, but does not require two separate external memories, or two external memory busses, in order to expand both data memory and program instruction memory. This allows most program instructions that are processed to obtain the speed advantages of simultaneous program instruction and data access. It also allows program memory space and data memory space to be expanded externally to the processor using only one external memory device that includes both program instructions and data and without using special purpose instructions.




According to the present invention, a processor includes a program memory space operable to store program instructions and data, a data memory space operable to store data, and mapping circuitry operable to map at least a portion of the program memory space to the data memory space. The program memory space may be internal to the processor. The processor may further comprise a page register operable to specify a location of the program memory space that is mapped to the data memory space.




In one aspect of the present invention, the processor may be operably connected to an external memory device operable to store program instructions and data. The processor may further comprise circuitry operable to map at least a portion of the external memory device to the data memory space. The external memory device may further comprise program memory space. The portion of the external memory device that is mapped to the data memory space may be separate from the program memory space in the external memory device or the portion of the external memory device that is mapped to the data memory space may overlap with the program memory space in the external memory device.




In one aspect of the present invention, at least a portion of the program memory space may be internal to the processor. The processor may further comprise a page register operable to specify a location of the program memory space that is mapped to the data memory space. The processor may be operably connected to an external memory device operable to store program instructions and data. At least a portion of the external memory device may be mapped to the program memory space that is mapped to the data memory space.











BRIEF DESCRIPTION OF THE FIGURES





FIG. 1

depicts a functional block diagram of an embodiment of a processor chip within which the present invention may find application.





FIG. 2

depicts a functional block diagram of a data busing scheme for use in a processor


100


, such as that shown in FIG.


1


.





FIG. 3

depicts an exemplary memory map of a data space memory, which may be implemented in the processor shown in FIG.


2


.





FIG. 4

depicts an exemplary block diagram of program memory space to data memory space mapping which may be implemented in the processor shown in

FIG. 2

, according to the present invention.











DETAILED DESCRIPTION




Overview of Processor Elements

FIG. 1

depicts a functional block diagram of an embodiment of a processor chip within which the present invention may find application. Referring to

FIG. 1

, a processor


100


is coupled to external devices/systems


140


. The processor


100


may be any type of processor including, for example, a digital signal processor (DSP), a microprocessor, a microcontroller, or combinations thereof. The external devices


140


may be any type of systems or devices including input/output devices such as keyboards, displays, speakers, microphones, memory, or other systems which may or may not include processors. Moreover, the processor


100


and the external devices


140


may together comprise a stand alone system.




The processor


100


includes a program memory


105


, an instruction fetch/decode unit


110


, instruction execution units


115


, data memory and registers


120


, peripherals


125


, data I/O


130


, and a program counter and loop control unit


135


. The bus


150


, which may include one or more common buses, communicates data between the units as shown.




The program memory


105


stores software embodied in program instructions for execution by the processor


100


. The program memory


105


may comprise any type of nonvolatile memory such as a read only memory (ROM), a programmable read only memory (PROM), an electrically programmable or an electrically programmable and erasable read only memory (EPROM or EEPROM) or flash memory. In addition, the program memory


105


may be supplemented with external nonvolatile memory


145


as shown to increase the complexity of software available to the processor


100


. Alternatively, the program memory may be volatile memory, which receives program instructions from, for example, an external non-volatile memory


145


. When the program memory


105


is nonvolatile memory, the program memory may be programmed at the time of manufacturing the processor


100


or prior to or during implementation of the processor


100


within a system. In the latter scenario, the processor


100


may be programmed through a process called in-line serial programming.




The instruction fetch/decode unit


110


is coupled to the program memory


105


, the instruction execution units


115


, and the data memory


120


. Coupled to the program memory


105


and the bus


150


is the program counter and loop control unit


135


. The instruction fetch/decode unit


110


fetches the instructions from the program memory


105


specified by the address value contained in the program counter


135


. The instruction fetch/decode unit


110


then decodes the fetched instructions and sends the decoded instructions to the appropriate execution unit


115


. The instruction fetch/decode unit


110


may also send operand information including addresses of data to the data memory


120


and to functional elements that access the registers.




The program counter and loop control unit


135


includes a program counter register (not shown) which stores an address of the next instruction to be fetched. During normal instruction processing, the program counter register may be incremented to cause sequential instructions to be fetched. Alternatively, the program counter value may be altered by loading a new value into it via the bus


150


. The new value may be derived based on decoding and executing a flow control instruction such as, for example, a branch instruction. In addition, the loop control portion of the program counter and loop control unit


135


may be used to provide repeat instruction processing and repeat loop control as further described below.




The instruction execution units


115


receive the decoded instructions from the instruction fetch/decode unit


110


and thereafter execute the decoded instructions. As part of this process, the execution units may retrieve one or two operands via the bus


150


and store the result into a register or memory location within the data memory


120


. The execution units may include an arithmetic logic unit (ALU) such as those typically found in a microcontroller. The execution units may also include a digital signal processing engine, a floating point processor, an integer processor, or any other convenient execution unit. A preferred embodiment of the execution units and their interaction with the bus


150


, which may include one or more buses, is presented in more detail below with reference to FIG.


2


.




The data memory and registers


120


are volatile memory and are used to store data used and generated by the execution units. The data memory


120


and program memory


105


are preferably separate memories for storing data and program instructions respectively. This format is a known generally as a Harvard architecture. It is noted, however, that according to the present invention, the architecture may be a Von-Neuman architecture or a modified Harvard architecture, which permits the use of some program space for data space. A dotted line is shown, for example, connecting the program memory


105


to the bus


150


. This path may include logic for aligning data reads from program space such as, for example, during table reads from program space to data memory


120


.




Referring again to

FIG. 1

, a plurality of peripherals


125


on the processor may be coupled to the bus


125


. The peripherals may include, for example, analog to digital converters, timers, bus interfaces and protocols such as, for example, the controller area network (CAN) protocol or the Universal Serial Bus (USB) protocol and other peripherals. The peripherals exchange data over the bus


150


with the other units.




The data I/O unit


130


may include transceivers and other logic for interfacing with the external devices/systems


140


. The data I/O unit


130


may further include functionality to permit in circuit serial programming of the Program memory through the data I/O unit


130


.





FIG. 2

depicts a functional block diagram of a data busing scheme for use in a processor


100


, such as that shown in

FIG. 1

, which has an integrated microcontroller arithmetic logic unit (ALU)


270


and a digital signal processing (DSP) engine


230


. This configuration may be used to integrate DSP functionality to an existing microcontroller core. Referring to

FIG. 2

, the data memory


120


of

FIG. 1

is implemented as two separate memories: an X-memory


210


and a Y-memory


220


, each being respectively addressable by an X-address generator


250


and a Y-address generator


260


. The X-address generator may also permit addressing the Y-memory space thus making the data space appear like a single contiguous memory space when addressed from the X address generator. The bus


150


may be implemented as two buses, one for each of the X and Y memory, to permit simultaneous fetching of data from the X and Y memories.




The W registers


240


are general purpose address and/or data registers. The DSP engine


230


is coupled to both the X and Y memory buses and to the W registers


240


. The DSP engine


230


may simultaneously fetch data from each the X and Y memory, execute instructions which operate on the simultaneously fetched data and write the result to an accumulator (not shown) and write a prior result to X or Y memory or to the W registers


240


within a single processor cycle.




In one embodiment, the ALU


270


may be coupled only to the X memory bus and may only fetch data from the X bus. However, the X and Y memories


210


and


220


may be addressed as a single memory space by the X address generator in order to make the data memory segregation transparent to the ALU


270


. The memory locations within the X and Y memories may be addressed by values stored in the W registers


240


.




Any processor clocking scheme may be implemented for fetching and executing instructions. A specific example follows, however, to illustrate an embodiment of the present invention. Each instruction cycle is comprised of four Q clock cycles Q1-Q4. The four phase Q cycles provide timing signals to coordinate the decode, read, process data and write data portions of each instruction cycle.




According to one embodiment of the processor


100


, the processor


100


concurrently performs two operations—it fetches the next instruction and executes the present instruction. Accordingly, the two processes occur simultaneously. The following sequence of events may comprise, for example, the fetch instruction cycle:




Q1: Fetch Instruction




Q2: Fetch Instruction




Q3: Fetch Instruction




Q4: Latch Instruction into prefetch register, Increment PC




The following sequence of events may comprise, for example, the execute instruction cycle for a single operand instruction:




Q1: latch instruction into IR, decode, and determine addresses of operand data




Q2: fetch operand




Q3: execute function specified by instruction and calculate destination address for data




Q4: write result to destination




The following sequence of events may comprise, for example, the execute instruction cycle for a dual operand instruction using a data pre-fetch mechanism. These instructions pre-fetch the dual operands simultaneously from the X and Y data memories and store them into registers specified in the instruction. They simultaneously allow instruction execution on the operands fetched during the previous cycle.




Q1: latch instruction into IR, decode, and determine addresses of operand data




Q2: pre-fetch operands into specified registers, execute operation in instruction




Q3: execute operation in instruction, calculate destination address for data




Q4: complete execution, write result to destination




An exemplary memory map of data space memory


300


is shown in FIG.


3


. Data space memory


300


includes a plurality of blocks of memory, divided into X address memory and Y address memory. Typically, data space memory


300


is implemented as random access read-write memory (RAM), so as to allow data to be read and written as necessary. However, read-only memory (ROM) may also be advantageously used for at least a portion of data space memory


300


. For example, constant data values, look up tables, etc., may be usefully stored in ROM. In the example shown in

FIG. 3

, X address memory includes memory blocks


302


,


304


,


306


, and


308


, while Y address memory includes memory block


310


. Data space memory


300


is split into two blocks, X address memory and Y address memory. A key element of this architecture is that the Y address memory space is a subset of the X address memory space, and is fully contained within the X address memory space. In order to provide an apparent linear addressing space, the X and Y address memory spaces would typically have contiguous addresses, although this is not an architectural necessity.




In the example shown in

FIG. 3

, memory block


302


includes a block of contiguous memory, starting at data memory location 0x0000. Memory block


302


is reserved in X address memory space and is directly addressable using memory direct instructions. The remaining X address memory and Y address memory spaces are indirectly addressable using other instructions. In the example shown in

FIG. 3

, Y address memory space


310


is located between two blocks of X address memory space, block


304


and


306


. However, this is only an example, as the Y address memory space


310


may be located anywhere within the non-reserved X address memory space. The partition between the X and Y address spaces is arbitrary and is determined by the memory decode shown in FIG.


2


. Both the X and Y address generator can generate any effective address (EA) within the range of data memory space


300


. However, accesses to memory addresses that are in the other address space, or to memory addresses that are not implemented with physical memory, will return fixed data, rather than data from memory.




Memory block


308


is shown in

FIG. 3

as being an X address memory block. Memory block


308


, which includes at least a portion of data memory space


300


, may be used as X address memory, Y address memory, or a mixture of X address memory and Y address memory. When used as X address memory, memory block


308


may be mapped into program memory space. This provides transparent access to constant data, such as stored constants, look up tables, etc., from the X address data memory space without the need to use special instructions. This feature also allows the mapping of a portion of data memory space into an unused area of program memory, and since all unused internal addresses are mapped externally, to the external memory bus. This is shown in

FIG. 4

, which is an exemplary block diagram of the program memory space to data memory space mapping. Data memory space block


308


, which is a portion of data memory space


300


is mapped to a data memory space page


402


in internal program memory space


404


. The location of data memory space page


402


in internal program memory space


404


is specified by page register


406


. Internal program memory space


404


is still used for program instruction access, as specified by program counter (PC)


408


.




External memory device


410


is connected to the external memory bus


412


of the processor. External memory device


410


includes external program/data memory space


414


. Since all unused internal addresses are mapped externally to the external memory bus, data memory space mapped page


402


is also mapped to external data memory space mapped page


416


, which is located in external program/data memory space


412


. If external memory device


410


is a RAM, then data may be read from and written to external data memory space mapped page


416


. External program/data space


414


may also include additional external program memory space


418


, which may be separate from external data memory space mapped page


416


, or which may overlap with external data memory space mapped page


416


.




While specific embodiments of the present invention have been illustrated and described, it will be understood by those having ordinary skill in the art that changes may be made to those embodiments without departing from the spirit and scope of the invention.



Claims
  • 1. A processor comprising:a program memory space operable to store program instructions and data; a data memory space operable to store data; mapping circuitry operable to map at least a portion of the program memory space to the data memory space, so that data accessible to the processor by addressing the data memory space is stored in the program memory space.
  • 2. The processor of claim 1, wherein the data accessible to the processor by addressing the data memory space that is stored in the program memory space is accessible using any data memory access instruction.
  • 3. The processor of claim 2, wherein the program memory space is stored in a memory included in the processor.
  • 4. The processor of claim 3, further comprising a page register operable to specify a location of the program memory space that is mapped to the data memory space.
  • 5. The processor of claim 2, wherein the processor is operably connected to a memory device that is not included in the processor, which is operable to store program instructions and data.
  • 6. The processor of claim 5, further comprising circuitry operable to map at least a portion of the memory device that is operably connected to the processor to the data memory space, wherein data accessible to the processor by addressing the data memory space is stored in the memory device that is operably connected to the processor.
  • 7. The processor of claim 6, wherein the memory device that is operably connected to the processor further comprises memory accessible to the processor by addressing program memory space.
  • 8. The processor of claim 7, wherein the portion of the memory device that is operably connected to the processor that is mapped to the data memory space is separate from the program memory space in the external memory device.
  • 9. The processor of claim 7, wherein the portion of the memory device that is operably connected to the processor that is mapped to the data memory space overlaps with the program memory space in the memory device that is operably connected to the processor.
  • 10. The processor of claim 2, wherein at least a portion of the program memory space is stored in a memory included in the processor.
  • 11. The processor of claim 2, further comprising a page register operable to specify a location of the program memory space that is mapped to the data memory space.
  • 12. The processor of claim 11, wherein the processor is operably connected to a memory device that is not included in the processor, which is operable to store program instructions and data.
  • 13. The processor of claim 12, wherein at least a portion of the program memory space that is mapped to the data memory space is stored in the memory device that is operably connected to the processor.
US Referenced Citations (105)
Number Name Date Kind
3781810 Downing Dec 1973 A
4398244 Chu et al. Aug 1983 A
4472788 Yamazaki Sep 1984 A
4481576 Bicknell Nov 1984 A
4488252 Vassar Dec 1984 A
4511990 Hagiwara et al. Apr 1985 A
4556938 Parker et al. Dec 1985 A
4626988 George Dec 1986 A
4730248 Watanabe et al. Mar 1988 A
4782457 Cline Nov 1988 A
4807172 Nukiyama Feb 1989 A
4829420 Stahle May 1989 A
4829460 Ito May 1989 A
4839846 Hirose et al. Jun 1989 A
4872128 Shimizu Oct 1989 A
4882701 Ishii Nov 1989 A
4941120 Brown et al. Jul 1990 A
4943940 New Jul 1990 A
4959776 Deerfield et al. Sep 1990 A
4977533 Miyabayashi et al. Dec 1990 A
4984213 Abdoo et al. Jan 1991 A
5007020 Inskeep Apr 1991 A
5012441 Retter Apr 1991 A
5032986 Pathak et al. Jul 1991 A
5038310 Akagiri et al. Aug 1991 A
5056004 Ohde et al. Oct 1991 A
5099445 Studor et al. Mar 1992 A
5101484 Kohn Mar 1992 A
5117498 Miller et al. May 1992 A
5122981 Taniguchi Jun 1992 A
5155823 Tsue Oct 1992 A
5197023 Nakayama Mar 1993 A
5197140 Balmer Mar 1993 A
5206940 Murakami et al. Apr 1993 A
5212662 Cocanougher et al. May 1993 A
5239654 Ing-Simmons et al. Aug 1993 A
5276634 Suzuki et al. Jan 1994 A
5282153 Bartkowiak et al. Jan 1994 A
5327543 Miura et al. Jul 1994 A
5327566 Forsyth Jul 1994 A
5379240 Byrne Jan 1995 A
5418976 Iida May 1995 A
5448703 Amini et al. Sep 1995 A
5448706 Fleming et al. Sep 1995 A
5463749 Wertheizer et al. Oct 1995 A
5469377 Amano Nov 1995 A
5471600 Nakamoto Nov 1995 A
5497340 Uramoto et al. Mar 1996 A
5499380 Iwata et al. Mar 1996 A
5504916 Murakami et al. Apr 1996 A
5548544 Matheny et al. Aug 1996 A
5568412 Han et al. Oct 1996 A
5596760 Ueda Jan 1997 A
5600813 Nakagawa et al. Feb 1997 A
5619711 Anderson Apr 1997 A
5642516 Hedayat et al. Jun 1997 A
5689693 White Nov 1997 A
5694350 Wolrich et al. Dec 1997 A
5696711 Makineni Dec 1997 A
5706460 Craig et al. Jan 1998 A
5715470 Asano et al. Feb 1998 A
5737570 Koch Apr 1998 A
5740419 Potter Apr 1998 A
5748516 Goddard et al. May 1998 A
5764555 McPherson et al. Jun 1998 A
5774711 Henry et al. Jun 1998 A
5778416 Harrison et al. Jul 1998 A
5790443 Shen et al. Aug 1998 A
5808926 Gorshtein et al. Sep 1998 A
5812439 Hansen Sep 1998 A
5825730 Nishida et al. Oct 1998 A
5826096 Baxter Oct 1998 A
5828875 Halvarsson et al. Oct 1998 A
5862065 Muthusamy Jan 1999 A
5867726 Ohsuga et al. Feb 1999 A
5880984 Burchfiel et al. Mar 1999 A
5892697 Brakefield Apr 1999 A
5892699 Duncan et al. Apr 1999 A
5894428 Harada Apr 1999 A
5909385 Nishiyama et al. Jun 1999 A
5917741 Ng Jun 1999 A
5930159 Wong Jul 1999 A
5930503 Drees Jul 1999 A
5938759 Kamijo Aug 1999 A
5941940 Prasad et al. Aug 1999 A
5943249 Handlogten Aug 1999 A
5951627 Kiamilev et al. Sep 1999 A
5951679 Anderson et al. Sep 1999 A
5991787 Abel et al. Nov 1999 A
5996067 White Nov 1999 A
6009454 Dummermuth Dec 1999 A
6014723 Tremblay et al. Jan 2000 A
6026489 Wachi et al. Feb 2000 A
6044392 Anderson et al. Mar 2000 A
6044434 Oliver Mar 2000 A
6058409 Kozaki et al. May 2000 A
6058410 Sharangpani May 2000 A
6058464 Taylor May 2000 A
6061780 Shippy et al. May 2000 A
6076154 Van Eijndhoven et al. Jun 2000 A
6101521 Kosiec Aug 2000 A
6115732 Oberman et al. Sep 2000 A
6128728 Dowling Oct 2000 A
6134574 Oberman et al. Oct 2000 A
6145049 Wong Nov 2000 A