EP1066559B1 - A digital signal processor reducing access contention - Google Patents

A digital signal processor reducing access contention Download PDF

Info

Publication number
EP1066559B1
EP1066559B1 EP99911150A EP99911150A EP1066559B1 EP 1066559 B1 EP1066559 B1 EP 1066559B1 EP 99911150 A EP99911150 A EP 99911150A EP 99911150 A EP99911150 A EP 99911150A EP 1066559 B1 EP1066559 B1 EP 1066559B1
Authority
EP
European Patent Office
Prior art keywords
data
instruction
set forth
memory
register
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP99911150A
Other languages
German (de)
French (fr)
Other versions
EP1066559A1 (en
Inventor
Gilbert C. Sih
Quizhen Zou
Sanjay K. Jha
Inyup Kang
Jian Lin
Quaeed Motiwala
Deepu John
Li Zhang
Haitao Zhang
Way-Shing Lee
Charles E. Sakamaki
Prashant A. Kantak
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qualcomm Inc
Original Assignee
Qualcomm Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/044,088 external-priority patent/US6496920B1/en
Priority claimed from US09/044,086 external-priority patent/US6425070B1/en
Application filed by Qualcomm Inc filed Critical Qualcomm Inc
Priority to EP04005665.7A priority Critical patent/EP1457876B1/en
Publication of EP1066559A1 publication Critical patent/EP1066559A1/en
Application granted granted Critical
Publication of EP1066559B1 publication Critical patent/EP1066559B1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0877Cache access modes
    • G06F12/0886Variable-length word access
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored program computers
    • G06F15/78Architectures of general purpose stored program computers comprising a single central processing unit
    • G06F15/7807System on chip, i.e. computer system on a single chip; System in package, i.e. computer system on one or more chips in a single package
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30007Arrangements for executing specific machine instructions to perform operations on data operands
    • G06F9/30036Instructions to perform operations on packed data, e.g. vector, tile or matrix operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30098Register arrangements
    • G06F9/30105Register structure
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30098Register arrangements
    • G06F9/30105Register structure
    • G06F9/30112Register structure comprising data of variable length
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30098Register arrangements
    • G06F9/3012Organisation of register space, e.g. banked or distributed register file
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30098Register arrangements
    • G06F9/30141Implementation provisions of register files, e.g. ports
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30145Instruction analysis, e.g. decoding, instruction word fields
    • G06F9/30149Instruction analysis, e.g. decoding, instruction word fields of variable length instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30145Instruction analysis, e.g. decoding, instruction word fields
    • G06F9/3016Decoding the operand specifier, e.g. specifier format
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30145Instruction analysis, e.g. decoding, instruction word fields
    • G06F9/3016Decoding the operand specifier, e.g. specifier format
    • G06F9/30167Decoding the operand specifier, e.g. specifier format of immediate specifier, e.g. constants
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3802Instruction prefetching
    • G06F9/3816Instruction alignment, e.g. cache line crossing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3885Concurrent instruction execution, e.g. pipeline, look ahead using a plurality of independent parallel functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3885Concurrent instruction execution, e.g. pipeline, look ahead using a plurality of independent parallel functional units
    • G06F9/3893Concurrent instruction execution, e.g. pipeline, look ahead using a plurality of independent parallel functional units controlled in tandem, e.g. multiplier-accumulator

Definitions

  • the present invention relates to a digital signal processor.
  • the present invention has particular, but not exclusive, application to digital signal processing using highly parallel, highly pipelined, processing techniques.
  • DSPs Digital Signal Processors
  • a digital signal is typically a series of numbers, or digital values, used to represent a corresponding analog signal.
  • DSPs are used in a wide variety of applications including audio systems such as compact disk players, and wireless communication systems such as cellular telephones.
  • a DSP is often considered to be a specialized form of microprocessor. Like a microprocessor, a DSP is typically implemented on a silicon based semiconductor integrated circuit. Additionally, as with microprocessors, the computing power of DSPs is enhanced by using reduced instruction set (RISC) computing techniques. RISC computing techniques include using smaller numbers of like sized instructions to control the operation of the DSP, where each instruction is executed in the same amount of time. The use of RISC computing techniques increases the rate at which instruction are performed, or the clock rate, as well as the amount of instruction pipelining within the DSP. This increases the overall computing power of the DSP.
  • RISC reduced instruction set
  • RISC based DSPs execute a greater number of instructions to perform a given task. Executing additional instructions increases the power consumption of the DSP, even though the time to execute those instructions decreases due to the improved clocking speed of a RISC based DSP. Additionally, using a greater number of instructions increases the size of the on-chip instruction memory within the DSP. Memory structures require substantial (often more than 50% of the total) circuit area within a DSP, which increases the size and cost of the DSP. Thus, the use of RISC based DSPs is less than ideal for low cost, low power, applications such as digital cellular telephony or other types of battery operation wireless communication systems.
  • Fig. 1 is a highly simplified block diagram of a digital signal processor configured in accordance with the prior art.
  • Arithmetic logic unit (ALU) 16 is coupled to ALU register bank 17 and multiply accumulate (MAC) circuit 26 is coupled to MAC register bank 27.
  • Data bus 20 couples MAC register bank 27, ALU register 17 and (on chip) data memory 10 .
  • Instruction bus 22 couples MAC register bank 27, (on-chip) instruction memory 12, MAC register bank 27 and ALU register bank 17.
  • Instruction decode 18 is coupled to MAC 26 and ALU 16, and in some prior art systems instruction decode 18 is coupled directly to instruction memory 12.
  • Data memory 10 is also coupled to data interface 11 and instruction memory 12 is also coupled to instruction interface 13.
  • Data interface 11 and instruction interface 13 exchange data and instructions with off-chip memory 6 .
  • instruction decode 18 During operation, the instructions in instruction memory 12 are decoded by instruction decode 18. In response, instruction decode 18 generates internal control signals that are applied to ALU 16 and MAC 26 .
  • the control signals typically cause ALU 16 to have data exchanged between ALU register bank 17 and data memory 10 or instruction memory 12. Also, the control signals cause MAC 26 to have instruction data exchanged between MAC register bank 27 and instruction memory 12 or data memory 10. Additionally, the control signals cause ALU 16 and MAC 26 to perform various operations in response to, and on, the data stored in ALU register bank 17 and MAC register bank 27 respectively.
  • instruction memory 12 may contain coefficient data for use by ALU 16 and MAC 26 and data memory 10 may contain data to be processed (signal data).
  • the coefficient data may be for implementing a frequency filter using the DSP, which is a common practice.
  • DSP digital signal processor
  • Additional instruction data within instruction memory 12 is also applied to instruction decode 18, either through instruction data bus 22 or through a direct connection.
  • the additional instruction data specifies the operation to be performed by MAC 26.
  • the results generated by MAC 26 are typically read back into data memory 10.
  • processing inefficiencies result from this prior art processing. These processing inefficiencies include, e.g., bus, or access contention, to instruction memory 12, which must supply instruction data to both MAC register 26 and instruction decode 18, as well as bus, or access contention, to data memory 10, which must both read out signal data and write in the output data. Additionally, in many instances, additional processing on the output data must be performed by ALU 16. This further aggravates access to data memory 10, and therefore creates contention for data bus 20, because the output data must be written from MAC register bank 27 into data memory 10, and then read out to ALU register 17. These read and write operation are performed over bus 20 and therefore consume additional bus cycles. Such inefficiencies reduce the processing performance of the DSP.
  • US-A-5,710,914 describes a digital signal processing method and system implementing pipelined read and write operations.
  • the system includes a program control unit, a memory, a data processing unit, and a dedicated bus for writing processed data from the data processing unit to the memory.
  • the dedicated write bus By using the dedicated write bus, the system goes some way to avoiding bus contention in a five stage pipeline operation involving fetch, decode, read, execute and write operations.
  • US-A-5,293,611 describes a digital signal processor utilising a multiply-and-add function for digital filter realisation.
  • the processor has a bypass signal path for transferring data without the intervention of any bus which is extended between the two ports of the data RAM.
  • sampling data items to be read out of a predetermined address of the data RAM can be transmitted to an internal bus and can simultaneously be written into the next address of the data RAM.
  • US-A-5 522 083 discloses a multiprocessor system wherein each processor comprises a register bank and has access to a first, second and third memory bank via a global bus and via a local bus. Global bus and local bus can be used simultaneously to provide data from the first and second memory banks to the processor and one of said buses can be used to write the result back to the third memory bank.
  • the present invention seeks to improve the performance and usefulness of a DSP by addressing the problems and inefficiencies listed above, as well as by providing other features and improvements described throughout the application.
  • the present invention aims to provide, a novel and improved method and circuit for digital signal processing.
  • a digital signal processor for processing signals in response to instruction data as set forth in claim 1.
  • the present invention provides a novel and improved method and circuit for digital signal processing.
  • signals, instructions, and data are preferably represented by electrical voltages, currents, current stores including charged, optical, or magnetic particles, or some combination thereof, the use of which is well known.
  • the use of various chemical and biological compounds to represent such signals, instructions, and data is generally consistent with the use of the present invention as well, although not preferred due to the difficulty to use, control and manipulate such items.
  • Fig. 2 is a block diagram of a portion of a digital signal processor (DSP) circuit configured in accordance with an exemplary embodiment of the invention.
  • Data memories 102 - 104 are coupled to data buses A, B and C respectively via address generation units (AGU) 105 - 107, and to data interface 100.
  • Data buses A, B and C are coupled to output ports PO1, PO2 and PO3 respectively of register bank 120, and to input ports PI1, PI2, and PI3 respectively of register bank 120 through multiplexers 122 - 126.
  • data buses A, B and C read and write data between data memories 102 - 104 and the registers within register bank 120.
  • three data buses and three data memories allows more data to be exchanged between the register banks and the data memories without creating bus contention.
  • three fetch operations can be performed simultaneously from the three memories 102 - 104 using the three data buses A, B and C.
  • three write operations can be performed simultaneously, as can any combination of three fetch and write operations.
  • a fourth data bus would allow even greater numbers of operations to be performed, and is consistent with some embodiments of the invention.
  • the use of only three data buses has particular advantages, because three buses facilitates performing many tasks commonly performed by a DSP such as filtering.
  • the addition of a fourth data bus does not provide the same incremental performance improvement as the addition of a third data bus, and requires the same amount of additional circuit area. Therefore, the addition of a fourth data bus provides incrementally less benefit than the addition of a third bus. So, in many embodiments of the invention the use of only three data buses is preferred.
  • Output ports PO4, PO5 and PO6 of register bank 120 are coupled to multiply accumulate (MAC) unit 128, the output of which is in turn coupled to input port PI4 of register bank 120.
  • Output ports PO7 and PO8 of register bank 120 are coupled to arithmetic logic unit (ALU) 130, the output of which is coupled to input port PI5 of register bank 120.
  • ALU arithmetic logic unit
  • Instruction memory 152 is coupled to instruction fetch unit 156 and instruction interface 150.
  • Instruction decoder 158 is coupled to instruction fetch unit 156, and to immediate bus Im1, immediate bus Im2 and immediate bus Im3, as well as immediate bus ImALU .
  • Immediate buses Im1, Im2 and Im3 are coupled to multiplexers 122, 124 and 126.
  • Immediate bus ImALU is coupled to ALU 130.
  • decoder 158 is coupled to the various subsystems shown by control connections (not shown for ease of drawing).
  • Register bank 120 contains eight (8) registers labeled L0-L3 and D0-D3.
  • Registers L0-L3 are forty (40) bit wide registers and can also be accessed in sixteen bit fragments via high word registers L0h - L3h and low word registers L01 - L31.
  • Registers D0 - D3 are thirty-two (32) bits wide, and can be accessed in sixteen bit fragments via subregisters R0-R7. In general, the registers and subregisters are referred to as simply "registers", with the particular nature of the registers made apparent by the particular register number provided.
  • One aspect of the invention is realized by having some registers coupled to, and therefore accessible by, multiple input and output ports.
  • this multiconnectedness is provided by the use of multiplexers coupled to the input of each register, and to each output port.
  • Other methods of providing multiconnectedness will be apparent, and are consistent with the use of some aspects of the present invention, including, e.g., using data buses and addressable memories.
  • the use of multiplexers is preferred in some embodiments because they provide rapid and controllable access to the various registers and ports.
  • Fig. 3 is a block diagram illustrating the connections between the set of registers within register bank 120 and the set of input ports PI1 - PI5.
  • the registers are defined as L0h - L3h, L01 - L31, and R0 - R7.
  • a register L0 is comprised of registers L0h and L01.
  • the registers L0h - L3h are 24 bits and registers and L01 - L31 and R0 - R7 are 16 bits, making registers L0 - L3 40 bits wide.
  • input ports PI3 - PI5 are comprised of input ports PI3h - PI5h of 24 bits and PI31 - PI51 of 16 bits for a total of 40 bits.
  • the input ports PI1 and PI2 are only 16 bits, and when used to write to registers L0h - L3h, write only to the least significant 16 of the 24 bits available.
  • registers L0 - L3 receive data from all the input ports, while other registers receive data from only some, or a portion, of the input ports.
  • all of the registers L0 - L3 receive data from all the input ports PI1 - P15, from multiplexers 500 - 514, with the 16 bit input ports being able to write to both the higher and lower registers within registers L0 - L3.
  • register L0 - L3 receive input from any bus A - C (corresponding to input ports PI0 - PI3) and from MAC unit 128 and ALU 130 (corresponding to input ports PI4 and PIS.)
  • Registers R0 - R7 receive input data from and bus A-C via multiplexers 516 - 530.
  • registers R0 - R7 receive input data from MAC unit 128 (input port PI4). Additionally, registers R0 - R3 receive input data from ALU unit 130 via multiplexers 516, 518, 524 and 526 .
  • the embodiment shown in Fig. 3 has various advantages. In particular, it provides sufficient connectivity between the input ports and registers to facilitate the most common operations, but the total connectivity is kept at a minimum to reduce the total circuit area required for implementing the circuit.
  • the output of MAC unit 128 is coupled only to long registers L0 - L3. This is beneficial because the result of multiply and accumulate operations generally exceed 32 bits, so coupling the output of MAC unit 128 to registers D0 - D3 provides minimal benefit.
  • ALU unit 130 can output to registers L0 - L3 and R0 - R3.
  • ALU unit 130 This enhances flexibility as data from ALU unit 130 may be written to a variety of registers, which is useful since ALU unit 13 performs a greater variety of operations making it useful to output data to a greater number of registers.
  • ALU unit 130 is not coupled to all the registers, and therefore unnecessary and excessive connectivity is avoided.
  • Fig. 4 is a block diagram illustrating the coupling of the output ports of register bank 120 to the registers when performed in accordance with one embodiment of the invention.
  • output port PO1 which outputs to BUS A, is coupled via multiplexer 540 to registers L0h - L3h, L0l - L3l and R0 - R7, which comprises all the available registers when accessed as subregisters.
  • output port PO2, which outputs to BUS B is coupled via multiplexer 542 to registers L0h - L3h, L01 - L3l and R0 - R7.
  • Output port PO4 which is coupled to the 40-bit input of MAC unit 128, is coupled via multiplexer 532 to registers L0 - L3. Since the values accumulated by MAC 128 tend to be large due to the nature of the multiply and accumulate operations performed, coupling output port PO4 to only the 40 bit "long" registers L0 - L3 provides an optimal coupling arrangement, because the utility derived from providing additional couplings to registers D0 - D3 is low, given that the accumulation of various multiply operations will typically exceed 32 bits.
  • Output port PO5 which is coupled to one 16 bit input of MAC unit 128, is coupled by multiplexer 534 to registers L0h - L3h, R0, R2, R4 and R6.
  • Output port PO6 which is coupled to the second 16 bit input port of MAC unit 128, is coupled to registers L0h - L3h, L01 - L31 and R0 - R7.
  • Output port PO7 which is coupled to an input of ALU 130, is coupled by multiplexer 546 to registers L0 - L3, L0h - L3h and R0 - R3, where L0h - L03h and R0 - R3 are output in conjunction with a set of logic zeros. That is, registers L0h - L03h and R0 - R3 are output to the bits 31-16 (bits numbered 0 - 39) bits of PO7, with bits 0-15 set to logic zero and bits 39 - 32 are sign extended using bit 31.
  • registers that are accessible by both the multiple data buses and the multiple processing units provides various advantages.
  • the registers provide an interface between the data buses and processing units, reducing the need to route each data bus to each processing unit. Reducing data bus routing saves circuit area and reduces chip cost.
  • coupling at least some (a set) of registers to multiple processing units allows multiple operations to be performed on the same data using the multiple processing units, without having to read and write the data over the data bus and to the memory. This saves bus cycles and therefore reduces bus contention. Instruction processing pipelining is also facilitated because data processed by a first processing unit during a first instruction cycle may then be processed further by a second processing unit during a second processing cycle within the same register.
  • pipelining is further enhanced by the use of two-phase clocked registers within register bank 120 .
  • the two-phase clocked registers are read on a first phase of the clock, and then written to on the second phase of the clock, within the same full clock ("processing") cycle.
  • processing full clock
  • data already processed by a first processing unit, such as the MAC 128, can be read out during the first clock phase and further processed by a second processing unit, such at ALU 130 , within the remaining portion of the processing cycle.
  • new data just processed by the MAC 128 is written into the same register, thus allowing complete pipeline processing between two processing units during one processing cycle. Once again, these operations are performed without running the data over any of the internal busses, and therefore increased bus contention is avoided.
  • instruction fetch unit 156 retrieves binary instructions from instruction memory 152, or if not available in instruction memory 152, from a memory located externally.
  • the external memory can take many forms well known in the art such as dynamic and static random access memory (DRAM and SRAM) or some derivative thereof, magnetic or optical hard disk memory or some other data storage medium well known in the art.
  • the instructions are of variable length and instruction fetch unit determines the length of the instruction and how much additional instruction data to fetch during each processing or clock cycle. Additionally, the instructions are stored in consecutive memory locations within the internal memory and external memory. The operation of fetch unit 156 and the storage of instruction data within memory and external memory are described in greater detail below.
  • Instruction decoder 158 receives the instructions retrieved by instruction fetch unit 156 and translates the instructions into control signals that are applied to one or more of the subsystems that make up the DSP including the data memories, register bank, MAC and ALU. Additionally, instruction decoder 158 may route immediate data that is contained within the receive instructions to the appropriate system via immediate buses Im1, Im2, Im3 or ImALU. Immediate data is typically numeric values stored within the instruction data that are used to perform operations on the data stored in data memories 102 - 104 or which may specify or modify an address.
  • Operations performed by the DSP of Fig. 2 include loading of data from a data memory into a register location via one of the data buses. Data may also be written from a register into a data memory. Also, MAC 128 or ALU 130 may perform operations on the data stored in one or more of the registers within register bank 120 , where the results are typically written back into a register within register bank 120.
  • the DSP architecture described above provides numerous advantages. For example, the use of three data buses facilitates the uninterrupted pipeline processing of data.
  • the data to be filtered (signal data) is stored in one data memory and the coefficients to be applied to that data are stored in the other data memory. Storing the signal data and coefficient data in the two narrower memories is preferred, as the results of the operations typically require more bits than the operands.
  • the coefficient and signal data are then read into register bank 120 and then multiplied and accumulated by MAC unit 128.
  • the results of these operations may be stored in a second register within register bank 120 , or overwritten into the register bank in which the input data was previously stored. Any results are then typically written from the register over the third bus (BUS C) into the wider (memory C) data memory.
  • the output data is written into a third memory via a third bus
  • the input data sets are read from first and second data memories via first and second data buses
  • little or no memory access conflict or bus contention occurs.
  • the processing of data may proceed uninterrupted, reducing the need to clock any of the memory subsystems or data buses at higher rates than the internal buses or other subsystems. This reduces power consumption while maintaining, or increasing, processing speed.
  • running the data through the DSP without interruption facilitates pipelined processing of the data where a number of different data values are processed differently at different stages within the DSP at any given time. Also, when combined with the use of parallel instructions as described below, significant processing flexibility can be achieved in combination with this efficient highly pipelined processing, thus providing a highly versatile, efficient and powerful DSP system.
  • data to be divided may be supplied as input data from memory C via data bus C, with the divisor supplied by another memory and bus, such as memory A and data bus A.
  • the result can then be stored in the remaining memory (memory B) by way of the remaining bus (data bus B).
  • data to be accumulated by MAC unit 130 is provided via a first memory and first bus (for example memory A and data bus A). After a set of accumulations is performed, the resulting data may be written to memory C via data bus C. Simultaneously, data to be logically shifted is provided from memory C via data bus C to ALU unit 130 during the processing cycles where data bus C is not carrying result data from MAC unit 128, which is most of the time such result data will only be available after a set of accumulate operations are performed. The logically shifted data is simultaneously written to memory B via data bus B.
  • the use of multiple data buses and memories in general, and particularly in conjunction with multiple processing units facilitates performing multiple operations by providing more possibilities for moving data within the DSP.
  • registers that are accessible by the multiple processing units, e.g., MAC unit 128 and ALU unit 130.
  • Registers accessible by multiple processing units allow data that must be processed by the processing units to be accessed without moving the data across any of the internal data buses. For example, data can be written to a register by a first processing unit, and then further processed by the second processing unit which may also access that register. This further relieves bus contention and congestion, and therefore maintains high data throughput.
  • results generated by MAC 26 increases in size (both in terms of absolute value and the number of bits used to represent the value) as the number of products over which accumulation is performed grows. Eventually, the result will have to be scaled, or "normalized", which typically calls for a logical shift operation by ALU unit 130.
  • the scaling operation can be performed simultaneously with the multiply and accumulate operations of the filtering.
  • unprocessed signal data and filtering coefficients are read from data memories 102 and 103 into registers within register bank 120 (for example L0h and L01).
  • MAC unit 128 reads the values previously stored in those registers (L0h and L01) and performs a multiply and accumulate operation where the output is written to a second register (for example L1).
  • ALU unit 130 reads the data previously stored in the second register (L1) and performs the scale operation writing the scaled value to a third register (for example L2).
  • the previous value stored in the third register (D0) is written into data memory 104 using bus C 112.
  • the particular operations may vary in accordance with the particular task being performed.
  • the use of highly parallel instructions that permit multiple operations to be performed further facilitates highly pipelined multi-instruction operations.
  • the highly parallel instructions allow specification of the different operations to be pipelined during each processing cycle.
  • this processing can all be done during a single processing cycle, where the data is read out from each register during the first clock phase, processed by the processing units, and the results written over the old data into the register during the second clock phase. It should be understood, that the same value is not subjected to all the steps in this process during a single clock cycle, but rather a set of values are pipelined through the DSP, each moving to the next step as the processing is performed.
  • Vocoding is the process of coding voice data. Vocoding requires many different types of operations to be performed, some of which can be performed independently, and therefore simultaneously. The use of multiple data buses and multiple processing units facilitates performing these operations.
  • Fig 5. is a table illustrating the packing of a set of variable length instructions within a portion of the addressable memory space of instruction memory 152 of Fig. 2 in accordance with one embodiment of the invention.
  • the variable length instructions may also be stored as shown in Fig. 2 within the external memory system to realize additional memory efficiencies.
  • Example addresses are shown in the left column, with each address pointing to a 32-bit data word, shown in the middle and right columns of memory 275.
  • the middle column represents a 16-bit high order subword and the right most column represents a sixteen-bit low order subword of each data word.
  • the high and low order subwords are not individually addressable in the preferred embodiment of the invention in order to reduce the amount of necessary address logic.
  • variable length instructions A-L are stored in the packed configuration shown.
  • Instruction A is a 48 bit instruction with the first two double-bytes A(1) and A(2) stored in address word 0x0000 and the third double-byte A(3) stored in the high order subword of address 0x0001.
  • Instruction B which follows instruction A, is a 32 bit instruction with the first double-byte B(1) stored in the low order word of address 0x0001 and the second double-byte B(2) stored in the high order subword of address 0x0002.
  • Instruction C is a 16-bit instruction with the first and only double-byte C(1) stored in the low order subword of address 0x0002.
  • this reduces the size or amount of memory 275 necessary to store a set of instructions by storing portions of different instructions within the same address word.
  • the third double-byte A(3) of instruction A is stored along with the first double-byte B(1) of instruction B.
  • variable length instructions By storing variable length instructions across word boundaries, or more particularly in consecutive locations within the memory address space, the amount of instruction memory required to store a given number of instructions is reduced. Reducing the amount of instruction memory reduces the size and cost of the die necessary to give the DSP a given amount of instruction caching capability.
  • the packing of instructions is further illustrated by placement of variable length instructions D - L within memory 275 as shown in Fig. 3.
  • packing all the instructions in consecutive locations is not necessary in some embodiments of the invention. For example, different embodiments of the invention pack only a substantial portion (e.g. 90% or more) of the instructions in consecutive locations within memory space. In other embodiment of the invention, only a significant portion (e.g. between 25 to 50%) of the instructions is advantageously packed in consecutive memory space. Other embodiments of the invention may use still other percentages of packed instructions.
  • consecutive locations are not necessary.
  • the instructions must simply be placed in a total memory space that is not substantially larger than the total amount of instruction data. This is preferably achieved by placing the instructions in adjacent locations within memory space, however, the instructions could be shuffled throughout memory space, so long as the instruction could be read out in intended order of execution.
  • this type of predetermined shuffling as a remapping of memory space, and that such remapping typically does not affect the operation of the invention, other than by adding generally undesirable complexity.
  • the packing scheme employed over a large set of instructions.
  • employing the packing scheme over at least ten instructions is preferred in some embodiments of the invention.
  • the particular packing scheme used in the exemplary embodiment of the invention is not necessary in some other embodiments of the invention.
  • some other embodiments of the invention may not have instructions in consecutive memory locations. Rather, instructions may separated by some small amount of memory space, including the use of an instruction separator code.
  • the small amount of memory space is less than the amount of memory space necessary to keep instruction boundaries on memory word boundaries.
  • the above-described packing is preferred in many instances, however, due to its simplicity, completeness and efficiency. In general, the chosen tradeoff between the completeness of packing and the complexity of the packing scheme can differ in different embodiments of the invention.
  • some embodiments of the invention employ the packing scheme on only portions of the instructions, but not the entire set of available instructions.
  • instruction packing may be performed only on sets of instructions which are used to perform a particular task or subroutine.
  • variable length instructions only consume the amount of data necessary to request the desired operations
  • highly packed instruction storage keeps the total memory equal to that consumed by the set of variable length instructions, and therefore at a minimum. Reduced memory size reduces chip size and cost.
  • the use of highly packed instructions and variable length instructions provides additional unexpected advantages when combined with other features of the architecture described above. For example, by reducing the size of instruction memory, additional circuit area is made available for the use of three data buses within the DSP, which provides the benefits as described above, including uninterrupted, highly pipelined, data processing, and the ability to perform multiple operations simultaneously within the DSP.
  • tightly packed instructions combines with the multibus architecture to provide the additional, unexpected, benefits of increased performance and efficiency.
  • FIG. 6 is a flow diagram of the operation of instruction fetch unit 156 when fetching instructions from instruction memory 152 in accordance with one embodiment of the invention.
  • the processing begins at step 200 and at step 202 a first set of instruction data is read from instruction memory 152.
  • a first set of instruction data is read from instruction memory 152.
  • two 32-bit words, or 64 bits, of instruction data are retrieved at step 202.
  • the first instruction contained in the 64 bits of retrieved instruction data is processed by instruction decoder 158.
  • the instruction may be 16, 32 or 48 bits long.
  • the instruction length is determined by a set of header bits contained in each instruction which indicate the instruction length as described in greater detail below.
  • Various other methods for specifying instruction length should be apparent including the use of codes that demarcate and separate two instructions, or the use of a super header instruction, which specifies the length of some set of instructions which are to follow.
  • the use of header bits is preferred in some instances because the instruction length information is kept in close proximity to the instruction, thereby reducing the need to store or maintain state information about the instruction processing.
  • step 206 After the first instruction contained within the 64 bits of retrieved instruction data is processed, it is determined at step 206 if 48 bits or more unprocessed instruction data remain in the 64 bits of retrieved instruction data. If 48 bits or more unprocessed instruction data remains, the next instruction contained in the remaining 48 bits of unprocessed data is processed again at step 204.
  • additional instruction data is loaded from instruction memory 152.
  • additional instruction data is loaded from the instruction memory to return the amount of unprocessed data stored in the instruction fetch unit to 48 bits. Ensuring the 48 bits of unprocessed data are stored within the instruction fetch unit ensures that at least one complete instruction is available to instruction decoder 158 .
  • instruction fetch unit retrieves a variable amount of data when less than 48 bits of unprocessed data remains depending on the particular amount of data that has been processed. In particular, if the amount of data processed is equal to or exceeds a data word (32 bits), an additional data word (32 bits) of new instruction data is retrieved. If the amount of data previously processed is equal to or exceeds two data words (64) bits, two new data words are retrieved by instruction fetch unit.
  • Determining the amount of data retrieved based on the number of words of data processed is preferred, because it keeps a sufficient amount of unprocessed data available to instruction decoder 158 while also allowing more efficient word-length access to the memory banks that make up the instruction memory. Once additional unprocessed instruction data is retrieved at step 206, the next instruction is processed within the total amount of unprocessed instruction data now available.
  • Fig. 7 is a block diagram of instruction fetch unit 156 and instruction memory 152 configured in accordance with one embodiment of the invention.
  • Instruction memory 152 is comprised of even memory bank 302 (RAM0) and odd memory bank 300 (RAM1) each of which reads and writes 32-bit data words.
  • the memory banks are labeled even and odd because they both are addressed within the same address space, but even addresses are directed to even memory bank 302 and odd addresses are directed to odd memory bank 300 .
  • Memory banks that read and write other word sizes including 8, 16, 24, 48 and 64 bits words may be used in alternative embodiments of the invention. Additionally different numbers of memory banks may be used including 1 - 8 memory banks. The use of two memory banks with 32-bits words, however, is preferred because it reduces the overall complexity while also allowing instruction data to be addressed in manageable chunks.
  • Control logic 304 causes data words to be read from memory banks 300 and 302 to instruction registers 306 and 307 .
  • the particular memory locations read are specified by address lines 310 and 314 and the reading of the instruction is controlled by enable lines 332, 315, 316 and 318.
  • the 32 bit outputs of instruction registers 306 and 307 are applied in 16 bit portions to inputs A, B, C and D of rotator 308 .
  • Rotator 308 outputs 48 bits of instruction data 324 .
  • the 48 bits of instruction data 324 are comprised of three of the four (3:4) inputs A, B, C and D with each input set containing 16 bits, as described in greater detail below.
  • control logic 304 loads instruction data from instruction memory banks 300 and 302 in accordance with the method described with reference to Fig. 5 .
  • control logic 304 first loads a total of 64 bits of unprocessed instruction data into instruction registers 306 and 307 by reading a 32 data word from both even memory bank 302 and odd memory bank 300. If a 16-bit instruction is processed, no new data is loaded because instruction registers 306 and 307 still contain 48 bits of unprocessed instruction data. If a 32-bit instruction is then processed, instruction register 306 is loaded with a 32-bit word of additional instruction data because less than 48 bits of unprocessed instruction data remains.
  • control logic 304 further configures rotator 308 using control signals 320 to output the next 48 next bits of instruction data received on inputs A, B, C and D based on 1) the location of the instruction data within the instruction address space, 2) the set of instruction data that has been processed, and 3) the length of the previous instruction processed.
  • rotator 308 is configured to output the set of 48 bits of instruction data next in line to be processed with the next-in-line bits of that instruction data in the most significant, or leftmost, position.
  • rotator 308 is configured to output the instruction data received on inputs C, D and A (CDA) in that order.
  • new instruction data is loaded into data registers 306 and 307 as described above, and rotator 308 is configured to continue to output the next-in-line instruction data on output 324 based on the size of the previous instruction processed.
  • the size of the previous instruction processed is conveyed to control logic 304 by header data 322, which is a copy of the first five bits of output 324.
  • header data 322 is a copy of the first five bits of output 324.
  • the size of the previous instruction is coded into two bits of state information I1 and I0 in accordance with Table 1 .
  • Full instruction formats I1 I0 Instruction Size 00 Branch/Stall/Reset 01 16-bit 10 32-bit 11 48-bit
  • rotator 308 is controlled by two select bits S1 and S0 that make up control 320 which are coded as set forth in Table 2 .
  • S1 S0 Rotator Output 00 ABC 01 BCD 10 CDA 11 DAB
  • the output of rotator 308 is left-rotated, or barrel-shifted.
  • the left-rotation is such that each input group (A, B, C and D) is shifted to the left on the output.
  • the input group that was at the left most position of the output is removed.
  • the input group previously not asserted at the output is subsequently output at the rightmost position.
  • rotator 308 begins with an output of ABC, and select bits S1 and S0 at 00. If a 16-bit instruction is received, the corresponding instruction length bits of I1 and I0 of 01 are added to S1 and S0 yielding a S1 and S0 of 01, which corresponds to an output 324 from rotator 308 of BCD.
  • An output of BCD is the next-in-line set of instruction data after the first 16-bits of instruction data (input A) have been processed.
  • next instruction is a 32-bit instruction
  • an instruction length I1 and I0 of 10 is added to the current S1 and S0 state of 01 yielding 11.
  • the resulting output is DAB, which corresponds to the next 48-bits of instruction data that have not been processed, configured with the next-in-line instruction data received on input D positioned in the most significant, or left most, position.
  • the instruction data from inputs B and C had been processed. It should be noted that during the processing of the previous 32-bit instruction new data is loaded into instruction register 307 in accordance with the instruction data loaded process described above.
  • control logic configures rotator 308 to output the next 48 bits of instruction data received from instruction registers 306 and 307 with the next instruction bits to be processed located in the left most position.
  • rotator 308 in addition to outputting the next 48 bits of unprocessed instruction data, rotator 308 also indicates the size of the next instruction to be processed to control logic 304. In particular, rotator 308 outputs an additional copy of the next 5 bits of instruction data to be processed to control logic 304. In the preferred embodiment of the invention, the length of the instruction is specified by the first five (5) bits of the instruction.
  • Fig. 8 is a block diagram of MAC unit 128 when configured in accordance with one embodiment of the invention.
  • Shift right 900 receives the 40-bit input to be accumulate and shifts the value by either 0 or 16 bits, with the output applied to one input of multiplexer 901.
  • the other input of multiplexer 901 receives the value 0x8000.
  • Multiplier 902 receives two 16 bits values to be multiplied along with sign bits from instruction decode 158 for a total of 17-bits for each input.
  • multiplier 902 is received by shift left 904, which shifts the output by 0, 1, 2 or 3 bits as specified by instruction decode 158.
  • Adder/subtractor 906 receives the output of multiplexer 901 and shift left 904. Adder/subtractor 906 performs addition or subtraction of the two input values as instructed by instruction decoder 158 and outputs the result, which is applied to register bank input port PI4 in the exemplary embodiment of the invention.
  • a shift right 16 unit 900 within MAC unit 128 provides additional utility over other types of MAC units. More specifically, the use of the shift right 16 unit 900 facilitates performing double precision operations in a reduced number of clock cycles. For example, to perform a double precision operation in which a 32 bit number (A) is multiplied with a 16 bits number (B), the low 16 bits (Al) of the 32 bit number are first multiplied with the 16 bit number B during a first clock cycle, yielding an intermediate value I that is stored in register bank 120 .
  • the intermediate value I is input into right shift 16 unit 900 and shifted right by 16 bits. Additionally, the 16 bit number B and the high 16 bits of the 32 bit number A (Ah) are multiplied, and the result added with the right shifted intermediate value I from shift right 16 unit 900 .
  • a double precision multiply is performed in two clock cycles rather than three.
  • many double precision operations require one or more variables to be shifted relative to the other variables, and as such allowing the shifting step to be performing during the same clock cycle as one of the multiply or accumulate operation reduces the number of cycles necessary to perform the double precision operation.
  • Fig. 9 is a block diagram illustrating the instruction hierarchy used in the exemplary embodiment of the invention.
  • Block 402 shows the variable length full instructions comprised of 16, 32 or 48 bits which control the operation of the DSP.
  • the variable length instructions are in turn made up of instruction fragments including general instruction fragments and memory move and program flow (MMPF) instructions as shown in block 403.
  • the general instruction fragments used in the exemplary embodiment of the invention include MAC8, MAC16, ALU8, ALU16, DMOV16, DMOV24 and DL40 instruction fragments.
  • the MMPF instruction fragments include OneMem11, TwoMem19, TwoMov19 and ThreeMem24 instruction fragments.
  • the MMPF instruction fragments are made up of the MMPF instruction subfragments shown in block 406.
  • the MMPF instruction subfragments include LD(A), LD(B), ST(A), ST(B), LS(C), DMOVA, DMOVB, and PF8.
  • the various full instructions, instruction fragments and instruction subfragments are described in greater detail below.
  • the DSP is controlled using full instructions having lengths of 16, 32 and 48 bits.
  • the full instructions are in turn formed by combining one or more instruction fragments.
  • the full instructions are configured to allow for consecutive storage within the instruction memory 152 and processing by the DSP.
  • the format and configuration of the full instructions are described below, followed by the format and configuration of the instruction fragments.
  • the DSP processes a full instruction each clock cycle. Thus, multiple operations can be performed during the processing of each full instruction, with particular operations determined by the particular set of instruction fragments selected.
  • the five (5) bit header used for each full instruction indicates the length of the full instruction, and some additional information as to the contents of the full instruction.
  • the format of the header used in the exemplary embodiment of the invention is provided in Table 4. Full instruction header formats. 5-bit Header Instruction Length 0 0 0 0 X 16-bit Instruction (2 types) 0 0 0 1 X 32-bit Instruction (2 types) 0 0 1 X X 48-bit Instruction (4 types) 0 1 X X X 32-bit Instruction (8 types) 1 X X X X 48-bit Instruction (16 types)
  • Each full instruction (16, 32 and 48 bit long) contains one or more instruction fragments.
  • Table 5 provides a list of the available instruction fragments in the exemplary embodiment of the invention. A more detailed description of the format and operation of the instruction fragments is provided after the discussion of the full instructions. Instruction fragments.
  • Tables 6 - 8 provide the various combinations of instruction fragments that may be used within 48, 32 and 16 bit full instructions in accordance with the exemplary embodiment of the invention. While other combinations of instruction fragments are consistent with the use and operation of the invention, certain features of the combination disclosed herein are preferred as discussed in greater detail below. Additionally, where all or part of full instructions are shown as "reserved,” no particular instruction combination is specified or used in the described embodiment, but future use of these full instruction combinations is contemplated.
  • Table 6 provides the formatting for 16 bit full instructions when performed in accordance with the exemplary embodiment of the invention described herein.
  • the full instruction is comprised of the five (5) bit header followed by eleven (11) instruction bits.
  • the header bits indicate the length of the instruction as well as some information about the type of instruction. For a header of 00000, the least significant three tail bits are used to further specify the operation performed. In particular, tail bits of 000 indicate the remaining eight bits contain a MAC8 instruction fragment. Tail bits of 001 indicate the remaining eight bits contain a ALU8 instruction fragment. For other tail bit combinations, no instructions are specified.
  • the remaining eleven (11) bits contain a OneMem11 instruction fragment.
  • the most common operations can be performed with the shortest full instruction. Since the shortest instruction requires the least amount of memory to store, the use of 16 bit full instruction as described reduces the amount of instruction-memory necessary to perform a particular set of operations. Thus, the overall size of the DSP, and therefore the cost and power consumption, is reduced as well.
  • the 16 bit instruction is typically used when conditions are such that only one, or a reduce number of, operations can be performed. Typically, the size of the instruction necessary to specify only one operation can be reduced, hence the use of the half-word, or 16 bit instruction for performing one operation. Additionally, the 16 bit instruction can be used for a MAC, ALU, memory move or program flow operation which encompasses almost all of the operations one would expect to perform.
  • Table 7 illustrates the instruction fragment combinations and associated formatting of a 32-bit full instruction when configured in accordance with one embodiment of the invention.
  • the five header bits indicate the length of the full instruction, as well as the particular combination of instruction fragments. For example, a header of 00010 indicates the remaining 27 instruction bits contain a ThreeMem27 instruction fragment, and a header of 00011 indicates the remaining 27 instruction bits contain a ALU8 instruction fragment followed by a TwoMem19 instruction fragment.
  • the least significant tail bits further indicate the combination of instruction fragments. For example, for a least significant tail bit of 0, the next two least significant bits indicate whether the remaining 24 bits contain a DMOV24, a ALU16 followed by a MAC8, or a MAC16 followed by a ALU8 instruction fragments.
  • Other tail bit states such as a least significant tail bit of 1, specify reserved combinations.
  • the thirty-two bit instruction allows many of the most commonly performed operations to be performed simultaneously, which facilitates pipelining while also reducing the instruction size. For example, it is common to perform two fetch operations and a multiply/accumulate operation for such applications as filtering.
  • the 32-bit instruction allows such sets of operations to be performed in pipelined fashion while not requiring a full 48 bits of instruction space.
  • the 32-bit instruction allows MAC and ALU operations to be performed simultaneously as well as program jump and call operations, also without the use of the largest instruction size.
  • Table 8 illustrates the instruction fragment combinations and format for 48 bit full instructions when performed in accordance with one embodiment of the invention.
  • header bits specify the length of the instruction as well as the particular instruction fragment combination. For example, header bits of 00100 indicate the 43 remaining instruction bits are comprised of DMOV24, MAC8 and OneMem11 instruction fragments. Header bits of 10011 indicate the 43 remaining bits are comprised of ALU16, MAC8 and TwoMem19 instruction fragments.
  • tail bits of 11111 the three least significant tail bits further indicate the instruction fragments contained in the remaining instruction bits. For example, tail bits of 000 indicate the remaining 40 instruction bits contain MAC16 and DMOV24 instruction fragments. Tail bits of 001 indicate the remaining 40 instruction bits contain MAC8, ALU8 and DMOV24 instruction fragments. Tail bits of 110 indicate the remaining 40 instruction bits contain a DL40 instruction fragment.
  • the instruction fragment combinations provided in 48 bit full instructions allow many operations to be performed simultaneously and therefore more rapidly than if performed serially. For example, several 48 bit full instructions allow ALU operations, MAC operations and memory operations to all be performed simultaneously.
  • the memory operations include load, store, and data move operations, and often allow multiple memory locations to be accessed at once.
  • the 48 bit instruction allows multiply operations to be performed in combination with ALU operations and data fetch and program flow operations, all in pipelined fashion. This can be useful for filtering when combined with scaling operations, which are often performed by performing a MAC operation followed by an ALU (such as shifting) operation. Other applications which use MAC and ALU operations include combining three or more streams of data.
  • the 48 bit instruction especially in combination with the use of the three buses architecture, facilities the pipelining of operations in these cases.
  • Varying the length of the instruction based on the number of operations that can be performed further increases the efficiency with which instruction memory is used. Any particular task has periods where multiple operations can be performed simultaneously, and other periods where fewer, or only one, operation can be performed. By adjusting the length of the instruction in accordance with the number of operations that can be performed simultaneously, the amount of instruction memory is reduced.
  • variable length instructions or tightly packed instructions, or both facilitate the use of a multiple bus architecture and a multi-access register bank, by making more circuit area available for implementing these features.
  • the combination of these aspects of the invention combine synergistically to simultaneously provide the benefits of improved performance and improved efficiency.
  • full instructions are comprised of a set of one or more instruction fragments grouped together in predefined ways.
  • the set of available instruction fragments in the exemplary embodiment of the invention is shown in Table 5.
  • the instruction fragments and the combinations made available using the full instructions provided in the exemplary embodiment of the invention are designed to allow the set of operations most likely to be performed together to be combined so that the amount of instruction memory necessary to perform a given operation is reduced.
  • the set of instruction fragments includes two types of MAC instruction fragments: MAC8 and MAC16.
  • MAC8 instruction fragments support signed-unsigned and signed-signed multiply types, with the results being stored in accumulators L0 or L1.
  • the MAC8 instruction fragment saves instruction RAM by allowing a MAC operation using a 16 bit full instruction, and for many parallel instruction combinations requiring MAC operations to be encoded into 32-bit instruction instead of 48-bit instructions.
  • the processing performed by a MAC8 instruction conforms to the following equation:
  • the MAC8 instruction fragment allows the contents of registers L0 or L1 to be summed with the product of registers R0, R2. R4 and R6 and R0, R1, R3, and R5, or set directly to the product of registers. Additionally, signed or unsigned multiplies may be specified.
  • the length of the instruction can be kept to 8 bits, allowing the MAC operations to be performed using a shorter, 8 bit, instruction fragment.
  • the particular operation performed by the MAC8 instruction is specified by the values of the eight bits that make up the instruction as set forth in Table 11.
  • MAC8 instruction fragment format 7 6 5 4 3 2 1 0 MAC Operation mac8Op1 mac8Op2 SU/SS
  • SU/SS specifies signed or unsigned multiply.
  • the codes for specifying various operations within the MAC8 instruction fragment are listed in Table 12 .
  • a MAC8 instruction of 0x99 places the sum the contents of register L0 with the unsigned product of registers R0 and R3 into register L0.
  • Equation (3) sets forth the operations that can be performed using the MAC16 instruction fragment.
  • every accumulator (L0 - L3) can be used as a destination, although not all combinations of accumulators are allowed in multiply-accumulate instructions.
  • the CPS field signals that a coprocessor should perform a particular operation in parallel.
  • the particular operation performed by the MAC16 instruction is specified by the values of the sixteen bits that make up the instruction are set forth in Table 13.
  • MAC16 instruction fragment format 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 0 MAC Operation macOp1 macOp2 mtype mshift CP S
  • the MAC16 instruction fragment allows left shifts of up to 3, and can perform round operations during straight multiplies (no accumulate), with the round occurring after the shift.
  • the accumulator to be added can be shifted down by 16 in parallel with a signed-signed multiply.
  • the CPS bit is the coprocessor strobe bit for indicating that data used in the MAC operation should be sent to a coprocessor.
  • the MAC8 instruction fragment performs a subset of the operations that can be performed by the MAC16.
  • the particular set of instructions selected for the MAC8 instruction fragment are the most commonly performed out of the set of operations that can be performed using the MAC16 instruction fragment. This saves program memory by allowing the majority of MAC operations to be performed using the MAC8 instruction fragment.
  • the 8 bit ALU8 instruction fragment comprises the ALU operations that are most commonly paralleled with MAC operations (MAC8 and MAC16) and which do no contain immediate data. All ALU8 shift operations are arithmetic shifts that use an internal shift register (SR) register to save instruction encoding bits.
  • SR shift register
  • the operations performed using the ALU8 instruction fragment are shown in Table 15.
  • ALU8 instruction fragment operations NOP ; NOP (needed for parallel combinations).
  • LD DETNORM ( LS ); Determine block normalization factor.
  • LD SET ( LS ); Copy accumulator (no saturation).
  • LD LS ⁇ SR ; Shift accumulator.
  • LD RND ( LS ⁇ SR ); Shift and round accumulator.
  • LD LD ⁇ ( LS ⁇ SR ); Accumulate shifted accumulator.
  • LD LS ⁇ LT; Add or subtract accumulators.
  • LS ⁇ LT Add/subtract accs result free (set flags).
  • LS is load source (L0 - L3) and LD is load destination (L0 - L3).
  • ALU8 instruction fragment The particular operations performed by the ALU8 instruction fragment are specified by the values of the eight bits that make up the instruction fragment as set forth in Table 16 .
  • ALU8 instruction fragment format 7 6 5 4 3 2 1 0 0 ALUOp LS LD 0 1 1 Sign LS LT 1 LD Sign LS LT
  • the ALU16 instruction fragment allows both arithmetic and logical shifts.
  • the particular operations performed by the ALU16 instruction fragment are set forth in Table 18 .
  • the format of the ALU16 instruction fragment is set forth in Table 19 .
  • the particular operations performed by the ALU16 instruction fragment are specified by the values of the bits that make up the instruction fragment as set forth in Table 20 .
  • the DMOV16 instruction fragment is a 16 bit instruction fragment for performing different data move, data inport and data outport operations as set forth in Table 21 .
  • the instruction OUTPORTA(port_addr) reads the value on Abus and outputs it to the designated port. By reading a value from memory A simultaneously, this instruction can be used to send a value directly from memory A to the port.
  • OUTPORTB(port_addr) operates similarly.
  • the DMOV24 instruction fragment is a 24 bit instruction fragment for performing different load/store register direct or load register immediate operations as set forth in Table 23 .
  • Table 24 provides the format and some codes used to perform the various operation available using the DMOV24 instruction fragment in accordance with the exemplary embodiment of the invention.
  • the 40-bit dual load instruction fragment (DL40) is a 40 bit instruction fragment for performing immediate load or address load operations.
  • the particular operations performed in the exemplary embodiment of the invention are as shown in Table 25 .
  • the format of the DL40 instruction fragment for each operation is provided in Table 26 .
  • Each memory move and program flow instruction (MMPF) fragment is comprised of a set of MMPF subfragments listed in Table 28.
  • the OneMem11 MMPF instruction fragment is used to perform single memory load and store operations, data move operations, and program flow operations. In the exemplary embodiment provided herein eight different operations are performed using the OneMem11 MMPF instruction fragment, with the particular operation indicated by the first three bits of eleven bit fragment as shown in Table 29, which lists the operations that can be performed using a OneMem11 data move instruction fragment.
  • OneMem11 instruction fragment format 10 9 8 7 6 5 4 3 2 1 0 0 0 0 0 LD(A) 0 0 1 ST(A) 0 1 0 LD(B) 0 1 1 ST(B) 1 0 0 LS(C) 1 0 1 DMOVA 1 1 0 DMOVB 1 1 1 PF8
  • TwoMem19 MMPF instruction fragment is a 19 bit instruction fragment that allows eight different combinations of memory load and store operations to be performed as set forth in Table 30 .
  • TwoMem19 instruction fragment format 18 17 16 15-8 7-0 0 0 0 LD(A) LD(B) 0 0 1 LD(A) ST(B) 0 1 0 LD(A) LS(C) 0 1 1 ST(A) LD(B) 1 0 0 ST(A) ST(B) 1 0 1 ST(A) LS(C) 1 1 0 LS(C) LD(B) 1 1 1 LS(C) ST(B)
  • the TwoMov19 MMPF instruction fragment is a 19 bit instruction fragment that allows eight different combinations of memory load and store operations along with data move operations as shown in Table 31 .
  • TwoMov19 instruction fragment format 18 17 16 15-8 7-0 0 0 0 LD(A) DMOVB 0 0 1 ST(A) DMOVB 0 1 0 DMOVA LD(B) 0 1 1 DMOVA ST(B) 1 0 0 DMOVA LS(C) 1 0 1 LS(C) DMOVB 1 1 0 DMOVA DMOVB 1 1 1 1 1 Reserved
  • the ThreeMem27 MMPF instruction fragment is a 27 bit instruction fragment that allows eight different combinations of memory load, memory store, and data operations to be performed as shown in Table 32 .
  • Equation (4) provides the operations performed by the LD(A) instruction subfragment.
  • Table 33 provides the format the LD(A) instruction subfragment in accordance with the exemplary embodiment of the invention.
  • LD(A) instruction subfragment format 7 6 5 4 3 2 1 0 dreg A0-A3 Amod
  • Equation (5) provides the operations performed by the LD(B) instruction subfragment.
  • Table 34 provides the format the LD(B) instruction subfragment in accordance with the exemplary embodiment of the invention.
  • LD(B) instruction subfragment format 7 6 5 4 3 2 1 0 dreg B0-B3 Bmod
  • Equation (6) provides the operations performed by the ST(A) instruction subfragment.
  • Table 35 provides the format the ST(A) instruction subfragment in accordance with the exemplary embodiment of the invention.
  • ST(A) instruction subfragment format 7 6 5 4 3 2 1 0 dreg A0-A3 Amod
  • Equation (7) provides the operations performed by the ST(B) instruction subfragment.
  • Table 36 provides the format the ST(B) instruction subfragment in accordance with the exemplary embodiment of the invention.
  • ST(A) instruction subfragment format 7 6 5 4 3 2 1 0 dreg B0-B3 Bmod
  • Table 37 lists the operations performed by the DMOVA instruction subfragment.
  • Table 38 provides the format the DMOVA instruction subfragment in accordance with the exemplary embodiment of the invention.
  • MMPF instruction fragments that can contain one or more instruction subfragments
  • the number of operations that can be performed using a full instruction is further enhanced.
  • a full instruction can cause arithmetic and MAC operations to be performed along with a set of up to three memory move and program flow operations.
  • the ability to perform this many operations using a single instruction further reduces the total number of instructions necessary to perform a given operation and therefore decreases the total instruction memory required on the DSP. Reducing the instruction memory decreases the die size and therefore the cost and power consumption of the DSP, which makes that DSP more suitable for a wide variety of applications include mobile wireless telephony.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Executing Machine-Instructions (AREA)
  • Advance Control (AREA)
  • Memory System (AREA)

Description

    BACKGROUND OF THE INVENTION I. Field of the Invention
  • The present invention relates to a digital signal processor. The present invention has particular, but not exclusive, application to digital signal processing using highly parallel, highly pipelined, processing techniques.
  • II. Description of the Related Art.
  • Digital Signal Processors (DSPs) are generally used for real time processing of digital signals. A digital signal is typically a series of numbers, or digital values, used to represent a corresponding analog signal. DSPs are used in a wide variety of applications including audio systems such as compact disk players, and wireless communication systems such as cellular telephones.
  • A DSP is often considered to be a specialized form of microprocessor. Like a microprocessor, a DSP is typically implemented on a silicon based semiconductor integrated circuit. Additionally, as with microprocessors, the computing power of DSPs is enhanced by using reduced instruction set (RISC) computing techniques. RISC computing techniques include using smaller numbers of like sized instructions to control the operation of the DSP, where each instruction is executed in the same amount of time. The use of RISC computing techniques increases the rate at which instruction are performed, or the clock rate, as well as the amount of instruction pipelining within the DSP. This increases the overall computing power of the DSP.
  • Configuring a DSP using RISC computing techniques also creates undesirable characteristics. In particular, RISC based DSPs execute a greater number of instructions to perform a given task. Executing additional instructions increases the power consumption of the DSP, even though the time to execute those instructions decreases due to the improved clocking speed of a RISC based DSP. Additionally, using a greater number of instructions increases the size of the on-chip instruction memory within the DSP. Memory structures require substantial (often more than 50% of the total) circuit area within a DSP, which increases the size and cost of the DSP. Thus, the use of RISC based DSPs is less than ideal for low cost, low power, applications such as digital cellular telephony or other types of battery operation wireless communication systems.
  • Fig. 1 is a highly simplified block diagram of a digital signal processor configured in accordance with the prior art. Arithmetic logic unit (ALU) 16 is coupled to ALU register bank 17 and multiply accumulate (MAC) circuit 26 is coupled to MAC register bank 27. Data bus 20 couples MAC register bank 27, ALU register 17 and (on chip) data memory 10. Instruction bus 22 couples MAC register bank 27, (on-chip) instruction memory 12, MAC register bank 27 and ALU register bank 17. Instruction decode 18 is coupled to MAC 26 and ALU 16, and in some prior art systems instruction decode 18 is coupled directly to instruction memory 12. Data memory 10 is also coupled to data interface 11 and instruction memory 12 is also coupled to instruction interface 13. Data interface 11 and instruction interface 13 exchange data and instructions with off-chip memory 6.
  • During operation, the instructions in instruction memory 12 are decoded by instruction decode 18. In response, instruction decode 18 generates internal control signals that are applied to ALU 16 and MAC 26. The control signals typically cause ALU 16 to have data exchanged between ALU register bank 17 and data memory 10 or instruction memory 12. Also, the control signals cause MAC 26 to have instruction data exchanged between MAC register bank 27 and instruction memory 12 or data memory 10. Additionally, the control signals cause ALU 16 and MAC 26 to perform various operations in response to, and on, the data stored in ALU register bank 17 and MAC register bank 27 respectively.
  • In an exemplary operation, instruction memory 12 may contain coefficient data for use by ALU 16 and MAC 26 and data memory 10 may contain data to be processed (signal data). The coefficient data may be for implementing a frequency filter using the DSP, which is a common practice. As the filtering is performed, both the signal data from data memory 10 and the coefficient data from instruction memory 12 are read into MAC register 27. Additional instruction data within instruction memory 12 is also applied to instruction decode 18, either through instruction data bus 22 or through a direct connection. The additional instruction data specifies the operation to be performed by MAC 26. The results generated by MAC 26 are typically read back into data memory 10.
  • Many processing inefficiencies result from this prior art processing. These processing inefficiencies include, e.g., bus, or access contention, to instruction memory 12, which must supply instruction data to both MAC register 26 and instruction decode 18, as well as bus, or access contention, to data memory 10, which must both read out signal data and write in the output data. Additionally, in many instances, additional processing on the output data must be performed by ALU 16. This further aggravates access to data memory 10, and therefore creates contention for data bus 20, because the output data must be written from MAC register bank 27 into data memory 10, and then read out to ALU register 17. These read and write operation are performed over bus 20 and therefore consume additional bus cycles. Such inefficiencies reduce the processing performance of the DSP.
  • US-A-5,710,914 describes a digital signal processing method and system implementing pipelined read and write operations. The system includes a program control unit, a memory, a data processing unit, and a dedicated bus for writing processed data from the data processing unit to the memory. By using the dedicated write bus, the system goes some way to avoiding bus contention in a five stage pipeline operation involving fetch, decode, read, execute and write operations.
  • US-A-5,293,611 describes a digital signal processor utilising a multiply-and-add function for digital filter realisation. The processor has a bypass signal path for transferring data without the intervention of any bus which is extended between the two ports of the data RAM. During the course of multiply-and-add processing, sampling data items to be read out of a predetermined address of the data RAM can be transmitted to an internal bus and can simultaneously be written into the next address of the data RAM.
  • Foley, P in "The MPACT media processor redefines the multimedia PC", Digest of Papers of Compcon (Computer Society Conference) 1996, Technologies for the Information Superhighway Santa Clara, Feb 25-28, 1996, CONF. 41, 311-318) describes multimedia processing using a single media processor, a high bandwidth RAMBUS RDRAM for media memory and a VLIW SIMD internal architecture.
  • US-A-5 522 083 discloses a multiprocessor system wherein each processor comprises a register bank and has access to a first, second and third memory bank via a global bus and via a local bus. Global bus and local bus can be used simultaneously to provide data from the first and second memory banks to the processor and one of said buses can be used to write the result back to the third memory bank.
  • SUMMARY OF THE INVENTION
  • The present invention seeks to improve the performance and usefulness of a DSP by addressing the problems and inefficiencies listed above, as well as by providing other features and improvements described throughout the application.
  • The present invention aims to provide, a novel and improved method and circuit for digital signal processing.
  • According to a first aspect of the present invention, there is provided a digital signal processor for processing signals in response to instruction data as set forth in claim 1.
  • Acccording to a second aspect, there is provided a method of processing data as set forth in claim 17.
  • Various embodiments of the invention are set out in the dependent claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The features, objects, and advantages of the present invention will become more apparent from the detailed description of exemplary embodiments of the invention set forth below when taken in conjunction with the drawings in which like reference characters identify correspondingly throughout and wherein:
  • FIG. 1 is a block diagram of a digital signal processor configured in accordance with the prior art;
  • FIG. 2 is a block diagram of a digital signal processor embodying the present invention;
  • FIG. 3 is a block diagram of the connections between the input ports and the registers of the register bank;
  • FIG. 4 is a block diagram of the connection between the register and the output ports of the register bank;
  • FIG. 5 is a diagram of a set of variable length instructions stored in memory space in accordance with an embodiment of the invention;
  • FIG. 6 is a flow chart illustrating the operation of the instruction fetch unit;
  • FIG. 7 is a block diagram of the instruction fetch unit when configured in accordance with an embodiment of the invention;
  • FIG. 8 is a block diagram of the MAC unit when configured in accordance with an embodiment of the invention;
  • FIG. 9 is a block diagram of the instruction hierarchy used in an embodiment of the invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The present invention provides a novel and improved method and circuit for digital signal processing. Throughout the application various references are made to signals, instructions, and data. These signals, instructions, and data are preferably represented by electrical voltages, currents, current stores including charged, optical, or magnetic particles, or some combination thereof, the use of which is well known. The use of various chemical and biological compounds to represent such signals, instructions, and data is generally consistent with the use of the present invention as well, although not preferred due to the difficulty to use, control and manipulate such items.
  • Additionally, reference is made to various aspects, benefits, features, or advantages of the invention (referred to herein collectively as aspects, when not referred to in the particular.) In some embodiments of the invention, these different aspects may be realized alone, without the presence of any of the other aspects of the invention. However, in other embodiments of the invention, two or more aspects of the invention may be realized together, to produce synergistic and unexpected advantages that are greater than those provided by embodiments of the invention that realize only one aspect of the two or more combined aspects of the invention.
  • I. DSP OPERATION AND INSTRUCTION STORAGE
  • Fig. 2 is a block diagram of a portion of a digital signal processor (DSP) circuit configured in accordance with an exemplary embodiment of the invention. Data memories 102 - 104 are coupled to data buses A, B and C respectively via address generation units (AGU) 105 - 107, and to data interface 100. Data buses A, B and C are coupled to output ports PO1, PO2 and PO3 respectively of register bank 120, and to input ports PI1, PI2, and PI3 respectively of register bank 120 through multiplexers 122 - 126. Preferably, data buses A, B and C read and write data between data memories 102 - 104 and the registers within register bank 120.
  • The use of three data buses and three data memories allows more data to be exchanged between the register banks and the data memories without creating bus contention. For example, three fetch operations can be performed simultaneously from the three memories 102 - 104 using the three data buses A, B and C. Similarly, three write operations can be performed simultaneously, as can any combination of three fetch and write operations.
  • The addition of a fourth data bus would allow even greater numbers of operations to be performed, and is consistent with some embodiments of the invention. However, the use of only three data buses has particular advantages, because three buses facilitates performing many tasks commonly performed by a DSP such as filtering. Thus, the addition of a fourth data bus does not provide the same incremental performance improvement as the addition of a third data bus, and requires the same amount of additional circuit area. Therefore, the addition of a fourth data bus provides incrementally less benefit than the addition of a third bus. So, in many embodiments of the invention the use of only three data buses is preferred.
  • Output ports PO4, PO5 and PO6 of register bank 120 are coupled to multiply accumulate (MAC) unit 128, the output of which is in turn coupled to input port PI4 of register bank 120. Output ports PO7 and PO8 of register bank 120 are coupled to arithmetic logic unit (ALU) 130, the output of which is coupled to input port PI5 of register bank 120.
  • Instruction memory 152 is coupled to instruction fetch unit 156 and instruction interface 150. Instruction decoder 158 is coupled to instruction fetch unit 156, and to immediate bus Im1, immediate bus Im2 and immediate bus Im3, as well as immediate bus ImALU. Immediate buses Im1, Im2 and Im3 are coupled to multiplexers 122, 124 and 126. Immediate bus ImALU is coupled to ALU 130. In addition to the data couplings described above, decoder 158 is coupled to the various subsystems shown by control connections (not shown for ease of drawing).
  • Register bank 120 contains eight (8) registers labeled L0-L3 and D0-D3. Registers L0-L3 are forty (40) bit wide registers and can also be accessed in sixteen bit fragments via high word registers L0h - L3h and low word registers L01 - L31. Registers D0 - D3 are thirty-two (32) bits wide, and can be accessed in sixteen bit fragments via subregisters R0-R7. In general, the registers and subregisters are referred to as simply "registers", with the particular nature of the registers made apparent by the particular register number provided.
  • One aspect of the invention is realized by having some registers coupled to, and therefore accessible by, multiple input and output ports. In one embodiment, this multiconnectedness is provided by the use of multiplexers coupled to the input of each register, and to each output port. Other methods of providing multiconnectedness will be apparent, and are consistent with the use of some aspects of the present invention, including, e.g., using data buses and addressable memories. However, the use of multiplexers is preferred in some embodiments because they provide rapid and controllable access to the various registers and ports.
  • Other aspects of the invention are realized in embodiments of the invention that use immediate data buses, which is the case in the exemplary embodiment provided herein. For example, data contained in the instruction data can be read into register bank 120 without the need to interface with memories 102 - 104. Thus, the additional data can be provided from the instruction processing system without interfacing with the data memories, further reducing bus contention.
  • Fig. 3 is a block diagram illustrating the connections between the set of registers within register bank 120 and the set of input ports PI1 - PI5. The registers are defined as L0h - L3h, L01 - L31, and R0 - R7. A register L0 is comprised of registers L0h and L01. In the context of FIGS. 3 and 4, the registers L0h - L3h are 24 bits and registers and L01 - L31 and R0 - R7 are 16 bits, making registers L0 - L3 40 bits wide. Similarly, input ports PI3 - PI5 are comprised of input ports PI3h - PI5h of 24 bits and PI31 - PI51 of 16 bits for a total of 40 bits. The input ports PI1 and PI2 are only 16 bits, and when used to write to registers L0h - L3h, write only to the least significant 16 of the 24 bits available.
  • As shown in Fig. 3, some registers receive data from all the input ports, while other registers receive data from only some, or a portion, of the input ports. In particular, all of the registers L0 - L3 receive data from all the input ports PI1 - P15, from multiplexers 500 - 514, with the 16 bit input ports being able to write to both the higher and lower registers within registers L0 - L3. Thus, register L0 - L3 receive input from any bus A - C (corresponding to input ports PI0 - PI3) and from MAC unit 128 and ALU 130 (corresponding to input ports PI4 and PIS.) Registers R0 - R7 receive input data from and bus A-C via multiplexers 516 - 530. However, none of registers R0 - R7 receive input data from MAC unit 128 (input port PI4). Additionally, registers R0 - R3 receive input data from ALU unit 130 via multiplexers 516, 518, 524 and 526.
  • The embodiment shown in Fig. 3 has various advantages. In particular, it provides sufficient connectivity between the input ports and registers to facilitate the most common operations, but the total connectivity is kept at a minimum to reduce the total circuit area required for implementing the circuit. For example, the output of MAC unit 128 is coupled only to long registers L0 - L3. This is beneficial because the result of multiply and accumulate operations generally exceed 32 bits, so coupling the output of MAC unit 128 to registers D0 - D3 provides minimal benefit. In another example, ALU unit 130 can output to registers L0 - L3 and R0 - R3. This enhances flexibility as data from ALU unit 130 may be written to a variety of registers, which is useful since ALU unit 13 performs a greater variety of operations making it useful to output data to a greater number of registers. However, ALU unit 130 is not coupled to all the registers, and therefore unnecessary and excessive connectivity is avoided.
  • Fig. 4 is a block diagram illustrating the coupling of the output ports of register bank 120 to the registers when performed in accordance with one embodiment of the invention. As shown, output port PO1, which outputs to BUS A, is coupled via multiplexer 540 to registers L0h - L3h, L0l - L3l and R0 - R7, which comprises all the available registers when accessed as subregisters. Similarly, output port PO2, which outputs to BUS B, is coupled via multiplexer 542 to registers L0h - L3h, L01 - L3l and R0 - R7. Output port PO3, which outputs to 40 bit wide BUS C, is coupled by multiplexer 530 to registers L0 - L3 and D0 - D3, which comprises all of the available registers when accessed as full registers.
  • Output port PO4, which is coupled to the 40-bit input of MAC unit 128, is coupled via multiplexer 532 to registers L0 - L3. Since the values accumulated by MAC 128 tend to be large due to the nature of the multiply and accumulate operations performed, coupling output port PO4 to only the 40 bit "long" registers L0 - L3 provides an optimal coupling arrangement, because the utility derived from providing additional couplings to registers D0 - D3 is low, given that the accumulation of various multiply operations will typically exceed 32 bits.
  • Output port PO5, which is coupled to one 16 bit input of MAC unit 128, is coupled by multiplexer 534 to registers L0h - L3h, R0, R2, R4 and R6. Output port PO6, which is coupled to the second 16 bit input port of MAC unit 128, is coupled to registers L0h - L3h, L01 - L31 and R0 - R7. By coupling one sixteen bit input of MAC unit 128 to all the available registers, while coupling the second 16 bit input port to a subset of the available registers, a useful compromise is achieved. In particular, when register space becomes limited, at least one piece of data to be processed can be placed in any available register. However, by limiting the number of registers connected to the other input, the total amount of connection circuitry is reduced, which facilitates providing other functions and features, such as higher connectivity among the other registers, input ports, and output ports.
  • Output port PO7, which is coupled to an input of ALU 130, is coupled by multiplexer 546 to registers L0 - L3, L0h - L3h and R0 - R3, where L0h - L03h and R0 - R3 are output in conjunction with a set of logic zeros. That is, registers L0h - L03h and R0 - R3 are output to the bits 31-16 (bits numbered 0 - 39) bits of PO7, with bits 0-15 set to logic zero and bits 39 - 32 are sign extended using bit 31. Output port PO8, which is coupled to another input of ALU 130, is also coupled by multiplexer 548 to registers L0 - L3, and registers R0 - R7 in conjunction with a set of logic zeros. Coupling the inputs of ALU 130 in this manner allows logic operations to be performed on all the available long registers L0 - L3, and therefore on large numbers, which is useful for many types of signal processing operations such as normalizing and scaling. Additionally, arithmetic operations can be performed between registers R0 - R7 and L0h - L3h and R0 - R7, which provides a high level of flexibility in terms of the set of registers that can be used, while also limiting the number of necessary connections, and therefore the required circuit area. It should be understood that the logical and arithmetic operations available are not limited to that described above.
  • The use of registers that are accessible by both the multiple data buses and the multiple processing units provides various advantages. For example, the registers provide an interface between the data buses and processing units, reducing the need to route each data bus to each processing unit. Reducing data bus routing saves circuit area and reduces chip cost.
  • Furthermore, coupling at least some (a set) of registers to multiple processing units allows multiple operations to be performed on the same data using the multiple processing units, without having to read and write the data over the data bus and to the memory. This saves bus cycles and therefore reduces bus contention. Instruction processing pipelining is also facilitated because data processed by a first processing unit during a first instruction cycle may then be processed further by a second processing unit during a second processing cycle within the same register.
  • However, typically not all data will require processing by multiple processing units, so other registers (other set) are accessible by only one processing unit, or by fewer than the total number of processing units where more than two processing units are present. The use of this other set of registers reduces the number of connections, and therefore circuit area, and thus an optimal balance between register connectivity and circuit area (and therefore between performance and efficiency) is provided.
  • Furthermore, in one embodiment of the invention, pipelining is further enhanced by the use of two-phase clocked registers within register bank 120. The two-phase clocked registers are read on a first phase of the clock, and then written to on the second phase of the clock, within the same full clock ("processing") cycle. Thus during a particular processing cycle, data already processed by a first processing unit, such as the MAC 128, can be read out during the first clock phase and further processed by a second processing unit, such at ALU 130, within the remaining portion of the processing cycle.
  • Additionally, during the second phase of the processing cycle, new data just processed by the MAC 128 is written into the same register, thus allowing complete pipeline processing between two processing units during one processing cycle. Once again, these operations are performed without running the data over any of the internal busses, and therefore increased bus contention is avoided.
  • Referring again to FIG. 2, during operation, instruction fetch unit 156 retrieves binary instructions from instruction memory 152, or if not available in instruction memory 152, from a memory located externally. The external memory can take many forms well known in the art such as dynamic and static random access memory (DRAM and SRAM) or some derivative thereof, magnetic or optical hard disk memory or some other data storage medium well known in the art. In the exemplary embodiment of the invention, the instructions are of variable length and instruction fetch unit determines the length of the instruction and how much additional instruction data to fetch during each processing or clock cycle. Additionally, the instructions are stored in consecutive memory locations within the internal memory and external memory. The operation of fetch unit 156 and the storage of instruction data within memory and external memory are described in greater detail below.
  • Instruction decoder 158 receives the instructions retrieved by instruction fetch unit 156 and translates the instructions into control signals that are applied to one or more of the subsystems that make up the DSP including the data memories, register bank, MAC and ALU. Additionally, instruction decoder 158 may route immediate data that is contained within the receive instructions to the appropriate system via immediate buses Im1, Im2, Im3 or ImALU. Immediate data is typically numeric values stored within the instruction data that are used to perform operations on the data stored in data memories 102 - 104 or which may specify or modify an address.
  • Operations performed by the DSP of Fig. 2 include loading of data from a data memory into a register location via one of the data buses. Data may also be written from a register into a data memory. Also, MAC 128 or ALU 130 may perform operations on the data stored in one or more of the registers within register bank 120, where the results are typically written back into a register within register bank 120.
  • The DSP architecture described above provides numerous advantages. For example, the use of three data buses facilitates the uninterrupted pipeline processing of data. During an exemplary filtering performed by the DSP, the data to be filtered (signal data) is stored in one data memory and the coefficients to be applied to that data are stored in the other data memory. Storing the signal data and coefficient data in the two narrower memories is preferred, as the results of the operations typically require more bits than the operands. The coefficient and signal data are then read into register bank 120 and then multiplied and accumulated by MAC unit 128. The results of these operations may be stored in a second register within register bank 120, or overwritten into the register bank in which the input data was previously stored. Any results are then typically written from the register over the third bus (BUS C) into the wider (memory C) data memory.
  • Because the output data is written into a third memory via a third bus, and the input data sets are read from first and second data memories via first and second data buses, little or no memory access conflict or bus contention occurs. Thus, the processing of data may proceed uninterrupted, reducing the need to clock any of the memory subsystems or data buses at higher rates than the internal buses or other subsystems. This reduces power consumption while maintaining, or increasing, processing speed.
  • Additionally, running the data through the DSP without interruption facilitates pipelined processing of the data where a number of different data values are processed differently at different stages within the DSP at any given time. Also, when combined with the use of parallel instructions as described below, significant processing flexibility can be achieved in combination with this efficient highly pipelined processing, thus providing a highly versatile, efficient and powerful DSP system.
  • It should be understood that the use of multiple buses increases the ability to move data around the DSP in a variety of additional ways that reduce bus conflict. For example, data to be divided may be supplied as input data from memory C via data bus C, with the divisor supplied by another memory and bus, such as memory A and data bus A. The result can then be stored in the remaining memory (memory B) by way of the remaining bus (data bus B).
  • In another exemplary operation facilitated by providing multiple data buses and memories, data to be accumulated by MAC unit 130 is provided via a first memory and first bus (for example memory A and data bus A). After a set of accumulations is performed, the resulting data may be written to memory C via data bus C. Simultaneously, data to be logically shifted is provided from memory C via data bus C to ALU unit 130 during the processing cycles where data bus C is not carrying result data from MAC unit 128, which is most of the time such result data will only be available after a set of accumulate operations are performed. The logically shifted data is simultaneously written to memory B via data bus B. Thus, the use of multiple data buses and memories in general, and particularly in conjunction with multiple processing units, facilitates performing multiple operations by providing more possibilities for moving data within the DSP.
  • As noted above, another aspect of the invention is realized by the use of registers that are accessible by the multiple processing units, e.g., MAC unit 128 and ALU unit 130. Registers accessible by multiple processing units allow data that must be processed by the processing units to be accessed without moving the data across any of the internal data buses. For example, data can be written to a register by a first processing unit, and then further processed by the second processing unit which may also access that register. This further relieves bus contention and congestion, and therefore maintains high data throughput.
  • Additionally, when combined with the use of parallel operation instructions, and parallel processing capability, as described in greater detail below, the ability to perform highly pipelined, multi-operation, processing of data is further enhanced. In contrast, typical pipelining involves staggering the different phases (i.e. fetch, decode, process) of a set of operations so that the processing time between the start of each instruction is reduced. Multi-operation pipelining provides the additional benefit of having data pass through a set of different operations, where those operations are performed simultaneously on different data sets. This multi-operation pipelining increases the number of instructions performed per processing cycle beyond the conventional instruction pipelining.
  • The synergy of the above-described architecture is further illustrated by the following exemplary processing. In an exemplary filtering operation (as also described above), results generated by MAC 26 increases in size (both in terms of absolute value and the number of bits used to represent the value) as the number of products over which accumulation is performed grows. Eventually, the result will have to be scaled, or "normalized", which typically calls for a logical shift operation by ALU unit 130.
  • In the described system, the scaling operation can be performed simultaneously with the multiply and accumulate operations of the filtering. During a processing cycle in which such simultaneous processing is performed, unprocessed signal data and filtering coefficients are read from data memories 102 and 103 into registers within register bank 120 (for example L0h and L01). Simultaneously, MAC unit 128 reads the values previously stored in those registers (L0h and L01) and performs a multiply and accumulate operation where the output is written to a second register (for example L1). Also simultaneously, ALU unit 130 reads the data previously stored in the second register (L1) and performs the scale operation writing the scaled value to a third register (for example L2). Also during the same processing cycle, the previous value stored in the third register (D0) is written into data memory 104 using bus C 112. As should be apparent, the particular operations may vary in accordance with the particular task being performed. As should also be apparent, the use of highly parallel instructions that permit multiple operations to be performed further facilitates highly pipelined multi-instruction operations. The highly parallel instructions allow specification of the different operations to be pipelined during each processing cycle.
  • Using the two-phase read-write operations, an example of which is provided above, this processing can all be done during a single processing cycle, where the data is read out from each register during the first clock phase, processed by the processing units, and the results written over the old data into the register during the second clock phase. It should be understood, that the same value is not subjected to all the steps in this process during a single clock cycle, but rather a set of values are pipelined through the DSP, each moving to the next step as the processing is performed.
  • Many other operations are facilitated by the various aspects of the invention described herein. For example, vocoding is the process of coding voice data. Vocoding requires many different types of operations to be performed, some of which can be performed independently, and therefore simultaneously. The use of multiple data buses and multiple processing units facilitates performing these operations.
  • Further advantages are provided from the use of a separate instruction memory and instruction decode. For example, simultaneous with the data processing described above, instructions are read from instruction memory 152 by instruction fetch 156, which in turn causes instruction decoder 158 to generate control signals to control the operation of the various other subsystems within the DSP (connections not shown for clarity and ease of drawing). Once again, the data buses do not have to carry instruction data, and therefore the signal data may be moved and processed without interruption from the instruction data. Thus, performance is further enhanced by separating the instruction processing from the data processing, which eliminates the need to consume data bus cycles for instruction data movement.
  • Fig 5. is a table illustrating the packing of a set of variable length instructions within a portion of the addressable memory space of instruction memory 152 of Fig. 2 in accordance with one embodiment of the invention. In some embodiments of the invention the variable length instructions may also be stored as shown in Fig. 2 within the external memory system to realize additional memory efficiencies. Example addresses are shown in the left column, with each address pointing to a 32-bit data word, shown in the middle and right columns of memory 275. The middle column represents a 16-bit high order subword and the right most column represents a sixteen-bit low order subword of each data word. The high and low order subwords are not individually addressable in the preferred embodiment of the invention in order to reduce the amount of necessary address logic.
  • Within memory 275, variable length instructions A-L are stored in the packed configuration shown. Instruction A is a 48 bit instruction with the first two double-bytes A(1) and A(2) stored in address word 0x0000 and the third double-byte A(3) stored in the high order subword of address 0x0001. Instruction B, which follows instruction A, is a 32 bit instruction with the first double-byte B(1) stored in the low order word of address 0x0001 and the second double-byte B(2) stored in the high order subword of address 0x0002. Instruction C is a 16-bit instruction with the first and only double-byte C(1) stored in the low order subword of address 0x0002.
  • As should be apparent from the storage location of instructions A - C, this reduces the size or amount of memory 275 necessary to store a set of instructions by storing portions of different instructions within the same address word. For example, the third double-byte A(3) of instruction A is stored along with the first double-byte B(1) of instruction B.
  • By storing variable length instructions across word boundaries, or more particularly in consecutive locations within the memory address space, the amount of instruction memory required to store a given number of instructions is reduced. Reducing the amount of instruction memory reduces the size and cost of the die necessary to give the DSP a given amount of instruction caching capability. The packing of instructions is further illustrated by placement of variable length instructions D - L within memory 275 as shown in Fig. 3.
  • It should be understood that packing all the instructions in consecutive locations, as described above, is not necessary in some embodiments of the invention. For example, different embodiments of the invention pack only a substantial portion (e.g. 90% or more) of the instructions in consecutive locations within memory space. In other embodiment of the invention, only a significant portion (e.g. between 25 to 50%) of the instructions is advantageously packed in consecutive memory space. Other embodiments of the invention may use still other percentages of packed instructions.
  • Also, the use of consecutive locations is not necessary. The instructions must simply be placed in a total memory space that is not substantially larger than the total amount of instruction data. This is preferably achieved by placing the instructions in adjacent locations within memory space, however, the instructions could be shuffled throughout memory space, so long as the instruction could be read out in intended order of execution. Those skilled in the art will recognize that this type of predetermined shuffling as a remapping of memory space, and that such remapping typically does not affect the operation of the invention, other than by adding generally undesirable complexity.
  • Similarly, it is preferable to have the packing scheme employed over a large set of instructions. For example, employing the packing scheme over at least ten instructions is preferred in some embodiments of the invention.
  • Additionally, the particular packing scheme used in the exemplary embodiment of the invention is not necessary in some other embodiments of the invention. For example, some other embodiments of the invention may not have instructions in consecutive memory locations. Rather, instructions may separated by some small amount of memory space, including the use of an instruction separator code. Preferably, the small amount of memory space is less than the amount of memory space necessary to keep instruction boundaries on memory word boundaries. The above-described packing is preferred in many instances, however, due to its simplicity, completeness and efficiency. In general, the chosen tradeoff between the completeness of packing and the complexity of the packing scheme can differ in different embodiments of the invention.
  • Also, as implied above, some embodiments of the invention employ the packing scheme on only portions of the instructions, but not the entire set of available instructions. For example, instruction packing may be performed only on sets of instructions which are used to perform a particular task or subroutine.
  • It should be noted that combining highly packed instruction storage with the use of variable length instructions further reduces the memory requirements of the DSP, because variable length instructions only consume the amount of data necessary to request the desired operations, and highly packed instruction storage keeps the total memory equal to that consumed by the set of variable length instructions, and therefore at a minimum. Reduced memory size reduces chip size and cost.
  • In addition to the benefit of reduced DSP size, and therefore reduced DSP cost, the use of highly packed instructions and variable length instructions provides additional unexpected advantages when combined with other features of the architecture described above. For example, by reducing the size of instruction memory, additional circuit area is made available for the use of three data buses within the DSP, which provides the benefits as described above, including uninterrupted, highly pipelined, data processing, and the ability to perform multiple operations simultaneously within the DSP. Thus, tightly packed instructions combines with the multibus architecture to provide the additional, unexpected, benefits of increased performance and efficiency.
  • The ability to store variable length instructions within consecutive locations of memory space is also made possible by providing a DSP that can fetch and process variable length instructions stored in such a configuration. Fig. 6 is a flow diagram of the operation of instruction fetch unit 156 when fetching instructions from instruction memory 152 in accordance with one embodiment of the invention. The processing begins at step 200 and at step 202 a first set of instruction data is read from instruction memory 152. In the exemplary embodiment of the invention, two 32-bit words, or 64 bits, of instruction data are retrieved at step 202.
  • At step 204 the first instruction contained in the 64 bits of retrieved instruction data is processed by instruction decoder 158. In the exemplary embodiment of the invention, the instruction may be 16, 32 or 48 bits long. The instruction length is determined by a set of header bits contained in each instruction which indicate the instruction length as described in greater detail below. Various other methods for specifying instruction length should be apparent including the use of codes that demarcate and separate two instructions, or the use of a super header instruction, which specifies the length of some set of instructions which are to follow. The use of header bits is preferred in some instances because the instruction length information is kept in close proximity to the instruction, thereby reducing the need to store or maintain state information about the instruction processing.
  • After the first instruction contained within the 64 bits of retrieved instruction data is processed, it is determined at step 206 if 48 bits or more unprocessed instruction data remain in the 64 bits of retrieved instruction data. If 48 bits or more unprocessed instruction data remains, the next instruction contained in the remaining 48 bits of unprocessed data is processed again at step 204.
  • If less than 48 bits of unprocessed instruction data remain in the retrieved instruction data at step 206, additional instruction data is loaded from instruction memory 152. Various methods for loading additional instructions are contemplated. In one embodiment of the invention, sufficient additional instruction data is loaded from the instruction memory to return the amount of unprocessed data stored in the instruction fetch unit to 48 bits. Ensuring the 48 bits of unprocessed data are stored within the instruction fetch unit ensures that at least one complete instruction is available to instruction decoder 158.
  • In a preferred embodiment of the invention, instruction fetch unit retrieves a variable amount of data when less than 48 bits of unprocessed data remains depending on the particular amount of data that has been processed. In particular, if the amount of data processed is equal to or exceeds a data word (32 bits), an additional data word (32 bits) of new instruction data is retrieved. If the amount of data previously processed is equal to or exceeds two data words (64) bits, two new data words are retrieved by instruction fetch unit.
  • Determining the amount of data retrieved based on the number of words of data processed is preferred, because it keeps a sufficient amount of unprocessed data available to instruction decoder 158 while also allowing more efficient word-length access to the memory banks that make up the instruction memory. Once additional unprocessed instruction data is retrieved at step 206, the next instruction is processed within the total amount of unprocessed instruction data now available.
  • Fig. 7 is a block diagram of instruction fetch unit 156 and instruction memory 152 configured in accordance with one embodiment of the invention. Instruction memory 152 is comprised of even memory bank 302 (RAM0) and odd memory bank 300 (RAM1) each of which reads and writes 32-bit data words. The memory banks are labeled even and odd because they both are addressed within the same address space, but even addresses are directed to even memory bank 302 and odd addresses are directed to odd memory bank 300.
  • Memory banks that read and write other word sizes including 8, 16, 24, 48 and 64 bits words may be used in alternative embodiments of the invention. Additionally different numbers of memory banks may be used including 1 - 8 memory banks. The use of two memory banks with 32-bits words, however, is preferred because it reduces the overall complexity while also allowing instruction data to be addressed in manageable chunks.
  • Control logic 304 causes data words to be read from memory banks 300 and 302 to instruction registers 306 and 307. The particular memory locations read are specified by address lines 310 and 314 and the reading of the instruction is controlled by enable lines 332, 315, 316 and 318. The 32 bit outputs of instruction registers 306 and 307 are applied in 16 bit portions to inputs A, B, C and D of rotator 308. Rotator 308 outputs 48 bits of instruction data 324. The 48 bits of instruction data 324 are comprised of three of the four (3:4) inputs A, B, C and D with each input set containing 16 bits, as described in greater detail below.
  • During operation, control logic 304 loads instruction data from instruction memory banks 300 and 302 in accordance with the method described with reference to Fig. 5. In particular, control logic 304 first loads a total of 64 bits of unprocessed instruction data into instruction registers 306 and 307 by reading a 32 data word from both even memory bank 302 and odd memory bank 300. If a 16-bit instruction is processed, no new data is loaded because instruction registers 306 and 307 still contain 48 bits of unprocessed instruction data. If a 32-bit instruction is then processed, instruction register 306 is loaded with a 32-bit word of additional instruction data because less than 48 bits of unprocessed instruction data remains. Loading the 32-bit instruction word, once again, places 48 bits of unprocessed instruction data in registers 306 and 307, with 16 unprocessed bits in register 307 and the next 32 in register 306. If a 48 bit instruction is then processed, no unprocessed instruction data remains, so both registers 306 and 307 are loaded with 32 bit words of instruction data, which is 64 bit of unprocessed instruction data, which is greater than the necessary 48 bits of instruction data. While loading a full 64 bits of instruction data is not particularly necessary, it is useful because it allows the use of two 32-bit word instruction memories and registers, which are common in the industry. The use of other methods of maintaining sufficient amounts of unprocessed instruction data is consistent with the use of some aspects of the present invention.
  • Once 64 bits of new instruction data is loaded into instruction registers 306 and 307, control logic 304 further configures rotator 308 using control signals 320 to output the next 48 next bits of instruction data received on inputs A, B, C and D based on 1) the location of the instruction data within the instruction address space, 2) the set of instruction data that has been processed, and 3) the length of the previous instruction processed. In particular, rotator 308 is configured to output the set of 48 bits of instruction data next in line to be processed with the next-in-line bits of that instruction data in the most significant, or leftmost, position.
  • For example, upon loading the first two words, or 64 bits, of instruction data into registers 306 and 307, if the instruction data in the even instruction register 307 is next-in-line, the output of rotator 308 is comprised of the instruction data received on input A, B and C (ABC) in that order. If the instruction data in the odd instruction register 306 is next in line, rotator 308 is configured to output the instruction data received on inputs C, D and A (CDA) in that order.
  • As instructions are processed, new instruction data is loaded into data registers 306 and 307 as described above, and rotator 308 is configured to continue to output the next-in-line instruction data on output 324 based on the size of the previous instruction processed. The size of the previous instruction processed is conveyed to control logic 304 by header data 322, which is a copy of the first five bits of output 324. As noted above, any predetermined method for specifying the instruction length to control logic 304 is consistent with the practice of the present invention, although the use of the first five bits is preferred because it allows the instruction length to be determined directly from the instruction data.
  • In an exemplary embodiment of the invention, the size of the previous instruction is coded into two bits of state information I1 and I0 in accordance with Table 1.
    Full instruction formats.
    I1 I0 Instruction Size
    00 Branch/Stall/Reset
    01 16-bit
    10 32-bit
    11 48-bit
  • Additionally, the configuration of rotator 308 is controlled by two select bits S1 and S0 that make up control 320 which are coded as set forth in Table 2.
    Rotator select control bits and output.
    S1 S0 Rotator Output
    00 ABC
    01 BCD
    10 CDA
    11 DAB
  • As should be apparent, as the state of S1 and S0 is incremented, the output of rotator 308 is left-rotated, or barrel-shifted. The left-rotation is such that each input group (A, B, C and D) is shifted to the left on the output. The input group that was at the left most position of the output is removed. The input group previously not asserted at the output is subsequently output at the rightmost position.
  • The state of S1 and S0, and therefore the configuration of rotator 308, is updated, or rotated, by an amount that varies in response to instructions of various lengths. It particular, the value (I1 I0), representing the length of the instruction being processed, is added to the control bits S1 and S0, and any carryout value is discarded. That is: S1(t+1), S0(t+1) = S1(t), S0(t) + I0, I1
  • For branch or reset condition the values of S1 and S0 are reset based on the particular instruction to which processing branches or resets, and therefore equation (1) is not utilized. Various methods for processing branch, reset and stall instructions are well known in the art, and because this processing is not particularly relevant to the invention, it is not described further.
  • In an exemplary processing, rotator 308 begins with an output of ABC, and select bits S1 and S0 at 00. If a 16-bit instruction is received, the corresponding instruction length bits of I1 and I0 of 01 are added to S1 and S0 yielding a S1 and S0 of 01, which corresponds to an output 324 from rotator 308 of BCD. An output of BCD is the next-in-line set of instruction data after the first 16-bits of instruction data (input A) have been processed.
  • If the next instruction is a 32-bit instruction, an instruction length I1 and I0 of 10 is added to the current S1 and S0 state of 01 yielding 11. The resulting output is DAB, which corresponds to the next 48-bits of instruction data that have not been processed, configured with the next-in-line instruction data received on input D positioned in the most significant, or left most, position. Previously, the instruction data from inputs B and C had been processed. It should be noted that during the processing of the previous 32-bit instruction new data is loaded into instruction register 307 in accordance with the instruction data loaded process described above.
  • If a 48 bit instruction is then processed, the state of the select bits S1 and S0 is increased by an instruction length I1 and I0 of 11, which yields an S1 and S0 of 10, discarding the carryout, which configures rotator 308 to output CDA. An output of CDA corresponds to the next 48 bits of instruction data to be processed, with the next-in-line bits configured in the most significant position. Previously the instruction data on input D, A and B was processed. It is once again noted that new instruction data is read into register 306 during the processing of the previous instruction. The logic described above to perform the associated instruction data processing is provided for purposes of example, and the use of other logic to perform the above described instruction data processing will be apparent and is consistent with the use of the invention.
  • Thus, in the exemplary embodiment, control logic configures rotator 308 to output the next 48 bits of instruction data received from instruction registers 306 and 307 with the next instruction bits to be processed located in the left most position. Those skilled in the art will recognize that presenting the instruction data to instruction processing unit 158 in any predetermined arrangement whereby the next set of instruction data to be processed is identifiable is also consistent with the use of the invention.
  • As noted above, in addition to outputting the next 48 bits of unprocessed instruction data, rotator 308 also indicates the size of the next instruction to be processed to control logic 304. In particular, rotator 308 outputs an additional copy of the next 5 bits of instruction data to be processed to control logic 304. In the preferred embodiment of the invention, the length of the instruction is specified by the first five (5) bits of the instruction.
  • Fig. 8 is a block diagram of MAC unit 128 when configured in accordance with one embodiment of the invention. Shift right 900 receives the 40-bit input to be accumulate and shifts the value by either 0 or 16 bits, with the output applied to one input of multiplexer 901. The other input of multiplexer 901 receives the value 0x8000. Multiplier 902 receives two 16 bits values to be multiplied along with sign bits from instruction decode 158 for a total of 17-bits for each input.
  • The output of multiplier 902 is received by shift left 904, which shifts the output by 0, 1, 2 or 3 bits as specified by instruction decode 158. Adder/subtractor 906 receives the output of multiplexer 901 and shift left 904. Adder/subtractor 906 performs addition or subtraction of the two input values as instructed by instruction decoder 158 and outputs the result, which is applied to register bank input port PI4 in the exemplary embodiment of the invention.
  • The use of a shift right 16 unit 900 within MAC unit 128 in some embodiments of the invention provides additional utility over other types of MAC units. More specifically, the use of the shift right 16 unit 900 facilitates performing double precision operations in a reduced number of clock cycles. For example, to perform a double precision operation in which a 32 bit number (A) is multiplied with a 16 bits number (B), the low 16 bits (Al) of the 32 bit number are first multiplied with the 16 bit number B during a first clock cycle, yielding an intermediate value I that is stored in register bank 120.
  • During a second clock cycle the intermediate value I is input into right shift 16 unit 900 and shifted right by 16 bits. Additionally, the 16 bit number B and the high 16 bits of the 32 bit number A (Ah) are multiplied, and the result added with the right shifted intermediate value I from shift right 16 unit 900. Thus, a double precision multiply is performed in two clock cycles rather than three. In general, many double precision operations require one or more variables to be shifted relative to the other variables, and as such allowing the shifting step to be performing during the same clock cycle as one of the multiply or accumulate operation reduces the number of cycles necessary to perform the double precision operation.
  • Three clock cycles would normally be required because the first multiply operation, the shift operation, and the second multiply operation each typically require a clock cycle. Therefore, the use of the shift circuit reduces the number of clock cycles required to perform the double precision multiply. Other double precision operations involving differently sized operands are also facilitated by the use of shift right unit 900.
  • II. INSTRUCTION CONFIGURATION A. Overview.
  • Fig. 9 is a block diagram illustrating the instruction hierarchy used in the exemplary embodiment of the invention. Block 402 shows the variable length full instructions comprised of 16, 32 or 48 bits which control the operation of the DSP. The variable length instructions are in turn made up of instruction fragments including general instruction fragments and memory move and program flow (MMPF) instructions as shown in block 403. The general instruction fragments used in the exemplary embodiment of the invention include MAC8, MAC16, ALU8, ALU16, DMOV16, DMOV24 and DL40 instruction fragments. The MMPF instruction fragments include OneMem11, TwoMem19, TwoMov19 and ThreeMem24 instruction fragments. The MMPF instruction fragments are made up of the MMPF instruction subfragments shown in block 406. The MMPF instruction subfragments include LD(A), LD(B), ST(A), ST(B), LS(C), DMOVA, DMOVB, and PF8. The various full instructions, instruction fragments and instruction subfragments are described in greater detail below.
  • B. Full Instructions.
  • In the preferred embodiment of the invention the DSP is controlled using full instructions having lengths of 16, 32 and 48 bits. The full instructions are in turn formed by combining one or more instruction fragments. The full instructions are configured to allow for consecutive storage within the instruction memory 152 and processing by the DSP. The format and configuration of the full instructions are described below, followed by the format and configuration of the instruction fragments. During operation, the DSP processes a full instruction each clock cycle. Thus, multiple operations can be performed during the processing of each full instruction, with particular operations determined by the particular set of instruction fragments selected.
  • The format of the three full instructions used in the exemplary embodiment of the invention are shown in Table 3.
    Figure 00250001
  • The five (5) bit header used for each full instruction indicates the length of the full instruction, and some additional information as to the contents of the full instruction. The format of the header used in the exemplary embodiment of the invention is provided in Table 4.
    Full instruction header formats.
    5-bit Header Instruction Length
    0 0 0 0 X 16-bit Instruction (2 types)
    0 0 0 1 X 32-bit Instruction (2 types)
    0 0 1 X X 48-bit Instruction (4 types)
    0 1 X X X 32-bit Instruction (8 types)
    1 X X X X 48-bit Instruction (16 types)
  • Each full instruction (16, 32 and 48 bit long) contains one or more instruction fragments. Table 5 provides a list of the available instruction fragments in the exemplary embodiment of the invention. A more detailed description of the format and operation of the instruction fragments is provided after the discussion of the full instructions.
    Instruction fragments.
    Field Explanation Width
    MAC8 8-bit MAC Operation 8
    ALU8 8-bit ALU Operation 8
    OneMem11 One Memory Operation 11
    MAC16 16-bit MAC Operation 16
    ALU16 16-bit ALU Operation 16
    DMOV16 Conditional Reg Move / Inport / Outport 16
    TwoMem19 Two Memory Operations 19
    TwoMov19 Two Memory/Data Move Operations 19
    DMOV24 Load/Store Direct / Load Addr / Jump 24
    ThreeMem27 Three Memory Operations 27
    DL40 Dual-Load 40
  • Tables 6 - 8 provide the various combinations of instruction fragments that may be used within 48, 32 and 16 bit full instructions in accordance with the exemplary embodiment of the invention. While other combinations of instruction fragments are consistent with the use and operation of the invention, certain features of the combination disclosed herein are preferred as discussed in greater detail below. Additionally, where all or part of full instructions are shown as "reserved," no particular instruction combination is specified or used in the described embodiment, but future use of these full instruction combinations is contemplated.
  • Table 6 provides the formatting for 16 bit full instructions when performed in accordance with the exemplary embodiment of the invention described herein. The full instruction is comprised of the five (5) bit header followed by eleven (11) instruction bits.
    Figure 00270001
  • The header bits indicate the length of the instruction as well as some information about the type of instruction. For a header of 00000, the least significant three tail bits are used to further specify the operation performed. In particular, tail bits of 000 indicate the remaining eight bits contain a MAC8 instruction fragment. Tail bits of 001 indicate the remaining eight bits contain a ALU8 instruction fragment. For other tail bit combinations, no instructions are specified.
  • For a header of 00001, the remaining eleven (11) bits contain a OneMem11 instruction fragment. By providing a 16 bit full instruction that allows ALU, MAC or memory move operations to be performed, the most common operations can be performed with the shortest full instruction. Since the shortest instruction requires the least amount of memory to store, the use of 16 bit full instruction as described reduces the amount of instruction-memory necessary to perform a particular set of operations. Thus, the overall size of the DSP, and therefore the cost and power consumption, is reduced as well.
  • The 16 bit instruction is typically used when conditions are such that only one, or a reduce number of, operations can be performed. Typically, the size of the instruction necessary to specify only one operation can be reduced, hence the use of the half-word, or 16 bit instruction for performing one operation. Additionally, the 16 bit instruction can be used for a MAC, ALU, memory move or program flow operation which encompasses almost all of the operations one would expect to perform.
  • Table 7 illustrates the instruction fragment combinations and associated formatting of a 32-bit full instruction when configured in accordance with one embodiment of the invention.
    Figure 00290001
  • As noted above, the five header bits indicate the length of the full instruction, as well as the particular combination of instruction fragments. For example, a header of 00010 indicates the remaining 27 instruction bits contain a ThreeMem27 instruction fragment, and a header of 00011 indicates the remaining 27 instruction bits contain a ALU8 instruction fragment followed by a TwoMem19 instruction fragment.
  • For a header of 01111, the least significant tail bits further indicate the combination of instruction fragments. For example, for a least significant tail bit of 0, the next two least significant bits indicate whether the remaining 24 bits contain a DMOV24, a ALU16 followed by a MAC8, or a MAC16 followed by a ALU8 instruction fragments. Other tail bit states, such as a least significant tail bit of 1, specify reserved combinations.
  • The thirty-two bit instruction allows many of the most commonly performed operations to be performed simultaneously, which facilitates pipelining while also reducing the instruction size. For example, it is common to perform two fetch operations and a multiply/accumulate operation for such applications as filtering. The 32-bit instruction allows such sets of operations to be performed in pipelined fashion while not requiring a full 48 bits of instruction space.
  • In addition, the 32-bit instruction allows MAC and ALU operations to be performed simultaneously as well as program jump and call operations, also without the use of the largest instruction size.
  • Table 8 illustrates the instruction fragment combinations and format for 48 bit full instructions when performed in accordance with one embodiment of the invention.
    Figure 00310001
  • The five header bits specify the length of the instruction as well as the particular instruction fragment combination. For example, header bits of 00100 indicate the 43 remaining instruction bits are comprised of DMOV24, MAC8 and OneMem11 instruction fragments. Header bits of 10011 indicate the 43 remaining bits are comprised of ALU16, MAC8 and TwoMem19 instruction fragments.
  • For header bits of 11111, the three least significant tail bits further indicate the instruction fragments contained in the remaining instruction bits. For example, tail bits of 000 indicate the remaining 40 instruction bits contain MAC16 and DMOV24 instruction fragments. Tail bits of 001 indicate the remaining 40 instruction bits contain MAC8, ALU8 and DMOV24 instruction fragments. Tail bits of 110 indicate the remaining 40 instruction bits contain a DL40 instruction fragment.
  • The instruction fragment combinations provided in 48 bit full instructions allow many operations to be performed simultaneously and therefore more rapidly than if performed serially. For example, several 48 bit full instructions allow ALU operations, MAC operations and memory operations to all be performed simultaneously. The memory operations include load, store, and data move operations, and often allow multiple memory locations to be accessed at once.
  • The 48 bit instruction allows multiply operations to be performed in combination with ALU operations and data fetch and program flow operations, all in pipelined fashion. This can be useful for filtering when combined with scaling operations, which are often performed by performing a MAC operation followed by an ALU (such as shifting) operation. Other applications which use MAC and ALU operations include combining three or more streams of data. The 48 bit instruction, especially in combination with the use of the three buses architecture, facilities the pipelining of operations in these cases.
  • This effectively increases the number of operations that can be performed in a single 48 bit full instruction to five (MAC, ALU, FETCH1, FETCH2, and STORE). The ability to simultaneously perform multiple instructions in the DSP, in general, is further enhanced by the use of a DSP with multiple internal buses for coupling the various processing systems within the DSP. Different set of data may be moved and accessed simultaneously using the different busses.
  • Varying the length of the instruction based on the number of operations that can be performed further increases the efficiency with which instruction memory is used. Any particular task has periods where multiple operations can be performed simultaneously, and other periods where fewer, or only one, operation can be performed. By adjusting the length of the instruction in accordance with the number of operations that can be performed simultaneously, the amount of instruction memory is reduced.
  • When combined with the use of tight instruction packing, an exemplary method of which is described above, the required instruction memory is further reduced. The use of variable length instructions or tightly packed instructions, or both, facilitate the use of a multiple bus architecture and a multi-access register bank, by making more circuit area available for implementing these features. Thus, the combination of these aspects of the invention combine synergistically to simultaneously provide the benefits of improved performance and improved efficiency.
  • C. Instruction Fragments
  • As noted above, full instructions are comprised of a set of one or more instruction fragments grouped together in predefined ways. The set of available instruction fragments in the exemplary embodiment of the invention is shown in Table 5. The instruction fragments and the combinations made available using the full instructions provided in the exemplary embodiment of the invention are designed to allow the set of operations most likely to be performed together to be combined so that the amount of instruction memory necessary to perform a given operation is reduced. A discussion of the operation and format of the various instruction fragments used in the exemplary embodiment of the invention follows.
  • C.1 Instruction Fragment Nomenclature
  • Throughout the following discussion of the instruction fragments and subfragments, the following abbreviations are used to refer to the registers listed in Tables 9 and 10 below. Additionally, the particular bit codes (mappings) used in the exemplary embodiment of the invention are shown to the left.
    Figure 00340001
    Figure 00350001
  • C.2 Instruction Fragment Description
  • The set of instruction fragments includes two types of MAC instruction fragments: MAC8 and MAC16. MAC8 instruction fragments support signed-unsigned and signed-signed multiply types, with the results being stored in accumulators L0 or L1. The MAC8 instruction fragment saves instruction RAM by allowing a MAC operation using a 16 bit full instruction, and for many parallel instruction combinations requiring MAC operations to be encoded into 32-bit instruction instead of 48-bit instructions. In general, the processing performed by a MAC8 instruction conforms to the following equation:
    Figure 00360001
  • As shown in equation (2), the MAC8 instruction fragment allows the contents of registers L0 or L1 to be summed with the product of registers R0, R2. R4 and R6 and R0, R1, R3, and R5, or set directly to the product of registers. Additionally, signed or unsigned multiplies may be specified. By limiting the number of registers for which the MAC operations may be performed using a MAC8 instruction, the length of the instruction can be kept to 8 bits, allowing the MAC operations to be performed using a shorter, 8 bit, instruction fragment.
  • The particular operation performed by the MAC8 instruction is specified by the values of the eight bits that make up the instruction as set forth in Table 11.
    MAC8 instruction fragment format.
    7 6 5 4 3 2 1 0
    MAC Operation mac8Op1 mac8Op2 SU/SS
  • SU/SS specifies signed or unsigned multiply. The codes for specifying various operations within the MAC8 instruction fragment are listed in Table 12.
    Figure 00360002
  • Thus, a MAC8 instruction of 0x99 places the sum the contents of register L0 with the unsigned product of registers R0 and R3 into register L0.
  • The MAC16 instruction fragment provides additional flexibility by allowing additional registers to be used in the multiply-accumulate operation. Equation (3) sets forth the operations that can be performed using the MAC16 instruction fragment.
    Figure 00370001
  • For example, every accumulator (L0 - L3) can be used as a destination, although not all combinations of accumulators are allowed in multiply-accumulate instructions. The CPS field signals that a coprocessor should perform a particular operation in parallel. The particular operation performed by the MAC16 instruction is specified by the values of the sixteen bits that make up the instruction are set forth in Table 13.
    MAC16 instruction fragment format.
    15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 0
    MAC Operation macOp1 macOp2 mtype mshift CP S
  • The codes for specifying various operations within the MAC16 instruction fragment are listed in Table 14.
    Figure 00380001
  • The MAC16 instruction fragment allows left shifts of up to 3, and can perform round operations during straight multiplies (no accumulate), with the round occurring after the shift. When an accumulation is performed, the accumulator to be added can be shifted down by 16 in parallel with a signed-signed multiply. The CPS bit is the coprocessor strobe bit for indicating that data used in the MAC operation should be sent to a coprocessor.
  • It should be noted that the MAC8 instruction fragment performs a subset of the operations that can be performed by the MAC16. The particular set of instructions selected for the MAC8 instruction fragment are the most commonly performed out of the set of operations that can be performed using the MAC16 instruction fragment. This saves program memory by allowing the majority of MAC operations to be performed using the MAC8 instruction fragment.
  • The 8 bit ALU8 instruction fragment comprises the ALU operations that are most commonly paralleled with MAC operations (MAC8 and MAC16) and which do no contain immediate data. All ALU8 shift operations are arithmetic shifts that use an internal shift register (SR) register to save instruction encoding bits. The operations performed using the ALU8 instruction fragment are shown in Table 15.
    ALU8 instruction fragment operations.
    NOP; NOP (needed for parallel combinations).
    LD = DETNORM(LS); Determine block normalization factor.
    LD = SET(LS); Copy accumulator (no saturation).
    LD = LS <<< SR; Shift accumulator.
    LD = RND(LS << SR); Shift and round accumulator.
    LD = LD ± (LS << SR); Accumulate shifted accumulator.
    LD = LS ± LT; Add or subtract accumulators.
    LS ± LT; Add/subtract accs result free (set flags).
    LS is load source (L0 - L3) and LD is load destination (L0 - L3).
  • The particular operations performed by the ALU8 instruction fragment are specified by the values of the eight bits that make up the instruction fragment as set forth in Table 16.
    ALU8 instruction fragment format.
    7 6 5 4 3 2 1 0
    0 ALUOp LS LD
    0 1 1 Sign LS LT
    1 LD Sign LS LT
  • The particular codes used to specify the operations performed using the ALU8 instruction fragment are set forth in Table 17.
    Figure 00390001
  • The ALU16 instruction fragment allows both arithmetic and logical shifts. The particular operations performed by the ALU16 instruction fragment are set forth in Table 18.
    Figure 00410001
  • The format of the ALU16 instruction fragment is set forth in Table 19.
    Figure 00420001
  • The particular operations performed by the ALU16 instruction fragment are specified by the values of the bits that make up the instruction fragment as set forth in Table 20.
    Figure 00430001
  • The DMOV16 instruction fragment is a 16 bit instruction fragment for performing different data move, data inport and data outport operations as set forth in Table 21.
    Figure 00440001
  • The format and codes used to perform the operations available using the DMOV16 instruction fragment are set forth in Table 22.
    Figure 00450001
  • The instruction OUTPORTA(port_addr) reads the value on Abus and outputs it to the designated port. By reading a value from memory A simultaneously, this instruction can be used to send a value directly from memory A to the port. OUTPORTB(port_addr) operates similarly.
  • The DMOV24 instruction fragment is a 24 bit instruction fragment for performing different load/store register direct or load register immediate operations as set forth in Table 23.
    Figure 00460001
  • Table 24. provides the format and some codes used to perform the various operation available using the DMOV24 instruction fragment in accordance with the exemplary embodiment of the invention.
    Figure 00470001
  • It should be noted that, for DMOV24 as well as other instruction fragments, some operations are encoded twice. For example, the formats specified in rows (i) and (j) encode the same operation, with one specifying the use of immediate bus Im1 and the other specifying the use of Immediate bus Im2. Encoding twice allows the instruction fragment to be combined with a greater variety of other instruction fragments, which may require the use of Immediate bus 1 or Immediate bus 2 as well.
  • The 40-bit dual load instruction fragment (DL40) is a 40 bit instruction fragment for performing immediate load or address load operations. The particular operations performed in the exemplary embodiment of the invention are as shown in Table 25.
    Figure 00480001
  • The format of the DL40 instruction fragment for each operation is provided in Table 26.
    Figure 00490001
  • As also shown in Table 5, four types of memory move and program flow instruction fragments are provided in the exemplary embodiment of the invention, a list of which is provided in Table 27.
    Figure 00490002
  • Each memory move and program flow instruction (MMPF) fragment is comprised of a set of MMPF subfragments listed in Table 28.
    Figure 00500001
  • The format and operation of the MMPF instruction fragments are discussed first, followed by a more detailed discussion of the format and operation of the MMPF subfragments.
  • The OneMem11 MMPF instruction fragment is used to perform single memory load and store operations, data move operations, and program flow operations. In the exemplary embodiment provided herein eight different operations are performed using the OneMem11 MMPF instruction fragment, with the particular operation indicated by the first three bits of eleven bit fragment as shown in Table 29, which lists the operations that can be performed using a OneMem11 data move instruction fragment.
    OneMem11 instruction fragment format.
    10 9 8 7 6 5 4 3 2 1 0
    0 0 0 LD(A)
    0 0 1 ST(A)
    0 1 0 LD(B)
    0 1 1 ST(B)
    1 0 0 LS(C)
    1 0 1 DMOVA
    1 1 0 DMOVB
    1 1 1 PF8
  • TwoMem19 MMPF instruction fragment is a 19 bit instruction fragment that allows eight different combinations of memory load and store operations to be performed as set forth in Table 30.
    TwoMem19 instruction fragment format.
    18 17 16 15-8 7-0
    0 0 0 LD(A) LD(B)
    0 0 1 LD(A) ST(B)
    0 1 0 LD(A) LS(C)
    0 1 1 ST(A) LD(B)
    1 0 0 ST(A) ST(B)
    1 0 1 ST(A) LS(C)
    1 1 0 LS(C) LD(B)
    1 1 1 LS(C) ST(B)
  • The TwoMov19 MMPF instruction fragment is a 19 bit instruction fragment that allows eight different combinations of memory load and store operations along with data move operations as shown in Table 31.
    TwoMov19 instruction fragment format.
    18 17 16 15-8 7-0
    0 0 0 LD(A) DMOVB
    0 0 1 ST(A) DMOVB
    0 1 0 DMOVA LD(B)
    0 1 1 DMOVA ST(B)
    1 0 0 DMOVA LS(C)
    1 0 1 LS(C) DMOVB
    1 1 0 DMOVA DMOVB
    1 1 1 Reserved
  • The ThreeMem27 MMPF instruction fragment is a 27 bit instruction fragment that allows eight different combinations of memory load, memory store, and data operations to be performed as shown in Table 32.
    ThreeMem27 instruction fragment format.
    26 25 24 23-16 15-8 7-0
    0 0 0 LS(C) LD(A) LD(B)
    0 0 1 LS(C) LD(A) ST(B)
    0 1 0 LS(C) ST(A) LD(B)
    0 1 1 LS(C) ST(A) ST(B)
    1 0 0 LS(C) DMOVA LD(B)
    1 0 1 LS(C) DMOVA ST(B)
    1 1 0 LS(C) LD(A) DMOVB
    1 1 1 LS(C) ST(A) DMOVB
  • Equation (4) provides the operations performed by the LD(A) instruction subfragment.
    Figure 00520001
  • Table 33 provides the format the LD(A) instruction subfragment in accordance with the exemplary embodiment of the invention.
    LD(A) instruction subfragment format.
    7 6 5 4 3 2 1 0
    dreg A0-A3 Amod
  • Equation (5) provides the operations performed by the LD(B) instruction subfragment.
    Figure 00520002
  • Table 34 provides the format the LD(B) instruction subfragment in accordance with the exemplary embodiment of the invention.
    LD(B) instruction subfragment format.
    7 6 5 4 3 2 1 0
    dreg B0-B3 Bmod
  • Equation (6) provides the operations performed by the ST(A) instruction subfragment.
    Figure 00520003
  • Table 35 provides the format the ST(A) instruction subfragment in accordance with the exemplary embodiment of the invention.
    ST(A) instruction subfragment format.
    7 6 5 4 3 2 1 0
    dreg A0-A3 Amod
  • Equation (7) provides the operations performed by the ST(B) instruction subfragment.
    Figure 00530001
  • Table 36 provides the format the ST(B) instruction subfragment in accordance with the exemplary embodiment of the invention.
    ST(A) instruction subfragment format.
    7 6 5 4 3 2 1 0
    dreg B0-B3 Bmod
  • Table 37 lists the operations performed by the DMOVA instruction subfragment.
    Figure 00530002
  • Table 38 provides the format the DMOVA instruction subfragment in accordance with the exemplary embodiment of the invention.
    Figure 00540001
    Thus, by providing MMPF instruction fragments that can contain one or more instruction subfragments, the number of operations that can be performed using a full instruction is further enhanced. For example, a full instruction can cause arithmetic and MAC operations to be performed along with a set of up to three memory move and program flow operations. The ability to perform this many operations using a single instruction further reduces the total number of instructions necessary to perform a given operation and therefore decreases the total instruction memory required on the DSP. Reducing the instruction memory decreases the die size and therefore the cost and power consumption of the DSP, which makes that DSP more suitable for a wide variety of applications include mobile wireless telephony.
  • Thus, a system and method for controlling a DSP using a highly parallel variable length instruction set has been described. The previous description of the preferred embodiments is provided to enable any person skilled in the art to make or use the present invention. The various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without the use of the inventive faculty. For example, while the system and method are described in the context of a DSP, various aspects thereof are applicable to general computing systems and devices.
  • Having thus described the invention by reference to preferred embodiments it is to be well understood that the embodiments in question are exemplary only and that modifications and variations such as will occur to those possessed of appropriate knowledge and skills may be made without departure from the scope of the invention as defined by claims.

Claims (32)

  1. A digital signal processor for processing signals in response to instruction data, said digital processor comprising:
    a register bank (120) for storing data to be processed;
    a first memory bank (102) for storing first data;
    a second memory bank (103) for storing second data;
    a third memory bank (104) for storing third data;
    a first data bus (108) for reading said first data into a first register in said register bank (120);
    a second data bus (110) for reading said second data into a second register in said register bank (120);
    a first processing unit (128,130) for processing data to be processed in said first register and said second register, and for writing a result to a third register in said register bank (120); and
    a third data bus (112) for writing said result to said third memory bank (104) from said third register,
    wherein said first, second and third data buses can be operated simultaneously to exchange data between the register bank and the first, second and third memory banks, respectively.
  2. The digital signal processor as set forth in claim 1, further comprising:
    a second processing unit (130) for generating a more processed result by processing said result, and for storing said result in a fourth register in said register bank (120), wherein said third data bus (112) is further for writing said more processed result from said fourth register to said third memory (104).
  3. The digital signal processor as set forth in claim 1, wherein said third data bus (112) is wider than said first data bus (108) and said second data bus (110).
  4. The digital signal processor as set forth in claim 2, wherein a first set of registers in said register bank (120) can write to both said first processing unit (128) and said second processing unit (130), and a second set of registers in said register bank (120) can write to one of the set of said first processing unit (128) or said second processing unit (130).
  5. The digital signal processor as set forth in claim 2, wherein a first set of registers in said register bank (120) can read from both said first processing unit (128) and said second processing unit (130), and a second set of registers in said register bank (120) can read from one of the set of said first processing unit (128) or said second processing unit (130).
  6. The digital signal processor as set forth in any preceding claim, wherein said instruction data comprises variable length instructions, said variable length instructions containing a variable number of instruction fragments, said digital signal processor further comprising:
    an instruction fetch unit (156) for fetching instruction data sufficient for at least one complete instruction of maximum length, said instruction data containing said variable length instructions;
    an instruction decoder (158) for decoding said instruction and generating control signals.
  7. The digital signal processor as set forth in claim 1, wherein:
    the first data bus (108) is further for writing data between the first memory bank (102) and the register bank (120);
    the second data bus (110) is further for writing data between the second memory bank (103) and said register bank (120);
    the third data bus (112) is further for writing data between a third memory bank (104) and said register bank (120), wherein
    said first data bus (108), said second data bus (110), and said third data bus (112) operate simultaneously.
  8. The digital signal processor as set forth in claim 7, further comprising:
    an instruction fetch unit (156) for fetching instructions of variable length, said instructions requesting a set of operations;
    an instruction decoder (158) for decoding said instructions of variable length and for causing said set of operations to be performed.
  9. The digital signal processor as set forth in claim 7, further comprising:
    a second processing unit (130) for processing data in said register bank (120) simultaneously with said first processing unit (128).
  10. The digital signal processor as set forth in claim 1, wherein said first bus (108) is narrower than said third bus (112).
  11. The digital signal processor as set forth in claim 1, further comprising a control system, for controlling said first data bus (108), said second data bus (110), and said third data bus (112).
  12. The digital signal processor as set forth in claim 2, wherein said first processing unit (128) is a multiply-accumulate unit, and said second processing unit (130) is an arithmetic logic unit.
  13. The digital signal processor as set forth in claim 2, wherein the processing units (128,130) are operable to receive data from each data bus (108,110,112).
  14. The digital signal processor as set forth in claim 2, wherein:
    a first set of the registers is operable to write data to both said first processing unit (128) and said second processing unit (130); and
    a second set of the registers is operable to write data to said first processing unit (128), but not said second processing unit (130).
  15. The digital signal processor as set forth in claim 1, wherein the first processing unit (128) comprises:
    a multiply accumulate unit, and the digital signal processor further comprises:
    a shift unit (900,904) coupled to an input of said multiply accumulate unit (128,900 to 906) for shifting input data in a first configuration and passing said input data through in a second configuration.
  16. The digital signal processor of claim 15, wherein said shift unit (900,904) shifts said input data and said multiply accumulate unit (128,900 to 906) performs said multiply accumulate operations during a single processing cycle.
  17. A method of processing data comprising:
    receiving first data from a first memory bank (102) via a first data bus (108);
    storing said first data in a first register (120);
    receiving second data from a second memory bank (103) via a second data bus (110); and
    storing said second data value in a second register (120);
    generating (128) a result using said first data and said second data;
    storing said result in a third register (120); and
    writing said result to a third memory bank (104) via a third data bus (112),
    wherein said first, second and third data buses can be operated simultaneously to exchange data between the registers and the first, second and third memory banks, respectively.
  18. The method as set forth in claim 17, wherein:
    said first data comprises a first data value; and
    said second data comprises a second data value.
  19. The method as set forth in claim 17 or 18, wherein said third data bus (112) is wider than said first data bus (108).
  20. The method as set forth in claim 17 or 18, wherein said third data bus (112) is wider than said first data bus (108) and said second data bus (110).
  21. The method as set forth in claim 17, further comprising:
    generating (128) said result using a first processing unit;
    generating (130) a second result from said first result using a second processing unit coupled to said third register (120).
  22. The method as set forth in claim 17 for operating a digital signal processor using variable length instructions, said variable length instructions having a set of instruction fragments each for requesting an operation, said method comprising:
    reading previously processed data from one of said registers (120) into a first processing unit (128) during a first clock phase of a first clock cycle;
    processing (128,130) said previously processed data based on a first instruction fragment from the set of instruction fragments, and during said first clock cycle, yielding twice processed data;
    processing (128,130) new data based on a second instruction fragment from said set of instruction fragments, and during said first clock cycle, producing new processed data;
    writing said new processed data into said one of said registers (120) during a second phase of said first clock cycle; and
    writing said twice processed data into a second of said registers (120) during said second phase of said first clock cycle.
  23. The method as set forth in claim 22, wherein said processing (128,130) said previously processed data is performed by a first processing unit, and said processing (128,130) new data is performed by a second processing unit.
  24. The method as set forth in claim 22, further comprising:
    reading instruction data containing said variable length instructions;
    determining a next instruction length;
    decoding an amount of data in said instruction data equal to said next instruction length.
  25. The method as set forth in claim 17 for operating a digital signal processor, said method further comprising:
    reading processed data from one of said registers (120) into a first processing unit during a first clock phase of a first clock cycle;
    processing (128,130) said processed data using said first processing unit during said first clock cycle yielding more processed data;
    processing (128,130) other data in a second processing unit during said first clock cycle producing new processed data; and
    writing said new processed data into said one of said registers (120) during a second phase of said first clock cycle.
  26. The method as set forth in claim 25, further comprising:
    writing said more processed data into a second of said registers (120) during said second phase of said first clock cycle.
  27. The method of claim 17 for controlling a digital signal processor using instructions stored in a memory (152), said memory having an address space, said method comprising:
    writing out a first data word of the memory containing a first part of a first instruction; and
    writing out a second data word of the memory containing a second part of said first instruction and a first part of a second instruction.
  28. The method as set forth in claim 27, wherein said first instruction is comprised of a plurality of instruction fragments, each instruction fragment for performing a particular operation.
  29. The method as set forth in claim 27, wherein said first instruction and said second instruction are different lengths.
  30. The method as set forth in claim 27, wherein said first instruction and said second instruction are comprised of sets of instruction fragments, wherein each instruction fragment is for performing a particular operation.
  31. The method as set forth in claim 30, wherein said instruction fragments include a first instruction fragment and a second instruction fragment, and said first instruction fragment requests a set of operations that are a subset of the set of operations performed by said second instruction fragment.
  32. The method as set forth in claim 31, wherein said first instruction fragment is shorter than said second instruction fragment.
EP99911150A 1998-03-18 1999-03-04 A digital signal processor reducing access contention Expired - Lifetime EP1066559B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP04005665.7A EP1457876B1 (en) 1998-03-18 1999-03-04 Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions

Applications Claiming Priority (13)

Application Number Priority Date Filing Date Title
US4408998A 1998-03-18 1998-03-18
US4410898A 1998-03-18 1998-03-18
US4408798A 1998-03-18 1998-03-18
US4410498A 1998-03-18 1998-03-18
US09/044,088 US6496920B1 (en) 1998-03-18 1998-03-18 Digital signal processor having multiple access registers
US09/044,086 US6425070B1 (en) 1998-03-18 1998-03-18 Variable length instruction decoder
US44087 1998-03-18
US44089 1998-03-18
US44108 1998-03-18
US44104 1998-03-18
US44088 1998-03-18
US44086 1998-03-18
PCT/US1999/004887 WO1999047999A1 (en) 1998-03-18 1999-03-04 A digital signal processor

Related Child Applications (1)

Application Number Title Priority Date Filing Date
EP04005665.7A Division EP1457876B1 (en) 1998-03-18 1999-03-04 Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions

Publications (2)

Publication Number Publication Date
EP1066559A1 EP1066559A1 (en) 2001-01-10
EP1066559B1 true EP1066559B1 (en) 2005-06-08

Family

ID=27556460

Family Applications (2)

Application Number Title Priority Date Filing Date
EP04005665.7A Expired - Lifetime EP1457876B1 (en) 1998-03-18 1999-03-04 Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions
EP99911150A Expired - Lifetime EP1066559B1 (en) 1998-03-18 1999-03-04 A digital signal processor reducing access contention

Family Applications Before (1)

Application Number Title Priority Date Filing Date
EP04005665.7A Expired - Lifetime EP1457876B1 (en) 1998-03-18 1999-03-04 Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions

Country Status (12)

Country Link
EP (2) EP1457876B1 (en)
JP (4) JP2002507789A (en)
KR (3) KR100896674B1 (en)
CN (2) CN1279435C (en)
AR (5) AR026081A2 (en)
AT (1) ATE297567T1 (en)
AU (1) AU2986099A (en)
CA (1) CA2324219C (en)
DE (1) DE69925720T2 (en)
DK (1) DK1066559T3 (en)
HK (2) HK1035594A1 (en)
WO (1) WO1999047999A1 (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1457876B1 (en) * 1998-03-18 2017-10-04 Qualcomm Incorporated Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions
JP4100300B2 (en) * 2003-09-02 2008-06-11 セイコーエプソン株式会社 Signal output adjustment circuit and display driver
JP4661169B2 (en) * 2003-11-14 2011-03-30 ヤマハ株式会社 Digital signal processor
JP4300151B2 (en) * 2004-04-19 2009-07-22 Okiセミコンダクタ株式会社 Arithmetic processing unit
US7246218B2 (en) * 2004-11-01 2007-07-17 Via Technologies, Inc. Systems for increasing register addressing space in instruction-width limited processors
US7337272B2 (en) * 2006-05-01 2008-02-26 Qualcomm Incorporated Method and apparatus for caching variable length instructions
US11768689B2 (en) 2013-08-08 2023-09-26 Movidius Limited Apparatus, systems, and methods for low power computational imaging
US10001993B2 (en) 2013-08-08 2018-06-19 Linear Algebra Technologies Limited Variable-length instruction buffer management
FR3021427B1 (en) * 2014-05-22 2016-06-24 Kalray VLIW TYPE INSTRUCTION PACKAGE STRUCTURE AND PROCESSOR ADAPTED TO PROCESS SUCH INSTRUCTION PACKAGE
WO2016016730A1 (en) * 2014-07-30 2016-02-04 Linear Algebra Technologies Limited Low power computational imaging
EP4116819A1 (en) * 2014-07-30 2023-01-11 Movidius Limited Vector processor
WO2018012828A1 (en) * 2016-07-13 2018-01-18 김태형 Multi-function calculation device and fast fourier transform calculation device
JP7384374B2 (en) * 2019-02-27 2023-11-21 株式会社ウーノラボ central processing unit
US11204768B2 (en) 2019-11-06 2021-12-21 Onnivation Llc Instruction length based parallel instruction demarcator

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4099229A (en) * 1977-02-14 1978-07-04 The United States Of America As Represented By The Secretary Of The Navy Variable architecture digital computer
US5293611A (en) * 1988-09-20 1994-03-08 Hitachi, Ltd. Digital signal processor utilizing a multiply-and-add function for digital filter realization
JP2791086B2 (en) * 1989-03-15 1998-08-27 富士通株式会社 Instruction prefetch device
EP0436341B1 (en) * 1990-01-02 1997-05-07 Motorola, Inc. Sequential prefetch method for 1, 2 or 3 word instructions
US5295249A (en) * 1990-05-04 1994-03-15 International Business Machines Corporation Compounding preprocessor for cache for identifying multiple instructions which may be executed in parallel
JP2560889B2 (en) * 1990-05-22 1996-12-04 日本電気株式会社 Microprocessor
JP2682761B2 (en) * 1991-06-18 1997-11-26 松下電器産業株式会社 Instruction prefetch device
US5438668A (en) * 1992-03-31 1995-08-01 Seiko Epson Corporation System and method for extraction, alignment and decoding of CISC instructions into a nano-instruction bucket for execution by a RISC computer
JPH06103068A (en) * 1992-09-18 1994-04-15 Toyota Motor Corp Data processor
JPH06250854A (en) * 1993-02-24 1994-09-09 Matsushita Electric Ind Co Ltd Instruction prefetching device
JP3168845B2 (en) * 1994-10-13 2001-05-21 ヤマハ株式会社 Digital signal processor
US5819056A (en) * 1995-10-06 1998-10-06 Advanced Micro Devices, Inc. Instruction buffer organization method and system
JP3655403B2 (en) * 1995-10-09 2005-06-02 株式会社ルネサステクノロジ Data processing device
US5710914A (en) * 1995-12-29 1998-01-20 Atmel Corporation Digital signal processing method and system implementing pipelined read and write operations
JP2806359B2 (en) * 1996-04-30 1998-09-30 日本電気株式会社 Instruction processing method and instruction processing device
EP1457876B1 (en) * 1998-03-18 2017-10-04 Qualcomm Incorporated Systems for and method of controlling a digital signal processor using a variable length instruction set, method of generating and storing said instructions and memory comprising said instructions

Also Published As

Publication number Publication date
JP6152558B2 (en) 2017-06-28
AU2986099A (en) 1999-10-11
KR100835148B1 (en) 2008-06-04
HK1035594A1 (en) 2001-11-30
AR026081A2 (en) 2002-12-26
JP2015028793A (en) 2015-02-12
CN1523491A (en) 2004-08-25
DK1066559T3 (en) 2005-10-03
HK1094608A1 (en) 2007-04-04
KR20010082524A (en) 2001-08-30
AR026078A2 (en) 2002-12-26
AR026082A2 (en) 2002-12-26
CA2324219A1 (en) 1999-09-23
AR026080A2 (en) 2002-12-26
KR100896674B1 (en) 2009-05-14
EP1457876A3 (en) 2006-11-02
JP2002507789A (en) 2002-03-12
KR20060040749A (en) 2006-05-10
EP1066559A1 (en) 2001-01-10
JP2010282637A (en) 2010-12-16
CN1279435C (en) 2006-10-11
KR100940465B1 (en) 2010-02-04
JP5677774B2 (en) 2015-02-25
CN1301363A (en) 2001-06-27
CA2324219C (en) 2011-05-10
EP1457876B1 (en) 2017-10-04
DE69925720T2 (en) 2006-03-16
DE69925720D1 (en) 2005-07-14
WO1999047999A1 (en) 1999-09-23
KR20060040748A (en) 2006-05-10
ATE297567T1 (en) 2005-06-15
AR026079A2 (en) 2002-12-26
EP1457876A2 (en) 2004-09-15
JP6300284B2 (en) 2018-03-28
JP2016146189A (en) 2016-08-12

Similar Documents

Publication Publication Date Title
JP5677774B2 (en) Digital signal processor
US6425070B1 (en) Variable length instruction decoder
US20070186079A1 (en) Digital signal processor with variable length instruction set
US6829696B1 (en) Data processing system with register store/load utilizing data packing/unpacking
JP2016146189A5 (en)
EP1512069B1 (en) An address generation unit for a processor
US20040078554A1 (en) Digital signal processor with cascaded SIMD organization
US7111155B1 (en) Digital signal processor computation core with input operand selection from operand bus for dual operations
EP2267596B1 (en) Processor core for processing instructions of different formats
US7340591B1 (en) Providing parallel operand functions using register file and extra path storage
US7107302B1 (en) Finite impulse response filter algorithm for implementation on digital signal processor having dual execution units
US6859872B1 (en) Digital signal processor computation core with pipeline having memory access stages and multiply accumulate stages positioned for efficient operation
US6820189B1 (en) Computation core executing multiple operation DSP instructions and micro-controller instructions of shorter length without performing switch operation

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20001017

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

RIN1 Information on inventor provided before grant (corrected)

Inventor name: ZHANG, LI

Inventor name: JOHN, DEEPU

Inventor name: MOTIWALA, QUAEED

Inventor name: LIN, JIAN

Inventor name: KANG, INYUP

Inventor name: JHA, SANJAY, K.

Inventor name: ZOU, QUIZHEN

Inventor name: SIH, GILBERT, C.

17Q First examination report despatched

Effective date: 20030221

RTI1 Title (correction)

Free format text: A DIGITAL SIGNAL PROCESSOR REDUCING ACCESS CONTENTION

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REF Corresponds to:

Ref document number: 69925720

Country of ref document: DE

Date of ref document: 20050714

Kind code of ref document: P

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050908

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050919

REG Reference to a national code

Ref country code: SE

Ref legal event code: TRGR

REG Reference to a national code

Ref country code: CH

Ref legal event code: NV

Representative=s name: KIRKER & CIE SA

REG Reference to a national code

Ref country code: DK

Ref legal event code: T3

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20051114

ET Fr: translation filed
PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20060309

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: SE

Payment date: 20091223

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: MC

Payment date: 20100105

Year of fee payment: 12

Ref country code: LU

Payment date: 20100323

Year of fee payment: 12

Ref country code: IE

Payment date: 20100114

Year of fee payment: 12

Ref country code: DK

Payment date: 20100208

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FI

Payment date: 20100301

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: AT

Payment date: 20100208

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20100402

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: BE

Payment date: 20100504

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: CY

Payment date: 20100115

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: CH

Payment date: 20110125

Year of fee payment: 13

BERE Be: lapsed

Owner name: *QUALCOMM INC.

Effective date: 20110331

REG Reference to a national code

Ref country code: NL

Ref legal event code: V1

Effective date: 20111001

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110331

REG Reference to a national code

Ref country code: DK

Ref legal event code: EBP

REG Reference to a national code

Ref country code: SE

Ref legal event code: EUG

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110304

Ref country code: AT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110304

Ref country code: FI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110304

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110331

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20111001

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110304

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110331

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20120331

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20120331

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110305

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20110304

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 18

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20170223

Year of fee payment: 19

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 19

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 69925720

Country of ref document: DE

Representative=s name: MAUCHER JENKINS PATENTANWAELTE & RECHTSANWAELT, DE

Ref country code: DE

Ref legal event code: R082

Ref document number: 69925720

Country of ref document: DE

Representative=s name: MAUCHER JENKINS, DE

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: IT

Payment date: 20170712

Year of fee payment: 19

Ref country code: FR

Payment date: 20170707

Year of fee payment: 19

Ref country code: DE

Payment date: 20170711

Year of fee payment: 19

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 69925720

Country of ref document: DE

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20180304

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20181002

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180304

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180304

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180331