US20090138685A1 - Processor for processing instruction set of plurality of instructions packed into single code - Google Patents

Processor for processing instruction set of plurality of instructions packed into single code Download PDF

Info

Publication number
US20090138685A1
US20090138685A1 US12/292,799 US29279908A US2009138685A1 US 20090138685 A1 US20090138685 A1 US 20090138685A1 US 29279908 A US29279908 A US 29279908A US 2009138685 A1 US2009138685 A1 US 2009138685A1
Authority
US
United States
Prior art keywords
instruction
instructions
packed
code
status
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/292,799
Inventor
Masami Makajima
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Electronics Corp
Renesas Electronics Corp
Original Assignee
Renesas Technology Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Renesas Technology Corp filed Critical Renesas Technology Corp
Priority to US12/292,799 priority Critical patent/US20090138685A1/en
Publication of US20090138685A1 publication Critical patent/US20090138685A1/en
Assigned to RENESAS ELECTRONICS CORPORATION reassignment RENESAS ELECTRONICS CORPORATION CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: NEC ELECTRONICS CORPORATION
Assigned to NEC ELECTRONICS CORPORATION reassignment NEC ELECTRONICS CORPORATION MERGER (SEE DOCUMENT FOR DETAILS). Assignors: RENESAS TECHNOLOGY CORP.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline or look ahead
    • G06F9/3836Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
    • G06F9/3853Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution of compound instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30007Arrangements for executing specific machine instructions to perform operations on data operands
    • G06F9/30036Instructions to perform operations on packed data, e.g. vector, tile or matrix operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/3017Runtime instruction translation, e.g. macros
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/3017Runtime instruction translation, e.g. macros
    • G06F9/30178Runtime instruction translation, e.g. macros of compressed or encrypted instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline or look ahead
    • G06F9/3818Decoding for concurrent execution
    • G06F9/3822Parallel decoding, e.g. parallel decode units

Definitions

  • This invention relates to a processor for processing a plurality of instructions in parallel, and in particular to a processor for processing an instruction set of a plurality of instructions packed into a single code.
  • a DSP digital signal processor
  • a DSP digital signal processor
  • ten thousand sets of data are to be arithmetically processed, for example, at least ten thousand cycles are required even if the operation on each set of data can be executed in a single machine cycle.
  • each set of data may be processed at high speed, but the time required for processing increases in proportion to the amount of data because the data processing is in series.
  • the processing performance can be improved by parallel operation.
  • a plurality of operation units are prepared and operated at the same time to process a plurality of sets of data at the same time.
  • the method called SIMD single instruction-multiple data streams
  • SIMD single instruction-multiple data streams
  • Document 1 (D. A. Patterson and J. L. Hennessy, “Computer Organization and Design”, Nikkei Business Publications) describes a method of decreasing the length of an instruction code to reduce the size of the instruction memory.
  • An object of this invention is to provide a processor in which the number of instructions capable of being executed in parallel per cycle is increased while at the same time reducing the capacity of an instruction memory.
  • a processor including an instruction memory for storing an instruction code, a converter for converting a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes, a decoder for decoding the plurality of the instruction codes converted by the converter, and a plurality of operation units for executing the operation corresponding to each of the plurality of the instruction codes in accordance with the decoding result of the decoder.
  • the converter converts a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes and the decoder decodes the plurality of the instruction codes converted by the converter, thereby causing the plurality of the operation units to execute the operations, the number of instructions that can be executed in parallel per cycle can be increased while at the same time reducing the capacity of the instruction memory.
  • FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for a processor according to a first embodiment of the invention
  • FIG. 2 is a block diagram showing a schematic configuration of the processor according to the first embodiment of the invention.
  • FIG. 3 is a flowchart for explaining the processing steps in a compiler for writing an instruction set in a conversion table 13 of the processor according to a second embodiment of the invention
  • FIG. 4 is a block diagram showing a schematic configuration of the processor according to a third embodiment of the invention.
  • FIG. 5 is a block diagram showing a schematic configuration of the processor according to a fifth embodiment of the invention.
  • FIG. 6 is a block diagram showing a schematic configuration of the processor according to a seventh embodiment of the invention.
  • FIG. 7 is a block diagram showing a schematic configuration of the processor according to an eighth embodiment of the invention.
  • FIG. 8 is a block diagram showing a schematic configuration of the processor according to a tenth embodiment of the invention.
  • FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jumping according to the tenth embodiment.
  • FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for the processor according to the first embodiment of the invention.
  • this instruction code includes a pack valid bit (V) indicating a packed instruction, a packed instruction, an operation code and an operand Mp of an instruction 1 , an operation code and an operand Ap of an instruction 2 and a mode bit (M) indicating the operation mode of the processor.
  • V pack valid bit
  • Mp mode bit
  • a “1 d@p 1 ” is an instruction for loading the data from an address indicated by a register p 1
  • an “add @p 2 ” is an instruction for adding the data stored at the address indicated by a register p 2 and the data stored in, for example, an accumulator.
  • the pack valid bit is a valid “1”
  • a plurality of instructions are selected in accordance with a 4-bit packed instruction as described later.
  • the packed instruction is “0100”, for example, four instructions including “mv r1, r5”, “inc r5, 2”, “mv r2, r3” and “inc r4, 1” are selected as shown in FIG. 1B .
  • the “mv” instruction is for data transfer between registers, and the “inc” instruction is for incrementing the contents of the register.
  • FIG. 2 is a block diagram showing a general configuration of a processor according to the first embodiment of the invention.
  • This processor includes an instruction memory 11 , an instruction queue 12 , a conversion table 13 for converting a packed instruction into a plurality of instruction codes and outputting the instruction codes, an instruction selector 14 for selectively outputting a plurality of converted instruction codes from conversion table 13 and a packed instruction from instruction queue 12 in accordance with the pack valid bit, an instruction decoder 15 , a register file 16 and ALUs 17 - 1 to 17 - 4 .
  • Instruction memory 11 stores a program to be executed by the processor.
  • Instruction queue 12 temporarily stores the instruction code fetched from instruction memory 11 .
  • the packed instruction is applied to conversion table 13 and instruction selector 14 as a 4-bit pre-conversion code.
  • the remaining 12 bits are applied directly to instruction decoder 15 .
  • Conversion table 13 includes a plurality of flip-flops 21 - 1 to 21 - n connected to a data bus 23 and a selector 22 for selecting the outputs of flip-flops 21 - 1 to 21 - n in accordance with a 4-bit pre-conversion code received from instruction queue 12 .
  • Flip-flops 21 - 1 to 21 - n each store a converted 32-bit code including a plurality of instructions through data bus 23 .
  • Data bus 23 is connected to the output port, for example, of a computer not shown, and the user can write contents of conversion table 13 from outside.
  • the flip-flops are arbitrary in number, and each corresponds to a 4-bit pre-conversion code. Specifically, in the case where the pre-conversion code is 4 bits, up to 16 flip-flops may be included.
  • the pre-conversion code is not limited to 4 bits, but the number of corresponding flip-flops can be increased by increasing the number of bits.
  • the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
  • instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15 , while in the case where the pack valid bit is invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 12 .
  • the 4-bit packed instruction is processed as a single instruction code.
  • Instruction decoder 15 includes a plurality of decode blocks 151 to 156 .
  • block 151 individually decodes the four instruction codes received from instruction selector 14 , and applies the decoding result to ALUs 17 - 1 to 17 - 4 .
  • the pack valid bit is an invalid “0”
  • the result of decoding the one instruction code received from instruction selector 14 is applied to any one of ALUs 17 - 1 to 17 - 4 and only one ALU is caused to execute the operation.
  • Blocks 152 and 153 decode the operation code of instruction 1 and the operation code of instruction 2 , respectively, and output the decoding result as a control signal to an operation array not shown.
  • Blocks 154 and 155 decode the operand portion of instruction 1 and the operand portion of instruction 2 , respectively, and output the decoding result as a control signal to an operation array not shown.
  • Block 156 decodes pack valid bit (v) and mode bit (M), and applies the decoding result as a control signal to each block of the processor.
  • Register file 16 is a group of registers for holding the data to be operated and the data constituting the result of operation.
  • ALUs 17 - 1 to 17 - 4 receive the decoding result from instruction decoder 15 , and while accessing the data held in register file 16 through common buses 18 - 1 and 18 - 2 , executes the parallel operation.
  • ALUs 17 - 1 to 17 - 4 can read the data at the same time from register file 16 through common buses 18 - 1 and 18 - 2 .
  • Register file 16 receives and stores the results of the operations of ALUs 17 - 1 to 17 - 4 through common buses 19 and 20 .
  • the decoding result applied from instruction decoder 15 to ALUs 17 - 1 to 17 - 4 includes an operand as shown in FIG. 1B .
  • conversion table 1 selects a plurality of instruction codes in accordance with the packed instruction
  • instruction selector 14 selects and outputs the converted code containing a plurality of the instruction codes output from conversion table 13
  • instruction decoder 15 decodes the plurality of the instruction codes and causes a plurality of the ALUs to execute the operations.
  • a plurality of instructions can be packed into a single instruction code and the instruction memory can be reduced in size, while at the same time increasing the number of instructions that can be executed in one cycle.
  • the operation can be executed at higher speed, and the power consumption required to fetch the instruction code can be reduced.
  • a plurality of the instructions stored in flip-flops 21 - 1 to 21 - n can be changed in accordance with the execution program.
  • the operation can be executed at higher speed.
  • the user writes an instruction set in conversion table 13 from outside.
  • a compiler implemented by the computer writes an instruction set in conversion table 13 .
  • the general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 2 . Therefore, similar or identical component parts of the configuration and functions are not described in detail again.
  • FIG. 3 is a flowchart for explaining the steps of the process executed by the compiler for writing an instruction set in conversion table 13 of the processor according to the second embodiment of the invention.
  • the compiler prepares a list of packed instructions (a plurality of instruction codes executed in parallel) from the instruction code described in a program (S 11 ), and generates a histogram according to the frequency of occurrence of the packed instruction (S 12 ).
  • the frequency of occurrence of “mv r1, r5; inc r1, 2; . . . ” is 26, the frequency of occurrence of “mv r2, r3; mv r3, r4; . . . ” is 154 and the frequency of occurrence of “inc r3, 2; inc r5, 1; . . . ” is 3.
  • the sorting is carried out (S 13 ), a plurality of the instruction codes are transferred to flip-flops 21 - 1 to 21 - n in the descending order of the frequency of occurrence (S 14 ) thereby to complete the process.
  • the plurality of the instruction codes like in the first embodiment, are transferred to conversion table 13 , for example, through the output port of the computer, not shown, connected to data bus 23 .
  • a packed instruction is assigned to the plurality of the instruction codes in accordance with which one of flip-flops 21 - 1 to 21 - n the plurality of the instruction codes are transferred to, and the program instruction is converted to a machine language using the particular packed instruction.
  • a plurality of the instruction codes determined as high in the frequency of occurrence by the compiler are packed and a plurality of the instruction codes are written in conversion table 13 .
  • a group of instructions high in compression efficiency can be packed and the instruction memory can be further reduced in size.
  • FIG. 4 is a block diagram showing a general configuration of the processor according to a third embodiment of the invention.
  • the processor according to this embodiment as compared with the processor according to the first embodiment shown in FIG. 2 , only the internal structure of conversion table 13 is different. The component parts of the configuration and functions similar or identical to those of the first embodiment, therefore, are not described in detail again.
  • Conversion table 13 includes a SRAM (static random access memory) and an associative memory 24 .
  • the 4-bit pre-conversion code output from instruction queue 12 is applied as an address of memory 24 .
  • Memory 24 has a width of 32 bits and a plurality of instruction codes are stored in each memory area.
  • a computer not shown is assumed to write a plurality of instruction codes sequentially in memory 24 through data bus 23 by outputting a plurality of the instruction codes to data bus 23 while controlling the address.
  • conversion table 13 is configured of memory 24 , and therefore, in addition to the effects described in the first embodiment, the hardware configuration can be simplified and reduced in size.
  • the processor according to a fourth embodiment of the invention is different from the processor according to the first embodiment shown in FIG. 2 only in that the data bus of the processor is connected to data bus 23 .
  • the processor writes a plurality of instructions in conversion table 13 through data bus 23 as a bus master at a predetermined timing such as resetting. In the process, until a plurality of the instructions are completely written in conversion table 13 , the processor executes the process with the pack valid bit set to invalid “0” according to the program, and upon completion of writing the plurality of the instructions in conversion table 13 , the execution of the packed instructions becomes possible.
  • the processor writes a plurality of instruction codes in conversion table 13 by itself as a bus master.
  • a plurality of instruction codes are not required to be transferred to conversion table 13 from an external source, and the processor can execute the process by itself.
  • the first embodiment as shown in FIG. 1 , is so configured that the instruction code includes packed and non-packed instructions.
  • the instruction code includes only packed instructions. Therefore, the instruction code contains no pack valid bit.
  • FIG. 5 is a block diagram showing a general configuration of the processor according to a fifth embodiment of the invention.
  • the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction selector 14 is eliminated, the output of selector 22 is applied directly to instruction decoder 15 and that the block in instruction decoder 15 for decoding the non-packed instructions is deleted.
  • the instruction code includes only the packed instruction.
  • the number of bits of the packed instruction can be increased and a greater number of instruction sets can be packed, thereby making it possible to execute a greater number of instructions in parallel.
  • a plurality of the instructions written in conversion table 13 contain an operand.
  • the instructions written in conversion table 13 contain no operand but only the operation code is packed.
  • the processor according to this embodiment has a general configuration similar to that of the first embodiment shown in FIG. 2 . Nevertheless, the operand of the instructions to be packed is included in the instruction code stored in instruction queue 12 . Instruction decoder 15 decodes the particular operand and applies the decoding result to ALUs 17 - 1 to 17 - 4 .
  • Conversion table 13 outputs a plurality of the instruction codes containing only the operation code to instruction selector 14 .
  • Instruction decoder 15 decodes the operation code of the instruction code received from instruction selector 14 and applies the decoding result to ALUs 17 - 1 to 17 - n.
  • the operand is not stored in flip-flops 21 - 1 to 21 - n in conversion table 13 .
  • the circuit size of conversion table 13 can be further reduced.
  • FIG. 6 is a block diagram showing a general configuration of the processor according to a seventh embodiment of the invention.
  • the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
  • conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
  • the packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and to instruction selector 14 on the other hand.
  • the remaining 12 bits are applied directly to instruction queue 12 .
  • the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
  • instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction queue 12 , while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction queue 12 .
  • Instruction queue 12 outputs the instruction code held therein to instruction decoder 15 and causes instruction decoder 15 to decode it at a predetermined timing.
  • conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
  • the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded is shortened.
  • the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded constitutes the critical path of the processor as a whole, the overall operating frequency of the processor can be improved.
  • FIG. 7 is a block diagram showing a general configuration of the processor according to an eighth embodiment of the invention.
  • the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction queue 12 is deleted.
  • the packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and applied to instruction selector 14 at the same time.
  • the remaining 12 bits are applied directly to instruction decoder 15 .
  • the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
  • instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15 , while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 15 .
  • instruction queue 12 is deleted, and therefore, in addition to the effects described in the first embodiment, the circuit size of the whole processor can be reduced.
  • a plurality of packed instruction codes are serially processed.
  • the general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 1 . Nevertheless, a predetermined bit of the packed instruction indicates parallel or serial execution, and instruction selector 14 changes the process by reference to the particular predetermined bit.
  • instruction selector 14 In the case where the packed instruction indicates the parallel execution, instruction selector 14 outputs the 32-bit converted code from selector 22 to instruction decoder 15 at a time. In the case where the packed instruction indicates the serial execution, on the other hand, instruction selector 14 serially outputs the 32-bit converted code from selector 22 in a plurality of cycles.
  • the instruction codes are sequentially output to instruction decoder 15 in 4 cycles, and the instruction decoder 15 decodes the 4 instructions sequentially.
  • the decoding result is continuously applied to one of ALUs 17 - 1 to 17 - 4 .
  • a plurality of packed instruction codes are processed serially.
  • the instructions serially executed can be also packed, thereby making it possible to further improve the instruction compression efficiency.
  • FIG. 8 is a block diagram showing a general configuration of the processor according to a tenth embodiment of the invention.
  • the processor according to this embodiment as compared with the processor according to the first embodiment shown in FIG. 1 , includes a jump destination instruction storage unit 31 for registering the instruction of the jump destination and a flip-flop 32 for holding the address of the jump destination.
  • jump instruction free of penalty is decoded by instruction decoder 15 during execution of a program by the processor.
  • An instruction registration control signal is output to jump destination instruction storage unit 31 .
  • the jump instruction free of penalty is assumed to include an unconditional branch instruction or a conditional branch instruction satisfying the branch conditions.
  • jump destination instruction storage unit 31 holds the particular jump destination instruction.
  • instruction decoder 15 decodes a jump instruction free of penalty
  • flip-flop 32 holds the jump destination address output from instruction decoder 15 .
  • Instruction selector 14 by reference to the jump destination address held in flip-flop 32 , stores the correspondence between the jump destination address and the jump destination instruction held in jump destination instruction storage unit 31 .
  • a jump instruction free of penalty is decoded by instruction decoder 15 , a similar process is executed if the particular jump destination instruction is not held in jump destination storage unit 31 .
  • instruction selector 14 determines that a jump destination instruction corresponding to the jump destination address held in flip-flop 32 is held in jump destination instruction storage unit 31 .
  • the particular jump destination instruction is read from jump destination instruction storage unit 31 and output to instruction decoder 15 .
  • the address of the instruction next to the jump destination instruction is applied to a program counter not shown.
  • FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jump according to the tenth embodiment of the invention.
  • the upper half of FIG. 9 shows the jump process according to this embodiment, in which the address of instruction 1 is issued in cycle T 1 and the address of instruction 2 is issued in cycle T 2 while at the same time instruction 1 is fetched.
  • Instruction 1 is a jump instruction free of penalty.
  • a jump instruction free of penalty is decoded by instruction decoder 15 , and instruction selector 14 , by referring to the jump destination address held in flip-flop 32 , reads the jump destination instruction from jump destination instruction storage unit 31 and outputs it to instruction decoder 15 .
  • the address of the instruction next to the jump destination instruction is applied to a program counter not shown, and this address is issued.
  • cycle T 4 the jump destination instruction is executed, the instruction next to the jump destination instruction is fetched, while at the same time the address of the next instruction but one is issued.
  • cycle T 5 and subsequent cycles a similar pipelining process is executed.
  • the instruction next to the jump instruction is stored in instruction queue 12 . Even in the case where the branch conditions of the conditional branch instruction fail to be met, therefore, the pipelining process similar to that shown in the upper half of FIG. 9 can be carried out by applying the particular instruction to instruction decoder 15 .
  • the lower half of FIG. 9 indicates the conventional jump process.
  • cycle T 1 the address of instruction 1 is issued
  • cycle T 2 the address of instruction 2 is issued while at the same time instruction 1 is fetched.
  • Instruction 1 is a jump instruction requiring the branch prediction.
  • cycle T 3 a jump instruction is decoded by instruction decoder 15 .
  • instruction decoder 15 it is assumed that the branch prediction fails and the instruction of the branch destination is required to be fetched again.
  • the instruction stored in instruction queue 12 is invalidated and processed as a NOP (non-operation) instruction.
  • cycle T 4 the NOP instruction is executed and the jump destination instruction is fetched, while at the same time the address of the instruction next to the jump destination instruction is issued.
  • cycle T 5 and subsequent cycles a similar pipelining process is executed. In this way, a one-cycle delay occurs as compared with the jump instruction free of penalty.
  • the jump destination instruction may be required to be fetched, as shown in the lower half of FIG. 9 , unless the jump instruction is held in jump destination instruction storage unit 31 . In such a case, a one-cycle delay occurs.
  • a failure of the branch prediction results in the requirement of a repeated instruction fetch, and therefore a jump penalty occurs.
  • no jump penalty occurs as long as a jump destination instruction is held in jump destination instruction storage unit 31 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Executing Machine-Instructions (AREA)
  • Advance Control (AREA)

Abstract

A conversion table converts a packed instruction (pre-conversion code) contained in the instruction code fetched from an instruction memory into a plurality of instruction codes (converted codes). An instruction decoder decodes the plurality of the instruction codes converted by a conversion table. A plurality of ALUs perform the operation in accordance with the decoding result of the instruction decoder. Therefore, the number of instructions that can be executed in parallel per cycle may be increased while at the same time the capacity of the instruction memory is reduced.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • This invention relates to a processor for processing a plurality of instructions in parallel, and in particular to a processor for processing an instruction set of a plurality of instructions packed into a single code.
  • 2. Description of the Background Art
  • In recent years, with the spread of portable terminal devices, the digital signal processing for processing a great amount of data such as voices and images at high speed has become increasingly important. A DSP (digital signal processor) is typically used as a semiconductor device exclusive to such digital signal processing. However, in the case where an amount of data to be processed is enormous, it is difficult to improve the performance dramatically even with the use of an exclusive DSP. Assuming that ten thousand sets of data are to be arithmetically processed, for example, at least ten thousand cycles are required even if the operation on each set of data can be executed in a single machine cycle. In other words, each set of data may be processed at high speed, but the time required for processing increases in proportion to the amount of data because the data processing is in series.
  • In the case where an amount of data to be processed is large, the processing performance can be improved by parallel operation. Specifically, a plurality of operation units are prepared and operated at the same time to process a plurality of sets of data at the same time. In the case where the same operation is performed on a plurality of sets of data, the method called SIMD (single instruction-multiple data streams) can be employed to reduce the area of the operation unit while maintaining a high parallel performance. Specifically, while a plurality of data processors are prepared, a high performance with a small area can be exhibited by providing a common control unit for interpreting an instruction and controlling the process.
  • Document 1 (D. A. Patterson and J. L. Hennessy, “Computer Organization and Design”, Nikkei Business Publications) describes a method of decreasing the length of an instruction code to reduce the size of the instruction memory.
  • Document 2 (Akira Nakamori, “Introduction to Microprocessor Architecture”, CQ Publishing), on the other hand, describes a method in which a plurality of slots are formed in one instruction format for parallel execution by VLIW (very long instruction word) in order to increase the number of instructions that can be executed per cycle.
  • However, reducing the size of the instruction memory as described in Document 1 and increasing the number of instructions that can be executed per cycle as described in Document 2 are in the relation of so-called tradeoff.
  • Specifically, according to the method described in Document 1, it is possible to reduce the size of the instruction memory, while the fact that a plurality of instructions are processed in series poses a problem that a number of cycles are required to execute the instructions. Taking an example where the instruction code length is 16 bits, the instruction length is short but four cycles are required to execute four instructions.
  • According to the method described in Document 2, on the other hand, more instructions may be executed per cycle, while the instruction length is increased and so is the size of the instruction memory. In the case where four slots of 16 bits are provided, for example, four instructions can be executed at the same time in a single cycle at the sacrifice of an extended instruction code length of 64 bits.
  • SUMMARY OF THE INVENTION
  • An object of this invention is to provide a processor in which the number of instructions capable of being executed in parallel per cycle is increased while at the same time reducing the capacity of an instruction memory.
  • According to one aspect of the invention, there is provided a processor including an instruction memory for storing an instruction code, a converter for converting a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes, a decoder for decoding the plurality of the instruction codes converted by the converter, and a plurality of operation units for executing the operation corresponding to each of the plurality of the instruction codes in accordance with the decoding result of the decoder.
  • In view of the fact that the converter converts a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes and the decoder decodes the plurality of the instruction codes converted by the converter, thereby causing the plurality of the operation units to execute the operations, the number of instructions that can be executed in parallel per cycle can be increased while at the same time reducing the capacity of the instruction memory.
  • The foregoing and other objects, features, aspects and advantages of the present invention will become more apparent from the following detailed description of the present invention when taken in conjunction with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for a processor according to a first embodiment of the invention;
  • FIG. 2 is a block diagram showing a schematic configuration of the processor according to the first embodiment of the invention;
  • FIG. 3 is a flowchart for explaining the processing steps in a compiler for writing an instruction set in a conversion table 13 of the processor according to a second embodiment of the invention;
  • FIG. 4 is a block diagram showing a schematic configuration of the processor according to a third embodiment of the invention;
  • FIG. 5 is a block diagram showing a schematic configuration of the processor according to a fifth embodiment of the invention;
  • FIG. 6 is a block diagram showing a schematic configuration of the processor according to a seventh embodiment of the invention;
  • FIG. 7 is a block diagram showing a schematic configuration of the processor according to an eighth embodiment of the invention;
  • FIG. 8 is a block diagram showing a schematic configuration of the processor according to a tenth embodiment of the invention; and
  • FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jumping according to the tenth embodiment.
  • DESCRIPTION OF THE PREFERRED EMBODIMENTS First Embodiment
  • FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for the processor according to the first embodiment of the invention. As shown in FIG. 1A, this instruction code includes a pack valid bit (V) indicating a packed instruction, a packed instruction, an operation code and an operand Mp of an instruction 1, an operation code and an operand Ap of an instruction 2 and a mode bit (M) indicating the operation mode of the processor. A “1 d@p1” is an instruction for loading the data from an address indicated by a register p1, and an “add @p2” is an instruction for adding the data stored at the address indicated by a register p2 and the data stored in, for example, an accumulator.
  • In the case where the pack valid bit is a valid “1”, a plurality of instructions are selected in accordance with a 4-bit packed instruction as described later. In the case where the packed instruction is “0100”, for example, four instructions including “mv r1, r5”, “inc r5, 2”, “mv r2, r3” and “inc r4, 1” are selected as shown in FIG. 1B. The “mv” instruction is for data transfer between registers, and the “inc” instruction is for incrementing the contents of the register.
  • In the case where the pack valid bit is invalid “0”, a 4-bit packed instruction is processed as one instruction.
  • FIG. 2 is a block diagram showing a general configuration of a processor according to the first embodiment of the invention. This processor includes an instruction memory 11, an instruction queue 12, a conversion table 13 for converting a packed instruction into a plurality of instruction codes and outputting the instruction codes, an instruction selector 14 for selectively outputting a plurality of converted instruction codes from conversion table 13 and a packed instruction from instruction queue 12 in accordance with the pack valid bit, an instruction decoder 15, a register file 16 and ALUs 17-1 to 17-4.
  • Instruction memory 11 stores a program to be executed by the processor. Instruction queue 12 temporarily stores the instruction code fetched from instruction memory 11. Of the 16-bit instruction code stored in instruction queue 12, the packed instruction is applied to conversion table 13 and instruction selector 14 as a 4-bit pre-conversion code. The remaining 12 bits are applied directly to instruction decoder 15.
  • Conversion table 13 includes a plurality of flip-flops 21-1 to 21-n connected to a data bus 23 and a selector 22 for selecting the outputs of flip-flops 21-1 to 21-n in accordance with a 4-bit pre-conversion code received from instruction queue 12.
  • Flip-flops 21-1 to 21-n each store a converted 32-bit code including a plurality of instructions through data bus 23. Data bus 23 is connected to the output port, for example, of a computer not shown, and the user can write contents of conversion table 13 from outside. The flip-flops are arbitrary in number, and each corresponds to a 4-bit pre-conversion code. Specifically, in the case where the pre-conversion code is 4 bits, up to 16 flip-flops may be included. The pre-conversion code is not limited to 4 bits, but the number of corresponding flip-flops can be increased by increasing the number of bits.
  • The 32-bit converted code selected by selector 22 is applied to instruction selector 14. In the case where the pack valid bit is a valid “1”, instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15, while in the case where the pack valid bit is invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 12. In this case, the 4-bit packed instruction is processed as a single instruction code.
  • Instruction decoder 15 includes a plurality of decode blocks 151 to 156. In the case where the pack valid bit is a valid “1”, block 151 individually decodes the four instruction codes received from instruction selector 14, and applies the decoding result to ALUs 17-1 to 17-4. In the case where the pack valid bit is an invalid “0”, on the other hand, the result of decoding the one instruction code received from instruction selector 14 is applied to any one of ALUs 17-1 to 17-4 and only one ALU is caused to execute the operation.
  • Blocks 152 and 153 decode the operation code of instruction 1 and the operation code of instruction 2, respectively, and output the decoding result as a control signal to an operation array not shown. Blocks 154 and 155, on the other hand, decode the operand portion of instruction 1 and the operand portion of instruction 2, respectively, and output the decoding result as a control signal to an operation array not shown.
  • Block 156 decodes pack valid bit (v) and mode bit (M), and applies the decoding result as a control signal to each block of the processor.
  • Register file 16 is a group of registers for holding the data to be operated and the data constituting the result of operation. ALUs 17-1 to 17-4 receive the decoding result from instruction decoder 15, and while accessing the data held in register file 16 through common buses 18-1 and 18-2, executes the parallel operation. ALUs 17-1 to 17-4 can read the data at the same time from register file 16 through common buses 18-1 and 18-2.
  • Register file 16 receives and stores the results of the operations of ALUs 17-1 to 17-4 through common buses 19 and 20. The decoding result applied from instruction decoder 15 to ALUs 17-1 to 17-4 includes an operand as shown in FIG. 1B.
  • As explained above, with the processor according to this embodiment, in the case where the pack valid bit is a valid “1”, conversion table 1 selects a plurality of instruction codes in accordance with the packed instruction, instruction selector 14 selects and outputs the converted code containing a plurality of the instruction codes output from conversion table 13, and instruction decoder 15 decodes the plurality of the instruction codes and causes a plurality of the ALUs to execute the operations. In this way, a plurality of instructions can be packed into a single instruction code and the instruction memory can be reduced in size, while at the same time increasing the number of instructions that can be executed in one cycle.
  • Also, the operation can be executed at higher speed, and the power consumption required to fetch the instruction code can be reduced.
  • Also, a plurality of the instructions stored in flip-flops 21-1 to 21-n can be changed in accordance with the execution program. By incorporating an instruction set suitable for the program in flip-flops 21-1 to 21-n, therefore, the operation can be executed at higher speed.
  • Second Embodiment
  • In the first embodiment of the invention, the user writes an instruction set in conversion table 13 from outside. According to the second embodiment, in contrast, a compiler implemented by the computer writes an instruction set in conversion table 13.
  • The general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 2. Therefore, similar or identical component parts of the configuration and functions are not described in detail again.
  • FIG. 3 is a flowchart for explaining the steps of the process executed by the compiler for writing an instruction set in conversion table 13 of the processor according to the second embodiment of the invention. First, the compiler prepares a list of packed instructions (a plurality of instruction codes executed in parallel) from the instruction code described in a program (S11), and generates a histogram according to the frequency of occurrence of the packed instruction (S12).
  • As indicated in step S12 of FIG. 3, the frequency of occurrence of “mv r1, r5; inc r1, 2; . . . ” is 26, the frequency of occurrence of “mv r2, r3; mv r3, r4; . . . ” is 154 and the frequency of occurrence of “inc r3, 2; inc r5, 1; . . . ” is 3.
  • Next, in accordance with the frequency of occurrence of the packed instruction, the sorting is carried out (S13), a plurality of the instruction codes are transferred to flip-flops 21-1 to 21-n in the descending order of the frequency of occurrence (S14) thereby to complete the process. The plurality of the instruction codes, like in the first embodiment, are transferred to conversion table 13, for example, through the output port of the computer, not shown, connected to data bus 23.
  • In the compiler, a packed instruction is assigned to the plurality of the instruction codes in accordance with which one of flip-flops 21-1 to 21-n the plurality of the instruction codes are transferred to, and the program instruction is converted to a machine language using the particular packed instruction.
  • As explained above, in the processor according to this embodiment, a plurality of the instruction codes determined as high in the frequency of occurrence by the compiler are packed and a plurality of the instruction codes are written in conversion table 13. In addition to the effects explained with reference to the first embodiment, therefore, a group of instructions high in compression efficiency can be packed and the instruction memory can be further reduced in size.
  • Third Embodiment
  • FIG. 4 is a block diagram showing a general configuration of the processor according to a third embodiment of the invention. In the processor according to this embodiment, as compared with the processor according to the first embodiment shown in FIG. 2, only the internal structure of conversion table 13 is different. The component parts of the configuration and functions similar or identical to those of the first embodiment, therefore, are not described in detail again.
  • Conversion table 13 includes a SRAM (static random access memory) and an associative memory 24. The 4-bit pre-conversion code output from instruction queue 12 is applied as an address of memory 24. Memory 24 has a width of 32 bits and a plurality of instruction codes are stored in each memory area.
  • A computer not shown is assumed to write a plurality of instruction codes sequentially in memory 24 through data bus 23 by outputting a plurality of the instruction codes to data bus 23 while controlling the address.
  • As explained above, in the processor according to this embodiment, conversion table 13 is configured of memory 24, and therefore, in addition to the effects described in the first embodiment, the hardware configuration can be simplified and reduced in size.
  • Fourth Embodiment
  • The processor according to a fourth embodiment of the invention is different from the processor according to the first embodiment shown in FIG. 2 only in that the data bus of the processor is connected to data bus 23. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again.
  • The processor writes a plurality of instructions in conversion table 13 through data bus 23 as a bus master at a predetermined timing such as resetting. In the process, until a plurality of the instructions are completely written in conversion table 13, the processor executes the process with the pack valid bit set to invalid “0” according to the program, and upon completion of writing the plurality of the instructions in conversion table 13, the execution of the packed instructions becomes possible.
  • As explained above, the processor according to this embodiment writes a plurality of instruction codes in conversion table 13 by itself as a bus master. In addition to the effects explained in the first embodiment, therefore, a plurality of instruction codes are not required to be transferred to conversion table 13 from an external source, and the processor can execute the process by itself.
  • Fifth Embodiment
  • The first embodiment, as shown in FIG. 1, is so configured that the instruction code includes packed and non-packed instructions. According to this embodiment, on the other hand, the instruction code includes only packed instructions. Therefore, the instruction code contains no pack valid bit.
  • FIG. 5 is a block diagram showing a general configuration of the processor according to a fifth embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction selector 14 is eliminated, the output of selector 22 is applied directly to instruction decoder 15 and that the block in instruction decoder 15 for decoding the non-packed instructions is deleted. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again.
  • As explained above, in the processor according to this embodiment, the instruction code includes only the packed instruction. In addition to the effects described in the first embodiment, therefore, the number of bits of the packed instruction can be increased and a greater number of instruction sets can be packed, thereby making it possible to execute a greater number of instructions in parallel.
  • Sixth Embodiment
  • In the first embodiment, a plurality of the instructions written in conversion table 13 contain an operand. According to the sixth embodiment, on the other hand, the instructions written in conversion table 13 contain no operand but only the operation code is packed.
  • The processor according to this embodiment has a general configuration similar to that of the first embodiment shown in FIG. 2. Nevertheless, the operand of the instructions to be packed is included in the instruction code stored in instruction queue 12. Instruction decoder 15 decodes the particular operand and applies the decoding result to ALUs 17-1 to 17-4.
  • Conversion table 13 outputs a plurality of the instruction codes containing only the operation code to instruction selector 14. Instruction decoder 15 decodes the operation code of the instruction code received from instruction selector 14 and applies the decoding result to ALUs 17-1 to 17-n.
  • As explained above, in the processor according to this embodiment, the operand is not stored in flip-flops 21-1 to 21-n in conversion table 13. In addition to the effects explained in the first embodiment, therefore, the circuit size of conversion table 13 can be further reduced.
  • Seventh Embodiment
  • FIG. 6 is a block diagram showing a general configuration of the processor according to a seventh embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that conversion table 13 is interposed between instruction memory 11 and instruction queue 12. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again.
  • The packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and to instruction selector 14 on the other hand. The remaining 12 bits are applied directly to instruction queue 12.
  • The 32-bit converted code selected by selector 22 is applied to instruction selector 14. In the case where the pack valid bit is a valid “1”, instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction queue 12, while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction queue 12.
  • Instruction queue 12 outputs the instruction code held therein to instruction decoder 15 and causes instruction decoder 15 to decode it at a predetermined timing.
  • As explained above, in the processor according to this embodiment, conversion table 13 is interposed between instruction memory 11 and instruction queue 12. In addition to the effects explained in the first embodiment, therefore, the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded is shortened. In the case where the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded constitutes the critical path of the processor as a whole, the overall operating frequency of the processor can be improved.
  • Eighth Embodiment
  • FIG. 7 is a block diagram showing a general configuration of the processor according to an eighth embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction queue 12 is deleted. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again.
  • The packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and applied to instruction selector 14 at the same time. The remaining 12 bits are applied directly to instruction decoder 15.
  • The 32-bit converted code selected by selector 22 is applied to instruction selector 14. In the case where the pack valid bit is a valid “1”, instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15, while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 15.
  • As explained above, in the processor according to this embodiment, instruction queue 12 is deleted, and therefore, in addition to the effects described in the first embodiment, the circuit size of the whole processor can be reduced.
  • Ninth Embodiment
  • According to the ninth embodiment, unlike in the first embodiment in which a plurality of packed instruction codes are processed in parallel, a plurality of packed instruction codes are serially processed.
  • The general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 1. Nevertheless, a predetermined bit of the packed instruction indicates parallel or serial execution, and instruction selector 14 changes the process by reference to the particular predetermined bit.
  • In the case where the packed instruction indicates the parallel execution, instruction selector 14 outputs the 32-bit converted code from selector 22 to instruction decoder 15 at a time. In the case where the packed instruction indicates the serial execution, on the other hand, instruction selector 14 serially outputs the 32-bit converted code from selector 22 in a plurality of cycles.
  • In the case where 4 instruction codes are output from selector 22, for example, the instruction codes are sequentially output to instruction decoder 15 in 4 cycles, and the instruction decoder 15 decodes the 4 instructions sequentially. The decoding result is continuously applied to one of ALUs 17-1 to 17-4.
  • As explained above, in the processor according to this embodiment, a plurality of packed instruction codes are processed serially. In addition to the effects explained in the first embodiment, therefore, the instructions serially executed can be also packed, thereby making it possible to further improve the instruction compression efficiency.
  • Tenth Embodiment
  • FIG. 8 is a block diagram showing a general configuration of the processor according to a tenth embodiment of the invention. The processor according to this embodiment, as compared with the processor according to the first embodiment shown in FIG. 1, includes a jump destination instruction storage unit 31 for registering the instruction of the jump destination and a flip-flop 32 for holding the address of the jump destination.
  • Assume that the jump instruction free of penalty is decoded by instruction decoder 15 during execution of a program by the processor. An instruction registration control signal is output to jump destination instruction storage unit 31. The jump instruction free of penalty is assumed to include an unconditional branch instruction or a conditional branch instruction satisfying the branch conditions.
  • In the case where a jump destination instruction is stored in instruction queue 12, jump destination instruction storage unit 31 holds the particular jump destination instruction. In the case where instruction decoder 15 decodes a jump instruction free of penalty, flip-flop 32 holds the jump destination address output from instruction decoder 15. Instruction selector 14, by reference to the jump destination address held in flip-flop 32, stores the correspondence between the jump destination address and the jump destination instruction held in jump destination instruction storage unit 31. In the case where a jump instruction free of penalty is decoded by instruction decoder 15, a similar process is executed if the particular jump destination instruction is not held in jump destination storage unit 31.
  • Next, assume that in the case where instruction decoder 15 decodes a jump instruction free of penalty, instruction selector 14 determines that a jump destination instruction corresponding to the jump destination address held in flip-flop 32 is held in jump destination instruction storage unit 31. The particular jump destination instruction is read from jump destination instruction storage unit 31 and output to instruction decoder 15. In the process, the address of the instruction next to the jump destination instruction is applied to a program counter not shown.
  • FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jump according to the tenth embodiment of the invention. The upper half of FIG. 9 shows the jump process according to this embodiment, in which the address of instruction 1 is issued in cycle T1 and the address of instruction 2 is issued in cycle T2 while at the same time instruction 1 is fetched. Instruction 1 is a jump instruction free of penalty.
  • In cycle T3, a jump instruction free of penalty is decoded by instruction decoder 15, and instruction selector 14, by referring to the jump destination address held in flip-flop 32, reads the jump destination instruction from jump destination instruction storage unit 31 and outputs it to instruction decoder 15. In the process, the address of the instruction next to the jump destination instruction is applied to a program counter not shown, and this address is issued.
  • In cycle T4, the jump destination instruction is executed, the instruction next to the jump destination instruction is fetched, while at the same time the address of the next instruction but one is issued. In cycle T5 and subsequent cycles, a similar pipelining process is executed.
  • The instruction next to the jump instruction is stored in instruction queue 12. Even in the case where the branch conditions of the conditional branch instruction fail to be met, therefore, the pipelining process similar to that shown in the upper half of FIG. 9 can be carried out by applying the particular instruction to instruction decoder 15.
  • The lower half of FIG. 9, on the other hand, indicates the conventional jump process. In cycle T1, the address of instruction 1 is issued, and in cycle T2, the address of instruction 2 is issued while at the same time instruction 1 is fetched. Instruction 1 is a jump instruction requiring the branch prediction.
  • In cycle T3, a jump instruction is decoded by instruction decoder 15. In this case, it is assumed that the branch prediction fails and the instruction of the branch destination is required to be fetched again. In this cycle, the instruction stored in instruction queue 12 is invalidated and processed as a NOP (non-operation) instruction.
  • In cycle T4, the NOP instruction is executed and the jump destination instruction is fetched, while at the same time the address of the instruction next to the jump destination instruction is issued. In cycle T5 and subsequent cycles, a similar pipelining process is executed. In this way, a one-cycle delay occurs as compared with the jump instruction free of penalty.
  • Even in the case of a jump instruction free of penalty, the jump destination instruction may be required to be fetched, as shown in the lower half of FIG. 9, unless the jump instruction is held in jump destination instruction storage unit 31. In such a case, a one-cycle delay occurs.
  • As explained above, with the processor according to this embodiment, in the case where a jump destination instruction is held in jump destination instruction storage unit 31, the particular jump destination instruction is applied to instruction decoder 15. Thus, the penalty in the cycle of jump can be eliminated, and the delay due to the irregularities of the pipeline is prevented.
  • Specifically, according to the prior art including the branch prediction, a failure of the branch prediction results in the requirement of a repeated instruction fetch, and therefore a jump penalty occurs. In the processor according to this embodiment, on the other hand, no jump penalty occurs as long as a jump destination instruction is held in jump destination instruction storage unit 31.
  • Although the present invention has been described and illustrated in detail, it is clearly understood that the same is by way of illustration and example only and is not to be taken by way of limitation, the spirit and scope of the present invention being limited only by the terms of the appended claims.

Claims (5)

1-13. (canceled)
14. A data processor comprising:
an instruction fetch unit;
an instruction decode unit;
an instruction execute unit; and
an instruction interpretation unit;
wherein the instruction fetch unit fetches an instruction from a memory, the instruction having at least a first part and a second part,
wherein the instruction interpretation unit interprets a binary pattern in the first part of the instruction as an instruction package containing a plurality of instructions by referring to a converting table when the second part of the instruction indicates a first status, or as a single instruction when the second part of the instruction indicates a second status,
wherein the instruction decode unit decodes each of the plurality of instructions of the instruction package interpreted by the instruction interpretation unit when the second part of the instruction indicates the first status, or decodes the single instruction when the second part of the instruction indicates the second status,
wherein the instruction execution unit executes the plurality of instructions or the single instruction decoded by the instruction decode unit, and
wherein the converting table is supplied from the outside data processor.
15. A data processor according to claim 14,
wherein the instruction execute unit executes the plurality of instructions whether in serial or in parallel in accordance with a third part of the instruction.
16. A data processor being adapted to perform each functions following:
an instruction fetching function that fetches an instruction from a memory,
an instruction interpreting function that interprets a binary pattern in a first part of the instruction as a plurality of instructions by referring a converting table when a second part of the instruction indicates a first status, or as a single instruction when the second part of the instruction indicates a second status,
an instruction decoding function that decodes each of the plurality of instructions when the second part of the instruction indicates the first status, or decodes the single instruction when the second part of the instruction indicates the second status,
an instruction executing function executes each of the plurality of instructions or the single instruction, and
a reading function that reads the converting table from outside the data processor.
17. A data processor according to claim 16,
wherein the instruction execution function that executes the plurality of instructions whether in serial or in parallel in accordance with a third part of the instruction.
US12/292,799 2005-09-29 2008-11-26 Processor for processing instruction set of plurality of instructions packed into single code Abandoned US20090138685A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/292,799 US20090138685A1 (en) 2005-09-29 2008-11-26 Processor for processing instruction set of plurality of instructions packed into single code

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2005-284145 2005-09-29
JP2005284145A JP4986431B2 (en) 2005-09-29 2005-09-29 Processor
US11/520,616 US7473293B2 (en) 2005-09-29 2006-09-14 Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator
US12/292,799 US20090138685A1 (en) 2005-09-29 2008-11-26 Processor for processing instruction set of plurality of instructions packed into single code

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/520,616 Continuation US7473293B2 (en) 2005-09-29 2006-09-14 Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator

Publications (1)

Publication Number Publication Date
US20090138685A1 true US20090138685A1 (en) 2009-05-28

Family

ID=37895576

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/520,616 Active 2027-01-25 US7473293B2 (en) 2005-09-29 2006-09-14 Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator
US12/292,799 Abandoned US20090138685A1 (en) 2005-09-29 2008-11-26 Processor for processing instruction set of plurality of instructions packed into single code

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/520,616 Active 2027-01-25 US7473293B2 (en) 2005-09-29 2006-09-14 Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator

Country Status (2)

Country Link
US (2) US7473293B2 (en)
JP (1) JP4986431B2 (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4444305B2 (en) * 2007-03-28 2010-03-31 株式会社東芝 Semiconductor device
JP2009288967A (en) * 2008-05-28 2009-12-10 Fujitsu Ten Ltd Signal processing apparatus, display control device, and register setting method for signal processing apparatus
US20110283089A1 (en) * 2009-01-28 2011-11-17 Harshal Ingale modularized micro processor design
US8970608B2 (en) * 2010-04-05 2015-03-03 Nvidia Corporation State objects for specifying dynamic state
US9378019B2 (en) 2011-04-07 2016-06-28 Via Technologies, Inc. Conditional load instructions in an out-of-order execution microprocessor
US9645822B2 (en) 2011-04-07 2017-05-09 Via Technologies, Inc Conditional store instructions in an out-of-order execution microprocessor
US9032189B2 (en) 2011-04-07 2015-05-12 Via Technologies, Inc. Efficient conditional ALU instruction in read-port limited register file microprocessor
US9336180B2 (en) 2011-04-07 2016-05-10 Via Technologies, Inc. Microprocessor that makes 64-bit general purpose registers available in MSR address space while operating in non-64-bit mode
US9043580B2 (en) 2011-04-07 2015-05-26 Via Technologies, Inc. Accessing model specific registers (MSR) with different sets of distinct microinstructions for instructions of different instruction set architecture (ISA)
US8880851B2 (en) 2011-04-07 2014-11-04 Via Technologies, Inc. Microprocessor that performs X86 ISA and arm ISA machine language program instructions by hardware translation into microinstructions executed by common execution pipeline
US9274795B2 (en) 2011-04-07 2016-03-01 Via Technologies, Inc. Conditional non-branch instruction prediction
US9141389B2 (en) 2011-04-07 2015-09-22 Via Technologies, Inc. Heterogeneous ISA microprocessor with shared hardware ISA registers
US9317288B2 (en) 2011-04-07 2016-04-19 Via Technologies, Inc. Multi-core microprocessor that performs x86 ISA and ARM ISA machine language program instructions by hardware translation into microinstructions executed by common execution pipeline
US9128701B2 (en) 2011-04-07 2015-09-08 Via Technologies, Inc. Generating constant for microinstructions from modified immediate field during instruction translation
US9176733B2 (en) 2011-04-07 2015-11-03 Via Technologies, Inc. Load multiple and store multiple instructions in a microprocessor that emulates banked registers
US8880857B2 (en) 2011-04-07 2014-11-04 Via Technologies, Inc. Conditional ALU instruction pre-shift-generated carry flag propagation between microinstructions in read-port limited register file microprocessor
US9898291B2 (en) 2011-04-07 2018-02-20 Via Technologies, Inc. Microprocessor with arm and X86 instruction length decoders
US8924695B2 (en) * 2011-04-07 2014-12-30 Via Technologies, Inc. Conditional ALU instruction condition satisfaction propagation between microinstructions in read-port limited register file microprocessor
US9146742B2 (en) 2011-04-07 2015-09-29 Via Technologies, Inc. Heterogeneous ISA microprocessor that preserves non-ISA-specific configuration state when reset to different ISA
US9244686B2 (en) 2011-04-07 2016-01-26 Via Technologies, Inc. Microprocessor that translates conditional load/store instructions into variable number of microinstructions
US9292470B2 (en) 2011-04-07 2016-03-22 Via Technologies, Inc. Microprocessor that enables ARM ISA program to access 64-bit general purpose registers written by x86 ISA program
US9672041B2 (en) 2013-08-01 2017-06-06 Andes Technology Corporation Method for compressing variable-length instructions including PC-relative instructions and processor for executing compressed instructions using an instruction table
JP2019095952A (en) * 2017-11-21 2019-06-20 ソニーセミコンダクタソリューションズ株式会社 Processor, information processing device and processing method
JP7506718B2 (en) * 2022-08-05 2024-06-26 たけおかラボ株式会社 Processor and program for controlling pipeline processing based on jump instructions

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5632024A (en) * 1993-06-08 1997-05-20 Hitachi, Ltd. Microcomputer executing compressed program and generating compressed branch addresses
US5636352A (en) * 1994-12-16 1997-06-03 International Business Machines Corporation Method and apparatus for utilizing condensed instructions
US20020120830A1 (en) * 2001-02-23 2002-08-29 Masahito Matsuo Data processor assigning the same operation code to multiple operations
US20030033504A1 (en) * 2001-08-07 2003-02-13 Hiromichi Yamada Micro-controller for reading out compressed instruction code and program memory for compressing instruction code and storing therein

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10116229A (en) * 1996-10-09 1998-05-06 Toshiba Corp Data processor
JP3585800B2 (en) * 1999-01-13 2004-11-04 株式会社東芝 Information processing equipment
JP2002215387A (en) * 2001-01-22 2002-08-02 Mitsubishi Electric Corp Data processor provided with instruction translator, and memory interface device
KR20040064713A (en) * 2001-11-26 2004-07-19 코닌클리즈케 필립스 일렉트로닉스 엔.브이. Wlim architecture with power down instruction
JP2004220188A (en) * 2003-01-10 2004-08-05 Sony Corp Data processor
JP4761808B2 (en) * 2005-04-11 2011-08-31 株式会社東芝 Microprocessor and control method thereof

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5632024A (en) * 1993-06-08 1997-05-20 Hitachi, Ltd. Microcomputer executing compressed program and generating compressed branch addresses
US5636352A (en) * 1994-12-16 1997-06-03 International Business Machines Corporation Method and apparatus for utilizing condensed instructions
US20020120830A1 (en) * 2001-02-23 2002-08-29 Masahito Matsuo Data processor assigning the same operation code to multiple operations
US20030033504A1 (en) * 2001-08-07 2003-02-13 Hiromichi Yamada Micro-controller for reading out compressed instruction code and program memory for compressing instruction code and storing therein

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Definition of "bit stream" retreived from google.com Dictionary on 11/7/2011. *
Definition of "stream" retreived from google.com Dictionary on 11/7/2011. *

Also Published As

Publication number Publication date
US7473293B2 (en) 2009-01-06
JP2007094813A (en) 2007-04-12
JP4986431B2 (en) 2012-07-25
US20070074010A1 (en) 2007-03-29

Similar Documents

Publication Publication Date Title
US7473293B2 (en) Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator
JP5313228B2 (en) System, method and software for preloading instructions from a variable length instruction set with appropriate predecoding
US20070106889A1 (en) Configurable instruction sequence generation
US20120204008A1 (en) Processor with a Hybrid Instruction Queue with Instruction Elaboration Between Sections
US7546442B1 (en) Fixed length memory to memory arithmetic and architecture for direct memory access using fixed length instructions
US9317285B2 (en) Instruction set architecture mode dependent sub-size access of register with associated status indication
JP2006313422A (en) Calculation processing device and method for executing data transfer processing
WO2008112207A2 (en) Software programmable timing architecture
US20230084523A1 (en) Data Processing Method and Device, and Storage Medium
US20050138327A1 (en) VLIW digital signal processor for achieving improved binary translation
US20200326940A1 (en) Data loading and storage instruction processing method and device
US8635434B2 (en) Mathematical operation processing apparatus for performing high speed mathematical operations
CN112559037B (en) Instruction execution method, unit, device and system
JP5335440B2 (en) Early conditional selection of operands
CN112540792A (en) Instruction processing method and device
JP2009524167A5 (en)
US9361109B2 (en) System and method to evaluate a data value as an instruction
JP4502532B2 (en) Data processing device
US7543135B2 (en) Processor and method for selectively processing instruction to be read using instruction code already in pipeline or already stored in prefetch buffer
CN116339832A (en) Data processing device, method and processor
US11210091B2 (en) Method and apparatus for processing data splicing instruction
US6829700B2 (en) Circuit and method for supporting misaligned accesses in the presence of speculative load instructions
US6988121B1 (en) Efficient implementation of multiprecision arithmetic
CN112130899A (en) Stack computer
US20050138331A1 (en) Direct memory access unit with instruction pre-decoder

Legal Events

Date Code Title Description
AS Assignment

Owner name: RENESAS ELECTRONICS CORPORATION, JAPAN

Free format text: CHANGE OF NAME;ASSIGNOR:NEC ELECTRONICS CORPORATION;REEL/FRAME:024990/0059

Effective date: 20100401

Owner name: NEC ELECTRONICS CORPORATION, JAPAN

Free format text: MERGER;ASSIGNOR:RENESAS TECHNOLOGY CORP.;REEL/FRAME:024990/0098

Effective date: 20100401

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION