US20090138685A1 - Processor for processing instruction set of plurality of instructions packed into single code - Google Patents
Processor for processing instruction set of plurality of instructions packed into single code Download PDFInfo
- Publication number
- US20090138685A1 US20090138685A1 US12/292,799 US29279908A US2009138685A1 US 20090138685 A1 US20090138685 A1 US 20090138685A1 US 29279908 A US29279908 A US 29279908A US 2009138685 A1 US2009138685 A1 US 2009138685A1
- Authority
- US
- United States
- Prior art keywords
- instruction
- instructions
- packed
- code
- status
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000006870 function Effects 0.000 claims description 13
- 238000006243 chemical reaction Methods 0.000 abstract description 41
- 238000000034 method Methods 0.000 description 23
- 238000010586 diagram Methods 0.000 description 16
- 230000000694 effects Effects 0.000 description 8
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3836—Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
- G06F9/3853—Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution of compound instructions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30003—Arrangements for executing specific machine instructions
- G06F9/30007—Arrangements for executing specific machine instructions to perform operations on data operands
- G06F9/30036—Instructions to perform operations on packed data, e.g. vector, tile or matrix operations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/3017—Runtime instruction translation, e.g. macros
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/3017—Runtime instruction translation, e.g. macros
- G06F9/30178—Runtime instruction translation, e.g. macros of compressed or encrypted instructions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3818—Decoding for concurrent execution
- G06F9/3822—Parallel decoding, e.g. parallel decode units
Definitions
- This invention relates to a processor for processing a plurality of instructions in parallel, and in particular to a processor for processing an instruction set of a plurality of instructions packed into a single code.
- a DSP digital signal processor
- a DSP digital signal processor
- ten thousand sets of data are to be arithmetically processed, for example, at least ten thousand cycles are required even if the operation on each set of data can be executed in a single machine cycle.
- each set of data may be processed at high speed, but the time required for processing increases in proportion to the amount of data because the data processing is in series.
- the processing performance can be improved by parallel operation.
- a plurality of operation units are prepared and operated at the same time to process a plurality of sets of data at the same time.
- the method called SIMD single instruction-multiple data streams
- SIMD single instruction-multiple data streams
- Document 1 (D. A. Patterson and J. L. Hennessy, “Computer Organization and Design”, Nikkei Business Publications) describes a method of decreasing the length of an instruction code to reduce the size of the instruction memory.
- An object of this invention is to provide a processor in which the number of instructions capable of being executed in parallel per cycle is increased while at the same time reducing the capacity of an instruction memory.
- a processor including an instruction memory for storing an instruction code, a converter for converting a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes, a decoder for decoding the plurality of the instruction codes converted by the converter, and a plurality of operation units for executing the operation corresponding to each of the plurality of the instruction codes in accordance with the decoding result of the decoder.
- the converter converts a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes and the decoder decodes the plurality of the instruction codes converted by the converter, thereby causing the plurality of the operation units to execute the operations, the number of instructions that can be executed in parallel per cycle can be increased while at the same time reducing the capacity of the instruction memory.
- FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for a processor according to a first embodiment of the invention
- FIG. 2 is a block diagram showing a schematic configuration of the processor according to the first embodiment of the invention.
- FIG. 3 is a flowchart for explaining the processing steps in a compiler for writing an instruction set in a conversion table 13 of the processor according to a second embodiment of the invention
- FIG. 4 is a block diagram showing a schematic configuration of the processor according to a third embodiment of the invention.
- FIG. 5 is a block diagram showing a schematic configuration of the processor according to a fifth embodiment of the invention.
- FIG. 6 is a block diagram showing a schematic configuration of the processor according to a seventh embodiment of the invention.
- FIG. 7 is a block diagram showing a schematic configuration of the processor according to an eighth embodiment of the invention.
- FIG. 8 is a block diagram showing a schematic configuration of the processor according to a tenth embodiment of the invention.
- FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jumping according to the tenth embodiment.
- FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for the processor according to the first embodiment of the invention.
- this instruction code includes a pack valid bit (V) indicating a packed instruction, a packed instruction, an operation code and an operand Mp of an instruction 1 , an operation code and an operand Ap of an instruction 2 and a mode bit (M) indicating the operation mode of the processor.
- V pack valid bit
- Mp mode bit
- a “1 d@p 1 ” is an instruction for loading the data from an address indicated by a register p 1
- an “add @p 2 ” is an instruction for adding the data stored at the address indicated by a register p 2 and the data stored in, for example, an accumulator.
- the pack valid bit is a valid “1”
- a plurality of instructions are selected in accordance with a 4-bit packed instruction as described later.
- the packed instruction is “0100”, for example, four instructions including “mv r1, r5”, “inc r5, 2”, “mv r2, r3” and “inc r4, 1” are selected as shown in FIG. 1B .
- the “mv” instruction is for data transfer between registers, and the “inc” instruction is for incrementing the contents of the register.
- FIG. 2 is a block diagram showing a general configuration of a processor according to the first embodiment of the invention.
- This processor includes an instruction memory 11 , an instruction queue 12 , a conversion table 13 for converting a packed instruction into a plurality of instruction codes and outputting the instruction codes, an instruction selector 14 for selectively outputting a plurality of converted instruction codes from conversion table 13 and a packed instruction from instruction queue 12 in accordance with the pack valid bit, an instruction decoder 15 , a register file 16 and ALUs 17 - 1 to 17 - 4 .
- Instruction memory 11 stores a program to be executed by the processor.
- Instruction queue 12 temporarily stores the instruction code fetched from instruction memory 11 .
- the packed instruction is applied to conversion table 13 and instruction selector 14 as a 4-bit pre-conversion code.
- the remaining 12 bits are applied directly to instruction decoder 15 .
- Conversion table 13 includes a plurality of flip-flops 21 - 1 to 21 - n connected to a data bus 23 and a selector 22 for selecting the outputs of flip-flops 21 - 1 to 21 - n in accordance with a 4-bit pre-conversion code received from instruction queue 12 .
- Flip-flops 21 - 1 to 21 - n each store a converted 32-bit code including a plurality of instructions through data bus 23 .
- Data bus 23 is connected to the output port, for example, of a computer not shown, and the user can write contents of conversion table 13 from outside.
- the flip-flops are arbitrary in number, and each corresponds to a 4-bit pre-conversion code. Specifically, in the case where the pre-conversion code is 4 bits, up to 16 flip-flops may be included.
- the pre-conversion code is not limited to 4 bits, but the number of corresponding flip-flops can be increased by increasing the number of bits.
- the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
- instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15 , while in the case where the pack valid bit is invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 12 .
- the 4-bit packed instruction is processed as a single instruction code.
- Instruction decoder 15 includes a plurality of decode blocks 151 to 156 .
- block 151 individually decodes the four instruction codes received from instruction selector 14 , and applies the decoding result to ALUs 17 - 1 to 17 - 4 .
- the pack valid bit is an invalid “0”
- the result of decoding the one instruction code received from instruction selector 14 is applied to any one of ALUs 17 - 1 to 17 - 4 and only one ALU is caused to execute the operation.
- Blocks 152 and 153 decode the operation code of instruction 1 and the operation code of instruction 2 , respectively, and output the decoding result as a control signal to an operation array not shown.
- Blocks 154 and 155 decode the operand portion of instruction 1 and the operand portion of instruction 2 , respectively, and output the decoding result as a control signal to an operation array not shown.
- Block 156 decodes pack valid bit (v) and mode bit (M), and applies the decoding result as a control signal to each block of the processor.
- Register file 16 is a group of registers for holding the data to be operated and the data constituting the result of operation.
- ALUs 17 - 1 to 17 - 4 receive the decoding result from instruction decoder 15 , and while accessing the data held in register file 16 through common buses 18 - 1 and 18 - 2 , executes the parallel operation.
- ALUs 17 - 1 to 17 - 4 can read the data at the same time from register file 16 through common buses 18 - 1 and 18 - 2 .
- Register file 16 receives and stores the results of the operations of ALUs 17 - 1 to 17 - 4 through common buses 19 and 20 .
- the decoding result applied from instruction decoder 15 to ALUs 17 - 1 to 17 - 4 includes an operand as shown in FIG. 1B .
- conversion table 1 selects a plurality of instruction codes in accordance with the packed instruction
- instruction selector 14 selects and outputs the converted code containing a plurality of the instruction codes output from conversion table 13
- instruction decoder 15 decodes the plurality of the instruction codes and causes a plurality of the ALUs to execute the operations.
- a plurality of instructions can be packed into a single instruction code and the instruction memory can be reduced in size, while at the same time increasing the number of instructions that can be executed in one cycle.
- the operation can be executed at higher speed, and the power consumption required to fetch the instruction code can be reduced.
- a plurality of the instructions stored in flip-flops 21 - 1 to 21 - n can be changed in accordance with the execution program.
- the operation can be executed at higher speed.
- the user writes an instruction set in conversion table 13 from outside.
- a compiler implemented by the computer writes an instruction set in conversion table 13 .
- the general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 2 . Therefore, similar or identical component parts of the configuration and functions are not described in detail again.
- FIG. 3 is a flowchart for explaining the steps of the process executed by the compiler for writing an instruction set in conversion table 13 of the processor according to the second embodiment of the invention.
- the compiler prepares a list of packed instructions (a plurality of instruction codes executed in parallel) from the instruction code described in a program (S 11 ), and generates a histogram according to the frequency of occurrence of the packed instruction (S 12 ).
- the frequency of occurrence of “mv r1, r5; inc r1, 2; . . . ” is 26, the frequency of occurrence of “mv r2, r3; mv r3, r4; . . . ” is 154 and the frequency of occurrence of “inc r3, 2; inc r5, 1; . . . ” is 3.
- the sorting is carried out (S 13 ), a plurality of the instruction codes are transferred to flip-flops 21 - 1 to 21 - n in the descending order of the frequency of occurrence (S 14 ) thereby to complete the process.
- the plurality of the instruction codes like in the first embodiment, are transferred to conversion table 13 , for example, through the output port of the computer, not shown, connected to data bus 23 .
- a packed instruction is assigned to the plurality of the instruction codes in accordance with which one of flip-flops 21 - 1 to 21 - n the plurality of the instruction codes are transferred to, and the program instruction is converted to a machine language using the particular packed instruction.
- a plurality of the instruction codes determined as high in the frequency of occurrence by the compiler are packed and a plurality of the instruction codes are written in conversion table 13 .
- a group of instructions high in compression efficiency can be packed and the instruction memory can be further reduced in size.
- FIG. 4 is a block diagram showing a general configuration of the processor according to a third embodiment of the invention.
- the processor according to this embodiment as compared with the processor according to the first embodiment shown in FIG. 2 , only the internal structure of conversion table 13 is different. The component parts of the configuration and functions similar or identical to those of the first embodiment, therefore, are not described in detail again.
- Conversion table 13 includes a SRAM (static random access memory) and an associative memory 24 .
- the 4-bit pre-conversion code output from instruction queue 12 is applied as an address of memory 24 .
- Memory 24 has a width of 32 bits and a plurality of instruction codes are stored in each memory area.
- a computer not shown is assumed to write a plurality of instruction codes sequentially in memory 24 through data bus 23 by outputting a plurality of the instruction codes to data bus 23 while controlling the address.
- conversion table 13 is configured of memory 24 , and therefore, in addition to the effects described in the first embodiment, the hardware configuration can be simplified and reduced in size.
- the processor according to a fourth embodiment of the invention is different from the processor according to the first embodiment shown in FIG. 2 only in that the data bus of the processor is connected to data bus 23 .
- the processor writes a plurality of instructions in conversion table 13 through data bus 23 as a bus master at a predetermined timing such as resetting. In the process, until a plurality of the instructions are completely written in conversion table 13 , the processor executes the process with the pack valid bit set to invalid “0” according to the program, and upon completion of writing the plurality of the instructions in conversion table 13 , the execution of the packed instructions becomes possible.
- the processor writes a plurality of instruction codes in conversion table 13 by itself as a bus master.
- a plurality of instruction codes are not required to be transferred to conversion table 13 from an external source, and the processor can execute the process by itself.
- the first embodiment as shown in FIG. 1 , is so configured that the instruction code includes packed and non-packed instructions.
- the instruction code includes only packed instructions. Therefore, the instruction code contains no pack valid bit.
- FIG. 5 is a block diagram showing a general configuration of the processor according to a fifth embodiment of the invention.
- the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction selector 14 is eliminated, the output of selector 22 is applied directly to instruction decoder 15 and that the block in instruction decoder 15 for decoding the non-packed instructions is deleted.
- the instruction code includes only the packed instruction.
- the number of bits of the packed instruction can be increased and a greater number of instruction sets can be packed, thereby making it possible to execute a greater number of instructions in parallel.
- a plurality of the instructions written in conversion table 13 contain an operand.
- the instructions written in conversion table 13 contain no operand but only the operation code is packed.
- the processor according to this embodiment has a general configuration similar to that of the first embodiment shown in FIG. 2 . Nevertheless, the operand of the instructions to be packed is included in the instruction code stored in instruction queue 12 . Instruction decoder 15 decodes the particular operand and applies the decoding result to ALUs 17 - 1 to 17 - 4 .
- Conversion table 13 outputs a plurality of the instruction codes containing only the operation code to instruction selector 14 .
- Instruction decoder 15 decodes the operation code of the instruction code received from instruction selector 14 and applies the decoding result to ALUs 17 - 1 to 17 - n.
- the operand is not stored in flip-flops 21 - 1 to 21 - n in conversion table 13 .
- the circuit size of conversion table 13 can be further reduced.
- FIG. 6 is a block diagram showing a general configuration of the processor according to a seventh embodiment of the invention.
- the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
- conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
- the packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and to instruction selector 14 on the other hand.
- the remaining 12 bits are applied directly to instruction queue 12 .
- the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
- instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction queue 12 , while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction queue 12 .
- Instruction queue 12 outputs the instruction code held therein to instruction decoder 15 and causes instruction decoder 15 to decode it at a predetermined timing.
- conversion table 13 is interposed between instruction memory 11 and instruction queue 12 .
- the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded is shortened.
- the delay time before the instruction code is transferred from instruction queue 12 to instruction decoder 15 and decoded constitutes the critical path of the processor as a whole, the overall operating frequency of the processor can be improved.
- FIG. 7 is a block diagram showing a general configuration of the processor according to an eighth embodiment of the invention.
- the processor according to this embodiment is different from the processor according to the first embodiment shown in FIG. 2 only in that instruction queue 12 is deleted.
- the packed instruction contained in the instructions fetched from instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and applied to instruction selector 14 at the same time.
- the remaining 12 bits are applied directly to instruction decoder 15 .
- the 32-bit converted code selected by selector 22 is applied to instruction selector 14 .
- instruction selector 14 selects the 32-bit converted code received from selector 22 and outputs it to instruction decoder 15 , while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output to instruction decoder 15 .
- instruction queue 12 is deleted, and therefore, in addition to the effects described in the first embodiment, the circuit size of the whole processor can be reduced.
- a plurality of packed instruction codes are serially processed.
- the general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in FIG. 1 . Nevertheless, a predetermined bit of the packed instruction indicates parallel or serial execution, and instruction selector 14 changes the process by reference to the particular predetermined bit.
- instruction selector 14 In the case where the packed instruction indicates the parallel execution, instruction selector 14 outputs the 32-bit converted code from selector 22 to instruction decoder 15 at a time. In the case where the packed instruction indicates the serial execution, on the other hand, instruction selector 14 serially outputs the 32-bit converted code from selector 22 in a plurality of cycles.
- the instruction codes are sequentially output to instruction decoder 15 in 4 cycles, and the instruction decoder 15 decodes the 4 instructions sequentially.
- the decoding result is continuously applied to one of ALUs 17 - 1 to 17 - 4 .
- a plurality of packed instruction codes are processed serially.
- the instructions serially executed can be also packed, thereby making it possible to further improve the instruction compression efficiency.
- FIG. 8 is a block diagram showing a general configuration of the processor according to a tenth embodiment of the invention.
- the processor according to this embodiment as compared with the processor according to the first embodiment shown in FIG. 1 , includes a jump destination instruction storage unit 31 for registering the instruction of the jump destination and a flip-flop 32 for holding the address of the jump destination.
- jump instruction free of penalty is decoded by instruction decoder 15 during execution of a program by the processor.
- An instruction registration control signal is output to jump destination instruction storage unit 31 .
- the jump instruction free of penalty is assumed to include an unconditional branch instruction or a conditional branch instruction satisfying the branch conditions.
- jump destination instruction storage unit 31 holds the particular jump destination instruction.
- instruction decoder 15 decodes a jump instruction free of penalty
- flip-flop 32 holds the jump destination address output from instruction decoder 15 .
- Instruction selector 14 by reference to the jump destination address held in flip-flop 32 , stores the correspondence between the jump destination address and the jump destination instruction held in jump destination instruction storage unit 31 .
- a jump instruction free of penalty is decoded by instruction decoder 15 , a similar process is executed if the particular jump destination instruction is not held in jump destination storage unit 31 .
- instruction selector 14 determines that a jump destination instruction corresponding to the jump destination address held in flip-flop 32 is held in jump destination instruction storage unit 31 .
- the particular jump destination instruction is read from jump destination instruction storage unit 31 and output to instruction decoder 15 .
- the address of the instruction next to the jump destination instruction is applied to a program counter not shown.
- FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jump according to the tenth embodiment of the invention.
- the upper half of FIG. 9 shows the jump process according to this embodiment, in which the address of instruction 1 is issued in cycle T 1 and the address of instruction 2 is issued in cycle T 2 while at the same time instruction 1 is fetched.
- Instruction 1 is a jump instruction free of penalty.
- a jump instruction free of penalty is decoded by instruction decoder 15 , and instruction selector 14 , by referring to the jump destination address held in flip-flop 32 , reads the jump destination instruction from jump destination instruction storage unit 31 and outputs it to instruction decoder 15 .
- the address of the instruction next to the jump destination instruction is applied to a program counter not shown, and this address is issued.
- cycle T 4 the jump destination instruction is executed, the instruction next to the jump destination instruction is fetched, while at the same time the address of the next instruction but one is issued.
- cycle T 5 and subsequent cycles a similar pipelining process is executed.
- the instruction next to the jump instruction is stored in instruction queue 12 . Even in the case where the branch conditions of the conditional branch instruction fail to be met, therefore, the pipelining process similar to that shown in the upper half of FIG. 9 can be carried out by applying the particular instruction to instruction decoder 15 .
- the lower half of FIG. 9 indicates the conventional jump process.
- cycle T 1 the address of instruction 1 is issued
- cycle T 2 the address of instruction 2 is issued while at the same time instruction 1 is fetched.
- Instruction 1 is a jump instruction requiring the branch prediction.
- cycle T 3 a jump instruction is decoded by instruction decoder 15 .
- instruction decoder 15 it is assumed that the branch prediction fails and the instruction of the branch destination is required to be fetched again.
- the instruction stored in instruction queue 12 is invalidated and processed as a NOP (non-operation) instruction.
- cycle T 4 the NOP instruction is executed and the jump destination instruction is fetched, while at the same time the address of the instruction next to the jump destination instruction is issued.
- cycle T 5 and subsequent cycles a similar pipelining process is executed. In this way, a one-cycle delay occurs as compared with the jump instruction free of penalty.
- the jump destination instruction may be required to be fetched, as shown in the lower half of FIG. 9 , unless the jump instruction is held in jump destination instruction storage unit 31 . In such a case, a one-cycle delay occurs.
- a failure of the branch prediction results in the requirement of a repeated instruction fetch, and therefore a jump penalty occurs.
- no jump penalty occurs as long as a jump destination instruction is held in jump destination instruction storage unit 31 .
Landscapes
- Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Executing Machine-Instructions (AREA)
- Advance Control (AREA)
Abstract
A conversion table converts a packed instruction (pre-conversion code) contained in the instruction code fetched from an instruction memory into a plurality of instruction codes (converted codes). An instruction decoder decodes the plurality of the instruction codes converted by a conversion table. A plurality of ALUs perform the operation in accordance with the decoding result of the instruction decoder. Therefore, the number of instructions that can be executed in parallel per cycle may be increased while at the same time the capacity of the instruction memory is reduced.
Description
- 1. Field of the Invention
- This invention relates to a processor for processing a plurality of instructions in parallel, and in particular to a processor for processing an instruction set of a plurality of instructions packed into a single code.
- 2. Description of the Background Art
- In recent years, with the spread of portable terminal devices, the digital signal processing for processing a great amount of data such as voices and images at high speed has become increasingly important. A DSP (digital signal processor) is typically used as a semiconductor device exclusive to such digital signal processing. However, in the case where an amount of data to be processed is enormous, it is difficult to improve the performance dramatically even with the use of an exclusive DSP. Assuming that ten thousand sets of data are to be arithmetically processed, for example, at least ten thousand cycles are required even if the operation on each set of data can be executed in a single machine cycle. In other words, each set of data may be processed at high speed, but the time required for processing increases in proportion to the amount of data because the data processing is in series.
- In the case where an amount of data to be processed is large, the processing performance can be improved by parallel operation. Specifically, a plurality of operation units are prepared and operated at the same time to process a plurality of sets of data at the same time. In the case where the same operation is performed on a plurality of sets of data, the method called SIMD (single instruction-multiple data streams) can be employed to reduce the area of the operation unit while maintaining a high parallel performance. Specifically, while a plurality of data processors are prepared, a high performance with a small area can be exhibited by providing a common control unit for interpreting an instruction and controlling the process.
- Document 1 (D. A. Patterson and J. L. Hennessy, “Computer Organization and Design”, Nikkei Business Publications) describes a method of decreasing the length of an instruction code to reduce the size of the instruction memory.
- Document 2 (Akira Nakamori, “Introduction to Microprocessor Architecture”, CQ Publishing), on the other hand, describes a method in which a plurality of slots are formed in one instruction format for parallel execution by VLIW (very long instruction word) in order to increase the number of instructions that can be executed per cycle.
- However, reducing the size of the instruction memory as described in
Document 1 and increasing the number of instructions that can be executed per cycle as described inDocument 2 are in the relation of so-called tradeoff. - Specifically, according to the method described in
Document 1, it is possible to reduce the size of the instruction memory, while the fact that a plurality of instructions are processed in series poses a problem that a number of cycles are required to execute the instructions. Taking an example where the instruction code length is 16 bits, the instruction length is short but four cycles are required to execute four instructions. - According to the method described in
Document 2, on the other hand, more instructions may be executed per cycle, while the instruction length is increased and so is the size of the instruction memory. In the case where four slots of 16 bits are provided, for example, four instructions can be executed at the same time in a single cycle at the sacrifice of an extended instruction code length of 64 bits. - An object of this invention is to provide a processor in which the number of instructions capable of being executed in parallel per cycle is increased while at the same time reducing the capacity of an instruction memory.
- According to one aspect of the invention, there is provided a processor including an instruction memory for storing an instruction code, a converter for converting a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes, a decoder for decoding the plurality of the instruction codes converted by the converter, and a plurality of operation units for executing the operation corresponding to each of the plurality of the instruction codes in accordance with the decoding result of the decoder.
- In view of the fact that the converter converts a packed instruction contained in an instruction code fetched from the instruction memory into a plurality of instruction codes and the decoder decodes the plurality of the instruction codes converted by the converter, thereby causing the plurality of the operation units to execute the operations, the number of instructions that can be executed in parallel per cycle can be increased while at the same time reducing the capacity of the instruction memory.
- The foregoing and other objects, features, aspects and advantages of the present invention will become more apparent from the following detailed description of the present invention when taken in conjunction with the accompanying drawings.
-
FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for a processor according to a first embodiment of the invention; -
FIG. 2 is a block diagram showing a schematic configuration of the processor according to the first embodiment of the invention; -
FIG. 3 is a flowchart for explaining the processing steps in a compiler for writing an instruction set in a conversion table 13 of the processor according to a second embodiment of the invention; -
FIG. 4 is a block diagram showing a schematic configuration of the processor according to a third embodiment of the invention; -
FIG. 5 is a block diagram showing a schematic configuration of the processor according to a fifth embodiment of the invention; -
FIG. 6 is a block diagram showing a schematic configuration of the processor according to a seventh embodiment of the invention; -
FIG. 7 is a block diagram showing a schematic configuration of the processor according to an eighth embodiment of the invention; -
FIG. 8 is a block diagram showing a schematic configuration of the processor according to a tenth embodiment of the invention; and -
FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jumping according to the tenth embodiment. -
FIGS. 1A and 1B are diagrams showing an example of the structure of an instruction code used for the processor according to the first embodiment of the invention. As shown inFIG. 1A , this instruction code includes a pack valid bit (V) indicating a packed instruction, a packed instruction, an operation code and an operand Mp of aninstruction 1, an operation code and an operand Ap of aninstruction 2 and a mode bit (M) indicating the operation mode of the processor. A “1 d@p1” is an instruction for loading the data from an address indicated by a register p1, and an “add @p2” is an instruction for adding the data stored at the address indicated by a register p2 and the data stored in, for example, an accumulator. - In the case where the pack valid bit is a valid “1”, a plurality of instructions are selected in accordance with a 4-bit packed instruction as described later. In the case where the packed instruction is “0100”, for example, four instructions including “mv r1, r5”, “inc r5, 2”, “mv r2, r3” and “inc r4, 1” are selected as shown in
FIG. 1B . The “mv” instruction is for data transfer between registers, and the “inc” instruction is for incrementing the contents of the register. - In the case where the pack valid bit is invalid “0”, a 4-bit packed instruction is processed as one instruction.
-
FIG. 2 is a block diagram showing a general configuration of a processor according to the first embodiment of the invention. This processor includes aninstruction memory 11, aninstruction queue 12, a conversion table 13 for converting a packed instruction into a plurality of instruction codes and outputting the instruction codes, aninstruction selector 14 for selectively outputting a plurality of converted instruction codes from conversion table 13 and a packed instruction frominstruction queue 12 in accordance with the pack valid bit, aninstruction decoder 15, aregister file 16 and ALUs 17-1 to 17-4. -
Instruction memory 11 stores a program to be executed by the processor.Instruction queue 12 temporarily stores the instruction code fetched frominstruction memory 11. Of the 16-bit instruction code stored ininstruction queue 12, the packed instruction is applied to conversion table 13 andinstruction selector 14 as a 4-bit pre-conversion code. The remaining 12 bits are applied directly toinstruction decoder 15. - Conversion table 13 includes a plurality of flip-flops 21-1 to 21-n connected to a
data bus 23 and aselector 22 for selecting the outputs of flip-flops 21-1 to 21-n in accordance with a 4-bit pre-conversion code received frominstruction queue 12. - Flip-flops 21-1 to 21-n each store a converted 32-bit code including a plurality of instructions through
data bus 23.Data bus 23 is connected to the output port, for example, of a computer not shown, and the user can write contents of conversion table 13 from outside. The flip-flops are arbitrary in number, and each corresponds to a 4-bit pre-conversion code. Specifically, in the case where the pre-conversion code is 4 bits, up to 16 flip-flops may be included. The pre-conversion code is not limited to 4 bits, but the number of corresponding flip-flops can be increased by increasing the number of bits. - The 32-bit converted code selected by
selector 22 is applied toinstruction selector 14. In the case where the pack valid bit is a valid “1”,instruction selector 14 selects the 32-bit converted code received fromselector 22 and outputs it toinstruction decoder 15, while in the case where the pack valid bit is invalid “0”, on the other hand, the 4-bit packed instruction is selected and output toinstruction decoder 12. In this case, the 4-bit packed instruction is processed as a single instruction code. -
Instruction decoder 15 includes a plurality of decode blocks 151 to 156. In the case where the pack valid bit is a valid “1”, block 151 individually decodes the four instruction codes received frominstruction selector 14, and applies the decoding result to ALUs 17-1 to 17-4. In the case where the pack valid bit is an invalid “0”, on the other hand, the result of decoding the one instruction code received frominstruction selector 14 is applied to any one of ALUs 17-1 to 17-4 and only one ALU is caused to execute the operation. -
Blocks instruction 1 and the operation code ofinstruction 2, respectively, and output the decoding result as a control signal to an operation array not shown.Blocks instruction 1 and the operand portion ofinstruction 2, respectively, and output the decoding result as a control signal to an operation array not shown. -
Block 156 decodes pack valid bit (v) and mode bit (M), and applies the decoding result as a control signal to each block of the processor. -
Register file 16 is a group of registers for holding the data to be operated and the data constituting the result of operation. ALUs 17-1 to 17-4 receive the decoding result frominstruction decoder 15, and while accessing the data held inregister file 16 through common buses 18-1 and 18-2, executes the parallel operation. ALUs 17-1 to 17-4 can read the data at the same time fromregister file 16 through common buses 18-1 and 18-2. -
Register file 16 receives and stores the results of the operations of ALUs 17-1 to 17-4 throughcommon buses 19 and 20. The decoding result applied frominstruction decoder 15 to ALUs 17-1 to 17-4 includes an operand as shown inFIG. 1B . - As explained above, with the processor according to this embodiment, in the case where the pack valid bit is a valid “1”, conversion table 1 selects a plurality of instruction codes in accordance with the packed instruction,
instruction selector 14 selects and outputs the converted code containing a plurality of the instruction codes output from conversion table 13, andinstruction decoder 15 decodes the plurality of the instruction codes and causes a plurality of the ALUs to execute the operations. In this way, a plurality of instructions can be packed into a single instruction code and the instruction memory can be reduced in size, while at the same time increasing the number of instructions that can be executed in one cycle. - Also, the operation can be executed at higher speed, and the power consumption required to fetch the instruction code can be reduced.
- Also, a plurality of the instructions stored in flip-flops 21-1 to 21-n can be changed in accordance with the execution program. By incorporating an instruction set suitable for the program in flip-flops 21-1 to 21-n, therefore, the operation can be executed at higher speed.
- In the first embodiment of the invention, the user writes an instruction set in conversion table 13 from outside. According to the second embodiment, in contrast, a compiler implemented by the computer writes an instruction set in conversion table 13.
- The general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in
FIG. 2 . Therefore, similar or identical component parts of the configuration and functions are not described in detail again. -
FIG. 3 is a flowchart for explaining the steps of the process executed by the compiler for writing an instruction set in conversion table 13 of the processor according to the second embodiment of the invention. First, the compiler prepares a list of packed instructions (a plurality of instruction codes executed in parallel) from the instruction code described in a program (S11), and generates a histogram according to the frequency of occurrence of the packed instruction (S12). - As indicated in step S12 of
FIG. 3 , the frequency of occurrence of “mv r1, r5; inc r1, 2; . . . ” is 26, the frequency of occurrence of “mv r2, r3; mv r3, r4; . . . ” is 154 and the frequency of occurrence of “inc r3, 2; inc r5, 1; . . . ” is 3. - Next, in accordance with the frequency of occurrence of the packed instruction, the sorting is carried out (S13), a plurality of the instruction codes are transferred to flip-flops 21-1 to 21-n in the descending order of the frequency of occurrence (S14) thereby to complete the process. The plurality of the instruction codes, like in the first embodiment, are transferred to conversion table 13, for example, through the output port of the computer, not shown, connected to
data bus 23. - In the compiler, a packed instruction is assigned to the plurality of the instruction codes in accordance with which one of flip-flops 21-1 to 21-n the plurality of the instruction codes are transferred to, and the program instruction is converted to a machine language using the particular packed instruction.
- As explained above, in the processor according to this embodiment, a plurality of the instruction codes determined as high in the frequency of occurrence by the compiler are packed and a plurality of the instruction codes are written in conversion table 13. In addition to the effects explained with reference to the first embodiment, therefore, a group of instructions high in compression efficiency can be packed and the instruction memory can be further reduced in size.
-
FIG. 4 is a block diagram showing a general configuration of the processor according to a third embodiment of the invention. In the processor according to this embodiment, as compared with the processor according to the first embodiment shown inFIG. 2 , only the internal structure of conversion table 13 is different. The component parts of the configuration and functions similar or identical to those of the first embodiment, therefore, are not described in detail again. - Conversion table 13 includes a SRAM (static random access memory) and an
associative memory 24. The 4-bit pre-conversion code output frominstruction queue 12 is applied as an address ofmemory 24.Memory 24 has a width of 32 bits and a plurality of instruction codes are stored in each memory area. - A computer not shown is assumed to write a plurality of instruction codes sequentially in
memory 24 throughdata bus 23 by outputting a plurality of the instruction codes todata bus 23 while controlling the address. - As explained above, in the processor according to this embodiment, conversion table 13 is configured of
memory 24, and therefore, in addition to the effects described in the first embodiment, the hardware configuration can be simplified and reduced in size. - The processor according to a fourth embodiment of the invention is different from the processor according to the first embodiment shown in
FIG. 2 only in that the data bus of the processor is connected todata bus 23. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again. - The processor writes a plurality of instructions in conversion table 13 through
data bus 23 as a bus master at a predetermined timing such as resetting. In the process, until a plurality of the instructions are completely written in conversion table 13, the processor executes the process with the pack valid bit set to invalid “0” according to the program, and upon completion of writing the plurality of the instructions in conversion table 13, the execution of the packed instructions becomes possible. - As explained above, the processor according to this embodiment writes a plurality of instruction codes in conversion table 13 by itself as a bus master. In addition to the effects explained in the first embodiment, therefore, a plurality of instruction codes are not required to be transferred to conversion table 13 from an external source, and the processor can execute the process by itself.
- The first embodiment, as shown in
FIG. 1 , is so configured that the instruction code includes packed and non-packed instructions. According to this embodiment, on the other hand, the instruction code includes only packed instructions. Therefore, the instruction code contains no pack valid bit. -
FIG. 5 is a block diagram showing a general configuration of the processor according to a fifth embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown inFIG. 2 only in thatinstruction selector 14 is eliminated, the output ofselector 22 is applied directly toinstruction decoder 15 and that the block ininstruction decoder 15 for decoding the non-packed instructions is deleted. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again. - As explained above, in the processor according to this embodiment, the instruction code includes only the packed instruction. In addition to the effects described in the first embodiment, therefore, the number of bits of the packed instruction can be increased and a greater number of instruction sets can be packed, thereby making it possible to execute a greater number of instructions in parallel.
- In the first embodiment, a plurality of the instructions written in conversion table 13 contain an operand. According to the sixth embodiment, on the other hand, the instructions written in conversion table 13 contain no operand but only the operation code is packed.
- The processor according to this embodiment has a general configuration similar to that of the first embodiment shown in
FIG. 2 . Nevertheless, the operand of the instructions to be packed is included in the instruction code stored ininstruction queue 12.Instruction decoder 15 decodes the particular operand and applies the decoding result to ALUs 17-1 to 17-4. - Conversion table 13 outputs a plurality of the instruction codes containing only the operation code to
instruction selector 14.Instruction decoder 15 decodes the operation code of the instruction code received frominstruction selector 14 and applies the decoding result to ALUs 17-1 to 17-n. - As explained above, in the processor according to this embodiment, the operand is not stored in flip-flops 21-1 to 21-n in conversion table 13. In addition to the effects explained in the first embodiment, therefore, the circuit size of conversion table 13 can be further reduced.
-
FIG. 6 is a block diagram showing a general configuration of the processor according to a seventh embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown inFIG. 2 only in that conversion table 13 is interposed betweeninstruction memory 11 andinstruction queue 12. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again. - The packed instruction contained in the instructions fetched from
instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and toinstruction selector 14 on the other hand. The remaining 12 bits are applied directly toinstruction queue 12. - The 32-bit converted code selected by
selector 22 is applied toinstruction selector 14. In the case where the pack valid bit is a valid “1”,instruction selector 14 selects the 32-bit converted code received fromselector 22 and outputs it toinstruction queue 12, while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output toinstruction queue 12. -
Instruction queue 12 outputs the instruction code held therein toinstruction decoder 15 and causesinstruction decoder 15 to decode it at a predetermined timing. - As explained above, in the processor according to this embodiment, conversion table 13 is interposed between
instruction memory 11 andinstruction queue 12. In addition to the effects explained in the first embodiment, therefore, the delay time before the instruction code is transferred frominstruction queue 12 toinstruction decoder 15 and decoded is shortened. In the case where the delay time before the instruction code is transferred frominstruction queue 12 toinstruction decoder 15 and decoded constitutes the critical path of the processor as a whole, the overall operating frequency of the processor can be improved. -
FIG. 7 is a block diagram showing a general configuration of the processor according to an eighth embodiment of the invention. The processor according to this embodiment is different from the processor according to the first embodiment shown inFIG. 2 only in thatinstruction queue 12 is deleted. The component parts of the configuration and functions similar or identical to those of the aforementioned embodiments, therefore, are not described in detail again. - The packed instruction contained in the instructions fetched from
instruction memory 11 is applied to conversion table 13 as a 4-bit pre-conversion code on the one hand and applied toinstruction selector 14 at the same time. The remaining 12 bits are applied directly toinstruction decoder 15. - The 32-bit converted code selected by
selector 22 is applied toinstruction selector 14. In the case where the pack valid bit is a valid “1”,instruction selector 14 selects the 32-bit converted code received fromselector 22 and outputs it toinstruction decoder 15, while in the case where the pack valid bit is an invalid “0”, on the other hand, the 4-bit packed instruction is selected and output toinstruction decoder 15. - As explained above, in the processor according to this embodiment,
instruction queue 12 is deleted, and therefore, in addition to the effects described in the first embodiment, the circuit size of the whole processor can be reduced. - According to the ninth embodiment, unlike in the first embodiment in which a plurality of packed instruction codes are processed in parallel, a plurality of packed instruction codes are serially processed.
- The general configuration of the processor according to this embodiment is similar to that of the processor according to the first embodiment shown in
FIG. 1 . Nevertheless, a predetermined bit of the packed instruction indicates parallel or serial execution, andinstruction selector 14 changes the process by reference to the particular predetermined bit. - In the case where the packed instruction indicates the parallel execution,
instruction selector 14 outputs the 32-bit converted code fromselector 22 toinstruction decoder 15 at a time. In the case where the packed instruction indicates the serial execution, on the other hand,instruction selector 14 serially outputs the 32-bit converted code fromselector 22 in a plurality of cycles. - In the case where 4 instruction codes are output from
selector 22, for example, the instruction codes are sequentially output toinstruction decoder 15 in 4 cycles, and theinstruction decoder 15 decodes the 4 instructions sequentially. The decoding result is continuously applied to one of ALUs 17-1 to 17-4. - As explained above, in the processor according to this embodiment, a plurality of packed instruction codes are processed serially. In addition to the effects explained in the first embodiment, therefore, the instructions serially executed can be also packed, thereby making it possible to further improve the instruction compression efficiency.
-
FIG. 8 is a block diagram showing a general configuration of the processor according to a tenth embodiment of the invention. The processor according to this embodiment, as compared with the processor according to the first embodiment shown inFIG. 1 , includes a jump destinationinstruction storage unit 31 for registering the instruction of the jump destination and a flip-flop 32 for holding the address of the jump destination. - Assume that the jump instruction free of penalty is decoded by
instruction decoder 15 during execution of a program by the processor. An instruction registration control signal is output to jump destinationinstruction storage unit 31. The jump instruction free of penalty is assumed to include an unconditional branch instruction or a conditional branch instruction satisfying the branch conditions. - In the case where a jump destination instruction is stored in
instruction queue 12, jump destinationinstruction storage unit 31 holds the particular jump destination instruction. In the case whereinstruction decoder 15 decodes a jump instruction free of penalty, flip-flop 32 holds the jump destination address output frominstruction decoder 15.Instruction selector 14, by reference to the jump destination address held in flip-flop 32, stores the correspondence between the jump destination address and the jump destination instruction held in jump destinationinstruction storage unit 31. In the case where a jump instruction free of penalty is decoded byinstruction decoder 15, a similar process is executed if the particular jump destination instruction is not held in jumpdestination storage unit 31. - Next, assume that in the case where
instruction decoder 15 decodes a jump instruction free of penalty,instruction selector 14 determines that a jump destination instruction corresponding to the jump destination address held in flip-flop 32 is held in jump destinationinstruction storage unit 31. The particular jump destination instruction is read from jump destinationinstruction storage unit 31 and output toinstruction decoder 15. In the process, the address of the instruction next to the jump destination instruction is applied to a program counter not shown. -
FIG. 9 is a diagram for explaining the pipelining process executed by the processor at the time of jump according to the tenth embodiment of the invention. The upper half ofFIG. 9 shows the jump process according to this embodiment, in which the address ofinstruction 1 is issued in cycle T1 and the address ofinstruction 2 is issued in cycle T2 while at thesame time instruction 1 is fetched.Instruction 1 is a jump instruction free of penalty. - In cycle T3, a jump instruction free of penalty is decoded by
instruction decoder 15, andinstruction selector 14, by referring to the jump destination address held in flip-flop 32, reads the jump destination instruction from jump destinationinstruction storage unit 31 and outputs it toinstruction decoder 15. In the process, the address of the instruction next to the jump destination instruction is applied to a program counter not shown, and this address is issued. - In cycle T4, the jump destination instruction is executed, the instruction next to the jump destination instruction is fetched, while at the same time the address of the next instruction but one is issued. In cycle T5 and subsequent cycles, a similar pipelining process is executed.
- The instruction next to the jump instruction is stored in
instruction queue 12. Even in the case where the branch conditions of the conditional branch instruction fail to be met, therefore, the pipelining process similar to that shown in the upper half ofFIG. 9 can be carried out by applying the particular instruction toinstruction decoder 15. - The lower half of
FIG. 9 , on the other hand, indicates the conventional jump process. In cycle T1, the address ofinstruction 1 is issued, and in cycle T2, the address ofinstruction 2 is issued while at thesame time instruction 1 is fetched.Instruction 1 is a jump instruction requiring the branch prediction. - In cycle T3, a jump instruction is decoded by
instruction decoder 15. In this case, it is assumed that the branch prediction fails and the instruction of the branch destination is required to be fetched again. In this cycle, the instruction stored ininstruction queue 12 is invalidated and processed as a NOP (non-operation) instruction. - In cycle T4, the NOP instruction is executed and the jump destination instruction is fetched, while at the same time the address of the instruction next to the jump destination instruction is issued. In cycle T5 and subsequent cycles, a similar pipelining process is executed. In this way, a one-cycle delay occurs as compared with the jump instruction free of penalty.
- Even in the case of a jump instruction free of penalty, the jump destination instruction may be required to be fetched, as shown in the lower half of
FIG. 9 , unless the jump instruction is held in jump destinationinstruction storage unit 31. In such a case, a one-cycle delay occurs. - As explained above, with the processor according to this embodiment, in the case where a jump destination instruction is held in jump destination
instruction storage unit 31, the particular jump destination instruction is applied toinstruction decoder 15. Thus, the penalty in the cycle of jump can be eliminated, and the delay due to the irregularities of the pipeline is prevented. - Specifically, according to the prior art including the branch prediction, a failure of the branch prediction results in the requirement of a repeated instruction fetch, and therefore a jump penalty occurs. In the processor according to this embodiment, on the other hand, no jump penalty occurs as long as a jump destination instruction is held in jump destination
instruction storage unit 31. - Although the present invention has been described and illustrated in detail, it is clearly understood that the same is by way of illustration and example only and is not to be taken by way of limitation, the spirit and scope of the present invention being limited only by the terms of the appended claims.
Claims (5)
1-13. (canceled)
14. A data processor comprising:
an instruction fetch unit;
an instruction decode unit;
an instruction execute unit; and
an instruction interpretation unit;
wherein the instruction fetch unit fetches an instruction from a memory, the instruction having at least a first part and a second part,
wherein the instruction interpretation unit interprets a binary pattern in the first part of the instruction as an instruction package containing a plurality of instructions by referring to a converting table when the second part of the instruction indicates a first status, or as a single instruction when the second part of the instruction indicates a second status,
wherein the instruction decode unit decodes each of the plurality of instructions of the instruction package interpreted by the instruction interpretation unit when the second part of the instruction indicates the first status, or decodes the single instruction when the second part of the instruction indicates the second status,
wherein the instruction execution unit executes the plurality of instructions or the single instruction decoded by the instruction decode unit, and
wherein the converting table is supplied from the outside data processor.
15. A data processor according to claim 14 ,
wherein the instruction execute unit executes the plurality of instructions whether in serial or in parallel in accordance with a third part of the instruction.
16. A data processor being adapted to perform each functions following:
an instruction fetching function that fetches an instruction from a memory,
an instruction interpreting function that interprets a binary pattern in a first part of the instruction as a plurality of instructions by referring a converting table when a second part of the instruction indicates a first status, or as a single instruction when the second part of the instruction indicates a second status,
an instruction decoding function that decodes each of the plurality of instructions when the second part of the instruction indicates the first status, or decodes the single instruction when the second part of the instruction indicates the second status,
an instruction executing function executes each of the plurality of instructions or the single instruction, and
a reading function that reads the converting table from outside the data processor.
17. A data processor according to claim 16 ,
wherein the instruction execution function that executes the plurality of instructions whether in serial or in parallel in accordance with a third part of the instruction.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/292,799 US20090138685A1 (en) | 2005-09-29 | 2008-11-26 | Processor for processing instruction set of plurality of instructions packed into single code |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005-284145 | 2005-09-29 | ||
JP2005284145A JP4986431B2 (en) | 2005-09-29 | 2005-09-29 | Processor |
US11/520,616 US7473293B2 (en) | 2005-09-29 | 2006-09-14 | Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator |
US12/292,799 US20090138685A1 (en) | 2005-09-29 | 2008-11-26 | Processor for processing instruction set of plurality of instructions packed into single code |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/520,616 Continuation US7473293B2 (en) | 2005-09-29 | 2006-09-14 | Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator |
Publications (1)
Publication Number | Publication Date |
---|---|
US20090138685A1 true US20090138685A1 (en) | 2009-05-28 |
Family
ID=37895576
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/520,616 Active 2027-01-25 US7473293B2 (en) | 2005-09-29 | 2006-09-14 | Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator |
US12/292,799 Abandoned US20090138685A1 (en) | 2005-09-29 | 2008-11-26 | Processor for processing instruction set of plurality of instructions packed into single code |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/520,616 Active 2027-01-25 US7473293B2 (en) | 2005-09-29 | 2006-09-14 | Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator |
Country Status (2)
Country | Link |
---|---|
US (2) | US7473293B2 (en) |
JP (1) | JP4986431B2 (en) |
Families Citing this family (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4444305B2 (en) * | 2007-03-28 | 2010-03-31 | 株式会社東芝 | Semiconductor device |
JP2009288967A (en) * | 2008-05-28 | 2009-12-10 | Fujitsu Ten Ltd | Signal processing apparatus, display control device, and register setting method for signal processing apparatus |
US20110283089A1 (en) * | 2009-01-28 | 2011-11-17 | Harshal Ingale | modularized micro processor design |
US8970608B2 (en) * | 2010-04-05 | 2015-03-03 | Nvidia Corporation | State objects for specifying dynamic state |
US9378019B2 (en) | 2011-04-07 | 2016-06-28 | Via Technologies, Inc. | Conditional load instructions in an out-of-order execution microprocessor |
US9645822B2 (en) | 2011-04-07 | 2017-05-09 | Via Technologies, Inc | Conditional store instructions in an out-of-order execution microprocessor |
US9032189B2 (en) | 2011-04-07 | 2015-05-12 | Via Technologies, Inc. | Efficient conditional ALU instruction in read-port limited register file microprocessor |
US9336180B2 (en) | 2011-04-07 | 2016-05-10 | Via Technologies, Inc. | Microprocessor that makes 64-bit general purpose registers available in MSR address space while operating in non-64-bit mode |
US9043580B2 (en) | 2011-04-07 | 2015-05-26 | Via Technologies, Inc. | Accessing model specific registers (MSR) with different sets of distinct microinstructions for instructions of different instruction set architecture (ISA) |
US8880851B2 (en) | 2011-04-07 | 2014-11-04 | Via Technologies, Inc. | Microprocessor that performs X86 ISA and arm ISA machine language program instructions by hardware translation into microinstructions executed by common execution pipeline |
US9274795B2 (en) | 2011-04-07 | 2016-03-01 | Via Technologies, Inc. | Conditional non-branch instruction prediction |
US9141389B2 (en) | 2011-04-07 | 2015-09-22 | Via Technologies, Inc. | Heterogeneous ISA microprocessor with shared hardware ISA registers |
US9317288B2 (en) | 2011-04-07 | 2016-04-19 | Via Technologies, Inc. | Multi-core microprocessor that performs x86 ISA and ARM ISA machine language program instructions by hardware translation into microinstructions executed by common execution pipeline |
US9128701B2 (en) | 2011-04-07 | 2015-09-08 | Via Technologies, Inc. | Generating constant for microinstructions from modified immediate field during instruction translation |
US9176733B2 (en) | 2011-04-07 | 2015-11-03 | Via Technologies, Inc. | Load multiple and store multiple instructions in a microprocessor that emulates banked registers |
US8880857B2 (en) | 2011-04-07 | 2014-11-04 | Via Technologies, Inc. | Conditional ALU instruction pre-shift-generated carry flag propagation between microinstructions in read-port limited register file microprocessor |
US9898291B2 (en) | 2011-04-07 | 2018-02-20 | Via Technologies, Inc. | Microprocessor with arm and X86 instruction length decoders |
US8924695B2 (en) * | 2011-04-07 | 2014-12-30 | Via Technologies, Inc. | Conditional ALU instruction condition satisfaction propagation between microinstructions in read-port limited register file microprocessor |
US9146742B2 (en) | 2011-04-07 | 2015-09-29 | Via Technologies, Inc. | Heterogeneous ISA microprocessor that preserves non-ISA-specific configuration state when reset to different ISA |
US9244686B2 (en) | 2011-04-07 | 2016-01-26 | Via Technologies, Inc. | Microprocessor that translates conditional load/store instructions into variable number of microinstructions |
US9292470B2 (en) | 2011-04-07 | 2016-03-22 | Via Technologies, Inc. | Microprocessor that enables ARM ISA program to access 64-bit general purpose registers written by x86 ISA program |
US9672041B2 (en) | 2013-08-01 | 2017-06-06 | Andes Technology Corporation | Method for compressing variable-length instructions including PC-relative instructions and processor for executing compressed instructions using an instruction table |
JP2019095952A (en) * | 2017-11-21 | 2019-06-20 | ソニーセミコンダクタソリューションズ株式会社 | Processor, information processing device and processing method |
JP7506718B2 (en) * | 2022-08-05 | 2024-06-26 | たけおかラボ株式会社 | Processor and program for controlling pipeline processing based on jump instructions |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5632024A (en) * | 1993-06-08 | 1997-05-20 | Hitachi, Ltd. | Microcomputer executing compressed program and generating compressed branch addresses |
US5636352A (en) * | 1994-12-16 | 1997-06-03 | International Business Machines Corporation | Method and apparatus for utilizing condensed instructions |
US20020120830A1 (en) * | 2001-02-23 | 2002-08-29 | Masahito Matsuo | Data processor assigning the same operation code to multiple operations |
US20030033504A1 (en) * | 2001-08-07 | 2003-02-13 | Hiromichi Yamada | Micro-controller for reading out compressed instruction code and program memory for compressing instruction code and storing therein |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10116229A (en) * | 1996-10-09 | 1998-05-06 | Toshiba Corp | Data processor |
JP3585800B2 (en) * | 1999-01-13 | 2004-11-04 | 株式会社東芝 | Information processing equipment |
JP2002215387A (en) * | 2001-01-22 | 2002-08-02 | Mitsubishi Electric Corp | Data processor provided with instruction translator, and memory interface device |
KR20040064713A (en) * | 2001-11-26 | 2004-07-19 | 코닌클리즈케 필립스 일렉트로닉스 엔.브이. | Wlim architecture with power down instruction |
JP2004220188A (en) * | 2003-01-10 | 2004-08-05 | Sony Corp | Data processor |
JP4761808B2 (en) * | 2005-04-11 | 2011-08-31 | 株式会社東芝 | Microprocessor and control method thereof |
-
2005
- 2005-09-29 JP JP2005284145A patent/JP4986431B2/en active Active
-
2006
- 2006-09-14 US US11/520,616 patent/US7473293B2/en active Active
-
2008
- 2008-11-26 US US12/292,799 patent/US20090138685A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5632024A (en) * | 1993-06-08 | 1997-05-20 | Hitachi, Ltd. | Microcomputer executing compressed program and generating compressed branch addresses |
US5636352A (en) * | 1994-12-16 | 1997-06-03 | International Business Machines Corporation | Method and apparatus for utilizing condensed instructions |
US20020120830A1 (en) * | 2001-02-23 | 2002-08-29 | Masahito Matsuo | Data processor assigning the same operation code to multiple operations |
US20030033504A1 (en) * | 2001-08-07 | 2003-02-13 | Hiromichi Yamada | Micro-controller for reading out compressed instruction code and program memory for compressing instruction code and storing therein |
Non-Patent Citations (2)
Title |
---|
Definition of "bit stream" retreived from google.com Dictionary on 11/7/2011. * |
Definition of "stream" retreived from google.com Dictionary on 11/7/2011. * |
Also Published As
Publication number | Publication date |
---|---|
US7473293B2 (en) | 2009-01-06 |
JP2007094813A (en) | 2007-04-12 |
JP4986431B2 (en) | 2012-07-25 |
US20070074010A1 (en) | 2007-03-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7473293B2 (en) | Processor for executing instructions containing either single operation or packed plurality of operations dependent upon instruction status indicator | |
JP5313228B2 (en) | System, method and software for preloading instructions from a variable length instruction set with appropriate predecoding | |
US20070106889A1 (en) | Configurable instruction sequence generation | |
US20120204008A1 (en) | Processor with a Hybrid Instruction Queue with Instruction Elaboration Between Sections | |
US7546442B1 (en) | Fixed length memory to memory arithmetic and architecture for direct memory access using fixed length instructions | |
US9317285B2 (en) | Instruction set architecture mode dependent sub-size access of register with associated status indication | |
JP2006313422A (en) | Calculation processing device and method for executing data transfer processing | |
WO2008112207A2 (en) | Software programmable timing architecture | |
US20230084523A1 (en) | Data Processing Method and Device, and Storage Medium | |
US20050138327A1 (en) | VLIW digital signal processor for achieving improved binary translation | |
US20200326940A1 (en) | Data loading and storage instruction processing method and device | |
US8635434B2 (en) | Mathematical operation processing apparatus for performing high speed mathematical operations | |
CN112559037B (en) | Instruction execution method, unit, device and system | |
JP5335440B2 (en) | Early conditional selection of operands | |
CN112540792A (en) | Instruction processing method and device | |
JP2009524167A5 (en) | ||
US9361109B2 (en) | System and method to evaluate a data value as an instruction | |
JP4502532B2 (en) | Data processing device | |
US7543135B2 (en) | Processor and method for selectively processing instruction to be read using instruction code already in pipeline or already stored in prefetch buffer | |
CN116339832A (en) | Data processing device, method and processor | |
US11210091B2 (en) | Method and apparatus for processing data splicing instruction | |
US6829700B2 (en) | Circuit and method for supporting misaligned accesses in the presence of speculative load instructions | |
US6988121B1 (en) | Efficient implementation of multiprecision arithmetic | |
CN112130899A (en) | Stack computer | |
US20050138331A1 (en) | Direct memory access unit with instruction pre-decoder |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: RENESAS ELECTRONICS CORPORATION, JAPAN Free format text: CHANGE OF NAME;ASSIGNOR:NEC ELECTRONICS CORPORATION;REEL/FRAME:024990/0059 Effective date: 20100401 Owner name: NEC ELECTRONICS CORPORATION, JAPAN Free format text: MERGER;ASSIGNOR:RENESAS TECHNOLOGY CORP.;REEL/FRAME:024990/0098 Effective date: 20100401 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |