WO2015073646A1 - Vector processing engine employing reordering circuitry in data flow paths between vector data memory and execution units, and related method - Google Patents
Vector processing engine employing reordering circuitry in data flow paths between vector data memory and execution units, and related method Download PDFInfo
- Publication number
- WO2015073646A1 WO2015073646A1 PCT/US2014/065412 US2014065412W WO2015073646A1 WO 2015073646 A1 WO2015073646 A1 WO 2015073646A1 US 2014065412 W US2014065412 W US 2014065412W WO 2015073646 A1 WO2015073646 A1 WO 2015073646A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- vector data
- data sample
- sample set
- vector
- input
- Prior art date
Links
- 239000013598 vector Substances 0.000 title claims abstract description 2634
- 238000012545 processing Methods 0.000 title claims abstract description 626
- 238000000034 method Methods 0.000 title claims abstract description 42
- 230000008569 process Effects 0.000 claims description 17
- 238000012805 post-processing Methods 0.000 abstract description 23
- 241001442055 Vipera berus Species 0.000 description 109
- 238000010586 diagram Methods 0.000 description 65
- 238000009825 accumulation Methods 0.000 description 51
- 230000035508 accumulation Effects 0.000 description 50
- 238000006243 chemical reaction Methods 0.000 description 50
- 108010003272 Hyaluronate lyase Proteins 0.000 description 32
- 230000002596 correlated effect Effects 0.000 description 27
- 230000007480 spreading Effects 0.000 description 14
- 238000004891 communication Methods 0.000 description 11
- 230000006870 function Effects 0.000 description 10
- 238000013461 design Methods 0.000 description 9
- 238000001228 spectrum Methods 0.000 description 7
- 230000000875 corresponding effect Effects 0.000 description 6
- 230000001419 dependent effect Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 239000000945 filler Substances 0.000 description 5
- 230000001934 delay Effects 0.000 description 4
- 238000007781 pre-processing Methods 0.000 description 4
- 230000000644 propagated effect Effects 0.000 description 4
- 230000003111 delayed effect Effects 0.000 description 3
- 238000001914 filtration Methods 0.000 description 3
- 238000000926 separation method Methods 0.000 description 3
- 238000007792 addition Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000009977 dual effect Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 239000002245 particle Substances 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 101100328463 Mus musculus Cmya5 gene Proteins 0.000 description 1
- RTAQQCXQSZGOHL-UHFFFAOYSA-N Titanium Chemical compound [Ti] RTAQQCXQSZGOHL-UHFFFAOYSA-N 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 239000000872 buffer Substances 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000005314 correlation function Methods 0.000 description 1
- 239000004744 fabric Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- VWDWKYIASSYTQR-UHFFFAOYSA-N sodium nitrate Chemical compound [Na+].[O-][N+]([O-])=O VWDWKYIASSYTQR-UHFFFAOYSA-N 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30181—Instruction operation extension or modification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored program computers
- G06F15/80—Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
- G06F15/8053—Vector processors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/76—Arrangements for rearranging, permuting or selecting data according to predetermined rules, independently of the content of the data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30003—Arrangements for executing specific machine instructions
- G06F9/30007—Arrangements for executing specific machine instructions to perform operations on data operands
- G06F9/3001—Arithmetic instructions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30003—Arrangements for executing specific machine instructions
- G06F9/30007—Arrangements for executing specific machine instructions to perform operations on data operands
- G06F9/30032—Movement instructions, e.g. MOVE, SHIFT, ROTATE, SHUFFLE
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30003—Arrangements for executing specific machine instructions
- G06F9/30007—Arrangements for executing specific machine instructions to perform operations on data operands
- G06F9/30036—Instructions to perform operations on packed data, e.g. vector, tile or matrix operations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3885—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units
- G06F9/3887—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units controlled by a single instruction for multiple data lanes [SIMD]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3885—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units
- G06F9/3887—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units controlled by a single instruction for multiple data lanes [SIMD]
- G06F9/38873—Iterative single instructions for multiple data lanes [SIMD]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3885—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units
- G06F9/3893—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units controlled in tandem, e.g. multiplier-accumulator
- G06F9/3895—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units controlled in tandem, e.g. multiplier-accumulator for complex operations, e.g. multidimensional or interleaved address generators, macros
- G06F9/3897—Concurrent instruction execution, e.g. pipeline or look ahead using a plurality of independent parallel functional units controlled in tandem, e.g. multiplier-accumulator for complex operations, e.g. multidimensional or interleaved address generators, macros with adaptable data path
Definitions
- VPEs PROCESSING ENGINES
- VPEs EMPLOYING FORMAT CONVERSION CIRCUITRY IN DATA FLOW PATHS BETWEEN VECTOR DATA MEMORY AND EXECUTION UNITS TO PROVIDE IN-FLIGHT FORMAT-CONVERTING OF INPUT VECTOR DATA TO EXECUTION UNITS FOR VECTOR PROCESSING OPERATIONS, AND RELATED VECTOR PROCESSOR SYSTEMS AND METHODS
- VPEs EMPLOYING FORMAT CONVERSION CIRCUITRY IN DATA FLOW PATHS BETWEEN VECTOR DATA MEMORY AND EXECUTION UNITS TO PROVIDE IN-FLIGHT FORMAT-CONVERTING OF INPUT VECTOR DATA TO EXECUTION UNITS FOR VECTOR PROCESSING OPERATIONS, AND RELATED VECTOR PROCESSOR SYSTEMS AND METHODS
- VECTOR PROCESSING ENGINES VPEs
- VPEs VECTOR PROCESSING ENGINES
- VPEs EMPLOYING MERGING CIRCUITRY IN DATA FLOW PATHS BETWEEN EXECUTION UNITS AND VECTOR DATA MEMORY TO PROVIDE IN-FLIGHT MERGING OF OUTPUT VECTOR DATA STORED TO VECTOR.
- DATA MEMORY, AND RELATED VECTOR PROCESSING INSTRUCTIONS, SYSTEMS, AND METHODS 124363U1 filed on November 15, 2013 and incorporated herein by reference in its entirety.
- the field of the disclosure relates to vector processors and related systems for processing vector and scalar operations, including single instruction, multiple data (SIMD) processors and multiple instruction, multiple data (MIMD) processors.
- SIMD single instruction, multiple data
- MIMD multiple instruction, multiple data
- wireless computing systems are fast becoming one of the most prevalent technologies in the digital information arena. Advances in technology have resulted in smaller and more powerful wireless communications devices.
- wireless computing devices commonly include portable wireless telephones, personal digital assistants (PDAs), and paging devices that are small, lightweight, and easily carried by users.
- portable wireless telephones such as cellular telephones and Internet Protocol (IP) telephones
- IP Internet Protocol
- a wireless telephone may include a digital still camera, a digital video camera, a digital recorder, and/or an audio file player.
- wireless telephones can include a web interface thai can be used to access the Internet.
- wireless communications devices may include complex processing resources for processing high speed wireless communications data according to designed wireless communications technology standards (e.g., code division multiple access (CDMA), wideband CDMA (WCDMA), and long term evolution (LTE)). As such, these wireless communications devices include significant computing capabilities,
- CDMA code division multiple access
- WCDMA wideband CDMA
- LTE long term evolution
- wireless computing devices become smaller and more powerful, they become increasingly resource constrained. For example, screen size, amount of available memory and tile system space, and amount of input and output capabilities may be limited by the small size of the device. Further, battery size, amount of power provided by the battery, and life of the battery are also limited. One way to increase the battery life of the device is to design processors that consume less power.
- baseband processors may be employed for wireless communications devices that include vector processors.
- Vector processors have a vector architecture that provides high-level operations that work on vectors, i.e. arrays of data.
- Vector processing involves fetching a vector instruction once and then executing the vector instruction multiple times across an entire array of data elements, as opposed to executing the vector instruction on one set of data and then re-fetching and decoding the vector instruction for subsequent elements within the vector. This process allows for a reduction in the energy required to execute a program, because among other factors, each vector instruction needs to be fetched fewer times. Since vector instructions operate on long vectors over multiple clock cycles at the same time, a high degree of parallelism is achievable with simple in-order vector instruction dispatch,
- FIG 1 illustrates an exemplary baseband processor 10 that may be employed in a computing device, such as a wireless computer device.
- the baseband processor 10 includes multiple processing engines (PEs) 12, each dedicated to providing function-specific vector processing for specific applications, in this example, six (6) separate PEs 12(G)-12(5) are provided in the baseband processor 10.
- the PEs 12(0)- 12(5) are each configured to provide vector processing for fixed X-bit wide vector data 14 provided from a shared memory 16 to the PEs 12(0)-12(5).
- the vector data 14 could be 512 bits wide.
- the vector data 14 ca be defined in smaller multiples of X-bit width vector data sample sets 18(0)-18(Y) (e.g., 16-bit and 32-bit sample sets).
- each PE 12(0)- 12(5) are capable of providing vector processing on multiple vector data sample sets 18 provided in parallel to the PEs 12(0)- 12(5) to achieve a high degree of parallelism.
- Each PE 12(0)- 12(5) may include a vector register file (VR) for storing the results of a vector instruction processed on the vector data 14.
- VR vector register file
- Each PE 12(0)- 12(5) in the baseband processor 10 in Figure 1 includes specific, dedicated circuitry and hardware specifically designed to efficie tly perform specific types of fixed operations.
- the baseband processor 10 in Figure 1 includes separate WCDMA PEs 12(0), 12(1) and LTE PEs 12(4), 12(5), because WCDMA and LTE involve different types of specialized operations.
- each of the PEs 12(0), 12(1), 12(4), 12(5) can be designed to include specialized, dedicated circuitry that is specific to frequently performed functions for WCDMA and LTE for highly efficient operation. This design is in contrast to scalar processing engines that include more general circuitry and hardware designed to be flexible to support a larger number of unrelated operations, but in a less efficient manner.
- Certain wireless baseband operations require subsequent processing operations to be performed on reordered data samples determined from previous processing operations. Tims, the output data samples resulting from previous processing operations must be reordered and restored after being initially stored in vector data memory, so that the data samples are reordered from subsequent processing operations. For example, subsequent processing operations may require previously processed data samples to be interleaved when stored back in data memory. As another example, subsequent processing operations may require previously processed data samples to be de-interleaved when stored back in data memory. For example, in CDMA processing operations, data samples representing a signal may need to be stored and interleaved according to even (i.e., on-time) and odd (i.e., late) phases of the signal.
- vector processors can include circuitry that performs post-processing reordering of output vector data from execution units after the output vector data is stored in vector data memory.
- the post-processed output vector data samples stored in vector data memory are fetched from the vector data memory, reordered, and stored back in the vector data memory. This post-processing delays the subsequent processing of the reordered vector data samples by execution units, and causes computational components in the execution units to be underutilized.
- Embodiments disclosed herein include vector processing engines (VPEs) employing reordering circuitry in data flow paths between execution units and vector data memory to provide in-flight reordering of output vector data stored to yector data memory.
- VPEs vector processing engines
- Related vector processor systems and methods are also disclosed.
- Reordering circuitry is provided in data flow paths between execution units and vector data memory in the VPE.
- the reordering circuitry is configured to reorder output vector data sample sets from execution units as a result of performing vector processing operations in-flight while the output vector data sample sets are being provided over the data flow paths from the execution units to the vector data memory to be stored.
- In-flight reordering of output vector data sample sets means that the output vector data sample set provided by execution units is reordered before being stored in yector data memory, so that the output vector data sample set is stored in the vector data memory in a reordered format
- the reordering of output vector data sample sets may include interleaving or de-interleaving of output vector data sample sets stored in the vector data memory, in this manner, the output vector data sample sets are stored in the reordered format in the vector data memory without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units.
- the efficiency of the data flow paths in the VPE are not limited by the reordering of output vector data sample sets.
- the subsequent vector processing in the execution units is only limited by computational resources rather than by data flow limitations when vector data outputs are to he stored in reordered format in vector data memory,
- a VPE configured to in-flight reorder a resultant output vector data sample set generated by at least one execution unit executing a vector processing operation.
- the VPE comprises at least one vector data file.
- the vector data file(s) is configured to provide a fetched input vector data sample set in at least one input data flow path for a vector processing operation.
- the vector data file(s) is also configured to receive a reordered resultant output vector data sample set from at least one output data flow path to be stored.
- the VPE also comprises at feast one execution unit provided in the at least one input data flow path.
- the execution unit(s) is configured to receive the input vector data sample set on the at least one input data flow path.
- the execution unit(s) is also configured to execute the vector processing operation on the input vector data sample set to provide a resultant output vector data sample set on the at least one output data flow path.
- the VPE also comprises at least one reordering circuitry.
- the reordering circuitry is configured to receive the resultant output vector data sample set on the at least one output data flow path.
- the reordering circuitry is also configured to reorder the resultant output vector data sample set into the reordered resultant output vector data sample set without the resultant output vector data sample set being stored in the at least one vector data file.
- Tire reordering circuitry is also configured to provide the reordered resultant output vector data sample set on the at least one output data flow path.
- a VPE configured to in-flight reorder a resultant output vector data sample set generated by at least one execution unit executing a vector processing operation.
- the VPE comprises at least one vector data file means.
- the vector data file means comprises a means for fetching an input vector data sample set in at least one input data flow path means for a vector processing operation.
- Tire vector data file means also comprises a means for receiving a reordered resultant output vector data sample set from at least one output data flow path means to be stored.
- the VPE also comprises at least one execution unit means provided in the at least one input data flow path means.
- the execution unit means comprises a means for receiving the input vector data sample set on the at least one input data flow path means.
- the execution unit means also comprises an execution means for executing the vector processing operation on the input vector data sample set to provide a resultant output vector data sample set on the at least one output data flow path means.
- the VPE also comprises at least one reordering means.
- the reordering means comprises a means for receiving the resultant output vector data sample set on the at least one output data flow path means.
- the reordering means also comprises a reordering means for reordering the resultant output vector data sample set into the reordered resultant output vector data sample set without the resultant output vector data sample set being stored in the at least one vector data file means.
- the reordering means also comprises a means for providing the reordered resultant output vector data sample set on the at least one output data flow path means.
- a method of in-flight reordering of a resultant output vector data sample set generated by at least one execution unit executing a vector processing operation comprises providing a fetched input vector data sample set from at least one vector data file in at least one input data flow path for a vector processing operation.
- the method also comprises receiving the input vector data sample set on the at least one input data flow path in at least one execution unit provided in the at least one input data flow path.
- the method also comprises executing the vector processing operation on the received input vector data sample set to provide a resultant output vector data sample set on at least one output data flow path.
- the method also comprises reordering the resultant output vector data sample set into the reordered resultant output vector data sample set in the at least one output data flow path without the resultant output vector data sample set being stored in the at least one vector data file.
- the method also comprises storing the reordered resultant output vector data sample set in the at least one output data flow path in the at least one vector data file.
- FIG. 1 is a schematic diagram of an exemplary vector processor that includes multiple vector processing engines (VPEs), each dedicated to providing function-specific vector processing for specific applications;
- VPEs vector processing engines
- FIG. 2 is a schematic diagram of an exemplary baseband processor that includes a VPE having programmable data path configurations, so that common circuitry and hardware provided in the VPE can be programmed in multiple modes to perform specific types of vector operations in a highly efficient manner for multiple applications or technologies, without a requirement to provide separate VPEs;
- FIG. 3 is a schematic diagram of a discrete finite impulse response (FIR) filter that may be provided in a filter vector processing operation supported by a VPE;
- FIR discrete finite impulse response
- FIG. 4 is a schematic diagram of an exemplary VPE employing tapped- delay lines to receive and provide shifted input vector data sample sets to execution units to be processed with filter coefficient data for providing precision filter vector processing operations with reduced re-fetching and power consumption;
- FIG. 5 is a flowchart illustrating an exemplary filter vector processing operation that can be performed in the VPE in Figure 4 according to an exemplary filter vector instruction
- FIG. 6A is a schematic diagram of filter tap coefficients stored in a register file in the V PE of Figure 4;
- Figure 6B is a schematic diagram of exemplary input vector data sample sets stored in a vector data file in the VPE in Figure 4;
- Figure 7 is a schematic diagram illustrating an exemplary tapped-deiay line and optional shadow tapped-delay line that can be provided in the VPE in Figure 4, wherein the exemplary tapped-delay lines each comprise a plurality of pipeline registers for receiving and providing, to execution units, an input vector data sample set from vector data memory and a shifted input vector data sample set, during filter vector processing operations performed by the VPE;
- Figure 8 is a schematic diagram illustrating more exemplary detail of the tapped-delay lines in Figure 7, illustrating exemplary detail of pipeline registers in data lanes, including intra-lane and inter-lane routing among the pipeline registers for shifting of input vector data samples in an input vector data sample set during a filter vector processing operation;
- Figure 9A is a schematic diagram of an input vector data sample set initially stored in a primary tapped-delay line in the VPE of Figure 4 as part of a first filter tap execution of an exemplary eight (8) tap filter vector processing operation;
- Figure 9B is a schematic diagram of filter tap coefficients stored in a register file and shadow input vector data sample set initially stored in a shadow tapped-delay line in the VPE of Figure 4 as part of a first filter tap execution of the exemplary eight (8) tap filter vector processing operation filter vector processing operation illustrated in Figure 9A;
- Figure 9C is a schematic diagram of shifted input vector data sample sets stored in the primary tapped-delay line and the shadow tapped-delay line, and the filter tap coefficients stored in a register file, in the VPE of Figure 4 as part of a second filter tap execution of the exemplary eight (8) tap filler vector processing operation;
- Figure 9D is a schematic diagram of the shifted input vector data sample sets stored in the primary tapped-delay line and the shadow tapped-delay line, and the filter tap coefficients stored in the register file, in the VPE of Figure 4 as part of an eighth filter tap execution of the exemplary eight (8) tap filler vector processing operation;
- Figure 10 is a schematic diagram of contents of accumulators of the execution units in the VPE of Figure 4 after the exemplary eight (8) tap filter vector processing operation has been fully executed;
- Figure 1 1 is a schematic diagram of an exemplary VPE employing tapped- delay lines to receive and provide shifted input vector data sample sets to execution units to be processed with sequence number data for providing precision correlation/covariance vector processing operations with reduced re-fetching and power consumption;
- Figures 12.A and 12.B are flowcharts illustrating exemplary correlation/covariance vector processing operations that can be performed in parallel in the VPE in Figure 1 1 with fetched interleaved on-time and late input vector data sample sets according to an exemplary correlation/covariance vector processing operation;
- Figure 13 is a schematic diagram of a correlation/covariance input vector data sample set stored in a register file in the VPE of Figure 11 ;
- Figure 14 is a schematic diagram illustrating an exemplary tapped-delay line and optional shadow tapped-delay line that can be provided in the VPE in Figure l i, wherein the exemplary tapped-delay lines each comprise a plurality of pipeline registers for receiving and providing, to execution units, an input vector data sample set from vector data memory and a shifted input vector data sample set, during a correlation/covariance vector processing operation performed by the VPE;
- Figure 15A is a schematic diagram of the input vector data sample set from the vector data file initially provided in the primary tapped-delay line in the VPE of Figure l i as part of a first processing stage of a correlation covariance vector processing operation;
- Figure 15B is a schematic diagram of a shadow input vector data sample set from the vector data file initially stored in the shadow tapped-delay line in the VPE of Figure 1 1 as part of a first processing stage of a correlation/covariance vector processing operation;
- Figure 15C is a schematic diagram of the shifted input vector data sample sets stored in the primary tapped-delay line and the shadow tapped-delay line and the shifted input vector data sample set stored in the register file, in the VPE of Figure 1 1 as part of a second processing stage of a correlation/covariance vector processing operation;
- Figure 15D is a schematic diagram of the shifted input vector data sample sets stored in the primary tapped-delay line and the shadow tapped-delay line, and the shifted input vector data sample set stored in the register file, in the VPE of Figure 1 1 as part of a fourteenth processing stage of a correlation/covariance vector processing operation;
- Figure 16 is a schematic diagram of contents of accumulators of the execution units in the VPE of Figure 1 1 after the exemplary correlation/covariance vector processing operation has been fully executed;
- Figure 17A is a diagram of exemplary vector data files showing a stored resultant filter output vector data sample set stored in the real and imaginary components of resultant filter output vector data samples stored separately;
- Figure I7B is a diagram of exemplary vector data files showing a stored resultant filter output vector data sample set stored with its even and odd resultant filter output vector data samples stored separately;
- Figures 18A and 18B are diagrams of exemplary interleaved vector data samples of a vector data sample set stored in a vector data file of a VPE in signed complex sixteen (16) bit format and complex eight (8) bit format, respectively;
- Figure 19 is a schematic diagram of an exemplary VPE employing format conversion circuitry configured to provide in-flight format-converting of input vector data sample set in at least one input data flow path between a vector data file and at least one execution unit without the input vector data sample set being required to be re- fetched from the vector data file, to provide a format-converted input vector data sample set to the at least one execution unit for executing a vector processing operation;
- Figure 20 is a flowchart illustrating exemplary in-flight format-converting of an input vector data sample set in the at least one input data flow path between the vector data file and the at least one execution unit that can be performed in the VPE of Figure 19;
- Figure 21 is a schematic diagram of an exemplary format conversion circuitry provided between tapped-delay lines and execution units in the VPE of Figure 19, wherein the format conversion circuitry is configured to provide in-flight format- converting of the input vector data sample set provided by the tapped-delay lines in the input data flow path to the execution units;
- FIG. 22 illustrates an exemplary vector instruction data format to provide programming to the VPE of Figure 19 to provide in-flight format conversion of the input vector data sample set in an input data flow path before receipt at execution units;
- Figure 23 is a schematic diagram of an exemplary VPE employing reordering circuitry configured to provide in-flight reordering of a resultant output vector data sample set in at least one output data flow path between at least one execution unit and at least one vector data file without the resultant output vector data sample set being stored in the at least one vector data file, to provide and store a reordered resultant output data sample set;
- Figure 24 is a flowchart illustrating exemplary in- flight de-interleaving of an output vector data sample set in the at least one output data flow path between the vector data file and the at least one execution unit in the VPE of Figure 23 to be stored in reordered form in the vector data file;
- Figure 25 is a schematic diagram of an exemplary VPE employing reordering circuitry in output data flow paths between executions units and a vector data file to provide in-flight reordering of output vector data sample sets stored to the vector data file;
- Figure 26A is a diagram of an exemplary vector data sample sequence representing a communications signal
- Figure 26B is a diagram of an exemplary code division multiple access (CDMA) chip sequence
- Figure 26C is a diagram of the vector data sample sequence in Figure 26A after being spread with the CDMA chip sequence in Figure 26B;
- Figure 26D is a diagram of despreading the spread vector data sample sequence in Figure 26C with the CDMA chip sequence in Figure 26B to recover the original vector data sample sequence in Figure 26A;
- Figure 27 is a schematic diagram of an exemplary VPE employing despreading circuitry configured to provide despreading of a resultant output vector data sample set in at least one output data flow path between at least one execution unit and at least one vector data file without the resultant output vector data sample set being stored in the at least one vector data file, to provide and store a despread resultant output vector data sample set;
- Figure 28 is a flowchart illustrating exemplary despreading of a resultant output vector data sample set in the at least one output data flow path between the at least one vector data file and the at least one execution unit in the VPE of Figure 27, to provide and store the despread resultant output vector data sample set in the at least one vector data file;
- Figure 29 is a schematic diagram of an exemplary despreading circuitry in output data flow paths between at least one execution unit and at least one vector data file in the VPE of Figure 27 to provide despreading of resultant output vector data sample sets to provide and store the despread resultant output vector data sample sets in the at least one vector data file;
- Figure 30 is a diagram of exemplary vector data samples to be merged, and illustrating the merged resultant vector data samples
- Figure 31 is a schematic diagram of an exemplary VPE employing merge circuitry configured to provide merging of a resultant output vector data sample set in at least one output data flow path between at least one execution unit and at least one vector data file without the resultant output vector data sample set being stored in the at least one vector data file, to provide and store a merged resultant output vector data sample set;
- Figure 32 is a flowchart illustrating exemplary add-merging of a resultant output vector data sample set in the at least one output data flow path between the vector data file and the at least one execution unit in the VPE of Figure 31 , to provide and store the add-merged resultant output vector data sample set in the vector data file;
- Figure 33 is a schematic diagram of an exemplary merge circuitry in output data flow paths between executions units and a vector data file in the VPE of Figure 31 to provide add-merging of resultant output vector data sample sets and storing of the add-merged resultant output vector data sample set in the vector data file;
- FIG. 34 is a schematic diagram of an exemplary merge circuitry in output data flow paths between executions units and a vector data file in the VPE of Figure 31 to provide maximum/minimum merging of resultant output vector data sample sets and storing of the maximunu'minimum-merged resultant output vector data sample sets in the vector data file;
- Figure 35 is a schematic diagram of exemplary vector processing stages that can be provided in a VPE, wherein certain of the vector processing stages include exemplary vector processing blocks having programmable data path configurations;
- Figure 36 is a flowchart illustrating exemplary vector processing of multiplier blocks and accumulator blocks, each having programmable data path configurations and provided in different vector processing stages in the exemplary VPE of Figure 35;
- Figure 37 is a more detailed schematic diagram of a plurality of multiplier blocks provided in a vector processing stage of the VPE of Figure 35, wherein the plurality of multiplier blocks each have programmable data path configurations, so that the plurality of multiplier blocks can be programmed in multiple modes to perform specific, different types of vector multiply operations;
- Figure 38 is a schematic diagram of internal components of a multiplier block among the plurality of multiplier blocks in Figure 37 having programmable data paths configurations capable of being programmed to provide multiply operations for 8- bit by 8-bit input vector data sample sets and 16-bit by 16-bit input vector data sample sets;
- Figure 39 is a generalized schematic diagram of a multiplier block and accumulator block in the VPE of Figure 38, wherein the accumulator block employs a carry-save accumulator structure employing redundant carry -save format to reduce carry propagation:
- Figure 40 is a detailed schematic diagram of exemplary internal components of the accumulator block of Figure 39, which is provided in the VPE of Figure 35, wherein the accumulator block has programmable data path configurations, so that the accumulator block can be programmed in multiple modes to perform specific, different types of vector accumulate operations with redundant carry-save format; and
- Figure 41 is a block diagram of an exemplary processor-based system that can include a vector processor that can include the VPEs disclosed herein to provide the vector processing circuits and vector processing operations, according to the embodiments disclosed herein.
- Embodiments disclosed herein include vector processing engines (VPEs) employing reordering circuitry in data flow paths between execution units and vector data memory to provide in-flight reordering of output vector data stored to vector data memory.
- VPEs vector processing engines
- Related vector processor systems and methods are also disclosed.
- Reordering circuitry is provided in data flow paths between execution units and vector data memory in the VPE.
- the reordering circuitry is configured to reorder output vector data sample sets from execution units as a result of performing vector processing operations in-flight while the output vector data sample sets are being provided over the data flow paths from the execution units to the vector data memory to be stored.
- In-flight reordering of output vector data sample sets means the output vector data sample set provided by execution units is reordered before being stored in vector data memory, so that the output vector data sample set is stored in vector data memory in reordered format.
- the reordering of output vector data sample sets may include interleaving or de-interleaving of output vector data sample sets stored in the vector data memory, in this manner, the output vector data sample sets are stored in the reordered format in the vector data memory without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units.
- the efficiency of the data flow paths in the VPE are not limited by the reordering of output vector data sample sets.
- the subsequent vector processing in the execution units is only limited by computational resources rather than by data flow limitations when vector data outputs are to he stored in reordered format in vector data memory,
- FIG. 2 is a schematic diagram of a baseband processor 20 that includes an exemplary vector processing unit 22, also referred to as a vector processing engine (VPE) 22.
- VPE vector processing engine
- the VPE 22 includes execution units 84 and other particular exemplary circuitry and functionality to provide vector processing operations including the exemplary vector processing operations disclosed herein.
- the baseband processor 20 and its VPE 22 can be provided in a semiconductor die 2.4.
- the baseband processor 20 includes a common VPE 22 that includes programmable data paths 26 that can be programmed to provide different programmable data path configurations.
- the programmable data paths 26 between the execution units 84 and vector data files 82 in the VPE 22 can be programmed and reprogrammed to provide different, specific types of vector processing operations in different operation modes without the requirement to provide separate VPEs 22 in the baseband processor 20.
- the baseband processor 20 in this non-limiting example is a 512-bit vector processor.
- the baseband processor 20 includes components in addition to the VPE 22 to support the VPE 22 providing vector processing in the baseband processor 20.
- the baseband processor 20 includes vector registers, also known as vector data files 82, that are configured to receive and store vector data 30 from a vector unit data memory (LMEM) 32.
- LMEM vector unit data memory
- the vector data 30 is X bits wide, with 'X' defined according to design choice (e.g., 512 bits).
- the vector data 30 may be divided into vector data sample sets 34.
- the vector data 30 may be 256- bits wide and may comprise smaller vector data sample sets 34(Y)-34(0). Some vector data sample sets 34(Y)-34(0) can be 16-bits wide as an example, and others of the vector data sample sets 34(Y)-34(0) can be 32-bits wide.
- the VPE 22 is capable of providing vector processing on certain chosen vector data sample sets 34(Y)-34(0) provided in parallel to the VPE 22 to achieve a high degree of parallelism.
- the vector data files 82 are also configured to store results generated when the VPE 22. processes the vector data 30. In certain embodiments, the VPE 22 is configured to not store intermediate vector processing results in the vector data files 82 to reduce register writes to provide faster vector instruction execution times. This configuration is opposed to scalar instructions executed by scalar processing engines that store intermediate results in registers, such as scalar processing digital signal processors (DSPs).
- DSPs scalar processing digital signal processors
- the baseband processor 20 in Figure 2 also includes condition registers 36 configured to provide conditions to the VPE 22 for use in conditional execution of vector instructions and to store updated conditions as a result of vector instruction execution.
- the baseband processor 20 also includes accumulate registers 38, a global register file 40 that includes global registers, and address registers 42.
- the accumulate registers 38 are configured to be used by the VPE 22 to store accumulated results as a result of executing certain specialized operations on the vector data 30,
- the global register file 40 is configured to store scalar operands for certain vector instructions supported by the VPE 22.
- the address registers 42 are configured to store addresses addressable by vector load and store instructions supported by the VPE 22 to retrieve the vector data 30 from the vector unit data memory 32, and store vector processing results in the vector unit data memory 32.
- the baseband processor 20 in this embodiment also includes a scalar processor 44 (also referred to as an "integer unit”) to provide scalar processing in the baseband processor 20 in addition to vector processing provided by the VPE 22. It may be desired to provide a central processing unit (CPU) configured to support both vector and scalar instruction operations based on the type of instruction executed for highly efficient operation.
- the scalar processor 44 is a 32-bit reduced instruction set computing (RISC) scalar processor as a non-limiting example.
- the scalar processor 44 includes an arithmetic logic unit (ALU) 46 for supporting scalar instruction processing in this example.
- ALU arithmetic logic unit
- the baseband processor 20 includes an instruction dispatch circuit 48 configured to fetch instructions from program memory 50, decode the fetched instructions, and direct the fetched instructions to either the scalar processor 44 or through a vector data path 53 to the VPE 22 based on instruction type.
- the scalar processor 44 includes general purpose registers 54 for use by the scalar processor 44 when executing scalar instructions.
- An integer unit data memory (DMEM) 56 is included in the baseband processor 20 to provide data from main memory into the general purpose registers 54 for access by the scalar processor 44 for scalar instruction execution.
- the DMEM 56 may be cache memory as a non- limiting example.
- the baseband processor 20 also includes a memory controller 58 that includes memory controller registers 60 configured to receive memory addresses from the general purpose registers 54 when the scalar processor 44 is executing vector instructions requiring access to main memory through memory controller data paths 62.
- a filter operation computes a quantized time-domain representation of the convolution of a sampled input time function and a representation of a weighting function of the filter. Convolution in the time domain corresponds to multiplication in a frequency domain.
- digital filters can be realized in the VPE 22 by an extended sequence of multiplications and additions carried out at a uniformly spaced sample interval.
- a discrete finite impulse response (FIR) filter can be implemented using a finite number (Y) of delay taps on a delay line with "Y" computation filter coefficients to compute a filter function.
- Figure 3 is a schematic diagram of an exemplary discrete FIR filter 64 that may be desired to be supported through a filter vector processing operation in the VPE 22 in Figure 2.
- a digitized input signal 66 (x[n]) can be filtered by passing digitized input signal samples (x[0J, x[I], ... x[n]) through delay structures called "filter delay taps" 68(1)-68(Y- 1).
- the filter delay taps 68(1 )-68(Y-l) shift clocked digitized input signal samples (i.e., x[0], x[l ], ...
- n is the number of input signal samples
- x[n] is the digitized input signal 66
- y[n] is the resultant filtered output signal 76
- the filter coefficients h(7) may be complex.
- the VPE 22 may receive filter coefficients (e.g., from the global register file 40). The VPE 22 may use the received filler coefficients directly to perform the FIR filter function, in which case the filter coefficients h(7) in the above equation may represent the received filier coefficients. Alternatively, the VPE 22. may compute the complex conjugates of the received filter coefficients before using them to perform the FIR filter function, in which case the filier coefficients h(/) in the above equation may represent the conjugates of the received filter coefficients.
- filtering operations such as the discrete FIR filter 64 in Figure 3
- filtering operations may be difficult to parallelize in vector processors due to the specialized data flow paths provided in a vector processor.
- the input vector data sample set e.g., the vectorized digitized input signal 66
- filter delay taps e.g., 68(1)-68(Y-1)
- the data flow path in a vector processor could be configured to provide the same number of multipliers (e.g., 70(0)-70(Y-l)) as filter delay taps (e.g., 68(1)-68(Y-1)) for efficient parallelized processing.
- multipliers e.g., 70(0)-70(Y-l)
- filter delay taps e.g., 68(1)-68(Y-1)
- other vector processing operations may require fewer multipliers thereby providing inefficient scaling and unde tilization of the multipliers in the data flow path. If the number of multipliers is reduced to be less than the number of filter delay taps to provide scalability, parallelism is limited by more re-fetches being required to memory to obtain the same input vector data sample set for different phases of the filter processing.
- FIG. 4 is a schematic diagram of an exemplary VPE 22.(1) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(1) in Figure 4 provides precision filter vector processing operations in the VPE 22( 1 ) with eliminated or reduced vector data sample re-fetching and reduced power consumption.
- the precision filter vector processing operations can be provided in the VPE 22(1) as compared to filter vector processing operations that require storage of intermediate results requiring vector data sample re-fetching, thereby increasing power consumption as a result.
- tapped-delay lines 78 are included in input data flow paths 80(0)- 80(X) between vector data files 82(0)-82(X) and execution units 84(0)-84(X) (also labeled "EU") in the VPE 22(1 ).
- 'X'-f-l is the maximum number of parallel input data lanes provided in the VPE 22(1) for processing of vector data samples in this example.
- the tapped-delay lines 78 are configured to receive an input vector data sample set 86(0)-86(X) on tapped-delay line inputs 88(0)-88(X) as a subset or all of the input vector data samples 86 of the input vector data sample set 86(0)-86(X) from a corresponding subset or all of the vector data files 82(0)-82(X).
- the input vector data sample set 86(0)-86(X) is comprised of ' ⁇ + ⁇ input vector data samples 86, which in this example are 86(0), 86(1 ), ... , and 86(X).
- the tapped-delay lines 78 store input vector data sample sets 86(0)-86(X) fetched from the vector data files 82(0)-82(X) to be processed by the execution units 84(0)-84(X) for a filter vector processing operation.
- the tapped-delay lines 78 are configured to shift the input vector data sample sets 86(0)- 86(X) for each filter delay tap (i.e., filter processing stage) of the filter vector processing operation according to a filter vector instruction to be executed by the VPE 22(1) to provide a shifted input vector data sample set 86S(0)-86S(X) to the execution units 84(0)-84(X).
- All of the shifted input vector data samples 86S comprise the shifted input vector data sample set 86S(0)-86S(X).
- the tapped-delay lines 78 provide the shifted input vector data sample 86S(Q)-86S(X) to execution unit inputs 90(0)-90(X) of the execution units 84(0) -84(X) during the filter vector processing operation.
- intermediate filter results based on operations performed on the shifted input vector data sample set 86S(0)-86S(X) for the filter taps of the filter vector processing operation do not have to be stored, shifted, and re-fetched from the vector data files 82(0)-82(X) during each processing stage of the filter vector processing operation performed by the VPE 22(1).
- the tapped-delay lines 78 can reduce power consumption and increase processing efficiency for filter vector processing operations performed by the VPE 22( 1).
- a processing stage in the VPE 22(1) which is also referred to as a ''vector processing stage," comprises circuitry and associated vector data paths that are designed to carry out a specific task or operation.
- a vector processing operation may executed by the VPE 2.2(1) in several different processing stages. Each processing stage may be performed oyer one or multiple clock cycles of the VPE 22(1). Consequently, execution of a vector processing operation in the VPE 22(1 ) may take many clock cycles to complete, since each processing stage of the vector processing operation may consume one or more clock cycles each.
- a processing stage may include the fetching of the input vector data sample set 86(0)-86(X) into the tapped-delay lines 78 in the VPE 22(1) in Figure 4.
- the vector processing stages in the VPE 22(1) can be pipelined.
- the execution units 84(0)-84(X) may include one or more pipeline stages that process the fetched input vector data sample set 86(0)-8 ⁇ ( ⁇ ).
- one pipeline stage in the execution units 84(0)-84(X) may include an accumulation stage comprised of accumulators configured to perform accumulation operations.
- another pipeline stage in the execution units 84(0)-84(X) may include a multiplication stage comprised of multipliers configured to perform multiplication operations.
- the execution units 84(0)-84(X) receive a filter coefficient 92 from among filter coefficients 92( ' 0)-92(Y-l) stored in the global register file 40 of Figure 2 for the filter vector processing operation, where ⁇ ' can equal the number of filter coefficients for the filter vector processing operation.
- the execution units 84(0)-84(X) are each configured to multiply one of the received filter coefficient 92(0), 90(1), ... 90(Y-1) with a shifted input vector data sample 86S(0), 868( 1), ...
- the intermediate filter vector data output sample sets are accumulated in each of the execution units 84(0)-84(X) (i.e., a prior accumulated filter output vector data sample is added to a current accumulated filter output vector data sample).
- This provides a final, resultant filter output vector data sample set 94(0)-94(X) provided by the execution units 84(0)- 84(X) on execution unit outputs 96(0)-96(X) on output data flow paths 98(0)-98(X), respectively, for each shifted input vector data sample 86S(0), 86S(i), ... 86S(X) in the shifted input vector data sample set 86S(0)-868(X),
- the resultant filter output vector data sample set 94(0)-94(X) is comprised of 'X+i ' resultant filter output vector data samples 94, which in this example are 94(0), 94(1), and 94(X).
- the resultant filter output vector data sample set 94(0)-94(X) is stored back in the respective vector data files 82(0)-82(X) for further use and/or processing by the VPE 22(1) without having to store and shift intermediate filter vector data output sample sets generated by the execution units 84(0)-84(X).
- the tapped-deiay lines 78 are programmable to be controlled according to the vector instruction being processed. If a filter vector instruction is not being processed, the tapped-deiay lines 78 can be programmed to not be included in the input data flow paths 80(0)-80(X) between the vector data files 82(0)-82(X) and the execution units 84(0)-84(X).
- the tapped-deiay lines 78 are configured to load and shift the input vector data sample set 86(0)-86(X) received from the vector data files 82(0)-82(X) to provide a shifted input vector data sample set 86S(0)-868(X) for each filter tap of the filter vector processing operation.
- the shifted input vector data sample set 86S(0)-86S(X) can be provided to the execution units 84(0)-84(X) for execution of a filter tap of the filter vector processing operation.
- the efficiency of the input and output data flow paths 80(0)-80(X), 98(0)-98(X) in the VPE 22(1) is not limited by the re-fetching delay of the shifted input vector data sample set 86S(0)-86S(X) from the vector data files 82(0)-82(X) during the filter vector processing operation.
- the shifted input vector data sample set 86S(0)-86S(X) is provided by the tapped-deiay lines 78 localized to the execution units 84(0)-84(X).
- the vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations. This means that the execution units 84(0)-84(X) are kept busy continuously, or substantially continuously, receiving the shifted input vector data sample set 86S(0)-86S(X) for performing vector processing operations without having to wait for the shifted input vector data sample set 86S(0)- 86S(X) to be fetched from the vector data flies 82(0)-82(X).
- the filter vector processing operations performed by the VPE 22(1) in Figure 4 may be more precise by employing the tapped-deiay lines 78, because output accumulations for intermediate filter processing stages in the execution units 84(0)- 84(X) do not have to be stored in the vector data files 82(0)-82(X). Storing of intermediate output vector data sample sets from the execution units 84(0)-84(X) in the vector data files 82(0)-82 X) may result in rounding. Thus, when the next intermediate output vector data sample set would be provided to the execution units 84(0)-84(X) for the vector processing operation, any rounding error would be propagated and added during each multiplication phase of the vector processing operation.
- the intermediate output vector data sample sets calculated by the execution units 84(0)-84(X) do not have to be stored in the vector data files 82(0)-82(X).
- the execution units 84(0)- 84(X) can accumulate prior intermediate output vector data sample sets with intermediate output vector data sample sets for next filter delay taps, because the tapped-delay lines 78 provide the shifted input vector data sample set 86S(0)--868(X) to the execution units 84(0) -84(X) during the vector processing operation to be processed, and the results are accumulated with prior vector data sample sets for prior filter delay taps.
- the VPE 22(1) in this embodiment is comprised of a plurality of vector data lanes (labeled VLANE0-VLANEX) 100(0)- 100(X)) for parallelized processing.
- Each vector data lane 100(0)- 00(X) contains a vector data file 82 and an execution unit 84 in this embodiment.
- the vector data file 82(0) therein is configured to provide the input vector data sample 86(0) on the input data flow path 80(0) to be received by the execution unit 84(0) for filter vector processing.
- the tapped-delay lines 78 are provided in the input data flow path 80(0) to shift the input vector data sample 86(0) and to provide the shifted input vector data sample 868(0) to the execution unit 84(0) for filter vector processing.
- the vector data file 82(0) is also configured to receive a resultant filter output vector data sample 94(0) provided by the execution unit 84(0) as a result of filter vector processing from the output data flow path 98(0) to be stored back in the vector data file 82(0) for a subsequent vector processing operation, as needed or desired according to the current or next vector instruction to be processed by the VPE 22(1 ).
- Any number of vector data lanes 100(0)- 100(X) may be provided in the VPE 22(1) as desired.
- the number of vector data lanes 100(0)-100(X) provided in the VPE 22(1) may be based on tradeoffs for parallelized vector processing for efficiency purposes versus the additional circuitry, space, and power consumption involved in providing additional vector data lanes 100(0)- lOO(X).
- sixteen vector data lanes 100 may be provided in the VPE 22(1 ), with each vector data lane 100 having a data width capability of thirty-two (32) bits, to provide for parallelized processing of up to 512 bits of vector data in the VPE 22(1 ).
- the vector data file 82(0) allows one or multiple samples of an input vector data sample 86(0) to be stored for vector processing.
- the width of the input vector data sample 86(0) is provided according to programming of the input vector data sample 86(0) according to the particular vector instruction being executed by the VPE 22( 1).
- the width of the input data flow path 80(0) is programmable and reprogrammable on a vector-instraction-by-vector-instruction basis, including on clock-cycle-by-clock-cycle basis for a given vector instruction to provide different widths of the input vector data sample 86(0) to the tapped-deJay lines 78 and the execution unit 84(0). in this manner, the vector data lane 100(0) can be programmed and reprogrammed to provide processing of different widths of the input vector data sample 86(0) depending on the type of vector instruction being executed.
- the vector data file 82(0) may be thirty-two (32) bits wide and capable of storing input vector data samples 86 that are also up to thirty-two (32) bits wide.
- An input vector data sample 86(0) may consume the entire width of the vector data file 82(0) (e.g., 32 bits), or may be provided in smaller sample sizes of the vector data file 82(0) width.
- the input vector data sample 86(0) size can be configured based on programming of the input data flo w path 80(0) configuration for the size of the input vector data sample 86(0) based on a vector instruction being executed by the VPE 22(1),
- the input vector data sample 86(0) may comprise two (2) separate 16-bit vector data samples for one vector instruction.
- the input vector data sample 86(0) may comprise four (4) 8-bit vector data samples in the vector data file 82(0) for another vector instruction, as opposed to one (1) 32-bit vector data sample, in another example, the input vector data sample 86(0) may comprise one (I) 32-bit vector data sample.
- the VPE 2.2(1 ) is also capable of programming and reprograming the output data flow path 98(0) for the vector data file 82(0) to receive different sizes of a resultant filter output vector data samples 94(0) provided by the execution unit 84(0) to the vector data file 82(0) for each vector instruction and/or each clock cycle of a given vector instruction.
- FIG. 5 is a flowchart illustrating an exemplary filter vector processing operation 102 that can be performed in the VPE 22( 1) in Figure 4 employing the tapped-delay lines 78 according to an exemplary filter vector instruction.
- the exemplary tasks performed in the filter vector processing operation 102 in Figure 5 will be described with reference to the examples provided in Figures 6A-10.
- the input vector data sample set 86(0)-86(X) to be processed in the filter vector processing operation 102 according to a filter vector instruction is fetched from vector data files 82(0)-82(X) into the input data flow paths 80(0)-80(X) for the filter vector processing operation 102 (block 104).
- the input vector data sample set 86(0)- 86(X) is multiplied by the filter coefficients 92(0)-92(Y-i) received from the global register file 40 in the execution units 84(0)-84(X).
- Figure 6A illustrates filter coefficients 92(0)-92(Y-l) (i.e., h7-h0) in the global register file 40.
- filter coefficients 92 stored in the global register file 40 providing eight (8) filter taps in the filter vector processing operation 102 to be performed. Note that in this example, the filter vector processing operation 102 from the discrete FIR filter 64 equation in Figure 3 discussed above is
- y[n] x[n]*h0 + x[n-l]*hl+ ... + x[n-7]*h7.
- Figure 6B illustrates an exemplary input vector data sample set 86(0)-86(X) stored in the vector data files 82(0)-82(X) in the VPE 22(1) in Figure 4 representing an input signal to be filtered by the filter vector processing operation 102.
- sample X0 is the oldest sample
- sample X63 is the most recent sample.
- sample X63 occurs in time after sample X0. Because each address of the vector data files 82(0)-82(X) is 16-bits wide, the first input vector data sample set 86(0)-86(X) stored in the vector data files 82(0)-82(X) spans ADDRESS 0 and ADDRESS 1, as shown in Figure 613.
- vector data files 82(0)-82(X) to provide input vector data samples 86 of 32-bit width to support the 32-bit width capability of the execution units 84(0)-84(X) in the VPE 2.2(1) example in Figure 4,
- there are sixty-four (64) total input vector data sample subsets i.e., X0-X63) each 8-bits in width totaling 512 bits that comprise the first input vector data sample set 86(0)-86(X).
- ADDRESS 2 and ADDRESS 3 store another, second input vector data sample set 86(0)-86(X) stored in the vector data files 82(0)-82(X).
- each vector data file 82(0)-82(X) is shown, which illustrate 256 total input vector data samples 86 (i.e., X0-X255), but such is not limiting.
- Either one, some, or all of the vector data lanes 100(0)-100(X) in the VPE 22(1) in Figure 4 can be employed to provide the filter vector processing operation 102 according to the programming of the vector instruction depending on the width of the input vector data sample set 86(0)-86(X) involved in the filter vector processing operation 102. If the entire width of the vector data files 82(0)-82(X) is required, all vector data lanes 100(0)-100(X) can be employed for the filter vector processing operation 102. Note that the filter vector processing operation 102 may only require a subset of the vector data lanes 100(G)- 100(X) that may be employed for the filter vector processing operation 1 02.
- the width of the input vector data sample set 86(0)-86(X) is less than the width of all vector data files 82(0)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the filter vector processing operation 102.
- the input vector data sample set 86(0)-86(X) employed in the filter vector processing operation 102 involves ail vector data lanes i0Q(0)-I00(X).
- a fetched input vector data sample set 86(0)-86(X) is provided to the input data flow paths 80(0)-80(X) from the vector data files 82(0)-82(X) to be loaded into the tapped delay-lines 78 as a current input vector data sample set 86(0)-86(X) (block 106).
- An input vector data sample set 86(0)-86(X) is loaded into the primary tapped-delay line 78(0) as the input vector data sample set 86(0)-86(X) to be processed by the execution units 84(0)-84(X) for the filter vector processing operation 102.
- the input vector data sample set 86(0)- 86(X) loaded into the primary tapped-delay line 78(0) is not shifted for the first filter tap operation of the filter vector processing operation 102.
- the purpose of the tapped-delay lines 78 is to provide shifting of the input vector data sample set 86(0)-86(X) to provide a shifted input vector data sample set 86S( ' 0)-86S(X) to the execution units 84(0)-84(X) for subsequent filter tap operations of the filler vector processing operation 102.
- the input vector data samples 86 are shifted in the primary- tapped delay line 78(0) to provide the shifted input vector data sample set 86S(0) ⁇ 86S(X) to the execution units 84 ⁇ 0)-84(X). In this manner, the input vector data sample set 86(0)- 86(X) does not have to be stored, shifted in the vector data files 82(0)-82(X), and re-fetched for each filter tap operation of the filter vector processing operation 102.
- a next input vector data sample set 86N(0)-86N(X) can also be loaded from the vector data files 82(0)-82(X) into the shadow tapped-deiay line 78(1).
- the next input vector data sample set 86N(0)- 86N( ' X) is shifted into the primary tapped-delay line 78(0) during the filter vector processing operation 102 to become at least part of the shifted input vector data sample set 86S(0)-86S(X).
- the primary tapped-delay line 78(0) can have the shifted input vector data sample set 86S(0)-86S(X) available during the filter vector processing operation 102 without fetching delay that would otherwise be incurred if the execution units 84(0)-84(X) were required to wait until the next input vector data sample set 86N(0)-86 (X) to be executed for the filter vector processing operation 102 was fetched from the vector data files 82(0)-82(X) into the primary tapped-delay line 78(0).
- Figure 7 illustrates the exemplary tapped-delay lines 78 that can be provided in the VPE 22(1) in Figure 4,
- the tapped-delay lines 78 comprise the shadow tapped-delay line 78(1) and the primary tapped-delay line 78(0).
- the primary- tapped delay line 78(0) in this example is comprised of a plurality of 8-bit primary pipeline registers 120 to allow resolution of input vector data samples 86 down to 8-bits in length.
- the first input vector data sample set 86(0)-86(X) processed by the execution units 84(0)-84(X) will be un-shifted in this example for the first filter tap of the filter vector processing operation 102, as will be discussed in regard to Figure 9A below.
- the execution units 84(0)-84(X) process subsequent filter taps for the filter vector processing operation 102, the input vector data samples 86 in the input vector sample set 86(0)-86(X) stored in the primary tapped-delay line 78(0) are shifted in the primary pipeline registers 120(0)- 120(4X+3), as indicated by the arrows in Figure 7, to become the shifted input vector data sample set 86S(0)-86S(X).
- the execution units 84(0)-84(X) are fully utilized by receiving and performing the filter vector processing operation 1 02 of the shifted input vector data sample set 86S(0)-86S(X) without having to store and shift the input vector data sample set 868(0)- 86S(X), and re-fetch the shifted input vector data sample set 86S(0)-86S(X) from the vector data files 82(0)-82(X).
- the primary pipeline registers 120(0)- 120(4X4-3) collectively are the width of the vector data files 82(0)-82(X) in Figure 4.
- the vector data files 82(0)-82(X) being 512-bits in width with "X" equal to fifteen (15)
- the primary tapped-delay line 78(0) is capable of storing the entire width of one (1) input vector data sample set 86(0)-86(X).
- the input vector data sample set 86(0)-86(X) can be shifted in the primary pipeline registers 120(0)- 120(4X43) down to a vector data sample size of eight (8) bits for 8-bit filter vector processing operations.
- the input vector data sample set86(0)-86(X) can be shifted in the primary pipeline registers 120(0)- 120(4X43) by two (2) primary pipeline registers 120 at a time, ⁇ 01 ⁇ 1 ]
- the shadow tapped-delay line 78(1 ) is also provided in the tapped-delay line 78.
- the shadow tapped-delay line 78(1) can be employed to latch or pipeline a next input vector data sample set 86N(0)-86N(X) from the vector data files 82(0)-82(X) for a subsequent vector processing operation.
- a next input vector data samples 86N from the next input vector data sample set 86N(0)- 86N(X) is shifted from the shadow tapped-delay line 78(1) into the primary tapped- delay line 78(0) as each filter tap for the filter vector processing operation 1 02 is executed by the execution units 84(0)-84(X).
- the shadow tapped-delay line 78(1) is also comprised of a plurality of 8-bit shadow pipeline registers 122 to allow resolution of input vector data samples 86 down to 8-bits in length similar to the primary tapped- delay line 78(0).
- the shadow pipeline registers 122(0)- 1 22(4X43) provided in the shadow tapped-delay line 78(1 ) collectively are the width of the vector data files 82(0)-82(X), which is 512-bits in this example.
- the shadow pipeline registers 122(0)-122(4X+3) of the shadow tapped- delay line 78(1 ) are also capable of storing the entire width of one (1) input vector data sample set 86(0) ⁇ 86(X).
- the number of shadow pipeline registers 122(0)-i22(4X+3) included in the primary tapped-delay line 78(0) is four times the number of vector data lanes 100(0)-100(X), which total sixteen (16) in this example (i.e., X ⁇ l 5).
- the number of shadow pipeline registers 122 also totals sixty-four (64) in this example for a total of 512 bits (i.e., 64 registers x 8 bits each).
- next input vector data sample set 86N(0)-86N(X) can be shifted down to a vector data sample size of eight (8) bits for 8-bit filter vector processing operations.
- Figure 8 is a schematic diagram illustrating selected primary pipeline and shadow pipeline registers 120, 122 present in the primary and shadow tapped-delay lines 78(0), 78(1) i Figure 7.
- Figure 8 is provided to facilitate discussing an example of shifting input vector data samples 86 between primary and shadow pipeline registers 120, 122.
- the input vector data samples 86 can also be shifted within the primary and shadow tapped-delay lines 78(0), 78(1) as well as from the shadow tapped-delay line 78(1 ) to the primary tapped-delay line 78(0).
- the pipeline registers 120, 122 are each 8- bits wide in this example to allow for input vector data sample 86 shifting at 8-bits of resolution if desired. This will be discussed in more detail below.
- the primary and shadow tapped-delay lines 78(0), 78(1 ) are also capable of performing 16-bit and 32-bit shifting resolution of input vector data samples 86, as will also be discussed in more detail below,
- Figure 8 illustrates shifting of input vector data samples 86 into primary pipeline registers 120(4X+3), i20(2X+l ), 120(4X+2), and 120(2X) that form the storage registers for input vector data sample 86S(X) in the primary tapped- delay line 78(0) in Figure 7.
- Primary pipeline registers 120(4X+3) and 120(4X+2) are registers 631 and B30, respectively, in the primary tapped-delay line 78(0) in Figure 7.
- Primary pipeline registers 120(2X+1 ) and 12Q(2X) are registers A31 and Aj 3 ⁇ 4 respectively, in the primary tapped-delay line 78(0) in Figure 7.
- primary pipeline registers 120(4X+3) and 120(4X+2) for registers 631 and B30 are configured to receive shifted input vector data samples 86 from adjacent shadow pipeline registers 122 in the shadow tapped-delay line 78(1).
- shadow pipeline registers 12.2(0), 122(1) for registers AO and A' i are illustrated as being configured to shift input vector data samples 86 into primary pipeline registers 120(4X+3) and 120(4X4-2) for B31 and B3.3.
- primary pipeline registers 120(2X4-3) and 120(2X+2) for registers Bi and Bo are illustrated as being configured to shift input vector data samples 86 into adjacent primary pipeline registers 120(2X4-1) and 120(2X) for registers A31 and A30. Exemplary shifting of input vector data samples 86 between these registers will now be discussed.
- an input vector data sample selector is associated with each of the primary and shadow pipeline registers 120, 122.
- input vector data sample selectors 124(0)-124(4X+3) are provided to vector data loaded or shifted into primary pipeline registers 120(0)- 120(4X4-3), respectively, in the primary tapped-delay line 78(0).
- 26(4 X+3) are provided to vector data loaded or shifted into shadow pipeline registers 122(G)- 122(4X4-3), respectively, in the shadow tapped-delay line 78( 1).
- the input vector data sample selectors 124(0)- 124(4X4-3) and input vector data sample selectors 126(G)- 126(4X4-3) are each a multiplexor in this example.
- the input vector data sample selectors 124(G)-! 24(4X+3), 126(0)-! 26(4 X ⁇ 3) can each be controlled by data width shift control inputs 125 to select input vector data to either be loaded or shifted into the primary and shadow pipeline registers 120(0)- 120(4X+3), 122(0)-122(4X+3).
- the data width shift control inputs 125 can be configured by the VPE 22(1 ) in Figure 4 to cause the input vector data sample selectors 124(4X4-3), 124(4X4-2), 124(2X4-1 ), i24(2X) to select the load data flow paths 133 (4X+3), 133(4X+2), 133(2X4-1), 133(2X).
- Selecting the load data flow paths 133 (4X+3), 133(4X4-2), 133(2X+1), 133(2X) allows input vector data from the vector data files 82(0)-82(X) to be stored in the primary pipeline registers 120(4X+3), 120(4X4-2), 120(2X+1), 120(2X).
- Loading input vector data from the vector data files 82(0)-82(X) may be performed on a ne or next vector instruction to be processed by the VPE 22(1) as an example.
- the data width shift control inputs 125 can also be configured by the VPE 22(1) in Figure 4 to cause the input vector data sample selectors 126(1), 124(2X+3), 126(0), 124(2X+2) to select the input data flow paths 135(1 ), 133(2X4-3), 135(0), 133(2X4-2). Selecting the load data flow paths 135(1), 133(2X43), 135(0), 133(2X4-2) allows input vector data from the vector data files 82(0)-82(X) to be stored in the pipeline registers 122(1 ), 120(2X4-3), 124(0), 120(2X4-2).
- the data width shift control inputs 125 can be configured by the VPE 22( 1) in Figure 4 to cause the input vector data sample selectors 124(4X4-3), 124(4X4-2), 124(2X4-1), 124(2X) to select the input data flow paths 137(4X4-3), 137(4X4-2), 137(2X4-1), 137(2X) for vector data sample shifting.
- the data width shift control inputs 125 also cause the input vector data sample selectors 126(1 ), 124(2X43), 126(0), 124(2X+2) to select the input data flow paths 139(1), 137(2X+3), 139(0), 137(2X42) for vector data sample shifting.
- the input vector data sample selectors 124(4X4-3), 124(4X4-2), 124(2X4-1), 124(2X) and input vector data sample selectors 126(1), 124(2X4-3), 126(0), 124(2X4-2) each include output data flow paths 141(4X4-3), 141(4X4-2), 141(2X4-1), 14i(2X) and 143(1), 141(2X4-3), 143(0), 124(2X4-2), respectively, that allow vector data to be shifted to other registers.
- the output data flow paths shown in Figure 8 are part of the output data flow paths 141(0)-14i(4X+3) and 143(0)-143(4X+3) that are now shown in total, but included for the input vector data sample selectors 124(0)- 124(4X4-3) in the primary tapped-delay line 78(0) and the input vector data sample selectors 126(0)- 126(4X4-3) in the shadow tapped-delay line 78(1), respectively.
- the input vector data sample selectors 124(4X4-3), 124(4X4-2), 124(2X+1 ), 124(2X) and input vector data sample selectors 126( 1), 124(2X4-3), 126(0), 124(2X+2) are configured to select the input data flow paths 137(4X+3), 137(4X+2), 137(2X+i), 137(2X), 139(1), 137(2X+3), 139(0), 137(2X+2), respectively.
- the vector data in primary pipeline register 120(2X4-1 ) i.e., A 3 !
- the vector data in primary pipeline register 12.0(2X4-3) (i.e., Bj) is shifted on outpui data flow path 141(2X+3) to primary pipeline register ! 20(4X4-2) (i.e., B30), as illustrated in Figure 8.
- the vector data in shadow pipeline register 122(1) (i.e., A'i) is shifted on output data flow path 143(1) to shadow pipeline register 122(0) (i.e., A'o), as illustrated in Figure 8.
- the vector data in primary pipeline register 120(2X4-2) (i.e., Bo) is shifted on outpui data flow path 141(2X4-2) 10 primary pipeline register 120(2X4-1 ) (i.e., A3 , as illustrated in Figure 8.
- the input vector data sample selectors 124(4X+3), 124(4X4-2), 124(2X+1), 124(2X) and input vector data sample selectors 126(1 ), 124(2X4-3), 126(0), 124(2X4-2) are configured to select the input data flow paths 145(4X4-3), 145(4X42), 145(2X4-1), 145(2X), 147( 1), 145(2X4-3), 147(0), 145(2X4-2), respectively.
- the vector data in primary pipeline register 120(2X+2) (i.e., Bo) is shifted on output data flow path 141 (2X4-2) to primary pipeline register 120(2X) (i.e., A30), as illustrated in Figure 8.
- the vector data in shadow pipeline register 122(0) (i.e., AO) is shifted on output data flow path 143(0) to primary pipeline register 120(4X4-2) (i.e., B30), as illustrated in Figure 8.
- the vector data in primary pipeline register 120(2X4-3) (i.e., Bj) is shifted on output data flow path 141(2X4-3) to primary pipeline register 120(2X4- 1) (i.e., A31), as illustrated in Figure 8.
- the vector data in shadow pipeline register 122(1) (i.e., A' i) is shifted on output data flow path 143(1) to primary pipeline register 120(4X4-3) (i.e., B 31 ), as illustrated in Figure 8.
- primary pipeline register 120(4X4-3) i.e., B 31
- the vector data stored in the primary pipeline registers 120(0)- 120(4X+3) and the shadow pipeline registers 122(0)-122(4X+3) can be shifted in two (2) 16-bit vector data shift operations, if desired.
- the tapped-delay lines 78(0), 78(1 ) are configurable to be selectively provided or not provided in the input data flow paths 80(0)-80(X) between the vector data files 82(0)-82(X) and the execution units 84(0)- 84(X) based on a programmable input data path configuration for the tapped-delay lines 78(0), 78(1) according to a vector instruction to be executed.
- the tapped-delay lines 78(0), 78(1) can be configured to not latch the input vector data sample sets 86(0)-86(X).
- the input vector data sample sets 86(0)-86(X) can be provided frora the vector data files 82(0)-82(X) to the respective execution units 84(0)- 84(X) by bypassing the primary and shadow tapped-delay lines 78(0), 78(1 ).
- This programmable data path configuration farther allows the primary and shadow tapped- delay lines 78(0), 78(1) to be provided or not provided in the input data flow paths 80(0)-80(X).
- the primary and shadow tapped-delay lines 78(0), 78(1) can be programmed to be provided or not provided in the input data flow paths 80(0)-80(X)for each vector instruction, as desired.
- Figure 9A illustrates an input vector data sample set 86(0)-86(X) loaded from the vector data files 82(0)-82(X) into the primary tapped-delay line 78(0) during a first clock cycle (CYCLE0) of a filter vector processing instruction.
- the primary tapped-delay line 78(0) and the shadow tapped-delay line 78(1) are shown in simplified form from Figure 7.
- the global register file 40 is also shown.
- the first input vector data sample set 86(0)-86(X) is loaded into the primary tapped-delay line 78(0) as input vector data samples X0-X63.
- a special vector instruction may be supported to load the first input vector data sample set 86(0)-86(X) into the primary tapped-delay line 78(0) (and also the shadow-tapped delay line 78(1), as discussed in more detail below).
- This first input vector data sample set 86(0)-86(X) was stored in ADDRESSES 0 and 1 in the vector data files 82(0)-82(X) shown in Figure 6B.
- XO, XI, X32, and X33 form the first input vector data sample 86(0), only because of the storage pattern of the vector data files 82(0)-82(X) in the VPE 22( 1) in Figure 4 for this example.
- Figure 9B illustrates a next input vector data sample set 86N(0)-86N(X) loaded into the shadow tapped-delay line 78(1) during a second clock cycle (CYCLE! ) of a filter vector processing instruction.
- the next input vector data sample set 86N(0)- 86N(X) is loaded into the shadow tapped-delay line 78(1 ) after the first input vector data sample set 86(0)-86(X) from the vector data files 82(0) » 82(X) is loaded into the primary tapped-delay line 78(0) to setup the execution of a filter processing operation.
- This next input vector data sample set 86N(0)-86N(X) is loaded into the shadow tapped- delay line 78( 1) as input vector data samples X64-X127.
- This next input vector data sample set 86N(0)-86N(X) was stored in ADDRESSES 2 and 3 in the vector data files 82(0)-82(X) shown in Figure 6B.
- X64, X65, X96, and X97 form the first input vector data sample 86(0), only because of the storage pattern of the vector data files 82(0)-82(X) in the VPE 22(1) in Figure 4 for this example.
- Other patterns could be provided to group the input vector data samples 86 together to form the input vector data sample set 86(0)-86(X).
- the first filter coefficients 92(0) from the global register file 40 are also shown as provided in a register ("C") to the execution units 84(0)"84(X) in Figure 9B for use in the filter vector processing operation 102.
- the next input vector data samples 86N stored in the shadow pipeline registers 122 are also shifted in the shadow pipeline registers 122 of the shadow tapped-delay line 78(1).
- the input vector data sample 86 stored in the first shadow pipeline register 122(0) in Figure 7 is shifted into the last primary pipeline register 120(4X+3) of the primary tapped-delay line 78(0) during each shift.
- the execution units 84(0)-84(X) can perform the filler vector processing operation 102 without any further input vector data sample sets 86(Q)-86(X) being re-fetched from the vector data files 82(0)-82(X).
- the execution units 84(0)-84(X) can then be provided with the previous next input vector sample data set 86N(0)-86N(X) stored in the primary tapped-delay line 78(0) as the shifted input vector data sample set 868(0)- 86S(X) for a next vector processing operation if unprocessed input vector data sample 86S are present in the tapped-delay lines 78(0), 78(1),
- Another exemplary rationale for providing the shadow tapped-delay line 78(1) is as follows. If a current filter vector processing operation 102 involves more input vector data samples 86 than can be pro vided in the width of the vector data lanes 1 0(0)-1 00(X), an additional input vector data sample set 86(0)-86(X) loaded into the shadow tapped-delay line 78(1) will be available to the execution units 84(0)-84(X) during the filter vector processing operation 102 without delay.
- next input vector data sample sets 86N(0)-86N(X) loaded into the shadow tapped-delay line 78(1) are shifted into the primary tapped-delay line 78(0).
- the next input vector data sample set 86N(0)-86N(X) for use in vector processing by the execution units 84(G)- 84(X) is available without delay.
- the execution units 84(0)-84(X) can continue to be fully utilized during the filter vector processing operation 102 regardless of whether a single fetched input vector data sample set 86(0)-86(X) of the width of the vector data files 82(0)-82(X) is sufficient to perform the entire filter vector processing operation 102.
- the first input vector data sample set 86(0)-86(X) and next input vector data sample set 86N(0)-86N(X) are loaded into the primary tapped-delay line 78(0) and the shadow tapped-delay line 78(1), respectively, the first input vector data sample set 86(0)-86(X) provided in the primary tapped-delay line 78(0) is provided to the respective execution units 84(0)-84(X) to be processed in a first processing stage of the filter vector processing operation 102 (block 108 in Figure 5).
- the first input vector data sample set 86(0)-86(X) is shifted in the primary tapped-delay line 78(0) to become the shifted input vector data sample set 86S(0)-86S(X) to be processed by the execution units 84(0)-84(X) after the first input vector data sample set 86(0)-86(X) is processed by the execution units 84(0)-84(X).
- the shifted input vector data sample 868(0) is provided to the execution unit 84(0)
- the shifted input vector data sample 868(1 ) is provided to the execution unit 84(1), and so on, as illustrated in the VPE 22(1) in Figure 4.
- subsequent shifted input vector data sample sets 86S(0)-86S(X) for the filter vector processing operation 102 are multiplied by the current filter coefficient 92(1)-92(Y-1).
- the execution units 84(0)"84(X) accumulate the resultant filter vector output vector data sample set 94(G)- 94(X) with the prior resultant filter output vector data sample set 94(G) -94(X) calculated by the execution units 84(0)-84(X) to provide the new prior resultant filter output vector data sample set 94(0)-94(X) (block i 12 in Figure 5). in the first processing stage of the filter vector processing operation 102, there is no prior resultant filter output vector data sample set.
- the samples stored in the tapped- delay lines 78(0) and 78(1) are shifted within the tapped-delay lines 78(0), 78(1 ) to provide a next shifted input vector data sample set 86S(0)-86S(X) for the filter vector processing operation 102 (block 1 18 in Figure 5).
- the shifted input vector data sample set 86S(O)-86S(X) is provided for calculating a next resultant filter output vector data sample set as an intermediate result to be accumulated with the prior resultant filter output vector data sample set until the filter vector processing operation 102 is complete.
- Figure 9C illustrates the contents of the tapped-delay lines 78 when the input vector data sample set 86(0)-86(X) is shifted in a second processing stage of the filter vector processing operation 102 to become the next shifted input vector data sample set 86S(0)-86S(X) for the next filter processing operation y[n] - x[n-6] * Ii6.
- the shifted input vector data sample set 86S(0)-86S(X) in the primary tapped-delay line 78(0) is shifted in the primary pipeline registers 120(0)- 120(4X+3) according to the width of input vector data sample shifting prescribed by the vector instruction being executed.
- sample X2 is shifted in shifted input vector data sample 86S(0), as illustrated in Figure 9C.
- the new shifted input vector data sample set 86S(0)-86S(X) is provided to the execution units 84(0)-84(X) for execution for the next filter tap of the filter vector processing operation 102.
- the filter coefficient 92 provided to the execution units 84(0)-84(X) is also the next filter coefficient 92, which is "h6" in this example.
- the process repeats by providing the shifted input vector data sample set 86S(0)-86S(X) from the primary iapped-delay line 78(0) to the execution units 84(0)-84(X) (block 108 in Figure 5) to be multiplied with the next filter coefficient 92 (block 110 in Figure 5).
- the resultant filter output vector data sample set 94(0)-94(X) is accumulated with the prior resultant filter output vector data sample set 94(0)-94(X) (block 1 12 in Figure 5).
- Figure 9D illustrates the state of the input vector data samples 86 present in the iapped-delay lines 78(0), 78(1 ) during the last processing stage of the exemplary filter vector processing operation 102.
- the filter vector processing operation 102 described above employs each of the vector data lanes 100(0)-100(X) in the VPE 22(1) to provide the filter vector processing operation 102, such is not required.
- the filter vector processing operation 102 may only require a subset of the vector data lanes 100(0)- 100(X) to be employed for the filter vector processing operation 102.
- the width of the input vector data sample set 86(0)-86(X) may be less than the width of all vector data files 82(0)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the filter vector processing operation 102.
- the iapped-delay lines 78(0), 78( 1) in Figure 7 may need to be modified to shift the next input vector data sample set 86N(0)-86N(X) from the shadow tapped-delay line 78(1) to the primary tapped-delay line 78(0) as shifted input vector data sample set 86S(0)-86S(X) in a vector data lane 100 prior to reaching the end vector data lane 100(X).
- Figure 10 is a schematic diagram of contents of accumulators (i.e., the resultant niter output vector data samples 94) in the execution units 84(0)-84(X) in the VPE 22(1 ) of Figure 4 after the exemplary eight (8) lap filter vector processing stages in the above example have been fully executed according to y[n] - x[n]*h0 +x[n-l]*hl + . . . x[n-7]*h7
- Accumulators Acc0-Acc3 are shown in Figure 10, because in this example, each execution unit 84(0) ⁇ 84(X) has four accumulators disposed in parallel for each vector data lane 100(0)-100(X).
- the accumulated resultant output vector data samples can be provided on the output data flow paths 98(0)--98(X) to the vector data files 82(0)-82(X) as the collective resultant filter output vector data sample set 94(0)- 94(X) to be stored therein for further analysis and/or processing.
- a specialized vector instruction may be supported by the VPE 22( 1) to move rows of the resultant filter output vector data sample set 94(G)-94(X) from the vector data files 82(0)-82(X) to the vector unit data memory 32 of Figure 2, if desired.
- ⁇ vector processing operations other than the filter vector processing operation 102 can also enjoy processing efficiencies in a VPE by use of tapped-delay lines 78 like or similar to that provided in the VPE 22(1) in Figure 4 discussed above.
- another specialized vector processing operation that involves shifting of input vector data sample sets 86 in a VPE is a correlation/covariance vector processing operation (referred to herein as "correlation vector processing operation").
- correlation vector processing operation it may be desired to employ vector processing to provide correlation operations to choose the direct spread-spectrum code (DSSC) (i.e., chip sequence) for demodulating a user signal in a CDMA system to provide good separation between the user signal and signals of other users in the CDMA system.
- DSSC direct spread-spectrum code
- the separation of the signals is made by correlating the received signal with the locally generated chip sequence of the desired user. If the signal matches the desired user's chip sequence, the correlation function will be high and the CDMA system can extract that signal. If the desired user's chip sequence has little or nothing in common with the signal, the correlation should be as close to zero as possible (thus eliminating the signal), which is referred to as cross-correlation. If the chip sequence is correlated with the signal at any time offset other than zero, the correlation should be as close to zero as possible. This is referred to as auto- correlation, and is used to reject multi-path interference. [00125] However, correlation operations may be difficult to parallelize in vector processors due to the specialized data flow paths provided in vector processors.
- the input vector data sample set representing the signal to be correlated is shifted between delay taps, the input vector data sample set is re-fetched from the vector data file, thus increasing power consumption and reducing throughput.
- the data flow path could be configured to provide the same number of multipliers as delay taps for efficient parallelized processing.
- other vector processing operations may require fewer multipliers thereby providing inefficient scaling and underutilization of the multipliers in the data flow path, if the number of multipliers is reduced to be fewer than the number of delay taps to provide scalability, parallelism is limited by more re-fetches being required to memory to obtain the same input vector data sample set for different phases of the correlation processing.
- Figure 1 1 is a schematic diagram of another exemplary VPE 22(2) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(2) in Figure 1 1 is configured to provide precision correlation vector processing operations in the VPE 22(2) with eliminated or reduced vector data sample re-fetching and reduced power consumption.
- the precision correlation vector processing operations can be provided in the VPE 22(2) as compared to correlation vector processing operations that require storage of intermediate results requiring vector data sample re-fetching, thereby increasing power consumption as a result.
- the tapped-delay lines 78 included in the VPE 22(1) in Figure 4 are also included in the input data flow paths 80(0)-80(X) between the vector data files 82(0)-82(X) and execution units 84(0)-84(X) (also labeled "EU") in the VPE 22(2).
- 'X'+l is the maximum number of parallel input data lanes provided in the VPE 22(2) for processing of vector data samples in this example.
- the tapped-delay lines 78 are configured to receive an input vector data sample set 86(0)-86(X) on tapped-delay line inputs 88(0)- 88(X) as a subset or all of input vector data samples 86 of the input vector data sample set 86(0)-86(X) from a corresponding subset or all of the vector data files 82(0)-82(X), All the input vector data samples 86 comprise the input vector data sample set 86(G)- 86(X).
- the input vector data sample set 86(0)-86(X) from the vector data files 82(0)-82(X) is correlated in the VPE 22(2) with a reference vector data sample set 130(0)-130(X) to provide a resultant correlated output vector data sample set 132(G)- 132(X).
- the reference vector data sample set 130(0)- 130(X) is comprised of ' ⁇ + reference vector data samples 130, which in this example are 130(0), 130(1 ), and 13Q(X).
- the resultant correlated output vector data sample set 132(0)-132(X) is comprised of 'X+l * resultant correlated output vector data samples 132, which in this example are 132(0), 132(1 ), ... , and 132(X).
- the tapped-delay lines 78 shift the input vector data sample set 86(0)-86(X) for each con'elation delay tap (i.e., correlation processing stage) of the correlation vector processing operation according to a correlation vector instruction to be executed by the VPE 22(2) to provide a shifted input vector data sample set 86S(0)-86S(X).
- Ail of the shifted input vector data samples 86S comprise the shifted input vector data sample set 86S(0)-86S(X).
- the tapped-delay lines 78 shift the input vector data sample set 86(0)-86(X) to provide a shifted input vector data sample set 86S(0)-86S(X), to execution unit inputs 90(0)-9Q(X) of the execution units 84(0)-84(X) during a correlation vector processing operation.
- intermediate correlation results based on the operations performed on the shifted input vector data sample set 86S(0)-86S(X) do not have to be stored, shifted, and re-fetched from the vector data files 82(0)-82(X) during each processing stage of a correlation vector processing operation performed by the VPE 22(2),
- the tapped- delay lines 78 can reduce power consumption and increase processing efficiency for a correlation vector processing operation performed by the VPE 22(2).
- the execution units 84(0)-84(X) also receive the reference vector data sample 130 from among the reference vector data sample set 130(0)-130(X) stored in a sequence number generator (SNG) 134 for the correlation vector processing operation.
- the execution units 84(0)-84(X) are configured to correlate the reference vector data sample set 130(0)- 130(X) with the input vector data sample set 86(0)-86(X) as part of the correlation vector processing operation.
- the sequence number generator 134 could also be a register or other file.
- the sequence number generator 134 is provided in this embodiment to provide the reference vector data sample set 130(0)-130(X), because the con'elation vector processing operation in this example is for a CDMA con'elation vector instruction.
- the reference vector data sample set 130(0)-130(X) is provided as a generated chip sequence for use in signal extraction from the input vector data sample set 86(0)-86(X) if the correlation between the reference vector data sample set 130(0)-13G(X) and the input vector data sample set 86(0)-86(X) is high.
- the correlation vector processing operation for a CDMA vector correlation instraction could provide a correlation between on-time input vector data samples 86 in the input vector data sample set 86(0)-86(X) and late input vector data samples in the input vector data sample set 86(0)-86(X).
- the on-time input vector data samples 86 in the input vector data sample set 86(0)-86(X) may be the even input vector data samples 86 in the input vector data sample set 86(0)-86(X) (e.g., 86(0), 86(2), 86(4), ... 86(X-i)).
- the late input vector data samples 86 in the input vector data sample set 86(0)-86(X) may be the odd input vector data samples 86 in the input vector data sample set 86(0)-86(X) (e.g., 86( 1), 86(3), 86(5), ... 86(X)).
- the on- time input vector data samples 86 may be the odd input vector data samples 86
- the late input vector data samples 86 may be the even input vector data samples 86.
- an on-time correlation vector processing operation may be provided according to the following:
- n is the number of input signal samples
- x[n] is the digitized input signal 66
- y[n] is the reference signal
- / is the sample number
- a late correlation vector processing operation may be provided according to the following:
- y[n] is the reference signal
- the reference signal y[n] (i.e.. reference vector data samples) may be complex.
- the VPE 22(2) may receive a reference signal (e.g., from the sequence number generator 134).
- the VPE 22(2) may use the received reference signal directly to perform the on-time and late correlation operations, in which case the reference signal y[n] in the above equations may represent the received reference signal.
- the VPE 22(2) may compute the complex conjugate of the received reference signal before using the reference signal to perform the on-time and late correlation operations, in which case the reference signal y[n] in the above equations may represent the conjugate of the received reference signal.
- the execution units 84(0)-84(X) are each configured to multiply the reference vector data sample set 130(0)-130(X) with the shifted input vector data samples 86S(0), 86S(1), ... 86S(X) of the shifted input vector data sample set 86S(0)-86S(X) during each processing stage of the correlation vector processing operation to provide intermediate correlation output vector data samples in the execution units 84(0)-84(X).
- the intermediate correlation output vector data sample sets are accumulated in each of the execution units 84(0) -84(X) (i.e., prior accumulated correlation output vector data sample is added to current correlation output vector data sample).
- This provides the final, resultant correlated output vector data sample set 132(G)- 132(X) provided by the execution units 84(0)-84(X) on execution unit outputs 96(G)-96(X) on the output data flow paths 98(Q)-98(X), respectively, for each input vector data sample set 86(0), 86(1), ... 86(X) to be stored back in the respective vector data files 82(0)-82(X) for further use and/or processing by the VPE 22(2) without having to store and shift intermediate correlation output vector data sample sets generated by the execution units 84(0)-84(X).
- the same components and architecture provided in the VPE 22(2) in Figure 1 1 is provided in the VPE 22(1 ) in Figure 4.
- the sequence number generator 134 is added and multiplexed by a multiplexor 136 with the global register file 40 that can provide the filter coefficients 92(0)-92(Y- 1 ) or other data to be processed with the reference vector data sample set 130(0)-130(X).
- the VPE 22(2) in Figure 1 1 can provide both the aforementioned filter vector processing operations and correlation vector processing operations discussed here and in more detail below by control of the multiplexor 136.
- the multiplexor 136 can be controlled by a selector signal 138 that is controlled based on the vector instruction being executed by the VPE 22(2).
- the selector signal 138 can be configured to provide filter coefficients 92(0)-92(Y-l) from the global register file 40 to be provided to the execution units 84(0)-84(X).
- the selector signal 138 can be configured to select the reference vector data sample set 130(0)-130(X) from the sequence number generator 134 to be provided to the execution units 84(0)-84(X).
- the tapped-delay lines 78(0), 78( 1) are programmable to be controlled according to the vector instruction being processed. If a correlation vector instruction or other instruction that does not employ the tapped-delay lines 78 is not being processed, the tapped-delay lines 78 can be programmed to not be included in the input data flow paths 80(0)-80(X) between the vector data files 82(0)-82(X) and the execution units 84(0)-84(X).
- two tapped-delay lines 78 are provided, a primary tapped-delay line 78(0) and a shadow tapped-delay line 78(1), with the shadow-tapped delay line 78(1 ) being optional in this embodiment.
- a separate shifting process would have to be performed to provide the shifted intermediate input vector data sample set again to the execution units 84(0)-84(X), thereby increasing latency and consuming additional power.
- the efficiency of the input and output data flow paths 80(0)- 80(X), 98(0)-98(X) in the VPE 22(2) are not limited by the re-fetching delay of the shifted input vector data sample set 86S(0)-86S(X) from the vector data files 82(0)- 82(X) during a correlation vector processing operation.
- the shifted input vector data sample set 86S(0)-86S(X) is provided by the tapped-delay lines 78 localized to the execution units 84(0)-84(X).
- the vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations.
- the correlation vector processing operations performed by the VPE 22(2) in Figure I i may be made more precise by employing the tapped-delay lines 78, because output accumulations for intermediate correlation processing stages in the execution units 84(0)-84(X) do not have to be stored in the vector data files 82(0)- 82(X). Storing of intermediate vector data sample sets from the execution units 84(0)- 84(X) in the vector data files 82(0)-82(X) may result in rounding. Thus, when the next intermediate vector data sample set would be provided to the execution units 84(0)- 84(X) for the vector processing operation, any rounding error would be propagated and added during each multiplication phase of the vector processing operation.
- the intermediate correlation output vector data sample sets calculated by the execution units 84(0)-84(X) do not have to be stored in the vector data files 82(0)-82(X).
- Prior intermediate correlation output vector data sample sets can be accumulated with intermediate correlation output vector data sample sets for next correlation output vector data sample sets, because the tapped-delay lines 78 provide the shifted input vector data sample sets 86S(0)-86S(X) to the execution units 84(0)-84(X) during the vector processing operation to be processed, and the results accumulated with prior vector data sample sets for prior correlation output vector data sample sets.
- FIGS 12A and 12B are flowcharts illustrating an exemplary correlation vector processing operation 140 that can be performed in the VPE 22(2) in Figure 1 1 employing the tapped-delay lines 78 according to an exemplary correlation vector instruction.
- Figures 12A and 12B are flowcharts illustrating an exemplary correlation/covariance vector processing operations that can be performed in parallel in VPE 22(2) in Figure 1 1 with fetched interleaved on-time and late input vector data sample sets according to an exemplary correlation/covariance vector processing operation.
- the input vector data sample set 86(0)-86(X) to be processed in a correlation vector processing operation 140 according to a correlation vector instruction is fetched from the vector data files 82(0)- 82(X) into the input data flow paths 80(0)-80(X) for a correlation vector processing operation 140 (block 142), As discussed above with regard to the VPE 22(2) in Figure 1 1 , the input vector data sample set 86(0)-86(X) is multiplied by the reference vector data sample set 130(0)-130(X) received from the sequence number generator 134 in the execution units 84(0)-84(X).
- Figure 13 illustrates the reference vector data sample set I30(0)-130(X) in the sequence number generator 134.
- Figure 6B previously discussed above illustrated an exemplary input vector data sample set 86(0)- 86(X) stored in the vector data files 82(0)-82(X), which is also applicable in this example and thus will not be re-described here.
- Either one, some, or all of the vector data lanes 100(0)- 100(X) in the VPE 22(2) in Figure 1 1 can be employed to provide the correlation vector processing operation 140 according to the programming of the vector instruction depending on the width of the input vector data sample set 86(0)-86(X) and the reference vector data sample set 130(0)-130(X) to be correlated in the correlation vector processing operation 140. if the entire width of the vector data files 82(0)-82(X) is required, all vector data lanes 100(0)- 100(X) can be employed for the correlation vector processing operation 140. Note that the correlation vector processing operation 140 may only require a subset of the vector data lanes 100(0)-100(X) that may be employed for the correlation vector processing operation 140.
- the width of the input vector data sample set 8 ⁇ (0)-8 ⁇ ( ⁇ ) is less than the width of all vector data files 82(0)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the correlation yector processing operation 140.
- the input vector data sample set 86(0)-86(X) and the reference vector data sample set 130(0)-130(X) employed in the correlation vector processing operation 140 involves all yector data lanes 100(0)-100(X) in the VPE 22(2).
- a fetched input vector data sample set 86(0) -86(X) is provided into the input data flow paths 80(0)-80(X) from the vector data files 82(0)-82(X) to be loaded into the tapped delay-lines 78 as a first input vector data sample set 86S(0)-86(X) for the correlation vector processing operation 140 (block 144).
- An input vector data sample set 86(0)- 86(X) is loaded into the primary tapped- delay line 78(0) as the input vector data sample set 86(0)-86(X) to be processed by the execution units 84(0)-84(X) for the correlation vector processing operation 140.
- the input vector data sample set 86(0)-86(X) loaded into the primary tapped-delay line 78(0) is not shifted for the first operation of the correlation vector processing operation 140.
- a next input vector data sample set 86N(Q)-86 (X) can also be loaded into the shadow tapped-delay line 78(1 ) as a next input vector data sample set 86N(0)-86N(X) to be processed by the execution units 84(0)-84(X).
- the purpose of the tapped-delay lines 78 is to provide shifting of the input vector data sample set 86(0)-86(X) to provide a shifted input vector data sample set 86S(0)-86S(X) to the execution units 84(0)-84(X) for subsequent correlation operations during operation of the coixelation vector processing operation 140.
- the input vector data samples 86 are shifted in the primary-tapped delay line 78(0) to provide the shifted input vector data sample set 86S(0)-86S(X) to the execution units 84(0)-84(X). In this manner, the input vector data sample set 86(0)-86(X) does not have to be stored, shifted in the vector data files 82(0)-82(X), and re- fetched for each correlation operation of the correlation vector processing operation 140.
- Figure 14 illustrates the exemplary tapped-delay lines 78 that can be provided in VPE 22(2) in Figure 11 .
- the tapped-delay lines 78 comprise the shadow tapped-delay line 78(1) and the primary tapped-delay line 78(0).
- the primary-tapped delay line 78(0) in this example is comprised of a plurality of 8 -bit primary pipeline registers 120 to allow resolution of input vector data samples 86 down to 8-bits in length.
- the first input vector data sample set 86(0)-86(X) processed by the execution units 84(0)-84(X) will be un-shifted in this example for the first correlation operation of the correlation vector processing operation 140.
- the input vector data samples 86 in the input vector data sample set 86(0)-86(X) stored in the primary tapped-delay line 78(0) are shifted in the primary pipeline registers 120(0)- 120(4X+3), as indicated by the arrows in Figure 14, to become the shifted input vector data sample set 86S(0)-86S(X).
- the execution units 84(0)-84(X) are fully utilized by receiving and performing the correlation vector processing operation 140 of the shifted input vector data sample set 868(Q)-86S(X) without having to store, shift, and re-fetch the input vector data sample set 86 (0)-86 (X) from the vector data files 82(0)-82(X).
- the number of shifts performed in the primary and shadow tapped-delay lines 78(G), 78(1 ) for the correlation vector processing operation 140 will be dependent on the number of samples to be correlated, if the number of input vector data samples 86 in the input vector data sample set 86(0)-86(X) fetched into the primary tapped-delay line 78(0) and shadow tapped-delay line 78( 1) from the vector data files 82(0)-82(X) is greater than the number of correlation operations in the correlation vector processing operation 140, the execution units 84(0)-84(X) can perform the correlation vector processing operation 140 without any further input vector data sample sets 86(0)-86(X) being re-fetched from the vector data files 82(0)-82(X).
- the number of correlation operations in the correlation vector processing operation 140 is greater titan the number of input vector data samples 86 in the input vector data sample set 86(0)- 86(X) fetched into the primary tapped-delay line 78(0) and shadow tapped-delay line 78(1) from the vector data files 82(0)-82(X), additional input vector data sample sets 86(0)-86(X) can be fetched from the vector data files 82(0)-82(X) as part of the correlation vector processing operation 140.
- the primary pipeline registers 120(0)- 120(4X+3) collectively are the width of the vector data files 82(0)-82(X),
- the vector data files 82(0)-82(X) being 512-bits in width with "X" equal to fifteen ( 15)
- there will be sixty-four (64) total primary pipeline registers 120(0)- 120(63) each eight (8) bits in width to provide a total width of 512 bits (i.e., 64 registers X 8 bits each).
- the primary tapped-delay line 78(0) is capable of storing the entire width of one (1) input vector data sample set 86(0)-86(X).
- the input vector data sample set 86(0)-86(X) can be shifted down to a vector data sample size of eight (8) bits for 8-bit correlation vector processing operations, if larger input vector data sample 86 sizes are desired for a correlation vector processing operation 140, such as 16 -bit or 32-bit samples for example, the input vector data sample set 86(0)-86(X) can be shifted in the primary pipeline registers 120(0)-120(4X+3) by two (2) primary pipeline registers 120 at a time.
- Figure 15A illustrates an input vector data sample set 86(0)-86(X) loaded from the vector data files 82(0)-82(X) into the primary tapped-delay line 78(0) during a first clock cycle (CYC-LEO) of a correlation vector processing instruction 140.
- the first input vector data sample set 86(0)-86(X) is loaded into the primary tapped-delay line 78(0) as input vector data samples X1-X32, but sixty- four (64) input vector data samples are provided.
- the primary pipeline registers 120(0)-120(2X+1 ) are loaded with on-time and late input vector data samples 86 from the input vector data sample set 86(0)-86(X).
- a special vector instruction may be supported to load the on-time and late input vector data samples of the input vector data sample set 86(0)-86(X) into the primary tapped-delay line 78(0) (and also the shadow- tapped delay line 78(1 ), as discussed in more detail later below).
- primary pipeline registers 12.2(0), 122( 1), 122(2X-K2), and 122(2X+3) collectively contain input vector data sample 86(0).
- Primary pipeline registers 122(0), 122(1 ) contain on-time input vector data sample 86OT(0), which are X(0) and X(l), where "OT” means “on- time.”
- Primary pipeline registers 122(2X+2), 122(2X+3) contain late input vector data samples 86L(0), which are X(l) and X(2), where "L” means 'late.”
- This input vector data sample 86 storage pattern in the primary tapped-delay line 78(0) is repeated for the other primary pipeline registers 122(2)-122(2X+ 1) and 122(2X+4)-122(4X+3) (see Figure 14).
- the shadow tapped-delay line 78(1) is also provided in the tapped-delay line 78.
- the shadow tapped-delay line 78(1) can be employed to latch or pipeline a next input vector data sample set 86N(0)-86N(X) from the vector data files 82(0)-82(X) for a subsequent vector processing operation.
- the shadow tapped-delay line 78(1) is also comprised of a plurality of 8-bit shadow pipeline registers 122 to allow resolution of input vector data samples down to 8-bits in length similar to the primary tapped-delay line 78(0).
- the shadow pipeline registers 122 collectively are the width of the vector data files 82(0)-82(X), which is 532-bits in this example, so that the shadow tapped-delay line 78(1) is also capable of storing the entire width of one (1 ) input vector data sample set 86(0)-86(X) just like the primary tapped- delay line 78(0).
- the number of shadow pipeline registers 122(0)- 122(4X+3) included in the primary tapped-delay line 78(0) is four times the number of vector data lanes 100(0)- 100(X), which total sixteen (16), each vector data lane 100(0)-100(X) capable of supporting 32-bits each in this example.
- the number of primary pipeline registers 120 also totals sixty-four (64) in this example for a total of 512 bits (i.e., 64 registers x 8 bits each).
- Figure 15B illustrates a next input vector data sample set 86N(0)-86N(X) loaded into the shadow tapped-delay line 78(1) during a second clock cycle (CYCLEl ) of a correlation vector processing instruction 140.
- the next input vector data sample set 86N(0)-86N(1) is loaded into the shadow tapped-delay line 78(1 ) after the first input vector data sample set 86(0)-86(X) from the vector data files 82(0)-82(X) is loaded into the primary tapped-delay line 78(0) to setup the execution of a correlation vector processing operation 140,
- This next input vector data sample set 86N(0)-86N(X) is loaded into the shadow tapped-delay line 78(1 ) as input vector data samples X(32j- X(63), with both on-time and late input vector data samples 860T, 86L.
- X(32) and X(33) form the on-time input vector data samples 86QT of the input vector data sample 86(0)
- X(33) and X(34) form the late input vector data samples 861. of the input vector data sample 86(0), like the storage pattern provided in the primary tapped-delay line 78(0) discussed above. Other patterns could be provided to group the input vector data samples 86 together to form the input vector data sample set 86(0)-86(X).
- the reference vector data samples 130 correlated during a first processing stage of the correlation vector processing operation 140 from the reference vector data sample set 130(0)- 130(X) from the sequence number generator 134 are also shown as provided in a register ("C") to the execution units 84(0)-84(X) in Figure 15B for use in the coiTelation vector processing operation 140.
- the shift pattern provided between the tapped-delay lines 78(0) and 78(1 ) in Figure 14 is different than the shift pattern provided between the tapped-delay lines 78(0) and 78(1) in Figure 7, As shown in Figure 14, the on-time input vector data samples 860T are shifted from the shadow pipeline register 122(0) in the shadow tapped-delay line 78(1) to primary pipeline register 120(2X+1) in the primary tapped-delay line 78(0).
- the late input vector data samples 86L are shifted from the shadow pipeline register 122(2X+2) in the shadow tapped-delay line 78(1) to primary pipeline register 120(4X+3) in the primary tapped-delay line 78(0).
- the on-time input vector data samples 860T and late input vector data samples 860T are kept segregated from each other in the tapped-delay lines 78(0), 78(1) as the shifting of input vector data samples 86 occurs during the correlation vector processing operation 140.
- the correlation vector processing operation 140 processing stages progress in the execution units 84(0)-84(X), eventually, the entire next input vector data sample set 86 (0)-86N(X) initially stored in the shadow tapped-delay line 78(1 ) is shifted fully into the primary tapped-delay line 78(0) to be provided to the execution units 84(0)- 84(X) for processing.
- the execution units 84(0)-84(X) can then be provided with the previously next input vector sample data set 86N(0)-86N(X) stored in the primary tapped-delay line 78(0) as the current input vector data sample set 86(0)-86(X) for a next correlation vector processing operation 140, if desired, without delay.
- the first input vector data sample set 86(0) ⁇ 86(X) and next input vector data sample set 86N(0)-86N(X) are loaded into the primary tapped-delay line 78(0) and the shadow tapped-delay line 78(1), respectively, as shown in Figure 15B
- the first input vector data sample set 86(0)-86(X) provided in the primary tapped-delay line 78(0) is provided to the respective execution units 84(0)-84(X) to be processed in a first processing stage of the correlation vector processing operation 140 (block 146 in Figure 12A).
- the first input vector data sample set 86(0)-86(X) becomes the current input vector data sample set 86(0)-86(X) being processed by the execution units 84(0)-84(X).
- the current input vector data sample 86(0) is provided to execution un t 84(0), the current input vector data sample 86(1) is provided to execution unit 84(1), and so on, as illustrated in the VPE 22(2) in Figure 1 1.
- the reference vector data input samples 130(0)-130(X) to be correlated with the input vector data sample set 86(0)-86(X) are provided to the execution units 84(0)-84(X) in the current processing stage of the correlation vector processing operation 140 (block 148 in Figure 12A).
- the result of the correlation is a current on-time correlation output vector data sample set R(OT)[n] and a current late correlation output vector data sample set R(L)[nj.
- the execution units 84(0)-84(X) then accumulate each current resultant correlation vector data sample set with its corresponding prior resultant correlation vector data sample set calculated by the execution units 84(0)-84(X) to provide the new prior input vector data sample sets 86(0)-86(X) (block 152 in Figure 12B).
- the first/ current resultant correlated output vector data sample set 132(0)- 132(X) will simply become the prior input vector data sample set 86(0)-86(X) for the second, next processing stage of the correlation vector processing operation 140.
- the shifted input vector data sample set 86S(G)-86S(X) is shifted in the tapped-delay lines 78(0), 78( 1) to the next position for the correlation vector processing operation 140 to provide the shifted input vector data sample set 86S(0)-86S(X) (block 156 in Figure 12B).
- the shifted input vector data sample set 86S(0)-86S(X) is provided for calculating a next resultant correlation output vector data sample set 132(0)-132(X) to be accumulated with the prior resultant correlation output vector data sample set 132(0)-132(X).
- Figure 15C illustrates the contents of the tapped-delay lines 78 when the input vector data sample set 86(0)-86(X) is shifted in a second processing stage of the correlation vector processing operation 140 to become the new shifted input vector data sample set 86S(0)-86S(X) for a next correlation processing operation 140
- R(OT)[n] y[2] * x[2+n] for on-time input vector data samples 86SOT and
- R(L)[n] y[3] * x[3+n] for late input vector data samples 86SL.
- the input vector data sample set 86(0)-86(X) in the primary tapped-delay line 78(0) is shifted by two input vector data samples 86.
- input vector data sample 860T(1) in Figure 15B of X(2) and X(3) is now shifted into input vector data sample 86S(0) in Figure 15C.
- the shifted input vector data sample set 86S(0)-86S(X) becomes the current input vector data sample set 86(0)- 86(X).
- the reference vector data samples 130 provided to the execution units 84(G)- 84(X) are also the reference vector data samples 130, which are Y(2) and Y(3) in this example.
- the process repeats by providing the next shifted input vector data sample set 86S(0)-86S(X) from the primary tapped- delay line 78(0) (and from a portion of the shadow tapped-delay line 78(1 )) to the execution units 84(0)-84(X) to be multiplied with the next reference vector data samples 130 (block 150 in Figure 12 A), with the resultant correlated output vector data sample set 132(0)- 132(X) being accumulated with the prior resultant con'elated output vector data sample set 132(0)-132(X) (block 152 in Figure 12B).
- Figure 15D illustrates the state of input vector data samples 86 present in the tapped-delay lines 78(0), 78(1 ) d uring the last processing stage of the exemplary correlation vector processing operation 140.
- there were sixteen (16) processing stages for the correlation vector processing operation 140 because the full data width of the tapped-delay lines 78 were employed for the input vector data sample set 86(0)-86(X), but split among on-time and late input vector data samples 860T, 86L.
- Y(30) and Y(31) are the last reference vector data samples 130(X) in the correlation vector processing operation 140 as shown in Figure 5D, which is reference vector data samples 130(15) in the example of Figure 13,
- the shifted input vector data sample set 868(0)-86S(X) has bee shifted sixtee (16) times (the width of the vector data lanes i00(0)- .100(X) in this example) such that input vector data samples X(30) and X(31 ) are stored in the shified input vector data sample 86S(0) in the primary tapped-delay line 78(0) in the final, sixteenth, processing stage for the correlation vector processing operation 140.
- Figure 16 is a schematic diagram of contents of accumulators (i.e., resultant correlated output vector data samples 132) in the execution units 84(0)-84(X) in the VPE 22.(2) of Figure I I after the exemplary sixteen (16) correlation vector processing stages in the above example have been fully executed.
- the resultant correlated output vector data sample sei is shown as 132(0)-132(X).
- Accumulators Acc0-Acc3 are shown in Figure 16, because in this example, each execution unit 84(0)-84(X) has four accumulators disposed in parallel for each vector data lane 100(0)- 100(X).
- the accumulated resultant output vector data samples can be provided on the output data flow paths 98(0)-98(X) to the vector data files 82(0)-82(X) as a collective resultant correlated output vector data sample set 132(G)- 132(X) to be stored therein for further analysis and/or processing.
- a specialized vector instruction may be supported by the VPE 22(2) to move rows of the resultant correlated output vector data sample set 132(0)- 132(X) from the vector data files 82(0)-82(X) to the vector unit data memory 32 (see Figure 2), if desired.
- Resultant output vector data sample sets provided by the execution units 84(0)-84(X), including the resultant filter vector output data sample sets 94(0)-94(X) and the resultant correlated output vector data sample sets 132(0)-132(X) described above, can be stored back in the vector data files 82(0)-82(X), 82(31) in different interleaved formats depending on the vector instruction executed by the VPE.
- 'X' is equal to thirty-one (31 ) in this example to provide the vector data files 82(G)-82(X), each thirty-two (32) bits width.
- a resultant output vector data sample set 158(0)-158(X), 158(31) can be stored in the vector data files 82(0)-82(X) separated by their real ("q") and imaginary ("i") components.
- the resultant output vector data sample set 158(0)-158(X) is comprised of ' ⁇ + ⁇ resultant output vector data samples 158, which in this example are 158(0), 158(1), ... , and 1 8(X).
- the resultant output vector data sample set 158(0)-158(X), 158(31) may be more efficient to store the resultant output vector data sample set 158(0)-158(X), 158(31) separated by their real ("q") and imaginary ("i") components for efficiency purposes, such as if a next vector instruction operates on real and imaginary components of the resultant output vector data sample set 158(0)-158(X), 158(31) as an input vector data sample set.
- the thirty -two (32) bit resultant output vector data sample 158 could be Y0 in Figure ⁇ 7 ⁇ .
- the imaginary component of Y0, YO.i 158(1), can be stored in ADDRESS '0' of vector data file 82(0), and the real component of Y0, YO.q 158(Q), can be stored in another ADDRESS, such as ADDRESS ⁇ .”
- the resultant output vector data sample set 158(0)- 158(X), 158(31 ) in Figure 17A could be stored in the vector data files 82(0)-82(X), 82.(31 ) interleaved by even and odd resultant output vector data samples. This is illustrated by example in Figure ⁇ 7 ⁇ .
- resultant output vector data sample Y0-Y31 158(0)- 158(X), 158(3 i) is stored in an interleaved format by even and odd vector data samples among ADDRESS '0' and ADDRESS ⁇ ' in vector data files 82(0)-82(31).
- Resultant output vector data sample Y0 158(0) is stored in ADDRESS '0' in vector data file 82(0).
- Resultant output vector data sample Yl 158( 1) is stored not in ADDRESS '0' in vector data file 82( 1), but in ADDRESS 'A' in vector data file 82(0).
- Resultant output vector data sample Y2 158(2) is stored in ADDRESS '0' in vector data file 82(1 ), and so on.
- Certain wireless baseband operations require data samples to be format- converted before being processed.
- the resultant output vector data sample sets 158(0)-158(X) stored in the vector data files 82(0)-82(X) in interleaved format in Figures 17A and 17B may need to be de-interleaved for a next vector processing operation.
- the resultant output vector data samples 158(0)-158(X) represent a CDMA signal
- the resultant output vector data samples 158(0)-158(X) may need to be de-interleaved to separate out even and odd phases of the signal.
- the de- interleaved signal may also be correlated with a locally generated code or sequence number in a correlation processing operation to determine if the CDMA system can extract the signal, such as with the exemplary correlation vector processing operation described above in regard to Figures 1 1 -16.
- Conventional programmable processors implement format conversion of data samples in multiple steps, which add cycles, power consumption, and data flow complications in vector data sample format conversions.
- Vector processors can pre-process the vector data samples to provide format conversions before the format-converted vector data samples are provided to execution units.
- the format-converted vector data samples are stored in vector data memory and re-fetched as part of a vector processing operation requiring data format conversion to be processed by execution units.
- this format pre-processing of the vector data samples delays the subsequent processing of the format-converted vector data samples by the execution units, and causes computational components in the execution units to be underutilized.
- FIGS 18A and 18B illustrate a vector data sample set D(0)-D(X) stored in vector data files 82(0)-82(X) in different formats.
- Figure I SA illustrates the vector data sample set D(G)-D(X) stored in signed complex (SC) sixteen-bit samples (SCI 6) and format-interleaved by real and imaginary components.
- the sixteen (16) bit real and imaginary components of thirty-two (32) bit vector data sample D(0) - D(0)(Q) and D(0)(I) are stored in thirty-two (32) bit vector data file 82(0).
- FIG. 18B illustrates the vector data sample set D(0)-D(X) stored in SC eight-bit samples (SC8) and format-interleaved by real and imaginary components.
- the eight (8) bit real and imaginary components of sixteen ( 16) bit vector data sample D(0)(I) - D(0)(1 )(Q), D(0)( i)(I), are stored in vector data file 82(0).
- the eight (8) bit real and imaginary components of sixteen (16) bit vector data sample D(0)(0) - D(0)(0)(Q), D(0)(0)(i), are also stored in thirty-two (32) bit vector data file 82(0).
- the eight (8) bit real and imaginary components of sixteen (16) bit vector data sample D(X)( 1) ⁇ D(X)(1)(Q), D(X)(1)(I), are stored in thirty-two (32) bit vector data file 82(X).
- the eight (8) bit real and imaginary components of sixteen (16) bit vector data sample D(X)(0) - D(X)(0)(Q), D(X)(0)(I) are also stored in thirty-two (32) bit vector data file 82(X).
- Figure 19 is a schematic diagram of another exemplary VPE 22(3) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(3) in Figure 19 is configured to provide in-flight format conversion (e.g., de- interleaving) of input vector data sample sets provided to execution units for vector processing operations in the VPE 22(3) with eliminated or reduced vector data sample re-fetching and reduced power consumption
- in-flight format conversion of input vector data sample sets means the input vector data sample set retrieved from vector data memory is format-converted without having to be stored and re-fetched from vector data memory before being provided to execution units for execution.
- format conversion circuitry 159(0)- 159(X) is included in each of the vector data lanes 100(G)- 100(X) between the vector data files 82(0)-82(X) and the execution units 84(0)-84(X),
- the input vector data sample set 86(0) ⁇ 86(X) from the vector data files 82(0)-82(X) is format-converted (e.g., de-interleaved) in the format conversion circuitry 159(0)- 159(X) in the VPE 22(3) to provide a format- converted input vector data sample set 86F(0) ⁇ 86F(X) to the execution units 84(0)- 84(X) for a vector processing operation that requires de-interleaving of the input vector data sample set 86(0)-86(X).
- All of the format-converted input vector data samples 86F comprise the format-converted input vector data sample set 86F(0)-86F(X) in this example.
- 'X'+l is the maximum number of parallel input data lanes provided in the VPE 22(3 ) for processing of input vector data samples 86 in this example.
- format conversion of the input vector data sample sets 86(G)- 86(X) in the VPE 22(3) does not require pre-processing, storage, and re-fetching from vector data files 82(0)-82(X), thereby reducing power consumption. Further, because the format conversion of the input vector data sample sets 86(0)-86(X) does not require pre-processing, storage, and re-fetching of tire format-converted input vector data sample sets 86(0)-86(X) from vector data file 82(0)-82(X), the execution units 84(G)- 84(X) are not delayed from performing vector processing operations.
- the efficiency of the data flow paths in the VPE 22(3) are not limited by format conversion pre-processing delays of the input vector data sample sets 86(0)-86(X).
- the format- converted (e.g., de-interleaved) input vector data sample sets 86F(0)-86F(X) are provided localized to the execution units 84(0)-84(X).
- the vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations,
- the format conversion circuitry 159(0)-159(X) can be included in the optional primary tapped- delay line 78(0). This arrangement provides the format conversion circuitry 159(0)- i59(X) in the input data flow paths 80(0)-80(X) between the vector data files 82(G)- 82(X) and execution units 84(0)-84(X) in the VPE 22(3) in Figure 19.
- the primary and shadow tapped- delay lines 78(0), 78(1) may be employed for the vector processing operation, requiring format-converted input vector data sample sets 86F(0)-86F(X) to be provided to the execution units 84( ' 0)-84(X), which in turn also requires format-converted, shifted input vector data sample sets, designated as 86SF(0)-86SF(X).
- FIG. 20 is a flowchart illustrating an exemplary de-interleaving format conversion vector processing operation 160 that can be performed in the VPE 22(3) in Figure 19 employing the format conversion circuitry 159(0)-159(X) according to an exemplary vector instruction requiring format conversion of the input vector data sample set 86(0)- 86(X).
- the input vector data sample set 86(0)-86(X) for a vector processing operation 160 is fetched from the vector data files 82(0)-82(X) into the input data flow paths 80(0)-80(X) (block 162).
- the format conversion for the vector processing operation 160 may be a de-interleaving vector processing operation 160 where the input vector data sample set 86(0)-86(X) is de-interleaved from its interleaved state in the vector data files 82(G)- 82(X) into de-interleaved input vector data sample set 86F(0)-86F(X).
- Either one, some, or all of the vector data lanes 100(0)-100(X) in the VPE 22(3) in Figure 19 can be employed to provide the vector processing operation 160 according to the programming of the vector instruction depending on the width of the input vector data sample set 86(0)-86(X) to be format-converted for the vector processing operation 160. If the entire width of the vector data fifes 82(0)-82(X) is required, all vector data lanes 100(G)- 100(X) can be employed for the vector processing operation 160. The vector processing operation 160 may only require a subset of the vector data lanes 100(0)-100(X) that may be employed for the vector processing operation 160.
- the width of the input vector data sample set 86(0)-86(X) is less than the width of all vector data files 82(0)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the vector processing operation 160.
- the input vector data sample set 86(0)-86(X) format-converted into input vector data sample set 86F(0)-86F(X) for the vector processing operation 160 involves all vector data lanes 1 00(G)- 100(X) in the VPE 22(3) in Figure 19.
- the fetched input vector data sample set 86(0)-86(X) is provided into the input data flow paths 80(0)-80(X) to the format conversion circuitry 159(0)-159(X) to be format- converted according to the vector processing operation 160 (block 164).
- the current input vector data sample set 86(0)-86(X) may optionally be loaded into the primary tapped- delay line 78(0) as the input vector data sample set 86(0)-86(X) to be format-converted before being provided to the execution units 84(0)-84(X) for the vector processing operation 160.
- a next input vector data sample set 86(0)-86(X) may also be optionally loaded into the shadow tapped-delay line 78(1) as a next input vector data sample set 86N(Q)-86N(X) to be processed by the execution units 84(0)- 84(X).
- the purpose of the tapped-delay lines 78 is to shift the input vector data sample set 86(0)-86(X) to shifted input vector data samples 86S(0)-86S(X) to be provided to the execution units 84(0)-84(X) during operation of a vector processing operation 160 operating on shifted input vector data samples 86S.
- the shifted format- converted input vector data sample set is designated as 86SF(Q)-86SF(X).
- the execution units 84(G)- 84(X) may next perform the vector processing operation 160 using the format-converted input vector data sample set 86F(0)-86F(X) (block 166).
- the execution units 84(0)-84(X) may be configured to provide multiplications and/or accumulation using the format- converted input vector data sample set 86F(0)-86F(X).
- the execution units 84(0)-84(X) can receive the shifted, format- converted input vector data sample set 86SF0)-86SF(X) during each processing stage of the vector processing operation 160 until the vector processing operation 160 is completed (block 168).
- a resultant output vector data sample set 172(0)-172(X) based on vector processing with a format-converted input vector data sample set 86F(0)-86F(X), or shifted, format-converted input vector data sample sets 86SF(0)-86SF(X), is provided in the output data flow paths 98(0)-98(X) to be provided and stored in the vector data files 82(0)-82(X) (block 170).
- the resultant output vector data sample set 172(G)- 172(X) is comprised of ' ⁇ + ⁇ resultant output vector data samples 172, which in this example are 172(0), 172(1 ), ... , and 172(X).
- Figure 21 is a schematic diagram of exemplary format conversion circuitry 159(0)-159(X) that receive shifted input vector data sample set 86S(0)-86S(X) from the primary tappcd-delay line 78(0).
- the format conversion circuitry 159(0)- 159(X) is provided on the output of the primary tapped-delay line 78(0) in the input data flow paths 80(0)-80(X).
- the exemplary format conversion circuitry 159(0)- 159(X) will now be described.
- the exemplary format conversion circuitry 159(0)- 159(X) will now be described. Exemplary detail of the internal components of the format conversion circuitry 159(0) is provided in Figure 2.1, but such is also applicable for format conversion circuitry 59(1)-159(X).
- the format conversion circuitry 159(0) in this example is configured to provide de-interleaving and sign extension of input vector data sample 86(0) or shifted input vector data samples 868(0) from the primary pipeline registers 120(0), 120(1) 120(2X+2), 120(2X+3) in vector data lane 100(0) to provide format- converted input vector data samples 86F(0) or shifted, format-converted input vector data samples 868F(0), respectively.
- each multiplexor 174(3)- 174(0) is configured to select either the portion of the shifted input vector data sample 86S(0) in the assigned primary pipeline register 12.0(0), 120( 1), , 120(2X+2), 120(2X+3), or the portion of the shifted input vector dat sample 868(0) to store in a primary pipeline register 120 adjacent to the assigned primary pipeline register 120(0), 120(1), 120(2X+2), 120(2X4-3).
- multiplexor 174(3)- 174(0) selections would be as follows. Multiplexor 174(3) would select the portion of the shifted input vector data sample 86S stored in its assigned primary pipeline register 120(0).
- multiplexor 174(2) would seleci the portion of the shifted input vector data sample 86S stored in primary pipeline register 120(1). This would provide a de-interleaved real portion of the input vector data sample 86S(0) (i.e., real [15:0]) in adjacent input data flow paths 80( ⁇ )(3), 80(0)(2). Similarly, multiplexor 174(0) would seleci the portion of the shifted input vector data sample 86S stored in its assigned primary pipeline register 120(2X+3). However, multiplexor 174(1) would select the portion of the shifted input vector data sample 86S stored in primary pipeline register 120(2X+2).
- Multiplexors 176(1), 176(0) provide the ability, to each multiplexor 174(3)- 174(0), to select a portion of the shifted input vector data sample 868(0) from a non-assigned, non-adjacent primary pipeline registers 120(0), 120(1), 120(2X+2)- 12Q(2X+3), as illustrated in Figure 21.
- the format conversion circuitry 159(0)-159(X) can also be configured to sign extend format-converted input vector data sample sets 86F(0)-86F(X).
- the format conversion circuitry 1 9(0)- 159(X) can be configured to sign extend the de-interleaved vector data samples by extending the most significant bits as '0"s for non-negative numbers, and as "F'"s for negative numbers.
- the format conversion circuitry 159(0)- 159(X) may have a sign extension (SC) input 178(0)-178(X) set according to the vector instruction being executed to indicate if sign extension is to be performed on the format- converted input vector data sample set 86F(0)-86F(X) or not.
- the SC inputs 178(0)- 178(X) can be proyided to sign extension circuitry 180(0)-180( ' X) provided in the format conversion circuitry 159(0)- 159(X) to perform the sign extension according to a programmable data path configuration provided by the SC inputs 178(G)- 178(X) according to the vector instruction being processed.
- the SC inputs 178(0)-178(X) can be configured and reconfigured for each vector instruction to provide for flexibility in the vector processing by the VPE 22(3).
- the programmable data paths in the format conversion circuitry 159(0)- 159(X) can be configured by the SC inputs 178(0)- i78(X) can be configured and reconfigured for each clock-cycle of a vector instruction, on a clock-cycle-by-clock- cycle basis if desired, to provide format conversion as desired, with full utilization of the execution units 84(0)-84(X), if desired.
- the format conversion circuitry 159(0)-159(X) does not have to be provided as part of the primary tapped-delay line 78(0).
- the primary and shadow tapped-delay lines 78(0), 78(1) are optional.
- the format conversion circuitry 159(G)- 159(X) could receive input vector data sample sets 86(0)-86(X) directly from the vector data files 82(0)-82(X). In this scenario as an example, with reference to Figure 21, the input vector data sample set 86(0)-86(X) could be loaded from the vector register files 82(0)-82(X) directly into the primary registers 120(0)- 120(4X+3).
- format conversion circuitry 159(0)-159(X) is provided on the output of the primary tapped-delay line 78(0) to format converted input vector data sample sets 86(0)-86(X), such is not required.
- the format conversion circuitry 159(0)-159(X) in Figure 21 could be provided on the input side of the primary and shadow tapped-delay lines 78(0), 78(1), such that the input vector data sample sets 86(0)-86(X) fetched from the vector data files 82(0)-82(X) are format converted in the format conversion circuitry 159(0)-159(X) prior to being loaded into the primary and shadow tapped-delay lines 78(0), 78(1).
- the input vector data sample sets 86(0)-86(X) would be stored as format-converted input vector data sample sets 86F(0)-86F(X) (or 86SF(0)-86SF(X) after shifting) in the primary and shadow tapped- delay lines 78(G), 78(1),
- the format-converted input vector data sample sets 86F(0)- 86F(X) (or 86SF(0)-86SF(X) after shifting) could then be provided directly from the primary tapped-delay line 78(0) directly to the execution units 84(0)-84(X) for execution in a vector processing operation.
- Figure 22 is a chart 182 that provides an exemplary data format of bits of a vector instruction to control programming of shifting and format conversion of input vector data sample sets 86(0)-86(X) in the VPE 22(3) in Figure 19,
- the data provided in the fields in the chart 182 provide programming to the VPE 22(3) to control whether the format conversion circuitry 159(0)-159(X) and/or tapped-delay lines 78 are included in the input data flow paths 80(0)-80(X) depending on if their functionality is needed for the vector instruction to be processed.
- a bias field 184 (BIAS SCI 6) is provided in bits [7:0] of a vector instruction or vector programming to indicate if a shift bias for arithmetic instructions is provided when using signed complex sixteen (16) bit format (SCI 6) by the tapped-delay lines 78.
- a first source data format conversion field 1 86 (DECIMATE SRC1) is provided in bit [16] of the vector instruction or vector programming to indicate if a first source data (i.e., input vector data sample set 86(G)- 86(X)) should be decimated (i.e., de-interleaved) and converted from SC8 to SC16 formal or not.
- a second source data format conversion field 1 88 (DECIMATE_SRC2) is provided in bit [17] of the vector instruction or vector programming to indicate if a second source data (i.e., input vector data sample set 86(0)-86(X)) should be decimated (i.e., de-interleaved) and converted from SC8 to SC16 format or not.
- An output data format field 190 (DEST FMT) is provided in bit [18] to indicate if an output source data (e.g. resultant output vector data sample set 172(0)- 172(X) in VPE 22(3) in Figure 19) should be stored in SCI 6 format or converted from SCI 6 to SC8 format and reordered when stored in the vector data files 82(0)-82(X).
- a phase format field 192 (DECIMATE PHASE) is provided in bit [19] to indicate if input source data (i.e., input vector data sample set 86(0)-86(X)) and output data (e.g., resultant output vector data sample set 172(0)-172(X) in VPE 22(3) in Figure 19) should be decimated (i.e., de- interleaved) along even (e.g., on-time) and odd (e.g., late) samples, which may be useful for CDMA- specific vector processing operations in particular, as previously described above and in Figure ⁇ 7 ⁇ .
- the execution units 84(0)-84(X) in the VPEs 22 perform vector processing on input vector data sample sets and provide resultant output vector data sample sets on the output data flow paths 98(0)-98(X) as a result
- subsequent vector processing operations may need to be performed on the resultant output vector data sample sets.
- the resultant output vector data sample sets may need to be reordered for subsequent vector processing operations.
- the resultant output vector data sample sets resulting from previous processing operations must be stored in the vector data files 82(0)-82(X), fetched for reordering, and re-stored in reordered formal in vector data files 82(0)-82(X).
- subsequent processing operations may require previously processed vector data samples to be interleaved when stored in the vector data files 82(0)--82.(X), as discussed above in Figures 17A and 17B.
- vector processors can include circuitry that performs post-processing reordering of output vector data from execution units after the output vector data is stored in vector data memory.
- the post-processed output vector data samples stored in vector data memory are fetched from the vector data memory, reordered, and stored back in the vector data memory. This postprocessing delays the subsequent processing of the reordered vector data samples by the execution units, and causes computational components in the execution units to be underutilized.
- Figure 23 is a schematic diagram of another exemplary VPE 22(4) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(4) in Figure 23 is configured to provide in-flight reordering of resultant output vector data sample sets 194(0)-194(X) provided by the execution units 84(0) -84(X) for vector processing operations to be stored in the vector data files 82(0)- 82(X) in the VPE 22(4) with eliminated or reduced vector data sample re-fetching and reduced power consumption.
- the resultant output vector data sample set 194(0)-194(X) is comprised of ' ⁇ + ⁇ resultant output vector data samples 194, which in this example are 194(0), 194(1 ), .... and ! 94; X ;.
- reordering could include interleaving of the resultant output vector data sample sets 194(0)- 194(X) before being stored in the vector data files 82(0)-82(X).
- reordering circuitry 196(0)- 196(X) is provided in the output data flow paths 98(0)-98(X) between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X) in each of the vector data lanes 100(0)- 100(X).
- the reordering circuitry 196(0)- 196(X) is configured based on programming according to a vector instruction to be executed to provide reordering of resultant output vector data sample set 194(0)- 194(X) as reordered resultant output vector data sample set 194R(0)- 194R(X) in the output data flow paths 98(0)-98(X).
- In-flight reordering of the resultant output vector data sample set 194(G)- 194(X) in the VPE 22(4) in Figure 23 means the resultant output vector data sample set i 94(G)- 194(X) provided by execution units 84(0)-84(X) is reordered as reordered resultant output vector data sample set 194R(0)-194R(X) before being stored in vector data files 82(0)-82(X).
- the resultant output vector data sample set 194(0)-194(X) is stored in vector data files 82(0)-82(X) in reordered format as the reordered resultant output vector data sample set 194R(0)- 194R(X).
- the reordering of resultant output vector data sample sets I94(0)-194(X) may include interleaving or de-interleaving of the resultant output vector data sample sets 194(0)-194(X) to be stored as the reordered resultant output vector data sample sets 194R(0)-194R(X) to in the vector data files 82(0)-82(X).
- the resultant output vector data sample set 194(0)- ⁇ 94(X) is not required to first be stored in the vector data files 82(G)-82(X), and then fetched from the vector data files 82(0)-82(X), reordered, and restored in vector data files 82(G)- 82(X).
- the resultant output vector data sample set 194(0)- 194(X) is reordered before being stored in the vector data files 82(0)-82(X).
- the resultant output vector data sample sets 194(0)-194 X) are stored in the reordered format in the vector data files 82(0)-82(X) without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units 84(G)-84(X).
- the efficiency of the data flow paths in the VPE 22(4) are not limited by the reordering of the resultant output vector data sample sets 194(G)- 194(X).
- the subsequent vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations when the resultant output vector data sample sets 194(G)- 194(X) are to be stored in reordered format as reordered resultant output vector data sample sets 194R(0)-I94R(X) in the vector data files 82(0)- 82(X).
- the VPE 22(4) that includes the reordering circuitry 196(G)- 196(X) can also optionally include the primary tapped-delay line 78(0) and/or the shado tapped-delay line 78(1).
- the operation of the tapped-delay lines 78(0), 78(1) was previously described above with regard to VPEs 22(1) and 22(2).
- the tappcd-delay lines 78(0), 78(1) may be employed for the vector processing operation requiring shifted input vector data sample sets 868(G)- 86S(X) to be provided to the execution units 84(0)-84(X).
- the reordering circuitry 196(0)-196(X) is configured to receive the resultant output vector data sample sets 194(G)- 194(X) on reordering circuitry inputs 198(G)- 198(X) on the output data flow paths 98(0)-98(X),
- the reordering circuitry 196(0)-196(X) is configured to reorder the resultant output vector data sample sets 194(0)-194(X) to provide the reordered resultant output vector data sample sets 194R(0)-194R(X).
- the reordering circuitry 196(0)- 196(X) is configured to provide the reordered resultant output vector data sample sets 194R(0)-194R(X) on reorder circuitry outputs 200(0)-200(X) in the output data flow paths 98(0)-98(X) to be provided to the vector data files 82(0)-82(X) for storage.
- Figure 24 is a flowchart illustrating an exemplary reordering of resultant output vector data sample sets 194(G)- 194(X) resulting from a vector processing operation 202 that can be performed in the VPE 22(4) in Figure 23 employing the reordering circuitry 196(G)- 196(X) according to an exemplary vector instruction requiring reordering of the resultant output vector data sample set 194(0)-194(X).
- the input vector data sample set 86(0)- 86(X) to be processed according to the vector processing operation 202 according to a vector instruction is fetched from the vector data files 82(0)-82(X) and provided in the input data flow paths 80(0)-80(X) (block 204 in Figure 24).
- the vector processing operation 202 can involve any vector processing operation desired according to the vector instruction to be executed. Non-limiting examples including the filter, correlation, and the format conversion vector processing operations described above.
- Either one, some, or all of the vector data lanes 100(0)-100(X) in the VPE 22(4) in Figure 23 can be employed to provide the vector processing operation 202 according to the programming of the vector instruction depending on the width of the input vector data sample set 86(G) -86(X) for the vector processing operation 2.02, If the entire width of the vector data files 82(Q)-82(X) is required, all vector data lanes 100(0)- 100(X) can be employed for the vector processing operation 202.
- the vector processing operation 202. may only require a subset of the vector data lanes 10Q(Q)-100(X).
- the width of the input vector data sample set 86(0)-86(X) is less than the width of all vector data files 82(G)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the vector processing operation 202.
- the fetched input vector data sample set 86(G)-86(X) is received from the input data flow paths 80(0) ⁇ 80(X) at the execution units 84(0)-84(X) (block 206 in Figure 24).
- the execution units 84(0)-84(X) perform vector processing on the received input vector data sample set 86(0)-86(X) according to the vector processing operation 202 provided according to a vector instruction (block 208 in Figure 24).
- the input vector data sample set 86(0)-86(X) may optionally be loaded into the primary tapped-delay line 78(0) as the input vector data sample set 86(0)-86(X) to be shifted during execution of the vector processing operation 202 during each processing stage of the vector processing operation 202 executed by the execution units 84(0)-84(X) that involve shifting of the input vector data sample set 86(0)-86(X).
- a next input vector data sample set 86 ⁇ ( ⁇ )-86 ⁇ ( ⁇ ) may also be optionally loaded into the shadow tapped-delay line 78(1) as a next input vector data sample set 86N(0)-86 (X) to be processed by the execution units 84(0)-84(X).
- the purpose of the tapped-delay lines 78 is to shift, the input vector data sample set 86(0)- 86(X) to shifted input vector data samples 86S(0)-86S(X) to be provided to the execution units 84(0)-84(X) during operation of a vector processing operation 202 operating on shifted input vector data samples 86S.
- the execution units 84(0)- 84(X) may be configured to provide multiplications and/or accumulation using the input vector data sample set 86(0)-86(X). If the tapped-delay lines 78 are employed to shift the format-converted input vector data sample set 86F(0)-86F(X) during the vector processing operation 202, the execution units 84(0)-84(X) can received the shifted input vector data sample set 86S(0)-86S(X) during each processing stage of the vector processing operation 202 until the vector processing operation 202 is completed, as previously described by example.
- the resultant output vector data sample set 194(0)-194(X) is stored in the vector data files 82(0)-82(X)
- the resultant output vector data sample set 194(0)-194(X) is provided to the reordering circuitry 196(0)-196(X) provided in the output data flow paths 98(0)-98(X) provided between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X).
- the reordering circuitry 196(0)- 196(X) is programmable to be included in the output data flow paths 98(0)-98(X) according to the vector instruction being executed, and if the vector instruction calls for reordering of the resultant output vector data sample set 194(0)- 194(X) to be stored in the vector data files 82(0)-82(X), as discussed below in more detail.
- the reordering circuitry 196(0)- 196(X) reorders the resultant output vector data sample set 194(0)-194(X) according to the reordering provided in the programming according to the vector instruction being executed without the resultant output vector data sample set 194(0)- 194(X) being stored in the vector data files 82(0)-82(X) (block 210 in Figure 24).
- the resultant output vector data sample set 194(0)- 194(X) does not have to first be stored in the vector data files 82(0)-82(X), re-fetched, reordered in a post-processing operation, and stored in reordered format in the vector data files 82(Q)-82(X), thereby providing delay in the execution units 84(0)-84(X).
- the resultant output vector data sample set 194(0)-194(X) is stored as the reordered resultant output vector data sample set 194R(0)-194R(X) in the vector data files 82(0)-82(X) without reordering post-processing required (block 212 in Figure 24).
- the resultant output vector data sample set 194(0)-194(X) may appear in a format like that provided in Figures 18A and 18B before being reordered by the reordering circuitry 196(0)-196(X).
- An example of the reordering circuitry 196(0)- 196(X) will now be described with regard to Figure 25. Exemplary detail of the internal components of the reordering circuitry 196(0)-196(X) is provided for one instance of the reordering circuitry 196(0) provided in vector data lane 100(0) is provided in Figure 25, but such is also applicable for reordering circuitry 196( 1)-196(X).
- the reordering circuitry 196(0) in this example is configured to reorder the resultant output vector data sample 194(0) provided by the execution unit 84(0) in the output data flow path 98(0) in vector data lane 100(0) to provide the reordered resultant output vector data sample 194R(0).
- four output vector data sample selectors 214(3)-214(0), provided in the form of multiplexors in this example, are provided in this example, which are arranged according to the bit widths of the execution unit outputs 96(0), which are four (4) in this example of eight (8) bit widths each 96(0)(3)-96(0)(0).
- Each output vector data sample selector 214(3)-2 4(0) is configured to select either the portion of the resultant output vector data sample 194(0) in the assigned execution unit output 96(0)(3)-96(0)(0), or a portion of the resultant shifted output vector data sample 1 94(0) from an execution unit output 96 adjacent to the assigned execution unit output 96( ' 0)(3)-96(G)(0).
- output vector data sample selector 214(3) -214(0) selections would be as follows. Output vector data sample selector 214(3) would select the resultant output vector data sample 194(0)(3) from execution unit output 96(G)(3) to provide on output data flow path 98(0 M ).
- output vector data sample selector 214(2) would select the portion of the resultant output vector data sample 194(G)(1) on execution unit output 96(G)(1) to provide on output data flow- path 98(G)(2). This would provide an interleaved real portion of the resultant shifted output vector data sample 194(0) (i.e.. real [31 :24], imaginary [23: 16]) in adjacent output data flow paths 98(G)(3), 98(G)(2), as reordered resultant output vector data sample 194R(0)(3), 194R(0)(2) of reordered resultant output vector data sample 194R(0).
- output vector data sample selector 214(0) would select the resultant output vector data sample 194(G)(0) from execution unit output 96(G)(0) to provide in output data flow path 98(G)(0).
- output vector data sample selector 214(1) would select the resultant output vector data sample 194(G)(2) on execution unit output 96(G)(2) to provide on output data flow path 98(G)(1 ). This would provide a reordered, interleaved resultant output vector data samples 194(G)(2), 194(0)(0) (i.e..
- Output vector data sample selectors 216(1), 216(0) also pro vided in the form of multiplexors, provide the ability to select between a resultant output vector data sample 194(G)(3)- 194(G)(0) from a non-assigned, non- adjacent execution unit output 96(G)(3)-96(G)(0), as illustrated in Figure 25.
- the reordering circuitry 196(G)- 196(X) could be provided as being programmable to be configured or reconfigured to not reorder resultant output vector data sample set 194(0)-194(X) according to the vector instruction to be executed.
- the reordering circuitry 196(G)- 196(X) may be programmed to provide for the output data flow paths 98(0)-98(X) to flow straight to the reordering circuitry 196(0)-196(X) without any reordering operations being formed.
- the output data format field 190 (DEST FMT) in chart 182 can be provided in bit [18] of a vector instruction as a non-limiting example to indicate if an output source data (e.g., resultant output vector data sample set 194(G)- 194(X) in VPE 22(4) in Figure 23) should be stored in SCI 6 format or converted from SCI 6 to SC8 format and re-ordered when stored in the vector data files 82(0)-82(X).
- an output source data e.g., resultant output vector data sample set 194(G)- 194(X) in VPE 22(4) in Figure 23
- a programmable reordering data path configuration input 218(0) in Figure 25 can be provided to the reordering circuitry 196(0) to program the reordering circuitry 196(0) to either reorder or not reorder the resultant output vector data samples 194(G)(3)-i 94(G)(0) in the output data flow path 98(0).
- Programmable reordering data path configuration inputs 218(1)-218(X) can also be similarly provided to the reordering circuitry 196(l)-i96(X) to program the reordering circuitry 196( 1)-196(X) to either reorder or not reorder the resultant output vector data sample sets 194(1)-194(X) in the output data flow paths 98(1)-98(X), respectively.
- the reordering circuitry 196(0)-196(X) can be programmed to not reorder the resultant output vector data sample sets 194(0)-194(X) if the vector ins truction does not provide for such processing to be performed.
- the programmable reordering data path configuration inputs 218(0)-218(X) can be configured and reconfigured for each vector instruction to provide for flexibility in the vector processing by the VPE 22(4),
- the programmable reordering data path configuration inputs 218(0)- 218(X) can be configured and reconfigured for each clock-cycle of a vector instruction, on a clock- cycle-by -clock- cycle basis, if desired, to provide reordering as desired, with full utilization of the execution units 84(0)-84(X), if desired.
- a data signal 220 that can be modulated using CDMA is illustrated in Figure 26 A.
- the data signal 220 has a period of 2T.
- the data signal 220 represents the data sequence 1010 in this example, where high signal levels represent a logical T and low signal levels represent a logical ' ⁇ ,' as illustrated in Figure 26A.
- the data signal 220 is spread by a chip sequence 222, such as chip sequence 222 in Figure 26B, which may be a pseudorandom code.
- the chip sequence 222 in this example, has a period that is ten (10) times smaller than the period of the data signal 220 to provide a chip sequence 222 having a spreading rate or factor of ten (10) chips for each sample of the data signal 220 in this example.
- the data signal 220 is exclusively ORed (i.e., XOR'ed) with the chip sequence 222 to provide a spread transmitted data signal 224, as illustrated in Figure 26C.
- Other data signals for other users transmitted in the same bandwidth with the spread transmitted data signal 224 are spread with other chip sequences that are orthogonal to each other and the chip sequence 222.
- the spread transmitted data signal 224 is correlated with the sequence numbers, as previously described above with regard to Figures 11-16. If there is a high correlation between the sequence number and the spread transmitted data signal 224, such as will be the case with chip sequence 222, the original data signal 220 can be recovered using the chip sequence associated with the high correlation sequence number.
- the spread transmitted data signal 224 is despread with the highly correlated chip sequence, which is chip sequence 222. in this example, to recover the original data signal 220 as recovered data signal 226 in Figure 26D.
- the despreadmg of the spread transmitted data signal 224 in Figure 26C can be performed in a despreading vector processing operation as an inner product between the spread transmitted data signal 224 and potential chip sequences, similar to the correlation vector processing operation described above with regard to the VPE 22(2) in Figure 1 1 , to determine a highly correlating chip sequence.
- the spread transmitted data signal 224 can be despread with the chip sequence 222. determined to have been used to CDMA modulate the original data signal 220, to provide the recovered data signal 226 in Figure 26D.
- the vector processors can include circuitry that performs despreading of spread signal vector data sequences after being output from execution units and stored in vector data memory.
- the spread signal vector data sequences stored in vector data memory are fetched from vector data memory in a post-processing operation, and despread with correlated spread code sequence or chip sequence to recover the original data signal.
- the despreaded vector data sequences which are the original data samples before spreading, are stored back in vector data memory.
- This post-processing operation can delay the subsequent vector operation processing by the execution units, and causes computational components in the execution units to be underutilized. Further, despreading of spread signal vector data sequences using a spreading code sequence is difficult to parallelize, since the spread signal vector data sequences to be despreaded cross over different data flow paths from the execution units.
- VPEs that include despreading circuitry provided in data flow paths between execution units and vector data memory in the VPE are provided.
- the despreading circuitry is configured to despread spread-spectrum sequences using an output vector data sample set from execution units in-flight while the output vector data sample set is being provided over the output data flow paths from the execution units to the vector data memory.
- In-flight despreading of output vector data sample sets means that the output vector data sample set provided by execution units is despread before being stored in vector data memory, so that the output vector data sample set is stored in vector data memory in a despread format.
- the despread spread-spectrum sequences can be stored in despread form in the vector data memory without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units.
- the efficiency of the data flow paths in the VPE may not be limited by the despreading of the spread-spectrum sequences.
- the subsequent vector processing in the execution units may only be limited by computational resources rather than by data flow limitations when despread spread-speciram sequences are stored in vector data memory.
- Figure 27 is a schematic diagram of another exemplary VPE 22(5) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(5) in Figure 27 is configured to provide in-flight despreading of resultant output vector data sample sets 228(0)-228(X) provided by the execution units 84(G)-84(X) with a code sequence for vector processing operations to be stored in the vector data files 82(0)-82(X) in the VPE 22(5) with eliminated or reduced vector data sample re-fetching and reduced power consumption.
- the resultant output vector data sample sets 228(0)-228(X) are comprised of 'X+P input resultant output vector data samples 228, which in this example are 228(0), 228(1), and 228(X).
- the code sequence could be a spread-spectrum CDMA chip sequence for a CDMA despreading vector processing operation, as a non-limiting example.
- the resultant output vector data sample set 228(0)-228(X) can be despread with a code sequence before being stored in the vector data files 82(0)-82(X).
- despreading circuitry 230 is provided in the output data flow paths 98(0)-98(X) between the execution units 84(0)- 84(X) and the vector data files 82(0)-82(X) in each of the vector data lanes 100(0)-100(X).
- the despreading circuitry 230 is configured based on programming according to a vector instruction to be executed to provide in-flight despreading of the resultant output vector data sample set 228(0)-228(X) with a code sequence provided as reference vector data sample set 130(G)- 130(X) generated by sequence number generator 134, as previously described above in Figures 1 1-16 with regard to correlation vector processing operations.
- a despread resultant output vector data sample set 229(0)-229(Z) is provided by the despreading circuitry 2.30 in the output data flow paths 98(0)-98(X).
- the despread resultant output vector data sample set 229(0)-229(Z) is comprised of ' ⁇ + ⁇ despread resultant output vector data samples 229, which in this example are 229(0), 229(1), ... , and 229(Z).
- In-flight despreading of the resultant output vector data sample set 228(0)-228(X) in the VPE 22(5) in Figure 27 means the resultant output vector data sample set 228(0)-228(X) provided by execution units 84(0)-84(X) is despread with a code sequence in the resultant vector data sample set 228(0)-228(X) before being stored in vector data files 82(0)-82(X). In this manner, the resultant output vector data sample set 228(0)-228(X) is stored in vector data files 82(0)-82(X) in despreaded form as despread resultant output vector data sample set 229(0)-229(X).
- the resultant output vector data sample set 228(0)-228(X) is not required to first be stored in the vector data files 82(0)-82(X), and then fetched from the vector data files 82(0)-82(X), despread, and restored in despreaded form in the vector data files 82(Q)-82(X).
- the resultant output vector data sample set 228(0)-228(X) is despreaded before being stored in the vector data files 82(0)-82(X).
- the despread resultant output vector data sample set 229(0)-229(Z) is stored in the vector data files 82(0)-82(X) without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units 84(0)-84(X).
- the efficiency of the data flow paths in the VPE 22(5) are not limited by the despreading of the resultant output vector data sample set 228(0)-228(X).
- the subsequent vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations when the resultant output vector data sample sets 228(0)-228(X) are stored in despreaded form as despreaded resultant output vector data sample sets 229(0)-229(Z) in the vector data files 82(0)- 82(X).
- the despreading circuitry 230 in the output data flow paths 98(0)-98(X) between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X), the resultant output vector data sample set 228(0)-228(X) does not have to cross vector data lanes 100 in the input data flow paths 80(0)-80(X) between tire vector data files 82(0)-82(X) and the execution units 84(0)-84(X).
- Providing data flow paths for despreading of input vector data samples 86 in an input vector data sample set 86(0)- 86(X) between different vector data lanes 100 would increase routing complexities.
- execution units 84(0)-84(X) may be underutilized while despreading operations are being performed in the input data flow paths 80(0)-80(X).
- despreading of the resultant output vector data sample set 228(0)-228(X) in the input data flow paths 80(0)-80(X) would require the resultant output vector data sample set 228(0)-228(X) to first be stored in the vector data files 82(0)-82(X) in the VPE 22.(5) in Figure 27, thereby increasing power consumption when re-fetched and despread and/or risking underutilization of the execution units 84(0)-84(X) that may be delayed while despreading operations are being performed.
- the despreading circuitry 230 is configured to receive the resultant output vector data sample set 228(G)- 228(X) on despreading circuitry inputs 232(0)-232(X) on the output data flow paths 98(0)-98(X).
- the despreading circuitry 230 is configured to despread the resultant output vector data sample set 228(G)-228(X) to provide the despread resultant output vector data sample set 229(0)-229(Z).
- the number of despread resultant output vector data samples 229 is ' ⁇ + ⁇ ' in the despread resultant output vector data sample set 229(0)-229(Z).
- the number of despread resultant output vector data samples 2.29 in the despread resultant output vector data sample set 229(G)- 229(Z) is dependent on the spreading factor used to despread the resultant output vector data sample set 228(0)-228(X).
- the despreading circuitry 230 is configured to provide the despread resultant output vector data sample set 229(0)-229(Z) on despreading circuitry outputs 234(0)-234(X) in the output data flow paths 98(0)-98(X) to be provided to the vector data files 82(0)-82(X) for storage.
- Figure 28 is a flowchart illustrating an exemplary despreading of resultant output vector data sample sets 228(0)- 2.28(X) resulting from a despread vector processing operation 236 that can be performed in the VPE 22(5) in Figure 27 employing the despreading circuitry 230 according to an exemplary vector instruction requiring despreading of the resultant output vector data sample set 228(0)-228(X).
- the input vector data sample set 86(0)- 86(X) to be processed according to the despread vector processing operation 236 according to a vector instruction is fetched from the vector data files 82(Q)-82(X) and provided in the input data flow paths 80(0)-80(X) (block 238 in Figure 28).
- Either one, some, or all of the vector data lanes 100(0)-100(X) in the VPE 22(5) in Figure 27 can be employed to provide the despread vector processing operation 236 according to the programming of the vector instruction depending on the width of the resultant output vector data sample set 228(0)-228(X) for the resultant despread vector processing operation 236.
- despread vector processing operation 236 involves performing despreading of ail the resultant output vector data samples 2.28 in the resultant output vector data sample set 228(0)-228(X), all vector data lanes 100(0)- 100(X) in the output data flow paths 98(0)-98(X) from the execution units 84(0)-84(X) can be employed for the despread vector processing operation 236.
- the despread vector processing operation 236 may only involve despreading a subset of resultant output vector data samples 228 in the resultant output vector data sample set 228(0)-228(X), thus only involving the vector data lanes 100 in the output data flow paths 98 corresponding to the subset of resultant output vector data samples 228.
- the fetched input vector data sample set 86(Q)-86(X) is received from the input data flow paths 80(0)-8Q(X) at the execution units 84(Q)-84(X) (block 240 in Figure 28).
- the execution units 84(0)-84(X) perform one or more vector processing operations on the received input vector data sample set 86(0)-86(X) according to the vector processing operation provided according to a vector instruction (block 242 in Figure 28).
- the execution units 84(0)-84(X) provide multiplications and/or accumulations using the input vector data sample set 86(0)-86(X) and the code sequence in the reference vector data sample set I 30(0)-130(X) for performing a vector processing operation to provide the resultant output vector data sample set 228(0)-228(X).
- the resultant output vector data sample set 228(0)-228(X) may be based on vector processing of the input vector data sample set 86(0)-86(X) with the reference vector data sample set 130(0)-130(X) is provided in the output data flow paths 98(0)-98(X) of the VPE 22(5) in Figure 27,
- the despreading vector processing operation 236 can be performed before the resultant output vector data sample set 228(0)-228(X) is stored in the vector data files 82(0)-82(X).
- the resultant output vector data sample set 228(0)-228(X) is provided to the despreading circuitry 230 provided in the output data flow paths 98(0)-98(X) provided between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X) in the VPE 22(5) in Figure 27.
- the despreading circuitry 230 is programmable to selectively despread resultant output vector data sample set 228(0)-228(X) in the output data flow paths 98(0)-98(X) according to the vector instruction being executed, and if the vector instruction calls for despreading of the resultant output vector data sample set 228(G)- 228(X) to be stored in the vector data files 82(0)-82(X).
- the despreading circuitry 230 despreads the resultant output vector data sample set 228(0)-228(X) according to the despreading programming according to the vector instruction being executed without the resultant output vector data sample set 228(0)-228(X) being stored in the vector data files 82(0)-82(X) (block 244 in Figure 28).
- the resultant output vector data sample set 228(0) -2.28 (X) does not have to first be stored in the vector data files 82(G)-82(X), re-fetched, despread in a post-processing operation, and stored in despreaded format in the vector data files 82(0)-82(X) thereby providing delay in the execution units 84(0)-84(X).
- the resultant output vector data sample set 228(0)-228(X) is stored as the despread resultant output vector data sample set 229(0)-229(Z) in the vector data files 82(0)-82(X) without despreading post-processing required (block 246 in Figure 28).
- Figure 29 is a schematic diagram of an exemplary despreading circuitry 230 that can be provided in the output data flow paths 98(0)-98(X) between the executions units 84(0)-84(X) and the vector data fifes 82(0)-82(X) in the VPE 22(5) of Figure 27.
- the despreading circuitry 230 is configured to provide despreading of the resultant output vector data sample set 228(0)-228(X) to provide the despread resultant output vector data sample set 229(0)-229(Z) for different spreading factors of repeated code sequences in the reference vector data sample set 130(0)- 130(X).
- the resultant output vector data sample set 228(0)-228(X) is provided from the execution unit outputs 96(0)- 96(X) to the despreading circuitry 230, as illustrated in Figure 27. It may be desired to despread the resultant output vector data sample set 228(0)-228(X) with different spreading factors of repeating sequence numbers in the reference vector data sample set 130(0)-130(X) generated by the sequence number generator 134 in Figure 27, because the spreading factor of the resultant output vector data sample set 228(0)-228(X) may be unknown.
- the despread resultant output vector data sample set 229(0)-229(Z) would contain eight (8) despread samples (i.e.. 32 samples / spreading factor of 4) after despreading of the resultant output vector data sample set 228(0)-228(X) is performed.
- the despread resultant output vector data sample set 229(0)-229(Z) would contain four (4) despread samples (i.e., 32 samples / spreading factor of 8) after despreading of the resultant output vector data sample set 228(0)-228(X) is performed.
- the despreading circuitry 230 is configured to despread the resultant output vector data sample set 228(0)-228(X) for a different number of spreading factors.
- the despreading circuitry 230 in this embodiment is configured to provide despread resultant output vector data sample set 229(G)-229(Z) for different spreading factors in one vector processing operation/one vector instruction.
- the despreading circuitry 2.30 contains an adder tree 248 coupled to the execution unit outputs 96(0)-96(X) to receive the resultant output vector data sample set 228(0)-228(X).
- the adder tree 248 of the despreading circuitry 230 is configured to receive each sample 228 of resultant output vector data sample set 228(0)-228(X) in their respective vector data lanes 10G(0)-J 00(X).
- a first adder tree level 248(1) is provided in the adder tree 248.
- the first adder tree level 248(1 ) is comprised of adders 250(0)-250(((X+I)*2)-l ), 250(7) to be able to spread the samples 228 in the resultant output vector data sample set 228(0)-228(X) by a spreading factor of four (4).
- Latches 251(0)-251(X) are provided in the despreading circuitry 230 to latch the resultant output vector data sample set 228(0)-228(X) from the output data flow paths 98(0)-98(X).
- each sample 228 in the resultant output vector data sample set 228(0)-228(X) is 32 bits wide and comprised of two (2) 16-bit comple vector data (i.e., first vector data according to format I8Q8 and second vector data according to format I8Q8).
- a spreading factor of four (4) could be applied to despread the four (4) vector data samples in two (2) resultant output vector data samples 228 in the resultant output vector data sample set 228(0)-228( ' X) into one despread resultant output vector data sample.
- adder 250(0) is configured to despread resultant output vector data samples 228(0) and 2.28(1 ) by a spreading factor of four (4) for those samples.
- adder 250(1) is configured to despread resultant output vector data samples 228(2) and 228(3) by a spreading factor of four (4) for those samples.
- Adder 250(((X+l)/2)-l), 250(7) is configured to despread resultant output vector data sample set 228(X-1) and 228(X) to provide a despread vector data sample set 252(0)-252(((X+l )/2)- l ), 252(7) with a spreading factor of four (4), A despread vector data sample set 252(0)-252(((X+ 1 )/2)- 1 ), 252(7) from despreading performed by the adders 250(((X+l)/2)-l), 250(7) is latched into latches 255(0)-255(((X+l) 2)-l), 255(7).
- the despread vector processing operation 236 requires a despreading of the resultant output vector data sample set 228(0)-228(X) by a spreading factor of four (4), as will be discussed in more detail below, the despread vector data sample set 252(0)- 252(((X+1 )/2) ⁇ 1 ), 252(7) can be provided as the despread resultant output vector data sample set 229(0)-229(Z), wherein ' ⁇ ' is seven (7).
- the despread vector processing operation 236 calls for a higher spreading factor (e.g...
- the despread vector data sample set 252(0)-252((( ⁇ +1)/2)- 1), 252(7) is not provided as despread resultant output vector data sample set 229(0)-229(Z)
- the despread vector data sample set 252(0)-252(((X+l)/2)- l), 252(7) is provided to a second adder tree level 248(2) to adders 254(0)-254(((X+l)/4)-l), 254(3).
- adder 254(0) is configured to performing despreading on despread vector data samples 252(0) and 252(1 ) to provide a resultant despread vector data sample 256(0) with a spreading factor of eight (8) for those samples.
- adder 254(1) is configured to perform despreading on despread vector data samples 252(2) and 252(3) to provide a resultant despread vector data sample 256(1) having a spreading factor of eight (8) for those samples.
- Adder 254((X+l)/4)-l ), 254(3) is configured to perform despreading on despread vector data sample set 252(((X+l )/4)-2), 252(((X+l)/4)-l), 252(3) to provide a resultant despread vector data sample 256(((X+l)/4)-l), 256(3) with a spreading factor of eight (8).
- the resultant despread vector data sample set 256(0)-256(((X ⁇ l)/4)- l), 256(3) from despreading performed by the adders 254(0)-254(((X ⁇ l )/4) ⁇ l ), 254(3) is latched into latches 257(0)-257(((X+I)/4)- 1), 257(3).
- the despread vector processing operation 236 requires a despreading of the resultant output vector data sample set 228(G)-228(X) by a spreading factor of eight (8), as will be discussed in more detail below, the despread vector data sample set 256(0)-256(((X+ 1 )/4)- 1 ), 256(3) can be provided as the despread resultant output vector data sample set 229(0)-229(Z), wherein 'Z' is three (3).
- the despread vector processing operation 236 calls for a higher spreading factor than eight (8) (e.g., 16, 32, 64, 128, 256)
- the despread vector data sample set 256(0)-256(((X+l)/4)- l), 256(3) is not provided as despread resultant output vector data sample set 229(0)-229(Z).
- the despread vector data sample set 256(0)-256(((X+l)/4)- 1), 256(3) is provided to a third adder tree level 248(3) to adders 258(G)-258(((X+l)/8)- l), 258(1 ).
- adder 258(0) is configured to performing despreading on despread vector data samples 256(0) and 256(1) to provide a spreading factor of sixteen (16) for those samples.
- adder 258(1) is configured to perform despreading on despread vector data samples 256(2) and 256(3) to provide a despread vector data sample set 260(0)-260(((X+ 1 )/8) - 1 ), 260(1) with spreading factor of sixteen (16).
- the despread vector data sample set 260(0)-260(((X+l)/8)-l), 260(1) from despreading performed by the adders 258(0)-258(((X+l)/8)-l), 258(1 ) is latched into latches 259(0)-259(((X+l )/8)-l ), 259(2).
- the despread vector processing operation 236 requires a despreading of the resultant output vector data sample set 228(0)-228(X) by a spreading factor of sixteen (16), as will be discussed in more detail below, the despread vector data sample set 260(0)-260(((X+l)/8)- 1), 256(1) can be provided as the despread resultant output vector data sample set 229(0)-229(Z), wherein 'Z' is one (1).
- the despread vector data sample set 26Q(Q)-260(((X T )/8)- l ), 260(1) is not provided as despread resultant output vector data sample set 229(0)-229(Z).
- the despread vector data sample set 260(0)-260(((X+l)/8)- l), 260(1 ) is provided to a fourth adder tree level 248(4) to adder 262.
- the adder 262 is configured to performing despreading on despread vector data samples 260(0) and 260(1) to provide a despread vector data sample 264 with a spreading factor of thirty -two (32),
- the despread vector data sample 264 from despreading performed by the adder 262 is latched into latches 266 and 268.
- the despread vector data sample 264 can be provided as a despread resultant output vector data sample 229.
- the despread vector processing operation 236 calls for a higher spreading factor than thirty -two (32) (e.g., 64, 128, 256)
- the despread vector data sample 264 is not provided as a despread resultant output vector data sample set 229.
- the despread vector data sample 264 remains latched into latch 268 without having to be stored in a vector data file 82.
- Another resultant output vector data sample set 228(0)-228(X) is loaded into latches 251(0)-251(X) over additional processing cycles to be despread using a spreading factor of thirty-two (32), as described above.
- the resulting despread vector data sample 264' is added by adder 270 in a fifth adder tree 248(5) to the previous despread yector data sample 264 to provide a despread vector data sample 272 having a spreading factor of sixty-four (64).
- Selector 273 controls whether the despread vector data sample 264 having a spread factor of thirty-two (32) or the despread vector data sample 264' ha ving a spread factor of sixty-four (64) is latched as despread vector data sample 272 is latched into latch 274. This same process of latching additional resultant output vector data sample sets 228(0)- 228(X) and despreading of same can be performed to achieve spread factors greater than sixty-four (64), if desired. The despread vector data sample 272 will eventually be latched into latch 274 as the desired despread resultant output vector data sample 229 according to the desired spreading factor for the despread vector processing operation 236.
- the despread resultant output vector data sample set 229(0)-229(Z) will need to be stored in the vector data files 82(0)- 82(X) in Figure 27.
- the despreading circuitry 230 in Figure 29 is also configured to load the despread resultant output vector data samples 229 provided as a result of performing the despreading vector processing operation 236 on resultant output vector data samples 228(0)-228(X) into latches 276(0)-276(X), to form the despread resultant output vector data sample set 229(G)- 229(Z).
- the despread resultant output vector data sample set 229(0)-229(Z) can be provided to the vector data files 82(0)-82(X) to be stored, in this manner, only one (1) write is required to the vector data files 82(0)-82(X) to store the despread resultant output vector data sample set 229(G)-229(Z) created by the despreading circuitry 230.
- the adder trees 248( 1)- 248(5) in the despreading circuitry 230 in Figure 29 can generate despread resultant output vector data samples 229 for all of spreading factors 4, 8.
- adders in adder trees not necessary to perform the despread vector processing operation 236 according to the desired spreading factor can be disabled or configured to add 0's.
- selectors 278(0)-278(((X+l)/4)-l), 278(3) are provided, as will now be discussed.
- selector 278(0) can select despread resultant output vector data samples 229 for any of spreading factors 4, 8, and 16 from adders 250(0), 254(0), 258(0), respectively, and spreading factors 32, 64, 128, 256 from adders 262, 270 based on the despread vector processing operation 236 being executed.
- Selector 278(1) can select despread resultant output vector data samples 229 for spreading factors 4, 8, and 16 from adders 250(1 ), 254(1), and 258(1) respectively, based on the despread vector processing operation 236 being executed.
- Selector 278(2) can select despread resultant output vector data samples 229 for spreading factors 4 and 8 from adders 250(2) and 254(2), respectively, based on the despread vector processing operation 236 being executed.
- Selector 278(3) can select despread resultant output vector data samples 229 for spreading factors 4 and 8 from adders 250(3) and 254(3), respectively, based on the despread vector processing operation 2.36 being executed.
- Selector 278(4) can select despread resultant output vector data samples 229 for spreading factors 4 and 8 from adder trees 248(1) and 248(2), respectively, based on the despread vector processing operation 236 being executed.
- Selectors are not provided to control the despread resultant output vector data samples 229 provided from adders 250(4)-250(7), because providing a spreading factor of eight (8) can be fully satisfied by selectors 278(0)-278(3).
- a series of data slicers 280(G)- 280(((X+l)/2)-l), 280(7) are provided to receive the despread resultant output vector data samples 229 selected by the selectors 278(0)-278(((X+l) 4)-l), 278(3) and adders 250(4)-250(((X+l)/2)-l), 250(7), respectively.
- Data slicers 280(0)-280(((X+l)/2)-l), 280(7) are configured to select whether its received despread resultant output vector data samples 229 will be characterized as a logical high level (e.g., a logical T) or a logical low level (e.g., a logical ⁇ ').
- the despread resultant output vector data samples 229 are then routed through connections to a crossbar 282 to the desired latch 276 among latches 276(0)-276(X) to be stored.
- the crossbar 282 provides for the flexibility to provide the despread resultant output vector data samples 229 according to the despread vector processing operation 236 to different latches 276(0)-276(X).
- despread resultant output vector data samples 229 can be stacked in latches 276(0)-276(X) among different iterations of despread vector processing operations 236 before being stored in the vector data files 82(0)-82(X).
- a despread resultant output vector data sample set 229(0)-229(Z) can be stacked in latches 276(G)- 276(X) among different iterations of despread vector processing operations 236 before being stored in the vector data files 82(0)-82(X).
- accesses to the vector data files 82(0)-82(X) to store despread resultant output vector data sample set 229(0)- 229(Z) can be minimized for operating efficiency.
- selectors 284(0)-284(X) coupled to crossbar 282 can be controlled to store the despread resultant output vector data sample 229 from data sheer 280(0) in any of latches 276(0)-276(X),
- Selectors 284(1), 284(3), 284(5), 284(7), 284(9), 284(1 1), 284(13), 284(15) coupled to crossbar 282 can be controlled to store despread resultant output vector data sample 229 from data slicers 280(1) to be stored in latches 276(1), 276(3), 276(5), 276(7), 276(9), 276(11 ), 276( 13), and 276(15).
- Selectors 284(2), 284(6), 284(10), 284(14) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(2) in latches 276(2), 276(6), 276(10), and 276(14).
- Selectors 284(3), 284(7), 284(1 1), 284(35) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(3) in latches 276(3), 276(7), 276(1 1 ), and 276( 15).
- Selectors 284(4) and 284(12) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(4) in latches 276(4) and 276(12).
- Selectors 284(5) and 284(13) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(5) to be stored in latches 276(5) and 276(13).
- Selectors 284(6) and 284(14) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(6) in latch 276(6) or 276(14).
- Selectors 284(7) and 284(15) coupled to crossbar 282 can be controlled to store a despread resultant output vector data sample 229 from data slicer 280(7) in latches 276(7) or 276(15).
- the despreading circuitry 230 can be programmed to be configured to perform or not perform despreading operations on the resultant output vector data samples 228(Q)-228(X) according to the vector instruction to be executed.
- a despread configuration input 286 in Figure 29 can be provided to the despreading circuitry 230 to perform despread operations on the resultant output vector data samples 228(0)-228(X) or to simply provide the resultant output vector data samples 228(0)-228(X) to the latches 276(0)-276(X), respectively, to be stored in tire vector data files 82(0)-82(X).
- the despreading circuitry 230 can be programmed to not despread the resultant output vector data sample sets 228(0)-228(X) if the vector instruction does not provide for such processing to be performed.
- the despread configuration input 284 can be configured and reconfigured for each vector instruction to provide for flexibility in the vector processing by the VPE 22(5) in Figure 27.
- the despread configuration input 284 can be configured and reconfigured for each clock- cycle of a vector instruction, on a clock- cycle-by-clock-cycle basis, if desired, to provide despreading as desired, with full utilization of the execution units 84(0)-84(X), if desired.
- Certain other wireless baseband operations require merging of data samples determined from previous processing operations for reasons other than despreading of spread spectrum data sequences. For example, it may be desired to accumulate vector data samples of varying widths that are wider than the data flow paths for the execution units 84(0) -84(X) provided by the vector data lanes 100(0)-100(X). As another example, it may be desired to provide a dot product multiplication of output vector data samples from different execution units 84(0)-84(X) to provide merging of output vector data in vector processing operations.
- the vector data lanes 100(0)-100(X) in the VPE could include complex routing to provide intravector data paths for crossing over vector data lanes 100(0)-100(X) to provide merged vector processing operations.
- Vector processors could include circuitry that performs postprocessing merging of output vector data stored in vector data memory from execution units.
- the post-processed output vector data samples stored in vector data memory are fetched from vector data memory, merged as desired, and stored back in vector data memory.
- this post-processing can delay the subsequent vector processing operations of the VPE, and cause computational components in the execution units to be underutilized.
- FIG. 30 For example, a two input vector data samples 290(0), 290(1) provided in vector data files 82(0), 82(1 ) in a VPE previously described are shown in Figure 30. it may be desired to add these two input vector data samples 290(0), 290(1) together, in this example, the sum of the two input vector data samples 290(0), 290(1) is ⁇ 1250314E,' which has a data width larger than either vector data lane 100(0) or 100(1).
- Data flow paths could be provided in the VPE 22 to provide vector data routing between vector data lanes 100(0), 100(1 ) to allow execution units 84(0), 84(1) to perform the execution of the sum of the two input vector data samples 290(0), 290(1) together, including providing carry logic between the two execution units 84(0), 84(1 ) across the vector data lanes 100(0), 100(1).
- the ability to cross all vector data lanes 100(0)-100(X) may be required to provide a scalar result of merged vector data samples, which may further increase complexity in data flow paths. However, as discussed above, this would add complexity in data flow paths thereby increasing complexity, and possibly reducing efficiency.
- VPEs that include merging circuitry provided in output data flow paths between execution units and vector data memory in a VPE.
- the merging circuitry is configured to merge output vector data samples from an output vector data sample set provided by the execution units in-flight while the output vector data sample set is being provided over the output data flow paths from the execution units to the vector data memory.
- In-flight merging of output vector data samples means the output vector data samples provided by execution units can be merged before being stored in vector data memory, so that the resulting output vector data sample set is stored in vector data memory in a merged format.
- the merged output vector data samples can be stored in vector data files without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units.
- the efficiency of the data flow paths in the VPE is not limited by vector data merging operations.
- the subsequent vector processing in the execution units is only limited by computational resources rather than by data flow limitations when merged vector data samples are stored in vector data memory .
- Figure 31 is a schematic diagram of another exemplary VPE 22(6) that can be provided as the VPE 22 in Figure 2.
- the VPE 22(6) in Figure 31 is configured to provide in-flight merging of resultant output vector data sample sets 292(0)-292(X) provided by the execution units 84(0)-84(X) with a code sequence for vector processing operations to be stored in the vector data files 82(0)-82(X) in the VPE 22(6) with eliminated or reduced vector data sample re-fetching and reduced power consumption.
- the resultant output vector data sample set 292(0)-292(X) is comprised of resultant output vector data samples 292(0), ... , 292(X).
- a merge vector processing operation could include adding resultant output vector data samples 292, determining a maximum vector data sample value among a plurality of resultant output vector data samples 292, or determining a minimum vector data sample value among a plurality of output vector data samples 292.
- the resultant output vector data samples 292 among the resultant output vector data sample set 292(0)-292(X) can be merged before being stored in the vector data files 82(0)-82(X).
- the merging circuitry 294 is configured based on programming according to a vector instruction to be executed to provide in-flight merging of the resultant output vector data samples 228 among the resultant output vector data sample set 228(0)- 228(X). Merged resultant output vector data samples 296(0)-296(Z) are pro vided by the merging circuitry 294 in the output data flow paths 98(0)-98(X).
- the merged resultant output vector data sample set 296(G)- 296(Z) is comprised of resultant output vector data samples 296, which are 296(0), and 296(Z) in this example.
- the number of merged resultant output vector data samples 296 in the merged resultant output vector data sample set 296(0)-296(Z) is dependent on the merging operations performed on the resultant output vector data sample set 292(0)- 292(X).
- In-flight merging of the resultant output vector data samples 292 in the VPE 22(6) in Figure 31 means that resultant output vector data samples 292. in the resultant output vector data sample set 292(0)-292 (X) provided by execution units 84(0)-84(X) can be merged together before being stored in vector data files 82(0)-82(X). In this manner, the merged resultant output vector data samples 296 of the merged resultant output vector data sample set 296(0)-296(Z) can be stored in vector data files 82(G)- 82(X) in merged form as merged resultant output vector data sample set 296(0)-296(Z).
- the resultant output vector data sample set 292(0)-292(X) is not required to first be stored in the vector data files 82(0)-82(X), and then fetched from the vector data files 82(0)-82(X), Desired resultant output vector data samples 292 are merged, and the resultant output vector data samples 292 are restored in merged form in the vector data files 82(0)-82(X).
- Resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X) can be merged before being stored in the vector data files 82(0)-82(X).
- merged resultant output vector data samples 296 from the merged resultant output vector data sample set 296(0)-296(Z) are stored in the vector data files 82(0)-82(X) without requiring additional post-processing steps, which may delay subsequent vector processing operations to be performed in the execution units 84(0)-84(X).
- the efficiency of the data flow paths in the VPE 22(6) are not limited by the merging of the resultant output vector data samples 292.
- the subsequent vector processing in the execution units 84(0)-84(X) is only limited by computational resources rather than by data flow limitations when the resultant output vector data samples 292 are stored in merged form in the vector data files 82(0)- 82(X).
- the merging circuitry 294 in the output data flow paths 98(0)-98(X) between the execution units 84(0)-84(X) and the vector data files 82(G)- 82(X), the resultant output vector data sample set 292(0)-292(X) does not have to cross vector data lanes 100 in the input data flow paths 80(0)-80(X) between the vector data files 82(0)-82(X) and the execution units 84(0)-84(X).
- Providing data flow paths for merging of input vector data samples 86 in an input vector data sample set 86(0)-86(X) between different vector data lanes 100 would increase routing complexities.
- execution units 84(0)-84(X) may be underutilized while merging operations are being performed in the input data flow paths 80(0)-80(X).
- merging of resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X) in the input data flow paths 80(0)-80(X) would require the resultant output vector data sample set 292(0)-292(X) to first be stored in the vector data files 82(G ⁇ -82(X) in the VPE 22(6) in Figure 31, thereby increasing power consumption when re-fetched and merged and/or risking underutilization of the execution units 84(0)-84(X) that may be delayed while merging operations are being performed.
- the merging circuitry 294 is configured to receive the resultant output vector data sample set 292(0)- 292(X) on merging circuitry inputs 30G(0)-300(X) on the output data flow paths 98(G)- 98(X).
- the merging circuitry 294 is configured to merge desired resultant output vector data samples 292 from the resultant output vector data sample set 292(G)-292(X) to provide the merged resultant output vector data sample set 296(0)-296(Z).
- 'Z' in the merged resuliani output vector data sample set 296(0)-296(Z) represents the bit width of the merged resultant output vector data sample set 296(0)-296(Z).
- 'Z' may be less than the bit width of resultant output vector data sample set 292(G) -292(X), represented by ' ⁇ ,' due to merging operations.
- the number of merged resultant output vector data samples 296 ' ⁇ + ⁇ in the merged resultant output vector data sample set 296(0)-296(Z) is dependent on the resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X) to be merged together.
- the merging circuitry 294 is configured to provide the merged resuliani ouiput vector data sample set 296(0)-296(Z) on merging circuitry outputs 301(0)-301(X) in the output data flow paths 98(0)-98(X) to be provided to the vector data files 82(0)-82(X) for storage.
- Figure 32 is a flowchart illustrating an exemplary merging of resultant output vector data samples 292 of the resultant output vector data sample sets 292(0)-292(X) resulting from a vector processing operation 302 that can be performed in the VPE 22(6) in Figure 31 employing the merging circuitry 294 according to an exemplary vector instruction requiring merging of the resultant output vector data samples 292.
- the input vector data sample set 86(0)- 86(X) to he processed according to a vector processing operation 302 according to a vector instruction is fetched from the vector data files 82(0)-82(X) and provided in the input data flow paths 80(0)-80(X) (block 304 in Figure 32).
- Either one, some, or all of the vector data lanes 100(0)-100(X) in the VPE 22(6) in Figure 31 can be employed to provide the vector processing operation 302 according to the programming of the vector instruction depending on the width of the input vector data sample set 86(0)-86(X) for the vector processing operation 302.
- ail vector data lanes 100(0)--1Q0(X) can be employed for the vector processing operation 302.
- the vector processing operation 302 may only require a subset of the vector data lanes 100(0)-100(X). This may be because the width of the input vector data sample set 86(0)-86(X) is less than the width of all vector data files 82(0)-82(X), where it is desired to employ the additional vector data lanes 100 for other vector processing operations to be performed in parallel to the vector processing operation 302.
- the fetched input vector data sample set 86(0)-86(X) is received from the input data flow paths 80(0)-80(X) at the execution units 84(0)-84(X) (block 306 in Figure 32).
- the execution units 84(0)-84(X) perform the vector processing operation 302 on the received input vector data sample set 86(0)- 86(X) according to the vector processing operation 302 provided according to a vector instruction (block 308 in Figure 32).
- the execution units 84(0)-84(X) can provide multiplications and/or accumulation using the input vector data sample set 86(0)--86(X) for the vector processing operation 302 to provide the resultant output vector data sample set 292(0)-292(X).
- a resultant output vector data sample set 292( ' 0)-292(X) based on the vector processing operation 302 carried out on the input vector data sample set 86(0)- 86(X) is provided in the output data flo paths 98(0)-98(X) of the VPE 22(6) in Figure 31.
- the resultant output vector data sample set 292(0)-292(X) is stored in the vector data files 82(0)-82(X)
- the resultant output vector data sample set 292(0)-292(X) is provided to the merging circuitry 294 provided in the output data flow paths 98(0)-98(X) provided between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X).
- the merging circuitry 294 is programmable to be included in the output data flow paths 98(0)-98(X) according to the vector instruction being executed, and if the vector instruction calls for merging of the resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X) to be stored in the vector data files 82(0)-82(X), as discussed below in more detail.
- the merging circuitry 294 merges the resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X) according to the vector instruction being executed without the resultant output vector data sample set 292(0)-292(X) being stored in the vector data files 82(G)-82(X) (block 310 in Figure 32).
- the resultant output vector data sample set 292.(0) - 292(X) does not have to first be stored in the vector data files 82(0)-82(X), re-fetched, merged in a post-processing operation, and stored in merged format in the vector data files 82(0)-82(X), thereby providing delay in the execution units 84(0)-84(X).
- the resultant output vector data sample set 292(0)-292(X) is stored as the merged resultant output vector data sample set 296(0)-296(Z) in the vector data files 82(0)-82(X) without merge post-processing required (block 312 in Figure 32).
- Figure 33 is a schematic diagram of an exemplary merging circuitry 294 that can be provided in the output data flow paths 98(0)-98(X) between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X) in the VPE 22(6) of Figure 31.
- the merging circuitry 294 is configured to provide merging of the resultant output vector data sample set 292(0)-292(X) to provide the merged resultant output vector data sample set 296(0)-296(Z).
- the resultant output vector data sample set 292(0)-292(X) is provided from the execution unit outputs 96(0)-96(X) to the merging circuitry 294, as illustrated in Figure 31.
- the merging circuitry 294 is configured to merge the resultant output vector data sample set 292(0)-292(X).
- the merging circuitry 294 in this embodiment is configured to provide a merged resultant output vector data sample set 296(0)-296(Z).
- the merging circuitry 294 contains an adder tree 318 coupled to the execution unit outputs 96(0)-96(X) to receive the resultant output vector data sample set 292(0)-292(X).
- the adder tree 318 of the merging circuitry 294 is configured to receive each sample 292 of resultant output vector data sample set 292(0)-292(X) in their respective vector data lanes 100(0)- 100(X).
- a first adder tree level 318(1) is provided in the adder tree 31 8.
- the first adder tree level 318(1 ) is comprised of merge circuits 320(0)-320(((X+l)/2)-l), 320(7) to he able to merge adjacent samples 292 in the resultant output vector data sample set 292(0)-292(X).
- Latches 321(0)-321(X) are provided in the merging circuitry 294 to latch the resultant output vector data sample set 292(0)-292(X) from the output data flow paths 98(0)-98(X).
- each sample 292 in the resultant output vector data sample set 292(0)-292(X) is 32 bits wide and comprised of two (2) 16-bit comple vector data (i.e., first vector data according to format I8Q8 and second vector data according to format I8Q8)
- a merging operation could be applied to merge the four (4) vector data samples in two (2.) resultant output vector data samples 292 in the resultant output vector data sample set 292(0)-292(X) into one merged resultant output vector data sample 296.
- adder 320(0) is configured to merge resultant output vector data samples 292(0) and 292(1).
- adder 320(1) is configured to merge resultant output vector data samples 292(2) and 292(3) for those samples.
- Adder 320(((X+l)/2)-l), 320(7) is configured to merge resultant output vector data sample set 292(X-i) and 292(X) to provide a merge vector data sample set 322(0)- 322(((X+l )/2 ⁇ - l ), 322(7).
- the merge vector processing operation 302 requires a merging of the resultant output vector data sample set 292(0)-292(X), as will be discussed in more detail below, the merge vector data sample set 322(0)-322(((X+l)/2)-l), 322(7) can be provided as the merge resultant output vector data sample set 296(0)-296(Z), wherein 'Z' is seven (7).
- the merge vector data sample set 322(0)- 322(((X+ l)/2)- l), 322(7) is not provided as merge resultant output vector data sample set 296(0)-296(Z).
- the merge vector data sample set 322(0)-322(((X+l)/2)- l), 322(7) is provided to a second adder tree level 318(2) to adders 324(0)-324(((X ⁇ l)/4)- l), 324(3).
- adder 324(0) is configured to perform merging on merge vector data samples 322(0) and 322(1) to provide a resultant merge vector data sample 326(0).
- adder 324(1) is configured to perform merging on merge vector data samples 322(2) and 322(3) to provide a resultant merge vector data sample 326(1).
- Adder 324((X+l)/4)- l), 324(3) is configured to perform merging on merge vector data sample 322(((X+l )/4)-2), 322(((X-;- l )/4)- l ), 322(3) to provide a resultant merge vector data sample 326(((X+l)/4)- l), 326(3).
- the merge vector data sample set 326(0)-326(((X-'-l)/4)-l), 326(3) can be provided as the merge resultant output vector data sample set 296(0)-296(Z), wherein 'Z' is three (3).
- the merge vector data sample set 326(0)- 326(((X+ l)/4)- l), 326(3) is not provided as merge resultant output vector data sample set 296(0)-296(Z)
- the merge vector data sample set 326(0)-326(((X+l)/4)-l), 326(3) is provided to a third adder tree level 31 8(3) to adders 328(0)-328(((X+l)/8)- l), 328(1).
- adder 328(0) is configured to performing merging on merge vector data samples 326(0) and 326(1 ) to pro vide a merge factor of sixteen (16) for those samples.
- adder 328(1) is configured to perform merging on merge vector data samples 326(2) and 326(3) to provide a merge vector data sample set 330(0)-330(((X+l )/8)-l ), 330(1 ) with merge factor of sixteen (16).
- the merge vector data sample set 330(0)-330(((X+l)/8)-l), 330(1) can be provided as the merge resultant output vector data sample set 296(0)-296(Z), wherein 'Z' is one (1).
- the merge vector data sample set 330(0)- 330((( ' X+l)/8)- l), 330(1) is not provided as merge resultant output vector data sample set 296(0)-296(Z).
- the merge vector data sample set 330(0)-330(((X+i)/8)- 1), 330(1) is provided to a fourth adder tree level 318(4) to adder 332.
- the adder 332 is configured to perform merging on merge vector data samples 330(0) and 330(1 ) to provide a merge vector data sample 334 with a merge factor of thirty- two (32).
- the merge vector data sample 334 from merging performed by the adder 332 is latched into latches 336 and 338.
- the merge vector data sample 334 can be provided as a merge resultant output vector data sample 296.
- the merge vector processing operation 302 calls for a higher merge factor than thirty-two (32) (e.g., 64, 128, 256)
- the merge vector data sample 334 is not provided as a merge resultant output vector data sample set 296.
- the merge vector data sample 334 remains latched into latch 338 without haying to be stored in a vector data file 82.
- Another resultant output vector data sample set 292(0)- 292(X) is loaded into latches 321 (0)-321 (X) over additional processing cycles to be merged using a merge factor of thirty-two (32), as described above.
- the resulting merge vector data sample 334' is added by adder 340 in a fifth adder tree 318(5) to the previous merge vector data sample 334 to provide a merge vector data sample 342 having a merge factor of sixty-four (64).
- Selector 343 controls whether the merge vector data sample 334 haying a merge factor of thirty-two (32) or the merge vector data sample 334' having a merge factor of sixty-four (64) is latched as merge vector data sample 342 is latched into latch 344.
- This same process of latching additional resultant output vector data sample sets 292(0) -292(X) and merging of same can be performed to achieve merge factors greater than sixty-four (64), if desired.
- the merge vector data sample 342 will eventually be latched into latch 344 as the desired merge resultant output vector data sample 296 according to the desired merge factor for the merge vector processing operation 302.
- the merging circuitry 294 in Figure 33 is also configured to load the merge resultant output vector data samples 296 pro vided as a result of performing the merging vector processing operation 302 on resultant output vector data samples 292(0)-292(X) into latches 346(0)-346(X), to form the merge resultant output vector data sample set 296(0)-296(Z).
- the merge resultant output vector data sample set 296(0)-296(Z) can be provided to the vector data files 82(0)-82(X) to be stored. In this manner, only one (I) write is required to the vector data files 82(0)- 82(X) to store the merge resultant output vector data sample set 296(0)- 296(Z) created by the merging circuitry 294.
- the adder trees 318(1 )-318(5) in the merging circuitry 294 in Figure 33 can generate merge resultant output vector data samples 296 for all of merge factors 4, 8, 16, and 32 regardless of what merge factor is called for in the merge vector processing operation 302. Alternatively, adders in adder trees not necessary to perform the merge vector processing operation 302 according to the desired merge factor can be disabled or configured to add 0's.
- selectors 348(0)-348(((X+l )/4)-l ), 348(3) are provided, as will now be discussed.
- selector 348(0) can select merge resultant output vector data samples 296 for any of merge factors 4, 8, and 16 from adders 320(0), 324(0), 328(0), respectively, and merge factors 32, 64, 128, 256 from adders 332, 340 based on the merge vector processing operation 302 being executed.
- Selector 348(1 ) can select merge resultant output vector data samples 296 for merge factors 4, 8, and 16 from adders 320(1 ), 324(1), and 328(1) respectively, based on the merge vector processing operation 302 being executed.
- Selector 348(2) can select merge resultant output vector data samples 296 for merge factors 4 and 8 from adders 320(2) and 324(2), respectively, based on the merge vector processing operation 302 being executed.
- Selector 348(3) can select merge resultant output vector data samples 296 for merge factors 4 and 8 from adders 320(3) and 324(3), respectively, based on the merge vector processing operation 302 being executed.
- Selectors are not provided to control the merge resultant output vector data samples 296 provided from adders 320(4)-320(7), because providing a merge factor of eight (8) can be fully satisfied by selectors 348(0)-348(3).
- the data slicers 350(0)- 350(((X+l)/2)- l), 350(7) provided for merge vector processing operations could be bypassed or configured to not perform data splicing on the received merge resultant output vector data samples 296 selected by the selectors 348(0)-348(((X+l)/4)-l), 348(3) and adders 32G(4)-320(((X+l)/2)- l ), 320(7), respectively.
- the merge resultant output vector data samples 296 are then routed through connections to a crossbar 352 to the desired latch 346 among latches 346(0)-346(X) to be stored.
- the crossbar 352 provides for the flexibility to provide the merge resultant output vector data samples 296 according to the merge vector processing operation 302 to different latches 346(0)- 346(X).
- merge resultant output vector data samples 296 can be stacked in latches 346(G)-346(X) among different iterations of merge vector processing operations 302 before being stored in the vector data files 82(0)-82(X).
- a merge resultant output vector data sample set 296(0)-296(Z) can be stacked in latches 346(0)-346(X) among different iterations of merge vector processing operations 302 before being stored in the vector data files 82(0)-82(X).
- accesses to the vector data files 82(0)-82(X) to store merge resultant output vector data sample set 296(0)-296(Z) can be minimized for operating efficiency.
- selectors 354(0)-354(X) coupled to crossbar 352 can be controlled to store the merge resultant output vector data sample 296 from selector 348(0) in any of latches 346(0)-346(X).
- Selectors 354(1), 354(3), 354(5), 354(7), 354(9), 354( 1 1 ), 354(13), 354(15) coupled to crossbar 352 can be controlled to store merge resultant output vector data sample 296 from selector 348(1) to be stored in latches 346(1), 346(3), 346(5), 346(7), 346(9), 346(1 1), 346( 13), and 346(15).
- Selectors 354(2), 354(6), 354(10), 354(14) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from selector 348(2) in latches 346(2), 346(6), 346(10), and 346(14).
- Selectors 354(3), 354(7), 354(11), 354(15) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from selector 348(3) in latches 346(3), 346(7), 346(1 1), and 346(15).
- Selectors 354(4) and 354(12) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from adder 320(4) in latches 346(4) and 346(12).
- Selectors 354(5) and 354(13) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from adder 320(5) to be stored in latches 346(5) and 346( 13).
- Selectors 354(6) and 354(14) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from adder 320(6) in latch 346(6) or 346(14).
- Selectors 354(7) and 354(15) coupled to crossbar 352 can be controlled to store a merge resultant output vector data sample 296 from adder 320(7) in latches 346(7) or 346(15).
- the adders could be configured to allow non-adjacent resultant output vector data samples 282 in the resultant output vector data sample set 292(0)-292(X) to be merged.
- adders in adder tree levels 318(1)-318(3) could be configured to simply pass merge resultant output vector data samples 292(0) with resultant output vector data samples 292(9) to adder tree level 318(4).
- the adder 332 in adder tree level 318(4) could then merge resultant output vector data sample 292(0) with resultant output vector data samples 292(9) to provide merged output vector data samples 296.
- Merging circuitry could also be provided in the output data flow paths 98(G)- 98(X) between the execution units 84(0)-84(X) and the vector data files 82(0)-82(X) that provide other types of vector merging operations other than vector and/or scalar adding.
- the merging circuitry 294 in Figure 33 could be configured to provide maximum or minimum vector and/or scalar merging operations.
- the adders in the adder tree levels 31 8(l)-318(5) in the adder tree 318 in Figure 33 could be replaced with maximum or minimum function circuitry, in other words, the circuitry would select to pass either the larger or the lesser of two resultant output vector data samples 292 from the resultant output vector data sample set 292(0)-292(X).
- the merging circuitry 294 could be configured to select vector data samples 290(1) if the merging circuitry 294 is configured to select the maximum vector data sample.
- adders 320(0)-320(((X+l)/2)-l), 320(7) in the first adder tree level 318(1) in Figure 33 could be replaced with maximum or minimum merge selectors adders 320'(0)-320'(((X+l)/2)- 1), 320'(7), as illustrated in Figure 34.
- Adders 324(0)-324(((X+l)/4)- l), 324(3) in the second adder tree level 318(2) could be replaced with maximum or minimum selectors 324'(0)-324'(((X+i)/4)- i), 324'(3), as illustrated in Figure 34.
- Adders 328(0)-328(((X+l)/8)-l), 328( 1) in the third adder tree level 318(3) could be replaced with maximum or minimum selectors 328'(0)-328'(((X+l)/8)-l), 328'(1), as illustrated in Figure 34.
- Adder 332 in the fourth adder tree level 318(4) could be replaced with maximum or minimum selector 332', as illustrated in Figure 34.
- Adder 340 in the fifth adder tree level 318(5) could be replaced with maximum or minimum selector 340', as illustrated in Figure 34. Note that in the merging circuitry 294 in Figure 34, the adders could be configured to select a maximum or minimum resultant output vector data sample 292 between non-adjacent resultant output vector data samples 292.
- adders in adder tree levels 318(I)-318(3) could be configured to simply pass merge resultant output vector data samples 292(0) with resultant output vector data samples 292(9) to adder tree level 31 8(4).
- the adder 332' in adder tree level 318(4) could then maximum merge the resultant output vector data sample 292(0) with resultant output vector data samples 292(9) to provide merged output vector data samples 264.
- the execution units 84(0)-84(X) are provided in the VPEs 22(1 )-22(6) to perform vector processing operations on input vector data sample sets 86(0)-86(X).
- the execution units 84(0)-84(X) also include programmable data path configurations that allow the execution units 84(0)-84(X) to provide multiple modes of operation with common circuitry and hardware for different vector processing operations. More exemplary detail regarding the execution units 84(0)-84(X) and their programmable data path configurations for pro viding multiple modes of operation with common circuitry and hardware are now discussed.
- Figure 35 illustrates an exemplary schematic diagram of an exemplary execution unit that can be provided for each of the execution units 84(0)- 84(X) in the VPEs 22(l)-22(6).
- the execution unit 84 includes a plurality of exemplary vector pipeline stages 460 having exemplary vector processing blocks that may be configured with programmable data path configurations.
- the programmable data path configurations provided in the vector processing blocks allow specific circuits and hardware to be programmed and reprogrammed to support performing different, specific vector processing operations on the vector data 30 received from the vector unit data memory 32 in Figure 2.
- certain vector processing operations may commonly require multiplication of the vector data 30 followed by an accumulation of the multiplied vector data results.
- Non-limiting examples of such vector processing includes filtering operations, correlation operations, and Radix- 2 and Radix-4 butterfly operations commonly used for performing Fast Fourier Transform (FFT) operations for wireless communications algorithms, where a series of parallel multiplications are provided followed by a series of parallel accumulations of the multiplication results.
- FFT Fast Fourier Transform
- the execution unit 84 in Figure 35 also has the option of fusing multipliers with carry-save accumulators to provide a redundant carry-save format in the carry-save accumulators. Providing a redundant carry -save format in the carry-save accumulators can eliminate a need to provide a carry propagation path and a carry propagation add operation during each step of accumulation.
- the MO multiply vector pipeline stage 460(1) of the VPE 22 will first be described.
- the MO multiply vector pipeline stage 460(1 ) is a second vector pipeline stage containing a plurality of vector processing blocks in the form of any desired number of multiplier blocks 462(A)- 462(0), each having programmable data path configurations.
- the multiplier blocks 462(A)-462(0) are provided to perform vector multiply operations in the execution unit 84.
- the plurality of multiplier blocks 462(A.)-462(0) are disposed in parallel to each other in the M0 multiply vector pipeline stage 460(1) for providing multiplication of up to twelve (12) multiply vector data sample sets 34(Y)-34(0).
- 'A' is equal to three (3), meaning four (4) multiplier blocks 462(3)-462(0) are included in the M0 multiply vector pipeline stage 460(1) in this example.
- the multiply vector data sample sets 34(Y)-34(0) are loaded into the execution unit 84 for vector processing into a plurality of latches 464(Y)-464(0) provided in an input read (RR) vector pipeline stage, which is a first vector pipeline stage 460(0) in the execution unit 84.
- RR input read
- the latches 464( 11)- 464(0) are configured to latch the multiply vector data sample sets 34(1 1)-34(0) retrieved from the vector registers (see the vector data files 28 of Figure 2) as vector data input sample sets 466(1 i)-466(0).
- each latch 464(1 1)-464(0) is 8- bits wide.
- the latches 464(1 1)-464(0) are each respectively configured to latch the multiply vector data input sample sets 466(1 1)-466(0), for a total of 96-bits wide of vector data 30 (i.e., 12 latches x 8 bits each).
- the plurality of multiplier blocks 462(3)-462(0) are configured to be able to receive certain combinations of the vector data input sample sets 466(1 1)-466(0) for providing vector multiply operations, wherein 'Y' is equal to eleven ( 1 1 ) in this example.
- the multiply vector data input sample sets 466(11)-466(0) are provided in a plurality of input data paths A3 -AO, B3-B0, and C3 - C0 according to the design of the execution unit 84.
- Vector data input sample sets 466(3)-466(0) correspond to input data paths C3-C0 as illustrated in Figure 35.
- Vector data input sample sets 466(7)-466(4) correspond to input data paths B3-B0 as illustrated in Figure 35.
- Vector data input sample sets 466(1 1 )-466(8) correspond to input data paths A3 -AO as illustrated in Figure 35.
- the plurality of multiplier blocks 462(3)- 462(0) are configured to process the received vector data input sample sets 466( 1 1)- 466(0) according to the input data paths A3-A0, B3-B0, C3-C0, respectively, provided to the plurality of multiplier blocks 462(3)-462(0), to provide vector multiply operations.
- programmable internal data paths 467(3)-467(0) provided in the multiplier blocks 462(3)-462(0) in Figure 35 can be programmed to have different data path configurations. These different data path configurations provide different combinations and/or different bit lengths of multiplication of particular received vector data input sample sets 466(1 1)-466(0) provided to the multiplier blocks 462(3)-462(0) according to the particular input data paths A3-A0, B3-B0, C3-C0 provided to each multiplier block 462(3)-462(0).
- the plurality of multiplier blocks 462(3)-462(0) provide vector multiply output sample sets 468(3)-468(0) as a vector result output sample set comprising a multiplication result of multiplying a particular combination of the vector data input sample sets 466(11)-466(0) together.
- the programmable internal data paths 467(3)-467(0) of the multiplier blocks 462(3)-462(0) may be programmed according to settings provided from a vector instruction decoder in the instruction dispatch circuit 48 of the baseband processor 20 in Figure 2.
- the vector instruction specifies the specific type of operation to be performed by the execution unit 84.
- the execution unit 84 can be programmed and reprogrammed to configure the programmable internal data paths 467(3)-467(0) of the multiplier blocks 462(3)-462.(0) to provide different types of vector multiply operations with the same common circuitry in a highly efficient manner.
- the execution unit 84 may be programmed to configure and reconfigure the programmable internal data paths 467(3)-467(0) of the multiplier blocks 462(3)-462(0) on a clock-cycle-by-clock-cycle basis for each vector instruction executed, according to decoding of the vector instructions in an instruction pipeline in the instruction dispatch circuit 48.
- the MO multiply vector pipeline stage 460(1) in the execution unit 84 is configured to process vector data input sample sets 466 every clock cycle, as a result, the multiplier blocks 462(3)-462(0) perform vector multiply operations on every clock cycle according to decoding of the vector instructions in an instruction pipeline in the instruction dispatch circuit 48.
- the multiplier blocks 462 can be programmed to perform real and complex multiplications. With continuing reference to Figure 35, in one vector processing block data path configuration, a multiplier block 462 may be configured to multiply two 8-bit vector data input sample sets 466 together. in one multiply block data path configuration, a multiplier block 462 may be configured to multiply two 16-bit vector data input sample sets 466 together, which are formed from a first pair of 8-bit vector data input sample sets 466 multiplied by a second pair of 8-bit vector data input sample sets 466. This is illustrated in Figure 38 and discussed in more detail below.
- multiplier blocks 462(3)- 462(0) provides flexibility in that the multiplier blocks 462(3)-462(0) can be configured and reconfigured to perform different types of multiply operations to reduce area in the execution unit 84 and possibly allow fewer execution units 84 to be provided in the baseband processor 20 to carry out the desired vector processing operations.
- the plurality of multiplier blocks 462(3)- 462(0) is configured to provide the vector multiply output sample sets 468(3)-468(0) in programmable output data paths 470(3)-470(0) to either the next vector processing stage 460 or an output processing stage.
- the vector multiply output sample sets 468(3)- 468(0) are provided in the programmable output data paths 470(3)-470(0) according to a programmed configuration based on the vector instruction being executed by the plurality of multiplier blocks 462(3 )-462(0).
- the vector multiply output sample sets 468(3)-468(0) in the programmable output data paths 470(3)-470(0) are provided to the Ml accumulation vector pipeline stage 460(2) for accumulation, as will be discussed below.
- the vector multiply output sample sets 468(3)-468(0) are provided to a plurality of accumulator blocks 472(3)-472(0) provided in a next vector processing stage, which is the Mi accumulation vector processing stage 460(2).
- Each accumulator block among the plurality of accumulator blocks 472(A)-472(0) contains two accumulators 472(X)(1 ) and 472(X)(0) (i.e., 472(3)(1), 472(3)(0), 472(2)(1), 472(2)(0), 472(T)(1), 472(1 )(0), and 472(0)(1), 472(0)(G)).
- the plurality of accumulator blocks 472(3 )-472(0) accumulate the results of the vector multiply output sample sets 468 3)-468( ⁇ ).
- the plurality of accumulator blocks 472(3)-472(0) can be provided as carry-save accumulators, wherein the carry product is in essence saved and not propagated during the accumulation process until the accumulation operation is completed.
- the plurality of accumulator blocks 472(3)-472(0) also have the option of being fused with the plurality of multiplier blocks 462(3)-462(0) in Figures 35 and 37 to provide redundant carry-save format in the plurality of accumulator blocks 472(3)-472(0).
- Providing redundant carry-save format in the plurality of accumulator blocks 472(3)-472(0) can eliminate a need to provide a carry propagation path and a carry propagation add operation during each step of accumulation in the plurality of accumulator blocks 472(3)-472(0).
- the Ml accumulation vector processing stage 460(2) and its plurality of accumulator blocks 472(3)-472(0) will now be introduced with reference to Figure 35,
- the plurality of accumulator blocks 472(3)- 472(0) in the Ml accumulation vector processing stage 460(2) is configured to accumulate the vector multiply output sample sets 468(3 )-468(0) in programmable output data paths 474(3)-474(0) (i.e., 474(3)(1 ), 474(3)(0), 474(2)(1), 474(2)(0), 474(1)( 1), 474(1)(0), and 474(0)(1 ), 474(0)(0)), according to programmable output data path configurations, to provide accumulator output sample sets 476(3)-476(0) (i.e., 476(3)( 1), 476(3)(0), 476(2)(1 ), 476(2)(0), 476(1)( 1), 476(1)(0), and 476(0)(i), 476(0)(0)) in either a next vector processing stage 460 or an output processing stage, in this example, the accumulator output sample sets 476(3)-476(0) are provided to an output processing stage, which is an ALU processing stage 460(3).
- the accumulator output sample sets 476(3)-476(0) can also be provided to the ALU 46 in the scalar processor 44 in the baseband processor 20 in Figure 2, as a non-limiting example.
- the ALU 46 may take the accumulator output sample sets 476(3)-476(0) according to the specialized vector instructions executed by the execution unit 84 to be used in more general processing operations.
- programmable input data paths 478(3)- 478(0) and/or programmable internal data paths 480(3)-480(0) of the accumulator blocks 472(3)-472(0) can be programmed to be reconfigured to receive different combinations and/or bit lengths of the vector multiply output sample sets 468(3 )-468(0) provided from the multiplier blocks 462(3 )-462(0) to the accumulator blocks 472(3)- 472(0).
- each accumulator block 472 is comprised of two accumulators 472(X)(1 ), 472(X)(0)
- the programmable input data paths 478(A)-478(0) are shown in Figure 35 as 478(3)(1), 478(3)(0), 478(2)(1 ), 478(2)(0), 478(1)(1), 478(1 )(0), and 478(0)(1), 478(0)(0).
- the programmable internal data paths 480(3)-48Q( 0) are shown in Figure 35 as 480(3)(1), 480(3)(0), 480(2)(1 ), 480(2)(0), 480(1)(1), 480(1)(0), 480(0)(1), 480(0)(0).
- the accumulator blocks 472(3)-472(0) can provide the accumulator output sample sets 476(3)-476(0) according to the programmed combination of accumulated vector multiply output sample sets 468(3)-468(0).
- the accumulator blocks 472(3 )-472.(0) can be configured and reconfigured to perform different types of accumulation operations based on the programming of the programmable input data paths 478(3)-478(0) and/or the programmable internal data paths 480(3)-480(0) to reduce area in the execution unit 84 and possibly allow fewer execution units 84 to be provided in the baseband processor 20 to cany out the desired vector processing operations.
- the programmable input data path 478 and/or the programmable internal data paths 480 of two accumulator blocks 472 may be programmed to provide for a single 40-bit accumulator as a non- limiting example.
- the programmable input data path 478 and/or the programmable internal data path 480 of two accumulator blocks 472 may be programmed to provide for dual 24-bit accumulators as a non- limiting example.
- the programmable input data path 478 and/or the programmable internal data path 480 of two accumulator blocks 472 may be programmed to provide for a 16-bit carry -save adder followed by a single 24-bit accumulator.
- the execution unit 84 can also be supported by the execution unit 84 according to the programming of the multiplier blocks 462(3)-462(0) and the accumulator blocks 472(3)- 472(0) (e.g., 16-bit complex multiplication with 16-bit accumulation, and 32-bit complex multiplication with 16-bit accumulation).
- the programmable input data paths 478(3)-478(0) and/or the programmable internal data paths 480(3)-480(0) of the accumulator blocks 472(3)-472(0) may be programmed according to settings provided from a vector instruction decoder in the instruction dispatch circuit 48 of the baseband processor 2.0 in Figure 2,
- the vector instruction specifies the specific type of operation to be performed by the execution unit 84.
- the execution unit 84 can be configured to reprogram the programmable input data paths 478(3)-478(0) and/or the programmable internal data paths 48Q(3)-480(G) of the accumulator blocks 472(3)-472(0) for each vector instruction executed according to decoding of the vector instruction in an instruction pipeline in the instruction dispatch circuit 48.
- a vector instruction may execute over one or more clock cycles of the execution unit 84, Also in this example, the execution unit 84 can be configured to reprogram the programmable input data paths 478(3)-478(0) and/or the programmable internal data paths 480(3)-480(0) of the accumulator blocks 472(3)-472(0) for each clock cycle of a vector instruction on a clock-cycle-by-clock-cycle basis.
- the programmable input data paths 478(3)- 478(0) and/or the programmable internal data paths 480(3)-480(0) of the accumulator blocks 472(3)-472(0) can be reconfigured each clock cycle during execution of the vector instruction.
- Figure 36 is a flowchart illustrating exemplary vector processing of the multiplier blocks 462(A)-462(0) and the accumulator blocks 472(A)(l)-472(0)(0) in the execution unit 84 in Figures 2 and 35 to provide additional illustration of the exemplary vector processing.
- the multiplier blocks 462(A)-462(0) and accumulator blocks 472(A)(i)-472(0)(0) each have programmable data path configurations and are provided in different vector processing stages in the exemplary execution unit 84 of Figures 2 and 35. For example, FFT vector operations involve multiply operations followed by accumulate operations.
- the vector processing involves receiving a plurality of multiply vector data sample sets 34(Y)-34(0) of a width of a vector array in an input data path among a plurality of input data paths A3-C0 in an input processing stage 460(0) (block 501 ).
- the vector processing then includes receiving the multiply vector data sample sets 34(Y)-34(0) from the plurality of input data paths A3-C0 in a plurality of multiplier blocks 462(A)-462(0) (block 503).
- the vector processing then includes multiplying the multiply vector data sample sets 34(Y) ⁇ 34(0) to provide multiply vector result output sample sets 468(A)-468(0) in multiply output data paths 470(A)-470(0) among a plurality of multiply output data paths 470(A)-470(0), based on programmable data path configurations for the multiplier blocks 462(A) ⁇ 462(0) according to a vector instruction executed by the vector processing stage 460(1) (block 505).
- the vector processing next includes receiving the multiply vector result output sample sets 468(A)-468(0) from the plurality of multiply output data paths 470(A)-470(0) in a plurality of accumulator blocks 472(A)(1)- 472(0)(0) (block 507).
- the vector processing next includes accumulating the multiply vector result output sample sets 468(A)-468(0) together to provide accumulator output sample sets 476(A)(l)-476(0)(0) based on programmable input data paths 478(A)(1 )- 478(0)(0), programmable internal data paths 480(A)(l)-480(0)(0), and programmable output data paths 474(A)(l)-474(0)(0) configurations for the accumulator blocks 472(A)( 1 )-472(0)(0) according to a vector instruction executed by the second vector processing stage 460(2) (block 509).
- the vector processing then includes providing the accumulator output sample sets 476(A)(l)-476(0)(0) in the programmable output data paths 474(A)(i)-474(0)(0) (block 51 1).
- the vector processing then includes receiving the accumulator output sample sets 476(A)( 1 )-476(0)(0) from the accumulator blocks 472(A)(l )-472(0)(0) in an output vector processing stage 460(3) (block 513).
- Figure 37 is a more detailed schematic diagram of the plurality of multiplier blocks 462(3)-462(0) in the M0 multiply vector processing stage 460(1 ) of the execution unit 84 of Figure 35.
- Figure 38 is a schematic diagram of internal components of a multiplier block 462 in Figure 37. As illustrated in Figure 37, the vector data input sample sets 466(1 1)-466(0) that are received by the multiplier blocks 462(3)-462(0) according to the particular input data paths A3-A0, B3-B0, C3-C0 are shown. As will be discussed in more detail below with regard to Figure 38.
- each of the multiplier blocks 462(3)-462(0) in this example include four (4) 8-bit by 8-bit multipliers.
- each of the multiplier blocks 462(3)- 462(0) in this example are configured to multiply a multiplicand input 'A' by either multiplicand input 'B' or multiplicand input 'C.”
- the multiplicand inputs ⁇ ," and 'B' or 'C that can be multiplied together in a multiplier block 462 are controlled by which input data paths A3-A0, B3-B0, C3-C0 are connected to the multiplier blocks 462(3)- 462(0), as shown in Figure 37.
- a multiplicand selector input 482(3)-482(0) is provided as an input to each multiplier block 462(3)-462(0) to control the programmable internal data paths 467(3)-467(0) in each multiplier block 462(3)-462(0) to select whet er multiplicand input 'B' or multiplicand input 'C is selected to be multiplied by multiplicand input 'A.'
- the multiplier blocks 462(3)-462(0) are provided with the capability for their programmable internal data paths 467(3)-467(0) to be reprogrammed to provide different multiply operations, as desired.
- input data paths A3 and A2 are connected to inputs AH and AL, respectively.
- Input AH represents the high bits of multiplicand input A
- AL means the low bits of input multiplicand input 'A
- Input data paths B3 and B2 are connected to inputs BH and BL, respectively
- input BH represents the high bits of multiplicand input ' ⁇
- AL represents the low bits of input multiplicand input ⁇ .”
- Input data paths C3 and C2 are connected to inputs CI and CQ, respectively.
- Input CI represents the real bits portion of input multiplicand input 'C in this example.
- multiplicand selector input 482(3) also controls whether the programmable internal data path 467(3) of multiplier block 462(3) is configured to perform 8-bit multiplication on multiplicand input 'A' with multiplicand input 'B' or multiplicand input 'C,' or whether multiplier block 462(3) is configured to perform i 6-bit multiplication on multiplicand input 'A' with multiplicand input 'B' or multiplicand input 'C in this example.
- the multiplier blocks 462(3)-462(0) are configured to each generate vector multiply output sample sets 468(3)-468(0) as carry 'C and sum 'S' vector output sample sets of the multiplication operation based on the configuration of their programmable internal data paths 467(3)-467(0).
- the carry 'C and sum 'S' of the vector multiply output sample sets 468(3) -468(0) are fused, meaning that the carry 'C and the sum 'S' are provided in redundant carry-save format to the plurality of accumulator blocks 472(3)-472(0) to provide redundant carry-save format in the plurality of accumulator blocks 472(3)-472(0).
- providing a redundant carry-save format in the plurality of accumulator blocks 472(3)-472(0) can eliminate a need to provide a carry propagation path and a carry propagation add operation during accumulation operations performed by the plurality of accumulator blocks 472(3)-472(0).
- multiplier blocks 462(3)-462(0) generating the vector multiply output sample sets 468(3)-468(0) as carry 'C and sum 'S' vector output sample sets of the multiplication operation based on the configuration of their programmable internal data paths 467(3)-467(0) are shown in Figure 37.
- multiplier block 462(3) is configured to generate carry COO and sum SOO as 32-bit values for 8-bit multiplications and carry C01 and sum S01 as 64-bit values for 16-bit multiplications.
- the other multiplier blocks 462(2)-462(0) have the same capability in this example.
- multiplier block 462(2) is configured to generate carry CIO and sum SI 0 as 32-bit values for 8-bit multiplications and carry CI 1 and sum SI 1 as 64-bit values for 16-bit multiplications.
- Multiplier block 462(1) is configured to generate carry C20 and sum S20 as 32-bit values for 8-bit multiplications and cany C21, and sum S21 as 64-bit values for i 6-bit multiplications.
- Multiplier block 462(0) is configured to generate cany C30 and sum S30 as 32-bit values for 8-bit multiplications and carry C31 and sum S31 as 64-bit values for 16-bit multiplications.
- Figure 38 is a schematic diagram of internal components of a multiplier block 462 in Figure 37 having programmable data path configurations capable of multiplying 8-bit by 8-bit vector data input sample set 466, and 16-bit by 16-bit vector data input sample set 466.
- the multiplier block 462 includes four 8x8-bit multipliers 484(3)-484(0) in this example. Any desired number of multipliers 484 could be provided.
- a first multiplier 484(3) is configured to receive 8-bit vector data input sample set 466A[H] (which is the high bits of input multiplicand input ⁇ ') and multiply the vector data input sample set 466 A[H] with either 8-bit vector data input sample set 466B[H] (which is the high bits of input multiplicand input ⁇ ') or 8-bit vector data input sample set 466C[I] (which is the high bits of input multiplicand input 'C').
- a multiplexor 486(3) is provided that is configured to select either 8-bit vector data input sample set 466B[H] or 8-bit vector data input sample set 466C[I] being providing as a multiplicand to the multiplier 484(3).
- the multiplexor 486(3) is controlled by multiplicand selector input 482[3], which is the high bit in the multiplicand selector input 482 in this embodiment.
- the multiplexor 486(3) and the multiplicand selector input 482[3] provide a programmable internal data path 467[0] configuration for the multiplier 484(3) to control whether 8-bit vector data input sample set 466B[H] or 8-bit vector data input sample set 466C[I] is multiplied with received vector data input sample set 466A[H].
- multipliers 484(2)-484(0) also include similar programmable internal data paths 467[2]-467[0] as provided for the first multiplier 484(3).
- Multiplier 484(2) includes the programmable internal data path 467 [2] having a programmable configuration to provide either 8-bit vector data input sample set 466B[H] or 8-bit vector data input sample set 466C[I] in the programmable internal data path 467[1] to be multiplied with 8-bit vector data input sample set 466A[L], which is the low bits of multiplicand input 'A.'
- the selection is controlled by multiplexor 486(2) according to the multiplicand selector input 482[2] in the multiplicand selector input 482 in this embodiment.
- Multiplier 484(1) includes programmable internal data path 467[1] programmable to provide either 8-bit vector data input sample set 466B[L], which is the low bits of multiplicand input ' ⁇ ,' or 8-bit vector data input sample set 466C[Q], which is the low bits of multiplicand input 'C in the programmable internal data path 467[1 ] to be multiplied with 8-bit vector data input sample set 466A[H], The selection is controlled by multiplexor 486(1) according to the multiplicand selector input 482[1] in the multiplicand selector input 482 in this embodiment.
- multiplier 484(0) includes programmable internal data path 467[0] programmable to provide either 8-bit vector data input sample set 466B[L] or 8- bit vector data input sample set 466C[Q] in the programmable internal data path 467[0J, to be multiplied with 8-bit vector data input sample set 466A[L].
- the selection is controlled by multiplexor 486(0) according to the multiplicand selector bit input 482[0] in the multiplicand selector input 482 in this embodiment.
- each multiplier 484(3)-484(0) can be configured to perform different bit length multiplication operations.
- each multiplier 484(3)-484(0) includes bit length multiply mode inputs 488(3)-488(0), respectively.
- each multiplier 484(3)-484(0) can be programmed in 8-bit by 8-bit mode according to the inputs that control the configuration of programmable data paths 490(3)-490(0), 491, and 492(3)-492(0), respectively.
- Each multiplier 484(3 )-484(0) can also be programmed to provide part of a larger bit multiplication operation, including 16-bit by 16-bit mode and 24-bit by 8-bit mode, according to the inputs that control the configuration of programmable data paths 490(3)-490(0), 491 , and 492(3) ⁇ 492(0), respectively.
- the plurality of multipliers 484(3)-484(0) as a unit can be configured to comprise two (2) individual 8-bit by 8-bit multipliers as part of the multiplier block 462.
- each multiplier 484(3)-484(0) is configured in 16-bit by 16-bit multiply mode according to configuration of the programmable data path 491
- the plurality of multipliers 484(3)-484(0) as a unit can be configured to comprise a single 16-bit by 16-bit multiplier as part of the multiplier block 462.
- the multipliers 484(3)-484(Q) are configured in 24-bit by 8-bit multiply mode according to coniigiiration of the programmable data paths 492(3)-492(0)
- the plurality of multipliers 484(3)-484(0) as a unit can be configured to comprise one (1) 16-bit by 24-bit by 8-bit multiplier as part of the multiplier block 462.
- the multipliers 484(3 )-484(G) in this example are shown as being configured in 16-bit by 16-bit multiply mode. Sixteen (16)-bit input sums 494(3), 494(2) and input carries 496(3), 496(2) are generated by each multiplier 484(3), 484(2), respectively. Sixteen (16)-bit input sums 494( 1), 494(0) and input carries 496(1), 496(0) are generated by each multiplier 484( 1), 484(G), respectively.
- the 1 -bit input sums 494(3), 494(2) and input carries 496(3), 496(2) are also provided to a 24-bit 4:2 compressor 515 along with 16-bit sums input 494(1), 494(0) and input carries 496(1), 496(0) to add the input sums 494(3)-494(0) and input carries 496(3)-496(G) together.
- the added input sums 494(3)-494(0) and input carries 496(3)-496(0) provide a single sum 498 and single carry 500 in 16-bit by 16-bit multiply mode when the programmable data path 491 is active and gated with the input sums 494(3)-494(0) and input carries 496(3)-496(0).
- the programmable data path 491 is gated by a first AND-based gate 502(3) with combined input sums 494(3), 494(2) as a 16-bit word, and by a second AND-based gate 502(2) with combined input carries 496(3), 496(2) as a 16-bit word to be provided to the 24-bit 4:2 compressor 515.
- the programmable data path 491 is also gated by a third AND-based gate 502( 1) with combined input sums 494(1 ), 494(0) as a 16-bit word, and by a fourth AND-based gate 502(0) with combined input carries 496(1 ), 496(0) as a 16-bit word to be provided to the 24-bit 4:2 compressor 515.
- the programmable output data path 470[0] is provided with the vector multiply output sample set 468 [0] as a compressed 32-bit sum SO and 32-bit carry CO partial product if the multiplier block 462 is configured in a 16-bit by 16-bit or 24-bit by 8-bit multiply mode.
- the programmable output data path 470[1] configuration is provided as the 16-bit input sums 494(3)-494(0) and corresponding 16-bit input carries 496(3)-496(0) as partial products without compression, if the multipliers 484(3)-484(0) in the multiplier block 462 are configured in 8-bit by 8-bit multiply mode.
- the programmable output data path 470[1] is provided as the 16-bit input sums 494(3 )-494(0) and corresponding 16-bit input carries 496(3)-496(0) as the vector multiply output sample sets 468[1 ] without compression if the multipliers 484(3)-484(0) in the multiplier block 462 are configured in 8-bit by 8 -bit multiply mode.
- the vector multiply output sample sets 468[0], 468[I], depending on a multiplication mode of the multiplier block 462, are provided to the accumulator blocks 472(3)-472(0) for accumulation of sum and cany products according to the vector instruction being executed.
- Figure 39 is a generalized schematic diagram of a multiplier block and accumulator block in the execution units 84(0)-84(X) described above, wherein the accumulator block employs a carry-save accumulator structure employing redundant carry-save format to reduce carry propagation.
- the multiplier blocks 462 are configured to multiply multiplicand inputs 466 [H] and 466[L] and provide at least one input sum 494 and at least one input cany 496 as a vector multiply output sample sets 468 in the programmable output data path 470.
- the at least one input sum 494 and the at least one input carry 496 in the vector multiply output sample sets 468 in the programmable output data path 470 are fused in redundant carry-save format to at least one accumulator block 472.
- the carry 496 in the vector multiply output sample sets 468 is provided as vector input cany 496 in carry-save format to the accumulator block 472.
- the input sum 494 and the input carry 496 in the vector multiply output sample sets 468 can be provided to a compressor 508 of the accumulator block 472, which in this embodiment is a complex gale 4:2 compressor.
- the compressor 508 is configured to accumulate the input sum 494 and the input carry 496 together with a previous accumulated vector output sum 512 and a previous shifted accumulated vector output carry 517, respectively.
- the previous shifted accumulated vector output carry 517 is in essence the saved carry accumulation during the accumulation operation.
- the compressor 508 is configured to accumulate the input sum 494 and the input carry 496 in a redundant form with the previous accumulated vector output sum 512 and previous shifted accumulated vector output carry 517, respectively.
- the shifted accumulated vector output carry 517 is generated by an accumulated vector output carry 14 generated by the compressor 508 by shifting the accumulated vector output carry 514 before the next accumulation of the next received input sum 494 and input carry 496 is performed by the compressor 508.
- the final shifted accumulated vector output carry 517 is added to the final accumulated vector output sum 512 by a single, final carry propagate adder 519 provided in the accumulator block 472 to propagate the carry accumulation in the final shifted accumulated vector output carry 51 7 to convert the final accumulated vector output sum 12 to the final accumulator output sample set 476 2's complement notation.
- the final accumulated vector output sum 512 is provided as accumulator output sample set 476 in the programmable output data path 474 (see Figure 35).
- Figure 40 is a detailed schematic diagram of exemplar internal components of an accumulator block 472 provided in the execution unit 84 of Figure 35.
- the accumulator block 472 is configured with programmable input data paths 478(3)-478(0) and/or the programmable internal data paths 480(3)-480(0), so that the accumulator block 472 can be programmed to act as dedicated circuitry designed to perform specific, different types of vector accumulation operations.
- the accumulator block 472 can be programmed to provide a number of different accumulations and additions, including signed and unsigned accumulate operations.
- Specific examples of the programmable input data paths 478(3)-478(0) and/or programmable internal data paths 480(3)-480(0) in the accumulator block 472 being configured to provide different types of accumulation operations are disclosed.
- the accimiuiator block 472 is configured to include carry-save accumulators 472[0], 472[1] to provide redundant carry arithmetic to avoid or reduce carry propagation to provide high speed accumulation operations with reduced combinational logic.
- FIG. 40 Exemplary internal components of the accumulator block 472 are shown in Figure 40.
- the accumulator block 472 in this embodiment is configured to receive a first input sum 494 ⁇ 0] and first input carry 496[0], and a second input sum 494[I] and second input carry 496[lj from a multiplier block 462 to be accumulated together.
- the input sums 494[0], 494[1] and input carries 496[0], 496[1] will be referred to as vector input sums 494[0], 494[1 ] and vector input carries 496[0], 496[1 ].
- the vector input sums 494[0], 494[1] and vector input carries 496[0], 496[1] in this embodiment are each 16-bits in length.
- the accumulator block 472 in this example is provided as two 24-bit carry-save accumulator blocks 472[0], 472[1], each containing similar components with common element numbers with '[0]' being designated for carry-save accumulator block 472[0], and with '[1]' being designated for carry-save accumulator block 472[1 ].
- the carry-save accumulator blocks 472 [0], 472[1] can be configured to perform vector accumulation operations concurrently.
- the vector input sum 494[0] and vector input carry 496[0] are input in a multiplexor 504(0) provided as part of the programmable internal data path 480[0].
- a negation circuit 506(0) which may be comprised of exclusive OR-based gates, is also provided that generates a negative vector input sum 494[0]' and negative vector input carry 496[0]' according to an input 521(0), as inputs into the multiplexor 504(0) for accumulation operations requiring a negative vector input sum 494 [0]' and negative vector input carry 496 [0]'.
- the multiplexor 504(0) is configured to select either vector input sum 494 [0] and vector input carry 496 [0] or the negative vector input sum 494 [0 J' and the negative vector input carry 496[0]' to be provided to a compressor 508(0) according to a selector input 510(0) generated as a result of the vector instruction decoding.
- the selector input 510(0) allows the programmable input data path 478 [0] of carry-save accumulator block 472 [0] to be programmable to provide either the vector input sum 494[0] and vector input carry 496[0], or the negative vector input sum 494 [0]' and the negative vector input carry 496[0]', to the compressor 508(0) according to the accumulation operation configured to be performed by the accumulator block 472.
- the compressor 508(0) of the carry- save accumulator block 472 [0] in this embodiment is a complex gate 4:2 compressor.
- the compressor 508(0) is configured to accumulate sums and carries in redundant carry-save operations.
- the compressor 508(0) is configured to accumulate a current vector input sum 494[0] and vector input carry 496[0], or a current negative vector input sum 494[0]' and negative vector input carry 496[0]', together with previous accumulated vector input sum 494 ⁇ 0] and vector input carry 496[0], or accumulated negative vector input sum 494[0]' and negative vector input carry 496[0]', as the four (4) inputs to the compressor 508(0),
- the compressor 508(0) provides an accumulated vector output sum 512(0) and accumulated vector output cany 514(0) as the accumulator output sample set 476 ⁇ 0] in the programmable output data path 474[0] (see Figure 35) to provide accumulator output sample sets 476(3)-476(0).
- the accumulated vector output carry 514(0) is shifted by a bit shifter 516(0) during accumulation operations to provide a shifted accumulated vector output carry 517(0) to control bit width growth during each accumulation step.
- the bit shifter 516(0) in this embodiment is a barrel-shifter that is fused to the compressor 508(0) in redundant carry- save format.
- the shifted accumulated vector output carry 517(0) is in essence saved without having to be propagated to the accumulated vector output sum 512(0) during the accumulation operation performed by the accumulator block 472[0J.
- power consumption and gate delay associated with performing a carry propagation add operation during each step of accumulation in the accumulator block 472 [0] is eliminated in this embodiment.
- Additional follow-on vector input sums 494[0] and vector input carries 496[0], or negative vector input sums 494[0]' and negative vector input carries 496[0]' can be accumulated with the current accumulated vector output sum 512(0) and current accumulated vector output carry 517(0).
- the vector input sums 494[0] and vector input carries 496[0], or negative vector input sums 494 [0]' and negative vector input carries 496[0]' are selected by a multiplexor 518(0) as part of the programmable internal data path 480[0] according to a sum-cany selector 520(0) generated as a result of the vector instruction decoding.
- the current accumulated vector output sum 512(0) and current shifted accumulated vector output carry 517(0) can be provided as inputs to the compressor 508(0) for can -save accumulator block 472[0] to provide an updated accumulated vector output sum 512(0) and accumulated vector output carry 514(0).
- the sum-carry selector 52.0(0) allows the programmable internal data path 480[0] of accumulator block 472[0] to be programmable to provide the vector input sum 494[0] and vector input carry 496 ⁇ 0] to the compressor 508(0) according to the accumulation operation configured to be performed by the accumulator block 472.
- Hold gates 522(0), 524(0) are also provided in this embodiment to cause the multiplexor 18(0) to hold the current state of the accumulated vector output sum 512(0) and shifted accumulated vector output carry 517(0) according to a hold state input 526(0) to control operational timing of the accumulation in the carry -save accumulator block 472 [0].
- the accumulated vector output sum 512(0) and shifted accumulated vector output carry 517(0) of carry-save accumulator block 472[0], and the accumulated vector output sum 512(1 ) and shifted accumulated vector output carry 17(1) of carry-save accumulator block 472[1], are gated by control gates 534(0), 536(0) and 534(1), 536(1), respectively.
- control gates 534(0), 536(0) and 534(1 ), 536(1 ) control the accumulated vector output sum 512(0) and shifted accumulated vector output carry 517(0), and the accumulated vector output sum 512(1) and shifted accumulated vector output carry 17(1 ), respectively, being returned to the compressors 508(0), 508(1).
- the accumulator block 472 can be configured in different modes.
- the accumulator block 472 can be configured to provide different accumulation operations according to a specific vector processing instruction with common accumulator circuitry illustrated in Figure 40.
- VPEs may be provided in or integrated into any processor-based device. Examples, without limitation, include a set top box, an entertainment unit, a navigation device, a communications device, a fixed location data unit, a mobile location data unit, a mobile phone, a cellular phone, a computer, a portable computer, a desktop computer, a personal digital assistant (PDA), a monitor, a computer monitor, a television, a tuner, a radio, a satellite radio, a music player, a digital music player, a portable music player, a digital video player, a video player, a digital video disc (DVD) player, and a portable digital video player.
- PDA personal digital assistant
- Figure 41 illustrates an example of a processor-based system 550.
- the processor-based system 550 includes one or more processing units (PUs) 552, each including one or more processors or cores 554.
- the PU(s) 552 may be the baseband processor 20 in Figure 2 as a non-limiting example.
- the processor 554 may be a vector processor like the baseband processor 20 provided in Figure 2 as a non-limiting example.
- the processor 554 may also include a VPE 556, including but not limited to the execution unit 84 in Figure 2.
- the PU(s) 552 may have cache memory 558 coupled to the processor(s) 554 for rapid access to temporarily stored data.
- the PU(s) 552 is coupled to a system bus 560 and can intercouple master and slave devices included in the processor-based system 550. As is well known, the PU(s) 552 communicates with these other devices by exchanging address, control, and data information over the system bus 560. For example, the PU(s) 552 can communicate bus transaction requests to a memory controller 562 as an example of a slave device. Although not illustrated in Figure 41, multiple system buses 560 could be provided, wherein each system bus 560 constitutes a different fabric.
- Other master and slave devices can be connected to the system bus 560. As illustrated in Figure 41 , these devices can include a memory system 564, one or more input devices 566, one or more output devices 568, one or more network interface devices 570, and one or more display controllers 572, as examples.
- the memory system 564 can include memory 565 accessible by the memory controller 562,
- the input device(s) 566 can include any type of input device, including but not limited to input keys, switches, voice processors, etc.
- the output device(s) 568 can include any type of output device, including but not limited to audio, video, other visual indicators, etc.
- the network interface device(s) 570 can be any devices configured to allow exchange of data to and from a network 574.
- the network 574 can be any type of network, including but not limited to a wired or wireless network, a private or public network, a local area network (LAIN), a wide local area network (WLAN), and the internet.
- the network interface device(s) 570 can be configured to support any type of communication protocol desired.
- the PUs 552 may also be configured to access the display controller(s) 572 over the system bus 560 to control information sent to one or more displays 578.
- the display controllers) 572 sends information to the display(s) 578 to be displayed via one or more video processors 580, which process the information to be displayed into a format suitable for the display(s) 578.
- the displa (s) 578 can include any type of display, including but not limited to a cathode ray tube (CRT), a liquid crystal display (LCD), a plasma display, etc.
- a processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine.
- a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- RAM Random Access Memory
- ROM Read Only Memory
- EPROM Electrically Programmable ROM
- EEPROM Electrically Erasable Programmable ROM
- registers hard disk, a removable disk, a CD-ROM, or any other form of computer readable medium known in the art.
- An exemplary storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium.
- the storage medium may be integral to the processor.
- the processor and the storage medium may reside in an ASIC.
- the ASIC may reside in a remote station.
- the processor and the storage medium may reside as discrete components in a remote station, base station, or server.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Advance Control (AREA)
- Executing Machine-Instructions (AREA)
- Complex Calculations (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020167015680A KR20160085335A (en) | 2013-11-15 | 2014-11-13 | Vector processing engine employing reordering circuitry in data flow paths between vector data memory and execution units, and related method |
CN201480062406.0A CN105765523B (en) | 2013-11-15 | 2014-11-13 | Use the vector processing engine for the circuit system that reorders with and related methods in data flow path between vector data memory and execution unit |
EP14812034.8A EP3069233A1 (en) | 2013-11-15 | 2014-11-13 | Vector processing engine employing reordering circuitry in data flow paths between vector data memory and execution units, and related method |
JP2016530912A JP2016541057A (en) | 2013-11-15 | 2014-11-13 | A vector processing engine (VPE) and associated vector processor that utilizes a reordering circuit in the data flow path between the execution unit and the vector data memory to provide in-flight reordering of output vector data stored in the vector data memory System and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/082,081 US9977676B2 (en) | 2013-11-15 | 2013-11-15 | Vector processing engines (VPEs) employing reordering circuitry in data flow paths between execution units and vector data memory to provide in-flight reordering of output vector data stored to vector data memory, and related vector processor systems and methods |
US14/082,081 | 2013-11-15 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015073646A1 true WO2015073646A1 (en) | 2015-05-21 |
Family
ID=52023626
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/065412 WO2015073646A1 (en) | 2013-11-15 | 2014-11-13 | Vector processing engine employing reordering circuitry in data flow paths between vector data memory and execution units, and related method |
Country Status (6)
Country | Link |
---|---|
US (1) | US9977676B2 (en) |
EP (1) | EP3069233A1 (en) |
JP (1) | JP2016541057A (en) |
KR (1) | KR20160085335A (en) |
CN (1) | CN105765523B (en) |
WO (1) | WO2015073646A1 (en) |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9495154B2 (en) | 2013-03-13 | 2016-11-15 | Qualcomm Incorporated | Vector processing engines having programmable data path configurations for providing multi-mode vector processing, and related vector processors, systems, and methods |
US9684509B2 (en) | 2013-11-15 | 2017-06-20 | Qualcomm Incorporated | Vector processing engines (VPEs) employing merging circuitry in data flow paths between execution units and vector data memory to provide in-flight merging of output vector data stored to vector data memory, and related vector processing instructions, systems, and methods |
US9792118B2 (en) | 2013-11-15 | 2017-10-17 | Qualcomm Incorporated | Vector processing engines (VPEs) employing a tapped-delay line(s) for providing precision filter vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods |
US9880845B2 (en) | 2013-11-15 | 2018-01-30 | Qualcomm Incorporated | Vector processing engines (VPEs) employing format conversion circuitry in data flow paths between vector data memory and execution units to provide in-flight format-converting of input vector data to execution units for vector processing operations, and related vector processor systems and methods |
US9619227B2 (en) | 2013-11-15 | 2017-04-11 | Qualcomm Incorporated | Vector processing engines (VPEs) employing tapped-delay line(s) for providing precision correlation / covariance vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods |
KR102240728B1 (en) * | 2015-01-27 | 2021-04-16 | 한국전자통신연구원 | Bit interleaver for 64-symbol mapping and low density parity check codeword with 64800 length, 4/15 rate, and method using the same |
KR102287614B1 (en) * | 2015-02-12 | 2021-08-10 | 한국전자통신연구원 | Bit interleaver for 16-symbol mapping and low density parity check codeword with 64800 length, 2/15 rate, and method using the same |
KR102287630B1 (en) * | 2015-02-17 | 2021-08-10 | 한국전자통신연구원 | Bit interleaver for 16-symbol mapping and low density parity check codeword with 16200 length, 3/15 rate, and method using the same |
CN109284823B (en) * | 2017-04-20 | 2020-08-04 | 上海寒武纪信息科技有限公司 | Arithmetic device and related product |
KR102343652B1 (en) | 2017-05-25 | 2021-12-24 | 삼성전자주식회사 | Method for aligning sequence for vector processor |
GB2569844B (en) * | 2017-10-20 | 2021-01-06 | Graphcore Ltd | Sending data off-chip |
GB2569271B (en) | 2017-10-20 | 2020-05-13 | Graphcore Ltd | Synchronization with a host processor |
GB2569775B (en) | 2017-10-20 | 2020-02-26 | Graphcore Ltd | Synchronization in a multi-tile, multi-chip processing arrangement |
US11277455B2 (en) | 2018-06-07 | 2022-03-15 | Mellanox Technologies, Ltd. | Streaming system |
GB2575294B8 (en) | 2018-07-04 | 2022-07-20 | Graphcore Ltd | Host Proxy On Gateway |
US20200106828A1 (en) * | 2018-10-02 | 2020-04-02 | Mellanox Technologies, Ltd. | Parallel Computation Network Device |
GB2579412B (en) | 2018-11-30 | 2020-12-23 | Graphcore Ltd | Gateway pull model |
US11625393B2 (en) | 2019-02-19 | 2023-04-11 | Mellanox Technologies, Ltd. | High performance computing system |
EP3699770A1 (en) | 2019-02-25 | 2020-08-26 | Mellanox Technologies TLV Ltd. | Collective communication system and methods |
CN110795687A (en) * | 2019-10-29 | 2020-02-14 | 南京宁麒智能计算芯片研究院有限公司 | Hierarchical segmentation system and method for autocorrelation algorithm |
US11750699B2 (en) | 2020-01-15 | 2023-09-05 | Mellanox Technologies, Ltd. | Small message aggregation |
US11252027B2 (en) | 2020-01-23 | 2022-02-15 | Mellanox Technologies, Ltd. | Network element supporting flexible data reduction operations |
US11876885B2 (en) | 2020-07-02 | 2024-01-16 | Mellanox Technologies, Ltd. | Clock queue with arming and/or self-arming features |
US11556378B2 (en) | 2020-12-14 | 2023-01-17 | Mellanox Technologies, Ltd. | Offloading execution of a multi-task parameter-dependent operation to a network device |
US11922237B1 (en) | 2022-09-12 | 2024-03-05 | Mellanox Technologies, Ltd. | Single-step collective operations |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0744686A1 (en) * | 1995-05-17 | 1996-11-27 | STMicroelectronics Limited | Manipulation of data |
GB2409063A (en) * | 2003-12-09 | 2005-06-15 | Advanced Risc Mach Ltd | Vector by scalar operations |
WO2006106342A2 (en) * | 2005-04-08 | 2006-10-12 | Icera Inc. | Data access and permute unit |
US20080140750A1 (en) * | 2006-12-12 | 2008-06-12 | Arm Limited | Apparatus and method for performing rearrangement and arithmetic operations on data |
GB2464292A (en) * | 2008-10-08 | 2010-04-14 | Advanced Risc Mach Ltd | SIMD processor circuit for performing iterative SIMD multiply-accumulate operations |
EP2455854A1 (en) * | 2010-11-18 | 2012-05-23 | Ceva D.S.P. Ltd. | System, device, and method for on-the-fly permutations of vector memories for executing intra-vector operations |
Family Cites Families (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5524256A (en) * | 1993-05-07 | 1996-06-04 | Apple Computer, Inc. | Method and system for reordering bytes in a data stream |
GB9801713D0 (en) * | 1998-01-27 | 1998-03-25 | Sgs Thomson Microelectronics | Executing permutations |
US7395538B1 (en) | 2003-03-07 | 2008-07-01 | Juniper Networks, Inc. | Scalable packet processing systems and methods |
GB2399900B (en) | 2003-03-27 | 2005-10-05 | Micron Technology Inc | Data reording processor and method for use in an active memory device |
US20050050303A1 (en) | 2003-06-30 | 2005-03-03 | Roni Rosner | Hierarchical reorder buffers for controlling speculative execution in a multi-cluster system |
JP2007034731A (en) | 2005-07-27 | 2007-02-08 | Toshiba Corp | Pipeline processor |
JP4686435B2 (en) * | 2006-10-27 | 2011-05-25 | 株式会社東芝 | Arithmetic unit |
US8140932B2 (en) | 2007-11-26 | 2012-03-20 | Motorola Mobility, Inc. | Data interleaving circuit and method for vectorized turbo decoder |
US8078834B2 (en) * | 2008-01-09 | 2011-12-13 | Analog Devices, Inc. | Processor architectures for enhanced computational capability |
KR20140092293A (en) * | 2011-10-17 | 2014-07-23 | 파나소닉 주식회사 | Adaptive equalizer |
US20140280407A1 (en) | 2013-03-13 | 2014-09-18 | Qualcomm Incorporated | Vector processing carry-save accumulators employing redundant carry-save format to reduce carry propagation, and related vector processors, systems, and methods |
US9275014B2 (en) | 2013-03-13 | 2016-03-01 | Qualcomm Incorporated | Vector processing engines having programmable data path configurations for providing multi-mode radix-2x butterfly vector processing circuits, and related vector processors, systems, and methods |
US9495154B2 (en) | 2013-03-13 | 2016-11-15 | Qualcomm Incorporated | Vector processing engines having programmable data path configurations for providing multi-mode vector processing, and related vector processors, systems, and methods |
US9880845B2 (en) | 2013-11-15 | 2018-01-30 | Qualcomm Incorporated | Vector processing engines (VPEs) employing format conversion circuitry in data flow paths between vector data memory and execution units to provide in-flight format-converting of input vector data to execution units for vector processing operations, and related vector processor systems and methods |
US20150143076A1 (en) | 2013-11-15 | 2015-05-21 | Qualcomm Incorporated | VECTOR PROCESSING ENGINES (VPEs) EMPLOYING DESPREADING CIRCUITRY IN DATA FLOW PATHS BETWEEN EXECUTION UNITS AND VECTOR DATA MEMORY TO PROVIDE IN-FLIGHT DESPREADING OF SPREAD-SPECTRUM SEQUENCES, AND RELATED VECTOR PROCESSING INSTRUCTIONS, SYSTEMS, AND METHODS |
US9619227B2 (en) | 2013-11-15 | 2017-04-11 | Qualcomm Incorporated | Vector processing engines (VPEs) employing tapped-delay line(s) for providing precision correlation / covariance vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods |
US9792118B2 (en) | 2013-11-15 | 2017-10-17 | Qualcomm Incorporated | Vector processing engines (VPEs) employing a tapped-delay line(s) for providing precision filter vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods |
US9684509B2 (en) | 2013-11-15 | 2017-06-20 | Qualcomm Incorporated | Vector processing engines (VPEs) employing merging circuitry in data flow paths between execution units and vector data memory to provide in-flight merging of output vector data stored to vector data memory, and related vector processing instructions, systems, and methods |
-
2013
- 2013-11-15 US US14/082,081 patent/US9977676B2/en active Active
-
2014
- 2014-11-13 CN CN201480062406.0A patent/CN105765523B/en not_active Expired - Fee Related
- 2014-11-13 WO PCT/US2014/065412 patent/WO2015073646A1/en active Application Filing
- 2014-11-13 JP JP2016530912A patent/JP2016541057A/en active Pending
- 2014-11-13 KR KR1020167015680A patent/KR20160085335A/en not_active Application Discontinuation
- 2014-11-13 EP EP14812034.8A patent/EP3069233A1/en not_active Withdrawn
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0744686A1 (en) * | 1995-05-17 | 1996-11-27 | STMicroelectronics Limited | Manipulation of data |
GB2409063A (en) * | 2003-12-09 | 2005-06-15 | Advanced Risc Mach Ltd | Vector by scalar operations |
WO2006106342A2 (en) * | 2005-04-08 | 2006-10-12 | Icera Inc. | Data access and permute unit |
US20080140750A1 (en) * | 2006-12-12 | 2008-06-12 | Arm Limited | Apparatus and method for performing rearrangement and arithmetic operations on data |
GB2464292A (en) * | 2008-10-08 | 2010-04-14 | Advanced Risc Mach Ltd | SIMD processor circuit for performing iterative SIMD multiply-accumulate operations |
EP2455854A1 (en) * | 2010-11-18 | 2012-05-23 | Ceva D.S.P. Ltd. | System, device, and method for on-the-fly permutations of vector memories for executing intra-vector operations |
Also Published As
Publication number | Publication date |
---|---|
EP3069233A1 (en) | 2016-09-21 |
CN105765523B (en) | 2018-07-17 |
US9977676B2 (en) | 2018-05-22 |
US20150143085A1 (en) | 2015-05-21 |
KR20160085335A (en) | 2016-07-15 |
JP2016541057A (en) | 2016-12-28 |
CN105765523A (en) | 2016-07-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9792118B2 (en) | Vector processing engines (VPEs) employing a tapped-delay line(s) for providing precision filter vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods | |
US9684509B2 (en) | Vector processing engines (VPEs) employing merging circuitry in data flow paths between execution units and vector data memory to provide in-flight merging of output vector data stored to vector data memory, and related vector processing instructions, systems, and methods | |
US9880845B2 (en) | Vector processing engines (VPEs) employing format conversion circuitry in data flow paths between vector data memory and execution units to provide in-flight format-converting of input vector data to execution units for vector processing operations, and related vector processor systems and methods | |
US9977676B2 (en) | Vector processing engines (VPEs) employing reordering circuitry in data flow paths between execution units and vector data memory to provide in-flight reordering of output vector data stored to vector data memory, and related vector processor systems and methods | |
US9619227B2 (en) | Vector processing engines (VPEs) employing tapped-delay line(s) for providing precision correlation / covariance vector processing operations with reduced sample re-fetching and power consumption, and related vector processor systems and methods | |
US20150143076A1 (en) | VECTOR PROCESSING ENGINES (VPEs) EMPLOYING DESPREADING CIRCUITRY IN DATA FLOW PATHS BETWEEN EXECUTION UNITS AND VECTOR DATA MEMORY TO PROVIDE IN-FLIGHT DESPREADING OF SPREAD-SPECTRUM SEQUENCES, AND RELATED VECTOR PROCESSING INSTRUCTIONS, SYSTEMS, AND METHODS | |
EP2972968B1 (en) | Vector processing engines having programmable data path configurations for providing multi-mode vector processing, and related vector processors, systems, and methods | |
WO2014164298A2 (en) | Vector processing engines having programmable data path configurations for providing multi-mode radix-2x butterfly vector processing circuits, and related vector processors, systems, and methods | |
WO2014164931A2 (en) | Vector processing carry-save accumulators employing redundant carry-save format to reduce carry propagation, and related vector processors, systems, and methods |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14812034 Country of ref document: EP Kind code of ref document: A1 |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
REEP | Request for entry into the european phase |
Ref document number: 2014812034 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2016530912 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20167015680 Country of ref document: KR Kind code of ref document: A |