CN1731374A - A cache prefetch module and method thereof - Google Patents

A cache prefetch module and method thereof Download PDF

Info

Publication number
CN1731374A
CN1731374A CNA2005100930779A CN200510093077A CN1731374A CN 1731374 A CN1731374 A CN 1731374A CN A2005100930779 A CNA2005100930779 A CN A2005100930779A CN 200510093077 A CN200510093077 A CN 200510093077A CN 1731374 A CN1731374 A CN 1731374A
Authority
CN
China
Prior art keywords
data
cache
image subblock
control information
bus interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CNA2005100930779A
Other languages
Chinese (zh)
Other versions
CN100378687C (en
Inventor
白锋
夏煜
王浩
怀千江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vimicro Corp
Original Assignee
Vimicro Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vimicro Corp filed Critical Vimicro Corp
Priority to CNB2005100930779A priority Critical patent/CN100378687C/en
Publication of CN1731374A publication Critical patent/CN1731374A/en
Application granted granted Critical
Publication of CN100378687C publication Critical patent/CN100378687C/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Memory System Of A Hierarchy Structure (AREA)

Abstract

The invention discloses a high speed buffer pre-access module which comprises a high speed buffer bus interface, a system chip bus interface, a control logic and data buffer area, wherein the control logic is used to send the picture sub-block data reading control information to system internal memory accruing to the intact picture sub-block data control information from the processor, preserve the read picture sub-block data into the data buffer area, and transmit the picture sub-block data to the high speed buffer; the data buffer area is used to store the intact picture sub-block data from the system internal memory. The invention also discloses a high speed buffer pre-access method.

Description

A kind of cache prefetch module and method thereof
Technical field
The present invention relates to the Data Access Technology field, particularly a kind of cache prefetch module and method thereof.
Background technology
In recent years, wait for the time of reading of data, proposed to use the technology of high-speed cache in order to reduce processor.Adopt high-speed cache System on Chip/SoC relevant portion as shown in Figure 1, this System on Chip/SoC comprises processor (CPU Core), high-speed cache, cache prefetch module, Installed System Memory, System on Chip/SoC bus.Wherein processor reading of data and control information of calculating and sending prefetch data to cache prefetch module from high-speed cache; What preserve in the high-speed cache is the data that processor will use; Cache prefetch module is used for being about to from the Installed System Memory read processor data of use, and data transmission is arrived high-speed cache; In store various data in the Installed System Memory; The System on Chip/SoC bus connects described cache prefetch module and Installed System Memory, and transmits various control informations and data.
Shown in Figure 2 is the structural representation of prior art high speed cache prefetching module, and this cache prefetch module comprises System on Chip/SoC bus interface, cache bus interface, data buffer and steering logic.Wherein the System on Chip/SoC bus interface is used for the connected system chip bus, be connected with Installed System Memory by the System on Chip/SoC bus, and in order to Data transmission and control information; The cache bus interface is used to connect high-speed cache, and in order to Data transmission and control information; The data buffer is used to store the view data of looking ahead from Installed System Memory, and it is made up of register, and the data buffer size is a cache lines (cache line), can store the data of 32 bytes; Steering logic is used for the prefetched instruction according to processor, the view data of a cache lines is read in control from Installed System Memory, and deposit the data buffer in, and the data of a cache lines in the data buffer are transferred to high-speed cache by the cache bus interface.Steering logic, the System on Chip/SoC bus interface, the data buffer, also exist the mutual of control information between the cache bus interface, thereby realize the transmission of data, when for example image data transmission in the data buffer being arrived high-speed cache, steering logic produces the control information to data buffer zone and cache bus interface, data buffer and cache bus interface are determined duty mutually according to the control information of steering logic then, when the data buffer is transmission (Transfer) state, the cache bus interface is that view data is transferred to high-speed cache from the data buffer through the cache bus interface when receiving (Receive) state.
In the Digital Image Processing process, every two field picture is divided into several image subblocks usually, the size of image subblock is generally 8 * 8 or 16 * 16 data points, the image subblock of 8 * 8 data points is corresponding to usually said piece (Block), and the image subblock of 16 * 16 data points is corresponding to usually said macro block (Macro Block).If image adopts the YUV4:2:2 coded format, each data point takies a byte, and the monochrome information of image subblock is respectively 8 * 8 bytes or 16 * 16 bytes so.Shown in Figure 3 is a picture frame of being made up of 99 macro blocks, and these 99 macro blocks are divided into 9 row 11 row, be labeled as respectively macro block 0, macro block 1 ... macro block 98, the size of each image subblock are 16 * 16 bytes, i.e. 256 bytes.
Suppose the image subblock of the N of processor in the pre-treatment frame shown in Figure 3 16 * 16 byte, the image subblock of N+1 16 * 16 bytes of need looking ahead, wherein N more than or equal to and 0 smaller or equal to 98.As shown in Figure 4, the look ahead process of an image subblock of aforementioned cache prefetch module is as follows in the prior art:
Step 401, steering logic receive the control information of 32 bytes in the 1st N+1 image subblock of looking ahead of processor, and this control information has comprised the address information of first byte data in preceding 32 byte datas of described N+1 image subblock.Processor was in idle condition and waited for the release of looking ahead this moment.
Step 402, after steering logic produces the address information of all the other 31 byte datas automatically according to the address information of above-mentioned first byte data, steering logic is sent the control information of reading this 32 byte data to Installed System Memory, this control information comprises the address information of preceding 32 byte datas of reading order and this N+1 image subblock, cache prefetch module occupies the System on Chip/SoC bus and carries out the 1st time and look ahead then, read preceding 32 byte datas of this N+1 image subblock, store the data buffer then into.Steering logic is transferred to high-speed cache with 32 byte datas of looking ahead for the 1st time more then.
This moment, processor still was in idle condition, waited for the data of N+1 image subblock.
Step 403, steering logic receives the control information of the 2nd N+1 the image subblock of looking ahead of processor, this control information has comprised the address information of first byte data in the 2nd 32 byte datas of described N+1 image subblock, and processor is in idle condition.
Step 404, after steering logic produces the address information of all the other 31 byte datas automatically according to the address information of above-mentioned first byte data, steering logic is sent the control information of reading this 32 byte data to Installed System Memory, this control information comprises the address information of the 2nd 32 byte datas of reading order and described N+1 image subblock, cache prefetch module occupies the System on Chip/SoC bus and carries out the 2nd time and look ahead then, read the 2nd 32 byte datas of this N+1 image subblock, store the data buffer then into.Steering logic is transferred to high-speed cache with 32 byte datas of looking ahead for the 2nd time more then.
At this moment, processor still is in the state of waiting for N+1 image subblock data.
Step 405, repeating step 403 be to step 404, finish to be prefetched to the 8th time for the 3rd time and to look ahead, and 32 byte datas of at every turn looking ahead, thus finish looking ahead to N+1 all 256 byte datas of image subblock.During this period, processor is in waiting status always.
Step 406, processor read and handle the data of N+1 image subblock from high-speed cache.
Look ahead flow process as can be seen from top, a cache lines is 32 byte datas because existing cache prefetch module can only be looked ahead at every turn, when image subblock data the time greater than 32 bytes, need repeatedly carry out just can the look ahead data of an image subblock of prefetch operation, processor is in waiting status for a long time during this, thereby causes image processing speed to slow down.
Summary of the invention
In view of this, the present invention proposes a kind of cache prefetch module and method thereof, in order to reduce the time that processor is waited for during view data is looked ahead at image subblock.
According to above-mentioned purpose, the invention provides a kind of cache prefetch module, comprise cache bus interface, System on Chip/SoC bus interface, steering logic and data buffer at least, wherein the cache bus interface is used to connect high-speed cache; The System on Chip/SoC bus interface is used for the connected system internal memory; Interconnective steering logic and data buffer are connected between cache bus interface and the System on Chip/SoC bus interface, described steering logic is used for the control information according to the sub-blocks of data of complete image of looking ahead of from processor, send the control information of reading the sub-blocks of data of described complete image to Installed System Memory, and the sub-blocks of data of the complete image that is read all is saved in described data buffer, and the sub-blocks of data of the complete image in the described data buffer is transferred to high-speed cache by a read operation; The data buffer is used to store the described sub-blocks of data of complete image that reads from Installed System Memory.
In technique scheme, described data buffer is made of register file.
Described data buffer capacity is more than or equal to the size of an image subblock.
Described control information of looking ahead the sub-blocks of data of complete image comprises the address information of first byte data in the described image subblock; Described steering logic is further used for producing according to the address information of first byte data in the described image subblock address information of remainder data in the described image subblock, and sends the control information of the described image subblock data that comprise described all data address information of image subblock to Installed System Memory.
The present invention also provides a kind of cache prefetch method, and this method may further comprise the steps:
A. receive processor and look ahead after the control information of the sub-blocks of data of complete image, cache prefetch module sends the control information of reading described image subblock data to Installed System Memory, and reads complete described image subblock data by a read operation;
B. described image subblock data are saved in the cache prefetch module;
C. described image subblock data transmission is arrived high-speed cache.
Described control information of looking ahead the sub-blocks of data of complete image comprises the address information of first byte data in the described image subblock; Cache prefetch module described in the steps A further comprised before Installed System Memory sends the control information of reading described image subblock data: the address information of first byte data produces the address information of remainder data in the described image subblock in according to described image subblock; Described cache prefetch module sends the address information that the control information of reading described image subblock data comprises all data in the described image subblock to Installed System Memory.
Further, comprise that processor reads the step of described image subblock data from described high-speed cache after the step C.
From such scheme as can be seen, the sub-blocks of data of complete image because cache prefetch module support of the present invention is once looked ahead, shortened the cycle of the image subblock data of looking ahead, reduced the time that processor is waited for during view data is looked ahead, thereby improved image processing speed.
Description of drawings
Fig. 1 is the position view of cache prefetch module in system;
Fig. 2 is the structural representation of prior art high speed cache prefetching module;
Fig. 3 is an image that comprises 11 * 9 image subblocks;
Fig. 4 is the schematic flow sheet of one the 256 sub-piece of byte image of looking ahead in the prior art;
Fig. 5 is the structural representation of cache prefetch module of the present invention;
Fig. 6 is the look ahead schematic flow sheet of one the 256 sub-piece of byte image of the present invention.
Embodiment
For making the purpose, technical solutions and advantages of the present invention clearer, the present invention is described in more detail by the following examples.
The cache prefetch module that the present invention proposes is the same with prior art in the position of digital image processing system, but the concrete structure of the cache prefetch module that the present invention proposes and function are unlike the prior art.
With reference to figure 5, cache prefetch module of the present invention comprises System on Chip/SoC bus interface, cache bus interface, data buffer and steering logic.Wherein the System on Chip/SoC bus interface is used for the connected system chip bus, be connected with Installed System Memory by the System on Chip/SoC bus, and Data transmission and control information.The cache bus interface is used to connect high-speed cache, and Data transmission and control information.Steering logic and data buffer are connected between System on Chip/SoC bus interface and the cache bus interface, and steering logic and data buffer interconnect between the two.The data buffer is used to store the view data of looking ahead from Installed System Memory, it is made up of register, unlike the prior art be, data buffer of the present invention size is at least the size of an image subblock, according to existing Mpeg-1, Mpeg-2, Mpeg-4, H.261, H.263 or H.264 wait the encoding and decoding treatment technology to know that the size of image subblock is generally 64 bytes or 256 bytes, in order not cause redundant waste, the size of the needs data buffer that can handle according to real image is set to 64 bytes or 256 bytes, and promptly the data buffer is made up of the register file that can store 64 bytes or 256 bytes.Compared with prior art, be equivalent to newly add register on the structure in the data buffer.The prefetched instruction of steering logic receiving processor, after the address information of first byte data produces the address information of remainder data of an image subblock automatically in the image subblock that sends over according to processor, send the control information of reading these image subblock data to Installed System Memory by the System on Chip/SoC bus interface, this control information comprises the address information of all data in reading order and the above-mentioned image subblock, from Installed System Memory, read the data of an image subblock, and deposit the data buffer in, and the data of an image subblock in the data buffer are transferred to high-speed cache by the cache bus interface.Same as the prior art, steering logic, the System on Chip/SoC bus interface, the data buffer, also exist the mutual of control information between the cache bus interface, thereby realize the transmission of data between each module, when for example image data transmission in the data buffer being arrived high-speed cache, steering logic produces the control information to data buffer zone and cache bus interface, data buffer and cache bus interface are determined duty mutually according to the control information of steering logic then, when the data buffer is transmission (Transfer) state, the cache bus interface is that view data is transferred to high-speed cache from the data buffer through the cache bus interface when receiving (Receive) state.
Suppose the image subblock of the N of processor in the pre-treatment frame shown in Figure 3 16 * 16 byte, the image subblock of N+1 16 * 16 bytes of need looking ahead, wherein N more than or equal to and 0 smaller or equal to 98.As shown in Figure 6, the look ahead process of an image subblock of high speed cache prefetching module of the present invention is as follows:
Step 601, steering logic receive the look ahead control information of N+1 complete image subblock of processor, and this control information comprises the address information of first byte data in described N+1 the image subblock.The image subblock because cache prefetch module support of the present invention is once looked ahead, so processor can send the control information of the image subblock of looking ahead to this cache prefetch module.Then, steering logic produces the address information of all the other 255 byte datas in this image subblock according to the address information of first byte data in described N+1 the image subblock.The method possibility that the steering logic of different vendor produces the remainder data address information is different, the simplest method is exactly to add one to 255 fixed numeric values on the address information of first byte data successively, thereby produces the address information of remaining 255 byte data.
Step 602, steering logic is sent the control information of reading this image subblock to Installed System Memory, this control information comprises the address information of 256 byte datas of reading order and this N+1 image subblock, cache prefetch module occupies the System on Chip/SoC bus and looks ahead then, read 256 byte datas of this N+1 image subblock, store the data buffer into.Steering logic 256 byte datas of again will this time looking ahead are transferred to high-speed cache then.
At this moment, the data of current N the image subblock of processor processing are perhaps handled the data of N image subblock, are in the state of waiting for N+1 image subblock data.If processor is being handled the data of N image subblock, so after the data of intact N the image subblock of processor processing, just can directly from high-speed cache, read N+1 image subblock data, thereby do not need to wait for, compare with prior art, reduce the time that processor is waited for, improved the efficient of Flame Image Process.If processor has been handled the data of N image subblock, and be in the state of waiting for N+1 image subblock data, so after this time of cache prefetch module end looked ahead, just can from high-speed cache, read the data of N+1 image subblock, compare with prior art, reduce the time that processor is waited for equally, improved the efficient of Flame Image Process.
Step 603, processor read and handle the data of N+1 image subblock from high-speed cache.
The above only is preferred embodiment of the present invention, and is in order to restriction the present invention, within the spirit and principles in the present invention not all, any modification of being done, is equal to replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (7)

1, a kind of cache prefetch module comprises cache bus interface, System on Chip/SoC bus interface, steering logic and data buffer at least, and wherein the cache bus interface is used to connect high-speed cache; The System on Chip/SoC bus interface is used for the connected system internal memory; Interconnective steering logic and data buffer are connected between cache bus interface and the System on Chip/SoC bus interface,
It is characterized in that, described steering logic is used for the control information according to the sub-blocks of data of complete image of looking ahead of from processor, send the control information of reading the sub-blocks of data of described complete image to Installed System Memory, and the sub-blocks of data of the complete image that is read all is saved in described data buffer, and the sub-blocks of data of the complete image in the described data buffer is transferred to high-speed cache by a read operation;
The data buffer is used to store the described sub-blocks of data of complete image that reads from Installed System Memory.
2, cache prefetch module according to claim 1 is characterized in that, described data buffer is made of register file.
3, cache prefetch module according to claim 1 is characterized in that, described data buffer capacity is more than or equal to the size of an image subblock.
4, cache prefetch module according to claim 1 is characterized in that, described control information of looking ahead the sub-blocks of data of complete image comprises the address information of first byte data in the described image subblock;
Described steering logic is further used for producing according to the address information of first byte data in the described image subblock address information of remainder data in the described image subblock, and sends the control information of the described image subblock data that comprise described all data address information of image subblock to Installed System Memory.
5, a kind of cache prefetch method is characterized in that, this method may further comprise the steps:
A. receive processor and look ahead after the control information of the sub-blocks of data of complete image, cache prefetch module sends the control information of reading described image subblock data to Installed System Memory, and reads complete described image subblock data by a read operation;
B. described image subblock data are saved in the cache prefetch module;
C. described image subblock data transmission is arrived high-speed cache.
6, method according to claim 5 is characterized in that, described control information of looking ahead the sub-blocks of data of complete image comprises the address information of first byte data in the described image subblock;
Cache prefetch module described in the steps A further comprised before Installed System Memory sends the control information of reading described image subblock data: the address information of first byte data produces the address information of remainder data in the described image subblock in according to described image subblock;
Described cache prefetch module sends the address information that the control information of reading described image subblock data comprises all data in the described image subblock to Installed System Memory.
7, method according to claim 5 is characterized in that, comprises that further processor reads the step of described image subblock data from described high-speed cache after the step C.
CNB2005100930779A 2005-08-25 2005-08-25 A cache prefetch module and method thereof Expired - Fee Related CN100378687C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CNB2005100930779A CN100378687C (en) 2005-08-25 2005-08-25 A cache prefetch module and method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CNB2005100930779A CN100378687C (en) 2005-08-25 2005-08-25 A cache prefetch module and method thereof

Publications (2)

Publication Number Publication Date
CN1731374A true CN1731374A (en) 2006-02-08
CN100378687C CN100378687C (en) 2008-04-02

Family

ID=35963730

Family Applications (1)

Application Number Title Priority Date Filing Date
CNB2005100930779A Expired - Fee Related CN100378687C (en) 2005-08-25 2005-08-25 A cache prefetch module and method thereof

Country Status (1)

Country Link
CN (1) CN100378687C (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102110147A (en) * 2011-02-16 2011-06-29 北京数码大方科技有限公司 Cache-based CAD file data unloading method and device
CN101634971B (en) * 2009-09-01 2011-07-06 威盛电子股份有限公司 Method, device and computer system for extracting data in advance
CN103034455A (en) * 2012-12-13 2013-04-10 东南大学 Method and system for managing data information buffer based on pre-decoding and analyzing
CN103077129A (en) * 2012-12-31 2013-05-01 上海算芯微电子有限公司 Information processing method and device
CN103729142A (en) * 2012-10-10 2014-04-16 华为技术有限公司 Method and device for pushing stored data
CN104899824A (en) * 2014-03-05 2015-09-09 珠海全志科技股份有限公司 Processing method and system for image data in DRAM
CN107168660A (en) * 2016-03-08 2017-09-15 成都锐成芯微科技股份有限公司 Image procossing caching system and method
CN108701159A (en) * 2016-12-13 2018-10-23 谷歌有限责任公司 System and method for prefetching content item

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101526924B (en) * 2009-04-22 2010-09-08 东南大学 Method for accessing optimal digital signal processing chip data

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6754780B1 (en) * 2000-04-04 2004-06-22 Hewlett-Packard Development Company, L.P. Providing data in response to a read command that maintains cache line alignment
US6725337B1 (en) * 2001-05-16 2004-04-20 Advanced Micro Devices, Inc. Method and system for speculatively invalidating lines in a cache
US7089368B2 (en) * 2002-02-12 2006-08-08 Ip-First, Llc Microprocessor apparatus and method for exclusively prefetching a block of cache lines from memory
US7089371B2 (en) * 2002-02-12 2006-08-08 Ip-First, Llc Microprocessor apparatus and method for prefetch, allocation, and initialization of a block of cache lines from memory
US6895475B2 (en) * 2002-09-30 2005-05-17 Analog Devices, Inc. Prefetch buffer method and apparatus
US7165146B2 (en) * 2003-01-28 2007-01-16 Sun Microsystems, Inc. Multiprocessing computer system employing capacity prefetching
US7055005B2 (en) * 2003-04-07 2006-05-30 Hewlett-Packard Development Company, L.P. Methods and apparatus used to retrieve data from memory into a RAM controller before such data is requested
CN100489813C (en) * 2003-04-21 2009-05-20 智慧第一公司 Method for selective prefetch withdraw

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101634971B (en) * 2009-09-01 2011-07-06 威盛电子股份有限公司 Method, device and computer system for extracting data in advance
CN102110147B (en) * 2011-02-16 2012-10-10 北京数码大方科技有限公司 Cache-based CAD file data unloading method and device
CN102110147A (en) * 2011-02-16 2011-06-29 北京数码大方科技有限公司 Cache-based CAD file data unloading method and device
CN103729142B (en) * 2012-10-10 2016-12-21 华为技术有限公司 The method for pushing of internal storage data and device
US9632938B2 (en) 2012-10-10 2017-04-25 Huawei Technologies Co., Ltd. Method and apparatus for pushing memory data
CN103729142A (en) * 2012-10-10 2014-04-16 华为技术有限公司 Method and device for pushing stored data
WO2014056329A1 (en) * 2012-10-10 2014-04-17 华为技术有限公司 Memory data pushing method and device
CN103034455A (en) * 2012-12-13 2013-04-10 东南大学 Method and system for managing data information buffer based on pre-decoding and analyzing
CN103034455B (en) * 2012-12-13 2015-09-16 东南大学 Based on data message buffer memory management method and the system of Decoding Analysis in advance
CN103077129B (en) * 2012-12-31 2016-07-13 上海算芯微电子有限公司 Information processing method and device
CN103077129A (en) * 2012-12-31 2013-05-01 上海算芯微电子有限公司 Information processing method and device
CN104899824A (en) * 2014-03-05 2015-09-09 珠海全志科技股份有限公司 Processing method and system for image data in DRAM
CN107168660A (en) * 2016-03-08 2017-09-15 成都锐成芯微科技股份有限公司 Image procossing caching system and method
CN108701159A (en) * 2016-12-13 2018-10-23 谷歌有限责任公司 System and method for prefetching content item
CN108701159B (en) * 2016-12-13 2022-05-13 谷歌有限责任公司 System and method for prefetching content items

Also Published As

Publication number Publication date
CN100378687C (en) 2008-04-02

Similar Documents

Publication Publication Date Title
CN1731374A (en) A cache prefetch module and method thereof
CN1859325A (en) News transfer method based on chained list process
JP5603507B2 (en) Accelerating cache state migration on directory-based multi-core architectures
CN101048762A (en) Method and apparatus for transmitting memory pre-fetch commands on a bus
CN101031897A (en) Establishing command order in an out of order dma command queue
CN104937931A (en) Method and apparatus using software engine and hardware engine collaborated with each other to achieve hybrid video encoding
CN1632771A (en) Direct memory access control device and image processing system and transmission method
US8736629B1 (en) System and method for an efficient display data transfer algorithm over network
CN1825295A (en) Data transmission control device, image processing unit, and data transmission control method
CN101365138B (en) JPEG2000 image compression processing system
CN1851677A (en) Embedded processor system and its data operating method
CN1286038C (en) Dual-ring quene-based, non-interrupt PCI communication method
CN1762133A (en) Pre-empting low-priority traffic with high-priority traffic
CN1922591A (en) Intelligent pci bridging consisting of prefetching all data prior to sending data to requesting service
CN1834946A (en) I/O controller, signal processing system, and method of transferring data
CN1308840C (en) Method for acquisition of data in hard disk
CN1757018A (en) Data processing system with prefetching means
CN101527844A (en) Method for block execution of data to be decoded
CN110059024B (en) Memory space data caching method and device
CN1650259A (en) Integrated circuit with a non-volatile memory and method for fetching data from said memory
US8700859B2 (en) Transfer request block cache system and method
US8514237B2 (en) Two dimensional memory caching apparatus for high definition video
CN101308568B (en) Method and apparatus for production line real-time processing based FIFO
CN1298049C (en) Graphic engine chip and its using method
CN1484157A (en) Embedding system and instruction prefetching device and method thereof

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C17 Cessation of patent right
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20080402

Termination date: 20120825