GB2330476A - Adaptively-coded digital video signal recording equipment - Google Patents

Adaptively-coded digital video signal recording equipment Download PDF

Info

Publication number
GB2330476A
GB2330476A GB9828342A GB9828342A GB2330476A GB 2330476 A GB2330476 A GB 2330476A GB 9828342 A GB9828342 A GB 9828342A GB 9828342 A GB9828342 A GB 9828342A GB 2330476 A GB2330476 A GB 2330476A
Authority
GB
United Kingdom
Prior art keywords
data
picture
playback
gop
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB9828342A
Other versions
GB9828342D0 (en
GB2330476B (en
Inventor
Shuji Sotoda
Satoshi Kurahashi
Takahiro Nakai
Tadashi Kasezawa
Masato Nagasawa
Hiroyuki Oohata
Yoshinobu Ishida
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Corp
Original Assignee
Mitsubishi Electric Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP22962094A external-priority patent/JP3506504B2/en
Priority claimed from JP25209894A external-priority patent/JP3506507B2/en
Priority claimed from JP27210794A external-priority patent/JP3501521B2/en
Priority claimed from JP02827795A external-priority patent/JP3370468B2/en
Application filed by Mitsubishi Electric Corp filed Critical Mitsubishi Electric Corp
Priority claimed from GB9519553A external-priority patent/GB2293516B/en
Publication of GB9828342D0 publication Critical patent/GB9828342D0/en
Publication of GB2330476A publication Critical patent/GB2330476A/en
Application granted granted Critical
Publication of GB2330476B publication Critical patent/GB2330476B/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/115Selection of the code volume for a coding unit prior to coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/149Data rate or code amount at the encoder output by estimating the code amount by means of a model, e.g. mathematical model or statistical model
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/154Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/177Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a group of pictures [GOP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/18Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/184Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • H04N21/42646Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/84Television signal recording using optical recording
    • H04N5/85Television signal recording using optical recording on discs or drums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/78Television signal recording using magnetic recording
    • H04N5/782Television signal recording using magnetic recording on tape
    • H04N5/783Adaptations for reproducing at a rate different from the recording rate
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/806Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
    • H04N9/8063Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • Algebra (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

A video signal recording device for recording a digital video signal coded by using motion compensation prediction and orthogonal transformation(e.g.MPEG), includes means for adaptively changing the data rate of the coding with respect to the picture pattern. The coding data rate information is multiplexed with other additional information(e.g. audio and error correction code data) which is added to the data to be recorded.

Description

c
TITLE OF THE INVENTIOj,Z
Digital Video Signal Recording and Playback Device BACKGROUND OF INVENTION
Field of the Invention
2330476 The present invention relates to a digital video signal recording and playback device for recording and playing back a digital video signal, and more particularly to a digital video signal recording and playback device for recording and playing back on a medium such as an optical disc or the like a digital video signal coded on the basis of a motion compensation prediction and an orthogonal conversion.
Description of Related Art
FIG. 1 is a block circuit diagram of a conventional optical disc record and playback device shown in the Japanese Patent Application Laid-Open No.
HEI 4-1143)69 (1992). Referring to FIG. 1, reference numeral 201 denotes an AID converter for converting a video signal, an audio signal or the like into digital information. Reference numeral 202 denotes a data compressing circuit, 203 a frame sector converting circuit for converting compressed data into sector data which is equal to integer times of a frame cycle, 204 an error correction coder for adding 1 1 1 \ 1 the error correction signal to sector data, 205 a modulator for modulating inter..LEerence between codes in a recording medium into a predetermined modulation code to reduce the interference, 206 a laser driving circuit for modulating laser light in accordance with a mcdulati. on code, and 207 a laser output switch. Further, reference numeral 208 denotes an optical head for emitting laser light, 209 an actuator for tracking a light beam emitted from the optical head 208, 210 a traverse motor for sending the optical head 208, 211 a disc motor for rotazing. an optical disc 212, 219 a motor driving circuit, 220 a first control circuit and 221 a second control circuit. Further, reference numeral 213 denotes a playback amplifier for amplifying a playback signal from the optical head 208. Reference numeral 214 denotes a demodulator for obtaining data from a recorded modulation signal, 215 an error correction decoder, 216 a frame sector inverse converting circuit, 217 a data extending circuit for extending the compressed data, 218 a D/A converter for converting extended data into, for example, an analog video signal and an audio signal.
FIG. 2 is a block circuit diagram showing an inside structure of the data compressing circuit 202 in FIG. 1. In FIG. 2, a digital video signal inputted from the A/D converter 201 is inputted into a memory circuit '001 - A video sip, al 321 outputted from the memory circuit 301 1) 4 is given to a first input of a subzraczer 302 and a second input of a motion compensation predicting circuit 310. An output of the subtracter 302 is!I-iputted to a quantizer 304 via a DCT (discrete cosine transform) circuit 303. An output of the quantizer 304- is given to an -fer 306 via a variable-length encoder 305.
input of a transmission bull. 1 An output of the transmission buffer 306 is outputted. to the frame sector converting circuit 203. In The meantime, an output oi the Quantizer 304- is inputted to the inverse DCT circuit 308 via an inverse quantizer 3071. An output off the inverse DCT circuit 308 is given to a first input 01 an adder 309. JAn Output 322 of the adder 309 is given to a first input of a motion compensation predicting circuit 310. An output 323 of the motion compensation predicting circuit 310 is given to a second input of the adder ')09 and a second input of the subtracter 302.
FIG. 3 is a block circuit diagram showing an inside structure of the motion compensation predicting circuit 310 in FIG. 2. In FIG. 3, the output 322 of the adder 309 is given to an input terminal 401a while the output 321 of the memory circuit 301 is given to an input terminal 401b.
The signal 322 inputted from the input terminal 401a is inputted to a frame memory 404a or a frame memory 4041b via a switch 403. A reference picture outputted from the frame memory 4041a is given to a first input of 3 0 a motion vector detecting circuit 405a. The video signal 321 inputted from the input terminal 401b is inputted to a second input ot the motion vector detecting ci,-cu,it 405a. An output of the motion vector detecting circuit 405Sa. is inputted to a prediction mode selector 406. In the is me.;-nntlme, the reference picture outputted from the frame memory IX0141b given to a first input of a motion vector detecting circuit 4050. The video signal 321 inputted from the input terminal 401b is given to a second input of the motion vector detecting circuit 405b. The output of the motion vector detecting circuit 405b is given to the second input of the prediction mode selector 406. The video signal 321 inputted from the input terminal Olb is given to a third input of the prediction mode selector 406. A zero signal is given to a second input of a swizcher 407. A second output of the prediction mode selector 406 is given to a third input of the switcher 407. The output 323 of the switcher 407 is 011tputted from a output terminal 402.
FIG. 4 is a block circuit diagram showing an inside structure of the data extending circuit 217 in FIG. I. In FIG. 4, the video signal inputted from the frame sector inverse converting circuit 216 is inputted to a reception buffer 501. An output from the reception buffer 501 is inputted to a variable-length decoder 502, and the output therefrom is inversely quantized at an inverse quantizer 503. Then, the output is 1 4 subjected to an inverse discrete cosine transform at an inverse DCT circuit SG4. The output is given to a first input of an adder 506. In the meantime, the output of the reception buffer SO1 is given to a prediction data decoding circuit 'DOS while an output of the prediction data decoding circuit 505 is given to a second output of the adder 506. The output of the adder 506 is outputted to the D/A converter 218 via a memory circuit 5071.
ext, an operation of the device will be explained. As one hgh efficiency coding mode in the case of coding a video signal, there is an coding a2aorithm by means of a M11EG (Aoving Picture Expert Group) mode. This is a hybrid coding mcde which combines an inter-frame prediction coding using a motion compensation prediction and an intra-frame conversion coding. This conventional example uses a data compressing circuit 202 having a structure shown in FIG. 2 and adopts the aforementioned MPEG mode.
FIG. 5 shows a simplified data arrangement structure (layer structure) c::' ICEG mode. In FIG. 5, reference numeral 621 denotes a sequence layer comprising a group of pictures (hereinafter referred to as "GOP---) comprising a plurality of frame data items, 622 a GOP layer comprising several pictures (screens), 623 a slice which divides one screen into several blocks, 624 a slice layer which has several 0 macroblocks, 625 a macroblock layer, 626 a block layer which consists of 8 pixels X 8 pixels.
This macroblock layer 625 is a block which consists of a least unl.' of 8 pixels x 8 pixels, for example, in the YEG mode. This block is a unit for performing DCL At this time, a Total cl' 6 blocks, including adjacent four Y signal blocks, one Cb block which corresponds to the Y signal blocks in position, and one Cr block are referred to as macroblocks. A plurality 0, these macroblocks constitute a slice. In addition, the macroblocks constitute a minimum unit of a motion compensation prediction, and a motion vector for the motion compensation prediction is formed in macroblock units.
Subsequently, a process for the inter-frame prediction coding will be explained. FIG. 6 shows an outline of the inter-frame prediction coding. Pictures are divided into three types, namely an intra-frame coded plcLure (hereinafter referred to as an I picture), a one direction prediction coded plc-lure (hereinafter referred to as a P picture), and a both direction prediction coded picture (hereinafter refferred to as a B Picture).
For example, in the case where one picture out of IN pictures is set 'to I picture, one picture out of M pictures is set to P picture or I picture, (. X n -)-Lh picture constitutes an I picturee, (N X n - M X 6 1 4 m)-Lh picture (m _ 1) constitutes a P picture, pictures trom (N X n M x m 1)th picture to C X n - A X m M - 1)th picture constitute pictures, where n n-nd m are integers and Az this time, pictures from GN x n - 1)th picture to (N X n - N,.)T- picture are reterred to as a GOP (group c-ic pictures) in summan;- FIG. 6 shows a case in which symbols N and M are defined as X=15 and =3. In FTG. 6, z-e i picture is riot subjected to the interframe prediction buz only to the intra-frame conversion coding. The P picture is subjected to a prediction -from the I picture immediately before the P picture or from the P picture. For example, the 6th picture in FIG. 6 is a P pictu7e. The 6th picture is subjected to the prediction from the 3rd I picture. Further, the 9th P picture in FIG. 6 is subjected to the prediction from the 6th P picture. The B picture is subjected to the prediction from I picture or the P plc-Lure immediately before and after the B picture. For example, in FIG. 6, the 4th and 5th B pictures is subjected to the prediction both from the 3rd I picture and the 6th P picture. Consequently, the 4th and Sth pictures is subjected to coding after coding of the 6th picture.
Then, an operation of a data compressing circuit 202 will be explained in accordance with FIG. 2. The memory circuit 301 outputs the digital video picture signals which are inputted a-fter rearranging the i signals in -L'Ille coding order. In other words, as described above, for example, the first S picture is coded after the 3rd I picture in FIG. 6. Consequently, the order of pictures are rearranged. FIG. 7 shows an operation of this rearrangement. An picture sequence inputted as shown in FIG. 7A is outputted in the order shown in FIG. 7B.
Further, the video signal 321 outpuzzed from the memory circuit 301 is subjected to DC1 in the direction of space axis after a difference between pictures from the prediction picture 323 outputted from the motion compensation predicting circuit 310 at the subtraczer 302 to reduce the redundancv in the direction of the time axis. The converted coefficient is quantized and variable-length coded followed by being outputted via the transmission buffer 306. In the meantime, the quantized conversion coefficient is inversely Quantized and is subjected to an inverse DCL After that, the coefficient is added to zn'e prediction picture 323 at the adder 309 and a decoded picture 322 is obtained. The decoded picture 322 is inputted to the motion compensation predicting circuit 310 for the subsequent coding of pictures.
Subsequently, an operation of the motion compensation predicting circuit 310 will be explained in accordance with FIG. 3. The motion compensation predicting circuit 310 uses two reference pictures which are stored in the frame memory 44a and the frame memory 44b to 8 perform a motion comperi--.-atlOrl prediction of the video signal 321 outpuzted from the memory circuit 301 for outpurting the prediction picture 323.
In the beginning, in the case where the picture 39-9- ccded and decoded as described above is either an I picture or a P picture, this picture 322 is stored in the frame memory 404-a or the frame memory 404-b for coding the subse-quent picture. At this time, the switcher 403 is switched so that the frame memory out of the two frame memories 404a and 404b which is renewed prior to the other in time is selected. However, when the decoded picture 322 is a B picture, writing is not performed at the frame memory 40-1a and the frame memory 404.b.
For example, when lst and 2nd plc-Lures in FIG. 7 are coded bv such switching of the switcher 403, the Oth P picture and the 3rd 1 picture are stored in the frame memory 404a and frame memory 404b, respectively. Further, when the 6th P picture is coded and decoded, the frame memory 404a is rewritten into the decoded picture of the 67-h P picture.
Consequently, when the 4th and the Sth B pictures are coded, the eth P picture and the 3rd I picture are stored in the frame memories 404a and 4CL111b, respectively. Further, when the 9th P picture is coded and decoded, the frame memory 4041b is rewritten into the decoded picture Of the 9th P picture. As a consequence, when the 7th B picture and the 9 0 8th B picture are coded, the 6th P picture and the gzh P picture are stored in the frame memories 4011a and 404b, respectively.
Then the video signal 321 outputted from the memory circuit 301 is inputted to The motion compensation predicting circuit 310, the motion vector detecting circuits 405)a and 405b detect a motion vector on th e basis of a reference picture stored in the frame memories 404a and 404b and output a motion compensation prediction picture. in other words, the video signal 321 is divided into a plurality of blocks. Then a block is selected so that the prediction distortion becomes zlhe smallest in the reference Picture with respect to each block. Then the relative position of the block is outputted as the motion block, and at the same time this block is output-Led as the mozion compensation prediction picture.
In the meantime, the prediction mode selector 406 selects a picture where the prediction distortion is the smallest out of two IDO Zion compensation prediction pictures outputted from the motion vector detecting circuits 405a and 405-b or an average picture thereof. Then, Lhe selected picture is outputted as a predicted p'cture. At this time, when the video signal 321 is not a B picture, the wilon compensation prediction picture is always selected and outputted which corresponds to the reference picture which is inputted prior to the other before in 4 time. Further, the Prediction mode selector 406 selects either coding in pictures in which prediction is not performed or prediction coding by the selected prediction picture in such a manner that the selected coding has a better coding efficiency.
At this t-me, when the video signal 321 is an I picture, the coding in pictures is always selected. When the coding in pictures is selected, a signal representative of the coding in the picture mode is outputted as a prediction mode. In the meantime, when the prediction coding betln;een pictures is selected, a signal representative of a selected prediction picture is outputted as a prediction mode. The switcher 4071 outputs a zero signal when the prediction mode outputted from the prediction mode selector 406 is a mode of coding in pictures. If the prediction mode is not the mode of coding in pictures, the prediction mode selector 406 outputs the prediction picture.
It follows from the aforementioned fact that when the video signal 321 outputted from the memory circuit 301 is an I picture, the motion compensation predicting circuit 310 always outputs the zero signal as a prediction picture 323, the I picture is not subjected to the inter-frame pred-iction but to the intra-frame conversion coding. In the meantime, when the video signal output-Led from the memory circuit 301 is the 6th P picture in FIG. 6, the motion compensation prediction 1 1 circuit 310 performs the motion compensation prediction from the 3rd I picture in FIG. 6 and outputs the prediction picture 323. Further, when the video signal 321 outpuzzed from the memory circuit 301 is the 4th B Picture shown in FIG. 6, the motion compensation prediction circuit 310 performs the motion compensation prediction from the 3rd I picture and The 6th P picture shown in FIG. 6 and outputs the prediction picture 323.
Subsequently, an operation of the transmission buffer 306 will De explained. The transmission buffer 306 converts video data variablelength coded by the variable-length encoder 305 into a bitstream of the Y.EG video signal. Here, the stream of the VEG has a six layer structure shown in FIG. 5. Header information which is an ident if 1' cation code is added for a sequence layer 621, a GOP layer 622, a picture layer 623, a slice layer 624 and a block layer 626 to constitute the layer structure.
Further, the transmission buffer 306 decomposes a bitszream of a video signal and a blitstream of an audio signal into a plurality of packets respectively so that these packets are multiplexed including a synchronization signal thereby constituting a svstem stream of a VEG2PS (program stream). Here, the CEG2-PS consists of a pack layer and a packet layer as shown in FIG. 8. Then the header information is added to the packet layer and The Pack layer. In the conventional example, a 19 system stream is constituted so that data of one GOP portion o the video data is includedi.
Here, the pack laver has a structure in which the packet layer is bound at the upper layer of the packet layer. Each packet layer which constitutes the pack layer is referred to as a PES packet. In addition, header information c-E the pack layer shown in FIG. 8 includes an identification signal of a pack and a synchronous signal which constitutes a basis of a video signal and an audio signal.
In the meantime, in the packet which constitutes the packet layer, three kinds of PES packets exist as show-i-i in FIG. 9. Here, a second stag= packet shown n FIG. 9 is a v-deo/aud'o/pr'vate 1 packet 1 1 1 wherein a code for identifying the front of the packet and time stamp information or the like (PTS) and DTS) needed at the time of decoding each packet as header "information are added before the packet data.
However, the time stamp information PTS is a time control information of the reproduction output and is information for controlling a decoding order of data stream of each packet at the time of reproduction. Further, DTS is time control information at the start of decoding and is information for controlling the transmission order of decoding data.
I"he third stage packet shown in FIG. 9 is a private 2 packet where user data is written. Further, the lowest stage packet is a padding 1 n 1) packet where all the packet data is masked with 4T" The header cl information in the private 2 packet and pad'ng packet, is constituted cl' a start code of a packet and a packet length.
As described above, the videc data and audio data items are converted into a system stre,'-,m of the YEG2-PS by the transmission buffer 306 and is converted for each of the frame sectors. This information is subjected to error correction processing, and at the same time, the information is modulated to minimize the interference between codes on the disc and is recorded on the optical disc 212. At this time, for example, data amount for each cl' the GOP unit is set to an approximately the same amount. Then, it is apparent that the edition for each of the GOP unit can be made by distributing the data into sectors which are equal To integer times of the frame cycle.
SubseQuenzly, an operation at the time of playback will be explained. At the time cl' the playback, the video information recorded on the optical disc 212 is amplified with the playback amplifier 213. After the information is restored to digital data a- the demodulator 214 and Lhe error correction decoder 21D followed by being restored as pure original video data free of data such as an address and a parity at the frame sector inverse converting circuit 216. Then, the data is inputted into the data extending circuit 2171 which has a structure shown in FIG.
1 c 4. The system stre-am which consists of a MPEG2-PS is inputted to the transmission buffer 501.
At the transmission buffer 501, the system stream which is input-Led is decomposed into a pack unit. After that, each PES packet is decomposed in accordance with the header information -Lliereby reconstructing the bitstream of the video data and audio data which is decomposed in the PES packet unit. Further, with respect to the video data, the stream is decomposed to the block laver show-ti in FIG. 5 so that the block data and the motion vector data is decomposed and outputted.
The block data outputted from the transmission bufter 501 is inputted in accordance with the variable-length decoder 502 so -hat the variablelength data becomes a fixed length data, inversely quanzized and is subjected to the inverse DCT to be outputted to the adder 506. In the meantime, the prediction data decoding circuit 505 decodes the prediction picture in accordance with the motion vector outputted from the transmission buffer 501 to be outputted to the adder 506.
In this case, the prediction data decoding circuit 505, like the motion compensation predicting circuit 310, provides a frame memory lor storing the I picture and P picture data which is decoded by the adder 506. Incidenzally, with respect to a method for renewing the 0 reference picture data an explanation will be omitted because the method is the s;-:me as the case of coding the data.
The adder 506 adds the output of the Pred-iction data decoding circuit 505) and the output of the inverse KT circuit 504 to be outputted to zine memory circuit 507. Here, at the time of coding the data, the frame is rearranged in accordance with the order of coding the data as shown in FIG. 7 with respect to the video signals which are continuous in time. Therefore, in the memory circuit 507, the data inputted in the order shown in FIG. 7B is rearranged so that the picture data continues in time and is outputted to the D/A converter 218.
Subsequently, The picture retrieval and the high speed playback thereof will be shown in the case where data with such a coding structure is recorded on the optical disc. In the case where the coding structure shown in FIG. 6 is provided, the hi&h speed playback of the picture can be performed when the data is played back in the unit of the I picture. In this case, the track jump is performed immediately after the I picture is played back. Then, the following or the preceding GOP is accessed so that the I picture is played back there. In the case shown in FIG. 6, the high speed feeding playb,..-:,ck and r-zwLnding playback can be actualized by repeating such an operation.
However, since this GOP rate is a variable biz rate, it is 16 impossible to recognize at all where the front of the following C-OP is located. Consequently, the optical head is allowed to appropriately jump to locate the _Front of the GOP. Thus, it is impossible to determine which track should be accessed.
In addition, the I picture has a large amount of data. Thus, when only the I picture is played back in a continuous manner, like a special playback, the picture cannot be played back at a frequency of 30 Hz like a norwa! animated picture because of a limit on the reading speed from the disc. Even when the optical head jumps after the completion of the i picture playback, the intermission for the renewal to the following I picture becomes longer so that the operation lacks in smoothness.
The conventional digital video signal record and playback device, is constituted in ti'le aforementioned manner. In the case where a high speed playback is per-formed at several times sDeed by using the I picture and the P picture, the I picture data and the P picture data is 1 1 read after the front of the GOP is detected from bitstreams wn h are ic recorded on a recording medium or the like such as an optical disc or the like. Consequently, in the case where the data amount of -Lie I picture and the P picture become very large, or in the case where it takes much time to search the front of the GOP, time for reading the 0 data from the recording medium becomes insufficient. Thus there arises a problem in that the data all the I picture and the P picture cannot be read so that the high speed playback cannot be realized.
In the conventional digital video sign.al record and playback device, iz takes much time to input I picture data which has a large amount of data even when the high speed playback is performed only by using the I picture. Consequently the special playback which surpasses tens of times cannot be realized. In This case, a higher speed special playback can be realized by playing back one 1 picture for several GOP. There is a problem in that the interval for the renewal of played back picture will be prolonged so that the content of the picture will become vague.
Since the conventional video signal record and plk,rback device is coded as described above, only the I picture having a large amount of data is decoded at the time of the skip search (watching data through a rapid playback). Consequently, the optical head is allowed to JUMP without playing back data sufficient for the decoding. Other-wise, when a sufficient amount of data is played back, The time for the playback of data is long, the destination to which the GOP is to jump must be set to a considerably far place causing a problem that the number of scenes outputted to the screen becomes few.
is In addition, since the sector address of the following G0P cannot be recognized because of the variable rate, 1z is not verified whether or not the ^ront of the GOP is located at the track to which the jump is made. Consequently, there arises a problem in that a plurality of disc rotation a.r.. reQuired to locate the front of the GOP at the track of destination and the number of scenes which are outputted to the screen becomes much fewer at the time of the special playback. Further, there arises a problem in that if the sector address cp-n be recognized.
1 lable 'or judging to what extent data can be played back no means s aval 1 for the optical head jump with the result that no judgement can be made without passing through the video decoder, and the efficiency at which the optical heaci jumps is lowered.
As other conventional digital video signal record and playback devices, some devices a-re disclosed in, for example, Japanese Patent Application Laid-Open No. F1I 6-98314 (19/94), Japanese Patent Application Laid-Open No. M 6-78289 (1994) and the like. One example is shown in FIG. 10. In FIG. 10, reference numeral T175 denotes a video signal generator such as a camera, a V7 or the like, 7760 an audio signal generator such as a microphone, a V7 or the like, 752 a video signal encoder, '163 an audio signal encoder, 777 a system layer bitstream generator, '178 an error correction coder, 779 a digital modulator, '180 an 19 optical disc, '156 a playback amplifier, 786 a detector, '181 a digital demodulator, 758 an error corrector. a system stream processor, '82 a video signal decoder, 783 an audio signal decoder, a monitor and 785 a speaker.
Currently, optical discs generally used have a diameter of These optical discs are normally capable of recording 600 M byte or more data. uite recently, these optical discs are capable of recording vicleo signal and an audio signal for '14 minutes at a data rate of about 1.2 Mbps. At the time of data recording, a vicleo signal is inputted to the video signal encoder 762 from the video signal generator 775 for encoding the video signal. From the audio signal generator 776, the audio signal is inputted to the audio signal encoder '163 for encoding the audio signal. The process for multiplexing the header or the like to these two encoded signals is carried out by the system layer bitstream generator 717171. After the error correction code is appended by the error correction coder '178, the error correction signal is digitally modulated with a digital modulator 7179 thereby generating a bitstream for recording. This bitstream creates a mother disc with a recording means (not shown), and T-e content of the mother disc is copied to the olDtical disc 780 with the result that a commercially available video software disc is prepared- In a plavb,-.ck device for users, a signal obtained from the video software disc by --,he optical disc is amplified with the playback amplifier 736 to input, a playback signal to the detector '186. After this playback signal is detected with the detector '186, the digital demodulator 781 digitally demodulates the signal to correct errors with -A ignal area is exz--ace an error corrector I'D zer this, the video s ted from the s-ign.al which has been error corrected, and This extracted data is decoded at the video sip-,a! decoder 782 and outputted together with the audio signal decodedby the audio signal decoder 783 to the monitor 784 and the speaker 785 respectively.
A typical method for coding this video signal is an Y.EG 1 and an PEL'- 2 referred to as an MPEG (Moving Picture Ex-perts Group) method which is an international standard coding method. A concrete example of coding method will be explained with respect to an example of...MPEG 2.
FIG. 11 shows a block diagram of an video signal coding part in a conventional digital signi a! record and playb,;-ack device for explaining the MPEG2 coding method. FIG. 12 is a block diagram of an video signal decoding Uniz in a conventional digital signal record and playback device for ex-Dlaining a decoding method. Further, FIG. 13 is a view showing a concept ef the mobile picture processing for the video signal coding in the conventional digital signal record and playback device for 91 0 explaining The grouping of mobile pictures according to the coding method of the PEG 2. Referring to FIG. 13, I2SPER designates ---, I an I picture, B a B picture and P a P picture. For example, in FIG. 13A, mobile pictures from I to the one immediately befcore the- appearance of another I are grouped in a definite number of frames. The number of frames of the pictures which constitute this group is normally 15 frames in many cases. However, the number is not limited to any specific number.
The GOP, a group O.L pictures which constitutes this group includes at least one frame of I picture which can be decoded completely in one frame. The LT also includes a P picture coded through the motion compensation prediction by one direction prediction cf- the time system on the basis of the i picture and a B picture coded by both direction prediction of the time system on the basis of the i picture and the P picture. Incidentally, the arrows in FIGS. 13A and 13B represent:)r-. diction relations.
In other words, the B picture can be coded and decoded only after the I picture and the P picture are prepared. The initial P picture in the GOP can be coded and decoded after the i picture before the P picture is prepared. The second P picture and P picture after that can be coded and decoded when the P picture immediately before the P picture is prepared. Consequently, in the absence of The I picture, either the P or B pictures cannot be coded and decoded.
Referring to FIG. 11, reference numeral '187 denotes a picture rearrang-r, 788 a scan converter, '189 an encoder buffer, '190 a mode determiner, '1702 a motion vector detector, '106 a subtracter, and '108 a DCT circuit which has a field memory, a frame memory, and DCT calculator. Reference numeral 710 denotes a quantizer, '114 an inverse quantizer, -116 an inverse DCT circuit, '118 an adder, 720 an imace memory, -122 a rate controller and 726 a variable-lengzh encoder.
Referring to FIG. 12, reference numeral 733 denotes a variablelength decoder, '136 an inverse DCT circuit, 737 an image memory, 788 an adder, 739 an inverse scan converter. Incidentally, the motion vector detector '102 and the mode determiner '190 combines together tO represent a motion vector detecting unit.
Subsequently, on the basis of FIGS. 11 though 13, an operation of a digital video signal record and playback device will be explained. Referring to FIG. 11, the picture rearranger 787 rearranges pictures for coding in an order shown in FIG. 13. Then the scan converter 788 converts the scan from the raster scan to the block scan. This picture rearrangement and the conversion processing from the raster scan to the block scan are generally referred to as preprocessing. The picture rearran.ger 787 and the scan converter '188 are generally referred to as 23 preprocessor. The inputted picture data is subjected to block scan in the order of encoding. 'When the picture is an 1 picture, the picture passes through the subtracter '106. WThen the picture is a P picture or a B picture, the picture is subtracted with the reference picture and the subtracter '106.
At this time, the motion vector detector '102 determines the motion direction and the motion quantity (the input of the original picture to this motion vector detector '102 may be a picture after the picture rearrangement or a picture after the block scan as an original picture, but the circuit size is smaller in the latter case. Further, the relerence picture must be inputted from the image memory 720 but the reference arrow in the drawing is omitted) with the result that a signal in the area in consideration of the portion of the direction and quantity from the imange memory 720 may be read. At this time, the mode determiner '190 determines whether both direction prediction is used or a one direction prediction may be used.
The substraction with the reference screen in consideration of the motion vector is performed at the subtracter 706. Even pictures with a small electric power are constituted so that the coding efficiency is hei"lzened. The output from the subtracier 706 is collected either in a unit of field or in a unit o-lc fIrame at the ? 1 circuit 708 to be subjected to a DCT process and conve-rze-d into data in a frequency component. This data is inputted to the quantizer '1710 where the weight is different for each of the frequenc.v. The data is scanned in a zlgzag manner in two dimensions over low frequency components and high frequency components to be subjected to a run length coding and a Hu'lfman coding.
This data which has been subjected To the run length coding and Huffman coding is controlled for variablelength coding so that a quantizing table is scaled by using the rate COntrollel- '1722 to allow the data to agree with a target code quantity.
The data That has been subjected to variable- length coding is normally outputted via the encoder buffer '189. The quantized data is brought back to the inverse quantizer 714 to be brought b,---ck to an original picture space data by the inverse DCT circuit 716 with the result that data which is the same as the decoded data is obtained by the adder 718 by adding the original picture space data to the data referenced by the subtracter '106.
FIG. 12 shows a schematic block structure of a decoder. The variablelength decoder 733 decodes picture data including header information such as the motion vector, the coding mode, The picture mode or the like. After this decoded data is quantized, the inverse DCT circuit '136 performs the inverse DCT calculation (incidentally, in FIG.
g- - 13 12, the inverse quantizer located in the front stage ol- the inverse KT circuit 736 is oxitted). BY referring to the picture data from the image memory 71371 in consideration of the motion vector, the motion compensation prediction is decoded by adding the picture, data that has been referred to with the data after the inverse DCT by the adder 738. This data is converted into a raster scan with the inverse scan converter '139 to obtain and outr)ut an interlace ' picture.
Further, in accordance with the variable transmission rate disc system inLreuced -:n;'the variable transmission rate disc system and the code quantity control method" in a publication of Yr. Sugiyaza et al, at the 1994 annual meeting of the Television Society, a proposal is made on a higher quality digital video signal encoding method. This is a method in which an encoding rate is fixed with one program (for exazple, a first set) so that each GOP is set at a rate depending on the difficulty of the design and encoded. FIG. 14 is a block diagram showing a video signal coding unit in a conventional digital signal record and plavb..-ck device. In FIG. 14, reference numeral 791 denotes a motion compensation 1 7 94 a prescictor, '192 a code amount memory, 793 a GOP rate setting unit, code amount assigning unit, 795 a subtracter, and '196 a code awunt counter and 797 a szlr-c'n. The GOP rate setting un-r- 'Q3 shown FIG. 111 is set to change The setting of The quan-Lizing value according to the 26 difficulty of the design pattern. In other words, while the switch 797 is connected to the virtual coding side, the output of the variable- length encoder '126 is inputted To the code amount counter '796 so that the code amount counter '196 counts the code amount to be stored in the code amount memony 792.
The GOP rate setting unit '193 determines the virtual code amount in the whole one program on the basis of The code amount stored in this code amount memory 792 to set and calculate the optimal encoding rate in each GOP. The code assignment at this time is calculated from the code amount assigning unit 7194- for the preparation of the actual encoding.
W 1 1 L hen the switch M is connected to the actual coding side, -he code amount assignment amount and the value of the code amount counter '1796 are compared so that the switch 797 is operated to control the Quantizer 710 on the basis of the actual code amount. In this manner, a small code amount is assigned To an easy design and a large code is assignee' to a difficult amount so that the coding difficulty that gradually changes in the program is absorbed. As a consequence, it has been reported that the picture quality of what is recorded at a rate of 3 Mbps by using this method is approximately the same as the picture quality of what l'S coded at a rate of 6 Mbps.
In consideration of the possibility of the skip search in the 2 T 0 digital video si&nal record and playback device using an optical disc, when the 1 picture and the P picture are played back ior fast rewinding even when the front of the GOP can be accessed at a high speed, the P picture is located at an appropriate position in the C-OP so zhat, there arises a need of operating the optical head while searching data on the bitstream. However, such a control cannot be made in time because of the time constant of a servo such as an actuator or the like. One GOP normally includes 15 frames of pictures, and in NTSC scanning method, 0.5 second is available for finding the front of the C-OP. However, in order to detect the front of a certain GOP, the bitstream requires the reading of 1/2 or more for reading 1/3 picture at a frame rate even when an attempt is made to read the I picture or the P picture at the time of the skip search with the result that the reading speed has to be set to 2.5 times faster or even faster than the normal speed when the head movement time is set to 200 milliseconds. This exceeds the response limit of the actuator. In a normal playback method, the skip search is substantially impossible to carry out.
In accordance with The conventional digital video signal record and playback device, the signal is coded in this manner. Thus, when an attempt is made to perform the skip search like a video tape recorder, a perfect playback picture cannot be obtained in the case 2S where the data is played back which does not allow obtaining a complete original picture from one picture data item like the B picture. Particularly, in the skip search, jerkiness (unnatural movement) is generated with respect to the output processing in the unit of frame. When a variable rate recording is performed with a good playback picture quality, there arises a problem in that the difficulty of accessing the front of the GOP itself increases since the position of the front address of the GOP changes, with the result that a space is formed in a disc area due to disurliform unit of the GOP.
SUMMARY OF THE rNVENTION
It is desirable to provide a digital video signal record and playback device which is capable of performing a special playback by using an I picture with a large data amount and obtaining a playback picture with a ood quality, 9 and a method for recording and playing back the same.
It is desirable to provide a digital video signal record and playback device which is capable of performing a high speed playback by using an 1 picture and a P picture with a large data amount and obtaining a playback picture with a good quality, and a method for recording and playing back the same.
It is desirable to provide a digital video signal record and playback device which is capable of realizing an improvement in the access characteristics of the GOP under the presupposition of adopting, the coding of a variable bit rate while obtaining a favorable skip search, and a method for C recording and playing back the same.
29 0 It is desirable to provide a digital video signal record and playback device which is capable of realizing an improvement in the access characteristics of the GOP and an effective use of space area on a storing medium under the presupposition of adopting the variable rate coding while performing a skip search, and a method for recording and playing back the same.
4 The present invention provides a digital video signal recording device as set out in claim 1. Preferably, data rate information is multiplexed with the additional information or is written in a predetermined region of the recording medium.
Another digital video signal record device includes judging means for judging the degree of picture deterioration at the time of coding and decoding on a basis of the motion compensation prediction and the orthogonal transform, adaptive coding means for coding a data rate by adaptively changing the rate on the basis of the judgement output from the judging means, information adding means for adding an audio signal, additional information such as header or the like, and error correction code, and a data rate setting means for setting a discrete value for the adaptively changed data rate. In the coding means for a variable rate, the rate is limited only to a limited value. Consequently, the data rate information of the GOP (which corresponds to the code amount of the GOP) can be represented with a small number of bits.
Still another digital video signal record device includes judging means forjudging the degree of picture deterioration at the time of coding and decoding on the basis of the motion compensation prediction and the orthogonal transform, adaptive coding means for coding a data rate by adaptively changing the rate on the basis of the judgement output from the judging means, information adding means for adding an audio signal, additional information such as header or the like, and error correction code, ? t wherein the device is so constituted that data rate information is multiplexed on the head or the like, or is written in a predetermined area on the recording medium. The data rate set information in the case where the picture data is coded at a variable rate is recorded on the recording medium apart from the video data. Consequently, the data rate information can be read in summary so that information can be recorded which allows immediate recording of the position of the predetermined GOP which occupies a disc.
The present invention also provides a digital signal playback device as set out in claim 2.
The invention also provides a recording medium as set out in claim 5.
A digital video signal playback device according to an embodiment of the invention switches over a playback mode between the normal playback mode and the special playback mode thereby extracting data rate information. At the time of the special playback mode, the position of the recording medium where data for the special playback exists is calculated on the basis of the data rate information at the time of the special playback mode. When the GOP with a different data rate is played back by extracting the data rate information of each GOP, the coded data divided into two are synthesized and decoded at the time of the normal playback. At the time of the special playback, the position of the GOP on the recording medium which is to be accessed is calculated. Then, the data to be accessed at least is played back to access the next target GOP. At this time, the position information on the recording medium where the COP to be accessed is calculated to facilitate the special playback and retrieval of a high quality variable rate.
In the aforementioned embodiment, a head position is controlled to a position on the recording medium depending on the result of the position calculation and the special playback rate. The position information on the disc where the COP which constitutes the access target is calculated on the basis of the special playback rate. The position of the optical head can be controlled to the position of the target COP depending on the special playback rate so that a high quality variable rate can be played back in a special mode at variety of speed.
The above and flirther features of the invention will more fully be apparent ftom the following detailed description with accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of a conventional optical disc record and playback device.
FIG. 2 is a block diagram of a video signal coding unit in a conventional MPEG.
FIG. 3 is a block diagram of a conventional motion compensation predicting circuit.
FIG. 4 is a block diagram of a video signal decoding unit in the conventional MPEG.
0 j FIG. 5 is a view showing a data arrangement structure of a video coding algorithm of the conventional MPEG.
FIG. 6 is a view showing an example of a GOP structure of a 3 4- video coding algorithm of the conventional VEG.
FIGS. 7A and 7B are views showing one example of a video bitstream of the conventional XPEG.
FIG. 8 is a view showing an example of a system stream in the PS in the conventional MPEC.
FIG. 9 is a view showing an example of a PES packet stream of the conventional XPEG.
FIG. 10 is a block diagram of the conventional digital signal record and Playback device.
FIG. 'Ll is a block diagram of a video signal coding unit in the conventional digital video signal record and playback device.
FIG. 12 is a block diagram of a video signal decoding unit in the conventional digital video signal record and playback device.
FIGS. 13A and 13B are views illustrating a concept 01 Mbile picture processing in the conventional digital signal record and Playback device.
FIG. 14 is a block diagram of a video signal coding unit in the conventional digital signal record and playback device.
FIG. 15 is a block diagram of a record system in a digital signal record and playback device according to embodiment 1.
FIG. 16 is a block diagram of a playback system in the digital 3,- k record and playback device according to embodiment 1.
FIG. 17 is a conceptual view for illustrating macro blocks.
FIG. 18 is a conceptual view for illustrating a screen division.
FIG. 19 is a conceptual view for illustrating a data arrangement.
FIGS. 20A through 20D are conceptual views for illustrating a Playback method in a special playback.
FIGS. 21A and 21B are conceptual views for a method for performing the special playback in the case where the data is exzended.
FIG. 22 is a conceptual view for illustrating a screen division according to embodiLment 3.
FIG. 23 is a conceptual view for illustrating a data arrangement according to embodiment 3.
FIGS. 24-A through 2,11E are conceptual views for illustrating a method for performing the special playback according to embodiment 3.
FIGS. 25A and 25B are conceptual view for illustrating an error correction block arrangement according to embodiment 3.
FIGS. 26A through 26D are conceptual views for illustrating a method for performing the special playback acCordina to embodiment 4.
FIGS. 27A throu h 27 are conceptual views for illustrating a gn method for performing the special playback in the case where the data 17) 1 ) llf. 4- interpolation in embodiment 4 is performed.
FIGS. 28 is a conceptual view for illustrating a data arrangement in embodiment 5.
FIGS. 29A through 29E are conceptual views for illustrating a playback method in a special playback according to embodiment 5.
FIG. 30 is a conceptual view for illustrating a data arrangement in embodiment 6.
FIGS. 31A through 31F are conceptual views for 'Illustrating a method for per'-rorming the special playback in embodiment 6.
FIGS. 32A and 32B are conceptual views Tor illustrating an error correction block arrangement in embodiment 6.
FIGS. 33A through 33G are conceptual views for illustrating a playback method in embodiment 7.
FIGS. 34-A through 34-F are conceptual views for illustrating a method for per-.Lorming the special playback in the c..-se where the data interpolation in embodiment 71 is performed.
FIG. 35 is a conceptual view for illustrating data arrangement in embodiment 8.
FIG. 36A through 36F are conceptual views for illustrating a method for performing the special plavback in embodiment 8.
FIG. 37 is a block diagram of a digital video signal coding unit S -) in embodiment 9.
FIG. 38 is a view showing a concept of a frequency division in embodiments 9 and 11.
FIG. 39 is a flowchart of a digital video signal coding processing in embodiment 9.
FIG. 40 is a view for illustrating a header in a bitstream in embodiment 9.
FIGS. 41A through 41D are views showing rearrangement of the bitstream in embodiment 9.
FIG. 42 is a view showing an example of address information of a system stream in embodiment 9.
FIG. 43 is a block diagram of a digital video signal decoding unit in embodiment 9.
FIG. 44. is a view showing a concept of decoding processing in embodiment 9.
in embodiment 10.
FIG. 45 is a flowchart of decoding processing in embodiment 9.
FIG. 45 is a block diagram of a digital video signal coding unit FIG. 47 is a block diagram of a digi-cal video signal decoding unit in embodiment 10.
FIG. 48 is a view showing an example of an area of a screen in 1 embodiment 10.
FIG. 49 is a view showing an example of a bitstream when the video daT-a is rearranged in the unit of area of the screen in embodiment 10.
FIG. 50 is a flowchart of a digital video signal coding processing in embodiment 10.
FIG. 5-1 is a view showing an example of address information o-E a system stream in embodiment 10.
FIG. Z52 is a view showing an example of a system stream in embodimenz 10.
FIGS. 53A through 53E are views showing an example of a playback screen in which the picture can be played back at the time of playback.
FIGS. SA-A through 5A1E are views showing an example of the Playback screen in embodiment 10 wherein only the central part of the screen is outputted at the time of the playback.
FIGS. 50-A and 55-B are views showing an example the playback screen in embodiment 10 in which an area in the central part of the screen is wWified and displayed.
FIG. 56 is a flowchart of a digital video decoding processing in embodiment 10.
0 I S a c k C'la97 am 01 a di g iza I s l pia l c c.;rig uni t in e' di 1 D DO lm en z 1 - - T-7 L L -1 rK. 58 is a -i7'Owcha-,z c' a digital v:d--o sig-,al coding Processing In 11.
FICS. v"D are views showing an example of a system Stream in embcdimenz II.
F71G. 60 is a view sL-Low-;..-Lg an example of address 1'n'Lorn.-z-icn of the svsz-zm S7-r-e-am in 11.
a dgital Y1 'e,,- signal 61 is a block diagram ol unit in 1-1.
FIG. 62 is a -L'-lowcharz c-ic a digital vIdec signal decoding processing in emcOCI-ment 71- cl -1 1 - (2 63 is a block---agram of a digizal, signal coding unit in 12.
FIC.. 6-1 is a view showing a concept C-I a resolution conversion in 12 on the scr=-=-,-1.
FIG. 65 "s a view illustrating one example of 'ata consztution rz-sul-Ls in embodiments 12, 19. and P11.
FIC. 66 is a block viow of a digital video sip.al coding unit in embcdiment H'.
Is a view illus-razing one example cl- a data IFZ 1-4.
0--MPIA IRIT21p R;o TRAtTP oolq ? ST cl -DII zugoTpoqwa uT ziun 1?u2is or.,p1A R;c jolq ? sT i SI1 -91:UGUIP09ma UT 2"UTSSGOOjd 2UTPOO I?U2IS C-9p1A peIT2TP lo ldgouo,) uoTzRjado ut, &-rz?jzsnlll i9TA.9 sT SCAt DII -91 ZUGTTpcqTG uT ziun 2uipoo 1?U21S OGPIA linTSTP ?,UTIOS Et12-21p 301q ? ST VS1 -SI1 zu9TTpoqma uT luTssaooid 2uTpoiap 1221S GG p 1 A 1 12' 121 p 2 zdaouoi uoTiejado u? &TzRjzsnlll IGTA ST EZI -gjl zugmTpoqxa uT ziun 2UTPOOGP I?U?TS OGPIA linT2TP ? p Igt>.121,>TP 001q ST VZZ '301q GUO UT PU2 Cl ZUGETPOqT@ UT DOIR TTE>GIZ5ZTq ? UT uGuodToo iouGnbGil ?;o GuTIno zugwGRmaj? o G-[dwxG auo Si-TTiRizsnlll SIGIA GI? Cl, tnoi- VI, SDIj i pn cl i SUgmrpoina UT ajugnbgs 2111SSGOOld TO G1duxa GUO 2UTIOS IGTA ? S! 01 'DII PUR sI 'ZT szuawlp<>qwG UT ?Z?D p-3poi o -iunoule lwTzsT.R-Is G-1 o aldw- 2x.3 u? 2uTz?jzsnlll taTA si ng ziun Suipm -1-2L2TS OGPIA 1?--)121p ? o MU.12,eTP iolq ST 99 -9iJ DOlq 1)j;0 9PISUI Ugl)TTGoj o IUGWG2LTil? 0 unit in embodiment 17.
FIG. 76 is a block diagram of a GOP address generator and a disc controller in embodiment 18.
FIG. 71 is a block diagram of a GOP address generator and a disc controller including a playback processing in embodiment 18.
FIG. 78 is a block diagram of a digital video signal decoding unit when the division by the frequency and the division by zille quantization in embodiment 19 are performed.
FIG. 79 is a block diagram of the digital video signal decoding unit when the division by the bit length in embodiment 19 is per-lormed.
FIG. 80 is a block diagram of the digital video signal decoding unit when the division bv the resolution in embodiment 19 is performed.
FIG. 81 is a block diagram of the digital video signal coding unit in embodiment 20.
FIG. 82 is a block diagram of the digital video signal decoding unit 'in embodiment 20.
FIGS. 83A and 83B are views illustrating a concept cl' the processing with the digital video signal record and playback device in embodiment 20.
FITC. P- is a block diagram of the digital video signal decoding unit when the division by the frequency or the division by the q quantization in embodiment 21 is performed.
FIG. 85 is a block diagram of the digital video signal decoding unit when the division by the bit length or the division by the quantization in embodiment 21 is performed.
FIG. 86 is a block diagram of a digital video signal decoding unit when the division by the resolution or the division by the quantization in embodiment 21 is performed.
FIG. 87 is a block diagram of the digital video signal decoding unit when the division by the frequency or the division by the quantization in embodiment 22 is performed.
FIG. 88 is a block diagram of the digital video signal decoding unit when the division by the bit length or the division by the quantization in embodiment 22 is performed.
FIGS. 89A and 89B are views showing the concept of processing at the time of skip search in embodiment 22.
FIGS. 90A and 90B are views showing a concept of processing at the time of the inverse playback in embodiment 22.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
The present invention will be explained in detail on the basis of the drawings showing embodiments. In the following description, Embodiments 1 to 11, especially Embodiments 1 to 4, 6, 7, 10 and 11, relate to the invention claimed in the parent application, GB 9519553.3, Embodiments 12 to 22 relate to the invention claimed in co-pending divisional application no. 9816673.9 and Embodiments 17 to 19 relate to the invention claimed in this application. Embodiment 1 -S c 1 1 imbc,d--lmenz 1 will be FIG.
15) is a D,-,C)ck diagram showing a recorcling SS..Sle-TE C1, a digital video s-,.g^Lal reCC7C Wild pla%.,b.---ck device in embodimenz 1. RelEerring to FIG. 115, a C'I'g-izal video Signal -rom an i7-LTz terminal 1 is - ' i ' i ec 1,-i;Dut-L a f-ormazzing 3. 7he vide-o s-;P,,a" zr,-cn Is ou 1 ' -!g C1 -L 3 is 'ii;Du-L-,ed -.c a " -,s7 ii-ipt-l-L c-: a sub-Lracz=r lrom Zhe t o rMA- z --. 1 ircul 1 1 11 4 and a second c-L a motion compensation Predlczl7,,g C-,1rCL,-;-L 11. An c L, z'iD u t c. 4 is -,17.Puz-.ed to a 5 via a DCT cl-,cu-;z 5. An c--,-: QLa,-iz-lzer 5 is inpuzzed -,c a Ir.ipLz c-LE a DUI, memory 12 via a encod--r 7. 1-ii zhe -,he cutput Of the 6 is also inpuzzed -cc an ir,,verse DCTI cv via an inv=rse 1 iven to 1 An ou-LPL- c-E c' inverse DC-1 9 's g- a f -i:- s zn:)LT z off an acide r 10 ,7 of the adder 10 is give,-, to a -Lcl'r.sz of The mozion compens.;-,-,lor pre-,,-,cT-ing circuit 11. A first output cf, the motion circuit 11 is givep to a se-Wirld input C-, the adder. 10 and a second input of the subtracter 4. a second output c-IF -Li-,e wzion compensation predicting circuit K is given To a second -.-'"le memorv 12. An o:-cpu-c c-,: buf', e. memory 19 is to a modulator. 14 via a for7zat encoder 1.3j. An ouzpL-L of the 1 disc c-, m ulazz- IL 's recorlded on a recording sLc' as an c -Lica t A- a the like via an output terminal 2.
FIG. 16 is a block circuit diagram showing a playback system in the digital video sip, al record and playback device according to embodiment 1. Referring to FIG. 16, video information read from the recording medium is inputted from an input terminal 20 to a demodulator 21. An output from the deffloculator 21 is inputted to the format decoder 23 via a buffer memory 22. The first output of a format decoder 23 is inputted to a variable-lengzh decoder 24, and inverselY quantized at an inverse quantizer 25). Then the output is subjected to the inverse DCT at an inverse DCT circuit 26 to be given to the f irst input of an adder 28. In the meantime, the second output of the format decoder 23 is inpuzte-d to a prediction data decoding circuit 271. Then, the output from the prediction data decoding circuit 27 is given to the second input of the adder 28. The output of the adder 28 is outputted from an output terminal 30 via an unformatzing circuit 29.
.ext, an operation of the device will be explained. The digital video signal is inputted from the input terminal 1 in the unit of line, and is supplied to the formatting circuit 3. Here, in the motion compensation predliction, one GOP is set to 15-frames as shown in FIG. 6 like the conventional example to perform prediction coding with one frame of I picture, 4 frames of P pictures (P1 through Pjd-), and ten 1 0 frames of B picti,r_zs (B1 through B10) In this case, in the formatting circuit 3, the video data inputted in a consecutive manner is rearranged and outputzed in zlne unit of frame in the order shown in FIG. 7.
Further, the data inputted in the un:lz of line is rearranged in Lhe c block of 8 X 8 P'.x=ls so that macroblocks (s'x blocks in total, such as adjacent four luminance signal Y blocks and two color C"fEerence signals Cr and Cb blocks which correspond in position to the Y block) is constituted. The data is outputted in the unit of macroblock. Here, The macroblocks are determined in the minimum unit of ine motion compensation prediction while the motion vector for the motion compensation prediction is determined in the unit of macroblock.
Further, with the formatting circuit 3, with respect to the I picture, one frame of video data is divided into three areas so th blocking is performed in this area in the unit of 8 x 8 pixels and the macroblock is constituted and outputted. Here, the three divided areas are set as arleas 1, 2 and 3 from the top of the screen as shown in FIG. 18. In FIG. 18, the area 2 located at the central part of the screen has a size of 720 pixels x 288 lines while the are,--s on both ends of the screen have a size of '120 pixels x 96 lines. in the meantime, in the P picture and the B picture, the blocking is performed without being divided into e,-,ch area and is ouiputted in tie unit of macroblock.
- c., An output of the rcult 3 is inputted to the tormatzing cl subtracter 4 and the motion compensation predicting C1.- LCul't 11. The operation of the subrracter 4, the DCT circuit 5, the quantizer 6, the variable-length encoder 7, the inverse quantizer 8, the inverse DCT circuit 9, the adder 10 and the motion compensation predicting circuit 11 is the same as the conventional embodiments, and the explanation thereof is omitted.
The video data outputted from the variable length encoder 7 and the motion vector outputted from the motion compensation predicting circuit 11 are inputted to the buffer memors, 12. In the buffer memory 12, the video data and the motion vector for one GOP portion are recorded and the data is subsequently outputzed in sequence to the format encoder 13. The output of the format encoder 13 is inputted to the modulator 14 and an error correction codes or the like are added and recorded on the recording medium such as an optical disc or the like.
In the format encoder 13, the video data for the one GOP portion is rearranged in the data arrangement as shown in FIG. 19 and is outputted to the modulator 14. Here, the I picture is divided into three areas as shown in FIG. 18. When the data of the I picture. corresponding to these areas 1 through 3 are set to 1 (1), 1 (2) and 1 (3), the data o the I picture is constituzed so that the data is recorded in the order of E2),!(I) and IM at the front part of the data string of one GOP portion.
Further, the address where th data of each picture area is stored at The front of the GOP is recorded as header information. The number of bytes which is occupied in the data format shown in FIG. 19 by the data in each area divided into three parts is recorded as header information. Consequently, depending on the number of bytes occupied by each area which is recorded in the header information, the end position ol each area can be recognized as a relative address from the front of the GOP at the time of Playback. Consequently, the optical head jumps to the Irlont address of the GOP in the unit of the definite time at the time of the special playback so that data can be read in each area in accordance with the header information from the front of the GOP.
With a general video signal record and playback device, on the data format at the time of data recording, the I picture is recorded in the unit of frame. In contrast, in FIG. 19, a priority is given to an area located at the central part of the screen out of the I picture data which is divided into three parts so that the area is located at the front of one GOP. Consequently, in the case where only a part of the area of the I picture can be decoded in a definite time at the time of a high speed playback, at least the playback picture at the central part -5 or of the screen c..-n be outputted.
Su-,sequent ly, an operation at the time of plavback will be explained in accordance with FIG. 16. The demodulator 21 performs an error correction processing so that the video signal recorded in a format shown in FIG. 19 in the buffer memory 22 is divided into the motion vector and the video data at the format decoder 23 to be outputted to the predicting data decoding circuit 27 and the variable-length decoder 24, respectively. Here, an operation at the time of the normal playback is the same as the conventional embodiment, and an explanation thereof is omitted.
At the time of a high speed playback, with respect to the data recorded in one GOP unit on the recording medium such as an optical disc or the like, the optical head jumps to the front of the one GOP in the unit of definite time so that the data part of the 1 picture is read in the unit of area in accordance with the header information recorded at the front so zn'at The data is demodulated at the demodulator 21 2-rld is input to the buffer memory 22. Here, in the case where data is read from the recording medium such as an optical disc or the like at the time of a high speed playback, waiting time for the disc rotation arises at the time of jumping to the front of the GOP even when the front address of the GOP which is recorded on the disc is known.
Consequently, when the high speed playback speed is 'increased, the time for reading the data on the disc becomes short. Since the waiting time for the disc rotation varies, it becomes impossible to read all the I picture data in a stable manner.
Consequently, when the high speed playback speed is increased, after only the data cl' the area 2 located at the central part of the screen is read the optical head jumps to the front of the subsequent GOP so that only the data in the area 2 that can be read is inputted to the buffer memory 22. In this case, the format decoder 23 decodes only the area 2 of the I plczure that can be read. On the other hand, the areas 1 and 3 whose data are not read are masked by the gray data, and a high speed playback picture is outputted. Consequently, in the case where one GOP is set to 15 frames, a 15 times speed special playback picture can be obtained.
FIG. 20 shows a playback picture in the case where a high speed playback is performed by playing back only the area 2 of the 1 picture from the nth GOP of one GOP up to the n13th GOP. In FIG. 20, the areas 1 and 3 on both ends of the screen in FIG. 20 are masked by the gray data. Further, when the information amount of the I picture is small and the disc rotation wait time is short, and time is available -for reading data in the areas 1 and 3, the data of the areas 1 and 3 is not decoded.
S^- 41 This is because all the data of one screen portion cannot be read in a stable way at the time of a high speed playback, and if a screen is outputted only when the data in the areas 1 and 3 can be read, these areas cannot be outputted in a definite interval so that a high speed playback picture becomes unnatural.
As described above, since the I picture used for the special Playback as shown in FIG. 19 is arranged so that a priority is given to the area located in the center of one screen is recorded on the recording medium at the front of G82 one COP, only the data of the area 2 located at the center is read for a high speed play back even when the high speed playback speed is increased with the result that the content of the playback picture is easy to see. Further, since only the data in the region of the area 2 is read from the recording medium, a higher speed special playback can be realized compared with a case in which the whole I picture is read.
In, the aforementioned embodiment 1, the I picture is divided into three areas in the vertical direction as shown in FIG. 18 and is recorded, the picture is not necessarily divided into three areas- The area may be divided into n areas (n >1) in The unit of slice defined in the international standard PEG for recording the data. Here, the slice has a one dimension structure of macroblocs of a arbitrary number of 1 0, -W, (on=- or me-e) so when the right end c-' ' e screen is attaineld, --,he to z.ne le-jo-L end one line below.
Embodiment 2 Next, 2.
zill be explained m respect to FIG. 21 is a conc--uz,-,al Y,-w for emilaining a mezhod for specla! Playback in the case where daza z-xzz-.-ision in embodimenz 2 is In e,,'DOC'imenz 1, -he I Picture is dIvided into three areas as s'icc-i in FIG. 18 so -L'na-L only Cat?- cf-Ln'e are,?- 2 located at, the c-;r-nzz-r oF zhz- are-n is read and plaved, 7.nus, with respect zo -me- a:"-=ns 1 and 3, the mask data is However, The 1 _z ' Caza 01 the area 2 is exz---nded to a size one screen as shown in FIG.
21.
In =-se, a-, z,ne zime ofE converting video signal into daza in -dhe uniz c-T!I'ne the uni^orma-L-L-i,-lg circi-,.:-, 219, the daza of Lhe area. 2 is inz-=:pcla7Ied to be ex-Len'ed to a size of one screen portion and is OUT uztz^:d. Tn the case of FIG. 21, -L:ne area 2 has a sIze - p - J.
C'I 7120 pIxels x 258 lines and is in line smmet".-ric in Vertical dIrections --From the center olc trIe screen.
Hp-=, at the -Lime of -Lhe special w'hen -Lhe upper half Part OF the area is s-=-, to 22a and the lower hal- -a r t 's seT zo AIR2b as show-, in FIC. 211A, AR2a and AR2b are extended by!.U:: zimes in the Si- 1 vertical direction respectively to synthesize playback pictures AR2a' and AR2b' as shown in FIG. 21B. With respect to the method for extending the pictures, when the data in the unit of each line of AR2a is defined as ATOM: line number 1 55 1 55 1441), and line data in the upper half part of the extended screen is set to DT(m) (1:-:: m:S 240), extension is made which is represented by the following expressions.
DT (3n - 2) = AT (2n 1) DT (3n - 1) = AT (2n - 1) DT (3n) = AT(2n) (n=l to 80) In the meantime, when the data in the unit c-LE each line of AR2b is defined as AB (1) 0: line number 1 55 1 55 144), and line data in the lower half part of the extended screen is set to DBW (1 t 240), extension is made which is represented by the following expressions.
DB (3n 2) = AB (2n -1) DB (3n 1) = A.B (2n - 1) DB (3n) AB (2n) (n = 1 to 80) As described above, only the data of the area, 2 located at the center of screen at the time of the high speed playback is read and is extended to a size of one screen portion and is outputted as a Playback picture. Consequently, since both ends of the playback picture at the 0 zime c-L: a high spe-ed is no-,. masked, -,,lie cease Lo be to zazch.
- - - - %c -1 pc in -,he ' embo-menT- 2, zhe screen is extend ' in lie.
7-11e vp-rz-ic-z Cl-lion by 1.-,Se-Lzing Cata simply in unit cl line data may be With respect zo the direcz-ion.
7=bod.,De,iz 3 :-z,bod imenz 3 zill be A S11'sZZM and a Plavb.---k S.?S-L=.i 0: -khe Video sipal and device 3 is the S.'aMe as embodizent 1 (See FICS. 1-3 and 16) - an operazler, o-f the device will be All d'.91.-Lal is in the unit cf, line flrom the z-ipu-, zerminial 1 and is suplied to c-ircL,-;-L 2. Here, in the motion compensanlon on= GOP is seto 15 frames like the conventional J. - -I - L 1 - - - example as shown -n FIG. 6. Then, -ine GOP is to T-,nz C -c p Pictures (P 1 prediction coding as one -7ame- 01 1 Pictulre, -four 17Wes 01 Lnroug P4) (B1 B10) In r-:5 Case, in in a continuous manner the 3, -,he data -nput-Lec like the con,enz-ional examcle is re,7--rrarig--c _in tille Ll-niz- c.L lrame in an m - 1 C r - order as shown in F7C. - an' is ouzputteA-. Further, zhe daZP- inPuttr--d -i 'I in the unit of line is rearranged in the unit of block having 8 x 8 pixels to constitute a macroblock (a total 0.1 six blocks of adjacent 'Lour luminance signal Y blocks and two color difference signals Cr and Cb blocks) shown in FIG. 17 so that the data is outputted in the unit of macroblock. Here, the macroblock is the minimum unit of The motion compensation prediction, and the motion vector for the motion compensation prediction is determined in the unit of macroblock.
Further, in the formatting circuit 3, the 1 picture is divided into five areas for each of 720 pixels X 96 lines in the vertical direction of one frame of video data. In this area, 8 X 8 pixels are blocked to constitute a macroblock for the output. In this case, divided five areas are defined as areas 1, 2, 3, 4 and 5. in the meantime, The P picture and the B picture are blocked without being divided into areas and is outputted in the unit of macroblock.
The output of the forwtting circuit 3 is inputted to the subtracter 4 and the motion compensation predicting circuit 11, but the operation of the subtracter 4, the DCT circuit 5, the quantizer 6, the variable length encoder 7 and the inverse quantizer 8, the inverse DCT circuit, the adder 10, and the motion compensation predicting circuit 11 is the same as the conventional embodiment and an explanation thereof is omitted.
5,5 The video data outputted from the variable-length encoder 7 and the motion vector outputted from the motion compensation circuit 11 are inputted to the bu-er memory 12. In the buffer memory 12, the video data and the motion vector of one GOP portion is recorded, and the data is ouzputted to the format encoder 13 in sequence. The output of the format coder 13 is inputted to a modulation circuit 14 so that an error correction code or the like is added and is recorded on the recording medium such as an optical disc or the like.
In the format encoder 13, the data of' the GOP portion is outputted to the modulator 14 by rearranging the video signal in the data arrangement as shown in FIG. 23. The I picture are divided into five areas as shown in FIG. 22 so that the data of the I picture corresponding to areas 1 through 5 are defined as I(1), IM, 1(3), 1(4) and I(E). In FIG. 9-3, the data of the I picture is constituted to be recorded in the order of 1(1), IM, IM, IM and 1(5) at the front of a data stream for one GOP so that priority is given to the area which comes to the center of the screen.
Further, in FIG. 23, the address where the data of each I pictures is stored is written as header information. As the header information, the number of bytes which the data in each area occupies on the data format is recorded, the area being obtained by dividing the 1 -, 1 picture into f ive parts. Consequently, at the time of the playback, it is possible to recognize the end position of each area as a relative address with respect to the iront of the GOP on the basis of the number of bytes occupied by each area which is recorded in header information at the time of the playback. As a consequence, the optical head jumps to the front address of the GOP in the unit of a definite time so that the data of the 1 picture can be read in the unit of area in accordance with the header information from the front of the GOP.
With a general video signal record and playback device in comwn use, in the data format at the time of recording, the I picture is recorded in the unit of frame. In contrast, in FIG. 23, a priority is given to an area 'Jocated at the central part of the screen out of the five areas obtained by dividing the I picture to be arranged at the front of one GOP with the result that the playback picture at least at the central part of the screen can be outputted even in the case where only the area in part of the I picture can be decoded.
Subsequently, an oDeration at the time of playback will be explained in accordance with FIG. 16. A video signal which is subjected to an error correction processing in the demodulator 21 and is recorded in a format c-" FIG. 23 in the buffer memory 22 is divided into the motion vector and the video data which are outputted to the prediction data 5-7 1 decoding circuit 27 and the variable-lengzh decoder 24, respectively. Here, the operation at the time of the normal playback, is the same as the conventional embodiments, a detailed explanation thereof is omitted.
At the time c-OL a high speed playback, with respect To the data recorded on the recording medium such as an optical disc or the like, the optical head jumps to the front of The GOP in the unit of a definite time to read the data part of the I picture in accordance with The header information and the data is demodulated at the demodulator 21 to be inputted to the buffer memory 22. However, in the case where the information awunt of the I picture is zw large to be read in a definite time, the data which has been half read is read to the last item of the data, and the optical head jumps to the front of the subsequent GOP to input only the data that can be read into the buff er memon, 22. In such a case, in the format decoder 23, only the area of the I picture that can be read is decoded and is outputted as a high speed playback picture. Consequently, when one GOP is set to 15 frames, a 15 times speed special playback picture can be obtained.
FIG. 24 shows a playback picture in the case where only the I picture of one GOP is played back. In this case, the information amount X of all areas of the I pictures is too large to be read from the recording medium, with respect to the area which cannot be re-ad, data of 4 the preceding area is held as they are to be outputted thereby synthesizing the high speed playback picture. In FIG. 24, in the case where the n-lth GOP area 5 and n3th GOP areas 1 and 5 cannot be read, the playback picture immediately before the playback picture is held as 11. is.
In this manner, the I picture used in the special playback as shown in FIG. 23 is positioned so that the priority is given to the area located at central part of the screen just above the screen to be recorded on the recording medium at the front of one GOP. Thus even when the whole I pictures cannot be read, the central part of the screen is given a priority in playback so that the content of the playback picture is easy to understand.
In the aforementioned embodiment 3, when the whole I pictures cannot be read, the playback picture is interpolated in the unit of are-a, the interpolation may not be made in the unit of area, but may be made in the unit of error correction block.
In this case, the demodulator 21 segments data into several byte-long packets with respect to the data arrangement shown in FIG. 23, and a error correction code is added to each packet. FIG. 25 shows an example of a case in which data in five areas inputted in a consecutive manner is divided into packets in the unit of error correction block S 0 1 -, unit. FIG. 25A shows a data. string before the packet division. FIG. 25B shows daza after the Packet divIsion. Five areas of the I picture are divided into packet's with a definite volume and the area IM is divided into packets from 1 through 1 and the 1(4) is divided into packets 1 through J for the input.
At the time of a high speed playback, the optical head jumps to the front of the C-OP in the unit of a definite time with respect to data recorded on the recording medium such as an optical. disc or the like in the unit of GOP to read the data portion of the I picture in the unit of area in accordance with the header information. The data portion is demodulated by the demodulator 21 to be input-Led to the buffer memory 22. However, - in the case where the whole I picture cannot be read in a definite time because the information amount of the I picture is large, the optical head jumps to the front of the subsequent GOP even when the one area portion of data is being read. Further, data which can be read is subjected to the error correction processing so that the data which can be error corrected is inpltted to the buffer memory 22. in this case, the format decoder 23 recognizes an address of the I picture area which can be decoded to the midway so that the data which can be read is decoded in the unit of macroblock and is outputted as a high speed playback picture. In this case, with respect to the macroblock which C,- 9 cannot be decoded, data of the preceding screen is held and outputted at it is.
In the aforeffientioned embodiment 3, data in each area of the I picture is divided into packets in a consecutive manner. However, data may be divided so that the data in two or more areas may riot be included in one packet. In this case, data in one area portion is closed in integer times of the error correction block with the result that the data can be rearranged in the unit of area immediately after The error correction processing. When data in e,-.ch area- is divided into the unit of packet, data is input-Led halfway to the last packet of each area so that the residual data is required to be placed in data masking (for example, all the data is masked to 'T').
In addition, in the aforementioned embodiment 3, a priority is given in the order of 3, 2, 4, 1 and 5. However, the order is not limited to this order. The order may be, for example, 3, 4, 2, 5 and 1.
In addition, in the aforementioned embodiment 3, the I picture is divided 'Into five areas in the horizontal direction and recorded as shown in FIG. 22. The data is not required to be divided into five areas, but the dat,. - may be divided into n areas (n > 1) in the unit of slice defined by the international standard XPEG. Here, the slice has a one dimensional structure of macro blocks with an arbitrary number of 'I i - k, lp-ng-is (One or more).
-rx rle slice is a belt which to zhe lez-L 1 - - 1 end one line below upon reac'lling the right end oi sc-ec-n.
Embodiment 4 emc,--dimenz -- will be z-xpla.,iiel wlzh respect zo d7-az1;7Lgs. FIG. 26 is a viez showIng a special- mezhc,d in em'DCc'i-m en t 4 In embo,-',-ment 3, a spee-ial is perormed FIG. 2,. However, -L'-iz s.ie-ia I - -- 1 playh-ack 1 playback P may be so lctu,-.,- as ^n In Z11lis Wse, the format, decoder 231 one screen 'DV playing e-ach one frOM th;- I P; 0-7 COP's which are continuous as shown in FIG. 26. For example, in FITIG. 25A, one screen pa-,;back P is sv-i-izhes--;-:.d ---OM zhe I p-.7-Ures a portion C1 Lhe 4-, L nth to C-OP so that the I Picture c--.: z1e WP is plaved b.ack in a7e---n 1, zh-z i picture ofF -Lhe n-t-3zh C10P is plaved back in area 2, t_he I -,iczu.re cf- z'.,ie n-2-Lh Cop is played b.-.--k in area 3, the I pic-,.ure 0-1f- the G0P is played back in area 4, and t' e j piczt,,re of the n-" IR L G0P is ulaved ba-- In area 5. Further, zo 7TC, 26, when an i azzentlon is paid to zhe are---. 5, the I c-E n-1zh, n-2t:rl --- GOP are back as -1he plaved back v-"Idea 7_ 0- dur' when the whole I piczl_ire W-PIn 1 ng a cefinize z-mz- beca-use -che "ii-or-maz-ion amounz- of T_ plc-ur is larg - (,-- j the data preceding by one Screen is held as it is and is outpuzzed to synthesize a higier speed playback picture. FIG. 27 is a playback picture when the n-10 GOP area 5, and n-30 GOP areas 1 and 5 cannot be read - in this case, since the data arrangement is recorded on the recording medium T giving a priority to the an area located at the central part of tie screen as shown in FIG. 23, the central portion OF the screen is given a priority in the playback even in the case Where the whole i picme cannot be read in terms of time CJ the result that it never happens Jaz the playback picture is hard ze see. FurCer, even in the case where daza in two or more areas connoz be read, one screen is divided inzo five areas. Since the frame played back in each area is differenc it is hard to detect that data is lacking in the playback picture.
Modimen: 5 Next, emboOment 5 will be explained with respect to Ce drawings. FIG. 28 is a view showing an arrangement structure of a digital video signal data according to embodiment 5. in embodimenz 3, the daza arrangement is written in the order of the areas 3, 2, 4, 1 and 5 with respect to the i picture as shown in FIG. 23. The arrangement may have a mucture shown in FIG. 28. in FIG. 28, when the data of the i plczure is recorded at the front portion of Ce data arrangement of one GOP portion, the area number at the front of each of the G0Ps is scrolled. In other words, as shown in FIG. 28, when the I picture data is recorded in the order of 1(5), 1(1), IM, IM and IM in the nth GOP, the I picture data is recorded in the order of 1(1), 1(2 IM, IN) and I(S) in the n-lth GOP. Further, IM comes first in the n-2th GOP. When the GOP number becomes n 1 3 and n-4 and ---, the front area is seQuentially scrolled and recorded in the order of 1(3), IM, I(I) and -.
Further, at the front of the GOP, the address where data in eh I picture is stored and information for re, i - ach 1 cognizing the kind of the front area are written as header information. As header information, the area number recorded at the front and the number of bytes indicating data amount occupied on the data format of each area as shown, in FIG. 2-1. Consequently, at the time of the playback, the data order of the I picture area and the end position of the each area on the recording m adium can be recognized as relative addresses with respect to the front of the GOP, with the front area number recorded in the header information and the number of bytes occupied by each area on the recording medium. Consequently, at the time of the special playback, the optical head jumps to the front address of the GOP in the unit of a certain period of time, so that the I picture data can be read for each L_ _ area from the 'front of the GOP in accordance with the header information.
In this case, the position where the I picture area divided into five areas is recorded is scrolled in the unit of GOP so that the area which cannot be decoded is not concentrated on the fixed position on the screen even when only part of the area of the I picture can be decoded at the time of the special playback.
At the time of the high speed playback, the optical head jumps to the front of the GOP in the unit of certain time with respect to the data recorded on the recording medium such as an optical disc or the like to read the data portion of the I picture in the unit of area in accordancewith the header information and is demodulated at the demodulator 21 and is inputted to the buffer memon 22. However, when the information amount of the I picture is too large to read the whole I picture in a certain time, the optical head jumps to the front of the subsequent GOP after reading to the last data in the area which is read halfways to input only the data in the area that can be read into the buffer memory 22. in this case, the format decoder 23 decodes only the area of the I picture that can be read, which is outputted as a high speed playback picture. Consequently, in the case where one GOP is set to 15 frames, a 15 times speed special playback picture is obtained.
1 S_ 1 FIG. 29 Shows a playb.-----k Picture in the case -inere the 1 p-lczure c-LE one GO? is played back in a hi& s;)eed T-his case, is to be recorded on the recordiling medium, an order as shol-n in TG. 28. Here in The case where zh.e amounz of the I Piczure is large, and the whole I be:aad in t-lme, zIne Caza c-, -rie prece-d-ing screen is held as -iz is and is so that a h-i&h speed Playback Plc-cire is syin--hesiz FIG. 29 shows a case in which GOP a.4maa 5 and n-3z'n C-OP 1 and 2 be ead C071,Ple-L-Zly. i-i t' is Cz-se, e data Of -die n- ---1 IS held as -i-,-'s.
above, the order ot7 record-ing the I -ic-Lre used for the special n'-, avbacK as sho-,n in FIG. 28 is scrolled in the unit of R C-OP. Conse-,uenzly, even i.n the case =here oniv some a-.e-.s c-'^ t, e I Picture can be C7ecy-e-" at zrie z.,;.me o the special pla.-bac, zhle area that can not be decoded is not concentrated on the lixed posi-nion on the screen.
Embodiment 6 Next, embod, imenz zM be explained 7ith respect zo C,7awi-,igs. FIG. 30 is a view showing a data C-IF a video data accordirig -LO 6. in this case, z,_le I p-icz,,j-e and the P p;cr-ure are "ive areas e-2Ch C-, 1 having 720 pixels x 9.6 lines so that each area is blocked in The unit of the macroblock and is coded as shown in FIG. 22. However, the P picture is divided into five area. The motion compensation prediction is performed and codz--l in such a manner that the retrieval scope of The reference pattern of the motion compensation prediction closes in the area. Here, the divided five areas are defined as areas 1, 2, 3, 4 and 5 from the top. Further, with respect to The B picture, the motion compensation prediction is performed and coded without being divided into areas.
With the format encoder 13, the data of one GOP portion is used to rearrange the video signal with the data arrangement shown in FIG. 30 and is outputted to the modulator 14. Here, with respect to the 1 picture and the P picture, one screen is divided into five areas as shown in FIG. 22. The I picture, and the P1, P2, P3 and P4 pictures are defined as I0) through I(S) and PI(I) through Pi(S) (1=1 through 4). In FIG. 30, the data of the I picture, the P1, P2, P3 and P4 pictures is constituted to be recorded in the order of 3, 2, 4, 1 and D' from the front of the data string for one GOP portion, so that the area located at the central parz of the screen is given a priority. Further, in FIG. 30, the data amount of each area is recorded as header information at the front of one GOP so that the address of the data in each I picture 1- 2 and P picture area can be recognized.
At the time of the high speed playback, the optical head jumps to the front of the WP with respect to the data which is recorded in the unit of one GOP on the recording medium such as optical disc or the like with the result that the data portion of the of the 1 picture and the P picture is read in the unit of area and is demodulated by the demodulator 21 and is inputted to the buffer memory 22. However, when the information amount Of the I picture and the P picture is too large to read the whole i pictures and P pictures in a def inite time, the areas read halfways are read to the last. Then the optical head jumps to the iront of the GOP in the unit of a certain time so that only the data that can be read is input-Led to the buffer memory 22. In this case, the format decoder 23 decodes only the areas of the I picture and the P picture that can be read and then output the da-L,:- as a high speed playback picture. Consequently, in the case where the one GOP is set to 15 frames, a triple speed special playback picture can be obtained.
Further, since the area located at the central part of the screen is given a priority to be arranged at the f ronz of one GOP out of The I picture divided into five sections, at least a playback picture at the central part oF the screen can be outputted even in the case where only a Part of either the 1 picture or the P picture can be decoded.
* 4 Further the pictures are recorded on the recording medium in the order of the i picture, the P1 picture, the P2 picture, the P3 picture and the P4 picture. Consequently, it never happens that the reference data cannot be played back at the prediction data decoding circuit 27 even when all the data cannot be read.
FIG. 31 shows a play back picture in the case where a high speed playback of the picture is performed by playing back only the I picture and the P picture in one GOP. In this case, when the whole I picture and the whole P pictures cannot be read from the recording medium in a definite time because the information amount of;he I picture and the P pictures are large, the data of the preceding screen is held as it is and outputted to synthesize a hign speed playback picture with respect to the area that cannot be read. FIG. 31 shows the case where the areas 3, 4 and 5 of the P4 of the nth WP cannot be read In this case, the data of the preceding screen is held as it is.
As described above, as shown in FIG. 30, the I picture and the P pictures are played back to output a special playback picture at the time of the special playback by collecting and arranging the I picture and the P picture used at the time of special playback in the unit of area at the front of the one GOP. Further, in the case where the whole I picture and the whole P picture cannot be read because of time limit, (. C? Lhe data of the preceding screen is interpolated to allow output of the playback picture.
in the aforementioned embodiment 6, in the case where the whole I picture and the whole P picture cannot be read, the playback picture is interpolated in zhe unit of area. However, the interpolation may not be performed in The area unit, but it may be performed in the unit of error correction code.
In this case, the demodulator 21 segments t,"le data into packets X c, several bytes with respect to the data arrangement shown in FIG. 30 so that an error correction code is added to each of the packets. FIG. 32 shows a case in which the data of five areas inputted to FIG. 32 in a continuous manner is divided into packets of error correction blocks. FIG. 32A shows the data string before the packet division. FIG. 32B shows the data after the packet division. In FIG. 32, the data is divided into ith through ith packets in the area PIM.
At the time of the high speed playback, the optical he,--,d JUMPS To the front of the GOP in the unit of a definite time with respect to the data which is recorded in the unit of GOP on the recording medium such as an optical disc or the like with the result that the data portion of the I picture is read in the unit of area in accordance with the header information and is demodulated at the demodulator 21 and is inputted to 9 v However, case where:i-i--oruzT:on amount -he bu"z--=-r mew. 1 n zn, e.1 i - -1 -1---L X Of Z,e T pcz,,1r-z so large that the whole T picz,_-z and zh- whole P be read in a definite time, the head iimc)s to th e f GOP even in the mi 'sz of r ding da-- 'n one arz.
ronT- c- -Ln.e nex-L c L 1 - a portion. the C'az-n that has been read is to n error correction processing, and zhe data zat = be er-,cr. cor-reczed is input-L-.--' to z.-ie bd-fer memory 22.
In this case, the formaz decocler 23 recognizes the a,-,2,-d,-ess of the I piczi.,re and -Lhe P z-La-L can be docec 11 - 1 -ch e 'a - can be reed Is d.ecDdc-z- In -Lhe ec a, -zavs so zl-la-L c La z.ha-L 10c, and, is cuzpuz:-:- as a high spe' Unit 0 macro.5u 7-rl-'s w=se, W= respect zo the inacroblock thaz camnoz be 'ec 'ed, -dhe a OC data a-f the Preced-ing screen is held at 1-z is and is ne:zion c=pensation in the a-,-or--:,menz.:Loned emb.c--',lmenT- 6n -,.' = -- ' n e-ach are-aa, bu-L _r le-d 1 C1 Z Z e -C c 0 0 _:
1 r;-rzr-ieval is set to be clcs-z,2, 1 it is not always --e,-ui:red zo be closed.
7 Nexz, embodiment 7 zill be explained 1 -,g a mzz' od for a r:--SDecz ZO FIG. 33. FIG. 33 is a view s' ow-', - Irl- specia-1 c-"---vb,-ck to embodiment 71. In enDc-ment 6, the special in a playback methoA- as in FIG. 31.
the may be iDer-LEc-,med so thaz zille playback Plc-Lure is / 1 1 - -1, outputted as shown in FIG. 33. In this case, the format encoder 23 synthesizes one screen by playing back areas one by one from continuous five frames as shown in FIG. 33. In FIG. 33A, a playback picture of one screen portion is synthesized from the I picture, the P1 through P4 pictures. Further, in FIG. 33A, the P4 picture is Played back in the area 1, the P3 picture is played back in the area 2, the P2 picture is played back in the area 3, and The P1 picture is played back in the area 4 and the I picture in the area 5. Further, in FIG. 33, the area 5 is noted with the Passage of time. The played back video data includes The I picture of the nth GOP, the P!, P2, P3 and P4, and the I picture of the n:-lzh GOP picture, and the P1 picture.
Further, 'In the case where the information amount of the I picture and the P picture is so large that all the I picture and the P picture cannot be read in a definite time, the data of the preceding screen is held as it is and is outputted to synthesize a high speed playback picture. FIG. 34 shows a playback picture in the case where the areas 1, 4 and 5 of the nl_n GOP cannot be read. In this case, as shown in FIG. 30, with respect to the data string, the area located at the center on the screen is given the priority to be recorded on the recording medium with the result that it never happens that the playback picture is hard to see because the central part of the screen in given 1 1 - e da Lhe Priorit,.- _in Further, even in the c-p-se w, -h, za in two or Ecre a =.s can net be read, one scralen areas anc -Lne i_rame;)!aye-, '----ck in each area -1s di-E,:- il is hard to see that the data is lack-i7.g in the playbeack Dicture.
Embodimenz 8 .'exz, em'.-o--:lmenz ^0 Zill be exilained.
i Z 7TG. 35 is a view a n respect- z c- 1 1 - 1 data arrangeEenZ StrUCZUre _in eMDOd-,linenZ 8. 1 - - ' 'imen7 ', z, ii -mDor-- 1 0 h= daza a r r ang e m e n z i s w r -1 e n i -, i z'-i e ord e r c f- th e a 7 aa s 3, 2, 4, 1 a n C' 5) a S SL-1071-1 In 30, bu- -Lne a--.-,ai-igement isay have a as si^tow-l in FIC. 35).
T7, L 1. FIG. 35, when the daza in zhe I and -che P pictures are at of -Lne daza ---Eor one C-OF, por-lion, the area llubz-7 at the -front is scrolled -ifor ep---',i of -,.,he In other words, as sliown in FIG. 28, in the nth GOP, -he I daza is recorded in Wle order of P1(2), PIM, P-1(4), PI(,5) wid W2.
Further, In the P2 picture, P2(3) comes to the in '03 Picture, anc the P!areas are scrolled and recorded sequen-Lially like P3(4) and P4(5).
Fu-7-zher, at the -E:-onz of -he GOP, the address where -che data oE th e I p lczL're P P-zure are szored, and information. f07 0 identifying The kind of the area at the front of each frame are recorded as header information. Here, as header 1'nformation, the area number which is recorded at the f ront of each area and the number of bytes indicating the da-L,,- amount in each area which is divided into five parts are recorded. Consequently, the optical device Jumps to the front of the GOP in the unit of a certain time at the time of the special playback so that the data can be read in The unit of area in accordance with the header intormation.
In this case, since the position where the I picture and P picture weas divided into -five parts are scrolled in the unit of frame, it never happens that the area which is not decoded is not concentrated on the fixed position on the sCreen even in the case where only a part of the 1 picture and -L' e P picture can be decoded.
areas o. the I"i At the time of the high speed special playback, the data which is recorded on the recording medium such as an optical disc or the like in the unit of one GOP is read in the unit of area unit in accordance with header information. Then, the data is demodulated by the demodulator 21 and is inputted to the buffer memory 22. However, when the information amount o-1 the I picture and the P picture is so large that the whole I picture and the whole P picture Cannot be read in a definite time, the data is read to the last with respect to the area read halfways. Then, 1 -ihe he-.d J=Ps to tIhe o -,,he C-OP to:,pLz data only of the 1 22 are-P C-2-.n be Impuzzed into the bu-i^ier 22. In 7-,-is Wse, -L".
1 23 decodes only the area of' zne i plc 7-d r -P and the P Tormaz picture, arid is ou-,;Duc-L=-d as a high spe_=c' plav'C.-Ick piezure.
1G. 36 shlows a playback Picture In z' e case w'here only Lhe I picture and -L,e P picture in one GOP are played baCk -1for a speed plavb,--c',, in z',-i-s case, when the data -Lhe i and The 'I picture -:S so la-7gz-t-,nat zhe wnole i Pictu7. a-nd E):
1. L C-In'-IC-L be read in a ZIM;Z, a nlg-,"i seed', playback is bv holding and cuzpuzz-:l7,g the dazz of -z.'ne precedIng as lz ls. "r- i G. 3 5 shows a case -,:herz- the areas 3, 4 and 5 ofF the P!- of' -Lne iiz'n C- OP cannot be re-p-d. In zh-is case, 71he of' the prec--d ing screen is Irleld as it is.
reco-,'i7.g the i picture . S des r-, d above, since the c use- fo-7 zie sDec,-'al. play',,ac'K is scrolled -L'-,e oE CrOP a s,o-.7, in FIG. 35, area. zl"li.ch cannot be decoded is not on the 11xed:.cs.z;on on zhe screen even in the case. w',-ier- only a Part of the areas o zhe I can be decoded at tme zime c: special play',.,---ck.
Embodiment 9 Next, 9 zill be ex-:)lain v 1 with r;-=s:-lecz to Flin- 37. FIG. 37 is a block clagra7, c-rj, -Lne recording i g ta Side a video signal coding Processing uniz in a d- L ":5- 1 0 video record and playback device wherein the KT block is divided into stages in the low frequency area and the high frequency area so that only the low frequency area is located at the front of the GOP. In FIG. 371, reference numeral 51 denotes a buffer memory, 52 a subtracter, 53 a DCT circuit, 54 a quantizer, 55 a variable- length encoder, 560 an inverse quantizer, 57 an inverse DCT circuit, 5,8 an adder, 59 a motion compensation predicting circuit, 60 a counter for counting the number of events and a code amount, 61 a format encoder, and 65 an input terminal.
Next, an opera-Lion of the device will be explained. The video data to be inputted is an interlace picture which has an effective screen size of horizontal 7041 pixels and vertical 480 pixels. Here, the operation of the subtracter 52, the DCT circuit 53, the quantizer 541, the variable- lengzh encoder 55, the inverse quantizer 156, the inverse DCT circuit 571, the adder 58 and the motion compensation predicting circuit 59 are the same as the counterparts shown in conventional embodiments. Thus, an explanation thereof will be omitted.
An operation of the variable- 1 ength encoder 55 will be explained with respect to FIG. 38. FIG. 38 shows a data arrangement of DCT coefficients inside of a DCT block. In FIG. 38, a low frequency component 1S located at the upper left portion and the data of the DCT coefficient of a high frequency component is located at the lower right 0 portion. The data of the low frequencv coefficient up to a specific position (the end of the events) (for example, the hatching part in FIG. 38) out of the data of the DCT coefficient arranged in the DCT block is coded into a variable-length code as a variable-length area and is outputted to the format encoder 61. Then, the variable-length coding is applied to the data of the DCT coefficient after the data of the DCT coefficient at the aforementioned position. In other words, The data in the space frequency area is coded through partizioning.
A boundary between the low frequency area and the high frequency area is referred to as a breaking point. The breaking point is set to assume a predetermined code amount in the loz-frequency area which the optical head can assess at the time of the special playback. The X variable-length encoder 55 divides the DCT coefficient into the lowfrequency area and the high-frequency area in accordance with The breaking point to be outputted to the format encoder 61.
The determination of the coding area is performed at the boundary of the event. It goes without saying that the determination can be made by other methods. For example, the determination of the coding area can be made at the boundary of the fixed number of events. The data may be divided with the quantizer 54- with the quantization data subjected to a rough quantization with the quantizer 54 and a J J differential value between a fine quantization and a rough quantization. Further, the data may be divided with coding of a picture whose space resolution is thirined to a half level with the buffer memory and a coding of a differential picture between a picture whose resolution is brought back from the half level and a picture with an original resolution. In other words, the data division is not limited to the division of the frequency area. It goes without saying that the high efficiency coded data of the picture may be divided by the quanzization and the division of the space resolution.
A t this time, wore -imDertant data as a picture is a low frequency area daza in the division by the frequency. When the division is the division by quantization, the data is subjected to a rough quantization for coding. Then the data is divided with the space resolution, the thinned picture is coded. By decoding only these important data, a decoded picture can be obtained which can be easily perceived by man. In this manner, one high efficiency coded data is divided into more lundamental and important Clata, and other data (this process is referred to as hlerarchization). An error correction code is added and modulation is performed to be recorded on a disc.
In this manner, since only the low-frequency component of the I picture and the P picture are divided, reading and playing back only 7-3 1 1 ly these low-frequency components at the time c' the special plaYback large reduces the amount of data which is read at the time of the special playb. -nck. As a consequence, time for reading data from the medium becomes shorter so that a high speed playback of a smooth movement can be realized at the time of skip search. Further, when only the i picture and the P picture a,-e arranged in a consecutive manner, the data of the lowfrequency component of the I picture and the P picture can be easily read from the disc to be decc,-ed. In This case, more efficient data structure can be made by extracting and arranging only the lowfreQuency comporieriz than bv arranging the whole area of the I picture and the P picture at the front of the GOP.
Next, an operatIon of the format encoder 61 will be explained. FIG. 39 is a flowchart showing a movement of a format encoder. In the beginning, when the encoding operation is started, it is Judged that the encoding mode is in a hierarchical mode or not. when the mode is not the hierarchical mode, information is inserted into a system stream which information is representative of the fact that the mode is a nonhierarchical mode to follow the conventional stream structure. In the case of the hierarchic,-,l mode, the setting of the sequence header is confirmed. Specifically, the data c-LE the sequence scalable extension is confirmed. When the data is written correctly, the front of the picture -, 1 is recognized so that The i picture and four P pictures are separated into the data of the loz-frequency component and z'ne data of The high area component to detect respective data lengths.
In the meantime, the length of the data of the B picture is detected for each oi the pictures. Further, a packet is prepared where only address information is recorded in the case where the data of the low-frequency component o-E the I picture and the P pictures area is arranged to follow the fron-L of the GOP. In this packet, address information for the low-frequency component of the I picture and the P picture, the high-frequency component of the I picture of the four P pictures and ten B pictures are contained so that the data length for the respective data is recorded.
Consequently, the front position of respective data stream is obtained as a relative address with respect to the front of the GOP header from this data length. The packet containing this address information and the lowfrequency component of the I picture and four P pictures as well as the remaining data are sequentially arranged to be formatted Out of them, the confirmation of the scalable mode on the scalable extension of the aforementioned sequence header refers to the confirmation of the scalable mode setting which is dezermined in the 1 syntax oi the VEC-2 of FIG. 40 and the confirmation of the description of the priority break point on the slice header. The priority break point is located at a predetermined number of events of FIG. 40 (corresponding to the aforementioned breaking point) and refers to the data representative of the boundary between the divided low-frequency component and the high- frequency component.
When a scramble mode assumes 'TV, it is shown that the Iollowing bitstream is a bitstream of a data partitioning. It is also shown that the bitstream which is divided into the low-frequency component and the high-frev,,uency component continues. When the B picture consists of the low-frequency component so that no high-frequency component is generated, the B picture is not divided.
One example of the bitstream which is generated in this manner is shown in FIG. 41. FIG. 41A shows a bitstream which is not hierarchized. When the bitszream is hierarchized with a circuit shown in FIG. 37, the bitstream is divided and hierarchized as shown in FIG. 41B. When this data is arranged in an array in consideration of this special playback, the low-frequency component of the i picture and the P picture is arranged at the front of the GOP as shown in FIG. 41C.
FIG. 41D shows a data arrangement in the case where address information is contained in the private packet as shown in a flowchart 0 in FIG. 39. In th,.s case, the address inforwation may be represented with a relative address with respect to the front of the GOP header as described above. However, the address information may be represented in such a manner that which byte of which packet is the front of each picture. It goes without saying that the address information mav be represented with a sector address on the disc as well.
FIG. 42 shows an example in which address 'information is contained in a private packet. When a packetized elementary stream (which is referred to as PES) packet is used as a private packet, the strean M is specified in BF (hexadecimal number representation). A-E-Ler describing the packet length, the byte MSB is set to 1 and the subsequent bit is set to 0 so that the code does not become the same as all The start codes (start code of packet and the start code of the 1 --on, the bitstre.am). Then the hierarchical mode, the kind of hierarchizat 1 kind of picture to be used at the time of the special playback, and the number of the start addresses or the like are described with the remaining six bits.
After that, the 21 bit long address information is described so that the GOP data amount can be represented up to a maximum length of 2 Mbytes. However, 100 (represented in binary representation) is inserted into the first 3 bits out of 21 bits of data so that the data does not ) 2 0 become te same as the front 24- bits 000001 (hexadecimal representation) 0, the start code as described above. Here, the start address includes a start address o the low-frequency component of the I picture, a start address of the high-frequency component of four P pictures and a start address of the high-frequency component of the I picture, the highfrequency component of four P pictures, and a start address of ten B pictures. Further, a sector address on a disc where the data of the preceding and succeeding GOP is recorded is added to jumD an optical head at the time of the special playback.
When the 1 bit parity is added to the 21 bit addresses, the reliability of the data is heightened. In this case, 10 (binary representation) mav be added to the front with respect to 21 bits---1 bit. Further, in consideration of the high-speed times of the special playback, the variation in the high-speed times of the special plavback is widened when the sector address of the several front and rear G0Ps is adcled as well as the address of the preceding and succeeding GOP. Further, it is shown that address information is described in the private two packets of the PES packet. It goes without saying that the sector address may be written on other user areas or the like such as private descripter of a program stream map or the like.
The playback sicle of embodiment 9 will be explained with . 1, s 0 respect to FIG. 43 and FIG. 414. FIG. 43 is a block diagram of a digital v"deo sipal decoding processing part. In FIG. 43, reference numeral '1 denotes a program stream header detector, 72 a PES packet header detector, '13 a video bitstream generator, '14 a data rearranger, 75 an address memory, 76 a mode switcher, '17 a variable- 1 erigzh decoder, 78 a switch, 719 an inverse quantizer, 80 an inverse DCT circuit, 81 an adder, 82 a prediction data decoding circuit, 83 a frame memcry and 84 a decodable determiner. FIG. 44 is a view showing an opera-Lion concept of FIG. 43.
Next, an operation of FIG. 43 will be explained with respect to FIG. 45. FIG. 45 is a flowchart showing an operation of a format decoder at the time of the playback. The bitstream outputted from the ECC is detected the header of a program stream to be divided into each PES packet. Further, the PES packet header is detected to differentiate a private packet including address information and a video packet.
In the case of a private packet, the address information contained in the packet is extracted and stored. In the meantime, in the case ol the video packet, the bitstream of the video data is extracted. Here, in the case of the normal playback, the data of the low-frequency componenz and the high-frequency component is extracted from the bitstream stream of the video data with respect to the I picture and the P picture so that the data is rearranged and a playback picture is outpuzzed. In the meantime, at the time of the special playback, only the low-frequency component c-LO the video data is extracted and played back. Here, after the low-'Lrn-uency component is played back, the optical head is allowed to jump to the front of the subsequent GOP.
In this case, when these addresses are described in the video stream, zille address information is extracted and stored after being converted into the bitstream. Consequently, in the case where the address information is described in the private descripter of the program stream map, the address information is extracted and stored at a level of the detection of the program stream header. It goes without saying that the address information may be either a relative address or an absolute address.
In actuality, a mode signal of such as skip search and a normal continuous playback or the like is inputted to the mode switcher '16. In the meantime, a playback signal from a disc or the like is amplified by an amplifier so that the signal is played back with a clock which has been subjected to a phase synchronization and outputted from a PLI, or the like. Next, a di-fferentiat ion operation is performed for digital demodulation. Then, after an error correction processing is performed, the program stream header detector 71 obtains data information that 11 1 _; follows the header. Further, the PES packet header detector 72 detects, for example, address
informazion for each picture described in the private packet of the PES packet and address information of the data for the special playback and the iniormation is stored in the address memory Here, the PE75 packet for the audio, the PES packet for such as characters or the like and the PES packet for the video are class-i4ied so that only the packet for the video is outputted to the video bitstream generator 2 Here, the video bitstream generator '13 erases added information from the PES packet and constructs a bitstream. Specifically, the data such as each kind of control code and the time stamp is eliminated. After this, in accordance with the address information obtained from the address memory '15, with the output of the mode switcher 76, the bitstream is rearranged at the time of the normal playback by the data rearranger '14.
The output (control signal) of The mode switcher 76 is supplied to the data rearranger '14 and the decodable determiner 84. The data rearranger 74 either reconstructs the data before the division from the lowfrequency component and the high-frequency component divided and hierarchized by obtaining the control signal. Ozherz-ise, only the low51,9 frequency component is outpuzted to the variable- length decoder 77. In other words, each of the low-frequency components is synthesized with the high-frequency component at the time of the normal playback so that the device is operated in such a manner that the data is rearranged in an order of the original picture. At the time of the special playback, either the low-frequency component only of the I picture or the lowfrequency component of the 1 picture and the P picture is outputted depending on the high-speed times.
At the time of the special playback that allows the passage only of the low-frequencv component, the time stamp is not used. In contrast, the variable-length decoder 11 extracts the boundary of the events in the lowfrequency component area designated by the priority break points of the slice header together with the decodable determiner 84 so that the data is decoded up to the boundary and is outputted to the switch '18. This switch '18 is connected so that zero is not inserted at the time of the normal playback. In the meantime, at the time of the special playback, the switch 78 is controlled with the decodable determiner 84 so that zero is inserted into the high-frequency component area after the priority breo-k point at the time of the special playback.
The aforementioned operation will be explained with respect to FIG. 44. Referring to FIG. 44, when the partitioning breaking point is E1 through E3, E1 -Lrough E3 is stored in the stream of the low-frequency component. E4 through EOS are stored in the stream of The high-frequency component. In the stream of the low-frequency component, the low frequency component data in the subsequent DCT block following E3 is stored.
Here, at -dhe time of the normal playback, the data rearranger 74 extracts the data E1 through E3 from the iog-frequency component stream and the data E4 through EOB from the high-frequency component stream. Further, the data rearranger '14 extracts the data respectively to reconstruct the DCT data in seQuency. In contrast, at the time of the special playback, the data rearranger 74 extracts the data E1 through E3 followed by variable- length decoding by the variable-length decoder 771, the decodable determiner 84 detects the priority break point so that zero is inserted into a portion provided with a hatching shown in FIG. 44 to constitute a DCT block by using only a low-frequency component.
The data which is converted into the DCT block is decoded in accordance with the motion vector. Here, an explanation of the decoding by the motion vector will be omitted because the decoding is the same as the conventional example. However, the reference used in the decoding of the P picture at, the time of the special playback is decoded by using the I picture or the P picture which is decoded only with the low- b frequency component.
1he data which is decoded in the unit of block is inputted to the frame memory 83. Here, the frame memory 83 restores the picture in the original order of the structure of the GOP, and outputs through the conversion from the block scan to the raster sc-nn. Incidentally, the frame memory 83 can be comwnly used with the memory incorporated in the prediction data decoding circuit 82.
The coding area is defined at the boundary cfL the events, but it goes without saving that the definition of the boundary can be made by other methods. In other words, the high efficiency coded data of the picture may be divided either with the quantization or the division of the space resolution in addition to the division of the frequency area.
1z this time, data more important as a picture is data of the lowfrequency area in the case of the frequency division. In the case of the division of the Quantization, the data refers to data coded by a rough quantlization. In the case of the data divided with the space resolution, the data refers to the data obtained by coding the thinned picture. In this case, in the playback picture decoded by using only these data items, the area which can be easily perceived by men is defined as important data. In other words, one high efficiency coded data is divided into basic and important data and data which is not so IMPOrtw71 (T-:LS P:-OCz-SS!S refe77ed to as 50 that only basic and daza cant be pia-ved back a-L -L.h-zime of -L!--. sDecial playback when T-'1=- data is played back from the disc.
7 ibes a canse, in which the side 9 descri, czr-respc-,ics to zi-le playback side. It is also considered in the case where the record and the playback is combined in a set such as a hard disc or the only T-.he Playback side is consic';-z:.ed on zine Z;nn Z-aZ the cp is recorded accor'anice ziTh, conc - Presuppcs '---1. iL _za 1 - G - - - epL 01 -1he available compact disc Cr the!"%:Z.
Nex-L, em'_-w--C'-men-L 10 zill be expla-ined.
FI.G. 1-5 is a block dlagram showing a record 0' Zhe Video signal ank' according to 10 Of the Present invenzion. Like numerals in FIG. 405 denote like P--r-L5 Or jr,g in FIG. 37. Reference numeral 65 denotes an input corresponcl 1 1 z -r- 7m inal, 51 a be-Eler MeTC7v., 52 a sub-Lraczc-r, 53 a KT circuit, SA- a Quantizz-7, 55 an inverse quantizer, E57 an Inverse D-7- circuit, 58 an =z adder, 512d a motion compensation predicting c-i-,rlc.11t, U a variablelength %r C' CZ7.
62 an rea.rra:ger, and 61 a enc '_- KG. 4-11: is a block circu- diagram showing a play'Dack system of L 1 - _ 1-W J a vide-o S-:P'al record and playback device az-c-d:-,ig zo /c, E ', Db 0 d 1, 3 e: ', z 10 - Like numerals in FILG. 1-7 denotes 11"k-= parts ol- parts in FTG. 4. Rz numeral -171 denotes a pr-,gram stream header detector, 172 a PIES Packet detector, 73 a V1Geo b-,zsz--eam generazar, 8.D an arezz rearranger, i an address memory, '16 a mode 77 a variable-leng-Lh 79 an invel.rse Quantizer, an inve-7-se DCT circuit, 81 an adder, 82 a data decoding Circuit, and 83 a mewry.
an ODEraZIOn 01 eTDOC-IMen will be 71he a'ig,zal VIdeo sig:-,.al Is in the uniz of 1 -n- f- m -,.' e Input l J. 1 1 11- 1, 10.71 - =-I. L tn 7 L terminal E5), and is suppli ' -o ' e bLf-er memory 151.
operation -from zz- bu-::7-^er memorv 51 to the encoder 55 is the same as the a---cremenzionec' example, and an e\planazion thereof will be omizzed.
The area 62 re.;-::.rranges data respect to tne I picture in a b-i-Ls7.7-e..-m ofF video OUT-PLIZLe. in U.1e Un-iZ Of C-OP from v 1 = CC, - n L Lne enc:- 55, so z' at an area locazed at the central Part Of- the screen is arranged at the front off the bl Here, the I Picture "is divided inzo three are,.-s as sillozm in FIG. AS. -Ine data of i the I to the areas 1 through 3 are de-Eined, as 1(1), 1(2), and. 17(3). However, each area shown, in FIG. 4 is a collecz-,lcn of a pl.u-,allzv ol ?EG slice lavers. I. -,TC.
11 ' -. 48, z,e area. 1 Q and 3 consists of six slices and the area 2 consists of 18 slices.
In actuality, the area rearranger 62 detects the slice header of the I picture on the bitstre-pm and classifies each slice into three areas shown in FIG. 48 thereby forming a bitstream for each of the areas for rearranging the bitstreams arranged for each of the areas. In other words, as shown in FIG. 49, the bitstreams are rearranged in the unit of area so that the b-izs,-ream is arranged in the order of 1(2), 1(3) and 1(1) at the front of the GOP. Further. the rearranged blitstreams are outputted to the format encoder 61 in the unit of GOP.
Nexz, an operation of the format encoder 61 will be explained in accordance with FIG. 50.
FIG. 50 is a flowchart showing an algorithm for,or-wttliig the video data into the PES packet in the unit of GOP. In the case of the screen central part priority mode, the picture header of the bitstreams to be inputted is detected and the picture information is detected. Here, in the case of the I picture, the central parts of the screen IM, IM and I(I) shown in FIG. 49 are extracted and respective data lengths are detected so that the data length of each area thus detected is converted into a binary number of 24 bit width thereby preparing an address information. On the other hand, the data lengths are detected in the unit of picture with respect to the P picture and the B picture so that the data lengths are converted into a binary -01.
number of 21- bit (3 bytes) width thereby preparing address information.
Further, the formatting unit collects the input address information and the bitstreams of the video data into two kinds of the PES packets. In other words, the PES packet having only the address information and the PES packet having only the audio are constituted.
Consequently, when one GOP consists of 15 frames as shown in FIG. 6, there are l'$ kinds of pictures as the address information, such as three kinds of I pictures, four kinds of P pictures, 10 kinds of B pictures. Further, as address information at the time of the special playback, there are two kinds of address information of the preceding and succeeding GOP on the disc (absolute addresses on the disc) These items of the address information are collected in one packet and are formatted as the PES packet. In actuality, these items of address information are collected in one packet and are formatted as a private 2 packet of The PES packet shown in FIG. 51. In FIG. 51, the absolute address on the preceding and succeeding GOP on the disc is arranged at the front of the packet data. Then, the address information of each picture is arranged in order. However, since 3 bytes (24 bits) long information is assigned to each address information, the packet has a length of 57 bytes.
In the meantime, with respect to one GOP portion of bitstreams other than the address data, the bitstreams are formatted into PES c) _) packets (video packets) by dividing the bitstreams into a plurality of packets and adding header information such as synchronous signals or the 1 ike.
In addition, the format encoder 61 divides The bitstreams of the inputted audio data into PES packets to constitute an YEG2-PS system strew together with the PES packets of the video data. In actually, as shown in FIG. 52, the bitstreams of one GOP portion of the video data and 1 1 1 and arranged into a Lhe bitstreams of the aud'o data are d'v'ded plurality of p-.ckets in one pack. In this case, a packet representative of the aferementioned address information is arranged at the front packet of the system stream as shown in FIG. 52. Subsequently, the device is constituted in such a manner that the packet conzaining the bitstream at the central part of the screen of the 1 picture is arranged.
Next, an operation at the time of the playb---ck will be explained with respect to FIG. 47. In FIG. 47, since an opera-Lion of the program stream header detector 71, the PES packet header detector 72, the video bitstream generator 713 and the rode switcher '16 is the same as conventional examples, an explanation thereof will be omitted.
In the decoded video bitstreams, the data az. the central part of the screen of the I picture is located at the front of the bitstream. Consequently, the area rearranger 855 rearranges the I picture data in cl 't- the order of I(I), IM and IM for each area in accordance with the data length of bits-Lre.-,ms of IM, IM and I(I) which is outputted from the address memory 75. The rearranged bitstreams are inputted to the variablelength decoder 717 to be decoded into the block data, the motion vector or the like. Here, since an operation which follows the variable-length decoding at the time of the normal playback is the same as conventional examples, an explanation thereof will be omitted.
In a high speed playback, since one GOP portion of data is assigned to one pack of a system stream as described above, there is _dered a method by wh;ch an optical head "umps to the front address consi of each GOP when reading data from a disc to read only the data of the I picture which is arranged at the front of the system stream so that the optical head jumps to the front of the subsequent GOP. In such a case, the PES packet is detected which has a record of address information arranged at The front of th system stream to control the disc drive by decoding the address on the disc of the subsequent WP and the address information of the I picture.
In the case shown in FIG. 6. when all the I pictures in each GOP can be read within one frame, a 15 times high-speed playback can be realized. When the I pictures in each GOP are read within 2 frames, a 7.5 times high-speed playback can be realized. In this manner, when the j higher speed of playback can be realized, time for reading data from the disc becomes shorter.
Further, in the case where data is read from the recording medium such as an optical disc or the like, even if the front address is known, there arises disc rotation waiting time at the time when the optical head jumps to a location of the disc where the data is actually recorded. Further, when the video signal is coded with a variable rate, the amount of information of The I picture is not definite and time required for reading the I picture also varies. Consequently, when the speed in the high-speed playback becomes higher, time for rea.ding data on the disc becomes shorter. Further, since the waiting time for the disc rotation is not definite, it becomes impossible to read stably the whole data of the I picture.
Consequently, in embodiment 10, the optical head jumps to the front of the GOP in the unit of a definite time with respect to the data recorded in the unit of GOP on the recording medium such as an optical disc or the like at the time of the high-speed playback. Thus the data part of the I picture is read from the disc. In this case, even if the whole data of the I picture cannot be read, the optical head jumps to the front of the subsequent GOP. In other words, the optical head jumps to the front address of each GOP in the unit of a certain time to read data ( ke 1.
as much as possible from the 'ront of the system stream and then jumps to the front of the subsequent GOP.
In this case, the PES packet including the address on the disc or the like of the subsequent GOP and the PES packet including the data at the central part o-L: the I picture are arranged at the front part of the system stream. Consequently, even in the case where the whole data of the I picture can not be read at the time of the special playback, at least the address on the subsequent GOP disc and the data at the central part of the I picture can be decoded, the address and the data being required for controlling the disc drive.
In the case where only the central part of the screen can be decoded at The time of the special playback, only the data which can be decoded bv the area rwrranger 835 is outputted to the variable-length decoder '17 so that the variable-length decoded video data is Inputted to the frame memory 83 through the inverse quantization and inverse DCT. In the meantime, the area rearranger 85 inputs the area information which cannot be decoded into the frame memory 83. With respect to the area which cannot be decoded, the data outputted in the preceding frame is held as it is and is output-Led.
FIG. 53 shows one example of playback picture in the case where a highspeed playback is performed by playing back only tille I pictures /1 0 from the nth GOP to the n-4th GOP. FIG. 53A shows a case in which the whole 1 picture cp-n be decoded. FIG. 53B shows a case "n which areas 2 and 3 can be decoded. In the area 1 which cannot be decoded, the value in the preceding --"'rame is held as it is and outputted. In additlon, FIG. 53C shows a case in which only the area 2 can be decoded. In the areas 1 and 3, the value in the preceding frame is held as it is.
Here, in the general video signal record and Playback device, a format I's adopted in which the I picture is recorded in the unit of frame at the time of recording. In contrast, in FIG. 52, the area located at the central part of the screen out of The I picture data which is divided into three parts is arranged at the front of one GOP by giving a priority to the area. ConseQuently, even in the case where the area of onls, part of the I picture can be read from The disc in a definite time at the time of the special playback, the playba-ack picture at least at the central part of the screen can be outputted.
As described above, in embodiment 10, as shoz-ii in FIG. 52, with respect to the I picture for use in the special playback, data of the area located at the center of the screen is arranged at the front of one GOP so thaz the area is given a priority to be recorded on the recording medium so that The area 2 located at the central part of The screen is given a priority to be played back even when the speed in high-speed 1 playback is high so that the content of the high speed playback picture is easy to see. Further, the special playback is performed in which the optical head jumps to the 'Llront of the GOP in the unit of a definite time with the result that an output screen can be renewed at a predetermined high-speed.
Incidentally, the aforementioned embodiment may be constituted so that data of an area that can be decoded at the time of the special playback is all outputted, and for the area whose daza cannot be decoded, the data of the preceding frame is held as it is. However only the central part of the screen may be played back at the time of the special playback.
In this case, the area rearranger 85 decodes the data only of the area of the I picture which is read from the disc. With respect to the areas 1 and 3 whose data is not decoded, for example, it is masked by a gray data to output a high speed playback picture at the frame memory FIG. 54 shows a playback picture in the case where only the area 2 of The I picture from the nth GOP to the n-4th GOP is played back for the high speed playback. In FIG. 54, the areas 1 and 3 on both ends of the screen in FIG. 5A. are masked by gray data. Further, even in the case where the information amount of the I picture is small, the waiting C7 1 time for the disc rotation is short, and sufficient time is available for reading the data of the areas 1 and 3, the data of the areas 1 and 3 are not decoded.
This is because the high speed playback picture becomes unnatural if the data of the areas 1 and 3 are outputted on the screen only when they can be read, and the areas 1 and 3 are not renewed in a certain interval. Consequently, when only the central part of the screen of the I picture is played back at The time of The special playback, the area to be renewed becomes constant so that the playback picture becomes free from unnaturalness.
rurther, in the aforementioned embodiment, only the area of the central part of the I picture which can be decoded at the time of the special playback is displayed to mask both ends of the screen. However, the central part of the screen may be extended to a size of one screen and outputted.
In this case, in the f rame memory 83, the data of the decoded area 2 is extended to a size of one screen as shown in FIG. 55. However, in the case of FIG. 55, the central part (FIG. 5SA) of the are.;-:, 2 surrounded by a dot line is extended to double the size by linear interpolation in the horizontal and vertical directions. In other words, in the case of FIG. 53-, the part surrounded by a dot line has a 1 or size of horizontal 360 pixels X vertical 240 lines. This dot part is extended bv linear c-Interpolation to a size of one screen consisting of horizontal '120 pixels X vertical 480 lines.
Accordingly, when only data of the area of the central part of the screen is decoded at the time of the special playback to extend the central part to a size of one screen. The area whose data is ouTputted become small. In this way, however, the masked part at the both ends of the screen which is conspicuous when only the central part of the screen is outputted can be eliminated.
In the aforementioned embodiment, only the central part of the screen of the I picture is given priority to be arranged on the bitstream. However, another constitution is also possible in which the central part of the screen of the P picture as well as that of I picture are given priority. In this case, the data of the central part of the screen of the P picture is arranged after the bitstream of the I picture.
In the aforementioned embodiment, the picture data is rearranged in the unit of area after the data is converted into the bitstreams. However, the picture data may not necessarily be rearranged after the data is converted into the bitstreams. The picture data may be rearranged before the data is converted into the bitstreams.
FIG. 56 sl"iows a flowchart of the playback side of embodiment j C, 10.
77e n7-ocecu-e c-l the lol;char-L is alrea'- desc,-,,bed above and is omitted 1-0 is described by side with the side. There is also considered a case in which the C- record and pla,,:ack consz"zuze a pair like a ard ' se. 7r, er s a 1 so conside.red a case z'nere -L',.ie playbanck side on a prz.s,.,nn-esizicn that data is recorded ip. with The supposition 11ke a Clu-7-7enz compact disc.
71urzhz-:., -it goes Without sa.ving that the daza re?-rrapge-.Dent c:-the S c r le le r, W1 7 h e un, -. c,: a:-e.;- w-'n be re-;.!"ze--:n---hedata C,eco--L!^g 82 and zhe --E--ame memory 83 ",v usip-g the dal- D ' - 1 rl - - lower 8-' 1- loing slice vertical pes-ition of -L' e si:cz start code in the slice - Embod -r- 't 1 exz, emDoA--imenz 11 zill be explained.
FIG. 5-7 shows a Liglizal videez) signal coding processIng un-1z in a record and Playback device where-in the DCT block is inzo a loz---.--,eqliency are and a araa. Furzher, FIG. 3-1 is a bicc diagram on the recording si 'e in z-i,h the screen is 1G - ILIC - d: o ';:)1ural i tv 0-, a-ea lc- cc-r"-L:-a, Dart - -. L. S so a t L o-1 the screen oi: the area, is given a priorizv to be arrang d at the front 1 - 1 ec 1 of the COP - L7 F1, C. - Z57, reference numeral 62 denczes an a7aa ( (-! L a rearranger. Like parts and corresponding parts in FIG. 57 are denoted by like numerals and an explanation thereof will be omitted.
,.ext, an operation of the device will be explained. This video data to be inputted includes an effective screen size with horizontal 704 pixels x vertical 480 pixels. The motion compen-c,.:--L-;on and The DCT are used to apply high efficiency coding to the picture data. Here, the operation up to the data division and hierarchization is the same as embodiment 9, and an explanation thereof will be omitted.
Embodiment 11 is the same as embodiment 9 in that the data mav be divided with the quantification and space resolution as well as with the frequency area with respect to the division hierarchization. In embodiment 11, important data further divided and hierarchized with the data rearranger 62 is divided for each area of the screen as shown in embodiment 10 so that the central part of the screen is given a priority to be arranged at the front cf: the GOP. In other words, the data is divided into important data and data which is not important so that the data is recorded on the disc in the priority order which is preliminarily determined in an area.
In this manner, the low-frequency components of the I picture and the P pictures are divided so that the central part of the screen is given a priority in the arrangement. When onIv the central part of the i, -3 0 screen of these low-frequency components are read and played back at the time of the special playback, the data amount which is read at the time of special playback will be largely decreased. Consequently, an allowance can be made for the reading speed from the recording medium so that an extremely fast skip search can be actualized at a speed of more than ten times speed or tens of times speed.
Here, the central part of the screen of the low-frequency of screen is arranged at the f-,ont of the GOP, and the data cl' the P pictures are arranged following the data of the peripheral part of the screen in the low-frequency area of the I picture with the result that a high-speed playback can be realized at more than ten times speed or tens oi times of speed by playing back only the central part of the screen in the low-frequency of the I pict-Ure. Further, The central part of the screen for the low-frequency component of the I picture and the P picture -:)r the special playback has a small amount of data so that the data in The central part of the screen = be easily read and decoded frow the disc. Thus a high-speed playback can be realized at several times speed In other words, since the data amount at the central part of the screen for the low-frequency component of the I picture and the P picture the has a smaller amount of data than that of entire low-frequency component, the special playback can be realized at a speed faster than embodiment Q+ 1 g.
Next, an operation of the area rearranger 62 and the format encoder 61 will be explained. FIG. 58 is a flowchart thereof. In the beginning, when the encoding is started, the slice header of the I picture of the lowfrequency component partition is detected so that each slice is classified into three areas shown in FIG. 48. Then bitstreams for each area are prepared for rearranging the bitstreams collected for each area. In other words, the data is rearranged for each area so that the bitszreams are arranged in the order of a lowfrequency area. 1(2), a low- frequency area IM, and a low-fre,-uency area IM at the front cl' the GOP with respect to the low-frequency area I picture like FIG. 49.
Then, in the case of the screen central part priority mode, the picture header of the bitstreams to be inputted is detected to detect the picture information. Here, in the case of the low-frequency area I picture, the central part E2) of the low-frequency area screen, the low-frequency area 1(3) and the low-frequency area I(1) are extracted to detect the data length thereby preparing address information from the data length of each aren-. In the meantime, in the case of the P picture and the B picture the data length is detected in the unit of picture thereby preparing address information. In modes other than the screen central to S 0 part priority rwde, the operation follows embodiment 9.
Next, the hierarchical mode is Judged. In modes other than the hierarchical mode, information is inserted "into the system streams which represents that the mode is non-hierarchical thereby following the In the case c.^ the hierarchical mode, the setting of the sequence header is confirmed. Specifically, the data of the secuence scalable extension is confirmed. In the case structure of the conventional streaw where the daz,.- is correctly described, the front of The picture is recognized with the picture header so that the low-frequency area data in the I picture and the P picture rearranged in the screen area is extracted and the data length is detected. in the meantime, the data length of the B picture is detected for each picture. Further, a packer- is prepared wherein only address in-formation is
recorded in the case where the screen central part of the low-frequency area of the I picture and The P pictures is collected at the front of the GOP. This packet includes the screen central part of the low--,Lreque,, icy area part of The I picture and the P picture, the peripheral part of the screen, the high-frequency area part of the 1 picture and P pictures, and address information of the B picture so that the da-L,- length O-LO respective data is recorded. Consequently, the front position of res, oective data streams are obtained as relative address with respect to i 0 - 1 c the front of the GOP header.
FIG. 59 shows bitstreams prepared in this manner. As shown in FIG. 59C, the low-frequenc-v areas of the I picture and the P pictures rearranged in the unit of area are arranged at the front of the C-OP. Consequently, FIG. 59D shows a case in which the rearranged data of lowfrequency areas are packetted so that the address information is arranged in the private 2 packet as shown in the flowchart of FIG. 58. In this case, the address information may be represented with a relative address with respect to the front of the GOP head as described above. Otherwise, the address information may be represented in such a manner that which bvte of which packet falls on the front of each picture. It goes without saying that the address information may be represented with a sector address on the disc in addition to iz.
FIG. 60 shows an example in the case where the address information is contained in the private 2 packet. In the case where the PES packet is adopted as the private 2 packet, the stream ID is set, so that the hierarchical mode, the kind of hierarchization, the kind of picture used at the time of the special playback and the number of start addresses are described. Here, the start address refers to the start address of the screen central part of the low-frequency area of the I picture, the start address of screen peripheral part of the low- r '? frequency area of the I picture, the start address c-E the remaining B picture.
Further, a sector address of the preceding and succeeding GOP on the disc is added for allowing the optical head to jump at the time of the special playback. In this case, when the sector address cl' the several front and rear GOP is further added in addition to the address X 0.1 the preceding and succeeding GOP in consideration of the high-speed times at the time of the special playback, the variation of high-speed times cl' the special playback will be widened. Further, it is shown that the address information is described in the private 2 packet of the PES packet. It goes without saying that the address information may be described in the private descripter of the program stream map, other user areas or the like.
The playback side of the device in embodiment 11 will be described in accordance with FIG. 61. FIG. 61 is a block diagram of the digital video signal decoding unit. Like parts or corresponding parts in the figure are denoted by like numerals, and an explanation thereof will be omitted.
Next, an operation of FIG. 61 will be explained in accordance with FIG. 62. FIG. 62 is a flowchart showing an operation of the format decoder at the time cl' the playback. The bitstream outputted from the 103 9 ECC is detected the header of the program strew and is separated for each of the PES packet. Further, the bitstream is detected the header of the PES packet -EC differentiate the private packet containing the address information and the video packet.
In the case of The private packet, the address information contained in the packet is extracted and stored. In the meantime, in the case of the video packet, the bitstreams of the video packet is extracted. Further, in the case of the private packet and normal playback, or in the case of the video packet, the data of the lowfrequency component and the highfrequency component is extracted from the bitszreams of the video data of the I picture and the P pictures so that the data is rearranged for outputting a playback picture.
In the meantime, in the case of the private packet and the special playback, it is judged in the beginning whether or not time is available for playing back the whole low-frequency I picture. In the case where time is available for the playback, it is further judged whether time is available for playing back the low-frequency P pictures. The aforementioned two or one judgements are made. Thus, in the case where time is available for playing back the low-frequency I picture and P pictures, the I picture and the P pictures are plaved back. In the case where time is available for playing back the whole low-frequency I 10.7 0 1 11 picture but time is riot available for playing back the low-frequency P pictures, only the lowfrequency 1 picture is played back. Further. in the case where time is not available for playing back the whole I picture, the central part off the screen of the low-frequency I picture is played back. In The aforementioned three cases, the optical head is allowed to JUMP to the front of the next GOP.
In the case where these addresses are described in the bitstre.ams, The address information is extracted and stored after the bitstreams are formed. In the case where these addresses are described in the private descripter of the program stream map, The address information is extracted and stored at the level of detecting The program stream header. It goes without saying that the address information may be either relative address of the program or the absolute address of the program.
In actually, as shoi-n in FIG. 61, the mode signal for skip search, normal continuous Playback or the like is inputted to the mode switcher 76 from the microcomputer. In the meantime, the playback signal from the disc is amplified by an amplifier and the signal is played back with a clock which is outputted from a PLL and in which the phase is synchronized. Then, the signal is digitally modulated and an error is corrected to restore a program stream. Further, information is obtained (0 4 as to the data which follows the header by The program stream header detector 71 for dezecting each head of the program st.re-am.
Further, the address "information for each picture and special playback data (low-irequencv data and data arranged by the area of the screen) which are described in the private 2 packet of the PES packet is detected by the PES packet header detector 72 and the information is stored in the address memory '15. Here it is judged whether the PES packet is an audio PES packet, a PES packet such as characters, or a video PES packet so that only the video PES packet is ouzpu-L,ed to the video bitstream generator 713. The video bitstream generator 73 eliminates the header removal of the PES packet to output The bitstreams. After this, in accordance with the address information obtained -from the address memory 715), the data rearranger '14 rearranges the bitstreams outputted from the mode switcher '16 and outputs the bitstreams in the normal playback.
The output (control signal) from the mode switcher 76 is supplied to the data rearranger 74 and the decodable determiner 84. Here, the data rearranger '14 synthesizes the low-frequencv component and the highfrequency component hierarchized and rearranged for each of the areas and outputs the synthesized components. In the meantime, either the data only of the low-frequency component or the 'data only of the 1 0 low-frequency components at the central part of the screen are outputted to the variable-length decoder 717 at the time of the special playback. In other words, at the time of the normal playback, the low-frequency components of the I picture and the P pictures are rearranged in the order of areas on the screen. Then the low-frequency components are synthesized with the high-frequency components so that the device is operated to rearrange the data in the original order of pictures. At the time of the special Playhack, the area of the lowfreruency components of the I picture at the central part of the screen and the area of the lowfreQuency components of the I picture and the P picture at the central part of the screen are switched over to be output-Led. The time stamp of the PTS and The DTS are not used at the time of the special playback which uses only the low-frequency components.
In contrast, the variable-length decoder 771 extracts the boundary of the events in the low-frequency components region denoted by the priority break point of the slice header together with the decodable determiner 84 so that the data up to the boundary is decoded to be outputted to the switch 78. The switch 78 is connected so as not to insert 0 at the time of the normal playback. At the time of the special playback, the switch 78 is controlled by the deccdable determiner 84 so that 0 is inserted into the high-frequency components after the priority 1 1 111 1 - t_ bre..-k point.
An operation concept cl' The decoding of the low-frequency is the same as FIG. 44. An explanation thereof will be omitted. Further, at this time, the rearrangement on the screen area is the same as explained in embodiment 10. An explanation thereof will be omitted.
The coding area is defined at the boundary of the events, but it is needless to say that the boundary of the events may be defined by other methods. For example, the coding area may be divided by the end of a predetermined number of events, or the coding area may be defined by dividing the data by the data subjected to a rough quantiZation by the quantizer 54-, and a differential value between the rough quantization and a fine quantization. Further, the data mav be divided with the coding 0J. the picture whose space resolution is decreased to the half by thinning and the picture whose resolution has been restored to the original level from the hall' level and the differential picture with the picture with the original resolution. In other words, it goes without saying that the high efficiency coded data of the picture may be divided by the division cl' the quantization and the space resolution in addition to the division of the frequency region.
At this time, more important data as a picture refers to the data in the low frequency region in the case of the frequency division.
1/ in the division o-i by the quantization, the mere Important data refers to the data coded by the rough quantization. In the case of the data divided by the space resolution, the wore impertant data refers to the data obtained by coding a thinned picture. in such a case, with respect To the playback picture decoded by using only these items of data, a region which can be easily perceived by man constitutes the more important data. In other words, one high efficiency coded data is divided into more basic and more important data and the data which is not important so that the Cata which is basic and important is played back at the time of the playback from the disc.
Embodiment 11 is described by corresponding the recording side to the playback side. There may be a case where record and playback 1 1 1 Further, a case 's considered wherein constitute a pa r like a hard disc. 1 1 only the playback side is given on a supposition that the data is recorded in accordance with the presupposition like compact discs. Further, with respect to the component rearrange for e-,ch of the areas of the screen, it goes without saying that a method for outputzing a screen as shown in FIG. 54- and FIG. 55 in embodiment 10 is available. Further, it goes without saying that the rearrangement in the unit of area on the screen can also be realized with the pre-Aliction data decoding circuit 82 and the frame memory 83 when the data of the slice vertical c -1:co:.z 7 Further, in emb,^:iDenz only the basic pc S 1 -L ii c. 11 -L1-1 e --- -1; - n e -2 - - -1 - Cia.a C1: i is bY the area ofF sc--,--=-n.
is needless to sa-v. -Lraz -L'rle dazan mav be divided zith -LI"e lcw---,Ej- =--uency off -dhe p p-iczu--e or Others.
Embod-imen.rl 1? 12 "ill be ex^;.1alned With b;l-- L resuect zo FIG. SMI. 7TC.. ^3 is a block diagram shcwi.-,ig a video sigpial cc-:17,g processing in a d-ig1T-;-l video sipal recorl and 17, FIC. 53j, 11.01 and 1D4 C'=-711.0,te pr.eprocessc7--s, 102 and 1105 motion vector deze-c-tors, 1031 a rewlu- Lic-iII converter, 1-06 and 107 subz..-aczer-s, 1-08 and 109 DC-1 liO and 111 quanz-ize.rs, 1-12 and 113 va--iable-leng-,,h ercce'ers, 111- and 115 inverse quantizers, 116 and!!-'I 1n-,;e.-se KT circuits, 118 and accers, 120 and 121- liage- me:.-.7-ies, 122 and 123 rate conz-.oll----s, i21= a inverse converzer and 1-25 a data reconstruczor as a daza arranging me.ans. rurzhz-r, 1G. 63) s'i-cws a first encoding mcans an a seCOnd encodii-Lg means as one example. In particular, the s.jbz.-acz--1.r 1-06 ouzDuts a be-rl-ween tne --E.:Lrsz, means and c Lhe means in the course o' z-jo =x7, an operation of embc,(-,,iment 12 7M be The videc da7-, is to z'nz- reso!U7-10n Conve.riel- 103 in an Order cl, the 0 raster scan of the interlace. The inputted video data is filtered and Thinned for preventing repetitive noises in the high-frequency region with-the resolution converter 103. FIG. 641 is an explanatory. view explaining The concept of this resolution conversion on the picture. For example, in the case ot: the data of horizontal 1 'i pixels and vertical 480 pixels, the data is filtered followed by being thinned into horizontal 352 pixels and vertical 240 pixels with a half resolution thereof respectively thereby being converted into a low resolution screen data.
This low resolution screen data is converted from a raster scan into a block scan by being inputted into the preprocessor 1041. Here, the block scan means that the data is sent in an order of the block of DCT. The I picture is coded without performing a calculation between frames us the output of the frame memory for intra-frame coding.
In the case of the I picture, The image memory 121 which is an input of the subtracter 107 outputs nothing so that the video signal passes through the subtracter 107. This data is orthogonally converted into the frequency component by the DCT circuit 109. This orthogonally converted data is inputted into the quantizer 111 and quantized in an order of being scanned in a zlgzag manner from the low frequency region. Further, the quantized picture data is converted into an entropy code via 1 1, ( 1 4 the variable-length encoder 113 to be outputted to the data reconstituting device 125.
In the meantime, the data quantized by the quantizer 111 is subjected to the inverse quantization with the inverse quantizer 115. Then, the picture data is inversely converted into daza of a space component from a frequency component data by the inverse DCT circuit 117. The I picture is decoded without calculation between frames performed by using the output of the frame memory which is subjected to the intra-frame coding. Consequently, in the case of the I picture, since there is no input from the image memory 121 of the adder 119, the data passes through the adder 119. -An output of the adder 119 is used as data stored in the picture memory 121. At least, the I picture data, or the I picture data and the P picture data is required to be stored in the picture memory. That is because the data of I plc-Lure and P picture is needed for decoding the B picture normally at the XPEK1 and XPEG2 as reference data.
Further, the image memory 120 inputs the output from the adder 118 of the decoded data and the result of the restored number of pixels by interpolating the pixel by the resolution inverse converter 124 to store the decoded data of the picture averaged with a certain weight. With respect to this weighting, there is described a case in which a 1-1-7 1 weight of 1 is used as the output of the resolution inverse converter and a weight of 0 is used as the output of the adder 118 for simplicity.
Further, the input video data is bufiered bv the preprocessor 101 to be scan-converted from the raster scan to the block scan. Then, the video data is subtracted by the subtracter 106 from the data of the image memory 120 which stores a signal subjected to the aforementioned low resolution processing (this is referred to as resolution residual component). The resolution degree residual component is orzhogonally converted into a frequency region to be converted into the scan from the low-frezuency region to be appropriately quantized by the quantizer 110. This data is coded into an entropy coded via the variable-length encoder 112 and is outpuzzed to the data reconstructor 125.
In the meantime, the data quantized by the quantizer 110 is inversely quantized by the inverse quantizer 114 and is inversely converted in the data in the space region at the inverse DCT circuit 116. The adder 118 adds the input from the image memory 120 which is the inversely converted data subjected to the low resolution processing with converted data with the output of the inverse DCT circuit 116 to obtain the result of decolll-ing c]': the data which is formed into two layers with the low resolution data and the data c-E the residual component as one example of the data other than the low resolution data. This L L _ 4 r layer is determined by the frequency of resolution conversion. It is possible to form the layer into three layers by performing two resolution conversions. In the same approach, it is possible to prepare data in any number of layers with The similar approach.
With, respect to cod-ing of the normal PEG, the I picture and the P picture are decoded and stored as a decoded data to code the B picture by per-forming a bidirectional prediction with the I picture and the P picture. In this manner, the I picture and the P picture are coded followed by the processing c-LE the B picture.
The aforementioned coding processing of the I picture, P pictures and B picture are performed with respect to both the low resolution component and the high resolution component. In this manner, a sequence can be constituted wherein the low resolution component R (hereinafter referred to as R component) and the resolution residual difference component S are arranged side by side. The operation is performed bv the data reconstructor 125 so that the data is arranged at a place such as the front of the GOP to which the optical head can favorably access. For example, the data is arranged as shown in sequence a cl' FIG. 65. when the data is rearranged as shown in FIG. 65 and the half of the area which is occupied by the L component, the low resolution component can be played back. The resolution residual 11 cl ell difference component has a smaller data amount than the non-resolution LI1 1 dual difference component, and the data can be e'-Eic'ently resi 11hierarchized. In other words, here, a first encoding means for coding in accordance with predetermined conditions and a second encoding means for coding The residual difference of coding using the f'L.7st encoding means as an example of a video information other than coded by the first encoding means out of video data are provided for an efficient hierarchization.
FIG. 65 is a view showing an example of the result of data constitution. In FIG. 65, a sequence a is a sequence generated by the coding processing of embodiment 12. A sequence b is a sequence generated by the coding processing in another embodiment. A sequence c is a sequence generated by the coding processing in iurzhe-. another embodiment. In the sequence b, symbol L denotes a low frequency component, and H a high frequency component. In the sequence C. symbol C denotes a component coded by a rough Quantization, and A a residual component bv the rough quantization, respectively. As shown, in the sequence a in FIG. 710, the aforementioned operation is performed with respect to only the I picture and P pictures. Only the component may be arranged in summary at the front of the GOP.
In This manner, when only the low resolution component is t ?.,- arranged in summary at the front of The GOP, the ratio of the L component occupying tne who!-- largely reduces so that an allowance can be made in the reading speed from the medium so that the skip search can be easily realized. In addition, like a sequence a, when only -.'rie R component of the I picture and zhe P picture are arranged in summarv at the front of the GOP, the operation is per-formed so that only the low resolution data of the I picture and the P picture are decoded. In the aforementioned embodiment, an explanation is given to a case in which the Thinning ratio is horizontal 1/2 times, and vertical 1/2 times. It goes without saying The ratio can be set to a value different from the aforementioned value, but an arbitrary ratio can be applied to the embodiment.
Further, The coding mode includes the XPEGI, XPEG2 and 1PEG or the like. in the hierarchization of the resolution, a common coding technique is not necessarily adopted. That is because when the data is coded by lowering the resolution, it is possible to suff iciently correspond to the coding with the XPEG1 mode. In addition, in the JPEG mode, the lamination of one frame on another constitutes a mobile picture. Consequently, it is possible to decode correctly the data even when the data occupies a specific position of the GOP. In addition, the explanation is given with respect to two degrees of resolution, but is goes without saying that larger number of hierarchies can be used. The 12 1 C. ' data - d112terenzial ccmpc7,=-nz may be co 'ed in -,he -LIrle o 11' the low comDonent is coded, wlzh the firs7_ means in FIG. 63; zhe ouzDu-- tErom zh--15 firls-L encoeding means is i;,n7,=-rpc lazed; the c compcnenz e p,,zu.re before pixzls an' an A da.a is obtained zich the 106;, d the ;.z-C-iierz.nz-lal ccm:,:nen: is coded by a d-i--El'.-r--nzial ccEponz-nt encoding leans - S nC71.n-- 70U rj- f rom t' ,-,e frame read -rom The image memory]V 'C7 g: L e :ct-on Wizh z'I'le of low resolution 1 1. 7 lu, zh e ec -ec 1 C razes daza:s ' to be stoi j mewry aj Cress) ad US-Ling the zille p-X"Ls.,-c gops 7 thout llaz an addl:.-Lg means may be prov.;d;-tLo a'd ay ing -L,- addizional 17,7-C71a-lion s=n as an auclo sipal, a header or -,'ne like, and an e--z- cc-7:,crion signal to zhle cDm::c,.nent.
Em'c,od-imenz 13 13 explained on 1 b,. " 5 0 - - c ed, i - S x Ill embodiment 13, the DCT block is divi Lhe 1. - - L C.:) all layers c-.,: a iow--E:equ=ncy region and a higli--EL7.eque,-icv region th Z 1g -.1 ' ---LLne only tille low region is arrair ed at Lhe 1 - 1; p 66) is a '--'-,ock d_iaz-cam 07- Zine digital video sipal cc g 1 - - - ing proc=-ss 7, rT-. PE, re-Ference nume-als 126 and 12, dznoze -;rsz var-abl i i c Gc --- - 1 1 - 1 - 1 le- 1 1 i c,- length encoder and a second variable-length encoder, respectively. Like parts or corresponding parts in FIG. 66 are denoted bv like numerals in FIG. 63, and an explanation thereof will be omitted.
Next, an operation will be explained. This interlace video data is a data item which has, for example, an effective screen size of horizontal 704Pixels and vertical 480 pixels. Since the I picture is decoded without performing calculations between frames using the output of the frame memory subjected to the intra-frame coding, the video data is passed thro ' and outputted. This video data is orthogonally converted into the frequency component by the KT circuit 108, and is converted into the block scan from the low frequency region. Then the video data is converted into a block scan from the low frequency region to be appropriately quantized by the quantizer 110.
The data arrangement of the KT coefficient inside of the XT block is shown in FIG. 67. In FIG. 67, a low frequency component is located at an upper part on the left and high frequency component is located at a lower part on the right. Out of the KT coefficient data arranged in this XT block, the KT coefficient data (for example, a hatched part in FIG. 67) in the low-frequency region up to the data of the KT coefficient at a specific position is entropy coded via the first variable-length encoder 126 as a low frequency region extracting means and j - '? is outputted to the data reconstructor 125. Further, the second variable- 1 ength encoder 127 performs the variable-length coding of the data of The DCT coef-ficient after the data of the DCT coefficient at the aforementioned specific position.
That is, in this manner, the data is partitioned and coded in the frequency region.
With respect to the coding of the motion vector and the DC component, the coding may be performed only by the -first variable- 1 ength encoder 126. The second variable- length encoder 127 is not required. That is because at the time of the normal playback, the output data of the first variable-length encoder 126 and the output of the second variable-length encoder 127 may be synthesized and coded.
The determination of the coding region is performed at the fixed position of The DC-1 coefficient. The determination can be made by other methods. For example, the coding region may be determined with the fixed number of events. In other words, a unit for providing a Hufflan code which is a variable-length code is an event. The coding region may be set with a predetermined number of events such as a unit of three or the like. In an example of the output bitstreams at the data reconstructor 125 with the arrangement of sequence b in FIG. 65, the low frequency region picture can be played back when only the first half of the lowfrequency region is read. The coding region may be determined in a i - - 1 c variable manner at the arrangement such as sequence b shown in FIG. '10.
In zhe meantime, the data quantized by the quantizer 110 is subjected to an inverse quantization. Then the data is inversely converted into the data in the space region by Lhe inverse DCT circuit 116. The I picture is decoded without performing The calculation between frames using the output of the frame memory subjected to the intra-frame coding. Consequently, in the case of the I picture, there is no input from the image memorv 120 cl' the adder 118. Consequently, the data is allowed to pass through the adder 118. The output of the adder 118 is used as data stored in the image memory 120.
At least, the I picture and the P pictures are required to be stored in the image memory. That is because the data of the I picture and the P picture is norm 1 ly-requi red as reference data for the decoding of the B picture normally at YEGs 1 and 2.
When constituted in this manner, the ratio of the L component largely reduces so that an allowance can be made in the reading from the medium, enabling to realize a skip search. Further, as to be described later, when only the I picture and the P picture are arranged in summary, the device c-an be operated so that only the data of the low-frequency component can be easily decoded. Since the data in the highfrequency region has smaller amount cl' data than all other regions of data, the
2 -41 1 1 '1 efficient constitution of data can be made possible than extracting data in the low-frequency region and storing the data before the data in all the regions.
When the coding of the I picture is ended by allowing the I picture to pass through the subtracter 106, the 8 picture is coded in the bidirectional prediction with the last P picture in the GOP preceding in terms of time. The output of the preprocessor 101 and the data from the memory cl' the reference frame (arrow in the drawing omitted) are compa. red with each other so that the motion vector is detected and the prediction mode and the frame structure are judged. On the basis cl' the result of judgment, the data of the reference frame memory in which the output of the preprocessor 101 and the Clata from the reference frame memory most favorably agree with each other is read as the data in the for-zard direction portion and the backward direction portion from the frame memory 120. Consequently, the data read in this manner and the output result of the preprocessor 101 cli the B picture are subjected to subtraction by the subtracter 106. (This result is referred to as timeresidual component with respect tO both the P picture and the I picture). This time residual component is subjected to the DC1 calculation so that the result is quantized and is subjected to the variable-length coding.
12-& Fmbodimenz 14 Embodiment!I- ill be ex-plained on the CC basis of 17TIG. 68. In 14, the daza is dv-,d ' into zhe rough L c,.uan-c.,zaz-ic,-i comDer.e-lz of zne DCT coel-flicier, and the rough rzsl 'lip-! Cl i -F-:
lerence coiripc.ne.",z as an example of zhe da.72 other than the rough C:)rponenz SO that ie rough cuaiiz-;zazion component is ng arrang--1A a: -Lhe ofE the C-OP. FIG. 68 is a block C'-agram a digital vide-o signal Processing wii-c. In F I C. eln, numeral -128 d-enozes a sLibz.-acter and 129 all T cl - - cor-?-espot-idi.-lg par..s In FK. SO' a-,- dep.ozed bv like -=,zrals in F7G. 63, and an explanat-lon zill be c w IL T- z ed.
1 1 1.
NexT, ' an CDerazion of embod imenz 114 wi 11 be explained. The -inpuz z:,-;s has, for exam;)- e, an sCreen Size of- 704 Pixels and verzilcal 480 Pixels. The I plicture is decoded w.;-Lhouz per-forming zhe calculation bet-jeen f7_ ames usIng -Lhe L _ Zn._ n- ouzPt., c memer.v z' Ich is subjectto the liitra--^r-a[De coding.
the Consequenzly, in zIne case oF the T p I c7-ure, no ' 1,..4n i ng, 5 input Led LO plc-Lure memcry is the input of zIne zille result thaz the vide-C signal;--sses through z-he 06 - This data is 1 - or-Lhogo7,,all; conve.rzed inzo the -Lcret-,enc-v compenenz and the DC, circuit 108, is conve7zed into -,he block scan frou: -Li"le lo- f-req,,1erics, region.
i)-7 Then, the quantizer 110 performs an appropriate rough quantization which reduces the coded data amount to less than half. This quantized data is coded into an entropy code via the variable-length encoder 112 to be outputted to the data reconstructor device 125).
In the meantime, the data quantized. by The quantizer 110 is subjected to inverse quantization (the result is referred to as the result of the rough quantization). The data subjected to the inverse quantization is sent to a different coding processing unit (part denoted by a dot line frame in FIG. 68). In the meantIme, the data is inversely converted into the data in the space region. Here, the coding of the I picture is described. Although there is no output from the image memory 121, in normal cases, the coding result at this coding processing unit is stored in the image memory 121 so that the data is subjected to motion vector detection at the motion vector detector 102, the prediction mode is determined and the DCT block mode is determined. The position data suitable for the determined mode is referred to, and is inputted to the subtraction input side of the subtracter 1071.
The output of the subtracter 1071 is subjected to the DCT to determine a residual difference (which is referred to as a rough quantization residual difference) with the result of the rough quantization and the subtracter 128. The rough quantization residual (.2 1 a difference is finely quantized (fine quantization on -,he same level as the normal coding in consideration of the coding amount control) to perform the variable-length coding while being inversely quanzized, subjected to the inverse DCT and decoded to be stored in the image memory 121. The result of this coding and the coding result of the rough quantization determine the allocation of necessary data and a header or the like is added.
As an e\axple of This output data, the sequence c shown in FIG. 65 is adopted, the decoding result of the picture which is subjected to the rough quantization is obtained only bv reading the first half of the GOP. In addition, since the data of the rough quantization residual difference is small compared with finely quantized data, a more dataefficient constitution can be obtained than storing the extracted rough quantization data before the fine quantization data.
Further, as another example, variable processing such as an arrangement of the sequence c shown in FIG. 70 may be performed. Thus constituted, the ratio of the C component (component coded by per-fiorming the rough quantization) out of the whole largely reduces so that an allowance can be made in the reading speed from the medium to enable a skip search or the like. Further, as will be described later, when the only the I picture and the P picture are arranged in summary, the device
1 1 71 1C2 0 of the invention is operated so that only the data which is subjected to the rough quantization of the I picture and the P pictures is decoded.
When the coding oi The I picture is ended by allowing the I picture to pass through the subtracter 106, the B picture is coded with bidirectional prediction with the last P picture in the preceding GOP in terms of time. An output of the preprocessor 101 is compared with the data (arrows are omitted in FIG. 70) from the memory of the reference frame so that the motion vector is detected and the prediction mode and the frame structure are judged. On the basis of the result of judgment, the data in the reference frame memory in which the output 0-LE the preprocessor 101 most favorably agrees with the data from the reference frame memory is read as data in the forward direction portion and the backward direction portion from the image memory 120 so that the data thus read and the output result of the preprocessor 101 of the B picture are subtracted by the subtracter 106 (this result is referred to as time residual difference component for both P picture and 8 picture). The data in the forward direction portion and in the backward direction portion is read from the picture memory 121 so that the data and the output of the preprocessor 101 is subtracted by the subtracter 1071 for orthogonal transform entropy coding. The same process is per-formed with respect to the P picture for coding the P picture.
(,0 A FIG. 69 is a view showing an example of a statistical amount of the coded data, the view showing a distribution of the code amount at The time when the number of frames in the GOP: N=15 and cycle of the I picture and the P picture: M=3. It is shown in FIG. 69 that the I picture and the P picture account for about '00% of the whole. When the hierarchy is divided with the resolution, the frequency, and the quantization at least with respect to this part or the I picture as described above, the code amount which is to be Played back turther reduces so the travel time of the optical head can be shortened thereby facilitating the realization of the functions such as the skip search or the like.
FIG. 70 shows a processing sequence in the above-mentioned case. In FIG. io, an arrangement of the I picture, P pictures and B picture of the original picture are coded so that the processing described in embodiments 12, 13 and 14 are performed only with respect to the I picture and the P pictures out of the aforementioned pictures while the B picture is coded without hierarchization. A sequence in which the I picture and the P picture are processed in accordance with the processing shown in embodiment 12 is referred to as sequence b while a sequence in which the I picture and the P picture are processed in accordance with the processing shown in embodiment 14 is referred to as l 0 sequence c.
In each sequence, the data is constituted by fixing and arranging in summary at the front of the GOP the I picture component and the P picture component of respective low resolution components (R), the low frequency components (L) and the rough quantizatlon component (C) by respective data reconstructor device 12j'.' With the sequence a, the low resolution picture of the I picture and the P picture can be decoded only with the low resolution component (in the sequence a of FIG. 70, core area portion after the data reconstruction) of the I picture and P picture with the result that the device can easily cope with skip search. Naturally, the data in the area other than the core area is not required to be arranged as shown in FIG. '10. It goes without saying That the data mav be arranged in an order of frame numbers at the time of encoding.
With respect to the sequence b, the low frequency component of the I picture and the P picture can be formed only with the low frequency component (core area part after data reconstitution in the sequence b of FIG. '10) so that the device is capable of easily coping with the skip search. With respect to the sequence c, rough quantization picture cf, the I picture and P picture can be decoded only with the rough quantization component of the I picture and P picture i S 4, ( The core area part after data reconstruction in The seQuence c of FIG. '10) so that the device can easily cope with the skip search. With respect to the sequence c, rough quantization picture of the I picture and P picture can be decoded only with the rough quantization component of the I picture and P picture (the core area part after data reconstruction in The sequence c of FIG. 70) so that the device can easily cope with the skip search.
For example, in a structure shown in FIG. 63, a coding loop including the reprocessor 104 is not used for the B picture so that the device may be operated in such a manner that the data is coded only with a coding loop including the preprocessor 101. In a structure shown in FIG. 66, all frequency components may be coded with the first variablelength encoder 126. Further, in a structure shown in FIG. 68, a fine quantization may be performed at the quantizer 110 for coding the data.
Most ideally, the basic data such as low frequency side data may be collected at the front of the GOP. It goes without saying that the data may be shifted a little so that the data may be overlapped with the front of the unit which constitutes an error correction code. Arranging the basic data corresponding to the unit of the error correction code in this manner can be practiced in the same manner in other embodimentsEmbodiment 15) I- 1- 1 4 1? :BUTZOUS 19TA 2 pu? Tun UT SS PC3:-:: D -1'2LE3T, S C.9 D ?l 51-TT no S E:,?-: 5-2T1) v D cl a ? G J? pu-e UT luGuoduloi,2UGntn-.:7 IUGTC-5j-1!2 U2 TO zdGOUC3 2 Slo S ci! 'Dii T DG U? 0Z ZDGdsn-j Z7.t P HILA nz)U912.7-77D 1010.) 10 EU JOIR 1)o GL:^. 'I?l27S G:)W2UTOMI ? O VA Oi HIA 5001!)G 9. U2..71 PG=OT 1. - ST GUO St0s 311 -tol -2 io U? ol ZI-JqdSG1 Lrl-,z (T-d) T?L2TS 1 1..1% 1 - 7. p 0 70103 9n TO 7T, 103 ? 10 _IIA J071c!)G ? PLr2 (A-E) io-lq -13G Gl 'I?L27S 93c2U7U-,11 ? 10 -., ---S.A307i T3G Gul 7A cl 1 L 1 1 - 1 L 1 -.tc)l no 1?'l S.tous Z 13-J G-Lout GLp-:o ol !A Giu=_291TTD 'I?L2,TS G:)u?U-LUM1 ? 70 01 wZ"., - v 3 - Guo.?UI S=OUS yT! SI1 Dolq Guo io uT uGnbG -, T Gul -1-0 U.-2 70 Glciz-2xG u? Pj-2 s,-)01R 1DG lo Gul su-,tous IGTA ? ST DU.2 cl -l) G ds:
GC IT72 L L:
cl ljGwTpoqtzE 4 reference numeral 130 denotes a mode switcher as a mode switching means, 131 a data rearranger as a data rearranging means, 132 a decodable determiner, 133 a variable-length decoder and 134 a switch. The decodable determiner 132 and the switch 134 constitut_= a data operating means. Reference numeral 133 denotes an inverse quantizer, 136 an inverse DCT circuit, 137 an image memory, 138 an adder, and 139 an inverse scan converter.
Next, an operation will explained. The data shown in FIG. 7 1 is a code arrangement assembled in 8 bits (1 byte) for example, in the vertical direction. In each macroblock information is described with f L 1 respect to the macroblock which is referred to as the macroblock header. This information refers to, for example, an increment address, a quantization scale code, a motion vector, a marker bit, a macroblock pattern or the like.
The coded data of each DCT block follows this macroblock header. A method:Eor embedding this data is constituted so that a byte is constituted zi-Lr'i bitszreams to arrange each byte in order. Since e-ach DCT block has a variable code length, the block boundary and the boundary between Che header and the data is not completed in the unit of bytes. It often happens that the boundary exists in the midst of one unit of bvte. The data in each block has a variable length, and a 13- ( _S 1 0 lower-'Lre-,,uency region is provided at a position nearer to the side of the macroblock head.
This data is divided into the a low frequency component (L) and a highfrequency component (H) to constitute a coded data as shown in FIG. 71B and 71C bv setting a fixed length code amount which is irrelevant to the event as a maximum value (the event is a unit for providing one variablelength code, and in the case of the DC component, the DC component constitutes one event while in the case of the AC component a combination of non-zero DCT coefficient and the run length constitutes one event for performing run length coding. One event completes with a code referred to as EOB at the end of the block) - Next, an operation shown in FIG. 72 will be explained. In the beginning, a mode signal is inputted from a microprocessor or the like to the mode switcher 130, the signal indicating that the skip search is being performed, or the normal continuous playback being performed. In the meantime, the playback signal from the disc is amplified with an amplifier, and is digitally demodulated to per-form an error correction by per-forming a differentiation operation from an output data obtained alter a signal playback is performed with a clock which is subjected to phase synchronization and is outputted from a PLL or the like, followed by separating an audio signal from a layer of a certain system which i ':S L, 1 constitutes video signal dat,:- and audio signal data. Then the bitstream of the video signal is extracted and is inputted to the data rearranger The output (control signal) of the mode switcher 130 is supplied n To the data rearranger 131 and t' e decodable determiner 132. The data rearranger 131 obtains a control signal and reconnects the data before division from an L component and an H component shown in FIG. 71, or outputs only the L component to the variable-lengzh decoder 133. The variable-length decoder 133 extracts a boundary of events in the L component region together with the decodable determiner 132. The portion up to The boundary is decoded and outputted to the switch 134. This switch 134 is connected so that no zero is inserted at the time of the normal playback. The switch 134, which is controlled with an output of the decodable determiner 132 inputs the decoded low-firequency component to the DCT block. In the meantime, the whole DCT block is constituted so that a zero is inserted into the highfrequency side of the DCT block.
At the time of decoding, the data of the DCT block constituted in the aforementioned manner is subjected to the inverse DCT process. Then, the reading of the image memory 137 is controlled in accordance with the cases of respective pictures to be added by the adder 138. In 1 1 3) the case of the I picture, the output of the adder 138 is passed through. In the case of the P picture, the P picture is corrected only by the motion vector of the I picture and P picture to be added. In the case of the B picture, the 3 picture is corrected by the motion vector Irom both the I picture and the 8 picture to be added.
Further, zne DC1 mode and the prediction mode motion vector at this time are controlled on the basis of information obtained by decoding the header code. In accordance with the aforementioned process, the data which is subjected to the motion compensation prediction is decoded and is stored in the image memory 137. The picture is restored LO the original constitution order of the GOP. The inverse scan converter 139 converts the buffering and the block scan into the raster scan in the output order of the picture.
Embodiment 15 is represented to be fixed in length even when the embodiment is shorter than The skip of the macroblock or the predetermined fixed length data. However, even when the length is shorter than the fixed length, the L component can be taken out with certitude by detecting the EOB every time. Consequently, it goes without saying that no problem is caused even wnen the L component data is connected to the subsequent block. Further, the EOB may be attached to the event demarcation as the L component daza when the length exceeds --- i 1 ' 1 c -'ne p rede ined 1 Fu r ther, it L.. L goes z-houz -L'-az an Cing me.:- ns 1 C 117-1 a:7 orriiaz_lrll-ll ad, or. add',-Lg ad"-Lonal SLI& as Pn audio signal, a h=_2c'er c- like and a correction cx'e is 'fill 1- -Lh e r provided to be zo zhe data in the region Zhough ncz 1 th e nar- 1.he fIgures in the explanaz"cn c- c cular,v s',noz-,, n - aforementioneld em'DOC'-merizs.
=b<>dlmenz is Next, 15 will be explained bv re-ferring to F71,C;. 73 and FIG. -14. FICS. -13.A and 733 ar-n a 1 - - - Z i ' a C agram 0 -1 a 1 v -1 d ec) S i gna 1 cod 1 p g p re, c e S S -.ng UP. 11 TE _n17C, showing an operiazion. co.-.ice,-z thereof. In FIG. 73A, reference numeral 11-0 1 denote a Fe-e, as means, a variable- e-lcoder 125 and a se--ond 12 7. pa--,-cs or COr-lP-S.DOn-'_-'ng;a-rzs are denated bv like numeral In 'KG. 63.
Next, an operazion will be explained. An --lnzerlace input p-czure da--,,is 7;-L,. conve7z a raste, i h the preprocessor 101 Scan in-La a block scan. -1he I Picture is decoded zi, a X X calculazion between frames using an output 0-1 The memc-75,. z'ri-ic'n is sub j ecze-d to z'n e zra---1:rame cod i rig - Consequenzly, In -j-ie wase of the I an inCLI-L 01 Picture, nothlrLg IS inDuzze-- zo the llP-ge Mlemory 17-0 zh c' zhe 106 so za- z.e videz signal pc-sses z-.-ough the subzra,,Z---.- 12 e) 1 106.
This data is orthogonally converted into the frequencv component by the DCT circuit 108, and is converted from the low frequency region into the block scan to be subjected to an appropriate quantization by the quantizer 110. A low-freQuencv region data up to the data of the DCT coefficient at a specific position out. of this quantized data is subject to entropy coding and outputzed to the data reconstructor 125 via the first variable-length encoc_r 126.
Further, the second variable-length encoder 127 performs the variablelength coding of the DCT coefficient alter the data located after the aforementioned specific position. With respect to the coding of the motion vector and the DC component, only the first variablelength encoder 126 may be used at least. It is required that the EOB is added both to the L component and to The H component even in one block so that the boundary of the L component changes at rates without limitations of codes. The boundary of the L component can be changed at a rate by temporarily arranging an EOB code at a demarcation part of the L component and the H component.
In the meantime, the data quantized by the quantizer 110 is subjected to the inverse quantization by the inverse quantizer 114 to be inversely converted into a space component data by the inverse DCT circuit i v (-, 1 1 0' 116.
Fhe I picture is decoded without performing a calculation between frames using an output of a frame memory which is subjected to the intra-frame coding. Consequently, in the case d' the I picture, since nothing is inputted from the image memory 120 to the adder 118, The data is allowed to pass through the adder 118. The output of the adder 118 is used as data stored in the image memory 120. Az le-pst the I picture, or the I picture data and the P picture data are required to be stored in the image memory 120. That is because the I picture and the P picture data are required for decoding the B picture normally in the SEG 1 and the XPEG 2 as reference data.
When the coding of the I picture is ended, the B picture is coded in The bidirectional prediction with the last P picture in the preceding GOP. Then, the output of the preprocessor 101 is compared with the data (arrows in the drawings omitted) from the reference frame memory to detect the motion vector and to judge the prediction mode and the frame structure. On the basis of the result of judgement, the data in the reference frame memory in which the output of the preprocessor 101 is most suitable with the data from the reference frame memory in the reference frame memory is re-ad from the image memory 120 together with the data in the forward direction portion and in the backward direction 1 Lt-( 0 portion. The data and the output result of the preprocessor 101 are subtracted by the subtracter 106 (rhe result is referred to as time residual difference component with respect to the P picture and the B picture). This time residual difference component is subjected to DCT process, quantizazlon and variable-length coding process.
When the data is divided into the low-frequency region and the highfrequency region, the rate becomes indefinite in the frequency component. Consexzuently, since the data rate in the low-frequency region does not become deflinlite, The scope in which an actuator of The head can be controlled cannot be completely compensated for. Here, the rate controller 140 renders the low-frequency component region variable. The rate controller 11-0 controls the rate so that a size of the low- frequency region becomes variable with respect to the target raze as shown in FIG. 73B.
In other words, while monitoring the output of the first variable-length encoder 126, the rate controller 140 reduces the size of the area occupied by data in the low-frequenc-v region when the monitored output is larger than the target rate set by the application. When the code amount of the first variable-length encoder 126 is small, the rate controller 140 enlarges the area of the low frequency region. In this manner, while monitoring the code amount, the rate controller 140 i, +_ _) c 1 c appropriately changes the setting of the occupied area. in the lowfrequency region with respect to the first variable-length encoder 126 and the second variable-length encoder 127.
AdditionallY, for example, a temporary coding may be performed to determine the standard for setting of the occupied area of the low- X frequency region from the result as to which region has larger number of codes and which region has smaller number of codes thereby setting the target rate.
FIG. '14 is a block diagram of a digital video signal decoding processing unit, a view showing the decoding processing for decoding data coded as described above. In FIG. '14, reference numeral 141 denotes an EOB retrieval unit. Like parts or corresponding parts are denoted by like numerals in FIG. 72A. A mode signal indicative of a st.aze such that the data is being skip searched or the normal continuous playback is being operated is inputted from a microcomputer or the like to the mode switcher 130. In the meantime, a playback signal from the disc is amplified by an amplifier so that the playback signal is differentiated with a clock which is subjected to PLL, for digital demodulation. An audio signal is separated from a system layer by conducting an error correction to extract video bitstreams to be inputted to the data rearranger 131. The output of the mode switcher 130 as mode switching 1 v 3 0 means is supplied to the data rearranger 131 and the decodable determiner 132. The data re-arranger 131 obtains this control signal to be operated so as zo connect the data before the division from the L component and H component shown in FIG. il. Otherwise, only the L component is outputted. to the variable- 1 ength decoder 133 which serves as a decoding means without being connected to the H component.
Theoretically, it never happens That the L component is severed in the midst of the events. In consideration of a case where the signal quality such as skip search or the like is not favorable, the boundary of the events is confirmed with the variable-length decoder 133 and the decodable determiner 132 so that the portion up to the boundary is decoded and outputted to the switch 134. The switch 134 is operated in such a manner that it is always turned on with respect to the playback data with a good signal quality like at the normal playback. Here, the decodable determiner 132 and the switch 134 constitute a data operating means.
The switch 134 is controlled with the decodable determiner 132 so that a zero is inserted into the high-frequency side of the of the block from the low-frequency component which has been successfully decoded thereby constituting the DCT block. Then, the data is subjected to inverse DCT so that the output of the adder 138 is passed through with I Y respect to the case of the I picture. In contrast, with respect to the P picture, the data is corrected and added by the portion of t,ne motion vector in the I picture of the reference. With respect to the B picture, the reading oil the image memory 137 is controlled and added by the adder 138 so that the B picture is corrected by the portion of the motion vector from the I picture and the P picture to be added. The DCT mode, and the prediction mode motion vector are controlled by decoding a code of the header. The data which is subjected to motion compensation prediction in this manner is decoded and is stored in the image memory 137. Then, the picture is rearranged into the original constitution order. The inverse scan converter 139 buffers the data and converts the data in the output order f rom the block scan to the raster scan.
Further, in the aforementioned explanation, an example is explained in which a size of the DCT coefficient is controlled. The numDer of events may be conzrolled instead. In this case, iz sometimes happens that the L component does not attain a predetermined number of events and EOB is added. However, since the EOB retrieval unit 141 monitors the appearance cl' the EOB, the L component can be detected with certitude. Here, in particular, the data is reconstituted on the basis of the data in the lowfrequency region, the data in the high-frequency region, and the EOB respectively. That is, the data rearranger 131 and 1 (-f C2:, 0- 1 the E08 uniz 11-1 constitutes a data recons-rl.-,jczing means.
Since -dhe energs, a-fter zhe DCT is nazu--allv Small, it goes ng 7-e 1. component c MOU sa., 1 j 11 an the H component are desilrably coded In --'iie same uzme.r with respect To the non-coded block which is not Cod ed to the H component, the data exc-"ld-iiig the L component is coded. BY setting -L"-le L component to 0, the H component may be coded. '15ince it is possible- to cope w11 M' the structure same as the d=-cod--%.- -Lhe r-,crma', -Lh-is one can be- in terms O-E Circuits.
Embodimenz 1,1 Ez,C-od i D en z 117 will- be enlained on the c.,--s-is Cf FIC. '15. FT-Cl. 75 is a block diagram S'nowing P- digizal video signial processing uniz. In FIG. 'TS, re--":=-re7,.ce numeral denotes a 143 a szitch, 144,1 a CecOce-11 j5 a second var:able-leng-,, de- er., 1,16 a -:irsz inverse L oc - 1 -- - 11-7 a second inverse quantizer' 1-48 and 11-Ou adders, 150 and 151 image memories, and 152 resolution inverse conier-ler. F1G. 'TS also shows a low rewluz-ion decoding uniz as de--OC'ing means. Like p..-rts or are denotedl by like numerals in FIC. 72.A and an explana-Lion is OMIt-LeQ.
Nexz, an operation oF embod; - 11 be explained. 'What is menZ!,, 71 c /, 1 1 4 shown in FIG. 75 may be considered as corresponding to the processing block of the video data of a playback signal from the disc in the case where the coded data as described in FIG. 68 is recorded on an optical disc or the like. A mode signal indicative of a state such that a skip Th search is being performed or normal continuous playback is being performed is inputted from a microcomputer or the like to the mode switcher 130 as a mode switching means. In the meantime, the playback signal from the disc is amplified at the amplifier and a playback signal is differentiated with a clock subjected to PLL for digital demodulation.
en an audio signal is separated from the system layer to extract video bitstreams.
7his extracted video bitstream is inputted to the multiplexer 142. The multiplexer 142 sends the data of the low resolution component to the second variable-lengzh decoder 145 while sending the other data to the first variable- 1 eng-th decoder 144 via the switch 143.
The switch 143 is controlled bv the mode switcher 130. As a mode, althoughonly the output of the playback picture of the low resolution component is demanded at the skip search or the like, the switch 143 is operated to suspend the sending of redundant data in the case where the resolution residual difference component is played back halfways. Further, at the time of the normal playback, the switch 143 i" 1 remains connected.
The second variable-length decoder 145) decodes a Huf-i'man code and the run length code to be inversely quantized by the second inverse quantizer 147 and is converted from a frequency component into a space component by the inverse DCT circuit 136.
With respect to the I picture, the converted data is passed through the adder 149 to be stored in the image memory. In The case of the P picture, the first frame of the P picture is read from the I picture stcred in the image memory and the P picture ot the second frame or after is referred to the preceding P picture stored in the image memory and corrected by the motion vector portion to be subjected to the motion compensation prediction by the adder 149. In the case c-10 the B picture, the same operation is performed on the basis of the I picture and the P picture.
In FIG. 715, a motion vector, a quantization parameter for inverse quantization and a prediction mode are outputted from the variablelength decoder. Such a motion vector, a quantization parameter and the a prediction mode are the same as shown in FIG. 74. A loop shown by a dot line block in FIG. 75) is a constitution unit for decoding a low resolution component. Since the decoding result is interpolated between pixels by the resolution inverse converter 152 as interpolation video generating 1 It 1.) c' means to compensate for the decoding result as resolution residual difference component, the decoding result is inputted to the image memory 150.
The decoding of the resolution residual component at the time of the normal playback is outputted as a picture by the inverse scan converter 139 in combination with the decoding result of the low resolution component (or in accordance with the division processing in the case where the decoding of the resolution residual component is performed by time division). The data can be decoded into the frequency component by the first variable-lengzh decoder 14-4 via the switch 143. The first inverse quantizer 146 inversely quantizes the data and the inverse DCT circuit 136 decodes data into the resolution residual difference component data in the space region.
Ihe image memory 150 refers to the pixel interpolation data of the low resolution component, and further the P picture refers to the I picture, the B picture refers to the I picture and the P picture so that the data is corrected in position by the motion vector portion with the result that the data is read from the image memory 150, and the motion compensation prediction is decoded by the adder 148.
7 rurther, in the case of the skip search, to prevent the resolution residual difference component from being played back I C 1-2 1 hal-'L'ways, superfluous data is prevented by the switch 143 from being outputted from the inverse DCT circuit 136, by suspend ng the output of the resolution residual difference. Consequently, only the pixel interpolated data o..L the low resolution component is outputted (the operation will be the same even if a switch is provided on the input part of the image memory 150) via the image memory 150 and via the inverse scan converter 139. Embodiment 18 Embodiment 18 will be explained with respect to FIG. 76 and FIG - 7 7. FIG. 7 6 is a block diagram showing a GOP address generating unit and a disc control unit, the diagram showing, in particular, a processing block in the case of recording the aforementioned rate information on a sequence header. In FIG. '16, reference numeral 153 denotes a register, 154 a GOP address calculator, and 155 an optical head/disc rotation control converter as a head position converting means and as a disc rotation control converting means. Further, FIG. 771 is a block diagram showing a GOP address generating unit and a disc control unit including a playback processing, the diagram showing, in Particular, a structure for performing a GOP playback from a disc on which the aforementioned rate "information is collected at several places. Referring to FIG. 71, reference numeral 156 denote a Playback S 1 4 amplifier, IDS7 a digital demodulator, 158 an error corrector, 159 a system layer processor, and 160 a rate data memory. The s-vszem layer processor 15)g and the rate data. memory 160 constitute a data rate information extracting means. Reference numeral 161 denotes a C-OP number counter. The GOP address calculator ISA- and the GOP number counzer 161 constitute a position information calculating means.
Next, an operation of embodiment 18 will be explained. An overall rate of one program can be optimized by rendering the raze per one C-OP variable as described in the conventional example with the result that the quality of the picture can be considerably improved. However, it is not apparent that the data falls on the front of the GOP until watching the data content. Also, in the case where it is desired that the software which has been matched hallway be played back again from that position, only way is to detect the starting position by retrieving the data on the disc minutely.
Here, in such a case, the rate control of the variable rate is set, in the beginning, to discrete rate goals such as 1 Xbits, 1.5 Mbits, 2 Xbits, 2.5) Mbits, 3 bizs or the like so that each of the rate information in all the GOP is recorded on a disc. In particular, it would be most effective when the rate information with respect zo each GOP is recorded on a TOC (Table of Content: a record region is assigned 1 _) i 1 0- to the very beginning of the disc so that information such as the tit the recording time or the like is recorded), a semi-TOC or the like.
Further, the rate information with respect to GOP may be assembled in the sequence header of video bitstreams. For example, Two hour software 14.4 k pieces of GOP. The rate information at this time may be represented with 3 bits when the rate information can be divided into five kinds of rates. Consequently, all the GOP rates can be recorded on the disc with 5. 4 k (14.4 k pieces x 3 bits /8 bits/ bytes).
A high speed access can be made to a desired GOP by storing the rate information of each of the GOP in the rate data memory 160 shown in FIG. 7 and adding up the information length corresponding to the value.
le, The device will be explained with respect to FIG. 76. The Huffman code, the run length code are decoded and the header is deciphered so that the motion vector and the kind of picture are judged.
In the meantime, the sequence header is 'decoded so that the rate information is inputted to the GOP address calculator 154- In addition, the address information of the GOP which is currently accessed to is stored in the register 153 so that the GOP f ront address for the next accessed is calculated and stored in the register 153. At the same time, by using the optical head/disc rotation control converter 155 up to -11 ' - 1 4 the front of the next GOP to be accessed, the position of the optical head is determined on the basis of the address. Then a control signal to the next access is calculated from a difference between the GOP which being. accessed and the front address to be accessed. Based on this control signal, the position control of the optical head actuator and the control of the disc rotation will be performed.
is The playback processing will be explained by referring to FIG.
The optical head and the optical head rotation are controlled so as to read the data either directly or indirectly from the TOC region or the region corresponding to the TOC region (after the rate information description address is designated, this address portion is accessed to read the rate information). Then, the playback signal from the optical head is amplified with a playback amplifier 156 to detect the wave of this signal by the digital demodulator 157 to differentiate the signal into the digital signal for digital demodulation.
The playback signal which is digitally demodulated into a digital signal is inputted into the error corrector 158 to correct an error included in the playback signal. The data after the error correction is separated into the audio bitstreams, video bitstreams and other data items by the system layer processor 159.
For example, it is judged which kind of data (AV (video and 1 J3 audio) daza, tp-xz daza and data such as prog.ram. Or -Lz- This S-ig-,al belongs zo cut and classif:Y the stream channel- i - II SUCM, a process, Lhe a-foremention raze in--c,-mazion "s stored in the raze daza memory 160.
7 ' - '. 1 1 -In conzrasz, in--,3riazion as to which number Cf: GOP is desired LO be processed is generazed, by using the C10P n w, er counter I'l. On the basis cf' -,-he address calculated bv the GOP address wiculazor the op-Lical Irlead ac-,.ua-.cr and disc rota-Lion speed cf: dis-- are Tn the explanation, an examcle is g;ven;n whIlich Lhe C-OP n=ber cotL-,.?,--zr 161 receives a Signal 'Erom -L, e svs-,em layer processor -159. in the cz-se whe.re a part which the user 7-1-F such as the a;-.rocess.,lrig or in the case where operation moves -from the playback to the skip se.,=---cil, -i-L would be more e-,.Llclen-. zo input address data from a variable- leng7-,.-l decodel. or the 1-1ke for processing the video bitstream.
Embod:Lmenz 19 Nex.rl, eui'DOC'-imen-- 1,91 Z'11 be explained i - e\7lained on FIC. '78, FIC. 79 and FIG. SO. An oper.ation will be here i nbe'-.,zz. FIG. 78 is a view showing a signal processing WT";z in the 1 L wase in -Li"le by the frequency of the signal J S - playback unit and the divilsion by the quantization are performed, the view being a block diagram of a structure used in the playback processing in the case where the rate information is collected and recorded at several places on the disc. Like parts or corresponding parts are denoted by like symbols in the aforementioned embodiments in the drawings.
The optical head or the optical he-ad rotation are controlled so that the data is either directly or indirectly read from the TOC region or the region corresponding to the TOC region (the rate information 0 description address is designated). The playback signal is amplified with the playback amplifier 156. Then, this signal is detected by the digital demodulator 157 to be differentiated for digital demodulation. Consequently, the playback signal which becomes a digital data is inputted to the error corrector 158 to correct an error included in the playback signal. The data free from errors is separated into audio bitstreams and video bitstreams by the system layer processor 159 and other data is processed as well.
For example, it is judged which kind of data (AV) data, text data, or binary data or the like such as programs) this signal belongs to divide and classify stream channels. Out of them, the aforementioned rate information is stored in the rate data memory 160. In contrast, information is generated as to which number of the GOP is desired to be processed is generated by using the GOP number counter 161. Then the address is calculated by the GOP address calculator 1,541. On the basis ol: The address calculated by the GOP address calculator 154, the optical head actuator and the rotation speed of the disc are controlled.
In this manner, at The time of skip playback, skipping is performed on the disc To find the address of the GOP to be accessed. Then the skipping is performed To make an access To a desired GOP, lowfrequencv region data obtained by using a structure described, for example, in embodiment 15 is played back and described on the screen while calculating the next address in the same manner.
A mode sipal indicative of the state such that the data is being skip searched or a normal playback is being continuously performed is inputted from a microcomputer to the mode switcher 130. As described above, the video bitstream is extracted to be inputted into the data rearranger 131. The output of the mode switcher 130 is supplied to the data rearranger 19. 1 and the deccodable determiner 132. The data rearranger 131 obtains such a control signal to operate so that the data before division is reconnected from the L component and the H component in FIG. 71. Otherwise, the data rearranger 131 outputs only the L component to the variable-length decoder 133 without connecting the L component to the 1 'S L) W H component.
In embodiment 19, theoretically it does not happen that the L component is cut in the midst of the event. However, considering the case where a signal having an unfavorable signal quality of the skip search or the like is decoded, the boundary of the event is confirmed with the variablelength decoder 133 and the decodable determiner 132 for assurance so that a portion up to the boundary is decoded and outputted to the switch 134.
The switch 134 is controlled with an output from the decodable determiner 13-21 so that a zero is inputted to the high-frequency side of the block from the low-frequency component which has been successfully decoded to constitute the DCT block. Then the reading of the image memory 137 is controlled and added by the adder 138 so that the data is subjected to the inverse DCT in such a manner That the output of the adder 138 is passed through in the case of the I picture and the data is corrected by the motion vector portion to be added in the case of the P picture and the data is corrected by the motion vector portion from the I picture and the P picture and is added in the case of the P picture.
Further, the DCT mode and the prediction mode motion vector at this time are controlled by decoding the header code. In this manner, the data subjected to the motion compensation prediction is decoded and 5-7 1 stored in the image memory 1371 to constitute the picture in the original constitution order. A- the inverse scan converter 139, the data is buffered to convert the data from the block scan to the raster scan in the output order of pictures. In addition, the switch 134 is not connected so that a zero is inserted at the time of the normal playback, but is controlled to operate and play back only the playback data.
Further, in the case where the data is divided and coded into the low frequency region and the high frequency region, there mav be cases in which a Quantization table which places an emphasis on the low C ide, a quantization table which 1 frequency si 1 1 places an emphasis on the high frequency side, and a fine quantization irrespective of the frequency region quite evenly with respect to one quantization table are prepared. Such a case can be realized when two sets of the variable-length decoder and the inverse quantizer are provided as can be seen in the local decoder shown in FIG. 68. At that time, the data rearranger 131 has to be a multiplexer.
Next, an operation of embodiment 19 will be explained on the basis of FIG. '19. FIG. '19 is a view showing a signal processing unit in the case where division by the bit length of the digital signal playback unit is performed, the view being a block diagram for explaining an embodiment with respect to the playback processing in the case where the 1 -s-g 1 Caforementioned rate information is collected and recorded particularly at several places on the disc. For example, at the very beginning of the playback of the disc which is a predetermined region on the recording medium, the optical head and the optical head rotation is controlled so that data is directly or indirectly read from the TOC region or a region corresponding to the TOC region (designating the rate information description address), and the playback signal from the optical head amplified with the playback amplifier 156 so that this signal is detected with the digital demodulator 157 to be differentiated into a digital signal for digital demodulation.
As a consequence, the playback signal which has become a digital data is inputted to the error corrector 158 to correct an error included in the playback signal. The data free from errors is separated into audio bitstreams and video bitstreams and other data items are processed as well.
* For example, this signal judges whether the data is video audio data, text data or binary data of programs or the like to cut and classify the stream channel. Out of such data, the aforementioned rate information is stored in the rate data memon., 160. In the meantime, information is generated by the GOP number counter 161 as to which number of the GOP is desired to be processed, and the address is I-S,7 1 calculated by the GOP address calculator 154 the rotation speed of the disc.
To control the actuator and In this manner, the address of the GOP to be accessed at the time of the skip playback is searched for by skipping on the disc. When The desired GOP is accessed, the next address is calculated in the same m=er and at the same time the data of the low frequency region obtained by using the structure described, for example, in embodiment 15 is played back To represent the data in one screen.
A mode signal indicative of the stare such that the skip search is being performed or a normal continuous playback is being performed is inputted to the mode switcher 130 from the microcomputer or the like. The video bitstream is extracted and is inputted to the data rearranger 131. An output of the mode switcher 130 is supplied to the data rearranger 131 and the decodable determiner 132. The data rearranger 131 obtains this control signal to operated to reconnect data before the division from the L component and the H component of FIG. 71. Otherwise, the data rearranger 131 outputs only the L component to the variable-length decoder 133 without connecting the L component to the H component.
The variable-length decoder 133 and the decodable determiner 132 extract the boundanv of the events in the L component SO that The ) L. -- i/', f portion up to the boundary is decoded and outputted to the switch 134. The switch 134 is controlled by the output of The decodable determiner 132 so that a zero is inserted to the high-frequency side of the block from the low-frequency component which has been successfully decoded to constitute a DCT block. The data is subjected to the inverse DCT. In the case of the I picture, an output of the adder 138 is passed through. In the case of the P picture, the picture is corrected by the motion vector portion within the I picture of the reference to be added. The reading of the image memory 137 is controlled and added by the adder 138 so that the data is corrected by the motion vector portion.
Further, the DCT mode and the prediction mode motion vector are controlled by decoding the header code. In this manner, the data subjected to the motion vector prediction is decoded and stored in the image memory 1371 to constitute the picture in the original order of the constitution of the GOP. The inverse scan converter 139 buffers the data to convert the data from the block scan to the raster scan. Further, the switch 134 is not connected to insert a zero at the time of the normal playback to perform connecting operation to play back only the playback data.
Next, an operation of FIG. 80 will be explained. FIG. 80 is a view showing a signal processing block in the case where the data is : 1 J p divided with the resolution of the digital signal playback Part, The view being a block diagram explaining an embodiment with respect to the playback processing in the case, in particular, where the aforementioned rate information is collected at several places on the disc.
The optical head and the rotation of the optical head is controlled so that the data is directly or indirectly (rate information description address designated) read from the TOC region or a region corresponding to the TOC region at the beginning of the disc playback, and the playback signal from the optical head is amplified by the playb-ack amplifier 156. This signal is detected with a digital demodulator 157 to be differentiated into the digital signal for digital demodulation.
Consequently, the playback signal which has become digital data is inputLed to the error corrector 158 in which an error included in the playback signal is corrected. The data free from errors is separated into audio bitstreams and video bitstreams by the system layer processor 159, and other data items are processed as well. For example, by judging whether the signal is video audio data, text data or binary data of programs or the like, the stream channel is cut and classified. Out of such data, the aforementioned rate information is stored in the rate data memory 160.
In the meantime, information is generated by the WP number f &- 1 4P counter 161 as to which number of the GOP is desired to be processed, and the address is calculated by the GOP address calculator 1P1 to control the optic,-:l head actuator and the rotation speed of the disc. In this manner, the address of the GOP to be accessed at the time oi the skip playback is searched for by skipping on the disc. When the desired GOP is accessed, the next address is calculated in the same manner and at the same rime the data of the low frequency region obtained by using the structure described,', for example, in embodiment 15 is played back to represent the data in one screen.
-A mode signal indicative of the state such that the skip search is being performed' or a normal continuous playback is being performed is The video bitstream is extracted and inputted to the multiplexer 142. The multiplexer 142 sends the low resolution component data to the second variablelength decoder 145 while sending other data items to the first variable-length decoder 144 via the switch 143. The switch 143 is controlled with the mode switcher 130. Despite that only the playback picture output of the low resolution component is demanded as a mode in skip search or the like, the switch 143 is operated so as to be turned off in the case where the resolution residual component 'is played back halfways. Further, when a playback operation is performed in which a inputted to the mode switcher 130 from a microcomputer or The like i (el. 3 1 good signal transmission qualitY is attained in such cases as the normal playback, the switch 143 is turned on.
The second variable-length decoder 145 decodes a Hufflan code and a runlength code. The data is subjected To inverse quantization by inverse Quantizer 147 to be convert -he second - _ed from a frequency region To the space region by the inverse DCT circuit 136. When the data is the I picture, the data passes through the adder 149 To be stored in the image memory. When the data is the P picture, the P picture is referred to from the image memory followed by being corrected in position by the motion vector portion to be read to decode the motion compensation prediction by the adder 149. When the data is the 8 picture, the same operation is per-formed with respect to the I picture and the P picture.
In FIG. 80, the motion vector, the quantization parameter for the inverse quantization, and the prediction mode are outputted from the variable length decoder. Since the flow of information is the same as FIG. 74, an explanation thereof is omitted. A loop on the low side of FIG. 75 is the decoding of the low resolution component. The decoding result is subjected to pixel interpolation by the resolution inverse converter 152 to compensate for the decoding result as the resolution residual difference, the data is inputted to the image memor-v 150. Embodiment 20 J &,- Next, embodiment 20 will be explained with reference to FIG. 81, FIG. 82 and FIG. 83. FIG. 81 is a block diagram for coding process. FIG. 82 is a block diagram for decoding process. In FIGS. 81 and 82, reference numeral 162 denotes a video signal encoder as encoding means. 163 an audio signal encoder, 154 and 167 memories, 165 and 168 memory controllers. The memory 1641 and the memory controller 168 constitute a data supply- ing means.
Further, the video signal encoder 162 and the memory controller 165 constitute a code amount comparing means. bitstream generator.
Reference numeral 166 denotes a system layer Reference numeral 169 denotes a variable-length decoder and 170 a decoded signal processor after the variable-length decoder. The variable-length decoder 169 and the decoded signal processor 170 serve as a data decoding means. The data rearranger 131 of FIG. 82 serves as a data reconstructing means.
In the beginning, an operation of a structure shown in FIG. 81 will be explained. Between the video signal encoder 162 and the system bitstream generator 166, the memory 164 is arranged. After the data is embedded between each of the G0Ps of coded video signals, each of the G0Ps is inputted to the system layer bitstream generator 166 while the audio signal is cocled with -he audio signal encoder 163 followed by being inputted to the system bitstream generator 166 along with a video 1,, 5- 1 signal to be subjected to an operation of adding headers or the like.
Here, data embedding operation in the memor%, 164 will be described. The memory controller 165 serves as a control circuit of the memory 164 to control the coded video signal controls video signal data so as to embed the space between each of the G0Ps. A sip, al processing will be explained by referring to FIG. 83 hereinbelow. FIG. 83 is a view illustrating a concept of processing at the digital video signal record and playback device. For example, in the case where (n-1)G0P end halfrays with respect to an access position oic an optical head or a control unit of an error correction to generate a space of a data region when nGOP generates superfluous data with respect to the access position of the optical head and the control unit of an error control, the device of the invention is controlled so that, as shown in FIG. 83A, the superfluous nGOP data is embedded in a space part after (n-1)GOP, and in The same manner a little amount of residual data of (n.1)G0P which cannot be embedded in the space because nGOP is embedded and (n12)G0P are embedded in a space part of (n3)GOP (the data is embedded in a direction from the left to the right on the paper).
Further, as another method of control, the superfluous data is k not sent in the bac ward d"rect'on as described above. As shown in FIG. 838, when (n.2)GOP exceeds the access position a little so that the 1 ', (, L - 1 (n-3)GOP ends halfways with respect to the access position of the optical head and the control unit of the error control, the device of the present invention is controlled so that the superfluous (n-3)GOP data is embedded in the space part after the (n-2)G0P data, and in -he same manner, the residual data of (n-2)GOP which cannot be embedded because (n-3)G0P is embedded is embedded in the space part of (n-1)G0P and (n-1)G0P which cannot be embedded is embedded in the space part o f nGOP (in the embedding direction from the right to the left on the paper).
Next, an operation of a structure shown in FIG. 82 will be explained. The memory 167 is controlled by the memorv controller 168 so that the data rearranged in accordance with the rule described with respect to the aforementioned FIGS. 83A and 83B is restored to the original state. For example, in the case where the data shown in FIG. 83A is restored, the device of the invention is operated so that the GOP data is restored to the original state such that nGOP portion which follows (n. 1)GOP is connected to a part after nGOP data on the left side of the paper followed by connecting (n-1)G0P data thereafter and then connecting (n- 1)G0P data following the (n2)GOP data.
This rearranging rule is required to be defined in advance as a formatting rule of a medium so that the rule is recorded as flag information in a well-organized region following, for example, the TOC 1 (e-? ? region. In the case where the rule is not defined, the rule must be clearly described somewhere on the medium. Embodiment 21 Next, embodiment 21 will be explained by referring to FIG. 84-, FIG. 85 and FIG. 86. FIG. 84 is ablock diagram representing a signal processing unit in The case where the division by the frequency at the digital signal playback part is performed, or the division by the quantizazion is performed. FIG. 85) is a block diagram representing a signal processing unit in the case where the division by the bit length at the digital signal playback part is performed, or the division by the quantization is performed. FIG. 86 is a block diagram representing a signal processing unit in the case where the division bv the resolution at the digital signal playback unit is performed, or the division by the ciuantization is performed. In FIGS. 84, 85) and 86, reference numeral 171 denotes an IP selection indicator, 172 a decodable determiner and 173 a switch. In FIG. 86, corresponding parts are shown as an example of the first decoding means, the second decoding means and the third decoding means. Like parts or corresponding parts in FIGS. 8A-, 85 and 86 are denoted by like numerals, and an explanation thereof will be omitted.
Next, an operation of embodiment 21 will be explained. In /VS a FIGS. 84and 85, an optical head or the rotation ot: -he optical head are controlled so that the data is read directly or indirectly (the rate information described address being designated) from a TOC region or a region corresponding to the TOC region. A playback signal from the optical head is amplified by the playback amplifier 156. This signal is detected by the digital demodulator 157 to be differentiated into a digital signal for digital demodulation. Consequently, the playback signal which has become digital data is inputted to the error corrector 158 to correct an error included in the playback signal. The data. free from errors is separated into audio bitstreams, and video bitstreams by the system layer processor 159 and other data items are processed as well. In FIGS. 84, 85 and 86, a control signal is inputted into the IP selection indicator I'll from the mode switcher 130 as a mode switching means for switching the decoding means on the basis of the special playback speed. Embodiment 21 is controlled so that the embodiment 21 is switched over between a mode of displaying only the I picture or a mode of displaying the I picture and the P picture with this control signal and the skip search speed.
When skip search speed is 100 times, the GOP must be outputted with a considerable thinning if both the I picture and the P picture are outputted on the screen. Consequently, the picture seems quite lu,7 6 unnatural with respect to the movement of the played back screen. In such a case, to remove the unnaturalness, the mode is required to be Switched to a mode of playing back only the I picture. The decodable determiner 132 (172 in FIG. 86) is designated to suspend rne decoding of not only the B picture but also the P picture (the switch 173 serves for this function in FIG. 86). At the sane time, the image memory 137 (150 and 151 in FIG. 86) is controlled to display only the I picture.
The screen display of the I picture and the P picture is normally favorable up to a speed of 15 times speed, but the screen display of only the I picture is more favorable at a speed of 15 times speed or more. That is because when the whole I picture and the whole P picture are displayed at a speed of 15 times speed, the continuity of the motion is extremely deteriorated since the GOP which can be played back in the subsequent process is located at a place of the 5th GOP from the GOP currently displayed even if the screen is renewed for each frame. Further, when number of frames in the GOP is N=15 and the I picture and the P picture has a cycle of X=3, the all the P pictures are decoded but only the I picture and the second frame of the P picture (third and the ninth frame in the GOP) are outputted, much fillner skip search can be performed.
As described above, the data state is divided and recorded by i /) /'-,' dividing -he data state on the basis of the predetermined condition of the following cases such as a case where the data is recorded by dividing the frequency region to a predetermined position of each GOP read from the recording medium for recording data, a case where the data is recorded by dividing the data by the resolution, and a case in which the data is divided by the quantization level to be recorded. Then when the first daza which is the basic data in the playback data and the second data excluding the loasic first data is played back from data collectively arranged, a decoding means is provided to obtain one of the playback pictures out of the following cases; a case in which all the first and the second data is decoded, and a case in which a playback picture is obtained which corresponds to the low-frequency region of the I picture and P plczure or the number of thinned pixels. Then, the decoding means to be used at the time of the special playback may be switched on the basis of the special playback speed.
It goes without saving that the setting of the way of displaying the I picture and the P picture may be changed at the time of the playback in the positive direction and at the time of the playback in the negative direction. Since the P picture can be decoded only in the positive direction of time, it is necessary to store screens which exist before the P picture LO be decoded at the time of The reverse 1 1 1 / 0 direction playback. Consequently, it is necessary to use superfluous memory for that portion. To facilitate the reverse direction playback without using such superfluous memory, The I picture and the P picture may be played back at the rime of the skip search in the positive direction and only the I picture is played back at the time of the reverse direction skip search. Embodiment 22 Embodiment 22 will be explained on the basis of FIGS. 871, 88, 89 and 90. FIG. 871 is a block diagram showing a signal processing unit in the case where the division by the frequency at the digital signal playback part is performed, or the division by the quantization is performed. FIG. 88 is a block diagram showing a signal processing unit in the case where the division by the bit length at the digital signal playback unit is performed. FIG. 89 is a view explaining a concept of processing at the time of the skip search. In FIGS. 87 and 88, reference numeral 174 denotes a field display controller. A system layer processor 159 serves as a video data extracting means. Further, FIGS. 87 and 88 shows corresponding parts as one example of video data decoding and playback means. Reference numeral 130 denotes a mode switcher as a mode switching means. Like parts or corresponding parts are denoted by like numerals in the aforementioned embodiments in the drawings.
1 4 Next, an operation of embodiment 22 will be explained. In FIGS. 87 and 88, an optical head and the rotation ofi the optical disc are controlled so that the data is directly or indirectly (the rate information described address being designated) read from a TOC region or a region corresponding to the TOC region. A playback signal from the optical head is amplified by the playback amplifier 156. This signal is detected by a digital demodulator 157 to be differentiated into a digital signal for digital demodulating. Consequently, the playback signal which has become digital data is inputted to the error corrector 158 to correct an error included in the playback signal. The data free from errors is separated into audio bitstreams, and video bitstreams by the system laver processor 159 and other data items are processed as well.
W 10 . en, ' r example, the I picture and the P picture are continuously outputted to the screen at the time of the skip search, the screen is outputted in the order denoted by arrows shown in FIG. 89. At this time, the even-field of the I picture and the odd-field of the P picture are continuous at the time of the skip search whereas there are four vacant fields between them in the encode data. In other words, the playback speed in encode data is five times as fast as the playback speed in the space between the odd-field and the even-field of the I picture. Therefore, the motion of the I picture varies in an unnatural
1 0 manner due to the change in the playback speed from one time speed to five times speed for each field.
This is replaced on the same screen as the odd field or the even field of the i picture. Otherwise, a screen is prepared by embedding the average of the upper and lower scan lines in consideration o' the interlace for output. 1 1 I L The image memory 13' shown in FIGS. 8' and 88 is controlled by using the field display controller 174 so as to constitute the same screen in the subsequent P picture. Consequently, a skip amount between fields which is a space between fields coded at the time ol recording data on each field of the picture which is played back can be uniformly obtained with the result that the Jerkiness (unnatural motion) becomes inconspicuous.
Further, FIG. 90 is a view explaining a concept ot processing at the time of the reverse playback, a view particularly showing a field order at the time of the reverse playback. Hereinafter, a field order at the time of the reverse playback will be explained on the basis of FIG. 90. At the time o-.lc the reverse playback, the reverse playback is performed in the unit of frame constituting a pair of the odd field and the even field. Specifically, when the operation moves from the odd number field to the even number field, the playback is performed in the same direction as time on the image (a playb,:-ck operation in which the
1 -)- 1 4 process a is traced in FIG. 910). When the operation moves from the even number field To the odd number field, the two field portion is reversely sent in a direction reverse to time on the image (a skip operation in which The process b is traced in FIG. 90A).
However., when the aforementioned playback process is taken, a three times speed playback is provided with the result that a playback picture is provided which moves in an awkward mariner to see so That the order of motion can not be felt smoothly. When the image memory 137 is controlled by the field displav controller 174 in FIGS. 87 and 88 so that the playback picture is displaved in the reverse direction one screen after another in the unit of field in the order of odd number, even number, odd number and even number as shown in FIG. 90B. Then since the skip amount between The fields can be almost uniformly obtained, the jerkiness becomes inconspicuous. However. with respect to the synchronous signal at this time, a normal odd number and even number field relation is required to be maintained without inverting the field synchronization signal.
The image memory 137 does not receive the output of the adder 138 as it is to take a display method independent for each of the fields, but the output of the adder 138 is received independently from the image memory 137. To pertorm such an operation, the order may be changed by
1 -2 _ 1 11 -, 0 k.
providing a buffer as a separate device. A memory may be used which has three ports in which an address control may be set independently. The aforementioned operation can be realized even when the data is multiplexed w ith a memory which can be operated at a very fast operation speed to read the data. Furziller, since the inverse scan converter 139 provides at least a memory of at least one f leld plus one splice, it goes without saying that such bu-L'-,1"ering function may be incorporated in the inverse scan converter 139.
Further, with respect to the slow playback cut of the special playback, the jerkiness becomes conspicuous when the same frame is outputted repetitively. Consequently, the frame is reconstituted and outputted so that the playback interval becomes equal. For example, in the case of the slow plavback at 1/3 times speed, it does not happen that, for example, aiter decoded I frame is outputted three times, the decoded B frame is outputted three times. Instead, the first one frame is constituted of an odd number field of the I frame. On the even number field side, the average of the upper and the lower lines may be taken.
In such a constitution, no picture shift appears in the vertical direction of the screen due to line shifts in the interlaced picture so that a stable picture can be obtained. The subsequent one frame outputs the original I frame, and the subsequent one frame 1' '2 4- 1 constitutes one frame with the odd number field of the I frame (on the even number field side, the average of the upper and lower lines may be taken). Then the subsequent one frame outputs The original B frame, and the subsequent one frame constitutes one frame with an even number field of the B frame (on the odd number field side the average of the upper and lower lines may be taken). Consequently, a slow playback can be realized at an equal interval in terms of time.
As this invention may be embodied in several forms without departing from the spirit of essential characteristics thereof, the present embodiments are therelore illustrative and not restrictive, since the scope of the invention is defined by the appended claims rather than by the description preceding them, and all changes that fall within the metes and bounds of the claims, or equivalence of such metes and bounds thereof are therefore intended to be embraced by the claims.
19) p

Claims (7)

  1. Claims
    A digital video signal recording device for recording on a recording medium a digital video signal coded by using a motion compensation prediction and an orthogonal transform, said device comprising:
    adaptive coding means for performing a coding with adaptively changing of a data rate on the basis of a picture pattern; and information adding means for adding an audio signal, additional information and an error correction code; wherein data rate information is recorded on the recording medium.
  2. 2. A digital video signal recording device as claimed in claim 1 wherein the data rate information is multiplexed with the additional information or Is written in a predetermined region of the recording medium.
  3. 3. A digital video signal playback device for reading data in which a data rate is adaptively variable in accordance with a picture pattern with respect to the data coded by using a motion compensation prediction and an orthogonal transform, said device comprising:
    data rate information extracting means for extracting data rate information recorded on the recording medium.
    - 1) c
  4. 4.
    A digital video signal playback device as claimed in claim 3 comprising:
    mode switching means for switching a playback mode between normal playback mode and special playback mode; and position calculating means for calculating a position on a recording medium where data for the special playback exists on the basis of data rate information output from said data rate information extracting means at the time of special playback mode.
  5. 5. A digital video signal playback device according to claim 4 further comprising: head position converting means for controlling a head position to a position on a recording medium in accordance with an output from said position calculating means and a special playback speed.
  6. 6. A recording medium comprising a recorded digital video signal coded by using motion compensation prediction and orthogonal trarisform, the coding having been performed with an adaptively changed data rate on the basis of the picture pattern, the recording medium further comprising recorded audio signal, additional information and an error correction code, and recorded data rate information.
  7. 7. A digital video signal playback device as claimed in any one of claims 2 to 4 and substantially as hereinbefore described as Embodiment 18 or Embodiment 19.
    1 'i 1
GB9828342A 1994-09-26 1995-09-25 Digital video signal recording and playback device Expired - Lifetime GB2330476B (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP22962094A JP3506504B2 (en) 1994-09-26 1994-09-26 Digital video signal recording and playback device
JP25209894A JP3506507B2 (en) 1994-10-18 1994-10-18 Digital video signal recording / reproducing device and digital video signal reproducing device
JP27210794A JP3501521B2 (en) 1994-11-07 1994-11-07 Digital video signal reproducing apparatus and reproducing method
JP02827795A JP3370468B2 (en) 1995-02-16 1995-02-16 Optical disk recording method, optical disk reproducing method and reproducing apparatus, and optical disk
GB9519553A GB2293516B (en) 1994-09-26 1995-09-25 Digital video signal record and playback device and method for recording and playing back the same

Publications (3)

Publication Number Publication Date
GB9828342D0 GB9828342D0 (en) 1999-02-17
GB2330476A true GB2330476A (en) 1999-04-21
GB2330476B GB2330476B (en) 1999-06-02

Family

ID=27517314

Family Applications (2)

Application Number Title Priority Date Filing Date
GB9828342A Expired - Lifetime GB2330476B (en) 1994-09-26 1995-09-25 Digital video signal recording and playback device
GB9816673A Expired - Lifetime GB2324925B (en) 1994-09-26 1995-09-25 Digital video signal recording device

Family Applications After (1)

Application Number Title Priority Date Filing Date
GB9816673A Expired - Lifetime GB2324925B (en) 1994-09-26 1995-09-25 Digital video signal recording device

Country Status (1)

Country Link
GB (2) GB2330476B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1032203A3 (en) * 1999-02-26 2002-11-06 Sony Corporation Data transmitting apparatus and method thereof, recording apparatus, and recording and reproducing apparatus
CN112581364B (en) * 2019-09-30 2024-04-09 西安诺瓦星云科技股份有限公司 Image processing method and device and video processor

Also Published As

Publication number Publication date
GB9828342D0 (en) 1999-02-17
GB2324925A (en) 1998-11-04
GB9816673D0 (en) 1998-09-30
GB2324925B (en) 1999-03-31
GB2330476B (en) 1999-06-02

Similar Documents

Publication Publication Date Title
US6134382A (en) Digital video signal record and playback device and method for giving priority to a center of an I frame
EP1995964B1 (en) Decoding variable-length encoded data
KR100291404B1 (en) Digital recorder
US5923814A (en) Methods and apparatus for performing video data reduction operations and for concealing the visual effects of data reduction operations
US5493456A (en) Method and apparatus for increasing the recording time of a digital video tape recorder
KR100284450B1 (en) Digital video signal recording apparatus and reproducing apparatus
JP3921841B2 (en) Signal processing apparatus and method, and recording apparatus, reproducing apparatus, and recording / reproducing apparatus
WO1995002946A1 (en) Decoding method and apparatus
CN1090871C (en) Apparatus for recording, recording/reproducing and reproducing digital video signals
EP0999551A1 (en) Data processing apparatus and data recording apparatus
US5673358A (en) Method for generating a reduced rate digital bitstream by selecting data as a function of the position within an image to which the data corresponds
US5692092A (en) Recording and playback system for variable speed playback wherein more a.c. coefficients are allotted to the macroblocks in the central part of an image frame
KR100721267B1 (en) Data recording apparatus, data recording/reproducing apparatus, data recording method, and data recording/reproducing method
GB2330476A (en) Adaptively-coded digital video signal recording equipment
GB2293516A (en) Digital video signal recording arrangement with enhanced skip search playback
JP3685780B2 (en) Digital video signal recording apparatus, recording method, reproducing apparatus, and reproducing method
JPH08140041A (en) Digital video signal recorder and reproducing device
WO1987004032A1 (en) Method of transmitting a video signal in sampled form
KR100225326B1 (en) Digital video signal record and playback device and method for recording and playing back the same
JP2005328512A (en) Digital video signal reproducer and reproducing method
JP2002026739A (en) Coder and decoder

Legal Events

Date Code Title Description
746 Register noted 'licences of right' (sect. 46/1977)

Effective date: 20110513

PE20 Patent expired after termination of 20 years

Expiry date: 20150924