Dividing an application of the application that the application is to be on November 2nd, 2006 applying date, and application number is 200580014078.8, by name " having the recording medium and correlation technique and the device that are used for the data structure that the managing literal caption data reproduces ".
Embodiment
The present in detail exemplary embodiment of the present invention shown in reference to the accompanying drawings.In all accompanying drawings, will use identical Reference numeral to represent same or similar part as far as possible.
In the present invention, master data is the heading message that is recorded in the CD, comprises video and voice data, offers the user by the author of coiling.Usually, video and voice data are based on the MPEG2 format record, and are called as main AV stream or master data.
The data that the auxiliary data representative is relevant with master data offer the user so that reproduce master data easily.Auxiliary data comprises caption information, interactive graphic stream, video picture graphical stream, acoustic information and has the supplemental audio data that browsable magic lantern is play.According to the feature of auxiliary data, by multiplexed with main AV stream and recorded, perhaps recorded based on MPEG2 or other suitable form as the stream file that is independent of main AV stream by auxiliary data based on the MPEG2 form for auxiliary data.
Captions normally are presented at the closed title of the presumptive area of display together with the video image that reproduces main AV data generation as auxiliary data.When the user wishes that captions with required language reproduce main AV data, by the language of captions is provided in the multilingual captions that provide at CD, the zone that the language and the video image of captions is presented at display together.Therefore, may provide captions with multiple form.Especially, the captions with the text data record are text subtitle.
Next,, elaborate the master data that writes down in the CD and comprise the record format of the auxiliary data of text subtitle data in conjunction with Fig. 1 and 2, and the file structure that is used to manage master data and auxiliary data.
Fig. 1 shows the file structure according to the various data on one exemplary embodiment of the present invention management dish.As shown, at least one BD catalogue BDMV is present under the root directory.In BD catalogue BDMV, comprised as general file (top file) information to guarantee index file index.bdmv and the obj ect file MovieObject.bdmv with user interaction.In addition, comprise playlist directory PLAYLIST, clip information directory CLIPINF, stream catalogue STREAM and auxiliary data directory AUXDATA among the BD catalogue BDMV.
Be used for being called as the video of " main AV stream " and the file of audio stream and be recorded in dish, flow among the catalogue STREAM and independently be present in such as the auxiliary flow of text subtitle (below be called text subtitle stream) according to specific format.Because text subtitle stream files and AV stream file are with MPEG2 form (as the MPEG2 transport packet) record, ' * .m2ts ' is used as the extension name (for example, 01000.m2ts, 02000.m2ts and 10001.m2ts) of each stream file.Alternately, for example, under the situation of text subtitle stream files, ' * .txtst ' can be used as extension name, because text subtitle stream has and the different auxiliary data feature of main AV stream.
In the BD standard, AV stream can be known as clip stream file.Relate to the present invention, text subtitle data will exist with the form of the file that separates with the AV stream file.For example in Fig. 1, text subtitle data exists as text subtitle stream files 10001.m2ts or 10001.txtst.
Clip information directory CLIPINF comprises clip information file * .clpi, and each clip information file and stream file are corresponding one by one.Clip information file * .clpi has the attribute information and the timing information of pairing stream file, and as management document.More clearly, the information in the clip information file comprises map information, makes time of developing stab the source packet number (SPN) that (PTS) is mapped to the source packet in the corresponding stream file.This mapping is called as entrance mapping or " EP_MAP ".
Stream file and corresponding clip information file can be called " montage " jointly.Therefore, the file in clip information directory CLIPINF " 01000.clpi " has the attribute information and the timing information of the file " 01000.m2ts " among the stream catalogue STREAM, and " 01000.clpi " and " 01000.m2ts " constitutes a montage.
Playlist directory PLAYLIST comprises play list file * .mpls, and each play list file has at least one playitems playitem PlayItem, points out the broadcast interval of a certain montage.Playitems playitem PlayItem comprises the timing information about the play start time In-Time of a certain montage that is used for playback and broadcast concluding time Out-Time, and by the montage of clip information file name sign is provided in the clip information file file-name field.The PTS information of utilization in In-Time and Out-Time information, the EP of the clip information file of appointment mapping allow search and obtain the specific stream address of corresponding stream file or the position (for example, SPN), makes the reproduction that the reproduction of playitems playitem has been caused montage.
Play list file * .mpls plays the montage of hope as the basic management file by at least one playitems playitem PlayItem is provided.In addition, play list file * .mpls also can provide sub-playitems playitem SubPlayItem to be used for managing for example reproduction of supplementary data, and described supplementary data can be reproduced (for example, browsable magic lantern is play) synchronously or asynchronous with playitems playitem PlayItem.For example, comprising under the situation that sub-playitems playitem is used for the playback text subtitle corresponding sub-playitems playitem and playitems playitem synchronized playback data.Yet, comprising under the situation that sub-playitems playitem is used for the voice data that the browsable magic lantern of playback plays, corresponding sub-playitems playitem and playitems playitem are asynchronous.
In one embodiment of the invention, the auxiliary data that comprises text subtitle will elaborate below by for example sub-playitems playitem SubPlayItems management.
Auxiliary data directory AUXDATA is the zone that is used for writing down separately the auxiliary data file that is used for playback.For example, for the playback of supporting that the user is more friendly, write down therein and be used to provide the audio files Sound.bmdv of click sound and font file * .font that text subtitle playback is used together or * .otf or the like.
Therefore, text subtitle stream 10001.m2ts as a kind of auxiliary data, can be recorded among the auxiliary data directory AUXDATA.
In addition, among the above-mentioned BD catalogue BDMV, index file index.bdmv and obj ect file MovieObject.bdmv exist as general file, to guarantee and user interaction.Index file index.bdmv has concordance list, and at user option menu information and heading message are provided.MovieObject.bdmv provides navigation command, for example, carries out playlist, and can be called by the selection of making in the concordance list.
Fig. 2 shows the file structure that is recorded in the CD and the form of relevant information.CD comprises file system information area, database community He Liu district.The system information record that is used for management document is in file system information area.In database community, write down play list file (* .m2ts) and clip information file (* .clpi), index file (index.bdmv) and the obj ect file (MovieObject.bdmv) of the stream file (* .m2ts) that is used to reproduce record.Equally, master data and auxiliary data are recorded in the stream district.As previously mentioned, master data comprises audio frequency, video and graph data, and auxiliary data comprises text subtitle.Equally, auxiliary data is separated record as unique file and master data in the stream district.That is, auxiliary data is recorded in the stream district and the discord master data is multiplexed.Correspondingly, user decision is based on being stored in the information regeneration master data in the database community He Liu district and the method for auxiliary data.
Next, the file structure of text subtitle stream files will be described in detail.In addition, detailed description definitions is used to reproduce the method for the management information of text subtitle, the method for file structure of text subtitle stream files that is used to make up the management information that comprises definition and the method and apparatus that is used to reproduce stream file.
Fig. 3 shows a display, and text subtitle stream and main AV stream are shown on a frame of video simultaneously.That is, it shows text subtitle with showing by reproducing the single frame of video that main AV miscarriage gives birth to, and in time that text subtitle and the frame of video that produced is synchronous in good time.
Fig. 4 is the view according to the file structure with fileinfo of one exemplary embodiment of the present invention, is used to illustrate the method for the reproduction of managing text subtitle.Text subtitle is by the sub-playitems playitem management in the playlist, and a sub-playitems playitem is managed the reproduction of a plurality of text subtitle clip.
As shown, play list file comprises the playitems playitem of the reproduction of at least one or the main AV montage of a plurality of management.Having under one or more text subtitles and the situation that main AV data are associated, text subtitle is managed according to sub-playitems playitem.For example, Fig. 4 shows English text subtitle clip 1 and the Korean caption montage 2 in play list file, and shows subtitle clip 1 and subtitle clip 2 is synchronous in time with main AV data.That is, subtitle clip and main AV data show together in the target time of developing.
Therefore, provide about the position of target time of developing, captions and the management information of size and be used for reproducing text subtitle.Next, elaborate the management information that is used for reproducing text subtitle.
Fig. 5 A shows the main AV stream of the reproduction that has the related text captions that is presented on the display to 5C, be used for setting forth the management information that is recorded in text subtitle stream.Fig. 5 A shows dialog information, and Fig. 5 B shows area information and Fig. 5 C shows style information.
Fig. 5 A shows the dialog information of the management information of the reproduction that is used to manage text subtitle.Dialog information is to be used for managing at least one or a plurality of text data in time of developing.
That is, time of developing is the time of reproducing corresponding subtitle, is expressed as PTS.By being defined as dialogue, managing captions easily and reproduce at the text subtitle that PTS shows.
For example, if text subtitle is shown as two row at PTS (k) and PTS (k+1), text subtitle is defined as same dialogue.In one embodiment, the text subtitle data in a dialogue is less than 100 character codes.
Fig. 5 B shows the area information in the management information.This zone is about the information in the dialogue of time of developing structure.Equally, the zone of similarly having used global style in the single frame of video of this region representation.In one embodiment, single dialogue has two zones at the most.That is, in this embodiment, talk with available single or two zone structures.
The reason that is limited to many two zones in single dialogue is to regulate the working load of decoding when text subtitle reproduces.Yet Qu Yu quantity can be restricted to maximum n (n>=2) in other embodiments.
Fig. 5 C shows the style information of management information.Style information is to be used for the information of the method for videotex captions about appointment, and comprises the style information position, size and the background color such as captions.Text is arranged or text rolls can be used as style information.
In one embodiment of the invention, style information is divided into global style and local style information is used for videotex captions by different way.
Global style is similarly to be applied to each regional style information, and it comprises each regional position and each regional size.This information is called as region style (region_style).Fig. 5 C shows two different region information style information and is applied to two zones (region#1 and #2).The region style of " position 1, size 1, color=blueness " is applied to zone 1 (region#1).The region style of " position 2, size 2, color=redness " is applied to zone 2 (region#2).
Local style information is a style information capable according to each text data in the zone or that each text data character is used.This information can be called as inline style " inline_style ".That is, inline style can be to be the style information of text data definition when region style changes.
For example, in Fig. 5 C, " inline style information (inline_style) " is applied to regional #1, and different inline style is applied to " mountain " in text data.
Fig. 6 shows the text subtitle stream files structure.For example, Fig. 6 shows the structure of the text subtitle stream files " 10001.m2ts " of Fig. 1.
In this embodiment, text subtitle stream makes up according to the MPEG2 transport stream.This transport stream comprises a plurality of transport packet, and same packet identifier (PID) is assigned to these transport packet (TP) such as " PID=0x18xx ".Therefore, optical disc replay method or device extract text subtitle stream effectively by only extracting the transport packet that is assigned " PID=0x18xx " from inlet flow.
Single information packetized elementary stream (PES) comprises a plurality of transport packet TP.In this embodiment, each dialogue forms single PES, thereby reproduces dialogue effectively.
First information packetized elementary stream (PES) in the text subtitle stream comprises the dialog style unit DSU (replacedly being called dialog style segment DSS) that contains one group of region style.And, whenever all comprise the dialogue visualization unit DPU (replacedly being called dialogue video picture fragment DPS) that contains real dialog information in second and the information packetized elementary stream subsequently with individual basic stream.
Therefore, the dialog information that Fig. 5 A defines in the 5C comprises DPU, and the style information of the dialog information link information (region_style_id) that comprises inline style information (inline style) and be linked to one of various region style of limiting among the DSU.
In one embodiment, dialog style unit comprises the region style set of maximum quantity.For example, dialog style unit comprises maximum 60 region style set, and they are distinguished by the unique region style identifications (region_style_id) that is assigned to each region style set.
Fig. 7 shows according to one exemplary embodiment of the present invention and is included in the dialog style unit (DSU) in the text subtitle stream and talks with visualization unit (DPU).The sentence structure back of text subtitle stream will be set forth in conjunction with Fig. 8.
In this embodiment, dialog style unit (DSU) comprises that maximum 60 region style set and region style set distinguish by region style identifications (region_style_id).Region style set comprises various region style and user's changeable style set, and region style set is come record according to region style identifications.The sentence structure of region information style information and user's changeable style set will be set forth the back in conjunction with Figure 10 A.
Dialogue visualization unit (DPU) comprises text data and PTS group, is one group of time of developing of text data.DPU also comprises and is applied to each regional style information.Therefore, DSU comprises the region_style_id that is used for region information style information is linked to the corresponding region.
As shown in Figure 7, DPU#1 comprises the dialogue video picture information in single zone.Because DPU#i comprises the region_style_id=k that is used for DPU#1 is linked to corresponding region pattern group, be region style set #k among the DPU so be applied to the region style of text data #1.DPU#2 comprises two zones.First regional orientation region style set #k (Region_style_id=k) of text data #1.Again, because the region style link information in the zone of DPU#2 2 is Region_style_id=n, so region style #n is applied to the text data #2 of second area.Similarly, region style set #n and region style set #m are applied to DPU#3 and DPU#4 respectively.
Similarly be applied to text data in the respective regions by the style information of region_style_id link as global style.Under the situation of text string in text data according to the specific style information change, local style information is applied to the corresponding text string.
Next, the sentence structure of DSU and DPU will be set forth.
Fig. 8 shows the sentence structure according to the text subtitle stream of an exemplary embodiment (Text_Subtitle_stream ()).As shown in Figure 8, text caption stream (Text_Subtitle_stream ()) comprises single DSU (dialog_style_unit ()) that has defined style information set and a plurality of DPU (dialog_presentation_unit ()) that comprised dialog information.That is, function T ext_Subtitle_stream () has comprised function d ialog_style_unit () and function d ialog_presentation_unit ().Text subtitle stream function T ext_Subtitle_stream () record, and the text subtitle stream utilization of record makes up by the single DSU of function d ialog_style_unit () definition with by a plurality of DPU that function d ialog_presentation_unit () defines.
Fig. 9 shows the sentence structure of dialog style unit (DSU).As previously mentioned, DSU is by function d ialog_style_unit () definition, and Fig. 9 shows the detailed structure of function d ialog_style_unit ().As shown in Figure 9, this dialog_style_unit () comprises function d ialog_styleset ().Function d ialog_styleset () has defined and the corresponding style information set of dialogue.Therefore, DSU comprises the style information set with the various style informations that are used for respective dialog.
Next, will set forth the various embodiment that the present invention defines style information set (dialog_styleset ()).These embodiment divide according to the normal place of definition text data and the method for dialogue size.Figure 10 A has showed function and the view of describing first embodiment to 12C, and Figure 13 A shows function and the view of describing second embodiment to 13C, and Figure 14 A has described to describe function and the view of the 3rd embodiment to 14G.
Figure 10 A has showed to have described according to the first embodiment of the present invention to 12C and has been used for function and view by the style information set of function d ialog_styleset () definition dialogue.First embodiment of definition style information set by the text box information that regional Chinese version data specified size and reference position are given in definition, describes its feature.
Figure 10 A shows according to the first embodiment of the present invention, the detailed structure of the dialog_styleset () sentence structure of definition among the function d ialog_style_unit () of Fig. 9.In function d ialog_styleset (), except function region style region_style (), also provide player style flag player_style_flag, region style to count number_of_region_styles, user's pattern is counted number_of_user_styles, user's changeable style set function user_changeable_styleset () and function palette (), is used to define the region information style information of dialogue.Whether mark player_style_flag points out to allow the user to change style information, and function user_changeable_styleset () provides style information to change scope.Function palette () provides colouring information.
As mentioned above, region information style information is the global style according to each zone definitions.Unique sign (region_style_id) is assigned to each region information style information.Therefore, the region information style information of dialogue is provided based on the style information set corresponding with the sign region_style_id that distributes.
Therefore, be applied to the sign region_style_id of the dialogue among the DPU by record, dialogue can be when using based on the creation dialogue among the dialog_styleset () style information set of same region_style_id definition obtain reproducing.
Next, the style information (Style Info) that comprises in the style information set that provides according to region_style_id will be set forth.
At first, define the position of picture frame respective regions and the information of size as being used to, region horizontal position region_horizontal_position, region vertical position region_vertical_position, peak width region_width and region height region_height are provided.Also provide regional background colour code region_bg_color_entry_id to be used to define the background colour of respective regions.
Secondly, the text box information of the text box size that is used for representing respective regions is provided.Text box is represented the zone of videotex data in the respective regions.
The size of text box is represented by TextField._width text_box_width and TextField._height text_box_height.The initial point of corresponding text box is provided by text box horizontal level text_box_horizontal_position and text box upright position text_box_vertical_position in the zone.
In addition, definition text rolling text_flow defines the direction of videotex, and the definition text is arranged text_alignment and defined the direction of arranging text, such as right a, left side and middle.Under the situation of text_flow, when a plurality of zone is arranged in dialogue in order to prevent the user from mixing up, be to have same text_flow with each zone definitions in the respective dialog.
By referring to Figure 11 A to 12C, will elaborate regulation below according to the text box information of text_flow and text_alignment.
As the style information that is included in the style information set, defined line space line_space and be used for controlling the spacing of respective regions between capable.Again, definition font ID font_id, font style font_style, font size font_size and font color identification font_color_entry_id are as the font information of the text data in the control respective regions.
Player_style_flag in dialog_styleset () is the style information information available whether of expression transcriber.For example, if player_style_flag is set as 1b (player_style_flag=1b), then allow in using dialog_styleset (), the definition style information, to use the style information of transcriber to reproduce text subtitle.If player_style_flag is set as 0b (player_style_flag=0b), then text subtitle is only by using the style information by dialog_styleset () definition to reproduce.
Figure 10 B shows the function " user_changeable_styleset () " of the middle definition of function " dialog_styleset () " among Figure 10 A.
By referring to Figure 10 B, function user_changeable_styleset () has defined the type and the change scope of the style information that can be changed by the user.Therefore, function " user_changeable_styleset () " is to be used for allowing the user change the style information of text subtitle data easily.
Yet in this embodiment, the style information that the user is variable is limited to font size and regional location.Therefore, region horizontal position direction region_horizontal_position_direction, region horizontal position increment region_horizontal_position_delta, region vertical position direction region_vertical_position_direction, region vertical position increment region_vertical_position_delta, font size increase and decrease font_size_inc_dec and font size increment font_size_delta in function user_changeable_styleset (), have been defined.Because text box position and line space change from the increase/minimizing according to font size, have defined the variable that is used for definition line spacing and text box position in function user_changeable_styleset ().
As mentioned above, function user_changeable_styleset () is according to the region_style_id definition, and a plurality of user's changeable style information (user_style_id) define in function user_changeable_styleset ().For example, if function user_changeable_styleset () according to the region_style_id=k definition, defines maximum 25 kinds of user's changeable style information among the function user_changeable_styleset () that defines in this exemplary embodiment.
The moving direction that provides region_horizontal_position_direction and region_vertical_position_direction to be used to define region_horizontal_position and region_vertical_position can change according to user's style identifications user_style_id.Equally, provide region_horizontal_position_delta and region_vertical_position_delta to be used for to be the change amount of unit to be defined in the once change of each direction moving area with the pixel.
For example, if region_horizontal_position_direction is defined as 0, respective regions moves right.If region_horizontal_position_direction is defined as 1, respective regions is moved to the left.Again, if region_vertical_position_direction is defined as 0, respective regions moves down.If region_vertical_position_direction is defined as 1, respective regions moves up.
Again, provide font_size_inc_dec to be used to define the direction of the increase/minimizing of font size, can change according to user_style_id; And to provide font_size_delta to be used for the pixel be the once change amount of the specific font size font_size of unit.
For example, if font_size_inc_dec is defined as 0, illustrating increases font_size, and if font_size_inc_dec is defined as 1, minimizing font_size be described.
Further, provide text box horizontal level text_box_horizontal_position and text box upright position text_box_vertical_position, be used to define the reference position of the text box that the increase/minimizing according to font_size changes.Text_box_horizontal_position and text_box_vertical_position can be by providing the same method of region_horizontal_position and region_vertical_position to define.For example, record text_box_horizontal_position and text_box_vertical_position are used to specify the change direction, and the horizontal increment text_box_horizontal_delta of recording text frame and text box vertical increment text_box_vertical_delta to be used for the pixel be that unit specifies once change amount.
Again, the TextField._width text_box_width and the TextField._height text_box_height that can similarly provide the increase/minimizing according to font_size to change with the described situation of top font_size.For example, recording text width of frame increase and decrease text_box_width_inc_dec and TextField._height increase and decrease text_box_height_inc_dec are used to specify the direction of increase/minimizing, and TextField._width increment text_box_width_delta and TextField._height increment text_box_height_delta to be used for the pixel be that unit specifies once change amount.
In addition, according to the increase/minimizing of the line of text spacing (line_space) of increase/minimizing of font_size, the same method of increase/minimizing of available and definition font_size defines.That is, record line space increase and decrease line_space_inc_dec and line space increment line_space_delta.
Figure 11 A and 11B show according to the text_box_horizontal_position, the text_box_vertical_position that define according to the first embodiment of the present invention shown in Figure 10 A and the example of text_flow videotex frame.
Promptly, Figure 11 A show definition text_box_horizontal_position, text_box_vertical_position and text_flow be used for reference area initial point videotex data (for example, leave the side-play amount of regional initial point), and not with reference to text_flow.In this situation, the initial point of text data becomes the left summit (left-top) (being also referred to as the upper left corner equally) of text box and need not cross reference file arrangement or text_flow.
Figure 11 B show when text_flow be from top to bottom the time, according to the text data in text_box_horizontal_position, text_box_vertical_position and the text_flow videotex frame.
As the last embodiment of Figure 11 A, the definition of the initial point of text box reference area.In this situation, the initial point of text data becomes the left summit of text box and need not the cross reference file arrangement.
Figure 12 A shows according to according to text_box_horizontal_psition, the text_box_vertical_position of the first embodiment of the present invention of Figure 10 A definition and another example of text_flow videotex frame to 12C.
That is, Figure 12 A show when text_flow be from left to right the time, definition text_box_horizontal_position, text_box_vertical_position and text_flow are used in left summit videotex data.Figure 12 B show when text_flow be from right to left the time, definition text_box_horizontal_position, text_box_vertical_position and text_flow are used in right summit videotex data.Figure 12 C show when text_flow be from top to bottom the time, definition text_box_horizontal_position, text_box_vertical_position and text_flow are used in right summit videotex data.
To shown in the 12C, the initial point of text data changes to right summit according to text_flow from left summit as Figure 12 A, need not cross reference file and arranges.
Figure 13 A shows the function d ialog_styleset () of definition in function d ialog_style_unit () according to a second embodiment of the present invention.Second embodiment of dialog_style_unit () has write down text filling information rather than text box information.That is, for the size of defined range Chinese version data, the recording text filling information is as region information style information.The text filling information comprises fills top padding_top, filling bottom padding_bottom, filling left end padding_left, and fills right-hand member padding_right.By the text filling information, can determine the overall dimensions of the text data in the zone.
Identical among information except the text filling information and first embodiment.Therefore, omit detailed description.
Figure 13 B shows the position of the text data that has the text filling information in the zone.According to padding_top, padding_bottom, the padding_left of the definition of the dialog_styleset () among second embodiment, and padding_right determines the position of regional Chinese version data.
Figure 14 A shows the function d ialog_styleset () that a third embodiment in accordance with the invention defines in function d ialog_style_unit ().The 3rd embodiment of dialog_styleset () has write down text horizontal position text_horizontal_position and text vertical position text_vertical_position rather than text box information and text filling information, and text_horizontal_position and text_vertical_position arrange text_alignment according to text and change.
That is, record text_horizontal_position and text_vertical_position are used for representing the reference position of regional Chinese version data.Reference position is defined by changing according to text_alignment, such as a left side, the neutralization right side.
Identical except among the information text_horizontal_position and the text_vertical_position and first embodiment.Therefore, omit detailed description.
If it is left-to-right or right to a left side that Figure 14 B shows to 14D that text rolls, and text_alignment be defined as a left side (Figure 14 B), in (Figure 14 C) or right (Figure 14 D), the position of text data.
That is, if text_alignment is defined as a left side as shown in Figure 14B, then first row that text_horizontal_position and text_vertical_position are defined as the specify text data begins the place and is positioned at left position.If text_alignment be defined as shown in Figure 14 C in, then first row that text_horizontal_position and text_vertical_position are defined as the specify text data begins the place and is positioned at the centre position.If text_alignment is defined as the right side shown in Figure 14 D, then the first row end that text_horizontal_position and text_vertical_position are defined as the specify text data is positioned at right end position.
Figure 14 E shows to 14G, is top to bottm if text rolls, and text_alignment be defined as (Figure 14 E), in when (Figure 14 F) or time (Figure 14 G), the position of text data.
If text_alignment be defined as shown in Figure 14 E on, then first row that text_horizontal_position and text_vertical_position are defined as the specify text data begins the place and is positioned at upper position.If text_alignment be defined as shown in Figure 14 F in, then first row that text_horizontal_position and text_vertical_position are defined as the specify text data begins the place and is positioned at the centre position.If text_alignment is defined as following shown in Figure 14 G, then the first row end that text_horizontal_position and text_vertical_position are defined as the specify text data is positioned at lower position.
As mentioned above, according to one of first, second and the 3rd embodiment, the reference position by definition record text data frame or text data shows text data with main AV stream.
Next, with reference to Figure 15 to 18, set forth the method for the text subtitle stream files that reproduces above-mentioned text subtitle stream files and display reproduction.
Figure 15 shows the view that is included in the information in dialogue visualization unit (DPU) and the dialog style message unit (DSU).As mentioned above, DSU has comprised one group of each regional region information style information that is used for each dialogue, and DPU has comprised dialog text information and video picture information.
Described dialog text information comprises region style identifier, has defined the inline style information that is used for each regional style information, text string and is used for text string.Region style identifier is the information that is used for respective regions is linked to the region information style information of dialog style message unit (DSU).
Video picture information among the DPU comprises that the time of developing that is used for showing each dialogue is stabbed (PTS) start time (PTS_start) and time of developing is stabbed the concluding time (PTS_end), be used in the palette lastest imformation (palette_update_flag) of the used color of dialogue procedure for displaying change and the palette information (palette that is used for upgrading) of the change during as color change.
That is, text subtitle stream is made up by dialog style information among the DSU and the dialogue video picture information among the DPU.And the information that comprises among DSU and the DPU is divided into three category informations according to the function of information.In other words, information is divided into the dialog text information that comprises text string; The composite signal that comprises the time, dialogue position and the color change that are used to show; And the information of playing up that text message is converted to graphical information.
As mentioned above, the information among the DSU is used as composite signal and plays up information.As shown in figure 15, information " region horizontal position " and " region vertical position " are used as composite signal, because " region horizontal position " and " region vertical position " represented the position in the zone in the display image.Equally, information " peak width and height ", " font and background colour ", " text rolls and arranges ", and " font name, pattern and size " be used as the information of playing up, because these information are used for expressing text according to each zone of each dialogue.Equally, information " text position " all is included in the information of playing up such as text box information, text filling information and text horizontal/vertical reference position information.
Next, elaboration is used to reproduce the method and apparatus of the text subtitle that comprises playback management information.
Figure 16 is the view of the optical disc reproducing apparatus that comprises decoding unit 40 that is used to reproduce text subtitle that illustrates according to one exemplary embodiment of the present invention.
As shown in figure 16, optical disc reproducing apparatus comprises microcomputer 3, system decoder 4, PID filtrator 5, audio decoding unit 10, video decoding unit 20, graphic decoder unit 30, caption decoding unit 40, selected cell 6 and mixed cell 50.System decoder 4 will be AV stream (for example, mpeg stream) from the signal decoding of optical disc replay under the control of microcomputer 3.Though for the sake of clarity do not illustrate, microcomputer 3 also according to the input that receives from user's (for example, by control knob or the telepilot on the device) and as mentioned above from the management information of optical disc replay, comes other element of control device.PID filtrator 5 receives AV stream, and according to extracting the separated flow of existence as the packet identifier in the transport packet of mpeg stream (PID) from system decoder 4, as video flowing, audio stream, graphical stream and text subtitle stream.Video decoding unit 20 receives the video flowing that extracts from PID filtrator 5, and this video flowing that extracts of decoding is used for reproducing the video flowing that this extracts.Audio decoding unit 10 receives the audio stream that extracts from PID filtrator 5, and this audio stream that extracts of decoding is used for reproducing the audio stream that this extracts.Graphic decoder unit 30 receives the graphical stream that extracts from PID filtrator 5, and this graphical stream that extracts of decoding is used for reproducing the graphical stream that this extracts.Equally, text subtitle decoder unit 40 receives the text subtitle stream that extracts from PID filtrator 5, and this text subtitle stream that extracts of decoding is used for reproducing the text subtitle stream that this extracts.
Optical disc reproducing apparatus can also or receive text subtitle from external source in addition.Therefore the optical disc reproducing apparatus of present embodiment can further comprise selected cell 6, is used to text subtitle decoder unit 40 to select input source.
If text subtitle and master data are together based on the mpeg format record, input source is exactly a PID filtrator 5.Therefore selected cell 6 is selected the input source of PID filtrators 5 as text subtitle stream, and the text subtitle stream that selected cell 6 will extract from PID filtrator 5 is delivered to text subtitle decoder unit 40, shown in the A of Figure 16.If optical disc reproducing apparatus receives text subtitle from external source, selected cell 6 is selected input source as text subtitle stream with external source, and selected cell 6 will be delivered to text subtitle decoder unit 40 from the text subtitle stream of external source, shown in the B of Figure 16.
Video decoding unit 20, audio decoding unit 10 and graphic decoder unit 30, each comprises transmission buffer 11,21,31 respectively, is used for buffered video, audio frequency and graphical stream.Video decoding unit 20 comprises that also Video Decoder 22 and video plane 23 are used for reading video flowing that the video flowing that extracts of transmission buffer 21 storages, decoding extract and produce video plane based on the video data of decoding, promptly are presented at the single video frame image on the display.Graphic decoder unit 30 comprises that also graphic decoder 32 and graphics plane 33 are used for decode stored and produce graphics plane in the graphical stream that extracts of transmission buffer 31 and based on the graph data of decoding, i.e. the single graph image frame that shows together with the video plane that produces.Graphic decoder unit 30 also further comprises color look-up table (CLUT) 34.CLUT34 comprises the transparency of groove (vale) that is used to control color and the graphics plane that produces, and is used for graphics plane and video plane overlapping.
Text subtitle decoder unit 40 comprises preload buffer 41, TXT-ST demoder 42, graphics plane (GP) 43 and CLUT 44.Text subtitle decoder unit 40 receives text subtitle data from selected cell 6, and received text subtitle data is preloaded in the preload buffer 41.Usually, the size that comprises the single text subtitle data of playback management information is the 0.5M byte.Therefore, the capacity of preload buffer 41 should be greater than the 0.5M byte.In order to provide seamless-reproduced, the capacity of preload buffer 41 should be more much bigger than 0.5M byte.For example, the size of preload buffer 41 can the standardization by the size of considering text subtitle data.
Text subtitle data is preloaded into preload buffer 41 and is based on following reason.Because it is less that the size of whole text subtitle compares, text subtitle data can be recorded in the independent file.By the individual files of text subtitle is provided, it can easily be read and is used to reproduce text subtitle effectively.In addition, by whole caption data is preloaded into preload buffer 41, controller buffer is convenient.
Text subtitle (TXT-ST) demoder 42 reads and is stored in the text subtitle data in the preload buffer 41 and the text caption data of decoding.Graphics plane 43 produces the subtitle graphic plane by using the text subtitle data of decoding, i.e. the single captions picture frame that shows together with graphics plane and video plane.The color on subtitle graphic plane and transparency be according to the value among the CLUT 44 control, is used for subtitle graphic plane and graphics plane and video plane overlapping.
The mixed cell 50 of optical disc reproducing apparatus comprises two totalizers 52 and 53, be used for and mix from the signal of video decoding unit 20, graphic decoder unit 30 and text subtitle decoder unit 40, and be used to produce video frame image with respective graphical plane and captions.Promptly, the plane of delineation that mixed cell 50 receiver, video decoding units 20 produce, the subtitle plane that graphics plane that graphic decoder unit 30 produces and text subtitle decoder unit 40 produce, and, produce the single video frame image that has corresponding subtitle by overlapping video plane, graphics plane and subtitle graphic plane.Basically, video plane becomes the background of single video frame image.Graphics plane and subtitle graphic plane overlap on the video plane.Overlapping order is that the feature according to graphics plane and subtitle graphic plane decides.For example, if graphics plane is the video picture figure, graphics plane is at first overlapped on the video plane by the totalizer in the mixed cell 50 52.The subtitle graphic plane that is produced by text subtitle decoder unit 40 is overlapped in the output of totalizer 52 by the totalizer in the mixed cell 50 53.Conversely, if the graphics plane from graphic decoder unit 30 is an interactive graphics, the subtitle graphic plane is at first overlapped on the video plane by the totalizer in the mixed cell 50 52, and graphics plane is overlapped in the output of totalizer 52 by the totalizer in the mixed cell 50 53 then.
Next, the method for reproducing text subtitle based on the text subtitle decoder unit 40 of text subtitle decoder 42 by utilizing will be elaborated.
Figure 17 shows the view according to the text subtitle decoder unit 40 of one exemplary embodiment of the present invention.The functional block of Figure 16 and 17 same reference numerals TYP.
Next, in conjunction with Fig. 1 to 16 and Figure 17, elaborate and reproduce a title according to one embodiment of present invention.
After CD was written into, optical disc reproducing apparatus read and stores the fileinfo that is used to reproduce with the reproduction of management information.When the user asked to reproduce title that writes down in the CD, optical disc reproducing apparatus reproduced the playlist relevant with this title.For example, if reproduced playlist as shown in Figure 4, this optical disc reproducing apparatus is by using video and audio decoding unit 10 and the 20 main AV data of reproducing by the playitems playitem management.Text subtitle stream by sub-playitems playitem management is stored in the preload buffer 41, and according to the specified font information of text subtitle stream, font file is extracted from CD.The font file that has extracted is stored in the font preload buffer 410.For example, if the user selects Korean caption from the text subtitle stream by sub-playitems playitem management, " text subtitle clip 2 " is stored in the preload buffer 41.Simultaneously, extract font file aaaaa.font, and aaaaa.font is stored in the font preload buffer 410 according to the font information of appointment in the clip information of " text subtitle clip 2 ".After extracting and storing corresponding text subtitle stream, text subtitle decoder unit 40 decoding text subtitle streams are used to produce the subtitle graphic plane, and this subtitle graphic plane is sent to mixed cell 50, be used for and the video frame image that produced by the main video data of handling the playitems playitem management shows these captions together.
The decoding of text subtitle stream was divided into for three steps.The first step splits text subtitle stream according to purpose.That is, from text subtitle stream, split out composite signal, play up information and dialog text information.The action that splits data is commonly referred to parsing.
In second step, be data bitmap by using the information of playing up that the text data that comprises in the dialog text information is played up.This is called plays up.
In the 3rd step, the text data of playing up is added graphics plane based on composite signal.That is, data bitmap adds graphics plane, is used to reproduce text subtitle.This is called synthetic.
In order to carry out analyzing step, text subtitle decoder 42 has comprised text subtitle (TXT_ST) processor 421, the composite signal that text subtitle processor 421 storage in the synthetic impact damper (DCB) 425 of dialogue has split and play up information, and the dialog text information that storage has split in dialog buffer (DB) 422.
Play up step in order to carry out, text subtitle decoder 42 has comprised text renderer 423 and dialogue presentation controller (DP) 426.As previously discussed, in this step, the text data in the dialog text information that has split is played up is data bitmap.Text renderer 423 receives the font information that is used for text data from font preload buffer 410, and receives the information of playing up that is used for text data from DP controller 426.Then, text renderer 423 is played up text data by the inline style information of using each text string that is used for text data and is data bitmap.After playing up, data bitmap is stored in the bitmap object buffer (BOB) 424.
Be stored in " object " among the BOB 426 and be each the regional text data in the dialogue that has been converted into bitmap.
Again, playing up step can carry out with at least two kinds of methods.As first method, text data can be played up behind the initial point confirming, initial point is exactly to determine that to being used for shown in the 14G text data wants the position of videotex data at one of first, second and the 3rd embodiment of the position in zone according to Figure 10 A.As second method, text data is at first played up, and the text data of playing up then is with respect to arranging to the initial point in the zone of one of first, second and the 3rd embodiment shown in the 14G according to Figure 10 A.
For synthesis step, the text subtitle decoder 42 that comprises DP controller 426 is arranged in the bitmap object of storing among the BOB 424 in the graphics plane 43 according to composite signal.Palette information in the composite signal is imported among the CLUT 44 and is used to control the change of color and transparency.Then, the subtitle graphic plane of generation is sent to the mixed cell 50 among Figure 16, is used for the video frame image that generation has captions is mixed on subtitle graphic plane and video plane.
As previously mentioned, the user can change style information in playing up step and synthesis step.In this situation, DP controller 426 receives users' input and imports to carry out according to the user and plays up and synthesize.Yet preferably restriction will be by the style information of user's change, because if the user has changed all style informations, text subtitle may not correctly reproduce.That is, can be limited to regional location and font size by the style information that the user changes.
Figure 18 shows according to one exemplary embodiment of the present invention, by analyzing step, play up the view of step and synthesis step decoding text subtitle.
That is, play up step and be by use split out in the analyzing step play up information and dialog text information is carried out, be used for arranging by playing up the bitmap object that step produces and carry out synthesis step by the composite signal that uses analyzing step to split out.
As mentioned above, text subtitle stream files is according to the method that makes up text subtitle stream files in high density compact disc of the present invention, is recorded in the CD in the mode of standardized information.Again, being recorded in text subtitle in the CD, reproduce the method and apparatus of text subtitle according to the present invention reproduced effectively.
Though the present invention discloses by the embodiment of limited quantity, those skilled in the art have benefited from this and disclose, and will understand a large amount of change based on this and variation.For example, when under some situations, consider that blue light ROM CD is described, and the invention is not restricted to this standard optical disc or CD.All these changes and variation all fall into the spirit and scope of the present invention.