WO2015083354A1 - Procédé de génération de fichier, procédé de lecture, dispositif de génération de fichier, dispositif de lecture, et support d'enregistrement. - Google Patents

Procédé de génération de fichier, procédé de lecture, dispositif de génération de fichier, dispositif de lecture, et support d'enregistrement. Download PDF

Info

Publication number
WO2015083354A1
WO2015083354A1 PCT/JP2014/005963 JP2014005963W WO2015083354A1 WO 2015083354 A1 WO2015083354 A1 WO 2015083354A1 JP 2014005963 W JP2014005963 W JP 2014005963W WO 2015083354 A1 WO2015083354 A1 WO 2015083354A1
Authority
WO
WIPO (PCT)
Prior art keywords
file
data
information
unit
streams
Prior art date
Application number
PCT/JP2014/005963
Other languages
English (en)
Japanese (ja)
Inventor
遠間 正真
智輝 小川
洋 矢羽田
山本 雅哉
村瀬 薫
小塚 雅之
Original Assignee
パナソニックIpマネジメント株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2014235068A external-priority patent/JP2015109131A/ja
Application filed by パナソニックIpマネジメント株式会社 filed Critical パナソニックIpマネジメント株式会社
Publication of WO2015083354A1 publication Critical patent/WO2015083354A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/12Formatting, e.g. arrangement of data block or words on the record carriers
    • G11B20/1217Formatting, e.g. arrangement of data block or words on the record carriers on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/84Television signal recording using optical recording
    • H04N5/85Television signal recording using optical recording on discs or drums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor
    • H04N5/913Television signal processing therefor for scrambling ; for copy protection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8233Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a character code signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor
    • H04N5/913Television signal processing therefor for scrambling ; for copy protection
    • H04N2005/91357Television signal processing therefor for scrambling ; for copy protection by modifying the video signal
    • H04N2005/91364Television signal processing therefor for scrambling ; for copy protection by modifying the video signal the video signal being scrambled

Definitions

  • This disclosure relates to a file generation method for generating an MP4 file.
  • the file format used in conventional optical discs is the MPEG2-TS (MPEG-2 Transport Stream) system defined by ISO / IEC 138181-1.
  • MPEG2-TS MPEG-2 Transport Stream
  • a file configured by multiplexing a video stream, an audio stream, and a subtitle stream in the MPEG2-TS file format is recorded on the optical disc.
  • a video stream, an audio stream, a subtitle stream, and the like are each divided into a plurality of 188-byte TS packets, multiplexed, and recorded on an optical disc.
  • This MPEG2-TS is optimized for media such as broadcast or optical discs that transmit or record data that is read and processed sequentially from the front. Therefore, even a consumer device having a relatively small buffer capacity can efficiently read, decode, and reproduce a stream.
  • the file format that is being used in the distribution of contents in the recent network is the MP4 system defined by ISO / IEC 14496-12.
  • MP4 employs an extremely flexible data structure on the premise that it is applied to a randomly accessible medium such as an HDD (Hard Disk Drive) or a flash memory.
  • streams such as a video stream, an audio stream, and a subtitle stream are divided into units of fragments of about several seconds, and these fragments are sequentially arranged to form one file.
  • a smartphone or tablet does not have an optical disk drive, but it is used as a terminal for receiving and playing back content distribution over a network by taking advantage of its high portability or the recent increase in screen size and definition. ing. For this reason, smartphones and tablets have many functions and processes corresponding to MP4, and their application to MPEG2-TS has not progressed much.
  • the file format of the content may be converted to MP4 (see, for example, Patent Document 1).
  • an MP4 file that is an MP4 file is generated.
  • a file generation method is a file generation method for generating an MP4 file, and a single MP4 file is obtained by integrating the two streams so that the two streams are continuously played back. Generated and stores information indicating a section in which the reproduction timing overlaps in each of the two streams in the generated MP4 file.
  • FIG. 1 is a diagram schematically showing an example of the structure of MPEG2-TS content stored on an optical disc.
  • FIG. 2 is a diagram for explaining a method of decrypting the aligned unit.
  • FIG. 3 is a diagram showing an internal structure of the aligned unit in a plain text state.
  • FIG. 4 is a diagram illustrating a method of creating an actual Elementary Stream from a plurality of TS Payloads.
  • FIG. 5 is a block diagram illustrating a configuration of the file generation device according to the first embodiment.
  • FIG. 6 is a diagram for explaining a method of generating an MP4 stream file from an MPEG2-TS stream file, a difference file, and a copy manifest file in the first embodiment.
  • FIG. 7 is a diagram for explaining a method of generating a difference file and a copy manifest file in the first embodiment.
  • FIG. 8 is a flowchart of the file generation method according to the first embodiment.
  • FIG. 9 is a diagram for explaining a file generation method according to the first modification of the first embodiment.
  • FIG. 10A is a diagram for explaining data encryption in the AES-CTR mode in the first modification of the first embodiment.
  • FIG. 10B is a diagram for explaining data decoding in the AES-CTR mode in the first modification of the first embodiment.
  • FIG. 11 is a diagram illustrating an example in which the MPEG-4 AAC access unit stored in the transport stream is stored in the MP4 file in the second modification of the first embodiment.
  • FIG. 12 is a diagram illustrating an example in which the MPEG-4 AVC access unit stored in the transport stream is stored in MP4 in the second modification of the first embodiment.
  • FIG. 13A is a diagram illustrating a storage example of a LATM header and a LATM payload in a TS packet in the second modification of the first embodiment.
  • FIG. 13B is a diagram illustrating an example of syntax of the AU_info table in the second modification of the first embodiment.
  • FIG. 13C is a diagram illustrating another example of syntax of the AU_info table in the second modification of the first embodiment.
  • FIG. 14 is a block diagram illustrating a configuration of the file generation device according to the second modification of the first embodiment.
  • FIG. 15A is a diagram showing a schematic structure of a NAL unit in the second modification of the first embodiment.
  • FIG. 15B is a diagram illustrating an example of the storage format of the NAL unit in MPEG2-TS in the second modification of the first embodiment.
  • FIG. 15C is a diagram illustrating an example of the storage format of the NAL unit in MP4 in the second modification of the first embodiment.
  • FIG. 16A is a diagram illustrating a configuration example of an access unit in a transport stream according to the second modification of the first embodiment.
  • FIG. 16B is a diagram illustrating an example of the syntax of the size information included in the size information NAL unit in the second modification of the first embodiment.
  • FIG. 16C is a diagram illustrating another example of the syntax of the size information included in the size information NAL unit in the second modification of the first embodiment.
  • FIG. 17 is a flowchart illustrating a processing operation in which the file generation device according to the second modification of the first embodiment generates an MP4 file.
  • FIG. 18 is a diagram illustrating a specific example of address designation when mode 2 is used in the third modification of the first embodiment.
  • FIG. 19 is a diagram illustrating an example of reading a continuous area exceeding the upper limit value of the copy size in the third modification of the first embodiment.
  • FIG. 20 is a diagram for describing a process of generating an MP4 file by copying data from an elementary stream in the third modification of the first embodiment.
  • FIG. 21 is a diagram illustrating an example of audio and video playback sections of two MP4 files that are played back continuously in the second embodiment.
  • FIG. 22A is a diagram for describing a method of generating one MP4 file by integrating playback sections in the second embodiment.
  • FIG. 22B is a block diagram of a file generation device according to Embodiment 2.
  • FIG. 22C is a flowchart of the file generation method in the second embodiment.
  • FIG. 22D is a block diagram of the playback device in the second embodiment.
  • FIG. 22E is a flowchart of the reproduction method in the second embodiment.
  • FIG. 23A is a diagram showing an example of a menu screen when an MP4 file is generated from content stored on an optical disc in the third embodiment.
  • FIG. 23B is a diagram for describing a method of generating an MP4 file using an optical disc and a network in the third embodiment.
  • FIG. 24 is a diagram illustrating an example of a copy manifest showing the size of the NAL unit, the PTS, and the DTS in the third embodiment.
  • FIG. 25 is a diagram illustrating an example of caption data stored at the end of the MP4 file in the third embodiment.
  • FIG. 26 is a diagram illustrating a case where subtitles with 2K resolution are scaled to 4K and displayed in the third embodiment.
  • FIG. 1 is a diagram schematically showing an example of the structure of MPEG2-TS content stored on an optical disc.
  • a Stream File is stored as content.
  • Stream File is XXXXXX. It is recorded with a file name of M2TS. A number is described in XXXXXX. When a plurality of contents are stored, it is possible to individually manage these contents by this number.
  • the Stream File is divided into a plurality of units each called 6144-byte Aligned Unit.
  • Aligned Unit is a unit of encryption. Note that the data amount of the Stream File may not necessarily be a multiple of 6144 bytes. If it is not a multiple of 6144 bytes, it is desirable to make the data amount of the stream file a multiple of 6144 bytes by a method such as storing NULL Data at the end of the content.
  • FIG. 2 is a diagram for explaining a method of decrypting the aligned unit.
  • the content on the optical disc is encrypted using the unit key Ku which is data.
  • the 6144-byte data included in the aligned unit is separated into the first 16-byte data and the remaining 6128-byte data, and the remaining 6128-byte data is encrypted.
  • AES_E When decrypting the Aligned Unit, AES_E first encrypts the first 16 bytes of data using the AES (Advanced Encryption Standard) encryption method using the unit key Ku. Next, an exclusive OR operation is performed on the data obtained by this encryption and the leading 16-byte data. AES_DCBC uses the result of this exclusive OR operation as a key, and decrypts the remaining 6128 bytes of data in AES-CBC (Cipher Block Chaining) mode. The first 16 bytes of data are added to the plaintext data obtained by this decryption. As a result, 6144 bytes of plaintext corresponding to the aligned unit is obtained.
  • AES Advanced Encryption Standard
  • FIG. 3 is a diagram showing the internal structure of the aligned unit in a plain text state.
  • Aligned Unit is composed of 32 source packets of 192 bytes each.
  • Each Source Packet is composed of a TP_extra_header that is a 4-byte header and a 188-byte Transport Packet that is a TS packet.
  • the 188-byte Transport Packet is composed of a 4-byte TS Header and a 184-byte TS Payload. Information indicating the attribute of TS Payload is described in TS Header.
  • TS Header from sync_byte (8bits), transport_error_indicator (1bit), payload_unit_start_indicator (1bit), transport_priority (1bit), PID (13bits), transport_scrambling_control (2bits), adaptation_field_control (2bits), and continuity_counter (4 bits) Composed.
  • the PID is information for identifying the type of elementary stream stored in the TS Payload, such as video or audio. Even when there are a plurality of types of audio, the type of audio of the elementary stream can be identified by this PID.
  • FIG. 4 is a diagram showing a method for creating an actual Elementary Stream from a plurality of TS Paloads.
  • a PES_Header and an Elementary Stream are configured by connecting a plurality of TS Payloads to which the same PID is assigned.
  • the first TS Payload of the plurality of TS Payloads is configured to include PES_Header.
  • a PES (Packetized Elementary Stream) or a PES packet is composed of at least a part of the PES_Header and the Elementary Stream.
  • the MPEG2-TS file (Stream File) is encrypted for each Aligned Unit. Therefore, in order to convert the file into an MP4 file, the above-described decryption is performed, and further re-encryption is performed.
  • a conversion device such as a smartphone or a tablet
  • plain text content is once created, there is a security problem.
  • a file generation method for generating an MP4 file, an original file configured in a file format different from MP4 is acquired, A difference file including data not included in the original file is acquired, a procedure file indicating a procedure for generating the MP4 file is acquired, and included in the difference file according to the procedure indicated in the procedure file
  • the MP4 file is generated by combining data and data included in the original file. For example, in the acquisition of the original file, the original file configured in the MPEG2-TS file format is acquired.
  • an MP4 file is generated by combining the data included in the difference file and the data included in the original file according to a predetermined procedure. Therefore, an MP4 file can be easily generated without returning an original file constituted by a file format such as MPEG2-TS to each stream such as a video stream or an audio stream. Even if the original file is encrypted, it is not necessary to decrypt and re-encrypt it. Therefore, it is possible to suppress the processing load for generating the MP4 file.
  • the acquisition of the original file, the difference file, and the procedure file may be acquired by reading the original file, the difference file, and the procedure file from an optical disc.
  • each of a plurality of parts included in the difference file is arranged so that a range of parts included in the difference file and a range of parts included in the original file are alternately arranged.
  • a range and each range of a plurality of parts included in the original file are described, and in the generation of the MP4 file, the parts indicated by the ranges are combined in the order of the ranges described in the procedure file. As a result, the MP4 file may be generated.
  • each part included in the MP4 file can be generated in order from the beginning of the MP4 file, and the MP4 file can be generated more easily because there is no backtracking.
  • Each range of the plurality of parts included in the difference file is described in the procedure file according to a data size, and each range of the plurality of parts included in the original file includes the start position and the data size of the part. And may be described in the procedure file.
  • the file generation method further records the MP4 file based on the data size of the MP4 file described in the procedure file. It may be determined whether there is enough free space on the medium.
  • the attribute of the MP4 file may be described in the procedure file, and the file generation method may further read the attribute described in the procedure file.
  • the MP4 file attribute is read from the procedure file before the MP4 file is generated, it can be determined in advance whether or not the desired MP4 file is generated.
  • the procedure file may describe a buffer size necessary for reproducing the MP4 file, and the file generation method may further read the buffer size described in the procedure file.
  • the buffer size necessary for reproducing the MP4 file is read from the procedure file, it is possible to easily determine whether or not the MP4 file can be reproduced without analyzing the MP4 file.
  • the file generation method further includes the procedure file
  • a file having a first file name described in a file may be specified as the original file, and a file having a second file name described in the procedure file may be specified as the difference file.
  • header information corresponding to MP4 that is data included in the difference file may be combined with data included in the original file.
  • the original file in a plain text state may be acquired, and in the generation of the MP4 file, the generated MP4 file may be encrypted.
  • the confidentiality of the data can be secured while leaving the data included in the original file as an MP4 file.
  • the part of the MP4 file corresponding to the part is generated and encrypted, and the part constituting the MP4 file is encrypted.
  • the part of the original file corresponding to the part may be deleted.
  • a file generation method for generating an MP4 file, wherein the two streams are played back in succession. Are combined to generate one MP4 file, and information indicating an interval in which reproduction timing overlaps in each of the two streams is stored in the generated MP4 file. For example, in the integration of the two streams, the two streams that are at least part of the original file each configured in the MP4 file format are integrated. Further, for example, in the integration of the two streams, the two streams each including audio data are integrated.
  • the playback device that plays back the MP4 file can easily specify the data of the overlap section from the MP4 file using the information.
  • the playback apparatus can appropriately play back the data by combining the data in the overlapping section. That is, an MP4 file suitable for overlapping reproduction can be generated.
  • the two streams when the section exists over a plurality of samples included in any one of the two streams, after deleting at least one of the plurality of samples
  • the two streams may be integrated.
  • time information indicating the time length of the section may be stored in the MP4 file as the information.
  • a playback device that plays back an MP4 file can easily specify the time length of the overlap section using the information.
  • the reproducing apparatus can appropriately reproduce the data within the specified time length by combining the data of the overlapping sections.
  • the time information may be stored in a traf in a moof in the MP4 file.
  • the playback device can appropriately acquire the stored time information.
  • the information may be acquired from a device or an optical disk that holds the information.
  • the playback method is a playback method for playing back an MP4 file, and extracts information indicating two sections in which playback timing overlaps in the playback target content from the MP4 file, The two sections in the content are specified based on the extracted information, and the decoding results for the data in the two sections are combined and output.
  • the playback device can easily specify the data of the overlap section from the MP4 file. As a result, the playback device can appropriately play back the data in the overlap section.
  • a recording medium is a recording medium that records an MP4 file, and the MP4 file is read by a computer and played back, and the playback timing of the content is Information indicating two overlapping sections.
  • a playback device that reads and plays back the MP4 file from the recording medium can easily identify the data of the above two sections from the MP4 file using the information. As a result, the playback device can appropriately play back the data by combining the data of those sections.
  • FIG. 5 is a block diagram showing a configuration of the file generation apparatus according to the present embodiment.
  • the file generation apparatus 10 in the present embodiment is an apparatus that generates an MP4 file, and includes an original file acquisition unit 11, a difference file acquisition unit 12, a procedure file acquisition unit 13, and a generation unit 14.
  • the original file acquisition unit 11 acquires an original file configured in a file format different from MP4.
  • the difference file acquisition unit 12 acquires a difference file including data that is not included in the original file.
  • the procedure file acquisition unit 13 acquires a procedure file indicating a procedure for generating an MP4 file.
  • the generation unit 14 generates an MP4 file by combining the data included in the difference file and the data included in the original file according to the procedure indicated in the procedure file. In other words, the generation unit 14 converts the original file into an MP4 file.
  • FIG. 6 is a diagram for explaining a method of generating an MP4 stream file from an MPEG2-TS stream file, a difference file, and a copy manifest file.
  • a stream file A (XXXX.M2TS) that is the above-mentioned original file and a copy manifest file (XXXX.X.) that is the above-described procedure file.
  • CMNF copy manifest file
  • XXXX.DMP4 difference file
  • the copy manifest file (XXXX.CMNF) describes a conversion procedure for processing the original file and converting it into an MP4 file.
  • the difference file (XXXX.DMP4) stores data necessary for conversion to an MP4 file.
  • the generating unit 14 converts the data of the stream file A (XXXX.M2TS) and the difference file (XXXX.M2TS) according to the description of the copy manifest file (XXXX.CMNF).
  • the stream file B (XXXX.MP4) which is an MP4 file is generated by alternately combining the data of XXXXX.DMP4). In other words, the generation unit 14 converts the stream file A (XXXX.M2TS) into the stream file B (XXXX.MP4).
  • the original file can be converted without being restored to an audio or video elementary stream (for example, a HE-AAC or MPEG-4 AVC stream). If the original file is encrypted, the conversion from the original file to the MP4 file can be easily performed without breaking the encryption.
  • an audio or video elementary stream for example, a HE-AAC or MPEG-4 AVC stream.
  • This copy manifest file (XXXX.CMNF) is “Input File A”, “Input File B”, “Output File”, “Output File Size”, “ATTRIBUTE”, “MP4 DECODER BUFFER SIZE” and “MP4 DECODER BUFFER SIZE”. including.
  • “Input File A” and “Input File B” indicate the file names of the two input files. The conversion process starts from a copy of the top part of the file indicated by “Input File A”.
  • “Output File” indicates an output file, that is, a file name of a generated MP4 file.
  • “Output File Size” indicates the data size of the output MP4 file. This data size is described with byte precision.
  • “ATTRIBUTE” indicates the attribute of each file, specifically, what kind of elementary stream is converted and what kind of file is generated.
  • “ATTRIBUTE” includes 4K video in which the converted MP4 file is compressed with HEVC (High Efficiency Video Coding) and 5.1ch English audio of AAC (Advanced Audio Coding). Indicates that Thus, before the conversion process, it is possible to confirm in advance what the converted file obtained when converting according to this copy manifest file is.
  • “ATTRIBUTE” may indicate the data structure of the MP4 file or the like, and may indicate the brand (that is, the type) of MP4 stored in “ftyp” of the MP4 file.
  • MP4 DECODER BUFFER SIZE indicates a minimum buffer size necessary for reproducing the converted stream file B (XXXX.MP4), which is an output MP4 file.
  • This buffer size is an amount that depends on what multiplexing rule is used to multiplex video and audio in the MP4 file that is the stream file B after conversion.
  • MP4 DECODER BUFFER SIZE may indicate what kind of decoder resources (memory amount and data transfer rate) are necessary to reproduce the stream file B after conversion.
  • “COPY MANIFEST” indicates the range of each part of the file indicated by “Input File A” and the range of each part of the file indicated by “Input File B”. Each of these portions is a portion that is sequentially copied and pasted to generate an MP4 file indicated by “Output File”. The range of each part is indicated by the copy start position and the copy size, or only by the copy size.
  • “COPY MANIFEST” is a file part indicated by “Input File A” and a part of the file indicated by “Input File B” are alternately copied and pasted. Indicates the range of the part.
  • the difference file (XXXX.DMP4) is copied and pasted from the beginning of the difference file for each part of the specified size. Therefore, in “COPY MANIFEST”, it is not necessary to specify the copy start position in order to indicate each part of the difference file (XXXX.DMP4). That is, the range of each part of the difference file is indicated only by the copy size (data size) without using the copy start position.
  • the stream file A (XXXX.M2TS) includes data that is not necessary for the converted stream file B (XXXX.MP4). Therefore, in “COPY MANIFEST”, in order to indicate each part of the stream file A that is the original file, the range of each part is indicated by the copy start position and the copy size.
  • the copy start position is a byte position from the beginning of the file
  • the copy size is a data size in bytes.
  • the generation unit 14 repeats copying and pasting alternately the portion included in the original file and the portion included in the difference file indicated by “COPY MANIFEST” according to the copy manifest file. That is, the generation unit 14 repeats alternately combining the part included in the original file and the part included in the difference file, which are indicated by “COPY MANIFEST”. As a result, an MP4 file that is the stream file B after conversion is generated.
  • the same converted stream file B (XXXX.MP4) can be generated regardless of the device used to convert the original file. That is, it becomes easy to reproduce the stream file B after conversion without any trouble on any device.
  • the file generation apparatus 10 may obtain the difference file (XXXX.DMP4) and the copy manifest file (XXXX.CMNF) from an optical disk such as a Blu-ray (registered trademark) disk, or both of them or one of them.
  • the file may be obtained via the Internet.
  • it is a special change to the file generation apparatus 10 to present the user with various selections such as converting the original file into a file in the latest file format or an MP4 file including another video and audio. It is possible without requiring.
  • the file generation apparatus 10 may perform the conversion while the stream file A that is the original file is encrypted, and after the decryption of the original file is performed, the file is generated by performing the conversion to the MP4 file.
  • the MP4 file may be encrypted again.
  • the file stream (XXXX.M2TS) which is the original file may be plain text. Since the difference file is composed of header information, it is not necessary to encrypt the difference file. However, if stream file A (XXXX.M2TS), which is the original file, or stream file B (XXXX.MP4), which is an MP4 file, is plain text, the entire file is expanded in memory or temporarily saved on the hard disk Operations such as doing this may not be allowed for security reasons.
  • the file generation apparatus 10 may delete the areas of the stream file A (XXXX.M2TS), which is the plaintext original file, in order from the area where the conversion to the MP4 file is completed. Further, when encrypting the stream file B (XXXXXX.MP4) which is an MP4 file, the file generation apparatus 10 generates a portion such as “Movie fragment” included in the MP4 file or a predetermined number of MP4 samples. Each time it is done, the generated portions may be encrypted in sequence. If an encryption method in which the data size does not change before and after encryption is used, the area of the data to be copied does not change regardless of the presence or absence of encryption.
  • the copy manifest file may indicate whether or not the stream file A (XXXX.M2TS) that is the original file is encrypted.
  • the copy manifest file can be converted after the encoded data multiplexed in the stream file A is encrypted or converted into an MP4 file after converting the encrypted encoded data into plain text. It may indicate whether or not
  • FIG. 7 is a diagram for explaining a method of generating a difference file and a copy manifest file.
  • the difference file (XXXX.DMP4) and copy manifest file (XXXX.CMNF) can be generated as follows.
  • the stream file A (XXXX.M2TS), which is the original file, is converted in format to generate the converted stream file B (XXXX.MP4), which is an MP4 file.
  • XXXX.MP4 which is an MP4 file.
  • at least one portion including data matching each other is searched between the stream file A and the stream file B which are the original files. At this time, the search is performed so that the data size of each part becomes the largest.
  • a difference file is generated by concatenating the remaining parts included in the stream file B other than the part searched as described above according to the order included in the stream file B.
  • the correlation result of each file at this time is recorded in the copy manifest file as “COPY MANIFEST”.
  • the above-described search is performed by sequentially acquiring the data included in each of the stream file A and the stream file B in the direction from the top to the back of each file and comparing the data. And the comparison is performed without returning to the reverse direction to the above-mentioned direction.
  • the above-described conversion process (copy process) can be performed sequentially, that is, by one continuous process, and speeding up or memory reduction can be realized.
  • a picture of a predetermined PTS (Presentation TimeStamp) included in the video is audio so that data of a picture having a large code amount such as an I picture does not underflow. May be multiplexed earlier in time than a frame of the same PTS as the predetermined PTS included in the.
  • the PTS of the first picture of the video and the PTS of the first frame of the audio in “Movie fragment” should be multiplexed or close to each other. Is common.
  • the conversion process may not be performed sequentially. As a result, conversion may occur while the pointer for reading or writing is sequentially moved back and forth. Therefore, whether or not the conversion process can be performed by one continuous process, or when it cannot be performed by one continuous process, the maximum data size that needs to be returned may be registered in the copy manifest file.
  • a specific threshold value may be provided. For example, only a portion including data that matches 8 bytes or more is registered in the copy manifest file. Even if each of the remaining portions other than that portion included in the stream file B includes data of less than 8 bytes that matches the data of the stream file A, the data included in those remaining portions is stored in the difference file. Stored. Further, this threshold value may be described in the copy manifest file as “MIN COPY SIZE: 8 bytes”.
  • the MPEG-2TS file may be converted into the MP4 file by using a copy manifest file indicating a multiplexing unit in the MP4 file.
  • a copy manifest file indicating a multiplexing unit in the MP4 file.
  • audio, video, or text is multiplexed as different “Movie fragments”.
  • the first and last audio frame and video picture DTS in decoding order and the byte offset value from the beginning of the MPEG-2TS file are registered together with the media identification information. May be. It should be noted that the registration order of “Movie fragment” is the same as the order of appearance of “Movie fragment” in the MP4 file.
  • FIG. 8 is a flowchart of the file generation method in the present embodiment.
  • the file generation method in the present embodiment is a method in which the file generation device 10 generates an MP4 file.
  • the original file acquisition unit 11 of the file generation device 10 acquires an original file configured in a file format different from MP4 (step S11).
  • the difference file acquisition unit 12 acquires a difference file including data that is not included in the original file (step S12).
  • the procedure file acquisition unit 13 acquires a procedure file indicating a procedure for generating an MP4 file (step S13).
  • generation part 14 produces
  • the original file acquisition unit 11 acquires an original file configured in the MPEG2-TS file format.
  • an MP4 file is generated by combining the data included in the difference file and the data included in the original file according to a predetermined procedure. Therefore, an MP4 file can be easily generated without returning an original file constituted by a file format such as MPEG2-TS to each stream such as a video stream or an audio stream. Even if the original file is encrypted, it is not necessary to decrypt and re-encrypt it. Therefore, it is possible to suppress the processing load for generating the MP4 file.
  • the original file acquisition unit 11, the difference file acquisition unit 12, and the procedure file acquisition unit 13 acquire the files by reading the original file, the difference file, and the procedure file from the optical disc, respectively.
  • the procedure file acquisition unit 13 acquires the files by reading the original file, the difference file, and the procedure file from the optical disc, respectively.
  • the ranges of the parts included in the difference file and the ranges of the parts included in the original file are alternately arranged so that the ranges of the parts included in the difference file, Each range of a plurality of parts included in the original file is described. Therefore, in step S14, the generation unit 14 generates an MP4 file by combining the parts indicated by the ranges in the order of the ranges described in the procedure file. Thereby, each part included in the MP4 file can be generated in order from the head side of the MP4 file, and the MP4 file can be generated more easily because it does not return.
  • each of the multiple parts included in the difference file is described in the procedure file according to the data size.
  • each range of the plurality of parts included in the original file is described in the procedure file by the start position and data size of the part.
  • step S14 the generation unit 14 further determines whether the medium has enough free space for recording the MP4 file based on the data size of the MP4 file described in the procedure file. Also good. Thus, since it is determined whether or not there is enough free space for recording the MP4 file on the medium, if there is no free space, processing such as canceling the generation of the MP4 file in advance may be performed. it can. That is, it is possible to prevent an error from occurring.
  • the MP4 file attributes are described in the procedure file. Therefore, the file generation apparatus 10 may further read out the attribute described in the procedure file.
  • the MP4 file attribute is read from the procedure file before the MP4 file is generated, it can be determined in advance whether or not the desired MP4 file is generated.
  • the procedure file describes the buffer size required for playing MP4 files. Therefore, the file generation device 10 may further read the buffer size described in the procedure file. Thus, if the buffer size necessary for reproducing the MP4 file is read from the procedure file, it is possible to easily determine whether or not the MP4 file can be reproduced without analyzing the MP4 file.
  • the file generation device 10 further specifies the file having the first file name described in the procedure file as the original file, and specifies the file having the second file name described in the procedure file as the difference file. May be. Thereby, the original file and difference file used for generating the MP4 file can be appropriately acquired.
  • step S14 the generation unit 14 combines header information corresponding to MP4 that is data included in the difference file with data included in the original file. Thereby, an MP4 file having appropriate header information of MP4 can be easily generated.
  • step S11 the original file acquisition unit 11 may acquire an original file in a plain text state, and in step S14, the generation unit 14 may encrypt the generated MP4 file.
  • the confidentiality of the data can be ensured while leaving the data contained in the original file as the MP4 file.
  • step S14 every time the part constituting the original file is acquired, the generation unit 14 generates and encrypts the part of the MP4 file corresponding to the part, and the part constituting the MP4 file is encrypted. Each time it is done, the portion of the original file corresponding to that portion may be deleted. Thereby, it is possible to prevent all the data included in the plaintext original file from being temporarily stored in the storage area, and to ensure the confidentiality of the data more reliably.
  • the MP4 file is generated using the difference file and the copy manifest file.
  • the MP4 file may be generated without using these files.
  • an MPEG2-TS stream file composed of a plurality of Source Packets each having a plaintext header is converted into an MP4 stream file using a counter.
  • an MP4 stream file is generated.
  • the MPEG2-TS stream file is an original file
  • the MP4 stream file is an MP4 file.
  • FIG. 9 is a diagram for explaining a file generation method in the present modification.
  • An MPEG2-TS stream file (that is, content) is composed of a plurality of source packets as described above.
  • the source packet is abbreviated as SP.
  • each source packet included in the MPEG2-TS stream file in this modification only the TS Payload portion of the data included in the source packet is encrypted. That is, among the data included in the source packet, TS_extra_header and TS Header are not encrypted and are plain text.
  • AES-CTR (CountTeR) mode is used for encryption.
  • encryption and decryption are performed using the counter value.
  • FIG. 9 using the value of the video counter (AES Counter for Video), multiple TS Payloads each containing video data are encrypted, and the value of the audio counter (AES Counter for Audio) is obtained.
  • AES Counter for Audio the value of the audio counter
  • the video counter counts only the video source packet data so that the count value increases in accordance with the arrow shown in FIG.
  • the audio counter counts only the audio source packet data so that the count value increases in accordance with the arrow shown in FIG. Details of encryption in the AES-CTR mode will be described later.
  • the file generation device can easily convert an MPEG2-TS stream file into an MP4 stream file by taking out only TS Payload from each of a plurality of Source Packets included in the MPEG2-TS stream file. it can.
  • FIG. 10A is a diagram for explaining data encryption in the AES-CTR mode.
  • a key and an initial value IV (Initial Vector) are used.
  • the ciphertext c1 corresponding to the first 16 bytes is generated by the exclusive OR operation between the value obtained by this encryption and the first 16 bytes of the data to be encrypted (m1 shown in FIG. 10A).
  • the ciphertext c2 corresponding to the next 16-byte data is generated.
  • a ciphertext is generated by performing an exclusive OR in bit units in the last block.
  • FIG. 10B is a diagram for explaining data decoding in the AES-CTR mode.
  • the same processing as that for encryption is performed on the data to be decrypted. That is, at the time of decryption, the process of encrypting the IV using the key is performed.
  • the file generation method is a file generation method for generating an MP4 file, which acquires an original file composed of a plurality of packets, and for each packet included in the original file, Among them, only the remaining encrypted payload portion excluding the header information in the plaintext state is acquired and combined to generate an MP4 file.
  • the original file configured in the MPEG2-TS file format is acquired.
  • an MP4 file can be easily generated without returning an original file constituted by a file format such as MPEG2-TS to each stream such as a video stream or an audio stream.
  • there is no need to decrypt and re-encrypt the original file Therefore, it is possible to suppress the processing load for generating the MP4 file.
  • the payloads of the plurality of packets including video data are encrypted using the counter value of the first counter for video
  • the audio data Each of the payloads of the plurality of packets including is encrypted using a counter value of a second counter for audio different from the first counter.
  • the first counter counts only data included in each of a plurality of packets corresponding to video from the beginning of the original file to the back, and the second counter proceeds from the beginning of the original file to the back. Thus, only the data included in each of the plurality of packets corresponding to the audio is counted.
  • the first counter is used as a video-dedicated counter, it is possible to easily decode a video elementary configured by combining the payloads of a plurality of packets corresponding to video in the MP4 file. it can.
  • the second counter is used as an audio-only counter, it is possible to easily decode an audio elementary configured by combining the payloads of a plurality of packets corresponding to audio in the MP4 file. it can.
  • the original file is composed of a plurality of aligned units, and each of the plurality of aligned units is composed of a plurality of source packets.
  • the plurality of packets included in the original file are a plurality of Source Packets included in each of the plurality of Aligned Units.
  • the payload is TS Payload
  • the header information is composed of TP_extra_header and TS Header.
  • the MP4 file is generated using the difference file and the copy manifest file.
  • the MP4 file may be generated without using these files.
  • an MPEG2-TS stream file including auxiliary information is converted into an MP4 stream file using the auxiliary information. As a result, an MP4 stream file is generated.
  • an access unit in MPEG2-TS includes both initialization information essential for decoding a video picture or audio frame, and encoded data of a picture or frame.
  • an access unit (referred to as a sample or MP4 sample in MP4) in an MP4 file is composed only of encoded data of a picture or a frame, and initialization information necessary for decoding is encoded data as header information of the MP4 file. Stored separately.
  • auxiliary information for reducing the amount of processing related to the conversion and a method of converting a multiplexed format using the auxiliary information explain.
  • the encoded data is encrypted, the amount of processing when decrypting and re-encrypting is large.
  • conversion to an MP4 file can be performed only by copying data without decrypting the encoded data.
  • the MPEG2-TS stream file that is the original file before conversion may be another TS (transport stream).
  • the original file is used not only for the TS specified by the MPEG-2 system but also for a TS (for example, Blu-ray (registered trademark) disc or video distribution) in which header information of a predetermined number of bytes is added to the TS packet.
  • TS for example, Blu-ray (registered trademark) disc or video distribution
  • the MP4 file generated by the conversion may be an MP4 file using “Movie fragment”, or an MP4 file not using “Movie fragment”.
  • the format of the file generated by the conversion is DEFF (Digital Entertainment Content Ecosystem) CFF (Common File Format) or MPEG-DASH (Dynamic Adaptive Stream) such as MPEG4 based on MP4. Good.
  • DEFF Digital Entertainment Content Ecosystem
  • CFF Common File Format
  • MPEG-DASH Dynamic Adaptive Stream
  • FIG. 11 is a diagram showing an example in which an MPEG-4 AAC access unit stored in a transport stream is stored in an MP4 file.
  • the MPEG-4 AAC access unit in the transport stream is composed of three types of data: LATM (Low Overhead Audio Transport Multiplex) header, PayloadLengthInfo (), and PayloadMux ().
  • LATM header includes initialization information necessary for decoding MPEG-4 AAC encoded data (also referred to as AAC data) such as the number of channels and sampling frequency. More specifically, the initialization information is stored in AudioSpecificConfig () in the LATM header.
  • PayloadLengthInfo () stores the size of PayloadMux (), and AAC data is stored in PayloadMux ().
  • AudioSpecificConfig () in the LATM header is stored in the sample entry in stsd in the moov of the MP4 file.
  • PayloadLengthInfo () and PayloadMux () are stored in mdat as sample data.
  • sample data is data stored in the sample.
  • sample data in mdat is referred to from moov, or when “Movie fragment” is used, sample data in mdat is referred to from moof.
  • an ADTS (Audio Data Transport Stream) header is used instead of a LATM header, and an access unit is composed of an ADTS header and AAC data (called raw_data_block ()).
  • the ADTS header is separated from the access unit, and at least ads_fixed_header () among the data included in the ADTS header is stored in the sample entry.
  • AAC data is stored in mdat as sample data.
  • FIG. 12 is a diagram illustrating an example in which an access unit of MPEG-4 AVC (Advanced Video Coding) stored in the transport stream is stored in MP4.
  • MPEG-4 AVC Advanced Video Coding
  • initialization information necessary for decoding such as Sequence Parameter Set (SPS) and Picture Parameter Set (PPS), is stored as part of the access unit.
  • SPS Sequence Parameter Set
  • PPS Picture Parameter Set
  • the initialization information is stored as the header information of the MP4 file.
  • the access unit shown in FIG. 12 constitutes an IDR (Instantaneous Decoder Refresh) picture.
  • NPS Network Adaptation Layer
  • SPS Network Adaptation Layer
  • a mode in which initialization information such as SPS and PPS can be included in MPEG-4 AVC sample data can be selected.
  • the mode is indicated by the identification information of the sample entry.
  • the identification information is “avc1” or “avc2”, it is prohibited to include the initialization information in the sample data.
  • the identification information is “avc3” or “avc4”, it is permitted to include initialization information in the sample data. Therefore, when the transport stream is converted into an MP4 file, when the above-described identification information in the MP4 file is set to “avc1” or “avc2”, the data stored in the mdat from the access unit in the transport stream is set.
  • the NPS unit of SPS and PPS or the NAL unit of FillerData used for stuffing is deleted from the inside.
  • the identification information is set to “avc3” or “avc4”
  • the SPS or PPS NAL unit does not have to be deleted. Therefore, whether to delete SPS and PPS may be switched according to the setting value of the identification information in the MP4 file.
  • HEVC High Efficiency Video Coding
  • MPEG-4 AVC High Efficiency Video Coding
  • initialization information may be included in the sample data. Therefore, when converting the transport stream into an MP4 file, whether to delete the initialization information from the data stored in the mdat from the access unit in the transport stream is determined according to the type of encoding method. The processing may be performed according to the determination result.
  • LATM payload As described above, when storing MPEG-4 AAC encoded data in an MP4 file, the LATM header is separated from the access unit. Furthermore, only PayloadLengthInfo () and PayloadMux () are stored in mdat as sample data. Hereinafter, PayloadLengthInfo () and PayloadMux () are collectively referred to as LATM payload.
  • the LATM header and the LATM payload are stored in separate TS packets. Thereby, the LATM header can be easily separated.
  • FIG. 13A is a diagram showing an example of storing the LATM header and the LATM payload in the TS packet. Stuffing is performed as necessary so that the data of the LATM header and the LATM payload are not mixed in the payload of the same TS packet.
  • the LATM header of access unit 1 is stored in the payload of the first TS packet.
  • stuffing is performed on the remaining area of the TS payload.
  • the PES packet is not described, but actually, the data of the access unit is stored in the payload of the PES packet, and the data of the PES packet is stored in the payload of the TS packet.
  • the payload_unit_start_indicator of the TS header is set to 1 in the TS packet including the head data of the PES packet. If it is ensured that the LATM header is included in the payload of the TS packet in which payload_unit_start_indicator is set to 1, whether or not the LATM header is included in the TS packet can be determined based on the value of payload_unit_start_indicator.
  • an AU_info table as auxiliary information may be arranged at the head of the payload of the PES packet.
  • This AU_info table includes the number of access units included in the payload of the PES packet, the LATM header and the size of the LATM payload in each access unit.
  • FIG. 13B is a diagram illustrating an example of the syntax of the AU_info table.
  • the AU_info table includes AU_info_identification_code, number_of_AU indicating the number of access units, and a size_of_LengthInfo indicating the size of the LATM header and LATM payload in the access unit.
  • AU_info_identification_code is a bit string specific to the AU_info table. By searching this code, it can be determined whether or not the AU_info table exists.
  • the AU_info table may have a data structure in which the Box data size and the Box type are combined. However, if the presence of the AU_info table is signaled by a descriptor in the transport stream or auxiliary data for conversion to an MP4 file prepared separately from the transport stream, the above code is omitted. Also good.
  • the AU_info table may also be used when storing one access unit in one PES packet.
  • a method for identifying data to be separated for conversion to an MP4 file by indicating the size or data offset of each component in the access unit is applied to MPEG-4 AVC encoded data. May be. That is, when storing MPEG-4 AVC encoded data in an MP4 file, the above-described method may be applied to separate the NPS units of SPS and PPS.
  • FIG. 13C is a diagram illustrating an example of the syntax of the AU_info table for indicating whether or not each constituent element in the access unit needs to be excluded from the sample data of the MP4 file.
  • the AU_info table includes AU_info_identification_code, number_of_data_unit, size_of_data_unit, and conversion_mode.
  • number_of_data_unit indicates the number of data units included in the access unit.
  • size_of_data_unit indicates the size of the data unit.
  • the conversion_mode is a conversion mode indicating how to handle the data unit at the time of conversion to the MP4 file. If the conversion_mode is 0, the conversion_mode indicates that the data unit is copied as sample data of the MP4 file. If the conversion_mode is 1, the conversion_mode indicates that the data unit is excluded from the sample data, that is, the data unit is not copied as the sample data.
  • the conversion_mode may indicate the handling of the data unit after the data unit is excluded. For example, the conversion_mode indicates that the data unit is excluded from the sample data and then stored in the sample entry.
  • information of each of the plurality of data units is stored in ascending order in the decoding order. For example, when one access unit of MPEG-4 AVC is stored as one PES packet, the data unit corresponds to a NAL unit, and the number of NAL units constituting the access unit is indicated by num_of_data_unit. Then, conversion_mode is set to 1 for each NPS unit of SPS and PPS. Each of the LATM header and the LATM payload may be regarded as a data unit. In this case, by setting the conversion_mode to 1 for the data unit corresponding to the LATM header, this AU_info table can be applied to the MPEG-4 AAC. When a plurality of access units are stored in the PES packet, number_of_data_unit indicates the total number of data units included in all the access units in the PES packet.
  • the AU_info table may be stored in the adaptation_field of the TS packet header (TS Header in FIG. 3) including the start portion of the PES packet.
  • the AU_info table may be stored as a part of encoded data such as NAL unit in MPEG-4 AVC or HEVC or SEI (Supplemental Enhancement Information).
  • the AU_info table can be stored for each access unit or for each random access.
  • the AU_info table may be stored for each PES packet, and only in the PES packet including the access unit leading in the decoding order in the random access unit.
  • a table may be stored.
  • the AU_info table stores information on all access units constituting the random access unit.
  • FIG. 14 is a block diagram showing a configuration of a file generation apparatus according to this modification.
  • the file generation apparatus 100 generates the MP4 file by converting the transport stream including the AU_info table into the MP4 file.
  • the file generation apparatus 100 includes an auxiliary information search unit 101, an auxiliary information analysis unit 103, and a generation unit 104.
  • the auxiliary information search unit 101 searches the AU_info table that is auxiliary information from the transport stream.
  • the auxiliary information analysis unit 103 determines whether or not an AU_info table exists based on the search result. Further, when the auxiliary information analysis unit 103 determines that the AU_info table exists, the auxiliary information analysis unit 103 analyzes the AU_info table.
  • the generation unit 104 generates an MP4 file based on the analysis result from the auxiliary information analysis unit 103.
  • the generation unit 104 includes a sample generation unit 104a and a sample entry generation unit 104b.
  • the sample generation unit 104a stores LATM payload data or NAL units other than SPS and PPS as sample data in the mdat of the MP4 file.
  • the sample entry generation unit 104b stores the LATM header data or each NPS unit of SPS and PPS in the sample entry in stsd in the moov of the MP4 file.
  • Such a file generation device 100 can easily convert a transport stream including the AU_info table described above into an MP4 file.
  • MPEG2-TS and MP4 have different storage locations for initialization information required for decoding. Furthermore, the storage format of NAL units in MPEG-4 AVC or HEVC differs between MPEG2-TS and MP4. Therefore, conversion from a transport stream to an MP4 file requires conversion of the storage format.
  • the storage format of the NAL unit in MPEG2-TS and MP4 will be described with reference to FIGS. 15A to 15C.
  • FIG. 15A is a diagram showing a schematic structure of a NAL unit.
  • the NAL unit is composed of a header and a payload.
  • type information indicating the type of data stored in the payload is stored.
  • FIG. 15B is a diagram showing an example of the storage format of the NAL unit in MPEG2-TS.
  • MPEG2-TS a unique bit string called a start code is added to the NAL unit as identification information in order to identify the boundary of the NAL unit (hereinafter, the format of such identification information is called a start code format).
  • a decoding device or the like can separate a desired NAL unit by searching the start code and the type information stored in the header of the NAL unit.
  • FIG. 15C is a diagram showing an example of the storage format of the NAL unit in MP4.
  • MP4 in order to identify the boundary of the NAL unit, a field indicating the data size of the NAL unit as identification information is added to the NAL unit (hereinafter, the format of such identification information is referred to as a NAL size format).
  • the field length of the field indicating the data size is stored in the AVCDecoderConfigurationRecord in the sample entry.
  • AVCDecoderConfigurationRecord is an area for storing initialization information at the time of decoding.
  • a decoding device or the like can separate a desired NAL unit based on the data size of the NAL unit.
  • MPEG2-TS and MP4 also differ in that the format of identification information indicating the boundary between NAL units is a start code format or a NAL size format. Therefore, when converting a transport stream into an MP4 file, it is necessary to convert identification information indicating the boundary of the NAL unit. Since the storage format of the NAL unit is defined for each encoding method, the conversion operation to the MP4 file may be switched with reference to the audio or video encoding method.
  • NAL unit data size is required. Therefore, it is desirable that the data size of the NAL unit constituting the access unit can be acquired in advance. By doing so, it is possible to reduce the processing amount when determining the data size of the NAL unit by searching the start code in the access unit of the start code format and detecting the NAL unit boundary.
  • size information indicating the data size of each NAL unit included in the access unit is stored at the head of the access unit in the transport stream in this modification.
  • FIG. 16A is a diagram illustrating a configuration example of an access unit in the transport stream according to the present modification.
  • size information indicating the data size of each NAL unit constituting the access unit multiplexed in the transport stream is stored at the head of the access unit.
  • this size information is stored in a newly defined size information NAL unit.
  • NAL unit types include private and user data. Therefore, one of these types is selected, and the selected type of NAL unit is used as the size information NAL unit. Since there is no need to encrypt the size information, this size information NAL unit is stored in plain text.
  • FIG. 16B is a diagram illustrating an example of the syntax of the size information included in the size information NAL unit.
  • the size information includes number_of_nalu and nal_size.
  • number_of_nalu indicates the number of NAL units constituting the access unit
  • nal_size indicates the data size of each NAL unit. Since the size information NAL unit is not stored in the MP4 file, the size information may not indicate the data size of the size information NAL unit itself.
  • the size information NAL unit is arranged in front of the NAL unit for signaling the head of the access unit called Access Unit Delimiter (AUD), but is arranged immediately after the AUD. Also good.
  • AUD Access Unit Delimiter
  • the size information of the size information NAL unit indicates the size of each NAL unit after the AUD. Therefore, the size information does not indicate the data size of AUD. However, since the data size of the AUD is fixed, the file generation device 100 may store the data size in advance. Further, similarly to the AU_info table, the size information may be arranged at the head of the payload of the PES packet.
  • FIG. 16C is a diagram illustrating another example of the syntax of the size information included in the size information NAL unit.
  • a variable length code such as zero_byte may be included in front of the NAL unit in the transport stream in addition to the start code. Therefore, as shown in FIG. 16C, the data size of the identification information including the start code of the NAL unit (prefix_size shown in FIG. 16C) may be stored in the size information. If the data size of the identification information is fixed, the data size of the identification information may be stored in the MPEG2-TS descriptor or auxiliary data at the time of conversion to the MP4 file. Further, in the NAL size format, the field length of the field indicating the data size of the NAL unit after conversion into the MP4 file may be indicated.
  • the identification information may be regarded as a data unit, and the contents of Nal_size_info may be indicated by an AU_info table.
  • a mode of converting the data structure of the data unit may be added in addition to the two operations of deleting the data unit and leaving it as it is by making conversion_mode multi-valued.
  • identification information for identifying the type of the data unit may be added to the AU_info table.
  • the type of NAL unit such as an SPS NAL unit, a PPS NAL unit, or a slice NAL unit in MPEG-4 AVC or HEVC can be determined based on such identification information.
  • a field indicating whether or not the data unit is encrypted may be provided separately.
  • Data units with different conversion_modes may be stored in different TS packets.
  • the conversion_mode is 2
  • the data unit stored in the sample entry as header information may be stored in the TS packet as plain text.
  • NAL unit of the slice data may be encrypted, and the other parts may be plain text. Since encryption is performed in units of TS packets, plaintext and encrypted parts cannot be mixed in the payload of TS packets, AUD has a small data size, and if AUD is stored as an independent packet, stuffing increases and multiplexing This is because efficiency decreases. Further, stuffing is necessary when the boundary of the NAL unit data of the encrypted slice is not aligned with the end of the payload of the TS packet. When performing stuffing, there is a method of using adaptation_field of a TS packet header, or inserting Filler Data NAL unit or Filler Data SEI into encoded data.
  • the adaptation_field of the TS packet header When the adaptation_field of the TS packet header is used, the adaptation_field needs to be in plain text, but switching the process of whether or not to encrypt a variable-length area is expensive. Therefore, when performing stuffing, it is desirable that the boundary of the NAL unit data is aligned with the end of the payload of the TS packet by using the stuffing data structure in the encoded data.
  • the user data storage SEI (User unregistered SEI) may be inserted into the adaptation field of the TS packet or the access unit, and the size information may be stored in the SEI.
  • the start code format can be converted to the NAL size format based on the size information of the NAL unit stored by any one or more of these methods. Further, information indicating whether size information is stored may be stored using a descriptor in MPEG2-TS.
  • FIG. 17 is a flowchart showing a processing operation in which the file generation device 100 according to this modification generates an MP4 file. Specifically, this flowchart shows an example of a processing operation for changing a transport stream into an MP4 file with reference to the AU_info table shown in FIG. 13C.
  • the auxiliary information search unit 101 of the file generation device 100 searches the AU_info table arranged at the beginning of the payload of the PES packet (step S101).
  • the auxiliary information analysis unit 103 determines whether or not an AU_info table exists based on the search result (step S102).
  • the auxiliary information analysis unit 103 includes the data size of each data unit included in the AU_info table and the conversion of each data unit.
  • the mode is acquired (step S103).
  • the sample generation unit 104a of the generation unit 104 regards the access unit separated from the PES packet as a data unit, and samples the data unit. The data is copied and pasted on mdat (step S105).
  • the sample generation unit 104a regards the payload of the PES packet as data for one access unit and separates it.
  • the sample generation unit 104a searches for the boundary of each access unit in the encoded data. To separate those access units.
  • the auxiliary information analysis unit 103 determines whether the conversion mode is 0 (step S104). If it is determined that the conversion mode is 0 (“Yes” in step S104), the sample generation unit 104a copies the data unit corresponding to the conversion mode as sample data and pastes it to mdat (step S105). On the other hand, if it is determined that the conversion mode is not 0 (“No” in step S104), the auxiliary information analysis unit 103 determines whether or not the conversion mode is 1 (step S106). If it is determined that the conversion mode is 1 (“Yes” in step S106), the sample generation unit 104a converts the data structure of the data unit corresponding to the conversion mode, and converts the converted data structure to the data structure.
  • the stored data unit is stored in mdat as sample data (step S107).
  • the sample generation unit 104a converts the format of the identification information of the NAL unit boundary from the start code format to the NAL size format.
  • the sample entry generation unit 104b stores the data unit corresponding to the conversion mode in the mdat without storing the data unit.
  • the sample entry generation unit 104b stores NPS units of SPS and PPS in MPEG-4 AVC in the sample entry.
  • the sample entry generation unit 104b separates AudioSpecificConfig () from the MPEG-4 AAC LATM header and stores it in the sample entry.
  • the portion stored in the sample entry is defined in advance in the encoding method, but auxiliary data for specifying the storage portion in the sample entry may be indicated in the AU_info table.
  • the address indicating the copy start position of the stream file described in the copy manifest file in the above embodiment may not be an absolute value from the beginning of the stream file.
  • the address indicating the copy start position may be a difference value from the address indicating the copy start position of the immediately preceding entry.
  • mode 1 data is copied alternately from the stream file and the difference file.
  • the range of data to be copied is described as (address of copy start position, copy size), and the bit length of each field is (int32, int8), for example.
  • mode 2 if the copy size is 0, copy is skipped. Since data can be copied continuously from the same file, the data length of the address can be shortened compared to mode 1. For example, the range of data to be copied is described as (address of copy start position, copy size), and the bit length of each field is (int16, int8), for example.
  • mode 3 instead of explicitly indicating the copy size in mode 2, data up to the end of the payload of the TS packet indicated by the address is copied.
  • the copy size field is omitted.
  • a difference value between index numbers of TS packets in the stream file may be used.
  • the range of data to be copied is described as (packet number, copy mode), and the bit length of each field is, for example, (int7, int1).
  • the copy mode indicates whether copying is skipped.
  • FIG. 18 is a diagram showing a specific example of address designation when mode 2 is used.
  • the start position 1 and the start position 2 which are copy start positions are the 210th byte and the 91428th byte, respectively.
  • the maximum value of the field length is 65535. Therefore, the difference value of the address cannot be expressed by one entry. Therefore, when jumping from the start position 1 to the start position 2, two entries (65535, 0) and (25683, 180) are used. By doing so, it is possible to acquire 180 bytes of data after jumping to the start position 2.
  • FIG. 19 is a diagram illustrating an example of reading a continuous area exceeding the upper limit of the copy size.
  • the address or copy size indicating the copy start position may be described in units of 4 bytes instead of units of 1 byte.
  • identification information indicating whether to copy from a difference file or a stream file is separately provided, so that entries for one file are not consecutively described, but entries for one file are consecutive. May be described.
  • FIG. 20 is a diagram for explaining a process of creating an MP4 file by copying data from an elementary stream.
  • an MPEG2-TS file in which video is multiplexed is 10 GB and the size of a TS packet is 192 bytes
  • the number of packets is 55924053, and the same number of entries is required.
  • the playback time length of this video is 2 hours and the frame rate is 30 Hz
  • the total number of frames is 216000. Therefore, when an entry is generated for each frame, the number of entries can be greatly reduced as compared with a case where an entry is generated for each packet.
  • an MPEG2-TS file is double-encrypted by AACS (Advanced Access Content System) and BD +, and data is interleaved and rearranged by BD +. Even in such a case, the order of the data constituting the encoded stream can be uniquely determined after decryption and separation of the encoded stream, so that the copying operation according to this method is possible.
  • AACS Advanced Access Content System
  • the entry consists of (file ID, copy start position address, copy size).
  • the bit length of each field is, for example, (int4, int22, int22 in 1-byte units).
  • the file ID is identification information of a file including data to be copied. For example, 0 is assigned to the difference file as the file ID, and a value of 1 or more is assigned to the file of the elementary stream.
  • the maximum size of one frame is about 4 Mbytes.
  • the maximum difference value of the address at the copy start position may be the maximum size of one frame, and 4 Mbytes can be expressed by 22 bits.
  • the copy size is 22 bits in order to express the maximum size of one frame.
  • the entry consists of (file ID, copy start position address, copy size).
  • the bit length of each field is, for example, (int4, int12, int16 in 1-byte units).
  • AVC or HEVC it is necessary to convert the header of the NAL unit from the start code format to the NAL size format, and it is necessary to copy data in units of payload of the NAL unit. For this reason, a field for the address of the copy start position is provided, but if the elementary stream in the TS can be copied as it is, the field of the address for the copy start position may be omitted.
  • one entry may indicate the entire NAL unit, and the header of the NAL unit may be converted into a NAL size format after copying. Also at this time, since the data can be read continuously, the address of the copy start position is unnecessary.
  • a plurality of MPEG2-TS files can be continuously reproduced by referring to a playlist, and a stream constraint condition at a file boundary is also defined. For example, when a plurality of files or playback sections are seamlessly connected, the playback sections of two audio to be connected may be overlapped.
  • MP4 it is possible to specify a plurality of MP4 files that are continuously played back using a playlist, and it is assumed that the same constraints are added.
  • FIG. 21 is a diagram showing an example of audio and video playback sections in two MP4 files that are played back continuously.
  • the audio playback sections overlap.
  • DTS Decode TimeStamp
  • PTS Physical Transport Stream
  • Information indicating whether or not the MP4 playback sections overlap or information indicating the overlapping playback sections is stored in the playlist or MP4 file.
  • the PTS of the first sample of audio and video in the MP4 file referred to by the play item in the playlist and the playback end time of the last sample are described.
  • the absolute time of the DTS or PTS of the first sample of the MP4 file may be stored in the moov of the MP4 file.
  • the absolute time of DTS or PTS of the first sample of Movie fragment or the first sample of each track in Movie fragment may be stored in moof or traf.
  • MP4 there is no overlap or gap in the video playback section, and it is guaranteed that PTS is continuous. In this case, it is only necessary to show the information about overlap only for audio or text.
  • XXX2. If the PTS of the first sample of the MP4 video is T0, XXX2.
  • An offset value between T0 and the reproduction end time of the last sample of the MP4 audio may be stored.
  • the offset value with the PTS of the first sample of MP4 is set to XXX2. You may store in MP4.
  • the playback device selects or synthesizes audio samples to be output based on these pieces of information related to overlap. Note that the video playback sections may be overlapped, or the audio or video playback sections may not be overlapped, but a gap may be
  • a plurality of files are continuously played using a playlist.
  • a single MP4 file may be generated by integrating playback sections that are played back continuously. .
  • FIG. 22A is a diagram for explaining a method of generating one MP4 file by integrating playback sections.
  • playback section 1 and playback section 2 are integrated to form XXX3.
  • MP4 is generated.
  • the audio in the playback section 1 is composed of, for example, 100 samples from sample 1-1 to sample 1-100.
  • the audio in the playback section 2 is composed of, for example, 100 samples from sample 2-1 to sample 2-100.
  • audio will be described as an example.
  • samples 2-1 to 2-100 are stored following samples 1-1 to 1-100, and each sample is played back in order during playback.
  • start time: PTS, end time: PTS + playback time length start time: PTS, end time: PTS + playback time length
  • sample 2-1 is deleted and XXX3.
  • An MP4 audio track is configured. Thereby, the overlap between samples is eliminated. The same applies to the case where the playback section of the plurality of samples included in the playback section 1 and the playback section of the plurality of samples included in the playback section 2 overlap.
  • ⁇ T be the time length of the overlapping playback sections.
  • the last sample in the playback section 1 and the first sample in the playback section 2 are different from each other, and ⁇ T is stored in the trof of the moof. This ⁇ T indicates that the playback section of ⁇ T from the beginning of the audio track in the movie fragment included in the playback section 2 overlaps with the playback section of the audio track of the immediately previous movie fragment.
  • a non-reproduction section corresponding to the gap section is generated by the edit list function in MP4 or the duration-is-empty flag set in the movie fragment.
  • an overlap section (overlapping playback section) or a gap section (a section where a gap exists) may be indicated in the metadata in the MP4 header or an external file.
  • MP4 including a gap section has higher affinity with conventional MP4 than MP4 including an overlap section. Therefore, when the MP4 file is generated by integrating the playback sections including the overlap sections, for example, in the playback section 2 of FIG. 22A, samples including at least a part of the overlap sections are deleted.
  • a non-reproduction section is generated. This method can be applied not only to audio but also to video and subtitles. Further, information indicating whether or not deletion is necessary, a sample that needs to be deleted, a time length of a gap generated as a result of deletion, and the like may be stored as auxiliary information such as a copy manifest.
  • Such a playback apparatus that plays back an MP4 file plays back audio data in the overlap section based on information such as the overlap section included in the MP4 file. That is, if the audio data to be reproduced is data in the overlap section, the playback apparatus synthesizes the decoding result of the audio sample including the data in each overlapped playback section, for example. Then, the playback device outputs the synthesized decoding result. Further, when integrating playback sections that are continuously played back, it is necessary to identify whether or not an overlap section or a gap section exists in the playback section to be integrated. Therefore, information indicating the presence / absence of an overlap section and the time length thereof may be acquired as auxiliary information in the optical disc or from the Internet.
  • FIG. 22B is a block diagram of the file generation apparatus according to the present embodiment.
  • the file generation apparatus 20 in the present embodiment is a file generation apparatus that generates an MP4 file, and includes an integration unit 21 and a storage unit 22.
  • the integration unit 21 generates one MP4 file by integrating the two streams so that the two streams are continuously reproduced.
  • the storage unit 22 stores, in the generated MP4 file, information indicating a section in which the playback timing overlaps in each of the two streams.
  • the above-mentioned two streams are the playback section 1 and the playback section 2 shown in FIG. 22A
  • the generated MP4 file is a file “XXX3.MP4” shown in FIG. 22A.
  • the section where the playback timing overlaps in each of the two streams is, for example, a section of ⁇ T shown in FIG. 22A.
  • FIG. 22C is a flowchart of the file generation method in the present embodiment.
  • the file generation method in the present embodiment is a file generation method for generating an MP4 file, and includes step S21 and step S22.
  • step S21 one MP4 file is generated by integrating the two streams so that the two streams are continuously reproduced.
  • step S21 information indicating a section where the reproduction timing overlaps in each of the two streams is stored in the generated MP4 file.
  • step S21 two streams that are at least part of the original file each configured in the MP4 file format are integrated.
  • two streams each including audio data are integrated.
  • the playback device that plays back the MP4 file can easily specify the data of the overlap section from the MP4 file using the information.
  • the playback apparatus can appropriately play back the data by combining the data in the overlapping section. That is, an MP4 file suitable for overlapping reproduction can be generated.
  • step S21 if the above-described section exists over a plurality of samples included in any one of the two streams, the two streams are deleted after deleting at least one of the plurality of samples. To integrate. Thereby, since the sample is deleted, the overlap section can be shortened. As a result, it is possible to reduce the burden of special processing by the playback device for the overlap section.
  • step S22 time information indicating the time length of the above-described section is stored in the MP4 file as the above-described information. That is, the time information indicating the above ⁇ T is stored in the MP4 file.
  • a playback apparatus that plays back an MP4 file can easily specify the time length of the overlap section using the information.
  • the reproducing apparatus can appropriately reproduce the data within the specified time length by combining the data of the overlapping sections.
  • step S22 the time information is stored in the traf in the moof in the MP4 file. Thereby, the reproducing
  • the information may be acquired from a device holding the above information via a communication network such as the Internet.
  • the information may be obtained from an optical disk on which the above information is recorded. Accordingly, the information can be easily stored in the MP4 file without generating information indicating the overlap section.
  • FIG. 22D is a block diagram of the playback device in the present embodiment.
  • the playback device 30 in the present embodiment is a playback device that plays back an MP4 file, and includes an extraction unit 31 and a synthesis unit 32.
  • the extraction unit 31 extracts, from the MP4 file, information indicating two sections in which playback timing overlaps in a playback target content (for example, an audio track).
  • the synthesizing unit 32 specifies two sections in the content based on the extracted information, and synthesizes and outputs the decoding results for the respective data in the two sections.
  • FIG. 22E is a flowchart of the reproduction method according to the present embodiment.
  • the reproduction method in the present embodiment is a reproduction method for reproducing an MP4 file, and includes step S31 and step S32.
  • step S31 information indicating two sections where the playback timing overlaps in the content to be played (for example, an audio track) is extracted from the MP4 file.
  • step S32 two sections in the content are specified based on the extracted information, and the decoding results for the respective data in the two sections are synthesized and output.
  • the playback device can easily specify the data of the overlap section from the MP4 file. As a result, the playback device can appropriately play back the data in the overlap section.
  • the recording medium in the present embodiment is a recording medium that records an MP4 file.
  • the MP4 file includes content (for example, an audio track) that is read and reproduced by a computer, and information indicating two sections in which reproduction timing overlaps.
  • the playback device that reads and plays back the MP4 file from the recording medium can easily specify the data of the two sections described above from the MP4 file using the information.
  • the playback device can appropriately play back the data by combining the data of those sections.
  • FIG. 23A is a diagram illustrating an example of a menu screen when an MP4 file is generated from content stored on an optical disc.
  • FIG. 23B is a diagram for explaining a method of generating an MP4 file using an optical disc and a network.
  • audio and subtitles in multiple languages are stored, and when the MP4 file is generated, the language stored in the MP4 file can be selected.
  • Japanese and English are selected for each of audio and subtitles from Japanese, English, Spanish, and Chinese stored in the disc.
  • the sample size in audio or subtitle differs for each language, the content of the copy manifest file depends on the sample size. For this reason, the number of copy manifest files increases in proportion to the combination of selected languages in audio or subtitles. Therefore, the MP4 file may always store audio and subtitle data in all languages, further store information indicating the language selected by the user, and select the user's desired language during playback. . By doing so, the copy manifest file can be made the same regardless of the selected language.
  • an audio encoding method such as AAC or AC3 may be selected depending on whether or not a device that plays back an MP4 file has support.
  • audio data of all encoding methods may be stored in the MP4 file. In the case of storing audio data of all encoding methods, at the time of reproduction, the encoding method is selected based on a user's selection operation or preset information of the reproduction device.
  • the audio and subtitles of all languages may be stored without selecting the language when generating the MP4 file, and the user may select it during playback.
  • the copy manifest file can be acquired via the network, it is not necessary to store the copy manifest file on the optical disc. In particular, when the number of copy manifest files increases to make it possible to select an arbitrary language, acquisition via a network is effective. Only the copy manifest file corresponding to the default language combination may be stored in the optical disc, and the copy manifest file corresponding to the other combination may be downloaded from the server.
  • audio or subtitles in a language that can be acquired from an optical disc or a network, and audio or subtitle languages included in the MP4 file are respectively acquired, and an external language out of the languages not included in the MP4 file is acquired.
  • the user may select and acquire a language that can be acquired from.
  • a list of externally obtainable audio is stored in an MP4 file or the like. If the playback apparatus cannot decode the audio encoding system in the MP4 file during playback of the MP4 file, the audio of the encoding system supported by the playback apparatus may be selected and acquired from the outside. At this time, the data acquired from the outside may be encoded data including only captions or audio, or an MP4 file. At this time, at the time of reproduction, the video etc. included in the original MP4 file and the newly acquired data are reproduced synchronously. Alternatively, a complete MP4 file including all of video, audio, and subtitles may be acquired from the outside.
  • the content stored on the optical disc may be an MP4 file instead of an MPEG2-TS file.
  • the MP4 file data stored in the optical disc may be copied or exported as it is to a device-bound or media-bound recording medium or device without performing the conversion process.
  • the key for encrypting the content may be changed.
  • a device that generates an MP4 file such as a BD player converts the content in the optical disc into the MP4 file by determining whether the content in the optical disc is in the MPEG2-TS format or the MP4 format based on the identification information of the data format.
  • An MP4 file may be generated by determining whether it is necessary.
  • auxiliary information such as a copy manifest file may indicate information indicating whether conversion to MP4 is necessary. Further, even if the content stored on the optical disc is in the MP4 format, the user may select a type such as audio or subtitles in a specific language in the content, a theater release version, or a director's cut version. Then, only the data selected from the MP4 file in the optical disk based on the selection result may be extracted to generate the MP4 file.
  • the file size can also be reduced by indicating the information for specifying the multiplexing unit of the output MP4 file in the copy manifest file, and the number of copy manifest files can be reduced. It is effective when there are many.
  • the information shown in the copy manifest it is possible to uniquely determine the unit of the movie fragment in the MP4 file and use language-independent information.
  • the information is, for example, the identification information of the PTS or DTS of the sample that is the head of the movie fragment, or the MPEG2-TS file from which the sample data is acquired.
  • the MP4 file header information cannot be included in a copy manifest file that can be used in common for different languages. Therefore, when converting to an MP4 file, the unit of the movie fragment is determined based on the copy manifest file, and the header information of the MP4 file is generated based on the PTS or DTS for each sample or the size of the sample. To do.
  • the size of the NAL unit constituting the access unit (corresponding to the sample of the MP4 file) and information indicating the PTS or DTS may be stored in the copy manifest.
  • FIG. 24 is a diagram showing an example of a copy manifest showing the NAL unit size, PTS, and DTS.
  • the MPEG2-TS start code format can be converted to the MP4 NAL size format using the size information.
  • the size information may indicate the size of the start code portion and the size of the NAL unit portion separately.
  • the byte length of the field indicating the size of the NAL unit may be the same as the byte length of the size portion in the NAL size format.
  • the data in the start code portion can be converted into the NAL size format by replacing the data indicating the size of the NAL unit portion.
  • This start code corresponds to the identification information shown in FIG. 15B and includes zero_byte.
  • the PTS or DTS may be expressed not in the PTS or DTS value in the MPEG2-TS PES header but in the data format used in the MP4 header.
  • the difference value between the DTSs of two samples that are consecutive in decoding order can be shown, and for PTS, the difference value between DTS and PTS can be shown.
  • information indicating the absolute value of the PTS or DTS of the first sample may be stored. Further, information for identifying a NAL unit to be deleted at the time of conversion to an MP4 file may be added.
  • header information such as ADTS and LATM is deleted from the sample data.
  • size of the header information is fixed, only one of the total value of the header information and the size of the payload data and the size of the payload data may be indicated.
  • the frame rate is fixed, only the default value may be indicated in the DTS information.
  • the playback time length for each sample may also be different.
  • a copy manifest file may be prepared for each audio encoding method.
  • subtitle data when subtitle data is stored in the MP4 file, the subtitle data may be stored together at the end of the file.
  • FIG. 25 is a diagram showing an example of caption data stored at the end of the MP4 file.
  • the movie fragment of the AV (Advanced Video) data is the same.
  • the video or audio movie fragment and the subtitle movie fragment are interleaved and stored, it is necessary to change the contents of the moof. This is because the position of the movie fragment of the AV data changes when the size of the caption data changes.
  • caption data is smaller in size than AV data. Therefore, it is also possible to reproduce the subtitle data of the entire content or the subtitle data included in a unit such as a chapter obtained by dividing the content in a memory collectively. At this time, if subtitle data is stored together at the end of the file, there is an advantage that the subtitle data can be easily obtained.
  • the caption data may be based on a text font or image data in a PNG (Portable Network Graphics) format or the like.
  • trun may be generated for each unit such as a chapter to improve accessibility to subtitle data included in a predetermined unit.
  • the trun may be generated so that the size of the caption data constituting the trun is equal to or smaller than the buffer size in accordance with the buffer size that holds the text data during reproduction.
  • subtitle data in a specific language can be easily obtained by storing each language in a different movie fragment.
  • information for specifying the language stored in the movie fragment is required. Therefore, for example, each language may be handled as a different track, and the track ID may be associated with the language.
  • the track ID is indicated by a box in the traf.
  • the information that associates the track ID with the language may be stored in a metadata storage box or the like in MP4, or may be management information different from the MP4 file.
  • the correspondence between languages and movie fragments can be applied to audio.
  • a random accessible sample in video, audio or subtitle is indicated by mfra.
  • the playback sections of video and audio movie fragments that are continuous with each other match, only the video random access points need be shown.
  • an audio sample having the same PTS, immediately before or immediately after can be obtained from the immediately following movie fragment.
  • the PTS of the first sample of video (V-1) matches the first sample of audio (A-1).
  • the text is stored at the end of the file, it is necessary to indicate a random access point independently for the text.
  • audio or subtitle data in all languages in the optical disc content is stored in the MP4 file.
  • a random access point may be indicated in mfra only for a track corresponding to the selected language.
  • the content stored on the optical disc has different resolutions for video and graphics.
  • the resolution of video is 4K
  • the resolution of graphics such as captions is 2K in order to reduce the processing amount.
  • FIG. 26 is a diagram illustrating a case in which subtitles with 2K resolution are scaled to 4K and displayed.
  • information for specifying a subtitle display area is required together with subtitle data and its resolution.
  • the display area is specified using, for example, the size and display position of a rectangular display area.
  • the information indicating the resolution of the track can indicate that the subtitle track is 2K and the video track is 4K.
  • the XML description part of the Timed Text is part of the XML (extensive data) that constitutes the Timed Text.
  • it may be stored in a box indicating metadata in the MP4 file.
  • the resolutions of the video and subtitles are acquired, and if the resolutions of both are different, the subtitles are scaled and displayed so as to match the video resolution.
  • the caption is image data
  • the image data is enlarged, and if it is text data, a size that matches the video resolution is selected.
  • the display area is also determined by calculation according to the scaling coefficient. Information indicating the display area after scaling according to the video resolution may be stored.
  • each component may be configured by dedicated hardware, or may be realized by executing a software program suitable for each component.
  • Each component may be realized by a program execution unit such as a CPU or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
  • the software that realizes the file generation apparatus and the like in the above-described embodiment and each modification causes the computer to execute each step included in the flowchart shown in FIG. 8 or FIG.
  • the file generation device and the playback device have been described based on each embodiment and each modification.
  • the present disclosure is limited to these embodiments and each modification. It is not a thing. Unless it deviates from the gist of the present disclosure, various modifications conceived by those skilled in the art are applied to each embodiment and each modification, and a form constructed by combining the components in each embodiment and each modification is also possible. It may be included within the scope of one or more embodiments.
  • a file constituted by MPEG2-TS is used as the original file.
  • the original file may be any file or transport stream other than the MPEG2-TS file as long as the file is configured in a file format different from MP4.
  • each of these playback sections may be a movie fragment of an MP4 file, It may be a stream.
  • Each of the above devices is specifically a computer system including a microprocessor, a ROM, a RAM, a hard disk unit, a display unit, a keyboard, a mouse, and the like.
  • a computer program is stored in the RAM or hard disk unit.
  • Each device achieves its functions by the microprocessor operating according to the computer program.
  • the computer program is configured by combining a plurality of instruction codes indicating instructions for the computer in order to achieve a predetermined function.
  • a part or all of the components constituting each of the above devices may be configured by one system LSI (Large Scale Integration).
  • the system LSI is an ultra-multifunctional LSI manufactured by integrating a plurality of components on a single chip, and specifically, a computer system including a microprocessor, ROM, RAM, and the like. .
  • a computer program is stored in the RAM.
  • the system LSI achieves its functions by the microprocessor operating according to the computer program.
  • a part or all of the constituent elements constituting each of the above devices may be constituted by an IC card or a single module that can be attached to and detached from each device.
  • the IC card or the module is a computer system including a microprocessor, a ROM, a RAM, and the like.
  • the IC card or the module may include the super multifunctional LSI described above.
  • the IC card or the module achieves its function by the microprocessor operating according to the computer program. This IC card or this module may have tamper resistance.
  • the present disclosure may be the method described above. Further, the present invention may be a computer program that realizes these methods by a computer, or may be a digital signal composed of the computer program.
  • the present disclosure also relates to a computer-readable recording medium such as a flexible disk, a hard disk, a CD-ROM, an MO, a DVD, a DVD-ROM, a DVD-RAM, a BD (Blu-ray ( (Registered trademark) Disc), or recorded in a semiconductor memory or the like.
  • the digital signal may be recorded on these recording media.
  • the computer program or the digital signal may be transmitted via an electric communication line, a wireless or wired communication line, a network represented by the Internet, a data broadcast, or the like.
  • the present disclosure may be a computer system including a microprocessor and a memory, and the memory may store the computer program, and the microprocessor may operate according to the computer program.
  • program or the digital signal is recorded on the recording medium and transferred, or the program or the digital signal is transferred via the network or the like and executed by another independent computer system. You may do that.
  • the data transmission method and the data reproduction method according to one or more aspects of the present disclosure have been described based on the embodiments, but the present disclosure is not limited to the embodiments. Unless it deviates from the gist of the present disclosure, one or more of the present disclosure may be applied to various modifications conceived by those skilled in the art in the present embodiment, or forms configured by combining components in different embodiments. It may be included within the scope of the embodiments.
  • the present disclosure has an effect that the processing load can be suppressed, and can be applied to, for example, a device that converts a file format from MPEG2-TS to MP4, and can be used for a device such as a smartphone or a tablet.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Television Signal Processing For Recording (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)

Abstract

L'invention concerne, dans un mode de réalisation, un procédé de génération de fichier permettant de générer un fichier MP4 et consistant en : une étape de génération d'un fichier Mp4 en intégrant deux flux de sorte que les deux flux sont lus de manière continue; et une étape de mémorisation, dans le fichier MP4 généré, des informations indiquant un intervalle pendant lequel les moments de lecture de chaque flux des deux flux se chevauchent. Il en résulte qu'il est possible de générer un fichier MP4 adapté à la lecture en chevauchement.
PCT/JP2014/005963 2013-12-03 2014-11-28 Procédé de génération de fichier, procédé de lecture, dispositif de génération de fichier, dispositif de lecture, et support d'enregistrement. WO2015083354A1 (fr)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201361911006P 2013-12-03 2013-12-03
US61/911,006 2013-12-03
US201361918131P 2013-12-19 2013-12-19
US61/918,131 2013-12-19
JP2014-235068 2014-11-19
JP2014235068A JP2015109131A (ja) 2013-12-03 2014-11-19 ファイル生成方法、再生方法、ファイル生成装置、再生装置および記録媒体

Publications (1)

Publication Number Publication Date
WO2015083354A1 true WO2015083354A1 (fr) 2015-06-11

Family

ID=53273139

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/005963 WO2015083354A1 (fr) 2013-12-03 2014-11-28 Procédé de génération de fichier, procédé de lecture, dispositif de génération de fichier, dispositif de lecture, et support d'enregistrement.

Country Status (1)

Country Link
WO (1) WO2015083354A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3094083A4 (fr) * 2014-01-10 2016-11-16 Panasonic Ip Man Co Ltd Procédé de génération de fichiers, dispositif de génération de fichiers et support d'enregistrement
CN114143491A (zh) * 2021-11-17 2022-03-04 深蓝感知(杭州)物联科技有限公司 一个单编5g记录仪录像分片生成方法

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004007610A (ja) * 2002-04-26 2004-01-08 Matsushita Electric Ind Co Ltd 逆多重化装置
WO2005015907A1 (fr) * 2003-08-08 2005-02-17 Matsushita Electric Industrial Co., Ltd. Dispositif et procede de traitement de donnees
JP2005136503A (ja) * 2003-10-28 2005-05-26 Sony Corp ファイル記録装置及びビデオエフェクトの編集方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004007610A (ja) * 2002-04-26 2004-01-08 Matsushita Electric Ind Co Ltd 逆多重化装置
WO2005015907A1 (fr) * 2003-08-08 2005-02-17 Matsushita Electric Industrial Co., Ltd. Dispositif et procede de traitement de donnees
JP2005136503A (ja) * 2003-10-28 2005-05-26 Sony Corp ファイル記録装置及びビデオエフェクトの編集方法

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3094083A4 (fr) * 2014-01-10 2016-11-16 Panasonic Ip Man Co Ltd Procédé de génération de fichiers, dispositif de génération de fichiers et support d'enregistrement
US9818446B2 (en) 2014-01-10 2017-11-14 Panasonic Intellectual Property Management Co., Ltd. File producing method, file producing device, and recording medium
CN114143491A (zh) * 2021-11-17 2022-03-04 深蓝感知(杭州)物联科技有限公司 一个单编5g记录仪录像分片生成方法

Similar Documents

Publication Publication Date Title
JP6467680B2 (ja) ファイル生成方法およびファイル生成装置
WO2016027579A1 (fr) Appareil de traitement d'informations, support d'enregistrement d'informations, procédé de traitement d'informations, et programme
JP2019024229A (ja) ファイル生成方法およびファイル生成装置
JP2013093755A (ja) 映像記録装置及び映像記録方法
JP7052733B2 (ja) 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム
JP2015109131A (ja) ファイル生成方法、再生方法、ファイル生成装置、再生装置および記録媒体
JP2017183762A (ja) 映像ストリーム生成方法、再生装置及び記録媒体
JP6500895B2 (ja) 情報処理装置、情報生成装置、情報記録媒体、および情報処理方法、並びにプログラム
WO2015083354A1 (fr) Procédé de génération de fichier, procédé de lecture, dispositif de génération de fichier, dispositif de lecture, et support d'enregistrement.
JP7215415B2 (ja) 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム
WO2018190122A1 (fr) Dispositif de traitement d'informations, procédé de traitement d'informations et programme
WO2015105037A1 (fr) Procédé de génération de fichiers, dispositif de génération de fichiers et support d'enregistrement
WO2016027426A1 (fr) Procédé de génération de flux vidéo, appareil de lecture, et support d'enregistrement
CN110462735B (zh) 信息处理设备、信息记录介质、信息处理方法和程序
JP6982829B2 (ja) 記録装置、記録方法および記録媒体
WO2015072127A1 (fr) Procédé de génération de fichier et appareil de génération de fichier
WO2016002127A1 (fr) Conversion de format mpeg -2-ts à mp4 sans déchiffrement
WO2015186479A1 (fr) Appareil de traitement d'informations, support d'enregistrement d'informations, procédé de traitement d'informations, et programme

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14867098

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14867098

Country of ref document: EP

Kind code of ref document: A1