CN112954435A - Receiving apparatus and receiving method - Google Patents

Receiving apparatus and receiving method Download PDF

Info

Publication number
CN112954435A
CN112954435A CN202110234610.8A CN202110234610A CN112954435A CN 112954435 A CN112954435 A CN 112954435A CN 202110234610 A CN202110234610 A CN 202110234610A CN 112954435 A CN112954435 A CN 112954435A
Authority
CN
China
Prior art keywords
data
unit
packet
mpu
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110234610.8A
Other languages
Chinese (zh)
Other versions
CN112954435B (en
Inventor
中泽太峰
井口贺敬
远间正真
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Intellectual Property Management Co Ltd
Original Assignee
Panasonic Intellectual Property Management Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Panasonic Intellectual Property Management Co Ltd filed Critical Panasonic Intellectual Property Management Co Ltd
Priority claimed from CN201680059146.0A external-priority patent/CN108141636B/en
Publication of CN112954435A publication Critical patent/CN112954435A/en
Application granted granted Critical
Publication of CN112954435B publication Critical patent/CN112954435B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/438Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving MPEG packets from an IP network
    • H04N21/4382Demodulation or channel decoding, e.g. QPSK demodulation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/02Arrangements for relaying broadcast information
    • H04H20/08Arrangements for relaying broadcast information among terminal devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/28Arrangements for simultaneous broadcast of plural pieces of information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/28Arrangements for simultaneous broadcast of plural pieces of information
    • H04H20/30Arrangements for simultaneous broadcast of plural pieces of information by a single channel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/40Arrangements for broadcast specially adapted for accumulation-type receivers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H60/00Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
    • H04H60/68Systems specially adapted for using specific information, e.g. geographical or meteorological information
    • H04H60/73Systems specially adapted for using specific information, e.g. geographical or meteorological information using meta-information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04JMULTIPLEX COMMUNICATION
    • H04J3/00Time-division multiplex systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4305Synchronising client clock from received content stream, e.g. locking decoder clock with encoder clock, extraction of the PCR packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • H04N21/4343Extraction or processing of packetized elementary streams [PES]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/438Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving MPEG packets from an IP network
    • H04N21/4385Multiplex stream processing, e.g. multiplex stream decrypting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4622Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet

Abstract

Provided are a receiving apparatus and a receiving method. The receiving device is provided with: a first processing unit that receives multiplexed data composed of first multiplexed data of a first multiplexing system and outputs the received multiplexed data; and a conversion unit that converts a multiplexing scheme of the first multiplexed data among the multiplexed data that is output into a second multiplexing scheme, and outputs converted data obtained by the conversion, wherein the conversion unit extracts a part of the first multiplexed data, performs a first conversion of storing a first packet constituting the first data into a second packet used in the second multiplexing scheme, and outputs first converted data composed of the second packet obtained by the first conversion.

Description

Receiving apparatus and receiving method
This application is a divisional application of a patent application having a chinese patent application number of 201680059146.0 (international application number PCT/JP2016/004020), entitled "receiving apparatus and receiving method", filed on 9/2/2016.
Technical Field
The present disclosure relates to a receiving apparatus and a receiving method.
Background
With the advanced development of broadcasting and communication services, ultra-high-definition moving image content such as 8K (7680 × 4320 pixels: hereinafter also referred to as "8K 4K") and 4K (3840 × 2160 pixels: hereinafter also referred to as "4K 2K") has been studied and introduced. The receiving apparatus needs to decode and display encoded data of received ultra-high definition moving pictures in real time, and the processing load when decoding moving pictures with a resolution of 8K or the like is large, and it is difficult to decode such moving pictures in real time by 1 decoder. Therefore, a method of performing decoding processing in parallel by using a plurality of decoders to reduce the processing load per 1 decoder and achieve real-time processing has been studied.
The encoded data is multiplexed by a multiplexing method such as MPEG-2TS (Transport Stream) or MMT (MPEG Media Transport) and then transmitted. For example, non-patent document 1 discloses a technique for transmitting encoded media data packet by packet in accordance with MMT.
Prior art documents
Non-patent document
Non-patent document 1: information technology-High efficiency coding and media delivery in heterologous environment-Part 1: MPEG Media Transport (MMT), ISO/IEC DIS 23008-1
Disclosure of Invention
However, there are various multiplexing methods used in the TV broadcast method, and for example, there are a TS method and an IP multiplexing method as the multiplexing method. Therefore, when the receiving apparatuses corresponding to the plurality of systems are mounted in the respective systems, the circuit scale becomes large, and the cost increases.
Therefore, one embodiment of the present disclosure provides a receiving apparatus and the like that can be easily implemented in common with existing installations and can be implemented at low cost.
Means for solving the problems
A receiving device according to an aspect of the present disclosure includes: a first processing unit that receives multiplexed data composed of first multiplexed data of a first multiplexing system and outputs the received multiplexed data; and a conversion unit that converts a multiplexing scheme of the first multiplexed data among the multiplexed data that is output into a second multiplexing scheme, and outputs converted data obtained by the conversion, wherein the conversion unit extracts a part of the first multiplexed data, performs a first conversion of storing a first packet constituting the first data into a second packet used in the second multiplexing scheme, and outputs first converted data composed of the second packet obtained by the first conversion.
In a reception method according to an aspect of the present disclosure, multiplexed data is received, and the received multiplexed data is output, the multiplexed data being composed of first multiplexed data of a first multiplexing scheme; the method includes converting a multiplexing scheme of the first multiplexed data among the multiplexed data to be output into a second multiplexing scheme, outputting converted data obtained by the conversion, extracting a part of the first data of the first multiplexed data during the conversion, performing a first conversion of storing a first packet constituting the first data into a second packet used in the second multiplexing scheme, and outputting first converted data constituted by the second packet obtained by the first conversion.
A receiving device according to an aspect of the present disclosure includes: a first processing unit that receives a broadcast signal in which multiplexed data is modulated, demodulates the received broadcast signal, and outputs the multiplexed data obtained by the demodulation, the multiplexed data being composed of at least first multiplexed data of a first multiplexing scheme and second multiplexed data of a second multiplexing scheme different from the first multiplexing scheme; and a conversion unit that converts a multiplexing scheme of the first multiplexed data among the multiplexed data that is output into the second multiplexing scheme, and outputs converted data obtained by the conversion.
In addition, a receiving apparatus according to an aspect of the present disclosure includes: a reception unit configured to receive conversion data in which first conversion data and second conversion data are multiplexed, the first conversion data being composed of a second packet in which a first packet constituting first multiplexing data of a first multiplexing scheme is stored and which is used in a second multiplexing scheme different from the first multiplexing scheme, and the second conversion data being composed of a second packet obtained by converting the first packet into the second multiplexing scheme; an inverse multiplexing unit that performs inverse multiplexing processing of inverse multiplexing the conversion data received by the receiving unit into the first conversion data and the second conversion data; a first decoding unit that extracts the first packet from the second packet constituting the first converted data obtained by the inverse multiplexing process, and performs a first decoding process based on the first multiplexing scheme on first data constituted by the extracted first packet; a second decoding unit configured to perform a second decoding process based on the second multiplexing scheme on second data configured by the second packet configuring the second converted data obtained by the inverse multiplexing process; and an output unit that outputs first decoded data obtained by the first decoding process and second decoded data obtained by the second decoding process.
The overall or specific aspects may be implemented by a method, a system, an integrated circuit, a computer program, or a computer-readable recording medium such as a CD-ROM, or any combination of a method, a system, an integrated circuit, a computer program, and a recording medium.
The receiving apparatus of the present disclosure can be easily implemented in common with existing installations, and can be implemented at low cost.
Drawings
Fig. 1 is a diagram showing an example of dividing a picture into slice segments (slice segments).
Fig. 2 is a diagram showing an example of a PES packet sequence in which picture data is stored.
Fig. 3 is a diagram showing an example of dividing a picture according to embodiment 1.
Fig. 4 is a diagram showing an example of dividing a picture according to a comparative example of embodiment 1.
Fig. 5 is a diagram showing an example of data of an access unit according to embodiment 1.
Fig. 6 is a block diagram of a transmitting apparatus according to embodiment 1.
Fig. 7 is a block diagram of a receiving apparatus according to embodiment 1.
Fig. 8 is a diagram showing an example of an MMT packet according to embodiment 1.
Fig. 9 is a diagram showing another example of the MMT packet according to embodiment 1.
Fig. 10 is a diagram showing an example of data input to each decoding unit according to embodiment 1.
Fig. 11 is a diagram showing an example of an MMT packet and header information according to embodiment 1.
Fig. 12 is a diagram showing another example of data input to each decoding unit according to embodiment 1.
Fig. 13 is a diagram showing an example of dividing a picture according to embodiment 1.
Fig. 14 is a flowchart of a transmission method according to embodiment 1.
Fig. 15 is a block diagram of a receiving apparatus according to embodiment 1.
Fig. 16 is a flowchart of a receiving method according to embodiment 1.
Fig. 17 is a diagram showing an example of an MMT packet and header information according to embodiment 1.
Fig. 18 is a diagram showing an example of an MMT packet and header information according to embodiment 1.
Fig. 19 is a diagram showing a configuration of an MPU.
Fig. 20 is a diagram showing a structure of MF metadata.
Fig. 21 is a diagram for explaining a data transmission sequence.
Fig. 22 is a diagram showing an example of a method for decoding without using header information.
Fig. 23 is a block diagram of a transmitting apparatus according to embodiment 2.
Fig. 24 is a flowchart of a transmission method according to embodiment 2.
Fig. 25 is a block diagram of a receiving apparatus according to embodiment 2.
Fig. 26 is a flowchart of an operation for specifying the MPU start position and NAL unit position.
Fig. 27 is a flowchart of an operation of acquiring initialization information based on a transmission order type and decoding media data based on the initialization information.
Fig. 28 is a flowchart of the operation of the receiving apparatus when the low-latency presentation mode is set.
Fig. 29 is a diagram showing an example of the transmission sequence of the MMT packet when transmitting the auxiliary data.
Fig. 30 is a diagram for explaining an example in which the transmission apparatus generates the auxiliary data based on the structure of moof.
Fig. 31 is a diagram for explaining reception of assistance data.
Fig. 32 is a flowchart of the receiving operation using the auxiliary data.
Fig. 33 is a diagram showing a configuration of an MPU configured by a plurality of movie fragments.
Fig. 34 is a diagram for explaining a transmission procedure of the MMT packet when the MPU having the configuration of fig. 33 is transmitted.
Fig. 35 is a diagram 1 for explaining an operation example of the receiving apparatus when 1 MPU is configured by a plurality of movie fragments.
Fig. 36 is a diagram 2 for explaining an operation example of the receiving apparatus when 1 MPU is configured by a plurality of movie fragments.
Fig. 37 is a flowchart illustrating the operation of the receiving method described with reference to fig. 35 and 36.
Fig. 38 is a diagram showing a case where non-VCL NAL units are individually aggregated as Data units.
Fig. 39 is a diagram showing a case where Data units are collectively obtained from non-VCL NAL units.
Fig. 40 is a flowchart of the operation of the receiving apparatus when packet loss occurs.
Fig. 41 is a flowchart of a receiving operation when the MPU is divided into a plurality of movie fragments.
Fig. 42 is a diagram showing an example of a prediction structure of a picture for each temporalld when temporal adaptability is realized.
Fig. 43 is a diagram showing a relationship between the Decoding Time (DTS) and the display time (PTS) of each picture in fig. 42.
Fig. 44 is a diagram showing an example of a prediction structure of a picture that requires a picture delay process and a picture reordering process.
Fig. 45 is a diagram showing an example in which an MPU configured in the form of MP4 is divided into a plurality of movie fragments and stored in an MMTP payload or an MMTP packet.
Fig. 46 is a diagram for explaining a calculation method and items to be studied of PTS and DTS.
Fig. 47 is a flowchart of a reception operation when calculating a DTS using information for calculating the DTS.
Fig. 48 is a diagram for explaining a method of depositing a data unit in MMT into a payload.
Fig. 49 is a flowchart of the operation of the transmission device according to embodiment 3.
Fig. 50 is a flowchart of the operation of the receiving apparatus according to embodiment 3.
Fig. 51 is a diagram showing an example of a specific configuration of a transmission device according to embodiment 3.
Fig. 52 is a diagram showing an example of a specific configuration of a receiving apparatus according to embodiment 3.
FIG. 53 shows a method of storing non-timed media in an MPU and a method of transferring non-timed media in an MMTP packet.
Fig. 54 is a diagram showing an example in which a plurality of pieces of divided data obtained by dividing a file are packed and transmitted for each piece of divided data.
Fig. 55 is a diagram showing another example in which a plurality of pieces of divided data obtained by dividing a file are packed and transferred for each piece of divided data.
Fig. 56 is a diagram showing the syntax of a loop for each file in the resource management table.
Fig. 57 is a flowchart of an operation of determining a divided data number in the receiving apparatus.
Fig. 58 is a flowchart of an operation of determining the number of divided data in the receiving apparatus.
Fig. 59 is a flowchart of an operation for determining whether to operate a slice counter in the transmission apparatus.
Fig. 60 is a diagram for explaining a method of determining the number of pieces of divided data and the number of pieces of divided data (in the case of using a slice counter).
Fig. 61 is a flowchart of the operation of the transmitting apparatus in the case of using the slice counter.
Fig. 62 is a flowchart of the operation of the receiving apparatus in the case of using the slice counter.
Fig. 63 is a diagram showing a service configuration in a case where the same program is transmitted by a plurality of IP data streams.
Fig. 64 is a diagram showing an example of a specific configuration of a transmitting apparatus.
Fig. 65 is a diagram showing an example of a specific configuration of a receiving apparatus.
Fig. 66 is a flowchart of the operation of the transmission device.
Fig. 67 is a flowchart of the operation of the receiving apparatus.
Fig. 68 is a diagram showing a reception buffer model in the case of using a broadcast transmission path based on the reception buffer model defined in ARIB STD-B60.
Fig. 69 is a diagram showing an example in which a plurality of data units are collectively stored in one payload.
Fig. 70 is a diagram showing an example of a case where a plurality of data units are collectively stored in one payload and a video signal in NAL size format is regarded as one data unit.
Fig. 71 is a diagram showing the structure of the payload of an MMTP packet whose data unit length is not shown.
Fig. 72 is a diagram showing an extended region given to a packet unit.
Fig. 73 is a flowchart of the operation of the receiving apparatus.
Fig. 74 is a diagram showing an example of a specific configuration of a transmitting apparatus.
Fig. 75 is a diagram showing an example of a specific configuration of a receiving apparatus.
Fig. 76 is a flowchart of the operation of the transmission device.
Fig. 77 is a flowchart of the operation of the receiving apparatus.
FIG. 78 is a diagram showing a protocol stack of the MMT/TLV scheme defined by ARIB STD-B60.
Fig. 79 is a diagram showing the structure of a TLV packet.
Fig. 80 is a diagram showing an example of a block diagram of a receiving apparatus.
Fig. 81 is a diagram for explaining a time stamp descriptor.
Fig. 82 is a diagram for explaining leap second adjustment.
Fig. 83 is a diagram showing a relationship between NTP time, MPU time stamp, and MPU presentation timing.
Fig. 84 is a diagram for explaining a method of correcting a time stamp on the transmission side.
Fig. 85 is a diagram for explaining a correction method of correcting a time stamp in a receiving apparatus.
Fig. 86 is a flowchart of the operation of the transmitting side (transmitting apparatus) in the case where the MPU timestamp is corrected in the transmitting side (transmitting apparatus).
Fig. 87 is a flowchart of the operation of the receiving apparatus in the case where the MPU timestamp is corrected in the transmitting side (transmitting apparatus).
Fig. 88 is a flowchart of the operation of the transmitting side (transmitting apparatus) in the case where the MPU timestamp is corrected in the receiving apparatus.
Fig. 89 is a flowchart of the operation of the receiving apparatus in the case where the MPU time stamp is corrected in the receiving apparatus.
Fig. 90 is a diagram showing an example of a specific configuration of a transmitting apparatus.
Fig. 91 is a diagram showing an example of a specific configuration of a receiving apparatus.
Fig. 92 is a flowchart of the operation of the transmitting apparatus.
Fig. 93 is a flowchart of the operation of the receiving apparatus.
Fig. 94 is a diagram showing an example of expansion of an MPU expansion timestamp descriptor.
Fig. 95 is a diagram for explaining a case where the MPU serial number is discontinuously generated by performing the adjustment of the MPU serial number.
Fig. 96 is a diagram for explaining a case where packet sequence numbers become discontinuous at the timing of switching from the normal device to the redundant system device.
Fig. 97 is a flowchart of the operation of the receiving apparatus when the MPU serial number or packet serial number is not consecutive.
Fig. 98 is a diagram for explaining a method of correcting a time stamp in the receiving apparatus when inserting leap seconds.
Fig. 99 is a diagram for explaining a method of correcting a time stamp in the leap second deletion time receiving apparatus.
Fig. 100 is a flowchart of the operation of the receiving apparatus.
Fig. 101 is a diagram showing an example of a specific configuration of a transmission/reception system.
Fig. 102 is a diagram showing a specific configuration of a receiving apparatus.
Fig. 103 is a flowchart of the operation of the receiving apparatus.
Fig. 104 is a diagram for explaining a method of correcting a time stamp in the leap second insertion transmission side (transmission apparatus).
Fig. 105 is a diagram for explaining a method of correcting the leap second time stamp in the leap second time erasure transmission side (transmission apparatus).
Fig. 106 is a flowchart of the operation of the transmitting apparatus described in fig. 104.
Fig. 107 is a flowchart of the operation of the receiving apparatus described in fig. 104.
Fig. 108 is a flowchart of the operation of the transmitting apparatus described in fig. 105.
Fig. 109 is a flowchart of the operation of the receiving apparatus described in fig. 105.
Fig. 110 is a diagram showing an example of a specific configuration of a transmitting apparatus.
Fig. 111 is a diagram showing an example of a specific configuration of a receiving apparatus.
Fig. 112 is a flowchart of an operation of the transmission apparatus (transmission method).
Fig. 113 is a flowchart of the operation of the receiving apparatus (receiving method).
FIG. 114 is a diagram showing the details of a protocol stack diagram of the MMT/TLV scheme defined in ARIB STD-B60.
Fig. 115 is a block diagram of a receiving apparatus.
Fig. 116 is a diagram showing a general broadcast protocol multiplexed in the MPEG-2 TS system.
Fig. 117 is a block diagram showing a receiving apparatus that receives broadcast signals multiplexed in the TS scheme.
Fig. 118A is a diagram showing a configuration of a conventional receiving apparatus in the case of processing a single system.
Fig. 118B is a diagram showing a configuration of a conventional receiving apparatus that independently processes a plurality of systems.
Fig. 118C is a diagram showing an example of the configuration of the receiving apparatus in the case of using the mode switching mechanism.
Fig. 118D is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism.
Fig. 118E is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism.
Fig. 118F is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism.
Fig. 119 is a diagram showing a modification of the configuration of the receiver in the case of using the mode switching mechanism.
Fig. 120 is a diagram showing an example of the details of the mode switching mechanism.
Fig. 121 is a diagram showing a flow of reception processing for performing reception using the reception apparatus shown in fig. 118F.
Fig. 122 is a diagram showing a process flow of the mode switching mechanism in fig. 119.
Fig. 123 is a diagram showing a processing flow of the mode processing means B in fig. 119.
Fig. 124 shows an example of a specific configuration of a receiving apparatus.
Fig. 125 is a diagram showing another specific example of the configuration of the receiving apparatus.
Fig. 126 is a diagram showing an operation flow (reception method) of the reception device.
Fig. 127 is a diagram showing an operation flow (reception method) of the reception device.
Detailed Description
A receiving device according to an aspect of the present disclosure includes: a first processing unit that receives a broadcast signal in which multiplexed data is modulated, demodulates the received broadcast signal, and outputs the multiplexed data obtained by the demodulation, the multiplexed data being composed of at least first multiplexed data of a first multiplexing scheme and second multiplexed data of a second multiplexing scheme different from the first multiplexing scheme; and a conversion unit that converts a multiplexing scheme of the first multiplexed data among the multiplexed data that is output into the second multiplexing scheme, and outputs converted data obtained by the conversion.
Such a receiving apparatus can be easily shared with existing installations and can be realized at low cost.
Further, the conversion unit may include: a first conversion unit that extracts first data that is a part of the first multiplexed data, performs first conversion of storing a first packet that constitutes the first data in a second packet that is used in the second multiplexing scheme, and outputs first converted data that is composed of the second packet obtained by the first conversion; a second conversion unit that extracts the first packet of second data constituting a remaining part of the first multiplexed data, performs second conversion of the extracted first packet into a second packet of the second multiplexing scheme, and outputs second conversion data constituted by the second packet obtained by the second conversion; and a multiplexing unit that performs multiplexing processing of multiplexing the first conversion data and the second conversion data that are output; the conversion unit outputs data obtained by the multiplexing process as the conversion data.
The first data and the second data may be data of different media.
The first conversion unit may assign a first identifier to the second packet of the first conversion data, the first identifier indicating the second packet obtained by the first conversion, the second conversion unit may assign a second identifier to the second packet of the second conversion data, and the second identifier indicating the second packet obtained by the second conversion.
Further, the present invention may further include: and a retransmission unit that retransmits the converted data output by the conversion unit to another receiving apparatus.
Further, the present invention may further include: and an accumulation unit that accumulates the conversion data output by the conversion unit in a storage device.
Further, the present invention may further include: and a retransmission unit that retransmits the converted data stored in the storage device to another receiving device.
Further, the present invention may further include: and a second processing unit that performs decoding processing for decoding the converted data output by the conversion unit, and outputs decoded data obtained by the decoding processing.
Further, the present invention may further include: a second processing unit that performs decoding processing for decoding the converted data output by the conversion unit and outputs decoded data obtained by the decoding processing, the second processing unit including: an inverse multiplexing unit that performs inverse multiplexing processing of inverse-multiplexing the conversion data output by the conversion unit into the first conversion data and the second conversion data; a first decoding unit configured to perform first decoding processing based on the first multiplexing scheme on the first data constituted by the first packet extracted from the second packet constituting the first converted data obtained by the inverse multiplexing processing; and a second decoding unit configured to perform second decoding processing based on the second multiplexing scheme on the second data configured by the second packet configuring the second converted data obtained by the inverse multiplexing processing, wherein the second processing unit outputs, as the decoded data, first decoded data obtained by the first decoding processing and second decoded data obtained by the second decoding processing.
Further, the receiving apparatus may further include: and an adjustment unit configured to perform adjustment for matching one of the first control information and the second control information with the other of the first control information and the second control information, using first control information of the first decoded data and second control information of the second decoded data.
The first control information may be first reference clock information, the second control information may be second reference clock information, and the adjusting unit may synchronize the first decoded data and the second decoded data by performing adjustment to match one of the first reference clock information and the second reference clock information with the other.
In addition, the first multiplexing mode may be an MMT/TLV mode, i.e., a moving picture experts group media transport/type length value mode, and the second multiplexing mode may be a TS mode, i.e., a transport stream mode.
A receiving device according to an aspect of the present disclosure includes: a reception unit configured to receive conversion data in which first conversion data and second conversion data are multiplexed, the first conversion data being composed of a second packet in which a first packet constituting first multiplexing data of a first multiplexing scheme is stored and which is used in a second multiplexing scheme different from the first multiplexing scheme, the second conversion data being composed of a second packet obtained by converting the first packet into the second multiplexing scheme; an inverse multiplexing unit that performs inverse multiplexing processing of inverse multiplexing the conversion data received by the receiving unit into the first conversion data and the second conversion data; a first decoding unit that extracts the first packet from the second packet constituting the first converted data obtained by the inverse multiplexing process, and performs a first decoding process based on the first multiplexing scheme on first data constituted by the extracted first packet; a second decoding unit configured to perform a second decoding process based on the second multiplexing scheme on second data configured by the second packet configuring the second converted data obtained by the inverse multiplexing process; and an output unit that outputs first decoded data obtained by the first decoding process and second decoded data obtained by the second decoding process.
These general and specific aspects may be implemented by a method, a system, an integrated circuit, a computer program, or a computer-readable recording medium such as a CD-ROM, or any combination of a method, a system, an integrated circuit, a computer program, or a recording medium.
The embodiments are described below in detail with reference to the drawings.
The embodiments described below are all general or specific examples. The numerical values, shapes, materials, constituent elements, arrangement positions and connection modes of the constituent elements, steps, order of the steps, and the like shown in the following embodiments are merely examples, and are not intended to limit the present disclosure. Among the components of the following embodiments, components that are not recited in the independent claims indicating the highest concept will be described as arbitrary components.
(knowledge forming the basis of the present disclosure)
Recently, high resolution of a display of a television, a smart mobile phone, a tablet terminal, or the like has been increasingly advanced. In particular, the broadcast in japan is scheduled for a service of 8K4K (resolution 8K × 4K) in 2020. In the case of a super-high resolution video such as 8K4K, it is difficult to decode the video in real time by a single decoder, and therefore, a method of performing decoding processing in parallel by a plurality of decoders has been studied.
Since the encoded data is multiplexed by a multiplexing system such as MPEG-2TS or MMT and transmitted, the receiving apparatus needs to separate the encoded data of the moving image from the multiplexed data before decoding. Hereinafter, the process of separating encoded data from multiplexed data is referred to as "inverse multiplexing".
When decoding processing is performed in parallel, it is necessary to allocate encoded data to be decoded to each decoder. When distributing the encoded data, the encoded data itself needs to be analyzed, and particularly in the content of 8K or the like, the bit rate is very high, and therefore, the processing load due to the analysis is heavy. Therefore, there are problems as follows: the inverse multiplexed portion becomes a bottleneck and cannot be reproduced in real time.
In the moving picture coding schemes such as h.264 and h.265 standardized by MPEG and ITU, a transmission apparatus codes a picture by dividing the picture into a plurality of regions called "slices" or "slice segments" and independently decoding each of the divided regions. Therefore, for example, in h.265, a receiving apparatus that receives a broadcast can achieve parallel decoding processing by separating data of each slice from received data and outputting the data of each slice to each decoder.
Fig. 1 is a diagram showing an example of dividing 1 picture into 4 slices in HEVC. For example, the receiving apparatus includes 4 decoders, and each decoder decodes any one of the 4 slice segments.
In conventional broadcasting, a transmitting apparatus stores 1 picture (access unit in MPEG system standard) in 1 PES packet, and multiplexes the PES packets into a TS packet sequence. Therefore, the receiving device needs to separate the respective clips by parsing the data of the access unit stored in the payload after separating the payload of the PES packet, and output the data of the respective separated clips to the decoder.
However, the inventors have found that: since the processing amount when analyzing data of an access unit to separate fragments is large, it is difficult to perform the processing in real time.
Fig. 2 is a diagram showing an example in which data of a picture divided into slices is stored in a payload of a PES packet.
As shown in FIG. 2, for example, data of a plurality of slices (slices 1-4) is stored in the payload of 1 PES packet. Further, the PES packets are multiplexed into a TS packet train.
(embodiment mode 1)
Hereinafter, a case where h.265 is used as an example of a coding method for a moving picture will be described, but the present embodiment can be applied to a case where another coding method such as h.264 is used.
Fig. 3 is a diagram showing an example of dividing an access unit (picture) in units of division in the present embodiment. The access unit is divided into 2 equal parts in the horizontal and vertical directions by a function called "tile (tile)" imported by h.265, and divided into 4 tiles in total. Further, the sliced segments create associations with tiles in a one-to-one correspondence.
The reason why the sample is equally divided into 2 parts in the horizontal and vertical directions will be described. First, in decoding, a line memory for storing data of horizontal 1 line is generally required, but if the resolution is ultrahigh, such as 8K4K, the size of the line memory increases because the size in the horizontal direction increases. In the installation of the receiving apparatus, it is preferable to be able to reduce the size of the line memory. In order to reduce the size of the line memory, division in the vertical direction is required. A data structure such as a tile is required in the vertical direction division. Tiles are utilized for these reasons.
On the other hand, since an image generally has a high correlation in the horizontal direction, the encoding efficiency is further improved if a large range in the horizontal direction can be referred to. Therefore, from the viewpoint of coding efficiency, it is preferable to divide the access unit in the horizontal direction.
By equally dividing the access unit by 2 in the horizontal and vertical directions, these 2 characteristics are taken into consideration, and both the mounting aspect and the encoding efficiency can be taken into consideration. When a single decoder can decode a 4K2K moving image in real time, the receiving apparatus can decode an 8K4K image in real time by dividing an 8K4K image 4 into equal parts and dividing each of the divided parts into 4K 2K.
Next, the reason why tiles obtained by dividing an access unit in the horizontal and vertical directions are associated with slice segments in a one-to-one correspondence will be described. In h.265, an access unit is composed of a plurality of units called "NAL (Network Adaptation Layer) units".
The payload of a NAL unit stores any one of an access unit delimiter indicating the start position of an access unit, an SPS (Sequence Parameter Set) as initialization Information at the time of decoding commonly used in Sequence units, a PPS (Picture Parameter Set) as initialization Information at the time of decoding commonly used in pictures, an SEI (Supplemental Enhancement Information) that is not necessary for the decoding process itself but is necessary for processing and display of the decoding result, and encoded data of slices. The header of the NAL unit contains type information for identifying data held in the payload.
Here, the transmitting apparatus can set the basic unit as a NAL unit when multiplexing encoded data in a multiplexing format such as MPEG-2TS, MMT (MPEG Media Transport: moving picture experts group Media Transport), MPEG DASH (Dynamic Adaptive Streaming over HTTP), or RTP (Real-time Transport Protocol). In order to store 1 slice segment in 1 NAL unit, it is preferable to divide an access unit into regions in units of slice segments. For this reason, the transmitting apparatus creates an association of tiles with slice segments in one-to-one correspondence.
As shown in fig. 4, the transmission apparatus can set 1 tile segment to tile 4 in a group. However, in this case, since all tiles are held in 1 NAL unit, it is difficult for the receiving apparatus to separate the tiles from the multiplexing layer.
In addition, although there are an independent slice that can be independently decoded and a reference slice that refers to the independent slice in the slice segment, a case where the independent slice is used will be described here.
Fig. 5 is a diagram showing an example of data of an access unit divided so that tiles match the boundaries of slice segments as shown in fig. 3. The data of the access unit includes NAL units storing access unit delimiters arranged at the head, NAL units storing SPS, PPS, and SEI arranged at the rear, and data of slices storing data of tiles 1 to 4 arranged at the rear. In addition, the data of the access unit may not contain part or all of the NAL units of SPS, PPS, and SEI.
Next, the configuration of the transmission device 100 according to the present embodiment will be described. Fig. 6 is a block diagram showing an example of the configuration of the transmission device 100 according to the present embodiment. The transmission device 100 includes an encoding unit 101, a multiplexing unit 102, a modulation unit 103, and a transmission unit 104.
The encoding unit 101 generates encoded data by encoding an input image, for example, in accordance with h.265. Further, for example, as shown in fig. 3, the encoding unit 101 divides an access unit into 4 slices (tiles) and encodes each slice.
The multiplexing unit 102 multiplexes the encoded data generated by the encoding unit 101. The modulation unit 103 modulates the data obtained by multiplexing. The transmitter 104 transmits the modulated data as a broadcast signal.
Next, the configuration of the receiving apparatus 200 according to the present embodiment will be described. Fig. 7 is a block diagram showing an example of the configuration of the receiving apparatus 200 according to the present embodiment. The reception device 200 includes a tuner 201, a demodulation unit 202, an inverse multiplexing unit 203, a plurality of decoding units 204A to 204D, and a display unit 205.
The tuner 201 receives a broadcast signal. The demodulation unit 202 demodulates the received broadcast signal. The demodulated data is input to the inverse multiplexing unit 203.
The inverse multiplexing unit 203 separates the demodulated data into division units, and outputs the data of each division unit to the decoding units 204A to 204D. Here, the division unit means a division region obtained by dividing an access unit, for example, a slice in h.265. Here, the 8K4K image is divided into 4K2K images. Therefore, there are 4 decoding units 204A to 204D.
The plurality of decoding units 204A to 204D operate in synchronization with each other based on a predetermined reference clock. Each Decoding unit decodes the coded data of the division unit in accordance with the DTS (Decoding Time Stamp) of the access unit, and outputs the Decoding result to the display unit 205.
The display unit 205 combines the plurality of decoding results output from the plurality of decoding units 204A to 204D to generate an output image of 8K 4K. The display unit 205 displays the generated output image according to a PTS (Presentation Time Stamp) of the access unit acquired separately. In addition, when merging the decoding results, the display unit 205 may perform filtering processing such as deblocking filtering on boundary regions of adjacent division units such as boundaries between tiles so that the boundaries are visually inconspicuous.
In the above description, the transmission device 100 and the reception device 200 that transmit or receive a broadcast are described as an example, but the content may be transmitted and received via a communication network. When the reception device 200 receives the content via the communication network, the reception device 200 separates multiplexed data from an IP packet received via a network such as ethernet.
In broadcasting, the transmission path delay from the transmission of a broadcast signal to the arrival at the receiving apparatus 200 is fixed. On the other hand, in a communication network such as the internet, a transmission path delay from data transmitted from a server to the reception device 200 is not constant due to the influence of congestion. Therefore, the receiving apparatus 200 often does not perform strict synchronous playback of a reference clock based on PCR in the broadcast MPEG-2 TS. Therefore, the receiving apparatus 200 may display the output image of 8K4K on the display unit according to the PTS without strictly synchronizing the decoding units.
Further, decoding processing for all the division units may not be completed at the time indicated by the PTS of the access unit due to congestion of the communication network or the like. In this case, the reception apparatus 200 skips the display of the access unit, or delays the display until the decoding of at least 4 division units is completed and the generation of the image of 8K4K is completed.
In addition, the content may be transmitted and received by using broadcasting and communication in combination. The method can also be applied to the reproduction of multiplexed data stored in a recording medium such as a hard disk or a memory.
Next, a multiplexing method of access units divided into slices when MMT is used as a multiplexing method will be described.
Fig. 8 is a diagram showing an example of packing data of an access unit of HEVC into an MMT. SPS, PPS, SEI, and the like are not necessarily included in the access unit, but their existence is illustrated in this example.
NAL units such as an access unit delimiter, SPS, PPS, and SEI arranged in an access unit before the first slice segment are grouped together and stored in the MMT packet # 1. The subsequent clips are stored in different MMT packets for each clip.
As shown in fig. 9, the NAL unit placed before the top slice in the access unit may be stored in the same MMT packet as the top slice.
Furthermore, if NAL units of End-of-Sequence or End-of-Bitstream, etc. indicating the End of a Sequence or stream are appended after the last slice segment, these are kept in the same MMT packet as the last slice segment. However, since NAL units such as End-of-Sequence and End-of-Bitstream may be inserted into the End point of the decoding process or the connection point of 2 streams, it is preferable that the receiving apparatus 200 be able to easily acquire these NAL units from the multiplex layer. In this case, these NAL units may also be stored in MMT packets different from slice segments. Accordingly, the receiving apparatus 200 can easily separate these NAL units from the multiplex layer.
Further, TS, DASH, RTP, or the like may be used as the multiplexing method. In these systems, the transmission apparatus 100 also stores different slice segments in different packets. This ensures that the receiving apparatus 200 can separate the slice from the multiplex layer.
For example, when TS is used, PES packetizing is performed on encoded data in units of slice segments. When the RTP is adopted, the encoded data is RTP packetized in units of slice segments. In these cases, as in MMT packet #1 shown in fig. 8, NAL units and slice segments arranged before the slice segments may be packetized separately.
When the TS is adopted, the transmission device 100 indicates a unit of data stored in a PES packet by using a data alignment descriptor or the like. Note that DASH is a system for downloading MP 4-format data units called "segments" by HTTP or the like, and therefore the transmission apparatus 100 does not packetize encoded data at the time of transmission. Therefore, the transmission device 100 may create a sub-sample in units of slice segments and store information indicating the storage location of the sub-sample in the header of the MP4 so that the reception device 200 can detect the slice segment in the multiplex layer under the MP 4.
The MMT packing of the sliced pieces is explained in detail below.
As shown in fig. 8, by packetizing encoded data, data commonly referred to in decoding of all slices in an access unit such as SPS and PPS is stored in an MMT packet # 1. In this case, the reception device 200 concatenates the payload data of the MMT packet #1 and the data of each slice, and outputs the resulting data to the decoding unit. In this way, the reception device 200 can easily generate input data to the decoding unit by concatenating the payloads of the plurality of MMT packets.
Fig. 10 is a diagram showing an example of generating input data to the decoding units 204A to 204D from the MMT packet shown in fig. 8. The inverse multiplexer 203 concatenates the payload data of the MMT packet #1 and the MMT packet #2 to generate data necessary for the decoder 204A to decode the slice 1. The inverse multiplexer 203 generates input data in the same manner for the decoders 204B to 204D. That is, the inverse multiplexer 203 concatenates the MMT packet #1 and the payload data of the MMT packet #3 to generate input data for the decoder 204B. The inverse multiplexer 203 concatenates the MMT packet #1 and the payload data of the MMT packet #4 to generate input data for the decoder 204C. The inverse multiplexer 203 concatenates the MMT packet #1 and the payload data of the MMT packet #5 to generate input data for the decoder 204D.
The inverse multiplexing unit 203 may remove NAL units necessary for non-decoding processing such as access unit delimiters and SEI from the payload data of the MMT packet #1, separate only NAL units of SPS and PPS necessary for decoding processing, and add the separated NAL units to the sliced data.
When the encoded data is packetized as shown in fig. 9, the inverse multiplexing unit 203 outputs an MMT packet #1 including the head data of the access unit in the multiplexing layer to the 1 st decoding unit 204A. The inverse multiplexing unit 203 analyzes the MMT packet including the header data of the access unit in the multiplexing layer, separates NAL units of the SPS and PPS, and adds the separated NAL units of the SPS and PPS to each data of the 2 nd and subsequent slices, thereby generating input data for each of the 2 nd and subsequent decoding units.
Further, it is preferable that the reception apparatus 200 be able to identify the type of data held in the MMT payload and the index number of the slice segment in the access unit when the slice segment is held in the payload, by using information contained in the header of the MMT packet. Here, the type of data refers to either data before a slice segment (NAL units arranged before the leading slice segment in an access unit are collectively referred to as this) or data of a slice segment. When a Unit obtained by slicing an MPU such as a slice segment is stored in an MMT packet, a mode for storing an MFU (Media Fragment Unit) is adopted. In the present mode, the transmission device 100 can set, for example, a Data unit (Data unit) which is a basic unit of Data in the MFU as a sample (a Data unit in the MMT, corresponding to an access unit) or a sub-sample (a unit obtained by dividing a sample).
At this time, the header of the MMT packet contains a field called "Fragmentation indicator" and a field called "Fragmentation counter".
The Fragmentation indicator indicates whether or not the Data stored in the payload of the MMT packet is Data obtained by fragmenting a Data unit, and when the Data is Data obtained by fragmenting, indicates that the fragment is the head or last fragment of the Data unit, or neither the head nor the last fragment. In other words, the Fragmentation indicator included in the header of a certain packet indicates (1) that only the packet is included in the Data unit as the basic Data unit; (2) dividing the Data unit into a plurality of packets for storage, wherein the packet is the head packet of the Data unit; (3) dividing the Data unit into a plurality of packets for storage, wherein the packets are packets except the head and the last of the Data unit; and (4) the Data unit is divided into a plurality of packets and stored, and the packet is the identification information of any one of 4 items of the last packet of the Data unit.
The Fragment counter is an index number indicating that the Data stored in the MMT packet is equivalent to the second Fragment in the Data unit.
Therefore, by setting the sample in the MMT as a Data unit by the transmission device 100 and setting the Data before the segment and each segment as a Data unit in a segment unit, the reception device 200 can identify the type of Data stored in the payload using information included in the header of the MMT packet. That is, the inverse multiplexer 203 can generate input data to the decoders 204A to 204D with reference to the header of the MMT packet.
Fig. 11 is a diagram showing an example when a sample is set as a Data unit and Data before a slice and the slice are packed as a slice of the Data unit.
The pre-slice segment data and the slice segments are divided into 5 slices of slice #1 to slice # 5. Each fragment is stored into an individual MMT packet. At this time, the values of the Fragmentation indicator and Fragmentation counter included in the header of the MMT packet are as shown in the figure.
For example, a Fragmentation identifier is a binary 2-bit value. The Fragmentation indicator at the head of the Data unit, i.e., the MMT packet #1, the Fragmentation indicator at the last MMT packet #5, and the Fragmentation indicators of the MMT packets #2 to #4, i.e., the packets therebetween, are set to different values, respectively. Specifically, the Fragmentation indicator of the MMT packet #1, which is the head of the Data unit, is set to 01, the Fragmentation indicator of the MMT packet #5, which is the last Data unit, is set to 11, and the Fragmentation indicators of the MMT packets #2 to #4, which are the packets therebetween, are set to 10. When the Data unit includes only 1 MMT packet, the Fragmentation indicator is set to 00.
The Fragment counter is 4, which is a value obtained by subtracting 1 from 5, which is the total number of fragments, in the MMT packet #1, and is sequentially decreased one by one in the subsequent packets, and is 0 in the last MMT packet # 5.
Therefore, the reception apparatus 200 can recognize the MMT packet storing the data before the segment by using either the Fragmentation indicator or the Fragmentation counter. Further, the reception apparatus 200 can recognize the MMT packet storing the nth Fragment by referring to the Fragment counter.
The header of the MMT packet additionally contains the sequence number of the Movie Fragment to which the Data unit belongs in the MPU, the sequence number of the MPU itself, and the sequence number of the sample to which the Data unit belongs in the Movie Fragment. The inverse multiplexer 203 can uniquely determine the sample to which the Data unit belongs by referring to these samples.
Furthermore, since the inverse multiplexing unit 203 can determine the index number of the Fragment in the Data unit from the Fragment counter or the like, even when a packet loss occurs, the Fragment stored in the Fragment can be uniquely specified. For example, even when the slice #4 shown in fig. 11 cannot be acquired due to a packet loss, the inverse multiplexing unit 203 knows that the slice received immediately after the slice #3 is the slice #5, and therefore can accurately output the slice segment 4 stored in the slice #5 to the decoding unit 204D, not the decoding unit 204C.
When a transmission path that ensures that no packet loss occurs is used, the inverse multiplexing unit 203 may periodically process the arriving packets without determining the type of data stored in the MMT packet or the index number of the slice with reference to the header of the MMT packet. For example, when an access unit is transmitted with 5 MMT packets in total of pre-slice data and 4 slice segments, the reception apparatus 200 can sequentially acquire pre-slice data and 4-slice data by determining pre-slice data of the access unit for which decoding is started and then sequentially processing the received MMT packets.
Next, a modification of the packing will be described.
The slice does not necessarily need to be divided in both the horizontal direction and the vertical direction within the plane of the access unit, and the access unit may be divided only in the horizontal direction or only in the vertical direction as shown in fig. 1.
Furthermore, when the access unit is divided only in the horizontal direction, tiles need not be employed.
The number of divisions in the plane of the access unit is arbitrary and is not limited to 4. However, the area size of the slice segment and the tile needs to be equal to or larger than the lower limit of the encoding standard such as h.265.
The transmission apparatus 100 may store identification information indicating the intra-plane division method of the access unit in the MMT message, the descriptor of the TS, or the like. For example, information indicating the number of divisions in the horizontal direction and the vertical direction in the plane may be stored. Alternatively, the unique identification information may be assigned to a dividing method such as dividing equally in the horizontal direction and the vertical direction 2 as shown in fig. 3, or dividing equally in the horizontal direction 4 as shown in fig. 1. For example, when the access unit is divided as shown in fig. 3, the identification information indicates pattern 2, and when the access unit is divided as shown in fig. 1, the identification information indicates pattern 1.
Further, information indicating restrictions on coding conditions associated with the in-plane division method may be included in the multiplex layer. For example, information indicating that 1 slice is composed of 1 tile may be used. Alternatively, information indicating that a reference block in the case of performing motion compensation at the time of decoding a slice or a tile is limited to a slice or a tile at the same position in a picture, or to a block in a predetermined range in an adjacent slice segment, or the like may be used.
Further, the transmission apparatus 100 may switch whether or not to divide the access unit into a plurality of slices according to the resolution of the moving image. For example, the transmission apparatus 100 may divide the access unit into 4 units when the moving image to be processed is 8K4K, and may not perform in-plane division when the moving image to be processed has a resolution of 4K 2K. By defining the division method in the case of the moving image of 8K4K in advance, the reception apparatus 200 can determine whether or not there is an in-plane division and division method by acquiring the resolution of the received moving image, and can switch the decoding operation.
The reception device 200 can detect the presence or absence of in-plane division by referring to the header of the MMT packet. For example, when the Data unit of MMT is set as a sample when the access unit is not divided, fragmentation of the Data unit is not performed. Therefore, the reception apparatus 200 can determine that an access unit is not divided when the value of the Fragment counter included in the header of the MMT packet is always zero. Alternatively, the reception apparatus 200 may detect whether or not the value of the Fragmentation indicator is always 01. The reception apparatus 200 can determine that an access unit is not divided even when the value of the Fragmentation indicator is always 01.
The receiving apparatus 200 can also cope with a case where the number of divisions in the plane of the access unit does not match the number of decoding units. For example, when the receiving apparatus 200 includes 2 decoding units 204A and 204B capable of decoding encoded data of 8K2K in real time, the inverse multiplexing unit 203 outputs 2 of 4 slices constituting encoded data of 8K4K to the decoding unit 204A.
Fig. 12 is a diagram showing an operation example when data subjected to MMT packetization as shown in fig. 8 is input to the 2 decoding units 204A and 204B. Here, it is preferable that the receiving apparatus 200 be able to directly combine and output the decoding results of the decoding units 204A and 204B. Therefore, the inverse multiplexer 203 selects the slice segment to be output to each of the decoders 204A and 204B so that the decoding results of the decoders 204A and 204B are spatially continuous.
The inverse multiplexing unit 203 may select a decoding unit to be used in accordance with the resolution, frame rate, or the like of the encoded data of the moving image. For example, when the reception apparatus 200 includes 4 decoding units of 4K2K, if the resolution of the input image is 8K4K, the reception apparatus 200 performs decoding processing using all 4 decoding units. Further, if the resolution of the input image is 4K2K, the reception apparatus 200 performs the decoding process with only 1 decoding section. Alternatively, even if the intra-plane division is 4, when 8K4K can be decoded in real time by a single decoding unit, the inverse multiplexing unit 203 merges all the division units and outputs the result to 1 decoding unit.
The receiving apparatus 200 may determine the decoding unit to be used in consideration of the frame rate. For example, there are cases where: when the receiving apparatus 200 includes 2 decoding units having an upper limit of 60fps of frame rate that can be decoded in real time at a resolution of 8K4K, 120fps of encoded data is input at 8K 4K. In this case, if the intra-plane is configured by 4 division units, slice 1 and slice 2 are input to the decoding unit 204A, and slice 3 and slice 4 are input to the decoding unit 204B, as in the example of fig. 12. Since each of the decoding units 204A and 204B can decode 120fps in real time as long as it is 8K2K (with a resolution of half of 8K 4K), the decoding processing can be performed by these 2 decoding units 204A and 204B.
Even if the resolution and the frame rate are the same, the amount of processing differs depending on the encoding system itself, such as the level (profile) or the level (level) in the encoding system, or h.264 or h.265. Therefore, the receiving apparatus 200 may select a decoding unit to be used based on the information. Further, when all the encoded data received through broadcasting or communication cannot be decoded, or when all the slices or tiles constituting the area selected by the user cannot be decoded, the reception apparatus 200 may automatically determine the slices or tiles that are decodable within the processing range of the decoding unit. Alternatively, the reception apparatus 200 may also provide a user interface for the user to select the decoded region. In this case, the reception apparatus 200 may display a warning message indicating that the entire area cannot be decoded, or may display information indicating the number of decodable areas, slices, or tiles.
The above method can also be applied to a case where MMT packets storing the same coded data in divided pieces are transmitted and received through a plurality of transmission paths such as broadcasting and communication.
The transmission device 100 may perform encoding so that the regions of the respective slices overlap each other in order to make the boundary of the division unit inconspicuous. In the example shown in FIG. 13, a picture of 8K4K is divided into 4 slices 1-4. For example, the cut fragments 1 to 3 are 8K × 1.1K, and the cut fragment 4 is 8K × 1K. Further, adjacent cut pieces overlap each other. In this way, motion compensation at the time of encoding can be efficiently performed at the boundary at the time of 4-division indicated by a dotted line, and therefore the image quality at the boundary portion can be improved. Thus, the image quality deterioration at the boundary portion is reduced.
In this case, the display unit 205 cuts out an 8K × 1K region from among the 8K × 1.1K regions, and combines the obtained regions. The transmission device 100 may encode the slice segments so as to be overlapped with each other and transmit the encoded slice segments separately while including information indicating the overlapping range in the multiplex layer or the encoded data.
In addition, the same method can be applied to the use of tiles.
The following describes a flow of operations of the transmission device 100. Fig. 14 is a flowchart showing an example of the operation of the transmission apparatus 100.
First, the encoding unit 101 divides a picture (access unit) into a plurality of slices (tiles) which are a plurality of areas (S101). Next, the encoding unit 101 encodes each of the plurality of slices so as to be independently decodable, thereby generating encoded data corresponding to each of the plurality of slices (S102). The encoding unit 101 may encode a plurality of slices by a single encoding unit, or may perform parallel processing by a plurality of encoding units.
Next, the multiplexing unit 102 stores the plurality of encoded data generated by the encoding unit 101 in a plurality of MMT packets, thereby multiplexing the plurality of encoded data (S103). Specifically, as shown in fig. 8 and 9, the multiplexer 102 stores a plurality of encoded data in a plurality of MMT packets so that the encoded data corresponding to different slice segments is not stored in 1 MMT packet. As shown in fig. 8, the multiplexer 102 stores control information that is used in common for all decoding units within a picture in an MMT packet #1 that is different from the MMT packets #2 to #5 in which a plurality of encoded data are stored. Here, the control information includes at least one of an access unit delimiter, an SPS, a PPS, and an SEI.
The multiplexer 102 may store the control information in the same MMT packet as any one of the plurality of MMT packets in which the plurality of encoded data are stored. For example, as shown in fig. 9, the multiplexer 102 may store the control information in the top MMT packet (MMT packet #1 in fig. 9) among a plurality of MMT packets in which a plurality of encoded data are stored.
Finally, the transmission apparatus 100 transmits a plurality of MMT packets. Specifically, the modulation unit 103 modulates the multiplexed data, and the transmission unit 104 transmits the modulated data (S104).
Fig. 15 is a block diagram showing an example of the configuration of the receiving apparatus 200, and is a diagram showing in detail the inverse multiplexing unit 203 shown in fig. 7 and the configuration of the subsequent stage thereof. As shown in fig. 15, the receiving apparatus 200 further includes a decode command unit 206. The inverse multiplexing unit 203 further includes a type discrimination unit 211, a control information acquisition unit 212, a slice information acquisition unit 213, and a decoded data generation unit 214.
The following describes the flow of the operation of the receiving apparatus 200. Fig. 16 is a flowchart showing an example of the operation of the receiving apparatus 200. Here, the operation for 1 access unit is shown. When the decoding processing of a plurality of access units is executed, the processing of the present flowchart is repeated.
First, the reception device 200 receives a plurality of packets (MMT packets) generated by the transmission device 100, for example (S201).
Next, the type discrimination unit 211 analyzes the header of the received packet to acquire the type of encoded data stored in the received packet (S202).
Next, the type discrimination unit 211 determines whether the data stored in the received packet is data before slicing or sliced based on the type of the acquired encoded data (S203).
When the data stored in the received packet is pre-sliced data (yes in S203), the control information acquisition unit 212 acquires pre-sliced data of the access unit to be processed from the payload of the received packet and stores the pre-sliced data in the memory (S204).
On the other hand, if the data stored in the received packet is fragmented data (no in S203), the receiving apparatus 200 determines whether or not the data stored in the received packet is encoded data of any one of the plurality of areas, using the header information of the received packet. Specifically, the slice information acquiring unit 213 analyzes the header of the received packet to acquire the index Idx of the slice stored in the received packet (S205). Specifically, the index number Idx is an index number within Movie Fragment of an access unit (sample in MMT).
The processing of step S205 may be performed in step S202.
Next, the decoded data generation unit 214 determines a decoding unit that decodes the slice segment (S206). Specifically, the index Idx is associated with a plurality of decoding units in advance, and the decoded data generator 214 determines the decoding unit corresponding to the index Idx acquired in step S205 as the decoding unit for decoding the slice segment.
As described in the example of fig. 12, the decoded data generation unit 214 may determine a decoding unit that decodes a slice segment based on at least one of the resolution of an access unit (picture), the method of dividing the access unit into a plurality of slices (tiles), and the processing capacity of a plurality of decoding units provided in the reception apparatus 200. For example, the decoded data generation unit 214 determines the access unit division method based on the identification information in the descriptor such as the MMT message and the section (section) of the TS.
Next, the decoded data generating unit 214 combines the control information, which is included in any one of the plurality of packets and is used in common for all the decoding units within the picture, with each of the plurality of pieces of encoded data of the plurality of slices, thereby generating a plurality of pieces of input data (combined data) to be input to the plurality of decoding units. Specifically, the decoded data generation unit 214 acquires fragmented data from the payload of the received packet. The decoded data generation unit 214 combines the data before the slice segment stored in the memory in step S204 with the acquired data of the slice segment, thereby generating input data to the decoding unit determined in step S206 (S207).
After step S204 or S207, if the data of the received packet is not the final data of the access unit (no in S208), the processing from step S201 onward is performed again. That is, the above-described processing is repeated until input data to the plurality of decoding units 204A to 204D corresponding to all the slices included in the access unit is generated.
The timing at which the packet is received is not limited to the timing shown in fig. 16, and a plurality of packets may be received in advance or sequentially and stored in a memory or the like.
On the other hand, if the data of the received packet is the final data of the access unit (yes in S208), the decode command unit 206 outputs the plurality of input data generated in step S207 to the corresponding decoding units 204A to 204D (S209).
Next, the plurality of decoding units 204A to 204D decode the plurality of input data in parallel according to the DTS of the access means, thereby generating a plurality of decoded images (S210).
Finally, the display unit 205 generates a display image by combining the plurality of decoded images generated by the plurality of decoding units 204A to 204D, and displays the display image according to the PTS of the access unit (S211).
The receiving apparatus 200 analyzes the payload data of the MMT packet storing header information of the MPU or header information of the Movie Fragment to acquire the DTS and PTS of the access unit. When the TS is used as the multiplexing scheme, the receiving apparatus 200 acquires the DTS and PTS of the access unit from the header of the PES packet. When the receiving apparatus 200 uses RTP as the multiplexing method, the DTS and PTS of the access unit are acquired from the header of the RTP packet.
When merging the decoding results of the plurality of decoding units, the display unit 205 may perform filtering processing such as deblocking filtering on the boundary between adjacent division units. Further, since the filtering process is not required when the decoding result of a single decoding unit is displayed, the display unit 205 may switch the process depending on whether or not the filtering process is performed on the boundary of the decoding results of a plurality of decoding units. Whether or not the filtering process is necessary may be predetermined in advance depending on the presence or absence of division. Alternatively, information indicating whether or not the filtering process is necessary may be separately stored in the multiplexing layer. In addition, information necessary for filtering processing such as filter coefficients is sometimes stored in SPS, PPS, SEI, or slice segments. The decoding units 204A to 204D or the inverse multiplexing unit 203 acquires these pieces of information by analyzing the SEI, and outputs the acquired pieces of information to the display unit 205. The display unit 205 performs filtering processing using these pieces of information. When these pieces of information are stored in the slice, the decoding units 204A to 204D preferably acquire these pieces of information.
In the above description, an example in which the types of data stored in a slice are 2 types, that is, pre-slice data and slice segment data, has been described, but the types of data may be 3 or more. In this case, discrimination according to the type is performed in step S203.
When the data size of the slice is large, the transmission device 100 may fragment the slice and store the fragment in the MMT packet. That is, the transmission apparatus 100 may slice the pre-slice data and the slice. In this case, if the access unit is set to be equal to the Data unit as in the packing example shown in fig. 11, the following problem occurs.
For example, when the slice segment 1 is divided into 3 slices, the slice segment 1 is divided into 3 packets of Fragment counter values 1 to 3 to transmit. In addition, after the Fragment segment 2, the Fragment counter value becomes 4 or more, and the correlation between the Fragment counter value and the data stored in the payload cannot be obtained. Therefore, the reception apparatus 200 cannot specify the packet storing the head data of the slice from the header information of the MMT packet.
In this case, the reception apparatus 200 may parse the data of the payload of the MMT packet and determine the start position of the slice. Here, as a format for storing NAL units in a multiplex layer, there are 2 types of formats called "byte stream format" in which a start code composed of a specific bit string is added immediately before a NAL unit header, and "NAL size format" in which a field indicating the size of a NAL unit is added.
The byte stream format is used in MPEG-2 systems, RTP, and the like. NAL size format is utilized in MP4, DASH using MP4, MMT, and the like.
When the byte stream format is used, the reception apparatus 200 analyzes whether or not the head data of the packet matches the start code. If the start data of a packet matches the start code, the receiving apparatus 200 can detect whether or not the data included in the packet is fragmented data by acquiring the type of NAL unit from the NAL unit header immediately following the NAL unit header.
On the other hand, in the NAL size format, the reception apparatus 200 cannot detect the start position of the NAL unit based on the bit string. Therefore, in order to acquire the start position of the NAL unit, the reception device 200 needs to sequentially read data corresponding to the size of the NAL unit from the leading NAL unit of the access unit and shift the pointer.
However, if the header of the MPU or Movie Fragment of the MMT indicates the size of a unit of a sub-sample and the sub-sample corresponds to the pre-slicing data or the slice segment, the reception apparatus 200 can determine the start position of each NAL unit based on the size information of the sub-sample. Therefore, the transmission apparatus 100 may include information indicating whether or not the information in units of subsamples exists in the MPU or the Movie Fragment in information acquired by the reception apparatus 200 at the start of data reception, such as the MPT in the MMT.
The data of the MPU is data expanded based on the MP4 format. MP4 has a mode in which parameter sets such as SPS and PPS of h.264 or h.265 can be stored as sample data and a mode in which they cannot be stored. Further, information for determining the mode is represented as an entry name of SampleEntry (sample entry). When the parameter set is included in the sample using the mode that can be saved, the reception apparatus 200 acquires the parameter set by the method described above.
On the other hand, when the mode that cannot be stored is used, the parameter set is stored as Decoder Specific Information (Decoder characteristic Information) in the SampleEntry or is stored using a stream for the parameter set. Here, since a stream for parameter sets is not generally used, it is preferable that the transmission apparatus 100 stores parameter sets in the Decoder Specific Information. In this case, the receiving apparatus 200 analyzes the SampleEntry transmitted as the metadata of the MPU or the metadata of the Movie Fragment in the MMT packet, and acquires the parameter set referred to by the access unit.
When saving a parameter set as sample data, the reception apparatus 200 can acquire a parameter set necessary for decoding with reference to only the sample data without referring to SampleEntry. At this time, the transmission apparatus 100 may not save the parameter set in SampleEntry. In this way, since the same SampleEntry can be used by the transmission device 100 in different MPUs, the processing load of the transmission device 100 at the time of MPU creation can be reduced. Further, there is an advantage that the reception apparatus 200 does not need to refer to the parameter set in SampleEntry.
Alternatively, the transmission apparatus 100 may store 1 default parameter set in SampleEntry and store the parameter set referred to by the access unit in the sample data. In the conventional MP4, since a parameter set is generally stored in SampleEntry, there is a possibility that a receiving device may stop reproduction when no parameter set exists in SampleEntry. This problem can be solved by using the above method.
Alternatively, the transmission device 100 may save the parameter set in the sample data only when the parameter set different from the default parameter set is used.
In addition, since both modes can save parameter sets into the SampleEntry, the transmitting device 100 can also always save parameter sets into the visual SampleEntry, from which the receiving device 200 always obtains parameter sets.
In the MMT standard, header information of MP4 such as Moov and Moof is transmitted as MPU metadata or movie clip metadata, but the transmission device 100 may not necessarily transmit MPU metadata and movie clip metadata. The receiving apparatus 200 may determine whether or not SPS and PPS are stored in the sample data based on services of ARIB (Association of Radio Industries and Businesses) standards, a type of resource, or whether or not MPU element transmission is performed.
Fig. 17 is a diagram showing an example of the case where Data units are set differently for Data before a slice and for each slice.
In the example shown in fig. 17, the data sizes of the pre-slicing segment data and the slicing segments 1 to 4 are Length #1 to Length #5, respectively. The respective field values of the Fragmentation indicator, Fragmentation counter, and Offset included in the header of the MMT packet are as shown in the figure.
Here, Offset is Offset information indicating a bit length (Offset) from the beginning of encoded data of a sample (access unit or picture) to which payload data belongs to the beginning byte of payload data (encoded data) included in the MMT packet. The value of Fragment counter is described as starting from a value obtained by subtracting 1 from the total number of fragments, but may start from another value.
Fig. 18 is a diagram showing an example of slicing a Data unit. In the example shown in fig. 18, the slice 1 is divided into 3 slices and stored in the MMT packet #2 to the MMT packet #4, respectively. At this time, the data sizes of the respective slices are set to Length #2_1 to Length #2_3, respectively, and the values of the respective fields are as shown in the figure.
As described above, when a Data unit such as a slice is set, the start of the access unit and the start of the slice can be determined as follows based on the field value of the MMT header.
The beginning of the payload in the packet whose Offset value is 0 is the beginning of the access unit.
The value of Offset is a value different from 0, and the beginning of the payload of a packet whose Fragmentation indicator has a value of 00 or 01 is the beginning of a slice.
When neither fragmentation of a Data unit nor packet loss occurs, the receiving apparatus 200 can determine the index number of the fragment stored in the MMT packet based on the number of fragment segments acquired after the start of the access unit is detected.
In addition, when the Data unit of the Data before the clip is fragmented, the reception apparatus 200 can detect the access unit and the beginning of the clip in the same manner.
When a packet loss occurs or when the SPS, PPS, and SEI included in Data before a slice are set to different Data units, the reception apparatus 200 can also determine the start position of a slice or a tile in a picture (access unit) by determining an MMT packet that stores the head Data of the slice based on the analysis result of the MMT header and then analyzing the header of the slice. The amount of processing involved in the analysis of slice headers is small, and the processing load is not a problem.
As described above, each of the encoded Data of the plurality of slices is associated with a basic Data unit (Data unit), which is a unit of Data stored in 1 or more packets, in a one-to-one manner. Further, each of the plurality of encoded data is stored in 1 or more MMT packets.
The header information of each MMT packet includes Fragmentation indicator (identification information) and Offset (Offset information).
The reception apparatus 200 determines the start of payload data included in a packet having header information including a Fragmentation indicator having a value of 00 or 01 as the start of encoded data of each slice. Specifically, the receiving apparatus 200 determines the start of payload data included in a packet having header information including Offset having a value other than 0 and a Fragmentation indicator having a value of 00 or 01 as the start of encoded data of each slice.
In the example of fig. 17, the head of the Data unit is either the head of the access unit or the head of the slice, and the value of the Fragmentation indicator is 00 or 01. The receiving apparatus 200 can also determine which of the access unit delimiter and the slice the head of the Data unit is by referring to the type of NAL unit, and can detect the head of the access unit or the head of the slice without referring to Offset.
As described above, when the transmitting apparatus 100 packetizes the NAL unit so that the start of the NAL unit must start from the start of the payload of the MMT packet, and thus also includes a case where the Data before the slice is divided into a plurality of Data units, the receiving apparatus 200 can detect the start of the access unit or the slice by analyzing the Fragmentation indicator and the NAL unit header. The type of NAL unit exists in the first byte of the NAL unit header. Therefore, the receiving apparatus 200 can acquire the type of NAL unit by additionally analyzing 1 byte data when analyzing the MMT packet header.
In the case of audio, the receiving apparatus 200 may detect the beginning of the access unit, and may determine whether the value of the Fragmentation indicator is 00 or 01.
As described above, when the encoded data encoded so as to be divisionally decodable is stored in the PES packet of the MPEG-2TS, the transmission device 100 can use the data alignment descriptor. Hereinafter, an example of a method of storing encoded data in a PES packet will be described in detail.
For example, in HEVC, the transmission device 100 can indicate which of an access unit, a slice segment, and a tile is data held in a PES packet by using a data alignment descriptor. The type of alignment in HEVC is specified as follows.
The type of alignment 8 represents a slice of HEVC. The type of alignment 9 represents a slice or access unit of HEVC. The type of alignment 12 represents a slice or tile of HEVC.
Therefore, the transmission device 100 can indicate that the data of the PES packet is either the slice or the pre-slice data by using, for example, type 9. Since a type indicating that the slice is not a slice is separately specified, the transmission device 100 may use a type indicating that the slice is not a slice.
The DTS and PTS included in the header of the PES packet are set only in the PES packet including the leading data of the access unit. Therefore, if the type is 9 and a field of DTS or PTS exists in the PES packet, the receiving apparatus 200 can determine that the entire access unit or the leading partition unit in the access unit is stored in the PES packet.
The transmitting apparatus 100 may use a field such as transport _ priority indicating the priority of a TS packet storing a PES packet containing the head data of an access unit so that the receiving apparatus 200 can distinguish the data included in the packet. The receiving apparatus 200 may determine the data included in the PES packet by analyzing whether or not the payload of the PES packet is an access unit delimiter. In addition, the data _ alignment _ indicator of the PES packet header indicates whether or not data is stored in the PES packet in these types. As long as the flag (data _ alignment _ indicator) is set to 1, the data held in the PES packet is guaranteed to be of the type indicated by the data alignment descriptor.
The transmission device 100 may use the data alignment descriptor only when PES packetizing is performed in units of divisible decoding such as slices. Accordingly, the receiving device 200 can determine that encoded data is PES-packetized in divisionally decodable units when the data alignment descriptor exists, and can determine that encoded data is PES-packetized in access unit units when the data alignment descriptor does not exist. In addition, when the data _ alignment _ indicator is set to 1 and no data alignment descriptor exists, the unit of PES packetization is specified as an access unit in the MPEG-2TS standard.
If the data alignment descriptor is included in the PMT, the receiving device 200 can determine that the PES packet is packetized in the divisible decoding unit, and can generate input data to each decoding unit based on the packetized unit. When the PMT does not include a data alignment descriptor and it is determined that parallel decoding of encoded data is necessary based on program information or information of another descriptor, the receiving apparatus 200 generates input data to each decoding unit by analyzing a slice header of a slice. When the encoded data can be decoded by the single decoding unit, the receiving apparatus 200 decodes the data of the entire access unit by the decoding unit. When information indicating whether or not the encoded data is formed of divisible and decodable units such as slices and tiles is separately indicated by a descriptor of the PMT, the receiving apparatus 200 may determine whether or not the encoded data can be decoded in parallel based on the analysis result of the descriptor.
Since the DTS and PTS included in the header of the PES packet are set only in the PES packet including the leading data of the access unit, when the divided access units perform PES packetization, the 2 nd and subsequent PES packets do not include information indicating the DTS and PTS of the access unit. Therefore, when performing decoding processing in parallel, the decoding units 204A to 204D and the display unit 205 use the DTS and PTS stored in the header of the PES packet including the leading data of the access unit.
(embodiment mode 2)
In embodiment 2, a method of storing data in NAL size format in an MPU based on MP4 format in MMT will be described. In addition, although a storage method for the MPU used for MMT is described below as an example, such a storage method can also be applied to DASH based on the MP4 format.
[ method of storing into MPU ]
In the MP4 format, a plurality of access units are collectively saved in 1 MP4 file. The MPU for MMT is able to save data for each media, including any number of access units, into 1 MP4 file. Since the MPU is a unit that can decode independently, for example, an access unit in units of GOPs is stored in the MPU.
Fig. 19 is a diagram showing a configuration of an MPU. The MPU starts with ftyp, mmpu, and moov, which are collectively defined as MPU metadata. The moov stores initialization information and MMT track (hint track) common to files.
In the moof, initialization information and size of each sample or sub-sample, information (sample _ duration, sample _ size, sample _ composition _ time _ offset) that can specify Presentation Time (PTS) and Decoding Time (DTS), data _ offset indicating the position of data, and the like are stored.
Further, the plurality of access units are respectively saved as samples in mdat (mdat box). Data other than the sample among moof and mdat is defined as movie fragment metadata (hereinafter referred to as MF metadata), and sample data of mdat is defined as media data.
Fig. 20 is a diagram showing a structure of MF metadata. As shown in fig. 20, the MF metadata is more specifically composed of type (type), length (length), and data (data) of the moof box (moof), and type (type) and length (length) of the mdat box (mdat).
When storing an access unit in MP4 data, there are a mode in which parameter sets such as SPS and PPS of h.264 or h.265 can be stored as sample data and a mode in which they cannot be stored.
Here, in the above-described mode that cannot be saved, parameter sets are saved to the Decoder Specific Information of SampleEntry of moov. In the above-described mode that can be stored, the parameter set is included in the sample.
MPU metadata, MF metadata, and media data are stored in the MMT payload, and a slice type (FT) is stored in the header of the MMT payload as an identifier that can identify these data. FT-0 denotes MPU metadata, FT-1 denotes MF metadata, and FT-2 denotes media data.
In addition, although fig. 19 shows an example in which MPU metadata units and MF metadata units are stored as Data units in the MMT payload, units such as ftyp, mmpu, moov, and moof may be stored as Data units in the MMT payload. Likewise, in fig. 19, an example in which sample units are saved as Data units into the MMT payload is illustrated. Alternatively, a Data unit may be configured in units of samples or NAL units, and such a Data unit may be stored in units of Data units in the MMT payload. Such Data unit may be further stored in units of fragmentation in the MMT payload.
[ conventional Transmission method and problems ]
Conventionally, when a plurality of access units are packaged in MP4 format, moov and moof are created at the time when all samples stored in MP4 are ready.
When the MP4 format is transmitted in real time by broadcasting or the like, for example, if the samples stored in 1 MP4 file are GOP units, moov and moof are created after storing time samples of GOP units, and therefore, a delay associated with encapsulation occurs. Due to such encapsulation on the transmitting side, the end-to-end delay always extends the GOP unit amount of time. Accordingly, it is difficult to provide services in real time, and in particular, when live content is transmitted, deterioration of services for viewers is caused.
Fig. 21 is a diagram for explaining a data transmission sequence. When MMT is applied to broadcasting, as shown in fig. 21 (a), if MMT packets are placed in the MMT packet in the order of the constituent MPUs and transmitted (MMT packets #1, #2, #3, #4, #5 are transmitted in the order of the constituent MPUs), a delay due to encapsulation occurs in the transmission of MMT packets.
In order to prevent this delay due to encapsulation, a method is proposed in which MPU header information such as MPU metadata and MF metadata is not transmitted (packets #1 and #2 are not transmitted, and packets #3 to #5 are transmitted in this order), as shown in fig. 21 (b). Further, the following method is considered: as shown in fig. 20 (c), the media data is transmitted without waiting for the MPU header information to be created, and the MPU header information is transmitted after the media data is transmitted (transmitted in the order of #3 to #5, #1, and # 2).
When the MPU header information is not transmitted, the receiving apparatus does not decode using the MPU header information, and when the MPU header information is transmitted after the comparison with the media data, the receiving apparatus waits for the acquisition of the MPU header information and then decodes.
However, in the conventional reception device conforming to MP4, decoding without MPU header information cannot be guaranteed. Further, when the receiving apparatus performs decoding by a special process without using the MPU header, the decoding process becomes complicated by the conventional transmission rule, and it is highly likely that real-time decoding becomes difficult. When the receiving apparatus waits for the acquisition of MPU header information and then performs decoding, the buffering of media data is necessary until the receiving apparatus acquires the header information, but the buffer model is not defined and decoding cannot be guaranteed.
Then, as shown in fig. 20 (d), the transmission device according to embodiment 2 stores only the common information in the MPU metadata, thereby transmitting the MPU metadata prior to the media data. The transmission device according to embodiment 2 transmits the MF metadata, which is delayed in generation, to the media data. Accordingly, a transmission method or a reception method capable of ensuring the decoding of media data is provided.
A receiving method when each of the transmission methods in fig. 21 (a) to (d) is used will be described below.
In each transmission method shown in fig. 21, MPU metadata, MF metadata, and media data are first structured in the order of MPU metadata, MF metadata, and media data.
After the MPU data is configured, when the transmitting apparatus transmits data in the order of MPU metadata, MF metadata, and media data as shown in (a) of fig. 21, the receiving apparatus can decode by any one of the following methods (a-1) and (a-2).
(A-1) the reception device acquires MPU header information (MPU metadata and MF metadata) and then decodes the media data using the MPU header information.
(A-2) the reception apparatus decodes the media data without using MPU header information.
Such methods all have the advantage that although a delay due to encapsulation occurs on the transmitting side, there is no need to buffer media data in order to acquire an MPU header in the receiving device. When buffering is not performed, a memory for buffering does not need to be mounted, and buffering delay does not occur. The method (a-1) is also applicable to a conventional receiving apparatus because it decodes using MPU header information.
When the transmitting apparatus transmits only media data as shown in fig. 21 (B), the receiving apparatus can perform decoding by the method (B-1) described below.
(B-1) the receiving apparatus decodes the media data without using MPU header information.
Note that, although not shown, when MPU metadata is transmitted prior to transmission of the media data in fig. 21 (B), decoding can be performed by the following method (B-2).
(B-2) the receiving apparatus decodes the media data using the MPU metadata.
Both of the methods (B-1) and (B-2) have advantages that no delay due to encapsulation occurs on the transmission side and that no buffering of media data is required for acquiring an MPU header. However, since the methods (B-1) and (B-2) do not decode the header information using the MPU, there is a possibility that special processing is required for the decoding.
When the transmitting apparatus transmits data in the order of media data, MPU metadata, and MF metadata as shown in fig. 21 (C), the receiving apparatus can decode by any one of the following methods (C-1) and (C-2).
(C-1) the reception device decodes the media data after acquiring MPU header information (MPU metadata and MF metadata).
(C-2) the receiving apparatus decodes the media data without using MPU header information.
When the method (C-1) is used, the media data needs to be buffered in order to acquire MPU header information. In contrast, when the method (C-2) is used, there is no need to perform buffering for acquiring MPU header information.
In addition, neither of the methods (C-1) and (C-2) described above causes a delay due to encapsulation on the transmission side. Further, since the method of (C-2) does not use MPU header information, there is a possibility that special processing is required.
When the transmitting apparatus transmits data in the order of MPU metadata, media data, and MF metadata as shown in fig. 21 (D), the receiving apparatus can decode by any one of the following methods (D-1) and (D-2).
(D-1) the receiving apparatus acquires the MF metadata after acquiring the MPU metadata, and then decodes the media data.
(D-2) the reception device decodes the media data without using the MF metadata after acquiring the MPU metadata.
In the case of the method (D-1), media data needs to be buffered in order to acquire MF metadata, and in the case of the method (D-2), buffering for acquiring MF metadata is not required.
Since the method of (D-2) does not perform decoding using MF metadata, there is a possibility that special processing is required.
As described above, there is an advantage that decoding is possible even in the conventional MP4 receiving apparatus when decoding is possible using MPU metadata and MF metadata.
In fig. 21, MPU data is structured in the order of MPU metadata, MF metadata, and media data, and in moof, position information (offset) for each sample or sub-sample is determined based on the structure. The MF metadata also includes data (size or type of box) other than the media data in the mdat box.
Therefore, when the receiving apparatus determines the media data based on the MF metadata, the receiving apparatus reconstructs the data in the order when the MPU data is configured regardless of the order in which the data is transmitted, and then decodes the data using moov of the MPU metadata or moof of the MF metadata.
In fig. 21, the MPU data is configured in the order of MPU metadata, MF metadata, and media data, but the MPU data may be configured in a different order from fig. 21 to determine the position information (offset).
For example, MPU data may be structured in the order of MPU metadata, media data, and MF metadata, and negative position information (offset) may be indicated in MF metadata. In this case, regardless of the order of transmitting data, the receiving apparatus reconstructs data in the order in which the MPU data was configured on the transmitting side, and then decodes the data using moov or moof.
Further, the transmitting apparatus may transmit, as signaling, information indicating an order in which the MPU data is configured, and the receiving apparatus may reconstruct the data based on the information transmitted as signaling.
As described above, the receiving apparatus receives the packaged MPU metadata, the packaged media data (sample data), and the packaged MF metadata in this order as shown in fig. 21 (d). Here, MPU metadata is an example of the first metadata, and MF metadata is an example of the second metadata.
Next, the receiving device reconstructs MPU data (file in MP4 format) including the received MPU metadata, the received MF metadata, and the received sample data. Then, the sample data included in the reconstructed MPU data is decoded using the MPU metadata and the MF metadata. The MF metadata is metadata including data (for example, length stored in mbox) that can be generated only after sample data generation on the transmitting side.
More specifically, the operation of the receiving apparatus is performed by each component constituting the receiving apparatus. For example, the receiving apparatus includes a receiving unit that receives the data, a reconstruction unit that reconstructs the MPU data, and a decoding unit that decodes the MPU data. The receiving unit, the generating unit, and the decoding unit are each realized by a microcomputer, a processor, a dedicated circuit, or the like.
[ method of decoding without using header information ]
Next, a method of decoding without using header information will be described. Here, a method of decoding without using header information in a receiving apparatus regardless of whether or not header information is transmitted in a transmitting side will be described. That is, this method can be applied to any of the transmission methods described with reference to fig. 21. However, some decoding methods are applicable only to a specific transmission method.
Fig. 22 is a diagram showing an example of a method for decoding without using header information. In fig. 22, only the MMT payload and MMT packet containing only media data are illustrated, and the MMT payload and MMT packet containing MPU metadata or MF metadata are not illustrated. In the following description of fig. 22, it is assumed that media data belonging to the same MPU are continuously transmitted. Note that, although the case where a sample is stored in the payload as media data is described as an example, it is needless to say that in the following description of fig. 22, NAL units may be stored, or fragmented NAL units may be stored.
In order to decode media data, a receiving device first has to acquire initialization information required for decoding. In addition, if the medium is a video, the receiving apparatus must acquire initialization information for each sample, or specify the start position of the MPU, which is a random access unit, and acquire the start positions of the sample and the NAL unit. In addition, the receiving apparatus needs to determine the Decoding Time (DTS) and Presentation Time (PTS) of the sample, respectively.
Thus, the receiving apparatus can perform decoding without using header information, for example, by using the following method. In addition, when NAL unit units or units obtained by slicing NAL units are stored in the payload, the following description may be given with "sample" replaced with "sample NAL unit".
< random Access (═ initial samples of a specific MPU) >
When the header information is not transmitted, the receiving apparatus has the following methods 1 and 2 in order to specify the top sample of the MPU. In addition, method 3 can be used when header information is transmitted.
Method 1 the receiving device takes samples contained in an MMT packet whose RAP _ flag is 1' in the MMT packet header.
Method 2 the receiving device takes samples of 'sample number 0' in the MMT payload header.
[ method 3] when at least either one of MPU metadata and MF metadata is transmitted before and after media data, a receiving device acquires a sample included in an MMT payload in which a slice type (FT) in an MMT payload header has been switched to the media data.
In methods 1 and 2, if a plurality of samples belonging to different MPUs are mixed in 1 payload, it is impossible to determine which NAL unit is a random access point (RAP _ flag is 1 or sample number is 0). Therefore, it is necessary to limit the samples of different MPUs to not coexist in 1 payload, or to limit the RAP _ flag to 1 when the last (or first) sample is a random access point when the samples of different MPUs coexist in 1 payload.
In order to acquire the start position of the NAL unit by the receiving apparatus, it is necessary to sequentially shift the read pointer of the data by the size amount of the NAL unit from the leading NAL unit of the sample.
When Data is fragmented, the receiving apparatus can specify a Data unit by referring to the fragment _ indicator or the fragment _ number.
< decision of DTS of sample >
The DTS of the sample is determined by methods 1 and 2.
[ method 1] A reception device determines the DTS of a start sample based on a prediction structure. However, since this method requires analysis of encoded data and may be difficult to decode in real time, the following method 2 is preferable.
[ method 2] separately transmitting the DTS of the start sample, and the reception device acquires the DTS of the transmitted start sample. As a method of transmitting the DTS of the top sample, for example, a method of transmitting the DTS of the MPU top sample using the MMT-SI, a method of transmitting the DTS of each sample using the MMT header extension area, or the like is available. The DTS may be an absolute value or a relative value with respect to the PTS. In addition, the transmitting side may transmit, as a signaling, whether or not the DTS of the start sample is included.
In addition, in both method 1 and method 2, the DTS of the subsequent samples is calculated as a fixed frame rate.
As a method of storing the DTS of each sample in the packet header, there is a method of storing the DTS of the sample included in the MMT packet in the 32-bit NTP timestamp field in the MMT packet header, in addition to using the extension region. When the DTS cannot be expressed by the number of bits of 1 packet header (32 bits), the DTS may be expressed by a plurality of packet headers, or the DTS may be expressed by combining the NTP timestamp field of the packet header and the extension field. When DTS information is not included, it is regarded as a known value (e.g., ALL 0).
< PTS decision of sample >
The receiving apparatus obtains the PTS of the top sample from the MPU time stamp descriptor of each asset included in the MPU. The receiving apparatus calculates the subsequent sample PTS as a fixed frame rate using a parameter indicating the display order of the samples such as POC. In this way, in order to calculate the DTS and PTS without using header information, it is necessary to perform transmission based on a fixed frame rate.
When the MF metadata is transmitted, the receiving apparatus can calculate the absolute values of the DTS and PTS from the relative time information of the DTS or PTS with respect to the header sample indicated by the MF metadata and the absolute value of the time stamp of the MPU header sample indicated by the MPU time stamp descriptor.
When the DTS and PTS are calculated by analyzing the encoded data, the receiving apparatus may calculate the DTS and PTS by using SEI information included in the access unit.
< initialization information (parameter set) >
[ case of video ]
In the case of video, parameter sets are saved into the sample data. In addition, when MPU metadata and MF metadata are not transmitted, it is ensured that a parameter set necessary for decoding can be acquired by simply referring to sample data.
As shown in fig. 21 (a) and (d), when MPU metadata is transmitted prior to media data, it may be defined that a parameter set is not stored in SampleEntry. In this case, the receiving device does not refer to the parameter set of SampleEntry but refers to only the parameter set within the sample.
When MPU metadata is transmitted prior to media data, a parameter set common to MPUs or a default parameter set may be stored in SampleEntry, and the receiving apparatus may refer to the parameter set of SampleEntry and the parameter set in a sample. By storing the parameter set in the SampleEntry, a conventional receiving apparatus that cannot reproduce the parameter set if the parameter set is not present in the SampleEntry can also perform decoding.
[ case of Audio ]
In the case of audio, the LATM header is required for decoding, and in MP4, the LATM header must be included in the Sample Entry (Sample Entry). However, when header information is not transmitted, it is difficult for the receiving apparatus to acquire the LATM header, and therefore the LATM header is separately included in control information such as SI. Additionally, the LATM header may also be incorporated into a message, table, or descriptor. In addition, the LATM head may be included in the sample.
The receiving apparatus acquires the LATM header from the SI or the like before decoding starts, and starts decoding of the audio. Alternatively, as shown in fig. 21 (a) and 21 (d), when transmitting MPU metadata prior to media data, the receiving apparatus can receive the LATM header prior to the media data. Therefore, even when MPU metadata is transmitted prior to media data, decoding can be performed by a conventional receiving apparatus.
< Others >
The transmission order or the type of transmission order may also be notified as control information for MMT headers or payload headers, or MPTs or other tables, messages, descriptors, etc. The type of transmission sequence here is, for example, 4 types of transmission sequences shown in fig. 21 (a) to (d), and identifiers for identifying the types may be stored in positions that can be obtained before decoding starts.
The type of the transmission sequence may be different between audio and video, or may be common between audio and video. Specifically, for example, audio may be transmitted in the order of MPU metadata, MF metadata, and media data as shown in fig. 21 (a), and video may be transmitted in the order of MPU metadata, media data, and MF metadata as shown in fig. 21 (d).
With the above method, the receiving apparatus can perform decoding without using header information. Further, even the conventional receiving apparatus can perform decoding when MPU metadata (fig. 21 (a) and 21 (d)) is transmitted prior to media data.
In particular, by transmitting the MF metadata after the media data ((d) of fig. 21), it is possible to decode even a conventional receiving apparatus without delay due to encapsulation.
[ constitution and operation of Transmission device ]
Next, the configuration and operation of the transmission device will be described. Fig. 23 is a block diagram of a transmission apparatus according to embodiment 2, and fig. 24 is a flowchart of a transmission method according to embodiment 2.
As shown in fig. 23, the transmission device 15 includes an encoding unit 16, a multiplexing unit 17, and a transmission unit 18.
The encoding unit 16 encodes the video or audio to be encoded, for example, in accordance with h.265, to generate encoded data (S10).
The multiplexing unit 17 multiplexes (packetizes) the encoded data generated by the encoding unit 16 (S11). Specifically, the multiplexing unit 17 packages sample data, MPU metadata, and MF metadata constituting a file in the MP4 format. The sample data is data obtained by encoding a video signal or an audio signal, the MPU metadata is an example of the first metadata, and the MF metadata is an example of the second metadata. The first metadata and the second metadata are both metadata used for decoding sample data, but the difference is that the second metadata includes data that can be generated only after the sample data is generated.
Here, the data that can be generated only after the generation of the sample data is, for example, data other than the sample data (data in the header of mdat, i.e., type and length shown in fig. 20) stored in mdat in the MP4 format. Here, the second metadata may include length as at least a part of the data.
The transmitter 18 transmits the packaged MP4 format file (S12). The transmission unit 18 transmits the MP4 format file by the method shown in fig. 21 (d), for example. That is, the packaged MPU metadata, the packaged sample data, and the packaged MF metadata are transmitted in this order.
The encoding unit 16, the multiplexing unit 17, and the transmitting unit 18 are each realized by a microcomputer, a processor, a dedicated circuit, or the like.
[ constitution of receiving apparatus ]
Next, the configuration and operation of the receiving apparatus will be described. Fig. 25 is a block diagram of a receiving apparatus according to embodiment 2.
As shown in fig. 25, the reception device 20 includes a packet filtering unit 21, a transmission order type determining unit 22, a random access unit 23, a control information acquiring unit 24, a data acquiring unit 25, a PTS/DTS calculating unit 26, an initialization information acquiring unit 27, a decode command unit 28, a decoding unit 29, and a presentation unit 30.
[ action 1 of the receiving apparatus ]
First, an operation of the reception device 20 for specifying the MPU start position and NAL unit position when the medium is a video will be described. Fig. 26 is a flowchart of such an operation of the receiving apparatus 20. Here, it is assumed that the transmission order type of MPU data is stored in the SI information by the transmission device 15 (multiplexing unit 17).
First, the packet filtering unit 21 performs packet filtering on a received file. The transmission order type determination unit 22 analyzes the packet-filtered SI information and acquires the transmission order type of MPU data (S21).
Next, the transmission order type determination unit 22 determines (determines) whether or not the MPU header information (at least one of the MPU metadata and the MF metadata) is included in the packet-filtered data (S22). When the MPU header information is included (yes in S22), the random access unit 23 detects switching of the slice type of the MMT payload header to the media data, and specifies an MPU start sample (S23).
On the other hand, when the MPU header information is not included (no in S22), the random access unit 23 specifies an MPU start sample based on the RAP _ flag of the MMT header or the sample number of the MMT payload header (S24).
The transmission order type determination unit 22 determines whether or not the MF metadata is included in the packet-filtered data (S25). When determining that the MF metadata is included (yes at S25), the data acquisition unit 25 acquires a NAL unit by reading the NAL unit based on the sample included in the MF metadata, the offset of the sub-sample, and the size information (S26). On the other hand, when determining that the MF metadata is not included (no in S25), the data acquisition unit 25 acquires NAL units by sequentially reading data of the size of the NAL units from the leading NAL unit of the sample (S27).
When the receiving apparatus 20 determines in step S22 that the MPU header information is included, the MPU start sample may be determined by the process of step S24 instead of step S23. When it is determined that the MPU header information is included, the process of step S23 and the process of step S24 may be used in combination.
If the reception device 20 determines in step S25 that MF metadata is included, it may acquire a NAL unit by the processing of step S27 without the processing of step S26. When it is determined that the MF metadata is included, the process of step S26 and the process of step S27 may be used in combination.
It is assumed that the MF metadata is included in the media data and the MF data is transmitted later in step S25. In this case, the receiving apparatus 20 may buffer the media data, wait until the MF metadata is acquired, and then perform the process of step S26, or the receiving apparatus 20 may determine whether or not to perform the process of step S27 without waiting for the MF metadata to be acquired.
For example, the receiving apparatus 20 may determine whether or not to wait for the MF metadata to be acquired based on whether or not a buffer having a buffer size capable of buffering the media data is held. The reception device 20 may determine whether or not to wait for the MF metadata to be acquired based on whether or not the end-to-end delay is small. The receiving apparatus 20 may perform the decoding process mainly by the process of step S26, and may use the process of step S27 when a process pattern such as a packet loss occurs.
When the transmission order type is determined in advance, step S22 and step S26 may be omitted, and in this case, the reception device 20 may determine a method of determining the MPU start sample and a method of determining the NAL unit, taking into account the buffer size and the end-to-end delay.
When the transmission sequence type is known in advance, the transmission sequence type determination unit 22 is not necessary in the reception device 20.
Further, although not described in fig. 26, the decode command unit 28 outputs the data acquired by the data acquisition unit to the decoding unit 29 based on the PTS and DTS calculated by the PTS/DTS calculation unit 26 and the initialization information acquired by the initialization information acquisition unit 27. The decoding unit 29 decodes the data, and the presentation unit 30 presents the decoded data.
[ action 2 of the receiving apparatus ]
Next, an operation in which the reception device 20 acquires initialization information based on the transmission order type and decodes media data based on the initialization information will be described. Fig. 27 is a flowchart of such an operation.
First, the packet filtering unit 21 performs packet filtering on a received file. The transmission sequence type determination unit 22 analyzes the packet-filtered SI information to acquire a transmission sequence type (S301).
Next, the transmission order type determination unit 22 determines whether MPU metadata is transmitted (S302). When it is determined that the MPU metadata is transmitted (yes in S302), the transmission order type determination unit 22 determines whether the MPU metadata is transmitted prior to the media data based on the result of the analysis in step S301 (S303). When the MPU metadata is transmitted prior to the media data (yes in S303), the initialization information acquisition unit 27 decodes the media data based on the common initialization information included in the MPU metadata and the initialization information of the sample data (S304).
On the other hand, when it is determined that the MPU metadata is transmitted after the transmission of the media data (no in S303), the data acquisition unit 25 buffers the media data until the MPU metadata is acquired (S305), and performs the process of step S304 after the MPU metadata is acquired.
When it is determined in step S302 that MPU metadata is not transmitted (no in S302), the initialization information acquisition unit 27 decodes the media data based only on the initialization information of the sample data (S306).
Only when the transmission side has guaranteed decoding of the media data based on the initialization information of the sample data, the process of step S306 is used without performing the process based on the determinations of step S302 and step S303.
The reception device 20 may determine whether or not to buffer the media data before step S305. In this case, the reception device 20 transitions to the process of step S305 when determining to buffer the media data, and transitions to the process of step S306 when determining not to buffer the media data. The determination of whether or not to buffer media data may be made based on the buffer size and the occupancy of the receiving apparatus 20, or may be made in consideration of the end-to-end delay, for example, by selecting a small end-to-end delay.
[ action 3 of the receiving apparatus ]
Here, details of a transmission method or a reception method when MF metadata is transmitted later than media data (fig. 21 (c) and fig. 21 (d)) will be described. Hereinafter, the case of fig. 21 (d) will be described as an example. Note that, only the method of (d) in fig. 21 is used for transmission, and the signaling of the transmission order type is not performed.
As described above, as shown in fig. 21 (d), when data is transmitted in the order of MPU metadata, media data, and MF metadata, the following 2 decoding methods can be performed:
(D-1) the reception device 20 acquires the MF metadata after acquiring the MPU metadata, and then decodes the media data.
(D-2) the reception device 20 decodes the media data without using the MF metadata after acquiring the MPU metadata.
Here, in D-1, buffering of media data for obtaining MF metadata is required, but decoding can be performed by a receiving apparatus conforming to MP4 in the related art because decoding can be performed using MPU header information. In D-2, although it is not necessary to buffer media data for obtaining MF metadata, decoding cannot be performed using MF metadata, and therefore special processing is required for decoding.
Further, in the method of fig. 21 (d), MF metadata is transmitted after media data, and thus there is an advantage that delay due to encapsulation can be prevented, thereby reducing end-to-end delay.
The receiving apparatus 20 can select the above 2 decoding methods according to the capability of the receiving apparatus 20 and the quality of service provided by the receiving apparatus 20.
The transmitting apparatus 15 must ensure that the decoding operation of the receiving apparatus 20 can reduce the occurrence of buffer overflow or underflow and perform decoding. As an element for defining a decoder model when decoding is performed by the method of D-1, for example, the following parameters can be used.
Buffer size for reconstruction of MPU (MPU buffer)
For example, for the buffer size ═ maximum rate × maximum MPU time × α, maximum rate (rate) means the rank of encoded data, the upper limit rate of the rank + overhead of the MPU header. The maximum MPU time is the maximum time length of a GOP when 1MPU is equal to 1GOP (video).
Here, the audio may be in units of GOP common to the video or in units of other units. α is a margin for not causing overflow, and may be multiplied by the maximum rate × the maximum MPU time or added. When multiplied, alpha is more than or equal to 1, and when added, alpha is more than or equal to 0.
The upper limit of the decode delay time from the input of data to the MPU buffer until the decoding is performed. (TSTD _ delay in STD of MPEG-TS)
For example, in transmission, the DTS is set so that the acquisition completion time of MPU data of the receiver is not more than the DTS, taking into account the upper limit values of the maximum MPU time and the decoding delay time.
The transmitter 15 may also assign a DTS and a PTS in accordance with a decoder model when decoding is performed by the method of D-1. Accordingly, the transmitting device 15 can transmit auxiliary information necessary for decoding by the D-2 method while ensuring that the receiving device performing decoding by the D-1 method performs the decoding.
For example, the transmitting device 15 can guarantee the operation of the receiving device that performs decoding by the D-2 method by transmitting, as signaling, the pre-buffering time of the decoder buffer when performing decoding by the D-2 method.
The pre-buffering time may be included in SI control information such as a message, a table, a descriptor, etc., or may be included in a header of the MMT packet or the MMT payload. Furthermore, the SEI within the encoded data may also be covered. The DTS and PTS for decoding by the D-1 method may be stored in the MPU time stamp descriptor and the SampleEntry, and the DTS and PTS or the pre-buffering time for decoding by the D-2 method may be described in the SEI.
The reception device 20 may select the decoding method D-1 when the reception device 20 corresponds to only the decoding operation conforming to MP4 using the MPU header, and may select either one of the methods when the reception device 20 corresponds to both of D-1 and D-2.
The transmitter 15 may assign a DTS and a PTS to ensure a decoding operation of one side (D-1 in the present description), and may transmit auxiliary information for assisting the decoding operation of the other side.
Further, comparing the case of the method using D-2 with the case of the method using D-1, there is a high possibility that the end-to-end delay becomes large due to the delay due to the pre-buffering of the MF metadata. Therefore, the receiving apparatus 20 may select the method of D-2 for decoding when it is desired to reduce the end-to-end delay. For example, the receiving device 20 may always utilize the D-2 approach when it is always desirable to reduce the end-to-end delay. The method of D-2 may be used only when the receiving apparatus 20 operates in a low-latency presentation mode in which presentation is performed with low latency, such as live content, channel selection, and channel switching (zapping).
Fig. 28 is a flow chart of such a receiving method.
First, the reception device 20 receives the MMT packet and acquires MPU data (S401). Then, the receiving apparatus 20 (transmission order type discriminating unit 22) determines whether or not to present the program in the low-delay presentation mode (S402).
When the program is not presented in the low-latency presentation mode (no in S402), the receiving apparatus 20 (the random access unit 23 and the initialization information acquisition unit 27) acquires random access and initialization information using header information (S405). The receiving apparatus 20(PTS/DTS calculating unit 26, decode instructing unit 28, decoding unit 29, presentation unit 30) performs decoding and presentation processing based on the PTS and DTS provided by the transmitting side (S406).
On the other hand, when presenting the program in the low-latency presentation mode (yes in S402), the receiving apparatus 20 (the random access unit 23 and the initialization information acquisition unit 27) acquires random access and initialization information by a decoding method that does not use header information (S403). The receiving apparatus 20 performs decoding and presentation processing based on auxiliary information for decoding without using PTS, DTS, and header information provided by the transmitting side (S404). In step S403 and step S404, the processing may be performed using MPU metadata.
[ transmitting/receiving method using auxiliary data ]
The above describes the transmission/reception operation when MF metadata is transmitted later than media data ((c) of fig. 21 and (d) of fig. 21). Next, a method will be described in which the transmitting device 15 transmits auxiliary data having a function of part of the MF metadata, so that decoding can be started earlier and end-to-end delay can be reduced. Here, an example in which the auxiliary data is further transmitted based on the transmission method shown in fig. 21 (d) will be described, but the method using the auxiliary data can also be applied to the transmission methods shown in fig. 21 (a) to (c).
Fig. 29 (a) is a diagram showing an MMT packet transmitted by the method shown in fig. 21 (d). That is, data is transmitted in the order of MPU metadata, media data, and MF metadata.
Here, sample #1, sample #2, sample #3, and sample #4 are samples included in the media data. Although the media data is stored in the MMT packet in sample units, the media data may be stored in the MMT packet in NAL unit units or in units obtained by dividing NAL units. In addition, sometimes multiple NAL units are aggregated and stored into MMT packets.
As described in D-1 above, in the method shown in (D) of fig. 21, that is, when data is transmitted in the order of MPU metadata, media data, and MF metadata, there is a method of acquiring the MPU metadata, then acquiring the MF metadata, and then decoding the media data. In the method of D-1, although it is necessary to buffer media data for obtaining MF metadata, the method of D-1 has an advantage that it can be applied to a receiving apparatus conforming to MP4 in the related art because decoding is performed using MPU header information. On the other hand, there is a disadvantage that the reception apparatus 20 must wait until the MF metadata is acquired to restart decoding.
In contrast, as shown in fig. 29 (b), in the method using auxiliary data, auxiliary data is transmitted prior to MF metadata.
The MF metadata includes information indicating the DTS, PTS, offset, and size of all samples included in the movie fragment. On the other hand, the auxiliary data includes information indicating the DTS, PTS, offset, and size of some of the samples included in the movie fragment.
For example, the MF metadata includes information of all samples (sample #1 to sample #4), whereas the auxiliary data includes information of a part of samples (sample #1 to sample # 2).
In the case shown in fig. 29 (b), since the samples #1 and #2 can be decoded by using the auxiliary data, the End-to-End delay becomes smaller than that of the D-1 transmission method. In addition, the auxiliary data may be included by combining the information of the samples, or the auxiliary data may be repeatedly transmitted.
For example, in fig. 29 (c), the transmission device 15 includes the information of sample #1 in the auxiliary information when the auxiliary information is transmitted at the timing of a, and includes the information of sample #1 and sample #2 in the auxiliary information when the auxiliary information is transmitted at the timing of B. When the transmission device 15 transmits the auxiliary information at the timing of C, the auxiliary information includes information of sample #1, sample #2, and sample # 3.
The MF metadata includes information of sample #1, sample #2, sample #3, and sample #4 (information of all samples in the movie fragment).
The assistance data does not necessarily need to be sent immediately after generation.
In addition, in the header of the MMT packet or MMT payload, a type indicating that auxiliary data is stored is specified.
For example, when auxiliary data is saved in the MMT payload using the MPU mode, a data type indicating that it is auxiliary data is specified as a fragment _ type field value (for example, FT ═ 3). The auxiliary data may be data based on the moof configuration, or may have another configuration.
When the auxiliary data is saved as a control signal (descriptor, table, message) in the MMT payload, a descriptor tag, table ID, message ID, and the like indicating that the auxiliary data is specified.
In addition, PTS or DTS may be stored in the header of the MMT packet or MMT payload.
[ example of generating auxiliary data ]
An example in which the transmitting apparatus generates the auxiliary data based on the structure of moof will be described below. Fig. 30 is a diagram for explaining an example in which the transmission apparatus generates the auxiliary data based on the structure of moof.
In a general MP4, moof is produced for movie clips as shown in fig. 20. The moof includes information indicating the DTS, PTS, offset, and size of samples included in the movie fragment.
Here, the transmission device 15 constructs an MP4(MP4 file) using only a part of the sample data constituting the MPU, and generates auxiliary data.
For example, as shown in fig. 30 (a), the transmission device 15 generates MP4 using only sample #1 among samples #1 to #4 constituting the MPU, with the head of moof + mdat as auxiliary data.
Next, as shown in fig. 30 (b), the transmission device 15 generates MP4 using the sample #1 and the sample #2 among the samples #1 to #4 constituting the MPU, with the head of moof + mdat as the next auxiliary data.
Next, as shown in fig. 30 (c), the transmission device 15 generates MP4 using the sample #1, the sample #2, and the sample #3 among the samples #1 to #4 constituting the MPU, with the head of moof + mdat as the next auxiliary data.
Next, as shown in fig. 30 (d), the transmission device 15 generates all MPs 4 in samples #1 to #4 constituting the MPU, with the head of moof + mdat being movie fragment metadata.
In addition, although the transmission device 15 generates the auxiliary data for each sample, the auxiliary data may be generated for each N samples. The value of N is an arbitrary number, and for example, when transmitting auxiliary data M times when transmitting 1 MPU, N may be equal to full sample/M.
The information indicating the shift of the sample in moof may be a shift value after ensuring that the sample inlet region for the subsequent number of samples is a NULL region.
In addition, the auxiliary data may be generated so as to be a structure of slicing the MF metadata.
[ example of receiving operation using auxiliary data ]
Reception of the assistance data generated as described in fig. 30 will be described. Fig. 31 is a diagram for explaining reception of assistance data. In fig. 31 (a), the number of samples constituting the MPU is 30, and the auxiliary data is generated and transmitted for each 10 samples.
In fig. 30 (a), the auxiliary data #1 includes sample information of samples #1 to #10, the auxiliary data #2 includes sample information of samples #1 to #20, and the MF metadata includes sample information of samples #1 to # 30.
Samples #1 to #10, samples #11 to #20, and samples #21 to #30 are stored in 1 MMT payload, but may be stored in sample units or NAL units, or may be stored in units of fragmentation or aggregation.
The reception device 20 receives packets of MPU elements, samples, MF elements, and auxiliary data, respectively.
The receiving device 20 concatenates the sample data in the order of reception (backward), and updates the current auxiliary data after receiving the latest auxiliary data. In addition, the receiving apparatus 20 can constitute a complete MPU by replacing the auxiliary data with the MF metadata at the end.
When receiving the auxiliary data #1, the receiving device 20 concatenates the data as in the upper stage of fig. 31 (b), thereby configuring the MP 4. Accordingly, the receiving apparatus 20 can analyze the samples #1 to #10 using the MPU metadata and the information of the auxiliary data #1, and can perform decoding based on the information of the PTS, DTS, offset, and size included in the auxiliary data.
When receiving the auxiliary data #2, the receiving apparatus 20 concatenates the data as in the middle of fig. 31 (b), thereby configuring the MP 4. Accordingly, the receiving apparatus 20 can analyze the samples #1 to #20 using the MPU metadata and the information of the auxiliary data #2, and can perform decoding based on the information of PTS, DTS, offset, and size included in the auxiliary data.
When receiving the MF metadata, the receiving apparatus 20 concatenates the data as in the lower stage of fig. 31 (b), thereby configuring the MP 4. Accordingly, the receiving apparatus 20 can analyze the samples #1 to #30 using the MPU metadata and the MF metadata, and can decode based on information of PTS, DTS, offset, and size included in the MF metadata.
When there is no auxiliary data, the reception device 20 can start acquiring information of the sample only after receiving the MF metadata, and therefore, it is necessary to start decoding after receiving the MF metadata. However, since the transmission device 15 generates and transmits the auxiliary data, the reception device 20 can acquire the information of the sample using the auxiliary data without waiting for the reception of the MF metadata, and thus can advance the decoding start time. Note that, by the transmitter 15 generating the moof-based auxiliary data described with reference to fig. 30, the receiver 20 can directly analyze the auxiliary data by the analyzer (parser) of the conventional MP 4.
Further, the newly generated auxiliary data or MF metadata contains information of a sample that overlaps with the auxiliary data transmitted in the past. Therefore, even when past auxiliary data cannot be acquired due to packet loss or the like, the MP4 can be reconstructed by using the newly acquired auxiliary data or MF metadata, and the information (PTS, DTS, size, and offset) of the sample can be acquired.
In addition, the auxiliary data does not necessarily need to include information of past sample data. For example, auxiliary data #1 may correspond to sample data #1- #10, and auxiliary data #2 may correspond to sample data #11- # 20. For example, as shown in fig. 31 (c), the transmitting apparatus 15 may sequentially transmit the full MF metadata as a Data unit and the unit fragmented from the Data unit as auxiliary Data.
In order to cope with packet loss, the transmission device 15 may repeatedly transmit the auxiliary data or the MF metadata.
The MMT packet and MMT payload storing the auxiliary data include an MPU serial number and a resource ID, as in the case of MPU metadata, MF metadata, and sample data.
The above-described reception operation using the auxiliary data will be described with reference to the flowchart of fig. 32. Fig. 32 is a flowchart of the receiving operation using the auxiliary data.
First, the receiving apparatus 20 receives the MMT packet and parses the packet header or the payload header (S501). Next, the reception device 20 analyzes whether the slice type is auxiliary data or MF metadata (S502), and updates the auxiliary data in the past if the slice type is auxiliary data (S503). At this time, if there is no past auxiliary data of the same MPU, the reception device 20 directly sets the received auxiliary data as new auxiliary data. The receiving device 20 then acquires and decodes a sample based on the MPU metadata, the auxiliary data, and the sample data (S504).
On the other hand, if the slice type is MF metadata, the receiving apparatus 20 overwrites the past auxiliary data with MF metadata in step S505 (S505). The receiving device 20 then acquires and decodes a sample in the form of a complete MPU based on the MPU metadata, MF metadata, and sample data (S506).
Although not shown in fig. 32, in step S502, the receiving apparatus 20 stores data in the buffer when the fragmentation type is MPU metadata, and stores data concatenated backward for each sample in the buffer when the fragmentation type is sample data.
When the auxiliary data cannot be acquired due to a packet loss, the reception device 20 can decode the sample by overwriting the latest auxiliary data or by using the past auxiliary data.
The transmission cycle and the number of times of transmission of the auxiliary data may be predetermined values. Information on the sending period or number (count, countdown) may be sent together with the data. For example, the Data unit header may store a time stamp such as a delivery cycle, a delivery count, and initial _ cpb _ removal _ delay.
The auxiliary data including information of the first sample of the MPU is transmitted 1 or more times before initial _ CPB _ removal _ delay, and thus the CPB buffer model can be satisfied. At this time, in the MPU timestamp descriptor, a value based on picture timing SEI is saved.
The transmission method using such an auxiliary data reception operation is not limited to the MMT method, and can be applied to a case where a packet having an ISOBMFF file format is streamed, such as MPEG-DASH.
[ Transmission method when 1 MPU is constituted by a plurality of movie fragments ]
In the above description of fig. 19 and the following, 1 MPU is configured by 1 movie segment, and here, a case where 1 MPU is configured by a plurality of movie segments is described. Fig. 33 is a diagram showing a configuration of an MPU configured by a plurality of movie fragments.
In fig. 33, samples (#1- #6) stored in 1 MPU are stored in 2 movie segments. The first movie fragment is generated based on samples #1- #3, generating the corresponding moof box. A second movie fragment is generated based on samples #4- #6, generating the corresponding moof box.
The headers of moof box and mdat box in the first movie fragment are saved as movie fragment metadata #1 into MMT payload and MMT package. On the other hand, the headers of the moof box and the mdat box in the second movie fragment are saved as movie fragment metadata #2 in the MMT payload and the MMT package. In fig. 33, the MMT payload storing movie fragment metadata is shaded.
The number of samples constituting the MPU and the number of samples constituting the movie clip are arbitrary. For example, 2 movie slices may be configured by setting the number of samples constituting the MPU to the number of GOP-unit samples and setting one-half of the number of GOP-unit samples as movie slices.
Note that, although an example is shown in which 2 movie fragments (moof box and mdat box) are included in 1 MPU, 3 or more movie fragments included in 1 MPU may be included instead of 2. The samples stored in the movie fragment may be divided into an arbitrary number of samples, instead of the number of equally divided samples.
In fig. 33, MPU metadata units and MF metadata units are stored as Data units in the MMT payload, respectively. However, the transmitting apparatus 15 may store Data units such as ftyp, mmpu, moov, and moof in the MMT payload in units of Data units, or may store Data units in the MMT payload in units of Data units. The transmission device 15 may store the MMT payload in units of aggregated Data units.
Further, in fig. 33, samples are held in MMT payload in sample units. However, the transmitting apparatus 15 may configure a Data unit not in units of samples but in units of NAL units or units obtained by combining a plurality of NAL units, and store the Data unit in the MMT payload in units of Data units. The transmitting apparatus 15 may store the Data units in units of fragments in the MMT payload, or may store the Data units in units of aggregates in the MMT payload.
In fig. 33, the MPU is configured by moof #1, mdat #1, moof #2, and mdat #2 in this order, and offset is given to moof #1 as mdat #1 to which the corresponding values are added. However, the offset may be assigned as mdat #1 before moof # 1. However, in this case, movie fragment metadata cannot be generated in the form of moof + mdat, and the headers of moof and mdat are separately transmitted.
Next, a transmission procedure of the MMT packet when transmitting the MPU having the configuration described in fig. 33 will be described. Fig. 34 is a diagram for explaining the transmission sequence of MMT packets.
Fig. 34 (a) shows a transmission sequence when the MMT packet is transmitted in the order of the constituent MPUs shown in fig. 33. Specifically, (a) of fig. 34 shows an example in which MPU elements, MF elements #1, media data #1 (samples #1 to #3), MF elements #2, and media data #2 (samples #4 to #6) are transmitted in this order.
Fig. 34 (b) shows an example in which the MPU element, media data #1 (samples #1 to #3), MF element #1, media data #2 (samples #4 to #6), and MF element #2 are transmitted in this order.
Fig. 34 (c) shows an example in which media data #1 (samples #1 to #3), MPU elements, MF elements #1, media data #2 (samples #4 to #6), and MF elements #2 are transmitted in this order.
MF primitive #1 is generated using samples #1- #3, and MF primitive #2 is generated using samples #4- # 6. Therefore, when the transmission method of fig. 34 (a) is used, a delay due to encapsulation occurs in transmission of sample data.
In contrast, when the transmission methods of fig. 34 (b) and 34 (c) are used, samples can be transmitted without waiting for the generation of an MF element, and thus the end-to-end delay can be reduced without causing a delay due to encapsulation.
In the transmission sequence in fig. 34 (a), since 1 MPU is divided into a plurality of movie fragments, and the number of samples stored in the MF element is reduced compared to the case of fig. 19, the delay amount due to packaging can be reduced compared to the case of fig. 19.
In addition to the method described here, for example, the transmission device 15 may connect the MF element #1 and MF element #2 and transmit them together at the end of the MPU. In this case, the MF elements of different movie fragments may be aggregated and stored in 1 MMT payload. Alternatively, the MF elements of different MPUs may be aggregated and stored in the MMT payload.
[ receiving method when 1 MPU is constituted by a plurality of movie fragments ]
Here, an operation example of the reception device 20 that receives and decodes the MMT packet transmitted in the transmission order described in fig. 34 (b) will be described. Fig. 35 and 36 are diagrams for explaining such an operation example.
The reception device 20 receives the MMT packet including the MPU element, the sample, and the MF element, which is transmitted in the transmission order shown in fig. 35, respectively. The sample data is concatenated in the order of reception.
The receiving apparatus 20 concatenates the data at T1, which is the time when the MF element #1 is received, as shown in (1) of fig. 36, and configures an MP 4. Accordingly, the receiving apparatus 20 can acquire samples #1 to #3 based on the MPU metadata and the information of the MF element #1, and can decode based on the information of PTS, DTS, offset, and size included in the MF element.
Further, the reception device 20 concatenates the data as shown in (2) of fig. 36 at T2, which is the time when the MF element #2 is received, and configures the MP 4. Accordingly, the receiving apparatus 20 can acquire samples #4 to #6 based on the MPU metadata and the information of the MF element #2, and can decode based on the information of the PTS, DTS, offset, and size of the MF element. The reception device 20 may also acquire samples #1 to #6 based on the information of MF metadata #1 and MF metadata #2 by concatenating data to form MP4 as shown in fig. 36 (3).
By dividing 1 MPU into a plurality of movie fragments, the time until the first MF element among the MPUs is acquired is shortened, and thus the decoding start time can be advanced. Further, the buffer size for accumulating samples before decoding can be reduced.
The transmission device 15 may set the division unit of the movie fragment so that the time from the transmission (or reception) of the first sample in the movie fragment to the transmission (or reception) of the MF element corresponding to the movie fragment is shorter than the time of initial _ cpb _ removal _ delay specified by the encoder. By setting in this manner, the reception buffer can conform to the cpb buffer, and low-delay decoding can be achieved. In this case, the absolute time based on initial _ cpb _ remove _ delay can be used for PTS and DTS.
The transmission device 15 may divide movie fragments at equal intervals or may divide subsequent movie fragments at intervals shorter than the previous movie fragment. Accordingly, the receiving apparatus 20 can receive the MF element including the information of the sample before the sample is decoded without fail, and can perform continuous decoding.
The following 2 methods can be used to calculate the absolute time of PTS and DTS.
(1) The absolute time of the PTS and DTS is determined based on the reception time (T1 or T2) of MF element #1 or MF element #2 and the relative time of the PTS and DTS included in the MF element.
(2) The absolute time of the PTS and DTS is determined based on the absolute time of the signaling transmitted from the transmitting side such as an MPU timestamp descriptor, and the relative time of the PTS and DTS included in the MF element.
The absolute time at which the (2-a) transmission device 15 transmits the signaling may be an absolute time calculated based on initial _ cpb _ removal _ delay specified by the encoder.
The absolute time at which the (2-B) transmission device 15 transmits the signaling may be an absolute time calculated based on a predicted value of the reception time of the MF element.
Further, MF #1 and MF #2 may be repeatedly transmitted. By repeatedly transmitting MF element #1 and MF element #2, the receiving apparatus 20 can acquire an MF element again even when the MF element cannot be acquired due to a packet loss or the like.
An identifier indicating the order of movie fragments can be saved in the payload header of the MFU containing the samples constituting the movie fragment. On the other hand, an identifier indicating the order of MF elements constituting a movie fragment is not included in the MMT payload. Accordingly, the reception apparatus 20 recognizes the order of MF elements through packet _ sequence _ number. Alternatively, the transmitting apparatus 15 may store the identifier indicating that the MF element belongs to the first movie fragment in the control information (message, table, descriptor), MMT header, MMT payload header, or Data unit header as signaling.
The transmitting device 15 may transmit the MPU elements, MF elements, and samples in a predetermined transmission order determined in advance, and the receiving device 20 may perform the receiving process based on the predetermined transmission order determined in advance. The transmitting apparatus 15 may transmit the transmission order as signaling, and the receiving apparatus 20 may select (determine) the reception process based on the signaling information.
The reception method described above will be described with reference to fig. 37. Fig. 37 is a flowchart illustrating the operation of the receiving method described with reference to fig. 35 and 36.
First, the reception device 20 determines (identifies) whether the data included in the payload is MPU metadata, MF metadata, or sample data (MFU) based on the slice type indicated by the MMT payload (S601, S602). When the data is sample data, the receiving apparatus 20 buffers the sample, and waits for the MF metadata corresponding to the sample to be received and decoding to be started (S603).
On the other hand, if the data is MF metadata in step S602, the reception device 20 acquires sample information (PTS, DTS, position information, and size) from the MF metadata, acquires a sample based on the acquired sample information, and decodes and presents the sample based on the PTS and DTS (S604).
In addition, although not shown, if the data is MPU metadata, the MPU metadata includes initialization information necessary for decoding. Therefore, the reception device 20 accumulates the initialization information and uses it for decoding sample data in step S604.
When the receiving apparatus 20 stores the received data (MPU metadata, MF metadata, and sample data) of the MPUs in the storing apparatus, the data is rearranged into the configuration of the MPUs described with reference to fig. 19 or 33 and then stored.
In the transmitting side, the MMT packet is given a packet sequence number to a packet having the same packet ID. In this case, the MMT packet including MPU metadata, MF metadata, and sample data may be rearranged in the transmission order and then given a packet sequence number, or may be given a packet sequence number in the order before rearrangement.
When the packet sequence numbers are given in the order before the rearrangement, the receiving apparatus 20 can rearrange the data into the constituent order of the MPUs based on the packet sequence numbers, and accumulation becomes easy.
[ method of detecting the beginning of an Access Unit and the beginning of a slice ]
A method of detecting the start of an access unit or the start of a slice segment based on information of an MMT header and an MMT payload header will be described.
Here, 2 examples of the case where non-VCL NAL units (access unit delimiter, VPS, SPS, PPS, SEI, and the like) are collectively stored as Data units in the MMT payload, and the case where non-VCL NAL units are stored as Data units in 1 MMT payload by aggregating the Data units.
Fig. 38 is a diagram showing a case where non-VCL NAL units are individually aggregated as Data units.
In the case of fig. 38, the beginning of an access unit is an MMT packet whose fragment _ type value is MFU, and is the beginning Data of an MMT payload containing Data units whose aggregation _ flag value is 1 and offset value is 0. At this time, the Fragmentation _ indicator value is 0.
In the case of fig. 38, the start of a slice segment is an MMT packet whose fragment _ type value is MFU, and is the start data of an MMT payload whose aggregation _ flag value is 0 and whose fragmentation _ indicator value is 00 or 01.
Fig. 39 is a diagram showing a case where non-VCL NAL units are grouped as Data units. In addition, the field value of the packet header is as shown in fig. 17 (or fig. 18).
In the case of fig. 39, with respect to the head of an access unit, the head data of the payload in a packet whose Offset value is 0 is the head of the access unit.
In the case of fig. 39, regarding the beginning of the clip, the beginning data of the payload of the packet whose Offset value is a value other than 0 and whose Fragmentation indicator value is 00 or 01 is the beginning of the clip.
[ receiving Process when packet loss occurs ]
In general, when data in the MP4 format is transmitted in an environment where packet loss occurs, the reception device 20 recovers packets by means of Application Layer FEC (Application Layer FEC) or packet retransmission control.
However, when packet loss occurs when AL-FEC is not used in a stream such as broadcast, the packet cannot be recovered.
The receiving apparatus 20 needs to restart decoding of video and audio after data is lost due to packet loss. For this reason, the receiving apparatus 20 needs to detect the beginning of an access unit or NAL unit and start decoding from the beginning of the access unit or NAL unit.
However, since the start of the NAL unit of MP4 format does not include a start code, the reception apparatus 20 cannot detect the start of an access unit or NAL unit even when parsing the stream.
Fig. 40 is a flowchart of the operation of the receiving apparatus 20 when a packet loss occurs.
The reception apparatus 20 detects packet loss by using a packet sequence number, a packet counter, or a Fragment counter of a header of the MMT packet or the MMT payload (S701), and determines which packet is lost according to a context (S702).
When the reception device 20 determines that packet loss has not occurred (no in S702), it constructs an MP4 file and decodes an access unit or NAL unit (S703).
When determining that packet loss has occurred (yes in S702), the reception device 20 generates NAL units corresponding to the NAL units for which packet loss occurs from dummy (dummy) data, and constructs an MP4 file (S704). The receiving apparatus 20 indicates the type of the NAL unit as dummy data when the dummy data is added to the NAL unit.
The receiving apparatus 20 detects the head of the next access unit or NAL unit by the method described with reference to fig. 17, 18, 38, and 39, inputs the detected head data to the decoder from the head data, and can restart decoding (S705).
When a packet loss occurs, the receiving apparatus 20 may restart decoding from the beginning of the access unit and the NAL unit based on the information detected from the packet header, or may restart decoding from the beginning of the access unit and the NAL unit based on the header information of the reconstructed MP4 file including the NAL unit of the dummy data.
When the reception device 20 accumulates the MP4 file (MPU), packet data (NAL unit or the like) lost due to packet loss may be acquired separately by broadcasting or communication and accumulated (replaced).
At this time, when the reception device 20 acquires a lost packet through communication, it notifies the server of information (packet ID, MPU serial number, packet serial number, IP stream number, IP address, and the like) of the lost packet, and acquires the packet. The reception device 20 may acquire packet groups before and after the lost packet at the same time, not limited to the lost packet.
[ method of composing movie fragments ]
Here, the method of forming a movie fragment will be described in detail.
As explained with reference to fig. 33, the number of samples constituting a movie fragment and the number of movie fragments constituting 1 MPU are arbitrary. For example, the number of samples constituting a movie fragment and the number of movie fragments constituting 1 MPU may be determined to be fixed to predetermined numbers or may be determined dynamically.
Here, by configuring movie fragments so as to satisfy the following conditions on the transmitting side (transmitting device 15), it is possible to ensure low-delay decoding by the receiving device 20.
The conditions are as follows.
The transmitting apparatus 15 generates and transmits MF elements by dividing sample data into units as movie fragments so that the receiving apparatus 20 can receive MF elements including information of an arbitrary sample (i)) before the decoding time (dts (i)) of the sample is determined.
Specifically, the transmitting device 15 constructs a movie fragment using samples (including the i-th sample) that have been encoded prior to the dts (i).
In order to ensure low-delay decoding, the number of samples constituting a movie fragment or the number of movie fragments constituting 1 MPU is dynamically determined, for example, by the following method.
(1) When decoding is started, the decoding time DTS (0) of Sample (0) at the beginning of a GOP is a time based on initial _ cpb _ removal _ delay. The transmitting apparatus constructs the first movie fragment using the already encoded samples at a timing prior to DTS (0). The transmitting device 15 generates MF metadata corresponding to the first movie fragment and transmits the MF metadata at a timing prior to DTS (0).
(2) The transmission device 15 constructs movie fragments so as to satisfy the above conditions in the subsequent samples.
For example, when the sample at the beginning of the movie fragment is the kth sample, the MF element of the movie fragment including the kth sample is transmitted before the decoding time dts (k) of the kth sample. The transmitting device 15 uses the kth sample to the I-th sample to form the movie fragment when the encoding completion time of the I-th sample precedes dts (k) and the encoding completion time of the (I +1) -th sample succeeds dts (k).
In addition, the transmitting device 15 may also use the kth sample to less than the ith sample to form a movie fragment.
(3) After the last sample of the MPU is encoded, the transmitting device 15 constructs a movie fragment from the remaining samples, and generates and transmits MF metadata corresponding to the movie fragment.
The transmission device 15 may constitute the movie fragment by using a part of the samples already coded, instead of constituting the movie fragment by using all the samples already coded.
In addition, the above description shows an example in which the number of samples constituting a movie fragment and the number of movie fragments constituting 1 MPU are dynamically determined based on the above conditions in order to guarantee low-latency decoding. However, the method of determining the number of samples and the number of movie clips is not limited to this method. For example, the number of movie clips constituting 1 MPU may be fixed to a predetermined value, and the number of samples may be determined so as to satisfy the above conditions. The number of movie fragments constituting 1 MPU and the time at which the movie fragments are divided (or the code amount of the movie fragments) may be fixed to predetermined values, and the number of samples may be determined so as to satisfy the above conditions.
In addition, when the MPU is divided into a plurality of movie fragments, information indicating whether or not the MPU is divided into a plurality of movie fragments, an attribute of the divided movie fragments, or an attribute of an MF element corresponding to the divided movie fragments may be transmitted.
Here, the attribute of a movie fragment is information indicating whether the movie fragment is the first movie fragment of the MPU, the last movie fragment of the MPU, or other movie fragments.
The attribute of the MF element is information indicating whether the MF element is an MF element corresponding to the first movie fragment of the MPU, an MF element corresponding to the last movie fragment of the MPU, an MF element corresponding to another movie fragment, or the like.
The transmitting device 15 may store and transmit the number of samples constituting the movie fragment and the number of movie fragments constituting 1 MPU as control information.
[ operation of the receiving apparatus ]
The operation of the receiving apparatus 20 based on the movie fragment configured as described above will be described.
The receiving apparatus 20 determines the absolute time of each of the PTS and DTS based on the absolute time transmitted as a signaling from the transmitting side, such as an MPU timestamp descriptor, and the relative time of the PTS and DTS included in the MF element.
When the MPU is divided based on the information on whether the MPU is divided into a plurality of movie fragments, the reception apparatus 20 performs processing as follows based on the attributes of the divided movie fragments.
(1) When a movie fragment is the first movie fragment of the MPU, the reception device 20 generates the absolute time of the PTS and DTS using the absolute time of the PTS of the first sample included in the MPU timestamp descriptor and the relative time of the PTS and DTS included in the MF element.
(2) When the movie fragment is not the first movie fragment of the MPU, the reception device 20 generates the absolute time of the PTS and DTS using the relative time of the PTS and DTS included in the MF element without using the information of the MPU timestamp descriptor.
(3) When the movie segment is the last movie segment of the MPU, the reception device 20 calculates the absolute time of the PTS and DTS of all samples, and then resets the calculation process of the PTS and DTS (addition process of relative time). The reset process may be performed in the movie fragment at the head of the MPU.
The receiving apparatus 20 may determine whether or not a movie fragment is divided as follows. The receiving apparatus 20 may acquire attribute information of the movie fragment as follows.
For example, the reception apparatus 20 may also determine whether or not to be split based on the identifier movie _ fragment _ sequence _ number field value indicating the order of movie fragments shown in the MMTP payload header.
Specifically, the receiving apparatus 20 may determine that the MPU is divided into a plurality of movie fragments when the number of movie fragments included in 1 MPU is 1, the movie _ fragment _ sequence _ number field value is 1, and a value equal to or greater than 2 is present in the field value.
The receiving apparatus 20 may determine that the MPU is divided into a plurality of movie fragments when the number of movie fragments included in 1 MPU is 1, the movie _ fragment _ sequence _ number field value is 0, and values other than 0 exist in the field value.
The attribute information of the movie fragment may be determined based on movie _ fragment _ sequence _ number in the same manner.
Note that instead of using the movie _ fragment _ sequence _ number, it is also possible to determine whether or not a movie fragment is divided, or attribute information of a movie fragment, by counting the transmission of movie fragments or MF elements included in 1 MPU.
With the above-described configuration of the transmitting apparatus 15 and the receiving apparatus 20, the receiving apparatus 20 can receive movie fragment metadata at intervals shorter than the MPU, and can start decoding with low delay. Further, decoding with low delay is possible by the decoding process based on the MP4 analysis method.
The reception operation when the MPU is divided into a plurality of movie fragments as described above will be described with a flowchart. Fig. 41 is a flowchart of a receiving operation when the MPU is divided into a plurality of movie fragments. In addition, the flowchart illustrates the operation of step S604 of fig. 37 in more detail.
First, the receiving apparatus 20 acquires MF metadata when the data type is MF metadata based on the data type indicated by the MMTP payload header (S801).
Next, the receiving apparatus 20 determines whether or not the MPU is divided into a plurality of movie fragments (S802), and when the MPU is divided into a plurality of movie fragments (yes in S802), determines whether or not the received MF metadata is metadata at the head of the MPU (S803). When the received MF metadata is the MF metadata at the beginning of the MPU (yes in S803), the reception apparatus 20 calculates the absolute times of the PTS and DTS from the absolute time of the PTS shown in the MPU timestamp descriptor and the relative times of the PTS and DTS shown in the MF metadata (S804), and determines whether the received MF metadata is the last metadata of the MPU (S805).
On the other hand, when the received MF metadata is not the MF metadata at the beginning of the MPU (no in S803), the reception apparatus 20 calculates the absolute times of the PTS and DTS using the relative times of the PTS and DTS shown in the MF metadata without using the information of the MPU timestamp descriptor (S808), and proceeds to the process of step S805.
If it is determined in step S805 that the MF metadata is the last MF metadata of the MPU (yes in S805), the reception device 20 calculates the absolute time of the PTS and DTS of all samples, and then resets the calculation processing of the PTS and DTS. When it is determined in step S805 that the MF metadata is not the last MF metadata of the MPU (no in S805), the reception apparatus 20 ends the process.
When it is determined in step S802 that the MPU is not divided into a plurality of movie fragments (no in S802), the reception device 20 acquires sample data based on MF metadata transmitted after the MPU and determines a PTS and a DTS (S807).
The receiving apparatus 20 finally performs decoding processing and presentation processing based on the determined PTS and DTS, although not shown.
[ problem and solution strategy thereof when dividing movie fragments ]
Thus, a method for reducing end-to-end delay by splitting movie fragments is described. Hereinafter, a problem newly generated when a movie fragment is divided and a solution thereof will be described.
First, as a background, a picture structure in encoded data will be described. Fig. 42 is a diagram showing an example of a prediction structure of each picture of temporalld (time ID) when temporal adaptability is realized.
In a Coding scheme such as MPEG-4AVC or HEVC (High Efficiency Video Coding), temporal adaptability (temporal adaptability) can be achieved by using B pictures (bidirectional reference prediction pictures) that can be referred to from other pictures.
The temporalld shown in fig. 42 (a) is an identifier of a hierarchy of the coding structure, and a larger value of temporalld indicates a deeper hierarchy. The square blocks represent pictures, Ix in each block represents an I picture (intra-picture prediction picture), Px represents a P picture (forward reference prediction picture), and Bx represent B pictures (bidirectional reference prediction pictures). The x in Ix/Px/Bx shows the display order, representing the order in which the pictures are displayed. Arrows between pictures indicate a reference relationship, and for example, a picture indicating B4 generates a prediction image using I0 and B8 as reference images. Here, one picture is prohibited from using another picture having a temporalld larger than its own temporalld as a reference image. The predetermined le level is for temporal adaptability, and for example, in fig. 42, if all pictures are decoded, a 120fps (frame per second) video is obtained, and if only a level having a temporalld of 0 to 3 is decoded, a 60fps video is obtained.
Fig. 43 is a diagram showing a relationship between the Decoding Time (DTS) and the display time (PTS) of each picture in fig. 42. For example, picture I0 shown in fig. 43 is displayed after the decoding of B4 is completed, so that no gap is generated in the decoding and display.
As shown in fig. 43, when B pictures are included in the prediction structure, the decoding order differs from the display order, and therefore, it is necessary to perform a picture delay process and a picture rearrangement (reordering) process after decoding the pictures in the receiving apparatus 20.
Although the above describes an example of a prediction structure of a picture in temporal adaptability, in some cases, when temporal adaptability is not used, a delay process and a reordering process of pictures may be required by the prediction structure. Fig. 44 is a diagram showing an example of a prediction structure of a picture that requires a picture delay process and a picture reordering process. In addition, the numbers in fig. 44 indicate the decoding order.
As shown in fig. 44, according to the prediction structure, the sample at the head in the decoding order may be different from the sample at the head in the presentation order, and in fig. 44, the sample at the head in the presentation order is the 4 th sample in the decoding order. Fig. 44 shows an example of a prediction structure, but the prediction structure is not limited to this structure. In other prediction structures, the sample that starts in the decoding order may be different from the sample that starts in the presentation order.
Fig. 45 is a diagram showing an example in which an MPU configured by the MP4 format is divided into a plurality of movie fragments and stored in an MMTP payload or an MMTP packet, as in fig. 33. The number of samples constituting the MPU or the number of samples constituting the movie segment is arbitrary. For example, 2 movie slices may be configured by setting the number of samples constituting the MPU to the number of GOP-unit samples and setting the number of samples of one-half of GOP-unit samples as movie slices. The 1 sample may be regarded as 1 movie fragment, or the samples constituting the MPU may not be divided.
Fig. 45 shows an example in which 1 MPU includes 2 movie fragments (moof box and mdat box), but the number of movie fragments included in 1 MPU may be different from 2. The number of movie fragments included in 1 MPU may be 3 or more, or may be the number of samples included in the MPU. Further, the samples stored in the movie fragment may be divided into an arbitrary number of samples, instead of the number of equally divided samples.
The movie fragment metadata (MF metadata) includes information on the PTS, DTS, offset, and size of a sample included in a movie fragment, and when decoding a sample, the receiving apparatus 20 extracts the PTS and DTS from the MF element including the information on the sample, and determines the decoding timing or presentation timing.
Hereinafter, for convenience of detailed description, the absolute value of the decoding time of the i sample is referred to as dts (i), and the absolute value of the presentation time is referred to as pts (i).
The information of the ith sample among the time stamp information stored in the moof of the MF element is specifically a relative value between the decoding time of the ith sample and the (i +1) th sample, and a relative value between the decoding time of the ith sample and the presentation time, and hereinafter, these are referred to as dt (i) and ct (i).
In the movie fragment metadata #1, the DT (i) and CT (i) of samples #1- #3 are included, and in the movie fragment metadata #2, the DT (i) and CT (i) of samples #4- #6 are included.
The PTS absolute value of the access unit at the beginning of the MPU is stored in an MPU timestamp descriptor, and the reception device 20 calculates the PTS and DTs based on the PTS _ MPU, CT, and DT of the access unit at the beginning of the MPU.
Fig. 46 is a diagram for explaining a method of calculating PTS and DTS when an MPU is configured with samples #1 to #10, and a problem.
Fig. 46 (a) shows an example in which the MPU is not divided into movie fragments, fig. 46 (b) shows an example in which the MPU is divided into 2 movie fragments of 5 sample units, and fig. 46 (c) shows an example in which the MPU is divided into 10 movie fragments per sample unit.
As described with reference to fig. 45, when PTS and DTs are calculated using the MPU timestamp descriptor and the timestamp information (CT and DT) in the MP4, the sample that is the top in the presentation order of fig. 44 is the 4 th sample in the decoding order. Therefore, the PTS held in the MPU timestamp descriptor is the PTS (absolute value) of the sample that is 4 th in decoding order. Hereinafter, this sample is referred to as "a sample". The sample at the beginning of the decoding order is referred to as "B sample".
Since the absolute time information relating to the time stamp is only information of the MPU time stamp descriptor, the reception apparatus 20 cannot calculate the PTS (absolute time) and DTS (absolute time) of other samples until the a sample arrives. The receiving apparatus 20 cannot calculate the PTS and DTS of the B sample.
In the example of fig. 46 (a), the a samples are included in the same movie fragment as the B samples and stored in 1 MF element. Therefore, the reception device 20 can determine the DTS of the B sample immediately after receiving the MF element.
In the example of fig. 46 (B), the a samples are contained in the same movie fragment as the B samples and stored in 1 MF element. Therefore, the reception device 20 can determine the DTS of the B sample immediately after receiving the MF element.
In the example of fig. 46 (c), the a samples are contained in different movie fragments from the B samples. Therefore, the receiving apparatus 20 cannot determine the DTs of the B sample unless it receives the MF element including the CT and DT of the movie fragment including the a sample.
Therefore, in the case of the example in fig. 46 (c), the reception apparatus 20 cannot start decoding immediately after the B sample arrives.
As described above, if a samples are not included in the movie fragments including B samples, the reception apparatus 20 cannot start decoding of B samples unless the MF element related to the movie fragment including a samples is received.
This problem occurs when the first sample in the presentation order does not match the first sample in the decoding order, and the movie fragment is divided to such an extent that the a sample and the B sample are not stored in the same movie fragment. This problem occurs regardless of whether the MF element is sent later or earlier.
As described above, when the first sample in the presentation order does not match the first sample in the decoding order, if the a sample and the B sample are not stored in the same movie fragment, the DTS cannot be determined immediately after the B sample is received. Then, the transmission device 15 separately transmits information enabling the reception side to calculate the DTS (absolute value) of the B sample or the DTS (absolute value) of the B sample. Such information may also be transmitted using control information, a header, or the like.
The receiving apparatus 20 calculates the DTS (absolute value) of the B sample using such information. Fig. 47 is a flowchart of a reception operation when calculating the DTS using such information.
The receiving apparatus 20 receives the movie fragment at the beginning of the MPU (S901), and determines whether or not the a sample and the B sample are stored in the same movie fragment (S902). When the information is stored in the same movie fragment (yes in S902), the reception apparatus 20 calculates the DTS using only the information of the MF element without using the DTS (absolute time) of the B sample, and starts decoding (S904). In step S904, the reception device 20 may determine the DTS using the DTS of the B sample.
On the other hand, if the a sample and the B sample are not stored in the same movie fragment in step S902 (no in S902), the receiving apparatus 20 acquires the DTS (absolute time) of the B sample, determines the DTS, and starts decoding (S903).
In the above description, an example has been described in which the absolute value of the decoding time and the absolute value of the presentation time of each sample are calculated using the MF element (time stamp information stored in moof of the MP4 format) in the MMT standard, but it is needless to say that the MF element may be replaced with any control information that can be used to calculate the absolute value of the decoding time and the absolute value of the presentation time of each sample. Examples of such control information include control information obtained by replacing the relative value ct (i) of the decoding time of the i-th sample and the (i +1) -th sample with the relative value ct (i) of the presentation time of the i-th sample and the (i +1) -th sample, or control information including both the relative value ct (i) of the decoding time of the i-th sample and the (i +1) -th sample and the relative value ct (i) of the presentation time of the (i +1) -th sample of the i-th sample.
(embodiment mode 3)
[ summary ]
In embodiment 3, a description will be given of a content transmission method and a data structure in the case of transmitting a content such as a video, audio, subtitle, and data broadcast by broadcasting. That is, a description will be given of a content transmission method and a data structure that are specialized for reproduction of a broadcast stream.
In embodiment 3, an example in which an MMT scheme (hereinafter, also simply referred to as MMT) is used as the multiplexing scheme is described, but other multiplexing schemes such as MPEG-DASH and RTP may be used.
First, details of a method of storing a payload in a Data Unit (DU) in an MMT will be described. Fig. 48 is a diagram for explaining a method of depositing a data unit in MMT into a payload.
In MMT, a transmitting apparatus stores a part of data constituting an MPU as a data unit in an MMTP payload, adds a header, and transmits the data. The MMTP payload header and the MMTP header are contained in the header. The unit of a data unit may be NAL unit or sample unit. In the case where the MMTP packet is scrambled, the payload becomes the object of the scrambling.
Fig. 48 (a) shows an example in which the transmitting apparatus stores a plurality of data units in a single payload in a lump. In the example of fig. 48 (a), a Data Unit Header (DUH) and a Data Unit Length (DUL) are assigned to the head of each of a plurality of Data units, and a plurality of Data units to which the Data Unit Header and the Data Unit Length are assigned are collected and stored in the payload.
Fig. 48 (b) shows an example of storing one data unit in one payload. In the example of fig. 48 (b), a data unit header is assigned to the head of a data unit and stored in the payload. Fig. 48 (c) shows an example in which one data unit is divided, and a data unit header is added to the divided data unit and stored in the payload.
The data units are classified into the following types: a time-MFU, which is a medium including information related to synchronization such as video, audio, and subtitles, a non-time-MFU, which is a medium such as a file not including information related to synchronization, MPU metadata, MF metadata, and the like, and a data unit header is determined according to the type of a data unit. In addition, no data unit header exists in the MPU metadata and the MF metadata.
The transmitting apparatus cannot group different types of data units as a principle, but may be specified to group different types of data units. For example, when the size of the MF metadata is small, such as when the MF metadata is divided into movie fragments for each sample, the number of packets can be reduced and the transmission capacity can be reduced by integrating the MF metadata with the media data.
When the data unit is an MFU, a part of information of the MPU, such as information configuring the MPU (MP4), is stored as a header.
For example, the header of the timed-MFU includes movie _ fragment _ sequence _ number, sample _ number, offset, priority, and dependency _ counter, and the header of the non-timed-MFU includes item _ iD. The meaning of the fields is shown in the standards ISO/IEC23008-1 or ARIB STD-B60, etc. The meaning of each field defined in such a standard is described below.
movie _ fragment _ sequence _ number represents the sequence number of the movie fragment to which the MFU belongs, and is also shown in ISO/IEC 14496-12.
sample _ number represents the sample number to which the MFU belongs, and is also shown in ISO/IEC 14496-12.
The offset represents an offset amount of the MFU in the sample to which the MFU belongs in byte units.
The priority indicates the relative importance of the MFU in the MPU to which the MFU belongs, and an MFU with a higher priority number is more important than an MFU with a lower priority number.
The dependency _ counter indicates the number of MFUs on which the decoding process depends (i.e., the number of MFUs that cannot perform the decoding process if the decoding process is not performed on the MFUs). For example, when a B picture or a P picture refers to an I picture when the MFU is HEVC, the B picture or the P picture cannot be decoded unless the I picture is decoded.
Therefore, when the MFU is a sample unit, the number of pictures referring to the I picture is indicated in the dependency _ counter in the MFU of the I picture. In the case where the MFU is a NAL unit, the dependency _ counter in the MFU belonging to an I picture shows the number of NAL units belonging to a picture referring to the I picture. Furthermore, in the case of a video signal that is temporally hierarchically encoded, the MFU of the extension layer depends on the MFU of the base layer, and therefore the number of MFUs of the extension layer is indicated in the dependency _ counter in the MFU of the base layer. The cost field cannot be generated if it is not after the number of MFUs to be relied upon is determined.
item _ iD represents an identifier that uniquely determines an item.
[ MP4 non-support mode ]
As described in fig. 19 and 21, as a method for transmitting an MPU in the MMT by the transmission apparatus, there are a method for transmitting MPU metadata or MF metadata before or after media data, and a method for transmitting only media data. In addition, there are a method of decoding using a reception apparatus or a reception method conforming to MP4 and a method of decoding without using a header.
As a method of transmitting data specialized for broadcast stream reproduction, for example, there is a method of transmitting data that does not support MP4 reconstruction in a receiving apparatus.
As a transmission method that does not support MP4 reconstruction in the receiving apparatus, for example, as shown in fig. 21 (b), a method is used in which metadata (MPU metadata and MF metadata) is not transmitted. In this case, the field value of the fragmentation type (information indicating the kind of the data unit) included in the MMTP packet is fixed to 2(═ MFU).
When metadata is not transmitted, as described above, in a receiving device or the like conforming to MP4, the received data cannot be decoded into MP4, but can be decoded without using metadata (header).
Therefore, the metadata is not information that is necessarily required for decoding and reproducing the broadcast stream. Also, the information of the data unit header in the timed-MFU illustrated in fig. 48 is information for reconstructing the MP4 in the receiving apparatus. Since the MP4 does not need to be reconstructed in the broadcast stream reproduction, information of a data unit header (hereinafter also referred to as a timed-MFU header) in the timed-MFU is not necessarily required for the broadcast stream reproduction.
The receiving device can easily reconstruct the MP4 by using the metadata and information for reconstructing the MP4 (hereinafter also referred to as MP4 configuration information) in the data unit header. However, even if only one of the metadata and the MP4 configuration information in the data unit header is transmitted, the receiving apparatus cannot easily reconstruct the MP 4. Only one of the transmission metadata and the information for reconstructing the MP4 provides little advantage, and the generation and transmission of unnecessary information increases the processing and reduces the transmission efficiency.
Therefore, the transmitting apparatus controls the data structure and transmission of the MP4 configuration information by the following method. The transmitting apparatus determines whether or not the MP4 configuration information is indicated in the data unit header based on whether or not the metadata is transmitted. Specifically, the transmitting apparatus indicates MP4 configuration information in the data unit header when metadata is transferred, and does not indicate MP4 configuration information in the data unit header when metadata is not transferred.
As a method of not indicating the MP4 configuration information in the data unit header, for example, the following method can be used.
1. The transmitting apparatus sets the MP4 configuration information as reserved (reserved) and does not operate. This can reduce the amount of processing on the transmitting side (the amount of processing of the transmitting device) for generating the MP4 configuration information.
2. The transmitting apparatus deletes the MP4 configuration information and performs header compression. This reduces the amount of processing on the transmitting side for generating the MP4 configuration information, and reduces the transmission capacity.
In addition, when deleting the MP4 configuration information and performing header compression, the transmission device may indicate a flag indicating that the MP4 configuration information is deleted (compressed). The flag is shown in a header (MMTP header, MMTP payload header, data unit header) or control information, etc.
The information indicating whether or not to transmit the metadata may be determined in advance, or may be separately transmitted to the receiving apparatus by signaling (signaling) in the header or the control information.
For example, information indicating whether or not to transmit metadata corresponding to the MFU may be stored in the MFU header.
On the other hand, the receiving apparatus can determine whether or not the MP4 configuration information is indicated based on whether or not the metadata is transmitted.
Here, when the order of data transmission (for example, the order of MPU metadata, MF metadata, and media data) is determined, the receiving apparatus may determine whether or not the metadata is received before the media data.
When the MP4 configuration information is indicated, the receiving apparatus uses the MP4 configuration information for the reconstruction of MP 4. Alternatively, the receiving apparatus can use the MP4 configuration information when detecting the head of another access unit or NAL unit.
The MP4 configuration information may be all or a part of the timed-MFU header.
In addition, the transmitting apparatus may similarly determine whether or not the item id is represented in the non-timed-MFU header based on whether or not the metadata is transmitted.
The transmitting apparatus may indicate the MP4 configuration information only in one of the time-MFU and non-time-MFU. When only one of the two parties indicates the MP4 configuration information, the transmitting apparatus determines whether to indicate the MP4 configuration information based on not only whether to transmit the metadata but also on whether to indicate the timed-MFU or the non-timed-MFU. In the receiving apparatus, it can be determined whether the MP4 configuration information is represented based on whether the metadata is transmitted or not and the timed/non-timed flag.
In the above description, the transmitting apparatus determines whether to represent MP4 configuration information based on whether or not metadata (both MPU metadata and MF metadata) is transmitted. However, the transmission device may be: when a part of the metadata (either the MPU metadata or the MF metadata) is not transmitted, the MP4 configuration information is not indicated.
The transmitting apparatus may determine whether or not to indicate the MP4 configuration information based on information other than the metadata.
For example, it is also possible to provide: modes such as MP4 support mode/MP 4 non-support mode are defined, and the transmitting apparatus indicates MP4 configuration information in the data unit header in the case of MP4 support mode, and does not indicate MP4 configuration information in the data unit header in the case of MP4 non-support mode. Further, it is also possible to: the transmitting apparatus transmits metadata and indicates MP4 configuration information in a data unit header when the MP4 supports the mode, and does not transmit metadata and does not indicate MP4 configuration information in a data unit header when the MP4 does not support the mode.
[ operation flow of Transmission device ]
Next, the operation flow of the transmission device will be described. Fig. 49 is an operation flow of the transmission device.
The transmitting apparatus first determines whether or not to transmit metadata (S1001). When the transmitting apparatus determines that the metadata is to be transferred (yes in S1002), the flow proceeds to step S1003, and MP4 configuration information is generated, stored in the header, and transferred (S1003). In this case, the transmitting apparatus also generates and transmits metadata.
On the other hand, when the transmitting apparatus determines that metadata is not to be transferred (no in S1002), the transmitting apparatus transfers the metadata without generating MP4 configuration information and storing the configuration information in a header (S1004). In this case, the transmitting apparatus does not generate metadata and does not transmit it.
Whether or not to transfer metadata in step S1001 may be determined in advance, or may be determined based on whether or not metadata is generated inside the transmission device and whether or not the metadata is transferred inside the transmission device.
[ operation flow of receiver ]
Next, the operation flow of the receiving apparatus will be described. Fig. 50 is an operation flow of the receiving apparatus.
The receiving apparatus first determines whether or not metadata is transmitted (S1101). By monitoring the slice type in the MMTP packet payload, it can be determined whether the metadata is transferred. In addition, whether to transmit or not may be determined in advance.
When the receiving apparatus determines that the metadata is transferred (yes in S1102), the receiving apparatus reconstructs the MP4 and performs decoding processing using the MP4 configuration information (S1103). On the other hand, if it is determined that the metadata is not transferred (no in S1102), the reconstruction processing of the MP4 is not performed, and the decoding processing is performed without using the MP4 configuration information (S1104).
Further, the receiving apparatus can perform detection of a random access point, detection of an access unit head, detection of a NAL unit head, and the like without using the MP4 configuration information by using the method described above, and can perform decoding processing, detection of packet loss, and recovery processing from packet loss.
For example, when non-VCL NAL units are grouped individually as data units as shown in fig. 38, the access unit header is the header data of the MMT payload with the aggregation _ flag value of 1. At this time, the Fragmentation _ indicator value is 0.
In addition, the beginning of the slice is the beginning data of the MMT payload whose aggregation _ flag value is 0 and whose fragmentation _ indicator value is 00 or 01.
The receiving apparatus can detect the beginning of the access unit and the slice based on the above information.
The receiving apparatus may analyze the NAL unit header in a packet including the head of a data unit whose fragmentation _ indicator value is 00 or 01, and detect whether the type of the NAL unit is an AU delimiter and whether the type of the NAL unit is a slice segment.
[ simple mode of broadcast ]
Although the above description has been made of a method in which MP4 configuration information is not supported in a receiving apparatus as a method for transmitting data specialized for broadcast stream reproduction, the method for transmitting data specialized for broadcast stream reproduction is not limited to this.
As a method of transmitting data specialized for broadcast stream reproduction, for example, the following method can be used.
The transmitting device may not use AL-FEC in a fixed reception environment of broadcasting. In case of not using AL-FEC, FEC _ type in MMTP header is always fixed to 0.
The transmitting apparatus may always use AL-FEC in the mobile broadcast reception environment and the UDP transmission mode of communication. In case of using AL-FEC, FEC _ type in MMTP header is always 0 or 1.
The transmitting apparatus may not perform bulk transfer of resources. When the bulk transfer of the resource is not performed, the location _ information indicating the number of transfer positions of the resource in the MPT may be fixed to 1.
The transmission device may not perform the mixed transmission of the resource, the program, and the message.
For example, a broadcast simple mode may be defined, and the transmitting apparatus may set the MP4 non-support mode in the broadcast simple mode, or may use the above-described data transmission method specialized for broadcast stream playback. Whether or not the broadcast mode is set may be determined in advance, or the transmitting apparatus may store a flag indicating that the broadcast mode is set as control information and transmit the control information to the receiving apparatus.
The transmitting apparatus may use the above-described data transmission method specialized for broadcast stream playback as the broadcast simple mode based on whether the MP4 unsupported mode (whether metadata is transmitted) is described in fig. 49, and when the MP4 unsupported mode is used.
In the broadcast simple mode, the receiving apparatus is set to the MP4 non-support mode, and can perform decoding processing without reconstructing the MP 4.
In the broadcast simple mode, the receiving apparatus determines the function specified for the broadcast, and can perform the receiving process specified for the broadcast.
Thus, in the broadcast simple mode, by using only the function specialized for the broadcast, it is possible to reduce not only processing unnecessary for the transmitting apparatus and the receiving apparatus but also transmission overhead by compressing unnecessary information without transmission.
In the case of using the MP4 non-support mode, presentation information for supporting an accumulation method other than the MP4 configuration may be shown.
Examples of accumulation methods other than the MP4 configuration include a method of directly accumulating MMT packets or IP packets, and a method of converting MMT packets into MPEG-2TS packets.
In addition, in the case of the MP4 non-supported mode, a format that does not comply with the MP4 configuration may be used.
For example, when the MP4 non-support mode is used as the data stored in the MFU, the data may be in a format in which a byte header is added instead of a format in which the size of an NAL unit is added to the beginning of an NAL unit in the MP4 format.
In MMT, a resource type indicating the type of the resource is described in 4CC registered in MP4REG (http:// www.mp4ra.org), and when HEVC is used as a video signal, "HEV 1" or "HVC 1" is used. "HVC 1" is a form that may also contain parameter sets among samples, "HEV 1" is a form that does not contain parameter sets among samples, but contains parameter sets in sample entries in MPU metadata.
In the case of the broadcast simple mode or the MP4 non-support mode, it can be also specified that the parameter set must be included in the sample without transmitting the MPU metadata and the MF metadata. Further, it can be defined that: whichever asset type represents "HEV 1" or "HVC 1" must take the form of "HVC 1".
[ supplement 1: transmitting device
As described above, a transmission apparatus that does not operate with MP4 configuration information set as reserved (reserved) when metadata is not transmitted can be configured as shown in fig. 51. Fig. 51 is a diagram showing an example of a specific configuration of a transmission device.
The transmission device 300 includes an encoding unit 301, an adding unit 302, and a transmission unit 303. The encoding unit 301, the giving unit 302, and the transmitting unit 303 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The encoding unit 301 encodes a video signal or an audio signal to generate sample data. The sample data is specifically a data unit.
The adding unit 302 adds header information including MP4 configuration information to sample data, which is data encoded as a video signal or an audio signal. The MP4 configuration information is information for reconstructing the sample data into a file in the MP4 format on the receiving side, and is information whose content differs depending on whether or not the presentation time of the sample data is decided.
As described above, the assigning unit 302 includes MP4 configuration information such as motion _ fragment _ sequence _ number, sample _ number, offset, priority, and dependency _ counter in the header (header information) of the time-MFU, which is an example of sample data (sample data including information related to synchronization) whose presentation time is determined.
On the other hand, the assigning unit 302 includes MP4 configuration information such as item _ id in the header (header information) of the non-timed-MFU, which is an example of sample data (sample data not including information related to synchronization) whose presentation time is not determined.
In the case where the metadata corresponding to the sample data is not transmitted by the transmission unit 303 (for example, in the case of fig. 21 (b)), the addition unit 302 adds header information that does not include MP4 configuration information to the sample data, depending on whether or not the presentation time of the sample data is determined.
Specifically, the assigning unit 302 assigns header information not including the first MP4 configuration information to the sample data when the presentation time of the sample data is determined, and assigns header information including the second MP4 configuration information to the sample data when the presentation time of the sample data is not determined.
For example, as shown in step S1004 of fig. 49, when the transmission unit 303 does not transmit metadata corresponding to sample data, the addition unit 302 essentially does not generate MP4 configuration information and essentially does not store the MP 3578 configuration information in a header (header information) by setting the MP4 configuration information to a reserved (fixed value). The metadata includes MPU metadata and movie fragment metadata.
The transmission unit 303 transmits the sample data to which the header information is added. More specifically, the transmission unit 303 packetizes and transmits the sample data to which the header information is added by the MMT scheme.
As described above, in the transmission method and the reception method specialized for reproduction of a broadcast stream, the reception apparatus side does not need to reconstruct a data unit to the MP 4. When the receiving apparatus does not need to be reconfigured to the MP4, unnecessary information such as MP4 configuration information is not generated, and processing by the transmitting apparatus is reduced.
On the other hand, the transmitting apparatus needs to maintain the compatibility with the standard so that necessary information needs to be transmitted, but does not need to separately transmit extra additional information or the like.
With the configuration of the transmission device 300, by setting the area in which the MP4 configuration information is stored to a fixed value or the like, it is possible to transmit only necessary information based on the standard without transmitting the MP4 configuration information, and thus it is possible to obtain an effect that redundant additional information is not transmitted. That is, the configuration of the transmitting apparatus and the processing amount of the receiving apparatus can be reduced. In addition, since useless data is not transmitted, the transmission efficiency can be improved.
[ supplement 2: receiving device
The receiving apparatus corresponding to the transmitting apparatus 300 may be configured as shown in fig. 52, for example. Fig. 52 is a diagram showing another example of the configuration of the receiving apparatus.
The reception device 400 includes a reception unit 401 and a decoding unit 402. The receiving unit 401 and the decoding unit 402 are realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The receiving unit 401 receives sample data, which is data obtained by encoding a video signal or an audio signal and to which header data including MP4 configuration information for reconstructing the sample data into a file in the MP4 format is added.
When the receiving unit does not receive the metadata corresponding to the sample data and the presentation time of the sample data is determined, the decoding unit 402 decodes the sample data without using the MP4 configuration information.
For example, as shown in step S1104 in fig. 50, when the metadata corresponding to the sample data is not received by the receiving unit 401, the decoding unit 402 executes the decoding process without using the MP4 configuration information.
This can reduce the configuration of the reception apparatus 400 and the amount of processing in the reception apparatus 400.
(embodiment mode 4)
[ summary ]
In embodiment 4, a method of storing an asynchronous (non-timed) medium such as a file, which does not include information related to synchronization, in an MPU and a method of transmitting the medium in an MMTP packet are described. In embodiment 4, an MPU in the MMT is described as an example, but the present invention is also applicable to DASH based on MP 4.
First, details of a method of storing non-timed media (hereinafter, also referred to as "asynchronous media data") in the MPU will be described with reference to fig. 53. FIG. 53 shows a method of storing non-timed media in an MPU and a method of transferring non-timed media in an MMT package.
The MPU for storing non-timed media is composed of boxes (box) such as ftyp, mmpu, moov, meta, etc., and stores information related to files stored in the MPU. A plurality of idat boxes can be stored in the meta box, and one file is stored as item in the idat box.
Part of ftyp, mmpu, moov and meta boxes form a data unit as MPU metadata, and item or idat boxes form a data unit as MFU.
After being summarized or fragmented, the data units are given a data unit header, an MMTP payload header, and an MMTP header and transmitted as MMTP packets.
Fig. 53 shows an example in which File #1 and File #2 are stored in one MPU. The MPU metadata is not divided, and the MFUs are divided and stored in the MMTP packet, but is not limited thereto, and may be aggregated or fragmented according to the size of the data unit. In addition, MPU metadata may not be transmitted, in which case only the MFU is transmitted.
Header information such as a data unit header shows itemID (an identifier that uniquely identifies an item), and an MMTP payload header or an MMTP header contains a packet sequence number (sequence number for each packet) and an MPU sequence number (sequence number of an MPU, a number unique within a resource).
The data structure of the MMTP payload header or MMTP header other than the data unit header includes an aggregation _ flag, a fragmentation _ indicator, a fragmentation _ counter, and the like, as in the case of the timed media (hereinafter, also referred to as "synchronous media data") described above.
Next, a specific example of header information in the case of dividing and packaging a file (Item) and MFU) will be described with reference to fig. 54 and 55.
Fig. 54 and 55 are diagrams showing an example in which a file is divided into a plurality of pieces of divided data, and the divided data are packed and transmitted. Fig. 54 and 55 specifically show information (packet sequence number, fragment counter, fragment identifier, MPU sequence number, and item ID) included in any one of the data unit header, MMTP payload header, and MMTP packet header, which is header information of each of the MMTP packets after the division. Fig. 54 is a diagram showing an example in which File #1 is divided into M pieces (M < (256)), and fig. 55 is a diagram showing an example in which File #2 is divided into N pieces (256< N).
The divided data number indicates an index of the divided data with respect to the beginning of the file, and this information is not transmitted. That is, the divided data number is not included in the header information. The divided data number is a number added to each packet corresponding to a plurality of divided data obtained by dividing a file, and is a number given by adding 1 to the first packet in ascending order.
The packet sequence number is the sequence number of a packet having the same packet ID, and in fig. 54 and 55, the divided data at the beginning of the file is a, and the divided data up to the end of the file is given consecutive numbers. The packet sequence number is a number given by adding 1 to the divided data from the beginning of the file in ascending order, and is a number corresponding to the divided data number.
The fragmentation counter indicates the number of pieces of divided data that are later than the divided data among pieces of divided data obtained by dividing one file. When the number of pieces of divided data obtained by dividing one file, that is, the number of pieces of divided data exceeds 256, the slice counter indicates the remainder obtained by dividing the number of pieces of divided data by 256. In the example of fig. 54, since the number of divided data is 256 or less, the field value of the slice counter is (M-divided data number). On the other hand, in the example of fig. 55, since the number of divided data exceeds 256, the remainder ((N-divided data number)% 256) is obtained by dividing (N-divided data number) by 256.
The fragment identifier indicates a state of division of data stored in the MMTP packet, and indicates a value indicating that the data is the first divided data, the last divided data, the other divided data, or one or more data units that are not divided. Specifically, the slice identifier is "01" in the first divided data, "11" in the last divided data, "10" in the remaining divided data, and "00" in the undivided data unit.
In the present embodiment, the remainder obtained by dividing the number of divided data by 256 is shown when the number of divided data exceeds 256, but the number of divided data is not limited to 256, and may be other numbers (predetermined numbers).
When a file is divided as shown in fig. 54 and 55, and a plurality of pieces of divided data obtained by dividing the file are transmitted by adding conventional header information to each of them, the following information does not exist in the receiving apparatus: the data stored in the received MMTP packet is the number of divided data (divided data number) in the original file, the number of divided data of the file, or information from which the divided data number and the number of divided data can be derived. Therefore, in the conventional transmission method, even if the MMTP packet is received, the divided data number or the number of divided data of the data stored in the received MMTP packet cannot be uniquely detected.
For example, as shown in fig. 54, when the number of divided data is 256 or less and the number of divided data is known in advance to be 256 or less, the number of divided data or the number of divided data can be specified by referring to the slice counter. However, when the number of pieces of divided data is 256 or more, the number of pieces of divided data or the number of pieces of divided data cannot be specified.
When the number of divided data of a file is limited to 256 or less, the maximum size of the file that can be transferred is limited to x × 256 bytes when the data size that can be transferred in one packet is x bytes. For example, in the broadcast, it is assumed that x is 4 kbytes, and in this case, the maximum size of a file that can be transferred is limited to 4k 256 to 1 mbyte. Therefore, when a file of more than 1 mbyte is desired to be transferred, the number of divided data of the file cannot be limited to 256 or less.
Further, for example, since the first divided data or the last divided data of a file can be detected by referring to the slice identifier, the number of MMTP packets can be counted until the MMTP packet including the last divided data of the file is received, or the number of divided data can be calculated by combining the MMTP packet including the last divided data of the file with the packet sequence number after the MMTP packet including the last divided data of the file is received. However, when receiving data from an MMTP packet including split data in the middle of a file (i.e., split data that is neither the first split data nor the last split data of the file), the number of split data or the number of split data of the split data cannot be determined. The split data number or the split data number of the split data can be determined only after receiving the MMTP packet including the last split data of the file.
The problem described in fig. 54 and 55 is to uniquely identify the divided data number and the number of divided data of the file at the time of starting reception of the packet including the divided data of the file from the middle, and the following method is used.
First, the divided data number is explained.
As for the split data number, a packet sequence number in the split data at the head of the file (item) is signaled.
The method for signaling transmission is stored in the control information of the management file. Specifically, the packet sequence number a of the divided data at the head of the file in fig. 54 and 55 is stored in the control information. The receiving device acquires the value of a from the control information, and calculates the divided data number from the packet sequence number indicated in the packet header.
The divided data number of the divided data is obtained by subtracting the packet sequence number a of the first divided data from the packet sequence number of the divided data.
The control information for managing the file is, for example, a resource management table defined in ARIB STD-B60. The resource management table shows file size, version information, and the like for each file, and stores and transmits the file in a data transmission message. Fig. 56 is a diagram showing the syntax of a loop for each file in the resource management table.
When the area of the existing file management table cannot be expanded, signaling may be performed using a 32-bit area of a part of the item _ info _ byte field that represents information of an item. In a part of the area of the item _ info _ byte, a flag indicating whether or not the packet sequence number in the divided data at the beginning of the file (item) is indicated may be included in, for example, a reserved _ future _ use field of the control information.
When a file is repeatedly transmitted in a data carousel or the like, a plurality of packet sequence numbers may be indicated, or a packet sequence number at the head of the file to be transmitted immediately thereafter may be indicated.
The packet sequence number of the divided data at the beginning of the file is not limited, and may be information associating the divided data number of the file with the packet sequence number.
Next, the number of divided data will be described.
The order of the loop for each file included in the resource management table may be defined as the file transfer order. Thus, since the packet sequence numbers at the head of two consecutive files in the transfer order are known, the number of divided data of a file transferred before can be determined by subtracting the packet sequence number at the head of the file transferred before from the packet sequence number at the head of the file transferred after. That is, for example, in the case where File #1 shown in fig. 54 and File #2 shown in fig. 55 are files that are consecutive in this order, the last packet sequence number of File #1 and the first packet sequence number of File #2 are given consecutive numbers.
Further, the number of divided data of a file may be specified by specifying a file dividing method. For example, when the number of pieces of divided data is N, the number of pieces of divided data can be inversely calculated from item _ size indicated in the resource management table by setting the size of each of the 1 st to (N-1) th pieces of divided data to L and defining the size of the nth piece of divided data as a mantissa (item _ size-L (N-1)). In this case, the number of pieces of divided data is the integer value obtained by carrying the mantissa (item _ size/L). In addition, the file division method is not limited thereto.
Alternatively, the number of pieces of divided data may be stored in the resource management table as it is.
In the receiving apparatus, by using the above method, control information is received, and the number of pieces of divided data is calculated based on the control information. Further, the packet sequence number corresponding to the divided data number of the file can be calculated based on the control information. When the reception timing of the packet of the divided data is earlier than the reception timing of the control information, the divided data number or the number of the divided data may be calculated at the timing when the control information is received.
In addition, when the divided data number or the number of divided data is signaled by using the above method, the divided data number or the number of divided data is not determined based on the slice counter, and the slice counter becomes useless data. Therefore, when information for specifying the number of divided data and the number of divided data is signaled by using the above-described method or the like in the asynchronous media transmission, the fragmentation counter is not used, or header compression may be performed. This can reduce the processing amount of the transmitting device and the receiving device, and can improve the transmission efficiency. That is, when asynchronous media is transmitted, the segment counter may be set to be reserved (invalidated). Specifically, the value of the slice counter may be a fixed value such as "0". In addition, the slice counter may be ignored when receiving asynchronous media.
When a synchronous medium such as video or audio is stored, the transmission order of the MMTP packets in the transmitting device matches the arrival order of the MMTP packets in the receiving device, and the packets are not retransmitted. In this case, if it is not necessary to detect a packet loss and reconstruct the packet, the slice counter may not be used. In other words, in this case, the slice counter may be set to be reserved (invalidated).
Further, without using the slice counter, detection of a random access point, detection of the head of an access unit, detection of the head of a NAL unit, and the like can be performed, and decoding processing, detection of packet loss, recovery processing from packet loss, and the like can be performed.
Further, in the transmission of real-time content such as live broadcasting, transmission with lower delay is required, and it is required to sequentially package and transmit encoded data. However, in the real-time content transmission, since the conventional slice counter cannot determine the number of pieces of divided data when the first piece of divided data is transmitted, a delay occurs after the transmission of the first piece of divided data is completed until the number of pieces of divided data is determined after all the encoding of the data unit is completed. In this case, by not using the slice counter using the above method, the delay can be reduced.
Fig. 57 is a flowchart of an operation of determining a divided data number in the receiving apparatus.
The receiving apparatus acquires control information in which information of a file is described (S1201). The reception device determines whether or not the control information indicates the packet sequence number at the beginning of the file (S1202), and when the control information indicates the packet sequence number at the beginning of the file (yes in S1202), calculates the packet sequence number corresponding to the divided data number of the divided data of the file (S1203). Then, after acquiring the MMTP packet storing the divided data, the receiving apparatus identifies the divided data number of the file based on the packet sequence number stored in the header of the acquired MMTP packet (S1204).
On the other hand, when the control information does not include the packet sequence number indicating the file head (no in S1202), the receiving apparatus acquires the MMTP packet including the last fragmented data of the file, and then specifies the fragmented data number using the fragment identifier and the packet sequence number stored in the packet header of the acquired MMTP packet (S1205).
Fig. 58 is a flowchart of an operation of determining the number of divided data in the receiving apparatus.
The receiving apparatus acquires control information in which information of a file is described (S1301). The receiving device determines whether or not information capable of calculating the number of divided data of the file is included in the control information (S1302), and if it is determined that information capable of calculating the number of divided data is included (yes in S1302), calculates the number of divided data based on the information included in the control information (S1303). On the other hand, when the receiving apparatus determines that the number of divided data cannot be calculated (no in S1302), after acquiring the MMTP packet including the last divided data of the file, the receiving apparatus specifies the number of divided data using the fragment identifier and the packet sequence number stored in the header of the acquired MMTP packet (S1304).
Fig. 59 is an operation flow for determining whether or not to use the slice counter in the transmitting apparatus.
First, the transmitting apparatus determines whether the medium to be transmitted (hereinafter also referred to as "media data") is a synchronous medium or an asynchronous medium (S1401).
If the result of the determination in step S1401 is a synchronized medium (synchronized medium in S1402), the transmitting apparatus determines whether the MMTP packets transmitted and received in the environment in which the synchronized medium is transmitted are in the same order and the reconstruction of the packet is not necessary when the packet is lost (S1403). When the transmitting apparatus determines that it is unnecessary (yes in S1403), it does not use the slice counter (S1404). On the other hand, when the transmitting apparatus determines that it is not unnecessary (no in S1403), it operates the slice counter (S1405).
If the result of the determination in step S1401 is an asynchronous medium (asynchronous medium in S1402), the transmission apparatus determines whether or not to use the fragmentation counter based on whether or not the number of divided data and the number of divided data are signaled using the method described above. Specifically, when the transmission device signals the number of divided data and the number of divided data (yes in S1406), the transmission device does not use the slice counter (S1404). On the other hand, when the transmission device does not signal the number of divided data or the number of divided data (no in S1406), the transmission device uses the slice counter (S1405).
In addition, when the transmitting apparatus does not use the slice counter, the transmitting apparatus may retain the value of the slice counter, or may perform header compression.
The transmitting apparatus may determine whether to perform signaling on the divided data number and the number of divided data based on whether or not the slice counter is used.
In addition, when the synchronous medium does not use the segment counter, the transmitting apparatus may signal the segment number and the number of segments in the asynchronous medium by using the above-described method. Conversely, the operation of the synchronized media may be determined based on whether the segment counter is used by the unsynchronized media. In this case, whether or not the slice is used can be the same operation in the synchronous media and the asynchronous media.
Next, a method of determining the number of pieces of divided data and the number of pieces of divided data (in the case of using a slice counter) will be described. Fig. 60 is a diagram for explaining a method of determining the number of pieces of divided data and the number of pieces of divided data (in the case of using a slice counter).
As described with reference to fig. 54, when the number of pieces of divided data is 256 or less and the number of pieces of divided data is known in advance to be 256 or less, the number of pieces of divided data or the number of pieces of divided data can be specified by referring to the slice counter.
When the number of divided data of a file is limited to 256 or less, the maximum size of the file that can be transferred is limited to x × 256 bytes when the data size that can be transferred in one packet is x bytes. For example, in the broadcast, it is assumed that x is 4 kbytes, and in this case, the maximum size of a file that can be transferred is limited to 4k 256 to 1 mbyte.
When the file size exceeds the maximum size of the transferable file, the file is divided in advance so that the size of the divided file becomes x × 256 bytes or less. Each of a plurality of divided files obtained by dividing a file is processed as one file (item), and is further divided into 256 or less, and divided data obtained by further dividing is stored in an MMTP packet and transferred.
Further, information indicating that the item is a divided file, the number of divided files, and the serial numbers of the divided files may be stored in the control information and transmitted to the receiving apparatus. These pieces of information may be stored in the resource management table, or may be represented by a part of the existing field item _ info _ byte.
When the item is one of a plurality of divided files obtained by dividing one file, the receiving apparatus identifies the other divided files and can reconstruct the original file. In addition, in the receiving apparatus, the number of divided data and the number of divided data can be uniquely determined by using the number of divided files of the divided file, the index of the divided file, and the slice counter in the control information. In addition, the number of pieces of divided data and the number of pieces of divided data can be uniquely determined without using a packet sequence number or the like.
Here, it is preferable that item _ ids of a plurality of divided files obtained by dividing one file are the same. When another item _ id is assigned, the item _ id of the top divided file may be indicated so as to uniquely refer to the file based on other control information or the like.
In addition, a plurality of divided files may necessarily belong to the same MPU. When a plurality of files are stored in the MPU, it is necessary to store a file obtained by dividing one file, instead of storing a plurality of types of files. The receiving apparatus can detect the update of the file by confirming the version information of each MPU even without confirming the version information of each item.
Fig. 61 is an operation flow of the transmitting apparatus in the case of using the slice counter.
First, the transmitting apparatus confirms the size of a file to be transferred (S1501). Next, the transmitting apparatus determines whether or not the file size exceeds x × 256 bytes (x is a data size that can be transferred for one packet, for example, MTU size) (S1502), and when the file size exceeds x × 256 bytes (yes in S1502), the transmitting apparatus divides the file so that the size of the divided file is smaller than x × 256 bytes (S1503). Then, the divided files are transferred as items, and information related to the divided files (for example, the divided files, serial numbers in the divided files, and the like) is stored in the control information and transferred (S1504). On the other hand, when the file size is smaller than x × 256 bytes (no in S1502), the file is transferred as an item as usual (S1505).
Fig. 62 is an operation flow of the receiving apparatus in the case of using the slice counter.
First, the receiving apparatus acquires and analyzes control information related to file transfer such as a resource management table (S1601). Next, the receiving apparatus determines whether or not the desired item is a divided file (S1602). When the receiving apparatus determines that the desired file is a divided file (yes in S1602), the receiving apparatus acquires information for reconstructing the file, such as the divided file or an index of the divided file, from the control information (S1603). Then, the receiving apparatus acquires the items constituting the divided file, and reconstructs the original file (S1604). On the other hand, when the receiving apparatus determines that the desired file is not a divided file (no in S1602), the receiving apparatus acquires the file as in the normal case (S1605).
In short, the transmitting apparatus signals the packet sequence number of the divided data at the beginning of the file. In addition, the transmitting apparatus signals information that can determine the number of divided data. Alternatively, the transmission apparatus defines a division rule capable of determining the number of divided data. In addition, the transmitting apparatus does not use the slice counter and performs reservation or header compression.
When the packet sequence number of the data at the beginning of the file is transmitted by signaling, the reception apparatus determines the split data number and the split data number based on the packet sequence number of the split data at the beginning of the file and the packet sequence number of the MMTP packet.
From another viewpoint, the transmitting apparatus divides a file, and divides and transmits data for each divided file. The information (sequence number, number of splits, etc.) associated with the split file is signaled.
The receiving device determines the number of the divided data and the number of the divided data according to the fragment counter and the serial number of the divided file.
This makes it possible to uniquely identify the divided data number and the divided data. Further, the divided data number of the divided data can be specified at the time of receiving the divided data in the middle, so that the waiting time can be reduced and the memory can be reduced.
Further, by not using the slice counter, the configuration of the transmitter/receiver apparatus can reduce the amount of processing and improve the transmission efficiency.
Fig. 63 is a diagram showing a service configuration in a case where the same program is transmitted by a plurality of IP data streams. The following examples are shown here: data of a part (video/audio) of a program having a service ID of 2 is transmitted by an IP data stream using an MMT scheme, and data having the same service ID but different from the part of the program is transmitted by an IP data stream using an advanced BS data transfer scheme (in this example, a different file transfer protocol, but the same protocol may be used).
The transmitting device multiplexes the IP data so that the receiving device can ensure that the data composed of a plurality of IP data streams is ready before the decoding time.
The receiving apparatus performs processing based on the decoding time using data composed of a plurality of IP data streams, thereby realizing guaranteed receiver operation.
[ supplement: transmitting apparatus and receiving apparatus
As described above, the transmission device that transmits data without using the slice counter can be configured as shown in fig. 64. Further, the receiving apparatus for receiving data without using the slice counter can be configured as shown in fig. 65. Fig. 64 is a diagram showing an example of a specific configuration of a transmitting apparatus. Fig. 65 is a diagram showing an example of a specific configuration of a receiving apparatus.
The transmission device 500 includes a dividing unit 501, a configuration unit 502, and a transmission unit 503. The dividing unit 501, the constituting unit 502, and the transmitting unit 503 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The reception device 600 includes a reception unit 601, a determination unit 602, and a configuration unit 603. The receiving unit 601, the determining unit 602, and the configuring unit 603 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The components of the transmission apparatus 500 and the reception apparatus 600 will be described in detail in the description of the transmission method and the reception method, respectively.
First, a transmission method will be described with reference to fig. 66. Fig. 66 shows an operation flow (transmission method) of the transmission device.
First, the dividing unit 501 of the transmission device 500 divides data into a plurality of pieces of divided data (S1701).
Next, the configuration unit 502 of the transmission device 500 configures a plurality of packets by giving header information to each of the plurality of divided data and packetizing the plurality of divided data (S1702).
Then, the transmission unit 503 of the transmission device 500 transmits the plurality of packets thus configured (S1703). The transmission unit 503 transmits the divided data information and the invalidated slice counter value. The divided data information is information for specifying the number of divided data and the number of divided data. The divided data number is a number indicating that the divided data is the second divided data among a plurality of divided data. The divided data number is the number of a plurality of divided data.
This can reduce the processing amount of the transmission device 500.
Next, a reception method will be described with reference to fig. 67. Fig. 67 shows an operation flow of the receiving apparatus (receiving method).
First, the reception unit 601 of the reception apparatus 600 receives a plurality of packets (S1801).
Next, the determination unit 602 of the reception apparatus 600 determines whether or not the divided data information is acquired from the plurality of received packets (S1802).
When the determination unit 602 determines that the divided data information is acquired (yes in S1802), the configuration unit 603 of the reception device 600 configures data from the received packets without using the value of the slice counter included in the header information (S1803).
On the other hand, when the determination unit 602 determines that the divided data information has not been acquired (no in S1802), the configuration unit 603 may configure data from a plurality of received packets using a slice counter included in the header information (S1804).
This can reduce the processing amount of the receiving apparatus 600.
(embodiment 5)
[ summary ]
In embodiment 5, a transmission method of transport packets (TLV packets) in the case of storing NAL units in a NAL size format in a multiplex layer is described.
As described in embodiment 1, when NAL units of h.264 and h.265 are stored in a multiplex layer, there are two types of storage. One is a form called "byte stream format" in which a start code composed of a specific bit string is attached immediately before the NAL unit header. The other is a form called "NAL size format" in which a field indicating the size of a NAL unit is attached. The byte stream format is used in MPEG-2 systems, RTP, and the like, and the NAL size format is used in MP4, DASH using MP4, MMT, and the like.
In the byte stream format, the start code is composed of 3 bytes, and an arbitrary byte (byte having a value of 0) can be added.
On the other hand, in the NAL size format in the general MP4, size information is represented by any one of 1 byte, 2 bytes, and 4 bytes. This size information is represented by the length sizeminusone field in the HEVC sample entry. The case where the value of this field is "0" indicates 1 byte, the case where the value of this field is "1" indicates 2 bytes, and the case where the value of this field is "3" indicates 4 bytes.
Here, in ARIB STD-B60 "MMT-based media transport stream scheme in digital broadcasting" standardized in 7 months 2014, when a NAL unit is stored in a multiplex layer and the output of an HEVC encoder is a byte stream, a byte start code is removed and the size of a byte NAL unit represented by 32 bits (unsigned integer) is added as length information immediately before the NAL unit. MPU metadata including the HEVC sample entry is not transmitted, and the size information is fixed to 32 bits (4 bytes).
In the ARIB STD-B60 "MMT-based media transport stream scheme in digital broadcasting", a pre-decoding buffer of a video signal is defined as a CPB in a reception buffer model that a transmitting device considers at the time of transmission in order to guarantee a buffering operation in a receiving device.
However, there are the following problems. In CPB in the MPEG-2 system and HRD in HEVC, it is prespecified that a video signal is in a byte stream format. Therefore, for example, when the rate control of transport packets is performed on the premise of the byte stream format to which a start code of 3 bytes is added, there is a possibility that a receiving apparatus that receives transport packets of NAL size format to which a size region of 4 bytes is added cannot satisfy the reception buffer model in ARIB STD-B60. In addition, the reception buffer model in ARIB STD-B60 does not indicate a specific buffer size and extraction rate, and therefore it is difficult to guarantee a buffering operation in the reception apparatus.
In order to solve the above problem, a reception buffer model for ensuring a buffering operation in a receiver is defined as follows.
Fig. 68 shows a reception buffer model defined by ARIB STD-B60, particularly in the case of using only the broadcast transmission path.
The receive buffer model includes a TLV packet buffer (first buffer), an IP packet buffer (second buffer), an MMTP buffer (third buffer), and a pre-decode buffer (fourth buffer). Here, in the broadcast transmission path, a jitter cancellation buffer or a buffer for FEC is not required, and therefore, is omitted.
The TLV packet buffer receives TLV packets (transport packets) from the broadcast transmission path, converts IP packets composed of variable-length headers (IP packet header, full header at the time of IP packet compression, and compressed header at the time of IP packet compression) and variable-length payloads stored in the received TLV packets into IP packets (first packets) having fixed-length IP packet headers obtained by header expansion, and outputs the converted IP packets at a fixed bit rate.
The IP packet buffer converts the IP packet into an MMTP packet (second packet) having a header and a variable-length payload, and outputs the MMTP packet obtained by the conversion at a fixed bit rate. The IP packet buffer may also be combined with the MMTP buffer.
The MMTP buffer converts the output MMTP packet into NAL units, and outputs the NAL units obtained by the conversion at a fixed bit rate.
The pre-decoding buffer sequentially accumulates the output NAL units, generates an access unit from the accumulated NAL units, and outputs the generated access unit to the decoder at a timing of a decoding time corresponding to the access unit.
In the reception buffer model shown in fig. 68, the MMTP buffer and the pre-decode buffer, which are buffers other than the TLV packet buffer and the IP packet buffer at the previous stage, are characterized by following the reception buffer model in the MPEG-2 TS.
For example, the MMTP buffer for video (MMTP B1) is composed of buffers corresponding to a Transport Buffer (TB) and a Multiplexing Buffer (MB) in MPEG-2 TS. The MMTP buffer for audio (MMTP Bn) is constituted by a buffer corresponding to the Transmission Buffer (TB) in the MPEG-2 TS.
The buffer size of the transmission buffer is set to a fixed value as in the MPEG-2 TS. For example, the MTU size is n times (n may be a decimal number or an integer and is 1 or more).
In addition, the MMTP packet size is specified such that the overhead rate of the MMTP packet header is smaller than the overhead rate of the PES packet header. Thus, the extraction rates RX1, RXn, RXs of the transmission buffer in the MPEG-2TS can be applied as they are from the extraction rate of the transmission buffer.
The size of the multiplexing buffer and the extraction rate are set to the MB size and RBX1 in the MPEG-2TS, respectively.
In addition to the above reception buffer model, the following restriction is also provided to solve the problem.
The HRD specification for HEVC presupposes the byte stream format, and MMT is NAL size format with a size region of 4 bytes appended to the beginning of the NAL unit. Therefore, in NAL size form at the time of encoding, rate control is performed to satisfy HDR.
That is, the transmitting apparatus performs rate control of the transport packet based on the reception buffer model and the restriction.
In the receiving apparatus, by performing the reception processing using the signal, the decoding operation can be performed without underflow or overflow.
The size region at the head of the NAL unit is considered even if the size region at the head of the NAL unit is not 4 bytes, and rate control is performed so as to satisfy the HRD.
The extraction rate of the TLV packet buffer (bit rate when the TLV packet buffer outputs the IP packet) is set in consideration of the transfer rate after the IP header extension.
That is, a TLV packet having a variable data size is input, and after removing the TLV header and expanding (restoring) the IP header, the transmission rate of the output IP packet is considered. In other words, the amount of increase and decrease of the head is considered for the input transfer rate.
Specifically, since the data size is variable, packets subjected to IP header compression and packets not subjected to IP header compression coexist, and the size of the IP header differs depending on the packet type such as IPv4 or IPv6, the transfer rate of the output IP packet is not unique. Therefore, the average packet length of the variable-length data size is determined, and the transfer rate of the IP packet output according to the TLV packet is determined.
Here, in order to specify the maximum transmission rate after the IP header is expanded, the transmission rate is determined assuming that the IP header is always compressed.
When the packet types of IPv4 and IPv6 are mixed or when the packet types are not defined separately, the transmission rate is determined assuming that the header size is large and the increase rate after header expansion is large for IPv6 packets.
For example, when all the IP packets stored in the S, TLV packets having the average packet length of the TLV packets input to the TLV packet buffer are IPv6 packets and header compression is performed, the maximum output transfer rate after TLV header removal and IP header expansion is expressed by the following mathematical formula (1):
[ mathematical formula 1]
Input rate x { S/(S + IPv6 head compression amount) } … … (1)
More specifically, the average packet length S of the TLV packet is set with reference to equation (2):
[ mathematical formula 2]
S-0.75 × 1500(1500 is assumed to be the maximum MTU size) … … (2)
And the IPv6 header compression amount is set to the following formula (3):
[ mathematical formula 3]
IPv6 header compression TLV header length-IPv 6 header length-UDP header length
=3-40-8 ……(3)
The maximum output transfer rate after removal of the TLV header and expansion of the IP header in the case of (4) is as shown in equation (4):
[ mathematical formula 4]
Input rate x 1.0417 ≈ input rate x 1.05 … … (4)
Fig. 69 is a diagram showing an example in which a plurality of data units are collectively stored in one payload.
In the MMT scheme, when data units are grouped, as shown in fig. 69, a data unit length and a data unit header are added before the data units.
However, for example, when a video signal in the NAL size format is stored as one data unit, as shown in fig. 70, 2 fields indicating the size are provided for one data unit, and the fields are repeated as information. Fig. 70 is a diagram showing an example of a case where a plurality of data units are collectively stored in one payload and a video signal in NAL size format is regarded as one data unit. Specifically, the size area at the head of the NAL size format (hereinafter referred to as "size area") and the data unit length field located before the data unit header in the MMTP payload header are both fields indicating the size and are duplicated as information. For example, in the case where the length of the NAL unit is L bytes, L bytes are indicated in the size region, and L bytes + "length of the size region" (bytes) are indicated in the data unit length field. Although the values indicated in the size field and the data unit length field do not completely match, the values can be said to be overlapping because one value can easily calculate the other value.
As described above, there are problems as follows: when data including size information of data therein is stored as a data unit and a plurality of the data units are collectively stored in one payload, the size information is duplicated, and therefore, the overhead rate is high and the transmission efficiency is low.
Therefore, when the transmitting apparatus stores data including size information of the data therein as data units and stores a plurality of the data units in one payload in a lump, it is conceivable to store the data units as shown in fig. 71 and 72.
As shown in fig. 71, it is conceivable that NAL units including size fields are stored as data units, and the MMTP payload header does not indicate the data unit length included in the past. Fig. 71 is a diagram showing the structure of the payload of an MMTP packet whose data unit length is not shown.
As shown in fig. 72, a flag indicating whether header compression is performed (that is, indicating whether the data unit length is indicated) and information indicating the length of the size area may be newly stored in the header. The position where the flag and the information indicating the length of the size area are stored may be indicated in units of data cells by a data cell header or the like, or may be indicated in units (packet units) in which a plurality of data cells are grouped. Fig. 72 shows an example in which a flag or information indicating the length of a size area is stored in an extended area provided for each packet unit. The storage location of the newly indicated information is not limited to this, and may be an MMTP payload header, an MMTP header, or control information.
On the receiving side, when the flag indicating whether the data unit length is compressed indicates that the data unit length is compressed, the length information of the size area inside the data unit is acquired, and the size area is acquired based on the length information of the size area, whereby the data unit length can be calculated using the acquired length information of the size area and the size area.
With the above method, the data amount can be reduced on the transmitting side, and the transmission efficiency can be improved.
However, the overhead may be reduced by reducing the size area without reducing the data unit length. In the case of reducing the size area, information indicating whether the size area is reduced or not and information indicating the length of the data unit length field may be stored.
Wherein the MMTP payload header also contains length information.
When NAL units including size fields are stored as data units, the payload size field in the MMTP payload header may be reduced regardless of whether they are grouped together.
In addition, when storing data not including a size area as a data unit, if the data unit is summarized and the data unit length is expressed, the payload size area in the MMTP payload header can be reduced.
When the payload size region is reduced, a flag indicating whether to reduce, length information of a reduced size field, or length information of an unreduced size field may be indicated in the same manner as described above.
Fig. 73 shows an operation flow of the receiving apparatus.
As described above, the transmitting apparatus stores NAL units including size fields as data units, and the data unit length included in the MMTP payload header is not indicated in the MMTP packet.
Hereinafter, a case will be described as an example in which a flag indicating whether or not the data unit length is indicated, and length information of the size area are indicated in the MMTP packet.
The receiving apparatus determines whether the data unit includes a size region and the data unit length is reduced based on the information transmitted from the transmitting side (S1901).
When it is determined that the data cell length is reduced (yes in S1902), the length information of the size region inside the data cell is acquired, and then the size region inside the data cell is analyzed to acquire the data cell length by calculation (S1903).
On the other hand, when it is determined that the data cell length has not been reduced (no in S1902), the data cell length is calculated from either the data cell length or the size region inside the data cell as usual (S1904).
However, when the flag indicating whether the data unit length is reduced or not and the length information of the size area are known in advance by the receiving apparatus, the flag and the length information may not be transmitted. In this case, the receiving apparatus performs the processing shown in fig. 73 based on predetermined information.
[ supplement: transmitting apparatus and receiving apparatus
As described above, the transmitting apparatus that performs rate control so as to satisfy the predetermined reception buffer model at the time of encoding can also be configured as shown in fig. 74. Further, the receiving apparatus that receives and decodes the transport packet transmitted from the transmitting apparatus can also be configured as shown in fig. 75. Fig. 74 is a diagram showing an example of a specific configuration of a transmitting apparatus. Fig. 75 is a diagram showing an example of a specific configuration of a receiving apparatus.
The transmission device 700 includes a generation unit 701 and a transmission unit 702. The generation unit 701 and the transmission unit 702 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The reception device 800 includes a reception unit 801, a first buffer 802, a second buffer 803, a third buffer 804, a fourth buffer 805, and a decoding unit 806. The receiving unit 801, the first buffer 802, the second buffer 803, the third buffer 804, the fourth buffer 805, and the decoding unit (decoder) 806 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The components of the transmission apparatus 700 and the reception apparatus 800 are explained in detail in the explanation of the transmission method and the reception method, respectively.
First, a transmission method will be described with reference to fig. 76. Fig. 76 shows an operation flow (transmission method) of the transmission device.
First, the generating unit 701 of the transmitting apparatus 700 generates a coded stream by performing rate control so as to satisfy the specification of a reception buffer model predetermined to guarantee the buffering operation of the receiving apparatus (S2001).
Next, the transmission unit 702 of the transmission device 700 packetizes the generated bit stream, and transmits a transport packet obtained by the packetization (S2002).
The reception buffer model used in the transmission device 700 has the configurations of the first to fourth buffers 802 to 805 having the configuration of the reception device 800, and therefore, the description thereof is omitted.
Thus, when data transfer is performed using a scheme such as MMT, the transmission device 700 can ensure the buffering operation of the reception device 800.
Next, a reception method will be described with reference to fig. 77. Fig. 77 shows an operation flow of the receiving apparatus (receiving method).
First, the reception unit 801 of the reception device 800 receives a transport packet including a fixed-length packet header and a variable-length payload (S2101).
Next, the first buffer 802 of the receiving apparatus 800 converts the packet composed of the variable-length header and the variable-length payload stored in the received transport packet into a first packet having a fixed-length header expanded by the header, and outputs the first packet obtained by the conversion at a fixed bit rate (S2102).
Next, the second buffer 803 of the reception apparatus 800 converts the first packet obtained by the conversion into a second packet composed of a header and a variable-length payload, and outputs the second packet obtained by the conversion at a fixed bit rate (S2103).
Next, the third buffer 804 of the reception apparatus 800 converts the output second packet into an NAL unit, and outputs the NAL unit obtained by the conversion at a fixed bit rate (S2104).
Next, the fourth buffer 805 of the reception apparatus 800 sequentially accumulates the outputted NAL units, generates an access unit from the accumulated NAL units, and outputs the generated access unit to the decoder at the timing of the decoding time corresponding to the access unit (S2105).
Then, the decoding unit 806 of the reception apparatus 800 decodes the access unit output from the fourth buffer (S2106).
In this way, the reception apparatus 800 can perform decoding operation without underflow or overflow.
(embodiment mode 6)
[ summary ]
In embodiment 6, a description will be given of a transmission method and a reception method in the case where leap second adjustment is performed on time information serving as a reference of a reference clock in an MMT/TLV transmission method.
FIG. 78 is a diagram showing a protocol stack of the MMT/TLV scheme defined by ARIB STD-B60.
In the MMT scheme, data such as video and audio is stored in a packet for each first data Unit of a plurality of MPUs (Media Presentation Unit) or MFUs (Media Fragment Unit), and an MMTP packet (MMTP packet) is generated as a predetermined packet by giving an MMTP header. Further, control information such as a control message in MMTP is also given to the MMTP header, and an MMTP packet that is a predetermined packet is generated. The MMTP packet is provided with a field for storing a 32-bit short format NTP (Network Time Protocol: specified in IETF RFC 5905), and can be used for QoS control of a communication line and the like.
Further, a reference clock of a transmitting side (transmitting apparatus) is synchronized with a 64-bit long format NTP defined in RFC 5905, and a Time Stamp such as a PTS (Presentation Time Stamp) or a DTS (Decode Time Stamp) is given to a synchronous medium based on the synchronized reference clock. Further, the reference clock information of the transmitting side is transmitted to the receiving side, and the receiving device generates a system clock in the receiving device based on the reference clock information received from the transmitting side.
Specifically, the PTS and DTS are stored in an MPU timestamp descriptor or an MPU extended timestamp descriptor, which is control information of the MMTP, in an MP table for each resource (asset), and are transmitted as a control message after being packetized into an MMTP packet.
The data packetized into the MMTP packet is given a UDP header or an IP header and encapsulated (encapsulated) into an IP packet. In this case, the IP header or UDP header includes a set of packets having the same source IP address, destination IP address, source port number, destination port number, and protocol type as an IP data flow. In addition, since the header of an IP packet of the same IP data flow is redundant, header compression is performed on a part of the IP packet.
As the reference clock information, 64-bit NTP time stamps are stored in an NTP packet and in an IP packet. In this case, in the IP packet storing the NTP packet, the source IP address, the destination IP address, the source port number, the destination port number, and the protocol type are fixed values, and the header of the IP packet is not compressed.
Fig. 79 is a diagram showing the structure of a TLV packet.
As shown in fig. 79, the TLV packet can include transmission control Information such as an IP packet, a compressed IP packet, an AMT (Address Map Table), or an NIT (Network Information Table) as data, and these data are identified by using a data type of 8 bits. In the TLV packet, a 16-bit field is used to indicate the data length (byte unit), and the data value is stored thereafter. In addition, TLV packets have 1-byte header information before the data type, which is deposited in a header area totaling 4 bytes. The TLV packet is mapped to a Transmission slot in the high BS Transmission scheme, and mapping information is stored in TMCC (Transmission and Multiplexing Configuration Control) Control information.
Fig. 80 is a diagram showing an example of a block diagram of a receiving apparatus.
In the receiving apparatus, first, the TLV packet is extracted by decoding and correcting transmission path encoded data in a demodulation means with respect to the broadcast signal received by the tuner. Then, in a TLV/IP Demultiplexing (DEMUX) mechanism, TLV demultiplexing processing or IP demultiplexing processing is performed. The demultiplexing process of the TLVs performs a process corresponding to the data type of the TLV packets. For example, in the case where the TLV packet has a compressed IP packet, the compressed header of the compressed IP packet is recovered. In IP multicast, processing such as header analysis of an IP packet or UDP packet is performed to extract an MMTP packet and an NTP packet.
The NTP clock generating means reproduces an NTP clock from the extracted NTP packet. In MMTP demultiplexing, filtering processing of components such as video and audio and control information is performed based on the packet ID stored in the extracted MMTP packet header. A time stamp descriptor stored in an MP table is acquired from a control information acquisition means, and PTS and DTS are calculated for each access unit by a PTS/DTS calculation means. The time stamp descriptor includes two descriptors, an MPU time stamp descriptor and an MPU extended time stamp descriptor.
The access unit playback means converts the video, audio, and the like filtered from the MMTP packet into data of a unit to be presented. The unit of data to be presented is, specifically, NAL units, access units, audio frames, presentation units of letters, and the like of video signals. The decoding presentation means decodes and presents the access unit at a timing when the PTS/DTS of the access unit matches, based on the reference time information of the NTP clock.
In addition, the structure of the receiving apparatus is not limited thereto.
The time stamp descriptor is explained next.
Fig. 81 is a diagram for explaining a time stamp descriptor.
The PTS and DTS are stored in an MPU timestamp descriptor of first control information, which is control information of the MMT, or an MPU extended timestamp descriptor of second control information, and are stored in an MP table for each resource, and transmitted after being packetized into an MMTP packet as a control message.
FIG. 81 (a) is a diagram showing the structure of an MPU time stamp descriptor defined by ARIB STD-B60. In the MPU timestamp descriptor, presentation time information (first time information) indicating a PTS (absolute value indicated by 64-bit NTP) of an AU whose presentation order is the top (first) (hereinafter referred to as "top AU") among a plurality of Access Units (AUs) stored in the MPU as a second data unit is stored for each of the MPUs. That is, presentation time information of the MPU given to the MPU is stored in control information of the MMTP packet and transmitted.
Fig. 81 (b) shows the structure of the MPU extended timestamp descriptor. The MPU extended time stamp descriptor stores information for calculating PTS and DTS of AUs included in each of the MPUs. The MPU extension timestamp descriptor includes relative information (second time information) with respect to the PTS of the leading AU of the MPU stored in the MPU timestamp descriptor, and the PTS and DTS of each of the plurality of AUs included in the MPU can be calculated based on both the MPU timestamp descriptor and the MPU extension timestamp descriptor. That is, the PTS and DTS of an AU other than the start AU included in the MPU can be calculated based on the PTS of the start AU stored in the MPU timestamp descriptor and the relative information stored in the MPU extended timestamp descriptor.
In other words, the second time information is time information for calculating the relativity of the PTS or DTS of each of the plurality of AUs together with the first time information. That is, the second time information is information indicating PTS or DTS of each of the plurality of AUs together with the first time information.
NTP is reference Time information based on Coordinated Universal Time (UTC). The UTC performs leap second adjustment (hereinafter referred to as "leap second adjustment") to adjust the difference between the time of the earth rotation and the astronomical time based on the earth rotation speed. Specifically, leap second adjustment is performed at 9 am in japan, and is adjustment for inserting or deleting 1 second.
Fig. 82 is a diagram for explaining leap second adjustment.
Fig. 82 (a) shows an example of leap seconds inserted during the japanese time. As shown in fig. 82 (a), in inserting leap seconds, the timing that originally became 9:00:00 after 8:59:59 in japanese time becomes 8:59:59, and 8:59:59 is repeated twice.
Fig. 82 (b) shows an example of leap second deletion in japanese time. As shown in fig. 82 (b), in the leap second deletion, the leap second is deleted during 1 second, in which the timing that originally became 8:59:59 after 8:59:58 in japan time becomes 9:00:00 and 8:59:59 seconds.
The NTP packet stores a 2-bit leaf _ indicator in addition to a 64-bit time stamp. The leap _ indicator is a flag for notifying that leap second adjustment is performed in advance, and indicates that leap second is inserted when the leap _ indicator is 1, and indicates that leap second is deleted when the leap _ indicator is 2. The notification may be performed in advance from the beginning of the month when leap seconds are performed, 24 hours before, or at any time. In addition, the leap _ indicator becomes 0 at the time (9:00:00) when leap second adjustment ends. For example, when the advance notice is given 24 hours ago, the leap _ indicator is indicated as "1" or "2" during the period from the time 9:00 of the day before leap second adjustment is performed to the time immediately before leap second adjustment is performed on the day of leap second adjustment (i.e., the time 8:59:59 seconds of the 1 st time when leap second is inserted, and the time 8:59:58 seconds when leap second is deleted).
Next, the problem of leap second adjustment will be described.
Fig. 83 is a diagram showing a relationship between NTP time, MPU time stamp, and MPU presentation timing. The NTP time is a time represented by NTP. The MPU timestamp is a timestamp indicating a PTS of the leading AU in the MPU. The MPU presentation timing is a timing at which the receiving apparatus is to present the MPU according to the MPU timestamp. Specifically, (a) to (c) of fig. 83 are diagrams showing the relationship between the NTP time, the MPU time stamp, and the MPU presentation time in the case where leap second adjustment does not occur, the case where leap second is inserted, and the case where leap second is deleted, respectively.
Here, a case will be described as an example where NTP time (reference clock) on the transmission side is synchronized with the NTP server, and NTP time (system clock) on the reception side is synchronized with NTP time on the transmission side. In this case, the receiving device performs reproduction based on the time stamp stored in the NTP packet transmitted from the transmitting side. In this case, since the NTP time of the transmitting side and the NTP time of the receiving side are synchronized with the NTP server, the adjustment of ± 1 second is performed when the leap second is adjusted. Note that the NTP time in fig. 83 is common to the NTP time on the transmission side and the NTP time on the reception side. Note that the description will be made as to the case where there is no transmission delay.
The MPU time stamp in fig. 83 indicates a time stamp of the head AU in presentation order among a plurality of AUs included in each of the MPUs, and is generated (set) based on the NTP time indicated by the base end of the arrow. Specifically, the presentation time information of the MPU is generated by adding a predetermined time (for example, 1.5 seconds in fig. 83) to the NTP time, which is the reference time information at the timing of generating the presentation time information of the MPU. The generated MPU timestamp is stored in an MPU timestamp descriptor.
In the receiving apparatus, the MPU is presented at an MPU presentation time based on the time stamp stored in the MPU time stamp descriptor.
In fig. 83, the playback time of 1 MPU is described as 1 second, but the playback time of 1 MPU may be other playback time, and may be 0.5 second or 0.1 second, for example.
In the example of fig. 83 (a), the receiving apparatus can sequentially present the MPUs #1 to #5 based on the time stamps stored in the MPU time stamp descriptor.
However, in fig. 83 (b), leap seconds are inserted, and thus the presentation times of MPU #2 and MPU #3 overlap. Therefore, if the receiving apparatus presents the MPUs based on the time stamps stored in the MPU time stamp descriptors, there are 2 MPUs presented in the same 9:00:00 second period, and it is not possible to determine which of the 2 MPUs should be presented. Further, since the MPU presentation time (8:59:59) indicated by the time stamp of MPU #1 exists 2 times, the receiving apparatus cannot determine which MPU presentation time should be presented among the 2 MPU presentation times.
In fig. 83 (c), leap seconds are deleted, and the MPU presentation time (8:59:59) indicated by the MPU timestamp of MPU #3 does not exist in the NTP time, so the receiving apparatus cannot present MPU # 3.
The receiving apparatus can solve the above-described problem without performing the decoding process and the presentation process of the MPU based on the time stamp. However, it is difficult for a receiving apparatus that performs processing based on the time stamp to perform different processing (processing not based on the time stamp) only when leap seconds occur.
Next, a method for solving the problem of leap second adjustment by correcting the time stamp on the transmission side will be described.
Fig. 84 is a diagram for explaining a method of correcting a time stamp on the transmission side. Specifically, (a) in fig. 84 shows an example of inserting leap seconds, and (b) in fig. 84 shows an example of deleting leap seconds.
First, the case of inserting leap seconds will be described.
As shown in fig. 84 (a), when leap seconds are inserted, the time immediately before leap seconds are inserted (i.e., 8:59:59 seconds before the 1 st time of NTP time) is defined as the a area, and the time after leap seconds are inserted (i.e., 8:59:59 seconds after the 2 nd time of NTP time) is defined as the B area. The a region and the B region are temporal regions, and are time periods or periods. The MPU timestamp in fig. 84 (a) is a timestamp generated (set) based on the time of the NTP at the timing when the MPU timestamp is given, as in the timestamp described in fig. 83.
Specifically, a method of correcting the time stamp on the transmission side when leap seconds are inserted will be described.
The following processing is performed on the transmission side (transmission apparatus).
1. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in (a) in fig. 84) is included in the area a and the MPU timestamp (value of the MPU timestamp before correction) is 9:00:00 or later, correction is performed for-1 second by subtracting 1 second from the MPU timestamp, and the corrected MPU timestamp is stored in the MPU timestamp descriptor. That is, when the MPU timestamp is generated based on the NTP time included in the area a and indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second. Here, "9: 00: 00" refers to a time (i.e., a time calculated by adding 9 hours to the UTC time) obtained by making the time that is the reference for leap second adjustment correspond to the japanese time. Further, correction information indicating that another correction is performed is transmitted to the receiving apparatus.
2. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in fig. 84 (a)) is in the B region, the MPU timestamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the B area, the MPU timestamp is not corrected.
The receiving apparatus presents the MPU based on the MPU timestamp and correction information indicating whether the MPU timestamp has been corrected (that is, whether information indicating that correction has been performed is included).
When the reception device determines that the MPU timestamp is not corrected (that is, when it determines that information indicating that correction is performed is not included), the reception device presents the MPU at a time when the timestamp stored in the MPU timestamp descriptor matches the NTP time (including both before and after correction) of the reception device. That is, when the MPU timestamp is the MPU timestamp stored in the MPU timestamp descriptor of the MPU transmitted before the corrected MPU, the MPU is presented at a timing when the MPU timestamp matches the NTP time before the insertion of the leap second (i.e., 8:59:59 seconds before the 1 st time). If the MPU timestamp stored in the MPU timestamp descriptor of the received MPU is a timestamp transmitted later than the corrected MPU timestamp, the MPU is presented at a timing that matches the NTP time after insertion of the leap second (i.e., 8:59:59 seconds after the 2 nd time).
When the MPU timestamp stored in the MPU timestamp descriptor of the received MPU is corrected, the receiving apparatus presents the MPU based on the NTP time after the leap second is inserted into the receiving apparatus (i.e., after 8:59:59 seconds of the 2 nd time).
The information indicating that the MPU timestamp value is corrected is transferred by storing it in a control message, a descriptor, a table, MPU metadata, MF metadata, an MMTP packet header, and the like.
Next, the leap second deletion will be described.
As shown in fig. 84 (b), when leap seconds are deleted, the time immediately before leap seconds are deleted (i.e., immediately before 9:00:00 of the NTP time) is the C area, and the time after leap seconds are deleted (i.e., immediately after 9:00:00 of the NTP time) is the D area. The C region and the D region are temporal regions, and are time periods or periods. The MPU timestamp in fig. 84 (b) is a timestamp generated (set) based on the NTP time at the timing when the MPU timestamp is given, as in the timestamp described in fig. 83.
Specifically, a method of correcting the time stamp on the transmission side when leap seconds are deleted will be described.
The following processing is performed on the transmission side (transmission apparatus).
1. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in (b) in fig. 84) is included in the C region and the MPU timestamp (value of the MPU timestamp before correction) indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second by adding 1 second, and the corrected MPU timestamp is stored in the MPU timestamp descriptor. That is, when the MPU timestamp is generated based on the NTP time included in the C area and indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second. Here, "8: 59: 59" refers to a time obtained by subtracting-1 second from a time obtained by making the time that becomes the reference for leap second adjustment correspond to the japanese time (i.e., a time calculated by adding 9 hours to the UTC time). Further, correction information indicating that another correction is performed is transmitted to the receiving apparatus. In this case, the correction information may not necessarily be transmitted.
2. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in fig. 84 (b)) is the D region, the MPU timestamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the D area, the MPU timestamp is not corrected.
In the reception apparatus, the MPU is presented based on the MPU timestamp. Further, if there is correction information indicating whether or not the MPU timestamp has been corrected, the MPU may be presented based on the MPU timestamp and the correction information.
By the above processing, even when the NTP time is adjusted by leap second, the receiving apparatus can present a normal MPU using the MPU timestamp stored in the MPU timestamp descriptor.
Further, the timing of assigning the MPU timestamp may be signaled in the a, B, C, or D regions, and notified to the receiving side. That is, the MPU timestamp may be signaled to notify the receiving side of whether it is generated based on the NTP time included in the a-region, generated based on the NTP time included in the B-region, generated based on the NTP time included in the C-region, or generated based on the NTP time included in the D-region. In other words, the identification information indicating whether the MPU timestamp (presentation time) is generated based on the reference time information before leap second adjustment (NTP time) may be transmitted. The identification information is given based on the leap _ indicator included in the NTP packet, and is therefore information indicating whether the time is the MPU timestamp set to the time from 9:00 of the day before leap second adjustment was performed to the time immediately before leap second adjustment was performed on the day before leap second adjustment was performed (i.e., the time of 8:59:59 seconds for the 1 st time when leap seconds were inserted, and the time of 8:59:58 seconds for the leap seconds were deleted). That is, the identification information is information indicating whether the MPU timestamp is generated based on the NTP time from a time earlier than the time immediately before leap second adjustment by a predetermined period (for example, 24 hours) to the immediately preceding time.
Next, a method for solving the problem of the leap second adjustment by correcting the time stamp in the receiving apparatus will be described.
Fig. 85 is a diagram for explaining a correction method of correcting a time stamp in a receiving apparatus. Specifically, (a) in fig. 85 shows an example of inserting leap seconds, and (b) in fig. 85 shows an example of deleting leap seconds.
First, the case of inserting leap seconds will be described.
As shown in fig. 85 (a), when inserting leap seconds, the time immediately before inserting leap seconds (i.e., 8:59:59 seconds before the 1 st time of the NTP time) is defined as the a area, and the time after inserting leap seconds (i.e., 8:59:59 seconds after the 2 nd time of the NTP time) is defined as the B area, as in fig. 84 (a). The a region and the B region are temporal regions, and are time periods or periods. The MPU timestamp in fig. 85 (a) is a timestamp generated (set) based on the time of the NTP at the timing when the MPU timestamp is given, as in the timestamp described in fig. 83.
Specifically, a method of correcting a time stamp in a reception device when leap seconds are inserted will be described.
The following processing is performed on the transmission side (transmission apparatus).
The MPU timestamp generated is not corrected, and the MPU timestamp is stored in an MPU timestamp descriptor and transmitted to the receiving apparatus.
Information indicating whether the timing of the time stamp given to the MPU is in the a region or the B region is transmitted to the receiving apparatus as the identification information. That is, identification information indicating whether the MPU timestamp is generated based on the NTP time included in the a area or the NTP time included in the B area is transmitted to the receiving apparatus.
The receiving apparatus performs the following processing.
In the receiving apparatus, the MPU time stamp is corrected based on the MPU time stamp and identification information indicating whether the time of the time stamp given to the MPU is in the A area or the B area.
Specifically, the following processing is performed.
1. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in (a) in fig. 85) is included in the area a and the MPU timestamp (value of the MPU timestamp before correction) is 9:00:00 or less, correction is performed for-1 second by subtracting 1 second from the MPU timestamp. That is, when the MPU timestamp is generated based on the NTP time included in the area a and indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second. Here, "9: 00: 00" refers to a time (i.e., a time calculated by adding 9 hours to the UTC time) obtained by making the time that is the reference for leap second adjustment correspond to the japanese time.
2. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in fig. 85 (a)) is in the B region, the MPU timestamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the B area, the MPU timestamp is not corrected.
When the MPU time stamp is not corrected, the MPU is presented at a time when the MPU time stamp stored in the MPU time stamp descriptor matches the NTP time (including before and after correction) of the receiving apparatus.
That is, in the case of an MPU timestamp received before the corrected MPU timestamp, the MPU is presented at a timing that matches the NTP time before leap second insertion (i.e., 8:59:59 seconds before the 1 st time). In the case of an MPU timestamp received later than the corrected MPU timestamp, the MPU is presented at a timing that matches the NTP time after insertion of leap seconds (8: 59:59 seconds after the 2 nd time).
When correcting the MPU timestamp, the corrected MPU timestamp is presented to the MPU based on the NTP time after leap seconds are inserted in the receiving device (i.e., 8:59:59 seconds after the 2 nd time).
The transmitting side stores identification information indicating whether the timing of the MPU time stamp is assigned to the a area or the B area in a control message, a descriptor, a table, MPU metadata, MF metadata, an MMTP header, and the like, and transmits the identification information.
Next, the leap second deletion will be described.
As shown in fig. 85 (b), when leap seconds are deleted, the time immediately before leap seconds are deleted (i.e., immediately before 9:00:00 of the NTP time) is defined as the C area, and the time after leap seconds are deleted (i.e., immediately after 9:00:00 of the NTP time) is defined as the D area, as in fig. 84 (b). The C region and the D region are temporal regions, and are time periods or periods. The MPU timestamp in fig. 85 (b) is a timestamp generated (set) based on the time of the NTP at the timing when the MPU timestamp is given, as in the timestamp described in fig. 83.
Specifically, a method of correcting a time stamp in a reception device when leap seconds are deleted will be described.
The following processing is performed on the transmission side (transmission apparatus).
The MPU timestamp generated is not corrected, and the MPU timestamp is stored in an MPU timestamp descriptor and transmitted to the receiving apparatus.
Information indicating whether the timing of the time stamp given to the MPU is in the C region or the D region is transmitted to the receiving apparatus as the identification information. That is, identification information indicating whether the MPU timestamp is generated based on the NTP time included in the C area or the D area is transmitted to the receiving apparatus.
The receiving apparatus performs the following processing.
In the receiving apparatus, the MPU time stamp is corrected based on the MPU time stamp and identification information indicating whether the timing of the time stamp given to the MPU is in the C region or the D region.
Specifically, the following processing is performed.
1. When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in (b) in fig. 85) is included in the C region and the MPU timestamp (the value of the MPU timestamp before correction) indicates 8:59:59 or later, the +1 second correction of adding 1 second to the MPU timestamp value is performed. That is, when the MPU timestamp is generated based on the NTP time included in the C area and indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second. Here, "8: 59: 59" refers to a time obtained by subtracting-1 second from a time obtained by making the time that becomes the reference for leap second adjustment correspond to the japanese time (i.e., a time calculated by adding 9 hours to the UTC time).
2. When the timing at which the MPU time stamp is given (timing indicated by the base end of the arrow in fig. 85 (b)) is the D region, the MPU time stamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the D area, the MPU timestamp is not corrected.
The receiving device presents the MPU based on the MPU time stamp and the corrected MPU time stamp.
By the above processing, even when the NTP time is adjusted by leap second, the receiving apparatus can present a normal MPU using the MPU timestamp stored in the MPU timestamp descriptor.
In this case, similarly to the case where the MPU timestamp is corrected on the transmitting side as described with reference to fig. 84, the timing of giving the MPU timestamp may be signaled in the a region, the B region, the C region, or the D region, and may be notified to the receiving side. The details of the notification are the same, and therefore, the description thereof is omitted.
The additional information (identification information) such as information indicating whether or not the MPU timestamp is corrected and information indicating whether the timing of adding the MPU timestamp is in the a region, the B region, the C region, or the D region described in fig. 84 and 85 may be effective when the leap _ indicator of the NTP packet indicates deletion (leap _ indicator ═ 2) or insertion (leap _ indicator ═ 1) of leap seconds. The setting may be made valid from a predetermined arbitrary time (for example, 3 seconds before leap second adjustment) or may be made dynamically valid.
The timing at which the validity period of the additional information ends or the timing at which the signaling of the additional information ends may correspond to the leaf _ indicator of the NTP packet, may be valid from a predetermined arbitrary time (for example, 3 seconds before leap second adjustment), or may be dynamically valid.
Further, since the 32-bit time stamp stored in the MMTP packet, the time stamp information stored in the TMCC, and the like are also generated and given based on the NTP time, the same problem arises. Therefore, even in the case of a 32-bit time stamp, time stamp information stored in TMCC, or the like, the time stamp can be corrected by the same method as in fig. 84 and 85, and the receiving apparatus can perform processing based on the time stamp. For example, when the above-described additional information corresponding to the 32-bit timestamp stored in the MMTP packet is stored, the additional information may be shown using an extension area of the MMTP packet header. In this case, the indication is additional information in the extension type of the multi-head type. In addition, the time when the leaf _ indicator is set up may also represent additional information using a part of bits among 32-bit or 64-bit time stamps.
In the example of fig. 84, the corrected MPU timestamp is stored in the MPU timestamp descriptor, but both the pre-correction and post-correction MPU timestamps (i.e., uncorrected MPU timestamps and corrected MPU timestamps) may be transmitted to the receiving apparatus. For example, the MPU timestamp descriptor before correction and the MPU timestamp descriptor after correction may be stored in the same MPU timestamp descriptor, or may be stored in 2 MPU timestamp descriptors. In this case, whether the MPU timestamp is before correction or after correction may be identified based on the arrangement order of the 2 MPU timestamp descriptors, the description order in the MPU timestamp descriptors, or the like. Further, the MPU timestamp descriptor may always store the MPU timestamp before correction, and when correction is performed, the corrected timestamp may be stored in the MPU extended timestamp descriptor.
In the present embodiment, the time of the japanese time (9 am) is described as an example of the NTP time, but the NTP time is not limited to the time of the japanese time. Leap second adjustments are corrected based on UTC time, all at once around the world. The japanese time is 9 hours faster than the UTC time, and is represented by a value of time (+9) relative to the UTC time. In this manner, a time adjusted to a time different according to the time difference corresponding to the position may be used.
As described above, by correcting the time stamp in the transmitting side or the receiving apparatus based on the information indicating the timing at which the time stamp is given, it is possible to realize normal reception processing using the time stamp.
In addition, although a reception device that continues decoding processing from a time sufficiently earlier than the leap second adjustment time may realize decoding processing and presentation processing that do not use a time stamp, a reception device that selects a station immediately before the leap second adjustment time cannot determine a time stamp, and may not present the time stamp until after the leap second adjustment is completed. In this case, by using the correction method of the present embodiment, the reception process using the time stamp can be realized, and channel selection can be performed even immediately before the leap second adjustment time.
Fig. 86 shows an operation flow of the transmitting side (transmitting apparatus) in the case where the transmitting side (transmitting apparatus) corrects the MPU timestamp described in fig. 84, and fig. 87 shows an operation flow of the receiving apparatus in this case.
First, an operation flow of the transmission side (transmission device) will be described with reference to fig. 86.
When leap seconds are inserted or deleted, it is determined whether the timing of the MPU time stamp is in the a, B, C, or D areas (S2201). Note that the case where leap seconds are not inserted or deleted is not shown.
Here, the a region to the D region are defined as follows.
And (2) area A: just before leap second insertion (8: 59:59 seconds before 1 st time)
And a B region: leap second inserted later (8: 59:59 after 2 nd time)
And a C region: leap second delete time immediately before (9:00:00 before)
And D region: leap second delete later time (after 9:00: 00)
If it is determined in step S2201 that the area a is present and the MPU timestamp indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second, and the corrected MPU timestamp is stored in the MPU timestamp descriptor (S2202).
Then, the correction information indicating that the MPU time stamp is corrected is signaled and transmitted to the receiving apparatus (S2203).
If it is determined in step S2201 that the area C is the area C and the MPU timestamp indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second, and the corrected MPU timestamp is stored in the MPU timestamp descriptor (S2205).
If it is determined in step S2201 that the area B or the area D is present, the MPU timestamp is stored in the MPU timestamp descriptor without being corrected (S2204).
Next, the operation flow of the receiving apparatus will be described with reference to fig. 87.
Based on the information signaled from the transmitting side, it is determined whether the MPU timestamp is corrected (S2301).
If it is determined that the MPU timestamp is corrected (yes in S2301), the receiving apparatus presents the MPU based on the MPU timestamp at the NTP time after leap second adjustment (S2302).
If it is determined that the MPU timestamp is not corrected (no in S2301), the MPU is presented based on the MPU timestamp at the NTP time.
When the MPU timestamp is corrected, an MPU corresponding to the MPU timestamp is presented in the section where the leap second is inserted.
On the other hand, when the MPU timestamp is not corrected, the MPU corresponding to the MPU timestamp is presented not in the leap second inserted section but in the leap second not inserted section.
Fig. 88 shows the operation flow of the transmitting side in the case where the MPU timestamp is corrected in the receiving apparatus described in fig. 85, and fig. 89 shows the operation flow of the receiving apparatus in this case.
First, an operation flow of the transmission side (transmission device) will be described with reference to fig. 88.
Based on the leaf _ indicator of the NTP packet, it is determined whether or not leap second adjustment (insertion or deletion) is performed (S2401).
When it is determined that leap second adjustment is to be performed (yes in S2401), the timing of the application of the MPU timestamp is determined, and signaling is performed on the identification information and transmitted to the receiving device (S2402).
On the other hand, if it is determined that leap second adjustment is not to be performed (no in S2401), the process ends with the normal operation.
Next, the operation flow of the receiving apparatus will be described with reference to fig. 89.
Based on the identification information signaled by the transmitting side (transmitting apparatus), it is determined whether the timing of assigning the MPU timestamp is in the a region, B region, C region, or D region (S2501). Here, the regions a to D are the same as defined above, and therefore, the description thereof is omitted. Note that, similarly to fig. 87, the case where leap seconds are not inserted or deleted is not shown.
When it is determined that the area a is present in step S2501 and the MPU timestamp indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second (S2502).
If it is determined in step S2501 that the area C is the area C and the MPU timestamp indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second (S2504).
If it is determined in step S2501 that the area B or the area D is present, the MPU timestamp is not corrected (S2503).
The receiving device further presents the MPU based on the corrected MPU timestamp in a process not shown.
When correcting the MPU timestamp, an MPU corresponding to the MPU timestamp is presented based on the MPU timestamp at the NTP time after leap second adjustment.
On the other hand, when the MPU timestamp is not corrected, the MPU corresponding to the MPU timestamp is presented not in the leap second inserted section but in the leap second not inserted section.
In short, the transmitting side (transmitting apparatus) determines, for each of the MPUs, a timing at which an MPU time stamp corresponding to the MPU is given. Then, when the determined time is the time immediately before leap second insertion and the MPU timestamp indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second. Further, the correction information indicating that the MPU time stamp is corrected is transmitted by signaling and transmitted to the receiving apparatus. When the determined time is the time immediately before leap second deletion and the MPU timestamp indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second.
The receiving device also presents the MPU based on the MPU timestamp indicated by the correction information transmitted by the signaling from the transmitting side (transmitting device) when the MPU timestamp is corrected, and based on the MPU timestamp at the NTP time after leap second adjustment when the MPU timestamp is corrected. If the MPU timestamp is not corrected, the MPU is presented based on the MPU timestamp at the NTP time before leap second adjustment.
The transmitting side (transmitting apparatus) determines, for each of the MPUs, a timing at which an MPU time stamp corresponding to the MPU is given, and signals the result of the determination. The receiving apparatus performs the following processing based on the information indicating the timing of giving the MPU time stamp, which is transmitted by the transmitting side signaling. Specifically, when the information indicating the timing indicates the time immediately before leap second insertion and the MPU timestamp indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second. When the information indicating the timing indicates the time immediately before leap second deletion and the MPU timestamp indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second.
When the MPU timestamp is corrected, the receiving device presents the MPU based on the MPU timestamp at the NTP time adjusted by the leap second. If the MPU timestamp is not corrected, the MPU is presented based on the MPU timestamp at the NTP time before leap second adjustment.
In this way, by determining the timing of the application of the MPU timestamp during leap second adjustment, the MPU timestamp can be corrected, and the receiving apparatus can determine which MPU should be presented, and can perform appropriate reception processing using the MPU timestamp descriptor or the MPU extension timestamp descriptor. That is, even when leap second adjustment is performed for NTP time, the receiving apparatus can present a normal MPU using the MPU timestamp stored in the MPU timestamp descriptor.
[ supplement: transmitter and receiver
As described above, the transmission device that stores data constituting the coded stream in the MPU and transmits the data may be configured as shown in fig. 90. A receiving device that receives an MPU that stores data constituting a coded stream may be configured as shown in fig. 91. Fig. 90 is a diagram showing an example of a specific configuration of a transmitting apparatus. Fig. 91 is a diagram showing an example of a specific configuration of a receiving apparatus.
The transmission device 900 includes a generation unit 901 and a transmission unit 902. The generation unit 901 and the transmission unit 902 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The receiving apparatus 1000 includes a receiving unit 1001 and a reproducing unit 1002. The reception unit 1001 and the reproduction unit 1002 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The components of the transmission apparatus 900 and the reception apparatus 1000 will be described in detail in the description of the transmission method and the reception method, respectively.
First, a transmission method will be described with reference to fig. 92. Fig. 92 shows an operation flow (transmission method) of the transmission device.
First, the generation unit 901 of the transmission device 900 generates presentation time information (MPU timestamp) indicating the presentation time of the MPU as the first data unit based on the NTP time received from the outside as the reference time information (S2601).
Next, the transmission unit 902 of the transmission device 900 transmits the MPU, the presentation time information generated by the generation unit 901, and identification information indicating whether the presentation time information (MPU timestamp) is generated based on the NTP time before leap second adjustment (S2602).
Thus, even when the leap second adjustment is performed, the receiving device that has received the information transmitted from the transmitting device 900 can reproduce the MPU based on the identification information, and thus can reproduce the MPU at an expected time.
Next, a reception method will be described with reference to fig. 93. Fig. 93 is an operation flow of the receiving apparatus (receiving method).
First, the receiving unit 1001 of the receiving apparatus 1000 receives the MPU, presentation time information indicating the presentation time of the MPU, and identification information indicating whether the presentation time information (MPU timestamp) is generated based on the NTP time before leap second adjustment (S2701).
Next, the playback unit 1002 of the reception device 1000 plays back the MPU received by the reception unit 1001 based on the presentation time information (MPU timestamp) and the identification information received by the reception unit 1001 (S2702).
Thus, even when leap second adjustment is performed, the reception apparatus 1000 can reproduce the MPU at an expected time.
In embodiment 6, a process representing a process including at least one of decoding and presentation is reproduced.
(modification of embodiment 6)
Next, a specific example of a specific signaling method of the additional information (identification information) described in fig. 84 and 85 will be described.
Here, a method of performing signaling by a combination of the MPU time stamp descriptor shown in (a) in fig. 81 and the MPU extended time stamp descriptor shown in (b) in fig. 81 is explained.
Fig. 94 is a diagram showing an example of expansion of an MPU expansion timestamp descriptor. The underlined fields in fig. 94 are fields to be newly added to the MPU extended timestamp descriptor shown in (b) in fig. 81.
The NTP _ leaf _ indicator in fig. 94 indicates, in the MPU extended timestamp descriptor as the 2 nd control information, a flag indicating whether or not additional information (identification information) related to leap second adjustment by the NTP is indicated. When this flag is set, MPU _ presentation _ time _ type in each cycle of the MPUs becomes valid.
MPU _ presentation _ time _ type in fig. 94 indicates whether MPU _ presentation _ time, which is the same serial number described in the MPU timestamp descriptor, is corrected when the transmitting side (transmitting apparatus) corrects the MPU timestamp as described in fig. 84. When the receiver apparatus corrects the MPU timestamp as described with reference to fig. 85, MPU _ presentation _ time _ type indicates whether MPU _ presentation _ time having the same sequence number as that described in the MPU timestamp descriptor is assigned to the a, B, C, or D region.
That is, the MPU extended timestamp descriptor stores identification information indicating whether the MPU timestamp is generated based on the time information before leap second adjustment (NTP time). The MPU extension timestamp descriptor stores identification information corresponding to each MPU in a loop of each of the MPUs.
In step S2602, the transmission unit 902 of the transmission device 900 transmits an MPU, an MPU timestamp descriptor as first control information storing presentation time information (MPU timestamp) generated by the generation unit 901, and an MPU extended timestamp descriptor as second control information storing identification information indicating whether the presentation time information is generated based on the time information before leap second adjustment (NTP time).
The receiving unit 1001 of the receiving apparatus 1000 receives the MPU, the MPU timestamp descriptor, and the MPU extended timestamp descriptor in step S2701. Then, the playback unit 1002 of the reception device 1000 plays back the MPU received by the reception unit 1001 based on the MPU timestamp descriptor and the MPU extended timestamp descriptor received by the reception unit 1001. In this manner, the reception apparatus 1000 can realize decoding processing using the MPU timestamp during leap second adjustment by analyzing both the MPU timestamp descriptor and the MPU extended timestamp descriptor. Thus, even when leap second adjustment is performed, the reception apparatus 1000 can reproduce the MPU at an expected time. Further, since the identification information is stored in the MPU extended timestamp descriptor, the function of signaling whether or not the MPU timestamp is generated based on the time information before leap second adjustment (NTP time) can be extended while maintaining the structure corresponding to the existing standard (MPEG). As described above, since the conventional configuration can be used, even when the signaling transfer function is extended, design changes can be minimized, and manufacturing costs of the transmitting side (transmitting apparatus) and the receiving apparatus can be reduced.
Note that the timing of setting the NTP _ leaf _ indicator may be the same as the leaf _ indicator in the NTP packet, or may be any time in the period in which the leaf _ indicator is set. That is, the NTP _ leaf _ indicator is determined as a value corresponding to the value of the leaf _ indicator in the NTP packet corresponding to the NTP time that is the reference for generating the MPU timestamp.
(modification 2 of embodiment 6)
The problem with leap second adjustment described in the present embodiment is not limited to the case of using the time based on the NTP time, but occurs similarly even when using the time based on the UTC time.
When the time system is a time system based on UTC and presentation time information (MPU timestamp) is generated and given based on the time system, the presentation time information can be corrected and processed using the presentation time information and signaling information (additional information and identification information) even at the time of leap second adjustment by using the same method as described above.
For example, in ARIB STD-B60, in order to notify a broadcasting station of a time (time specifying information) for specifying an action for an application acting in a receiver, an event message descriptor is stored in an MMTP packet when an event message is transmitted, and the event message descriptor indicates the time at which the event message occurs. In this case, data constituting the application is stored in the MPU as the first data unit and transmitted. As for the time at which the event message occurs, there are various time-specifying methods, represented by time _ mode. When the time at which the event message occurs is expressed in the time _ mode as UTC, identification information indicating whether the UTC time is time information given based on the time information before leap second adjustment or time information given based on the time information after leap second adjustment is signaled. That is, the identification information is information indicating whether or not presentation time information is generated based on the time information before leap second adjustment (NTP time), as in the identification information described in the present embodiment. Such identification information (signalling information) is preferably shown within the event message descriptor. For example, the reserved future use field of the event message descriptor can be used for signaling.
That is, in step S2601, the generation unit 901 of the transmission device 900 generates time designation information indicating the operation time of the application based on the NTP time received from the outside as the reference time information. Then, in step S2602, the transmission unit 902 of the transmission device 900 transmits the MPU, and the event message descriptor storing the time specification information indicating the operation time of the application (the time at which the event message occurred) generated by the generation unit 901, and the identification information indicating whether or not the time information is the time information before leap second adjustment as the control information.
The transmitting side (transmitting apparatus) may determine the time and program that require leap second adjustment, and prohibit the use of the time _ mode expressed by UTC and use the time _ mode that is not related to UTC in the time and program that require leap second adjustment.
Alternatively, the transmitting side (transmitting apparatus) may calculate the UTC time at which the event message occurred, determine whether the time at which the event message occurred was leap second adjusted, and, if the time at which the event message occurred was leap second adjusted, calculate the leap second adjusted UTC time and store the same in the event message descriptor. In addition, whether or not the presentation time information is leap second adjusted may be separately signaled.
Similarly, the UTC (Universal Time Coordinated) -NPT (Normal Play Time) Reference descriptor uses the UTC _ Reference field and the NPT _ Reference field to indicate the relationship between UTC and NPT. The time information indicated by the UTC _ Reference field or the NPT _ Reference field is signaled based on the time information given by the Reference time information before leap second adjustment (NTP time) or based on the time information given by the Reference time information after leap second adjustment. That is, the UTC-NPT reference descriptor is control information in which the presentation time information generated by the generation unit 901 of the transmission apparatus 900 and identification information indicating whether or not the presentation time information is time information before leap second adjustment (NTP time) are stored. The identification information can be signaled using a reserved field or the like within the UTC-NPT reference descriptor.
The method of specifying the presentation time of the Subtitle or character superimposition described in ARIB STD-B60 or ARIB STD-B62 is represented by TMD (time control mode) in the Additional identification information (Additional _ ARIB _ Subtitle _ Info). The additional identification information is deposited in the resource loop of the MPT.
In the case of a mode in which TMD (Time control mode) is a Time that starts with an MPU timestamp or NPT (Normal Play Time), the presentation Time of superimposition of subtitles or characters is indicated in an MPU extended timestamp descriptor or UTC-NPT reference descriptor. With the above signaling information, even at the time of leap second adjustment, the time information can be corrected and processed by the receiving apparatus using the time information and the signaling information.
When the time control mode is "0000" (UTC _ TTML description), the presentation time is expressed such that the timecode in the ARIB-TTML caption corresponds to UTC. In this case, the UTC time indicated by the time code is signaled based on the time information given by the time information before leap second adjustment or the time information given by the time information after leap second adjustment. The signaling information transmitted at this time is preferably represented within ARIB-TTML subtitles.
When the time control mode is "0010" (UTC _ TTML description), the time starts from the UTC time indicated by the reference _ start _ time field in the additional identification information. In this case, the signaling is performed in the same manner as in the case where the time control pattern is "0000" (described in UTC _ TTML) in the additional identification information. In this case, the transmitting side (transmitting apparatus) transmits, as the identification information, information indicating: the reference _ start _ time field indicates whether or not the UTC time is a time from a time (e.g., 9:00:00 of the day immediately before leap second adjustment) that is earlier than the time (e.g., 9:00:00) immediately before leap second adjustment by a predetermined period (e.g., 24 hours) to the immediately preceding time.
The time zone or program requiring leap second adjustment may be determined on the transmitting side (transmitting apparatus), and the time control mode expressed by UTC may be prohibited from being used, and the time control mode unrelated to UTC may be used.
The same applies to the MH-Expire descriptor.
(embodiment 7)
A general broadcasting station apparatus has a plurality of transmission systems, and 1 system is used as a general apparatus for an actual broadcasting service and the other systems are used as redundant system apparatuses for backup. This enables switching to a redundant system device when a service provided to a viewer has an error due to a transmitting device or the like. Here, in the MMT/TLV system, the following phenomenon occurs when switching to a redundant system device.
1. Adjustment of MPU sequence number
As a preprocessing for switching the redundant system devices, the adjustment of MPU serial numbers is performed. As shown in fig. 95, when the MPU serial number is adjusted, the MPU serial number may not be continuous before and after the adjustment. Fig. 95 shows an example of an MPU serial number duplication.
The access unit or the MPU timestamp corresponding to the access unit is the same before and after the adjustment of the MPU serial number, but since the same MPU serial number is given to different MPUs, the same MPU serial number is given between 2 MPUs.
Thus, the MPU serial number stored in the MMTP header, and the MPU serial number stored in the MPU timestamp descriptor and the MPU extended timestamp descriptor are repeated in different MPUs. Similarly, the MPU serial number may jump or return.
2. Packet sequence number discontinuity
Since the packet sequence number and the packet counter are different between the normal device and the redundant system device, the packet sequence number and the packet counter may be discontinuous when the normal device is switched to the redundant system device. As shown in fig. 96, the packet sequence number is continuous in the normal operation, and a skip, a repeat, and a return may occur. Fig. 96 is a diagram for explaining a case where the packet sequence numbers are not consecutive when switching from the normal system device to the redundant system device.
The following problems exist: in the case where MPU serial numbers or packet serial numbers become discontinuous as in 1 or 2 above, there is a possibility that a receiving apparatus operating in a belief of continuity of these serial numbers may malfunction.
The access unit and the MPU timestamp corresponding to the access unit maintain continuity even when the MPU serial number or packet serial number is not continuous, and continuous data is transmitted to the receiving apparatus.
The reception process in the case where the MPU sequence number or packet sequence number is not continuous as described with reference to fig. 95 and 96 will be described.
If the MPU serial number or packet serial number is not consecutive, the receiving device may malfunction if it performs the processing using the MPU serial number or packet serial number as described below.
As the processing using the MPU serial number, for example, processing for detecting an MPU boundary is performed in accordance with switching of the MPU serial number in the MMTP header.
As a process using the MPU serial number, for example, a time stamp corresponding to the MPU serial number is described in an MPU time stamp descriptor or an MPU extended time stamp descriptor, and decoding and/or presentation is performed using the time stamp.
As the processing using the packet sequence number, for example, the following processing is given: packet loss is detected using the packet sequence number, and it is determined that the packet is lost when the packet sequence number is skipped, and it is determined that the packet is exchanged when the packet sequence number is returned.
In order to prevent a malfunction due to the processing using the MPU serial number or packet serial number, the receiving apparatus determines that the MPU serial number or packet serial number is not continuous, and performs processing not using the MPU serial number or packet serial number when the receiving apparatus determines that the MPU serial number or packet serial number is not continuous.
Here, as a method of determining discontinuity of the MPU serial number or the packet serial number, for example, the following method can be used.
First, one of the above-described determination methods may consider signaling additional information indicating switching to a redundant system device. Thus, the reception device can perform reception processing without using the MPU sequence number or the packet sequence number based on the additional information.
As a method of signaling additional information, there is a method of storing and transmitting the additional information in a control message, a descriptor, a table, MPU metadata, MF metadata, an MMTP header, an extension header of an MMTP packet, and the like.
Next, a method for detecting discontinuity using other information in the receiving apparatus will be described.
For example, as information indicating the MPU boundary, there are a method of detecting switching of an MPU sequence number in the MMTP header and a method of using RAP _ flag, which is a flag indicating whether or not the MMTP header is a random access point. Further, MPU boundary detection is also possible by switching packets whose Sample _ number is 0, switching slice types, or the like. Further, by counting the number of access units, MPU boundary can also be detected.
Here, in a case where there is a possibility of performing switching of redundant system devices, MPU boundaries are determined not using switching of MPU serial numbers but using other methods. For example, a packet whose RAP _ flag is 1 is determined as the leading packet in the MPU. Here, when the MPU serial number of the packet subsequent to the MPU boundary and the MPU serial number of the packet previous to the MPU boundary do not change, it is determined that the MPU serial number overlaps with the MMTP packet.
Alternatively, a packet whose RAP _ flag is 1 is determined as the leading packet in the MPU. Here, when it is determined that the MPU sequence number of the packet at the head of the MMTP packet has jumped or returned from the MPU sequence number of the previous packet, it is determined that the MPU sequence number is discontinuous in the MMTP packet.
In addition, when discontinuity of the MPU sequence number in the MMTP payload header does not necessarily occur before discontinuity of the MPU sequence number or packet sequence number in the MPU timestamp descriptor or the MPU extended timestamp descriptor, the reception process not using the timestamp or the process not using the packet sequence number may be performed based on the determination result.
Further, when the discontinuity of the MPU sequence number in the MMTP payload header of the video packet is not necessarily caused before the discontinuity of the MPU sequence number in the MMTP payload header of the audio packet, the discontinuity of the MPU sequence number in the audio MPU timestamp descriptor or the MPU extension timestamp descriptor, or the discontinuity of the packet sequence number, the reception process using no audio timestamp or the process using no packet sequence number may be performed based on the detection result of the discontinuity of the MPU sequence number of the video packet. The determination result may be determined as the timing of the start of switching to the redundant system device.
The receiving apparatus may perform processing not using the MPU serial number or the packet serial number only when the additional information indicating the switching to the redundant system device is indicated by signaling the additional information indicating the switching to the redundant system device.
The receiving apparatus may perform the discontinuity detection processing of the RAP _ flag and the MPU serial number only when the additional information indicating the switching to the redundant system device is indicated by signaling the additional information indicating the switching to the redundant system device.
Here, the determination using the time stamp descriptor in the receiving apparatus is explained. Normally, a time stamp descriptor corresponding to an MPU serial number is described and is often transmitted a plurality of times, but when the time stamp value is updated although the MPU serial number is the same, the receiving apparatus determines that the MPU serial number is duplicated.
When the time stamp value is updated by the presentation time of 1 MPU, but the MPU serial number is updated by 2 or more, and the relationship between the time stamp value and the MPU serial number does not correspond, the receiving apparatus determines that the MPU serial number is not continuous.
In addition, when the MPU timestamp or the MPU extended timestamp is updated, version (version) in the MP table may not be updated. Alternatively, the version information may be separately defined and the version of the time stamp information may be transmitted separately from the version of the information other than the time stamp information. This makes it possible to reduce the frequency of version of information other than the time stamp information, and to reduce the reception processing.
In addition, when the MPU serial number is not continuous, the relationship between the MPU timestamp and the timestamp information is not trusted. Therefore, information indicating that the time stamp information is not authentic may also be signaled in the descriptor. For example, the same field as the NTP _ leaf _ indicator or MPU _ presentation _ time _ type described in fig. 94 may be used to indicate that the timing of giving the MPU timestamp is the timing before and after switching the redundant system and the timestamp is not trusted. Further, it may be assumed that the MPU time stamp descriptor is used in combination with the MPU time stamp descriptor, and the MPU extended time stamp descriptor indicates the time stamp descriptor.
With the above method, by determining that the MPU sequence number or packet sequence number is discontinuous in the receiving apparatus, it is possible to perform processing without using the MPU sequence number or packet sequence number, and it is possible to prevent a malfunction during synchronous playback.
The operation flow of the receiving apparatus in the case where the MPU sequence number or packet sequence number is not consecutive as described above will be described. Fig. 97 is an operation flow of the receiving apparatus in the case where the MPU serial number or packet serial number is not consecutive.
In the receiving apparatus, the MPU boundary is determined from the header or payload header in the MMTP packet. Specifically, the MPU boundary is determined by determining a packet whose RAP _ flag is 1 as a packet at the head of the MPU. Further, the MPU serial number is determined, and whether the number is increased by 1 from the MPU serial number of the previous packet is determined (S2801).
If RAP _ flag is 1 and it is determined that the MPU sequence number is increased by 1 (yes in S2801), a reception process using the MPU sequence number is performed as the MPU sequence number without discontinuity (S2803).
On the other hand, if RAP _ flag is 0 or if RAP _ flag is 1 but it is determined that 1 is not added to the MPU sequence number of the previous packet (no in S2801), the MPU sequence number is discontinued and a reception process is performed without using the MPU sequence number (S2802).
(embodiment mode 8)
In embodiment 6, description has been given mainly on correction of the MPU time stamp, and here, a method of correcting the PTS or DTS of each AU included in the MPU in the receiving apparatus is described. That is, a correction method for correcting a time stamp indicating a PTS or DTS of each of a plurality of other AUs other than the head AU among a plurality of AUs included in the MPU in the receiving apparatus will be described.
Fig. 98 is a diagram for explaining a method of correcting a time stamp in the receiving apparatus when inserting leap seconds. Fig. 98 is the same as fig. 85 (a).
Here, the following case will be described as an example: the NTP time on the transmitting side is synchronized with the NTP server, and the NTP time on the receiving side is synchronized with the NTP time on the transmitting side by being reproduced based on the time stamp stored in the NTP packet transmitted from the transmitting side. In this case, the +1 second adjustment is performed for both the NTP time on the transmission side and the NTP time on the reception side when leap seconds are inserted.
NTP time in fig. 98 is set as: the NTP time on the transmitting side and the NTP time on the receiving side are common. In addition, the following describes a case where there is no transmission delay. In addition, only the method of correcting the PTS of the AU in the MPU will be described below, and the DTS of the AU in the MPU can be corrected by using the same method. In fig. 98, the number of AUs included in one MPU is described as 5, but the number of AUs included in an MPU is not limited to this.
Similarly to the transmission method described in fig. 85 a, the NTP time immediately before insertion of leap second (8: 59:59 seconds before 1 st time) is defined as the a region, and the NTP time after insertion of leap second (8: 59:59 and after 2 nd time) is defined as the B region.
The following processing is performed on the transmission side (transmission apparatus).
As described in embodiment 6 with reference to fig. 83, an MPU timestamp is generated as presentation time information (first time information) of an MPU.
Instead of correcting the generated MPU timestamp, the MPU timestamp is stored in an MPU timestamp descriptor and transmitted to the receiving device. That is, the first time information is the time information that is not corrected by the transmitting side of the MPU at the time of leap second adjustment. Identification information indicating whether the timing of the MPU time stamp is in the a region or the B region is transmitted to the receiving apparatus. Specifically, an MMTP packet storing a plurality of MPUs is transmitted. The MMTP packet includes, as control information, an MPU timestamp descriptor including an MPU timestamp that is first time information, an MPU extended timestamp descriptor including relative information that is second time information, and identification information.
In fig. 98, the timing of the time stamp given to the MPU is indicated by the base end of the arrow.
The following processing is performed in the receiving apparatus.
As explained in embodiment 6, in the MPU extended time stamp descriptor, the second time information for calculating the PTS or DTS of each AU is stored as relative information with respect to the PTS of the leading AU of the MPU stored in the MPU time stamp descriptor. In this case, the relative information stored in the MPU extended timestamp descriptor is not a value obtained by considering the leap second insertion.
The receiving apparatus first receives an MMTP packet including the MPU, the first time information, the second time information, and the identification information transmitted from the transmitting side (transmitting apparatus), and calculates PTS (dts) of all AUs included in the MPU based on both the first time information stored in the MPU timestamp descriptor and the second time information stored in the MPU extended timestamp descriptor, as shown by "PTS per AU" in fig. 98.
Next, the receiving apparatus corrects the PTS of each AU. AU to be corrected is as follows: the timing of assigning an MPU time stamp to an MPU to which an AU belongs (the timing shown by the base end of the arrow in fig. 98) is the a region, and the PTS of each AU before correction is 9:00:00 and later. The AU to be corrected is an AU shown underlined in fig. 98, and is AU #5 in MPU #1 and AU #1 to AU #5 in MPU # 2. When the reception device determines that the AU corresponds to the PTS to be corrected, the reception device corrects the PTS of the AU for-1 second.
That is, in the correction, it is determined whether or not the following correction condition is satisfied for each of the plurality of AUs: the MPU time stamp of the MPU storing the AU is generated based on the reference time information before leap second adjustment, and the calculated PTS or DTS of the AU is a predetermined time (9: 00:00 in the case of insertion of leap seconds) or later. Then, in the correction, the PTS or DTS of the AU determined to be in conformity with the correction condition is corrected.
An AU whose MPU belongs to which an MPU time stamp is assigned at the timing of the a region and whose PTS is not corrected (AU before AU #4 in MPU #1 in fig. 98) is presented at a time that matches the NTP time before insertion of leap second (8: 59:59 seconds before the 1 st time). Further, for an AU whose PTS is corrected or AUs subsequent thereto (in fig. 98, AUs #5 and AUs subsequent to AU # 1), an AU is presented at a time that matches the NTP time after insertion of a leap second (8: 59:59 seconds and subsequent to the 2 nd time).
Identification information indicating the timing of applying the MPU time stamp needs to be presented at least to an MPU including the AU to be corrected. That is, identification information does not necessarily need to be presented to an MPU that does not include an AU to be corrected.
Fig. 99 is a diagram for explaining a method of correcting the time stamp in the receiving apparatus when the leap second is deleted. Fig. 99 is the same as fig. 85 (b).
Here, the following case will be described as an example: the NTP time on the transmitting side is synchronized with the NTP server, and the NTP time on the receiving side is synchronized with the NTP time on the transmitting side by being reproduced based on the time stamp stored in the NTP packet transmitted from the transmitting side. In this case, the adjustment is performed for-1 second for both the NTP time on the transmitting side and the NTP time on the receiving side when leap seconds are deleted.
NTP time in fig. 99 is set as: the NTP time on the transmitting side and the NTP time on the receiving side are common. In addition, the following describes a case where there is no transmission delay. In addition, only the method of correcting the PTS of the AU in the MPU will be described below, and the DTS of all AUs in the MPU can be corrected by using the same method. In fig. 99, the number of AUs included in one MPU is described as 5, but the number of AUs included in an MPU is not limited to this.
Similarly to the transmission method described in fig. 85, the NTP time immediately before leap second deletion (8:59:58 seconds) is set as the C area, and the NTP time after leap second deletion (9:00:00 and thereafter) is set as the D area.
The following processing is performed on the transmission side (transmission apparatus).
As described in embodiment 6 with reference to fig. 83, an MPU timestamp is generated as presentation time information (first time information) of an MPU.
Instead of correcting the generated MPU timestamp, the MPU timestamp is stored in an MPU timestamp descriptor and transmitted to the receiving device. That is, the first time information is the time information that is not corrected by the transmitting side of the MPU at the time of leap second adjustment. Identification information indicating whether the timing of the MPU time stamp is in the C region or the D region is transmitted to the receiving apparatus. Specifically, an MMTP packet storing a plurality of MPUs is transmitted. The structure of the MMTP packet is the same as the MMTP packet illustrated in fig. 98.
In fig. 99, the timing of the time stamp given to the MPU is indicated by the base end of the arrow.
The following processing is performed in the receiving apparatus.
As explained in embodiment 6, in the MPU extended time stamp descriptor, the second time information for calculating the PTS or DTS of each AU is stored as relative information with respect to the PTS of the leading AU of the MPU stored in the MPU time stamp descriptor. In this case, the relative information stored in the MPU extended timestamp descriptor is not a value obtained by considering the leap second deletion.
The receiving apparatus first receives an MMTP packet including the MPU, the first time information, the second time information, and the identification information transmitted from the transmitting side (transmitting apparatus), and calculates PTS (dts) of all AUs included in the MPU based on both the first time information stored in the MPU timestamp descriptor and the second time information stored in the MPU extended timestamp descriptor, as shown by "PTS per AU" in fig. 99.
Next, the receiving apparatus corrects the PTS of each AU. AU to be corrected is as follows: the timing of assigning the MPU time stamp of the MPU to which the AU belongs (the timing shown by the base end of the arrow in fig. 99) is the C region, and the PTS of each AU before correction is 8:59:59 and later. The AU to be corrected is an AU shown underlined in fig. 99, and is AU #5 in MPU #2 and AU #1 to AU #5 in MPU # 3. When the reception device determines that the reception device is an AU corresponding to the PTS to be corrected, the reception device corrects the PTS of the AU for +1 second.
That is, in the correction, it is determined whether or not the following correction condition is satisfied for each of the plurality of AUs: the MPU time stamp of the MPU storing the AU is generated based on the reference time information before leap second adjustment, and the calculated PTS or DTS of the AU is a predetermined time (8: 59:59 in the case of leap second deletion). Then, in the correction, the PTS or DTS of the AU determined to be in conformity with the correction condition is corrected.
The receiving apparatus can perform synchronous playback of all AUs based on the NTP time after leap second deletion and the PTS after correction.
Identification information indicating the timing of applying the MPU time stamp needs to be presented at least to an MPU including the AU to be corrected. That is, identification information does not necessarily need to be presented to an MPU that does not include an AU to be corrected.
Fig. 100 is an operation flow of the receiving apparatus described in fig. 98 and 99.
The operation flow of the transmitting side is the same as that shown in fig. 88, and therefore, the description thereof is omitted.
In the receiving apparatus, first, the time stamps (PTS and DTS) of all AUs are calculated using the MPU time stamp descriptor and the MPU extended time stamp descriptor (S2901).
Next, it is determined, for each AU, whether the timing of giving the MPU timestamp of the MPU to which the AU belongs is the a region, the B region, the C region, or the D region, based on the identification information signaled by the transmitting side (S2902). Here, since the regions a to D are the same as defined above, the description thereof is omitted. Note that, as in fig. 89, the case where leap seconds are not inserted or deleted is not illustrated.
When it is determined in step S2902 that the MPU timestamp of the MPU to which the AU belongs is the timestamp given based on the NTP time in the a region, it is determined whether or not the timestamp of the AU calculated in step S2901 is 9:00:00 and thereafter (S2903).
When it is determined that the time stamp of the AU is 9:00:00 or less (yes in step S2903), the time stamp value of the AU is corrected for-1 second (S2904).
When it is determined in step S2902 that the MPU timestamp of the MPU to which the AU belongs is the timestamp given based on the NTP time in the C region, it is determined whether or not the timestamp of the AU calculated in step S2901 is 8:59:59, and thereafter (S2905).
If it is determined that the time stamp of the AU is 8:59:59 or later (yes in step S2905), the time stamp value of the AU is corrected for +1 second (S2906).
On the other hand, when it is determined in step S2902 that the MPU timestamp of the MPU to which the AU belongs is the timestamp given based on the NTP time of the B area or the D area, when it is determined in step S2903 that the timestamp value of the AU is not 9:00:00 or less (no in step S2903), or when it is determined in step S2905 that the timestamp value of the AU is not 8:59:59 or less (no in step S2905), the timestamp of the AU is not corrected (step S2907), and the operation of the receiving apparatus is ended.
As described above, when the first time information indicating the absolute value of the time stamp of the MPU and the second time information indicating the relative value to the absolute value are transmitted separately as typified by the MPU time stamp descriptor and the MPU extended time stamp descriptor, the transmitting side transmits the identification information indicating the area indicating the NTP time that is the basis of the absolute value given to the time stamp together with the absolute value of the time stamp.
In a receiving apparatus, time stamps (PTS and DTS) of all AUs are calculated based on first time information indicating an absolute value of a time stamp of an MPU and second time information indicating a relative value, and then whether or not to correct the time stamp is determined based on each time stamp and identification information indicating an area to which NTP time that is a basis for giving the absolute value of the time stamp of the MPU that is a basis of the time stamp is added, thereby correcting the time stamp.
The receiving apparatus need not display additional information such as information indicating whether or not a time stamp is corrected for each AU, information indicating whether the timing of an MPU time stamp given to an MPU of an AU to which the AU belongs is an a region, a B region, a C region, or a D region, in all MPUs, but may display additional information in a section in which an MPU including the AU to be corrected is included.
For example, when the time difference between the time stamp (PTS or DTS) in an arbitrary AU and the NTP time that is the basis of the MPU time stamp given to the MPU to which the AU belongs is N seconds, the period during which the AU to be corrected exists is at least N seconds before leap second adjustment when leap seconds are inserted or leap seconds are deleted. Therefore, the area a for inserting leap seconds or the area C for deleting leap seconds needs to be an area that indicates at least a period of N seconds or more before leap second adjustment.
Conversely, the following restrictions may also apply: the time for inserting the leap second area a and deleting the leap second area C is defined as M seconds, and a time stamp is given so that N < M is inevitably obtained.
The insertion of the leap second in the B area or the deletion of the leap second in the D area does not necessarily require signaling, and the insertion of the leap second in the a area and the deletion of the leap second in the D area may be indicated at minimum by information indicating a transition from the a area to the B area and information indicating a transition from the C area to the D area.
In short, the receiving apparatus first calculates a time stamp of all the access units based on the first time information (absolute time information) and the second time information (relative time information), and corrects the time stamp of each of the access units based on the identification information indicating whether the NTP time, which is the basis of calculating the MPU time stamp of the MPU to which the access unit belongs, is before leap second adjustment. Specifically, when the identification information indicates the time immediately before insertion of the leap second and the time stamp of the access unit indicates 9:00:00 or later, the time stamp is corrected by-1 second. When the identification information indicates the time immediately before leap second deletion and the time stamp of the access unit indicates 8:59:59 or later, the time stamp is corrected by +1 second. In the case of an access unit with a corrected time stamp, the receiving device presents the access unit based on the reference time information after leap second adjustment.
In this way, in leap second adjustment, by determining the timing of assigning the MPU timestamp of the MPU to which the access unit belongs, the timestamps of all access units can be corrected, and appropriate reception processing can be performed using the MPU timestamp descriptor and the MPU extended timestamp descriptor. That is, even when leap second adjustment is performed with respect to reference time information that is a reference of reference clocks of the transmitting side and the receiving device, a plurality of access units stored in the MPU can be reproduced at an expected time.
[ supplement: transceiver system and receiver
As described above, the transmission/reception system including the transmission device that stores data constituting the coded stream in the MPU and transmits the data and the reception device that receives the transmitted MPU may be configured as shown in fig. 101. The receiving apparatus may be configured as shown in fig. 102. Fig. 101 is a diagram showing an example of a specific configuration of a transmission/reception system. Fig. 102 is a diagram showing an example of a specific configuration of a receiving apparatus.
The transmission/reception system 1100 includes a transmission device 1200 and a reception device 1300.
The transmission device 1200 includes a generation unit 1201 and a transmission unit 1202. The generation unit 1201 and the transmission unit 1202 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The generation unit 1201 generates first time information (MPU timestamp) indicating presentation time of the first data unit (MPU) based on reference time information (NTP time) received from the outside.
The transmission unit 1202 transmits: the first data unit (MPU), the first time information (MPU timestamp) generated by the generation unit 1201, the second time information (relative information) indicating Presentation Time (PTS) or Decoding Time (DTS) of each of the plurality of second data units (AUs) together with the first time information (MPU timestamp), and the identification information.
The reception device 1300 includes a reception unit 1301, a calculation unit 1302, and a correction unit 1303. The receiving unit 1301, the calculating unit 1302, and the correcting unit 1303 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The detailed description of each component of receiving apparatus 1300 is given in the description of the receiving method.
The reception method will be described with reference to fig. 103. Fig. 103 shows an operation flow of the receiving apparatus (receiving method).
First, the reception unit 1301 of the reception apparatus 1300 receives: a first data unit (MPU), first time information (MPU timestamp), second time information (relative information), and identification information (S3001).
Next, the calculation unit 1302 of the reception apparatus 1300 calculates Presentation Time (PTS) or Decoding Time (DTS) of each of the plurality of second data units (AU) received by the reception unit 1301, using the first time information (MPU timestamp) and the second time information (relative information) received by the reception unit 1301 (S3002).
Next, the correction unit 1303 of the reception device 1300 corrects Presentation Time (PTS) or Decoding Time (DTS) of each of the plurality of second data units (AU) calculated by the calculation unit 1302, based on the identification information received by the reception unit 1301 (S3003).
Thus, even when leap second adjustment is performed, the reception apparatus 1300 can normally reproduce a plurality of second data units stored in the first data unit.
In embodiment 8, the playback represents a process including at least one of decoding and presentation.
(embodiment mode 9)
In embodiment 8, a method of correcting a PTS or a DTS of each AU included in an MPU in a receiving apparatus is described, and in embodiment 9, a method of correcting the PTS or the DTS also in a transmitting apparatus is described. That is, a correction method for correcting a time stamp indicating a PTS or DTS of each of a plurality of other AUs other than the head AU among a plurality of AUs included in the MPU in the transmission apparatus will be described.
Fig. 104 is a diagram for explaining a method of correcting a time stamp in the transmitting side (transmitting apparatus) when leap seconds are inserted. Fig. 104 is the same as fig. 84 (a).
Here, the following case will be described as an example: the NTP time on the transmitting side is synchronized with the NTP server, and the NTP time on the receiving side is synchronized with the NTP time on the transmitting side by being reproduced based on the time stamp stored in the NTP packet transmitted from the transmitting side. In this case, the +1 second adjustment is performed for both the NTP time on the transmission side and the NTP time on the reception side when leap seconds are inserted.
NTP time in fig. 104 is set as: the NTP time on the transmitting side and the NTP time on the receiving side are common. In addition, the following describes a case where there is no transmission delay. In addition, only the method of correcting the PTS of the AU in the MPU will be described below, and the DTS of the AU in the MPU can be corrected by using the same method. In fig. 104, the number of AUs included in one MPU is described as 5, but the number of AUs included in an MPU is not limited to this.
The following processing is performed on the transmission side (transmission apparatus).
1. As described in embodiment 6 with reference to fig. 83, an MPU timestamp generation process for generating presentation time information (first time information) for an MPU is performed.
2. When the MPU time stamp application timing (timing indicated by the base end of the arrow in fig. 104) is included in the area a and the MPU time stamp (MPU time stamp value before correction) indicates 9:00:00 or later, the following correction processing is performed: the MPU timestamp values are corrected for-1 second and the corrected MPU timestamps are stored in an MPU timestamp descriptor. That is, when the MPU timestamp is generated based on the NTP time included in the area a and indicates 9:00:00 or later, the MPU timestamp is corrected for-1 second. Here, "9: 00: 00" refers to a time (i.e., a time calculated by adding 9 hours to the UTC time) obtained by making the time that is the reference for leap second adjustment correspond to the japanese time.
When the timing at which the MPU timestamp is given (the timing indicated by the base end of the arrow in fig. 104) is in the B region, the MPU timestamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the B area, the MPU timestamp is not corrected.
In the MPU extended timestamp descriptor, second time information for calculating the PTS or DTS of each AU is stored as relative information with respect to the PTS of the leading AU of the MPU stored in the MPU timestamp descriptor. In this case, the relative information stored in the MPU extended timestamp descriptor is not a value obtained by considering the leap second insertion.
Therefore, in the generation processing of the above 1, in the case where the MPU timestamp, which is the first time information, satisfies the correction condition, the MPU timestamp is corrected. Therefore, the first time information is time information to be corrected at the leap second adjustment. The correction condition is that the MPU timestamp is generated based on the NTP time before leap second adjustment and is 9:00:00 or less.
3. A transmission process is performed for transmitting an MMTP packet in which a plurality of MPUs are stored to a reception device. The MMTP packet includes: an MPU time stamp descriptor including an MPU time stamp as first time information, an MPU extended time stamp descriptor including relative information as second time information, and identification information are used as control information.
Here, the identification information is the following information: when the MPU timestamp does not satisfy the MPU timestamp correction condition in the generation process of 1 and the transmitting side does not correct the MPU timestamp, the receiving apparatus identifies the MPU including the AU requiring the leap second-inserted correction for the PTS or DTS when the receiving apparatus calculates the PTS or DTS for each AU in the MPU using the MPU timestamp and the relative information.
That is, the identification information is information indicating that there is a possibility that an AU (hereinafter referred to as "AU requiring correction") that is presented at a PTS requiring correction or a DTS requiring correction due to leap second adjustment with respect to the reference time information is included in the MPU and is an AU other than the head AU. In other words, the identification information may be information indicating whether or not the receiving apparatus determines whether or not the AU needs to be corrected and is included in the MPU.
The identification information is, for example, information indicating: whether the MPU timestamp is generated based on the NTP time before leap second adjustment in the generation process of step 1 and is not corrected in the correction process of step 2. That is, the identification information includes: the MPU timestamp is generated based on the NTP time before leap second adjustment, and the MPU timestamp is corrected.
As described above, when the identification information indicates that the MPU timestamp is generated based on the NTP time before leap second adjustment in the generation process and is not corrected in the correction process, it indicates that there is a possibility that the AU needs to be corrected and is included in the MPU. In contrast, when the identification information indicates that the MPU timestamp is information that is generated based on the NTP time before leap second adjustment or is corrected during the correction process during the generation process, it indicates that the AU requiring correction cannot be included in the MPU (i.e., the reception apparatus is not caused to determine whether the AU requiring correction is included in the MPU).
The processing on the transmission side (transmission apparatus) will be specifically described with reference to fig. 104.
In the example shown in fig. 104, the MPU #2 that meets the correction condition in the correction processing of 2 described above performs-1 second correction on the MPU timestamp of the MPU #2 on the transmission side, and does not perform correction on the MPU timestamps of other MPUs, as shown by the dashed line. The MPU timestamp calculated as described above is shown in an MPU timestamp descriptor, and in an MPU extended timestamp descriptor not shown, the pts (pts) (dts) of each AU in the MPU is shown by relative information with respect to the MPU timestamp, as in the conventional art.
In addition, MPU #1 in fig. 104 is an MPU including the AU that needs to be corrected. Here, fig. 104 shows a PTS for each AU calculated based on the MPU timestamp descriptor and the MPU extended timestamp descriptor. The PTS of AU #5 in MPU #1 was calculated to be 9:00:00.1, but actually is an AU that should be presented in 8:59:59 seconds 2 after insertion of leap seconds, and requires-1 second correction. That is, AU #5 in MPU #1 is a correction-required AU.
As described above, in the transmitting apparatus, an MPU that meets the condition that the application timing of the MPU time stamp (timing shown by the base end of the arrow in fig. 104) is the a region and that the correction condition is not satisfied (that is, the MPU time stamp before correction is 9:00:00 before the predetermined time and at least 1 or more of the pts (dts) of all AUs in the MPU calculated using the MPU time stamp before correction and the relative information becomes 9:00:00 or less) can be determined to include an AU that needs to be corrected, and information indicating that the MPU that includes the AU that needs to be corrected is signaled from the transmitting apparatus to the receiving apparatus. In other words, the MPU may also be referred to as an "MPU in which both the AU to be presented at the 1 st time in 8:59:59 seconds and the AU to be presented at the 2 nd time in 8:59:59 seconds exist within the MPU without correcting the MPU timestamp.
The following processing is performed in the receiving apparatus.
1. A reception process of receiving the MMTP packet transmitted by the transmitting side (transmitting apparatus) is performed.
2. A calculation process is performed for calculating PTS (partial Transmit sequence) and DTS (partial Transmit sequence) of all AUs based on MPU time stamp descriptors and MPU extended time stamp descriptors included in control information of an MMTP packet.
3. A determination process is performed to determine whether or not an AU to be corrected is included in an MPU indicated by identification information indicating that the AU to be corrected is likely to be included, based on the identification information indicating whether or not the MPU is likely to include the AU to be corrected, which is signaled by a transmitting side (transmitting apparatus). Specifically, it is determined whether or not the pts (dts) of all AUs included in the MPU is 9:00:00 and thereafter, and it is determined that an AU requiring correction is included in the MPU by determining an AU requiring correction after the pts (dts) is 9:00: 00. Then, for pts (dts) whose AU needs to be corrected, correction is performed for-1 second.
The processing of the receiving apparatus is specifically described with reference to fig. 104.
In fig. 104, since the MPU #1 transmitted by signaling from the transmitting side is an MPU that may include an AU that needs to be corrected, the PTS calculated based on the time stamp descriptor in the MPU #1 indicates that AU # 59: 00:00 or later is the subject of correction, and-1 second correction is performed. Thus, the PTS of AU #5 in the MPU #1 after correction becomes 8:59: 59.1.
In addition, when it is signaled from the transmitting side that it is impossible to include an MPU which requires AU correction, the receiving apparatus does not perform the above determination processing. Therefore, pts (dts) of the AUs calculated based on the MPU time stamp descriptor and the MPU extended time stamp descriptor are not corrected for the AUs stored in the MPU. Thus, the pts (dts) of all AUs can be calculated.
When the receiving device performs presentation or decoding using the time stamp corrected in either the transmitting device or the receiving device based on the information indicating whether the transmitting-side MPU time stamp is corrected or whether the pts (dts) is corrected in the receiving device, the receiving device must perform presentation or decoding based on the NTP time after the leap second insertion (8: 59:59 seconds and thereafter from the 2 nd time). On the other hand, when presentation or decoding is performed using the timestamp before correction, presentation or decoding is performed based on the NTP time before insertion of leap seconds (8: 59:59 seconds before the 1 st time). In the example of fig. 104, the presentation of NTP time before AU #4 in MPU #1 is based on the leap second insertion (8: 59:59 seconds before 1 st time), and the presentation of NTP time after AU #5 in MPU #1 is based on the leap second insertion (8: 59:59 seconds after 2 nd time).
In the present embodiment, the relative value to the MPU timestamp is stored in the MPU extended timestamp descriptor in the transmitting side without considering the insertion leap seconds as in the conventional art, but the relative value obtained by correcting the timestamp of AU #5 in MPU #1 by-1 second may be stored in the MPU extended timestamp descriptor in consideration of the correction by inserting leap seconds. At this time, the following situation is additionally signaled: this AU is corrected based on the time indication of NTP after insertion of leap seconds (8: 59:59 seconds after the 2 nd time) or the decoded AU.
The reception device calculates PTS or DTS of all AUs based on the MPU time stamp descriptor and the MPU extended time stamp descriptor. In this case, since the relative value is already corrected, it is not necessary to correct the time stamp of the AU in the receiving apparatus, and all AUs can be presented or decoded based on the information indicating whether the time stamp of the MPU is corrected and the information indicating whether the PTS or DTS of the AU in the MPU is corrected.
Fig. 105 is a diagram for explaining a method of correcting the time stamp in the transmitting side (transmitting apparatus) when leap seconds are deleted. Fig. 105 is the same as fig. 84 (b).
Here, the following case will be described as an example: the NTP time on the transmitting side is synchronized with the NTP server, and the NTP time on the receiving side is synchronized with the NTP time on the transmitting side by being reproduced based on the time stamp stored in the NTP packet transmitted from the transmitting side. In this case, the adjustment is performed for-1 second for both the NTP time on the transmitting side and the NTP time on the receiving side when leap seconds are deleted.
NTP time in fig. 105 is set as: the NTP time on the transmitting side and the NTP time on the receiving side are common. In addition, the following describes a case where there is no transmission delay. In addition, only the method of correcting the PTS of the AU in the MPU will be described below, and the DTS of the AU in the MPU can be corrected by using the same method. In fig. 105, the number of AUs included in one MPU is described as 5, but the number of AUs included in an MPU is not limited to this.
The following processing is performed on the transmission side (transmission apparatus).
1. As described in embodiment 6 with reference to fig. 83, an MPU timestamp generation process for generating presentation time information (first time information) for an MPU is performed.
2. When the MPU time stamp adding timing (timing indicated by the base end of the arrow in fig. 105) is included in the C region and the MPU time stamp (MPU time stamp value before correction) indicates 8:59:59 or later, the following correction processing is performed: the MPU timestamp value is corrected for +1 second and the corrected MPU timestamp is stored in an MPU timestamp descriptor. That is, when the MPU timestamp is generated based on the NTP time included in the C area and indicates 8:59:59 or later, the MPU timestamp is corrected for +1 second. Here, "8: 59: 59" refers to a time obtained by subtracting-1 second from a time obtained by making the time that is the reference for leap second adjustment correspond to the japanese time (i.e., a time calculated by adding 9 hours to the UTC time).
When the timing at which the MPU timestamp is given (timing indicated by the base end of the arrow in fig. 105) is in the D region, the MPU timestamp is not corrected. That is, in the case of the MPU timestamp generated based on the NTP time included in the D area, the MPU timestamp is not corrected.
In the MPU extended timestamp descriptor, second time information for calculating the PTS or DTS of each AU is stored as relative information with respect to the PTS of the leading AU of the MPU stored in the MPU timestamp descriptor. In this case, the relative information stored in the MPU extended timestamp descriptor is not a value obtained by considering the leap second deletion.
Therefore, in the generation processing of the above 1, in the case where the MPU timestamp, which is the first time information, satisfies the correction condition, the MPU timestamp is corrected. Therefore, the first time information is time information to be corrected at the leap second adjustment. In addition, the correction conditions are: the MPU timestamp is generated based on the NTP time before leap second adjustment and is 8:59:59 and later.
3. A transmission process is performed for transmitting an MMTP packet in which a plurality of MPUs are stored to a reception device. The MMTP packet includes: an MPU time stamp descriptor including an MPU time stamp as first time information, an MPU extended time stamp descriptor including relative information as second time information, and identification information are used as control information.
Here, the identification information is the following information: when the MPU timestamp does not satisfy the MPU timestamp correction condition in the generation process of 1 and the transmitting side does not correct the MPU timestamp, the receiving apparatus identifies the MPU including the AU requiring correction by leap second deletion with respect to the PTS or DTS when the receiving apparatus calculates the PTS or DTS for each AU in the MPU using the MPU timestamp and the relative information.
Here, the identification information is the same as the identification information explained when leap seconds are inserted, and explanation thereof is omitted.
The processing on the transmitting side (transmitting apparatus) will be specifically described with reference to fig. 105.
In the example shown in fig. 105, the MPU #3 that satisfies the condition for correcting the MPU timestamp in the correction processing of the above 2 is an MPU #3, and as shown by the dashed line, the MPU timestamp of the MPU #3 is corrected for +1 second on the transmission side, and the MPU timestamps of the other MPUs are not corrected. The MPU timestamp calculated as described above is shown in an MPU timestamp descriptor, and in an MPU extended timestamp descriptor not shown, the pts (pts) (dts) of each AU in the MPU is shown by relative information with respect to the MPU timestamp, as in the conventional art.
In addition, MPU #2 in fig. 105 is an MPU including the AU that needs to be corrected.
Here, fig. 105 shows a PTS for each AU calculated based on the MPU timestamp descriptor and the MPU extended timestamp descriptor. Although the PTS of AU #5 in MPU #2 is calculated to be 8:59:59.1, actually, it is an AU that should be presented 9:00:00 seconds after leap second is deleted, and correction of +1 second is necessary. That is, AU #5 in MPU #2 is a correction-required AU.
As described above, in the transmitting apparatus, an MPU that satisfies the condition that the timing of assigning the MPU timestamp (the timing shown by the base end of the arrow in fig. 105) is the C region and that does not satisfy the correction condition (that is, the MPU timestamp before correction is 8:59:59 before the predetermined time and at least 1 or more of the pts (dts) of all AUs in the MPU calculated using the MPU timestamp before correction and the relative information is 8:59:59 and thereafter) can be determined to include an AU that needs to be corrected, and information indicating that the MPU includes the AU that needs to be corrected is transmitted from the transmitting apparatus to the receiving apparatus by signaling. In other words, the MPU may be referred to as an "MPU in which there are both AUs to be presented 8:59:58 seconds before leap second deletion and AUs to be presented 9:00:00 seconds after leap second deletion" in the MPU without correcting the MPU timestamp.
The following processing is performed in the receiving apparatus.
1. A reception process of receiving the MMTP packet transmitted by the transmitting side (transmitting apparatus) is performed.
2. A calculation process is performed for calculating PTS (partial Transmit sequence) and DTS (partial Transmit sequence) of all AUs based on MPU time stamp descriptors and MPU extended time stamp descriptors included in control information of an MMTP packet.
3. A determination process is performed to determine whether or not an AU to be corrected is included in an MPU indicated by identification information indicating that the AU to be corrected is likely to be included, based on the identification information indicating whether or not the MPU is likely to include the AU to be corrected, which is signaled by a transmitting side (transmitting apparatus). Specifically, it is determined whether or not the pts (dts) of all AUs included in the MPU is 8:59:59, and then, it is determined that an AU requiring correction is included in the MPU by determining an AU having a pts (dts) of 8:59:59 and then, that is, an AU requiring correction. Then, for pts (dts) whose AU needs to be corrected, correction is performed for +1 second.
In fig. 105, since the MPU #2 that is signaled from the transmitting side is an MPU that may include AUs that need to be corrected, the PTS calculated based on the time stamp descriptor in the MPU #2 indicates that AU #5 after 8:59:59 is the subject of correction, and +1 second correction is performed. Thus, the PTS of AU #5 in MPU #2 becomes 9:00:00.1 after correction.
In addition, when it is signaled from the transmitting side that it is impossible to include an MPU which requires AU correction, the receiving apparatus does not perform the above determination processing. Therefore, pts (dts) of the AUs calculated based on the MPU time stamp descriptor and the MPU extended time stamp descriptor are not corrected for the AUs stored in the MPU. Thus, the pts (dts) of all AUs can be calculated.
In addition, the receiving apparatus can use the corrected pts (dts) to cue or decode the AU.
In the present embodiment, the explanation has been made on the transmitting side that the relative value to the MPU timestamp is stored in the MPU extended timestamp descriptor without considering the leap second deletion as in the conventional art, but the relative value obtained by correcting the timestamp of AU #5 in MPU #1 by +1 second may be stored in consideration of the correction due to the leap second deletion.
The reception device calculates PTS or DTS of all AUs based on the MPU time stamp descriptor and the MPU extended time stamp descriptor. In this case, since the relative value is already corrected, the reception apparatus does not need to correct the time stamp of the AU, and can present or decode all AUs.
In the leap second insertion or leap second deletion correction method described above, the transmission side (transmission apparatus) may be able to select which correction method to use. When the transmitting side (transmitting apparatus) selects the correction method, the transmitting side performs the correction using which method and transmits the signal to the receiving apparatus. The receiving apparatus can switch between the correction, presentation operation, and decoding operation in the receiving apparatus based on the information transmitted by the signaling.
Fig. 106 is an operation flow of the transmission device described in fig. 104. Fig. 107 is an operation flow of the receiving apparatus described in fig. 104.
In fig. 106, the transmission apparatus determines whether the time stamp application timing of the MPU is in the a region or the B region (S3101). Here, since the a region and the B region are the same as defined above, the description thereof is omitted. Note that, similarly to fig. 89, the case where leap seconds are not inserted is not illustrated.
If it is determined in step S3101 that the area B is present, the MPU timestamp is not corrected (S3102), and the process ends.
On the other hand, if it is determined that the area a is the area in step S3101, it is determined whether or not the MPU time stamp is 9:00:00 seconds or less (S3103).
When it is determined that the MPU timestamp is 9:00:00 or less (yes in step S3103), the MPU timestamp is corrected for-1 second, the corrected MPU timestamp is signaled to the receiving apparatus (S3104), and the process is terminated.
On the other hand, when it is determined that the MPU timestamp is not 9:00:00 or less (no in step S3103), the timestamps (PTS and DTS) of all AUs included in the MPU are calculated using the MPU timestamp descriptor and the MPU extended timestamp descriptor (S3105).
Next, it is determined whether or not an AU whose calculated time stamp is 9:00:00 or later is included in the MPU (S3106).
If it is determined that the calculated time stamp is contained in an AU from 9:00:00 or later (yes in S3106), the time stamp of the AU in the MPU needs to be corrected in the receiving apparatus and the signal is transmitted to the receiving apparatus (S3107).
On the other hand, when it is determined that the AU whose calculated time stamp is 9:00:00 or later is not included in the MPU (no in S3106), the process is ended.
In fig. 107, the receiving apparatus calculates time stamps of all AUs using the MPU time stamp descriptor and the MPU extended time stamp descriptor (S3201).
Next, it is determined whether or not the time stamp of the AU in the MPU needs to be corrected based on the identification information signaled (S3202).
When it is determined that the time stamps of AUs in the MPU need to be corrected (yes in step S3202), the time stamps of AUs whose time stamps are 9:00:00 and later are corrected for-1 second (S3205).
On the other hand, when it is determined that the time stamp of the AU in the MPU does not need to be corrected (no in step S3202), it is determined whether or not the MPU time stamp is corrected based on the identification information signaled (S3203).
If it is determined that the MPU timestamp is not corrected (no in step S3203), the AU is presented based on the NTP time before the leap second is inserted for the timestamp before the corrected timestamp, and based on the NTP time after the leap second is inserted for the timestamp after the corrected timestamp (S3204).
After step S3205 or when it is determined in step S3203 that the time stamp is corrected (yes in step S3203), the AU is presented based on the NTP time after the leap second insertion (S3206).
Fig. 108 is an operation flow of the transmitting apparatus described in fig. 105. Fig. 109 is an operation flow of the receiving apparatus described in fig. 105.
In fig. 108, the transmission apparatus determines whether the time stamp application timing of the MPU is in the C region or the D region (S3301). Here, the C region and the D region are the same as defined above, and therefore, the description thereof is omitted. Note that, similarly to fig. 89, the case where leap second deletion is not performed is not illustrated.
If it is determined in step S3301 that the area is D, the MPU time stamp is not corrected (S3302), and the process ends.
On the other hand, if it is determined that the area C is the area in step S3301, it is determined whether or not the MPU timestamp is 8:59:59, and thereafter (S3303).
If it is determined that the MPU timestamp is 8:59:59 or later (yes in step S3303), the MPU timestamp is corrected for +1 second (S3304), and the process ends.
On the other hand, when it is determined that the MPU timestamp is not 8:59:59 or later (no in step S3303), the timestamps (PTS and DTS) of all AUs included in the MPU are calculated using the MPU timestamp descriptor and the MPU extended timestamp descriptor (S3305).
Next, it is determined whether or not an AU whose calculated time stamp is 8:59:59 or later is included (S3306).
When it is determined that the AU is included after the calculated time stamp is 8:59:59 (yes in S3306), the time stamp of the AU in the MPU needs to be corrected in the receiving apparatus and the signaling is performed to the receiving apparatus (S3307).
On the other hand, if it is determined that the AU after the calculated time stamp is 8:59:59 is not included in the MPU (no in S3306), the process is ended.
In fig. 109, the receiving apparatus calculates time stamps of all AUs using the MPU time stamp descriptor and the MPU extended time stamp descriptor (S3401).
Next, it is determined whether or not the time stamp of the AU in the MPU needs to be corrected based on the identification information signaled (S3402).
When it is determined that the time stamp of an AU in the MPU needs to be corrected (YES in step S3402), the time stamps of AUs whose time stamps are 8:59:59 and later are corrected for +1 second (S3403).
After step S3403 or when it is determined in step S3402 that correction of the time stamp of the AU in the MPU is not necessary (no in step S3402), the AU is presented based on the NTP time.
Thereby, (i) the MPU, (ii) the MPU timestamp, and (iii) the identification information indicating: the second data unit presented at the presentation time that needs to be corrected or decoded at the decoding time that needs to be corrected, and the second data units other than the second data unit whose presentation order is the head among the plurality of second data units, may be included in the first data unit. Therefore, even when leap second adjustment is performed with respect to reference time information that is a reference of the reference clocks of the transmitting side and the receiving apparatus, the receiving apparatus can reproduce a plurality of second data units stored in the first data unit at an expected time.
[ supplement: transmitter and receiver
As described above, the transmission device that stores data constituting the coded stream in the MPU and transmits the data may be configured as shown in fig. 110. A receiving device that receives an MPU storing data constituting a coded stream may be configured as shown in fig. 111. Fig. 110 is a diagram showing an example of a specific configuration of a transmitting apparatus. Fig. 111 is a diagram showing an example of a specific configuration of a receiving apparatus.
The transmission device 1400 includes a generation unit 1401 and a transmission unit 1402. The generation unit 1401 and the transmission unit 1402 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The reception apparatus 1500 includes a reception unit 1501 and a determination unit 1502. The receiving unit 1501 and the determining unit 1502 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like.
The components of the transmission apparatus 1400 and the reception apparatus 1500 will be described in detail in the description of the transmission method and the reception method, respectively.
First, a transmission method will be described with reference to fig. 112. Fig. 112 is an operation flow (transmission method) of the transmission device.
First, the generation unit 1401 of the transmission device 1400 generates first time information indicating presentation time of an MPU as a first data unit, based on reference time information received from the outside (S3501).
Next, the transmission unit 1402 of the transmission device 1400 transmits: a first data unit (MPU), first time information (MPU timestamp), and identification information (S3502). Here, the first time information (MPU timestamp) is information generated by the generation unit 1401. In addition, the identification information indicates: the second data unit (access unit) presented at the Presentation Time (PTS) requiring correction or decoded at the Decode Time (DTS) requiring correction, and the other second data unit (access unit) other than the second data unit (access unit) whose presentation order is the head among the plurality of second data units (access units), may be included in the first data unit (MPU).
Thus, even when leap second adjustment is performed, the receiving apparatus that has received the information transmitted from the transmitting apparatus 1400 can reproduce the AU in the MPU based on the identification information, and therefore can reproduce the AU in the MPU at an expected time.
Next, a reception method will be described with reference to fig. 113. Fig. 113 shows an operation flow of the receiving apparatus (receiving method).
First, the reception unit 1501 of the reception apparatus 1500 receives: a first data unit (MPU), first time information (MPU timestamp), and identification information (S3601). The first timing information and the identification information are the same as those described in the transmission device 1400.
Next, the determination unit 1502 of the reception apparatus 1500 determines whether or not another second data unit (access unit) is included in the first data unit (MPU) whose identification information received by the reception unit 1501 indicates that another second data unit (access unit) is likely to be included (S3602).
Thus, even when leap second adjustment is performed, the reception apparatus 1500 can reproduce the AU in the MPU at an expected time.
(embodiment mode 10)
Fig. 114 is a diagram showing details of a protocol stack diagram of the MMT/TLV scheme defined in ARIB STD-B60, and is a diagram showing details of fig. 78.
In the MMT/TLV system, video, audio, subtitle/character superimposition, application, and the like are packetized into an MMTP packet and transmitted. The subtitles/character overlays are TTML encoded. As for the application, the HTML5 application is stored in the MMTP package.
Fig. 115 is a block diagram showing the receiving apparatus, and is a diagram showing fig. 80 in more detail.
The receiving apparatus 1600 decodes the transmission path encoded data in the demodulation means 1602 with respect to the broadcast signal received by the tuner 1601, and extracts the TLV packet by applying error correction or the like to the decoded data.
A TLV/IP/UDP Demultiplexing (DEMUX) unit 2603 performs TLV demultiplexing and IP/UDP demultiplexing. The TLV/IP/UDP demultiplexing mechanism 2603 performs processing corresponding to the data type in the TLV demultiplexing process. For example, in the case of a compressed IP packet, the compressed header is restored. The TLV/IP/UDP multiplexing mechanism 2603 handles TLV-SI (AMT, NIT) in the case that the data type is TLV-SI. The TLV/IP/UDP demultiplexing mechanism 2603 performs discarding or the like when the data type is null packet. The TLV/IP/UDP multiplexing unit 2603 performs processing such as header analysis of an IP packet or a UDP packet in the IP/UDP multiplexing processing, and extracts an MMTP packet and an NTP packet.
Descrambling mechanism 1604 descrambles the MMTP packets if they are scrambled. The descrambling unit 1604 acquires scramble key information for descrambling from the EMM part, ECM part, and the like in the control information analysis unit 1605. Here, in the case where the scramble packet is scrambled in units of IP packets, the descrambling unit 1604 descrambles the IP packets.
The NTP clock generation mechanism 1606 reproduces the NTP clock from the NTP packet.
The MMT demultiplexing unit 1607 performs filtering processing of components such as video, audio, subtitle, character superimposition, and application, or control information based on the packet ID stored in the MMTP packet header. The MMT demultiplexing means 1607 obtains the time stamp descriptor stored in the MP table (the time stamp descriptor is written when referring to both the MPU time stamp descriptor and the MPU extended time stamp descriptor) from the control information analysis means 1605, and calculates the PTS and DTS for each access unit in the PTS/DTS calculation means.
The media processing unit 1608 converts the video, audio, subtitle, character superimposition, application, and the like filtered from the MMTP packet into data of a unit to be presented. The unit to be presented is specifically a NAL unit or an access unit of a video signal, a presentation unit of an audio frame or a subtitle, or the like.
The decoding presentation mechanism 1609 decodes and presents the access unit at the time when the PTS/DTS of the access unit matches, based on the time information of the NTP clock. When the respective processing units are mounted as different LSIs, different devices, and different devices (e.g., televisions, displays, recording devices, set-top boxes), the decoding presentation mechanism 1609 connects the respective processing blocks using an external output mechanism, a communication mechanism, an interface mechanism 1610, and the like.
Here, the structure of the receiving apparatus 1600 is not limited to the structure shown in fig. 115. The IP multiplexing scheme includes, in addition to MMT, DASH/ROUTE scheme, TS over IP (TS over IP) scheme, and the like, and broadcast transmission schemes corresponding to the IP multiplexing scheme include DVB-T2, DVB-S2, DVB-C2, ATSC3.0, and the like. As a Layer2 (second Layer) Protocol which serves as an interface between the IP multiplexing scheme and the broadcast transmission scheme, there are a GSE scheme, an ALP (ATSC Link-Layer Protocol) scheme, and the like, in addition to the TLV scheme.
Further, an interface standard using an IP multiplexing method, a communication standard, a communication multicast, and the like are available. In these cases, the receiver apparatus has the same configuration as that shown in fig. 115.
Fig. 116 is a diagram showing a general broadcast protocol multiplexed in the MPEG-2 TS system (hereinafter referred to as "TS system").
In the TS system, video, audio, caption/character superimposition, and application are stored in PES or segment form, and are packetized and transmitted.
In conventional ARIB terrestrial digital broadcasting and BS digital broadcasting, ARIB caption coding is performed for caption/character superimposition. In addition, the application is data broadcast BML encoded.
Fig. 117 is a block diagram showing a receiving apparatus that receives broadcast signals multiplexed in the TS scheme.
The reception device 1700 decodes the transmission channel encoded data for the broadcast signal received by the tuner 1701 in the demodulation section 1702, and extracts a TS packet by applying error correction or the like to the decoded data.
The descrambling mechanism 1703 descrambles the scrambling if the TS packet is scrambled. The descrambling unit 1703 acquires scramble key information for descrambling from the EMM unit, ECM unit, and the like in the control information analysis unit 1704.
The reference clock generating unit 1705 extracts a PCR from the TS packet, and reproduces a PCR clock.
The TS demultiplexing section 1706 performs filtering processing of components such as video, audio, subtitle, character superimposition, and application, or control information based on a PID (Packet Identifier) stored in the TS Packet header.
The media processing unit 1707 converts video, audio, subtitles, character superimposition, application, and the like filtered from the TS packet into data of a unit to be presented, and further extracts a time stamp (PTS, DTS) from the PES packet header. The unit to be presented is a NAL unit or an access unit of a video signal, a presentation unit of an audio frame or a subtitle, or the like.
The decoding presentation means 1708 decodes and presents the access unit at the timing when the PTS/DTS of the access unit matches, based on the time information of the PCR clock. When each processing unit is mounted as a different LSI, a different device, or a different device (for example, a television, a display, a recording device, a set-top box), each processing block is connected using an external output unit, a communication unit, an interface unit 1709, and the like.
Here, the structure of the receiving apparatus 1700 is not limited to the configuration shown in fig. 117. As a transmission method corresponding to the TS method, there are a satellite broadcast, terrestrial broadcast, cable broadcast transmission method, interface standard, communication multicast method, and the like standardized by DVB, ISDB, ATSC, and the like. In these embodiments, the receiving apparatus has the same configuration as that of fig. 117.
The broadcast or communication multicast method includes two broad types of methods, i.e., an IP multiplexing method described in fig. 114 and 115 and a TS method described in fig. 116 and 117. The receiving apparatus includes a receiving apparatus corresponding to a single-system reception process, a receiving apparatus corresponding to a plurality of-system reception processes, and the like, and examples thereof are shown below.
Fig. 118A, 118B, 118C, 118D, 118E, and 118F are configuration examples of a receiving apparatus, the first mode is, for example, the MMT/TLV mode (first multiplexing mode) described in fig. 114 and 115, and the second mode is, for example, the TS mode (second multiplexing mode) described in fig. 116 and 117. However, the first mode may be the reverse of the second mode, or three or more modes may be mixed. The processing means a and the processing means B are processing means combining functions of a part of the plurality of processing means described in fig. 115 and 117. Here, an example of roughly dividing the processing means a and the processing means B is shown, and any of the plurality of processing means described in fig. 115 and 117 may be combined, or a processing means obtained by further dividing the processing of one processing means among the plurality of processing means described in fig. 115 and 117 may be combined, or a processing means obtained by combining a plurality of processing means may be composed of three or more processing means.
Fig. 118A is a diagram showing a configuration of a conventional receiving apparatus in a case where a single system is processed. In this case, the receiving apparatus processes the signal transmitted in the first mode by the first mode processing means a1811, and then processes the signal by the first mode processing means B1812.
Fig. 118B is a diagram showing a configuration of a conventional receiving apparatus that independently processes a plurality of systems. In this case, the receiving apparatus is configured such that the signal transmitted in the first mode is processed by the first mode processing means a1821, then processed by the first mode processing means B1822, and the signal transmitted in the second mode is processed by the second mode processing means a1823, then processed by the second mode processing means B1824.
In contrast, the configuration in the case of using the mode conversion mechanism as the conversion portion is as follows.
Fig. 118C is a diagram showing an example of the configuration of the receiving apparatus in the case of using the mode switching mechanism. In this case, the receiving apparatus processes the signal transmitted in the first mode by the first mode processing means a1831, converts the output of the first mode processing means a1831 into the second mode by the mode conversion means 1832, and processes and thereafter by the second mode processing means B1833.
Fig. 118D is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism. In this case, the receiving apparatus processes the signal transmitted in the first mode by the first mode processing means a1841, then processes a part of the output of the first mode processing means a1841 by the first mode processing means B1842, converts the remaining part into the second mode by the mode conversion means 1843, and processes the processed signal by the second mode processing means B1844.
Fig. 118E is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism. In this case, the receiving apparatus receives the signal transmitted in the first mode and the signal transmitted in the second mode, the signal transmitted in the first mode is processed by the first mode processing means a1851, the output of the first mode processing means a1851 is converted into the second mode by the mode converting means 1852, and the processed signal is processed by the second mode processing means B1854. On the other hand, in the receiving apparatus, the second mode processing means a1853 processes the signal transmitted in the second mode, and then the second mode processing means B1854 processes the output of the second mode processing means a 1853. That is, the receiving apparatus processes both the signal transmitted in the first mode and the signal transmitted in the second mode by the second mode processing means B.
Fig. 118F is a diagram showing another example of the configuration of the receiving apparatus in the case of using the mode switching mechanism. In this case, the receiving apparatus receives a signal transmitted in the first mode and a signal transmitted in the second mode, processes the signal transmitted in the first mode processing means a1861, then processes a part of the output of the first mode processing means a1861 in the first mode processing means B1862, converts the remaining part into the second mode in the mode conversion means 1863, and processes the second mode processing means B1865 and thereafter. On the other hand, in the receiving apparatus, the signal transmitted in the second mode is processed by second mode processing means a1864, and then the output of second mode processing means a1864 is processed by second mode processing means B1865. That is, the receiving apparatus processes a part of the signal transmitted in the first mode and the signal transmitted in the second mode by the second mode processing means B1865.
However, as shown in fig. 118D or 118F, when the signal transmitted in the first mode is processed by both of the first mode processing means B1842 and 1862 and the second mode processing means B1844 and 1865, information (for example, time information, clock information, or the like) for synchronization may be shared between the first mode processing means B1842 and 1862 and the second mode processing means B1844 and 1865. That is, in this case, adjustment for matching one process with the other process may be performed between the first process means B1842 and 1862 and the second process means B1844 and 1865.
For example, the first-mode processing means a1831, 1851, 1861 and the second-mode processing means a1853, 1864 which perform the preceding-stage processing in fig. 118C to 118F are first processing units which perform the following processing: the broadcast signal in which the multiplexed data is modulated is received, the received broadcast signal is demodulated, and the multiplexed data obtained by the demodulation is output. The multiplexed data referred to herein is data composed of at least first multiplexed data of a first multiplexing scheme and second multiplexed data of a second multiplexing scheme different from the first multiplexing scheme.
The mode conversion mechanisms 1832, 1843, 1852, 1863 are conversion units that convert the multiplexing mode of the first multiplexed data among the multiplexed data output from the first mode processing mechanisms a1831, 1851, 1861 as the first processing units into the second multiplexing mode and output the converted data obtained by the conversion.
For example, second mode processing means B1833, 1844, 1854, 1865 in fig. 118C to 118F, which perform processes after the mode conversion means 1832, 1843, 1852, 1863, are second processing units that perform decoding processing for decoding the converted data output by the mode conversion means 1832, 1843, 1852, 1863 and output decoded data obtained by the decoding processing.
The receiving apparatus having the mode switching mechanisms 1832, 1843, 1852, 1863 as above has the following effects.
For example, by providing the mode conversion mechanisms 1832, 1843, 1852, 1863, it is not necessary to perform all the subsequent processes by the first mode processing means B, and at least a part of the subsequent processes can be processed by the second mode processing means. For example, when only the second processing means B is installed as it is, the first processing means B may not be newly installed. Alternatively, for example, in the case where the second mode processing means B is a device (hardware or software) having high performance, the signal requiring only high-performance processing is converted by the mode conversion means and processed by the second mode processing means B, and on the other hand, only the signal not requiring high-performance processing is processed by the first mode processing means B, so that the first mode processing means B becomes unnecessary for high performance. Therefore, the circuit scale can be reduced, and the cost can be reduced. Here, the second-mode processing means B requiring highly accurate processing is installed by using hardware, and the first-mode processing means B not requiring high-performance processing is installed by using software in the second-mode processing means B, whereby the second-mode processing means B can be installed by using hardware having only the function of the second-mode processing means B. In addition, since various configurations can be flexibly realized, it is possible to realize mounting effective for compatibility with an existing system, a circuit scale, and cost reduction.
Fig. 119 is a diagram showing a modification of the configuration of the receiver in the case of using the mode switching mechanism.
Here, the first mode is an MMT/TLV mode, and the second mode is a TS mode.
As shown in fig. 119, the receiving apparatus 1900 includes a first mode processing means a1910, a mode switching means 1920, and a mode processing means B1930.
The first-mode processing means a1910 receives the MMT/TLV-mode signal, performs processing such as TLV multiplexing, outputs a part of the MMTP packet signal to the first-mode switching mechanism 1921 as a first switching unit, and outputs the remaining part of the MMTP packet signal to the second-mode switching mechanism 1922 as a second switching unit.
The mode conversion mechanism 1920 includes a first mode conversion mechanism 1921, a second mode conversion mechanism 1922, and a second mode multiplexing mechanism 1923 as a multiplexing unit. The first conversion mechanism 1921 stores the data structure of the MMTP packet in the TS packet, and outputs the TS packet to the second multiplexing mechanism 1923. That is, the first conversion means extracts a part of the first data of the first multiplexed data, performs first conversion of storing a first packet (MMTP packet) constituting the first data in a second packet (TS packet) used in the second multiplexing method, and outputs first converted data constituted by the second packet obtained by the first conversion.
The second conversion mechanism 1922 extracts data from the other MMT packet, converts the data into a TS packet, and outputs the TS packet to the second multiplexing mechanism. That is, the second mode converting mechanism 1922 extracts a first packet (MMTP packet) of the second data constituting the remaining part of the first multiplexed data, performs second conversion of the extracted first packet into a second packet (TS packet) of the second multiplexing mode, and outputs second converted data constituted by the second packet obtained by the second conversion.
Here, the first data and the second data may be data of different media. For example, the first data may be data representing video and audio, and the second data may be data representing subtitles, applications, and the like.
The second mode multiplexing unit 1923 multiplexes and outputs the data in the TS packet format output from the first mode switching unit 1921 and the second mode switching unit 1922 by using the TS method. That is, the second multiplexing unit 1923 performs multiplexing processing for multiplexing the first converted data and the second converted data that are output.
In this manner, the mode switching mechanism 1920 outputs a signal as one interface by multiplexing the signals using either one of the first mode and the second mode. That is, the mode switching mechanism 1920 outputs the data obtained by the multiplexing process as the switching data.
The second-mode processing means B1920 includes second-mode inverse multiplexing means 1931, first-mode processing means B1932, and second-mode processing means B1933 as inverse multiplexing units. The second-mode inverse multiplexing unit 1931 receives the TS-multiplexed signal and performs inverse multiplexing. That is, the second mode inverse multiplexing means 1931 performs inverse multiplexing processing for inversely multiplexing the conversion data output by the mode conversion means 1920 into the first conversion data and the second conversion data.
The second inverse multiplexing processing means 1931 converts the TS packet converted by the first conversion means 1921, which constitutes the first conversion data obtained by inverse multiplexing, into an MMTP packet, and outputs the MMTP packet to the first processing means B. The second inverse multiplexing processing means 1931 outputs the TS packet converted by the second conversion means 1922, which constitutes the second conversion data obtained by inverse multiplexing, to the second processing means B1933.
Then, the first-mode processing means B1932 processes the MMTP packet. Specifically, the first-mode processing means B1932 performs the first decoding process in the first multiplexing scheme on the first data constituted by the first packet (MMTP packet) extracted from the second packet (TS packet) constituting the first converted data obtained by the inverse multiplexing process.
Further, the second-mode processing means B1933 processes TS packets. Specifically, the second scheme processing means B1933 performs the second decoding process in the second multiplexing scheme on the second data constituted by the second packet (TS packet) constituting the second converted data obtained by the inverse multiplexing process.
The mode processing means B1930 may include: and an adjusting unit that adjusts the first decoded data and the second decoded data by performing adjustment to match one of the first control information and the second control information with the other using the first control information of the first decoded data and the second control information of the second decoded data. Specifically, the first control information is first reference clock information, the second control information is second reference clock information, and the adjusting unit synchronizes the first decoded data and the second decoded data by performing adjustment to match one of the first reference clock information and the second reference clock information with the other.
In this manner, the mode processing means B1930 can also output, as decoded data, the first decoded data obtained by the first decoding process and the second decoded data obtained by the second decoding process.
Fig. 120 is a diagram showing an example of the details of the mode switching mechanism.
As shown in fig. 120, the first-mode processing means a2010 includes a tuner 2011, a demodulation means 2012, a TLV/IP/UDP multiplexing means 2013, a first-mode descrambling means 2014, a control information analysis means 2015, a first reference clock generation means 2016, an MMT multiplexing means 2017, and the like. The tuner 2011, the demodulating unit 2012, the TLV/IP/UDP multiplexing unit 2013, the first descrambling unit 2014, the control information analyzing unit 2015, the first reference clock generating unit 2016, and the MMT multiplexing unit 2017 correspond to the tuner 1601, the demodulating unit 1602, the TLV/IP/UDP multiplexing unit 1603, the descrambling unit 1604, the control information analyzing unit 1605, the NTP clock generating unit 1606, and the MMT multiplexing unit 1607 of the receiving apparatus 1600 described in fig. 115, respectively, and therefore, the description thereof is omitted.
The mode conversion mechanism 2020 includes a media conversion mechanism (the first media conversion mechanism 2021 and the second media conversion mechanism 2022), a clock generation mechanism (the second reference clock generation mechanism 2023 and the differential information generation mechanism 2024), a second mode multiplexing mechanism 2025, a second mode scrambling mechanism 2026, and the like.
The mode switching mechanism 2020 may use a different media switching mechanism for each media. For example, the mode switching mechanism 2020 may be configured to: the second mode conversion mechanism 1922 of the mode conversion mechanism 1920 described with reference to fig. 119 is used for the first media (e.g., video and audio), and the first mode conversion mechanism 1921 of the mode conversion mechanism 1920 is used for the second media (e.g., subtitles).
The mode conversion means 2020 generates and multiplexes the difference information between the first reference clock information and the second reference clock in the reference clock information conversion means and the difference information generation means. That is, the adjustment unit described above may be provided in the mode conversion mechanism 2020.
A specific example of a mode conversion mechanism for converting a signal received in the MMT/TLV mode into the TS mode will be described below.
(video and audio)
A case where the mode conversion mechanism converts video data and audio data from the first mode to the second mode will be described.
In the case of conversion by the second conversion mechanism 1922, the video signal and the audio signal stored in the MMTP packet to be input are converted in units of access units, an Elementary Stream (ES) is generated, and then a PES header is added to the elementary stream to repackage the elementary stream. Specifically, the second conversion mechanism 1922 extracts data stored in the MMTP packet, divides the data when the MFUs are aggregated, and receives and combines all the data when the MFUs are fragmented.
When the MFU of a video signal is in NAL units, a plurality of NAL units are further combined to form an access unit. Here, when a video signal is transmitted with the resource type 'hvc 1', the access unit is configured by acquiring non-VCL NAL units from the MFU, and when a video signal is transmitted with the resource type 'hev 1', the access unit is configured by acquiring non-VCL NAL units from MPU metadata separately transmitted. Further, in the case of the NAL size format, a 4-byte NAL size area is deleted, and a byte start code is given to the area, so that the area is converted into a byte stream and output.
In the case where the MFU of the sound signal is in units of frames, the MFU is set in units of access units. When the MFU is in the form of a LATM/LOAS stream and is in the form of AudioMuxConfig (), synchronization information and length information may be added to convert the MFU into the form of audiogyncisterm (), and when the MFU is in the form of a Raw data stream, AudioMuxConfig () may be separately acquired from MPU metadata, control information, or the like and converted into the form of a LATM/LOAS stream. But also into other stream forms.
(time stamp information)
A case where the mode switching mechanism switches the time stamp information from the first mode to the second mode will be described.
Time stamp information (decoding time or presentation time, or information for calculating decoding time or presentation time) of video, audio, or the like is acquired from the received MMTP packet, and converted into PTS and DTS for each access unit. For example, the control information analysis means acquires presentation times of access units whose presentation order is the head in the MPU from the MPU timestamp descriptor, and calculates PTSs and DTSs of all access units in the MPU by using the MPU extended timestamp descriptor as well. The calculated DTS and PTS are stored in the header of a PES packet including the corresponding video or audio access unit, and a PES packet is generated and output. Note that, in the first conversion mechanism 1921, the MMTP packet storing the synchronous MPU/MFU may be stored in a PES packet or a TS packet and output, instead of being converted into a PES packet. The transport method at the time of output may be stored in a PES packet as a private stream, may be transported in a private area of a TS packet, may be segmented as control information, may be ethernet-transported in an IP packet, may be transported using a communication means such as USB or I2C, or may be any transport means.
(reference clock information)
A case where the mode conversion mechanism converts the reference clock information from the first mode to the second mode will be described. For example, the operations of the first reference clock generation means 2016, the second reference clock generation means 2023, and the differential information generation means 2024 shown in fig. 120 will be described.
In the MMT/TLV scheme, the time stamp is based on the transmission system clock synchronized with the UTC, and the time stamp value based on the system clock is stored in the NTP packet as reference clock information and transmitted in order to notify the system clock information. The receiving device reproduces the receiver system clock based on the reference clock information stored in the NTP packet in the first reference clock generating means 2016 of the first processing means a 2010. The mode switching unit 2020 may convert system clock information (first reference clock information) reproduced by an NTP packet into STC (second reference clock information), store a time stamp (PCR) thereof in an adjustment field of a TS packet, and output the resulting value, or may directly convert a time stamp value of an NTP packet into a PCR value and store the PCR value in the TS packet, and output the resulting value. For example, the manner conversion mechanism 2020 converts from NTP having accuracy of power of 2 to PCR having accuracy of 27 MHz. For example, the mode conversion unit 2020 extracts a specific 33bit from 64-bit NTP and sets the extracted result as PCR.
Leap second adjustment of reference clock information
Next, a leap second adjustment method in the case where leap second adjustment is performed on the reference clock information stored in the NTP packet will be described. When a leap _ indicator indicating leap second adjustment is indicated in an NTP packet, the receiving apparatus corrects PTS and DTS based on an MPU _ presentation _ time _ leap _ indicator stored in an MPU extended time stamp descriptor, and then stores the corrected PTS and DTS in a PES packet header. Alternatively, instead of correcting the PTS and DTS, mpu _ presentation _ time _ leap _ indicator information may be separately generated and output, or information indicating that leap second correction is necessary may be generated and output for the PTS and DTS of each access unit. The system processing means B1930 that received the PTS and DTS for which leap second correction was not performed corrects the PTS and DTS based on the information indicating that leap second correction is necessary. In the leap second adjustment, the leap second adjustment of ± 1 second may be performed simultaneously with the system clock of the receiver, or the time stamp may be corrected without performing the leap second adjustment on the system clock of the receiver. When the STC is discontinuous for ± 1 second at leap second adjustment, the PCR discontinuity flag in the TS packet header is set to valid. Note that, a System Time Clock (STC) of 27MHz may be independently generated in the receiving apparatus, and the time stamp value thereof may be transmitted as PCR. In this case, video and audio PTSs and DTSs are assigned to the independently generated STC and stored in a PES packet header. When the leap _ indicator indicating leap second adjustment is shown in the NTP packet, leap second adjustment may not be performed on the PTS and DTS based on the independently generated system clocks.
However, in the case where the first mode processing means B1932 and the second mode processing means B1933 are provided as in the mode processing means B1930 of fig. 119, the receiving apparatus may have two systems of clocks, an NTP clock generated based on an NTP packet and a System Time Clock (STC) generated independently, and the differential information generating means may generate differential information indicating a time correspondence relationship between the clocks. In this case, the generated difference information and a time stamp (PCR) based on the independently generated STC are multiplexed and output. The system processing means B1930 receives the PCR and reproduces the STC. Since the media processed by the second mode processing means B1933 is provided with PTS and DTS based on STC (second reference clock information), decoding and presentation are performed at timings at which the PTS and DTS match the second reference clock information. On the other hand, since the media processed by the first processing means 1910 is provided with PTS and DTS based on NTP (first reference clock information), the media is converted into time corresponding to the second reference clock information based on the difference information, and is decoded or presented at timing that matches the second reference clock information. When leap second adjustment is performed on the NTP clock, a leap _ indicator indicating that leap second adjustment is performed may be output together with the difference information.
The transport method at the time of output may be stored in a PES packet as a private stream, may be stored in a private area of a TS packet, may be segmented as control information, may be stored in an IP packet and may be ethernet-transferred, may be transferred using a communication means such as USB or I2C, or may be any transport means. In these cases, an identifier that can identify whether the information is information indicating the time correspondence between clocks or information indicating the leap second correction flag is provided and output. The leap second flag is also included in the reference _ start _ time in the caption or the UTC-NPT reference descriptor in the event message, and leap second information and difference information are generated and output by the same method.
When the data reception timings of the access unit and the time stamp information of the access unit are different when a PES packet is generated, one of the data is accumulated, and the PES packet is generated at a timing when both the data are aligned. For example, when it is guaranteed that the time stamp information of a certain access unit can be received N seconds before the decoding time of the access unit, the access unit is accumulated until N seconds before the decoding time of the access unit, and a PES packet is generated and output N seconds before the decoding time of the access unit.
(Caption)
The explanation will be given of a case where the mode switching mechanism switches from the first mode to the second mode by superimposing the subtitles and characters stored in the inputted MMTP packet on the TS packet.
The input data is, for example, in the form of TTML encoded subtitle data stored in a synchronous MPU/MFU and stored in an MMTP packet. The second mode conversion mechanism 1922 extracts TTML data from the MMTP packet and converts the TTML data into a segment format in the TS mode. The segment form is output in a data carousel manner.
The first conversion unit 1921 does not convert the data into the TS segment format, but stores the MMTP packet storing the synchronous MPU/MFU in the PES packet or the TS packet and outputs the PES packet or the TS packet. The transport method at the time of output may be stored in a PES packet as a private stream, may be stored in a private area of a TS packet, may be segmented as control information, may be stored in an IP packet and may be ethernet-transported, may be transported using a communication means such as USB or I2C, or may be any transport means.
(applications)
The description will be given of a case where the mode switching mechanism switches from the first mode to the second mode by storing application data stored in an inputted MMTP packet in a TS packet.
The input data is, for example, in the form of application data encoded in HTML5, stored in an asynchronous MPU/MFU, and stored in an MMTP packet. The second mode conversion unit 1922 extracts HTML data from the MMTP packet and converts the HTML data into a segment format in the TS mode. The segment form is output in a data carousel manner.
The first conversion unit 1921 does not convert the packet into the TS section format, but stores the MMTP packet in which the asynchronous MPU/MFU is stored in the PES packet or the TS packet and outputs the PES packet or the TS packet. The transport method at the time of output may be stored in a PES packet as a private stream, may be stored in a private area of a TS packet, may be segmented as control information, may be stored in an IP packet and may be ethernet-transported, may be transported using a communication means such as USB or I2C, or may be any transport means.
Among them, control information (e.g., MH-AIT, application control information descriptor, data transfer message, etc.) related to application transfer is also converted into a TS form and output. The control information inputted in the form of MMT message may be converted into TS segment form and outputted, or the control information inputted in the form of MMT message may be stored as private stream in PES packet, or may be stored in private area of TS packet and transmitted. However, only the M2 segment message may be converted into the TS segment format and output, and other messages may be output without being converted into the TS segment format.
(control information)
The TS system includes PAT and PMT, and the MMT/TLV system includes PLT, MPT, AMT, and the like, as control information to be an entry point of a service.
In the case where there is a component using the second mode conversion mechanism 1922, the control information is converted so as to include information of the component. For example, when video and audio are converted by the second conversion mechanism 1922 and subtitle and application are converted by the first conversion mechanism 1921, information on the service and information on the video and audio are converted from the MPT to the PMT. At the same time, a PAT representing PMT position information is generated and output.
When there is a module using the first conversion mechanism 1921, control information is stored in a PES packet or a TS packet and output.
(packet header conversion)
In the header of the MMTP packet, a packet counter indicating the continuity of the MMTP packet, and a packet sequence number indicating the continuity of each resource, a slice counter indicating the continuity of the MFU after fragmentation, or a slice identifier are shown. These fields indicating continuity shown in the inputted MMTP packet are monitored, and when a packet loss due to a transmission error or the like is detected, a continuity counter (continuity counter) in the TS packet is set to a discontinuous value and outputted. However, the transport error flag in the TS packet may be set to 1, and invalid data may be generated and output, or an empty TS packet may be inserted and output so as to match the transport rate. Alternatively, when a packet loss occurs, the data may be restored by copying the previous data, for example. The restored data is stored in a PES packet and a TS packet and output. In this case, a continuous value is set in the continuity index and output.
The packet ID (16 bits) indicated in the MMTP packet is converted into the PID (13 bits) of the TS packet. For example, a rule or table for associating packet IDs and usable areas of PIDs is prepared in advance, and conversion is performed according to the table. When the MMTP packet is stored as private data in the PES packet or the TS packet for each MMTP packet header using the first conversion mechanism 1921, a PID indicating that the MMTP packet is stored as it is with the data structure of the MMTP packet is assigned to the PID in the TS packet and output, regardless of the data stored in the MMTP packet. When receiving a TS packet to be output, if the TS packet is given a PID indicating that the data structure of the MMTP packet is maintained and stored as it is, the MMTP packet is extracted from the TS packet and MMTP multiplexing processing is performed.
In the case of video and audio, a payload unit start identifier is set to 1 in a TS packet including the head of an access unit. In an MMTP packet for transmitting a video, the start of an NAL unit is identified by whether fragment _ indicator is 00 or 01, and an MMTP packet including the start of an access unit is identified by whether the header type of the NAL unit is an AU delimiter, whereby a payload unit start identifier is set to 1 in the start packet when the MMTP packet is converted into a TS packet.
In an MMTP packet for transmitting a voice, the start of an access unit is identified based on whether a fragment _ indicator is 00 or 01, and a payload unit start identifier is set to 1 in the start packet when the MMTP packet is converted into a TS packet.
In the case where the MMTP packet is scrambled, descrambling is performed, and re-scrambling is performed in a scrambling manner in the TS manner.
When an MMTP packet is scrambled, a multi-type extension header is assigned to an extension header area in the MMTP packet, and if the multi-extension header type is an extension header for scrambling, information indicating whether the MMTP packet is scrambled or not and whether a scrambling key is an even key or an odd key is stored.
When the MMTP header is directly converted into the TS header, scrambling control bits in the extension header of the MMTP packet are converted into transport scrambling control bits in the TS packet. However, when the extension header for scrambling is not shown in the MMTP packet, the transport scrambling control bit in the TS packet is set to 0 as unscrambled.
In the case of re-scrambling, the first scrambling scheme may be used instead of the second scrambling scheme.
Fig. 121 is a diagram showing a flow of reception processing for performing reception using the reception apparatus shown in fig. 118F.
The receiving device 1860 receives signals transmitted in different plural ways (a signal transmitted in a first way and a signal transmitted in a second way) (S3701).
Next, the receiving device 1860 determines whether the received signal is of the first scheme or the second scheme (S3702).
When the received signal is of the second scheme (the second scheme in S3702), the receiving apparatus 1860 executes processing using the second scheme processing means a1864 (S3703).
On the other hand, when the received signal is of the first format (the first format in S3702), the receiving apparatus 1860 executes processing using the first format processing means a1861 (S3705).
Next, the receiving device 1860 determines whether the output to the first-mode processing means a1861 is processed by the first-mode processing means B1862 or processed by the second-mode processing means B1865 (S3706).
When it is determined that the processing is performed by the second mode processing means B1865 (yes at step S3706), the receiving device 1860 switches the first mode to the second mode in the mode switching means 1863 (S3707).
Then, the receiving device 1860 performs processing on both the signal of the second method processed by the second method processing means a1864 and the signal converted from the signal of the first method into the signal of the second method by the method conversion means 1863 by using the second method processing means B1865 (S3704).
On the other hand, when the reception device 1860 determines that the processing is performed using the first-mode processing means B1862 (no in step S3706), the processing is executed using the first-mode processing means B1962 (S3708).
As described above, in the receiving device 1860 that receives signals of a plurality of different modes (the first mode and the second mode), by providing the mode switching mechanism 1863, it is possible to execute processing in the second mode processing mechanism B1865 for any mode of signal.
However, when there is no signal to be processed by the first processing means B, the steps S3706 and S3708 may be omitted, and the output of the step S3705 may be switched to the reception processing flow of the step S3707.
Fig. 122 is a diagram showing a processing flow of the mode switching mechanism 1920 in fig. 119.
The mode switching mechanism 1920 acquires the signal output from the first mode processing means a1910 (S3801).
Next, it is determined whether or not the acquired signal is a signal processed by the second aspect processing means B1933 (S3802).
If the signal is processed by the second mode processing means B1933 (yes at step S3802), the mode conversion means 1920 converts the packet of the first mode into the packet of the second mode at the second mode conversion means 1922 (S3803). Then, the mode switching mechanism 1920 gives an identifier for identifying the component, the control information, or the like (S3804). That is, the second conversion mechanism 1922 adds a second identifier indicating that the second packet (TS packet) obtained by the second conversion to the second packet of the second conversion data.
On the other hand, when the signal is not processed by the second mode processing means B1933 (no in step S3802), the mode switching mechanism 1920 stores the packet of the first mode in the packet of the second mode in the first mode switching mechanism 1921 (S3805). The mode switching mechanism 1920 assigns an identifier indicating that the packet of the first mode is stored in the packet of the second mode (S3806). That is, the first conversion mechanism 1921 adds the first identifier indicating the second packet (TS packet) obtained by the first conversion to the second packet of the first conversion data.
The second scheme multiplexing means 1923 multiplexes the packets packed in the second scheme and given the identifiers in steps S3801 to S3806, and outputs the multiplexed packets to the scheme processing means B1930 (S3807).
Fig. 123 is a diagram showing a process flow of the mode processing means B1930 in fig. 119.
The mode processing means B1930 acquires the signal multiplexed in the second mode output from the mode switching mechanism 1920 (S3901).
Next, the mode processing means B1930 determines whether or not the packet is a packet storing the packet of the first mode, based on the packet identifier (S3902).
When the packet is a packet in which the packet of the first format is stored (yes in step S3902), the format processing means B1930 extracts the packet of the first format and executes the process by the first format processing means B1932 (S3903).
On the other hand, when the packet is not a packet in which the packet of the first format is stored (no in step S3902), the format processing means B1930 extracts the packet of the second format and executes the process by the second format processing means B1933 (S3904).
The receiving devices 1830 to 1860 in fig. 118C to 118F may be configured such that the second-mode transmission means serving as a retransmission unit, not shown, is included or replaced in the second-mode processing means B1833, 1844, 1854, 1865. In this case, the signal received in the first mode can be retransmitted in the second mode.
That is, the following retransmission unit may be provided: the converted data output from the mode conversion mechanisms 1832, 1843, 1852, 1863 serving as conversion units is retransmitted to other receiving apparatuses.
However, the accumulating apparatus may be configured such that the second-mode accumulating means, which is not shown, is included in or replaced in the second-mode processing means B1833, 1844, 1854, 1865 of the receiving apparatuses 1830 to 1860 of fig. 118C to 118F. In this case, the signal received in the first mode can be accumulated in the second mode.
That is, the following accumulation unit may be provided: the conversion data output by the mode conversion mechanisms 1832, 1843, 1852, 1863 serving as conversion units is stored in the storage device. The storage device is realized by an auxiliary storage device including, for example, a hard disk drive, a nonvolatile memory, and the like.
In the receiving devices 1830 to 1860 in fig. 118C to 118F, the processing is performed by two processing means, that is, the processing on the front stage side and the processing on the rear stage side. That is, when the retransmission unit is provided, the second-mode processing means B1833, 1844, 1854, 1865 and the first-mode processing means 1842, 1862 on the subsequent stage side may not be provided. That is, these second-mode processing means B1833, 1844, 1854, 1865 and first-mode processing means 1842, 1862 on the subsequent stage side may be configured to be provided in another receiving apparatus.
Specifically, the signal transmitted from the retransmission apparatus according to the second embodiment and the signal reproduced from the accumulation apparatus according to the second embodiment can be reproduced by the second embodiment processing means B1833, 1844, 1854, 1865 provided in the other reception apparatus. The accumulating apparatus may further include a retransmitting unit operable to retransmit the converted data accumulated in the storage apparatus to another receiving apparatus. However, in addition to the method of accumulating the signals processed by the mode conversion mechanisms 1832, 1843, 1852, 1863, the signals of the first mode before being processed by the mode conversion mechanisms may be accumulated, and the signals may be reproduced by the mode conversion mechanisms and the second mode processing mechanisms when reproducing the accumulated signals.
[ supplement: receiving device
The receiving apparatus may be configured as shown in fig. 124. The receiving apparatus may be configured as shown in fig. 125. Fig. 124 shows an example of a specific configuration of a receiving apparatus. Fig. 125 is a diagram showing another example of the specific configuration of the receiving apparatus.
The reception device 2100 includes a first processing unit 2101 and a conversion unit 2102. The first processing unit 2101 and the conversion unit 2102 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like. That is, each processing section constituting the receiving apparatus 2100 may be realized by software or hardware.
The receiving device 2200 includes a receiving unit 2201, an inverse multiplexing unit 2202, a first decoding unit 2203, a second decoding unit 2204, and an output unit 2205. The reception unit 2201, the inverse multiplexing unit 2202, the first decoding unit 2203, the second decoding unit 2204, and the output unit 2205 are each realized by, for example, a microcomputer, a processor, a dedicated circuit, or the like. That is, each processing unit constituting the receiving apparatus 2200 may be realized by software or hardware.
The detailed description of each component of the receiving apparatuses 2100 and 2200 is given in the description of the corresponding receiving method.
First, a reception method of the reception apparatus 2100 will be described with reference to fig. 126. Fig. 126 shows an operation flow of the receiving apparatus (receiving method).
First, the first processing unit 2101 of the reception device 2100 receives a broadcast signal in which multiplexed data composed of at least first multiplexed data of a first multiplexing system (MMT/TLV system) and second multiplexed data of a second multiplexing system (TS system) different from the first multiplexing system is modulated, demodulates the received broadcast signal, and outputs the demodulated multiplexed data (S4001).
Next, the conversion section 2102 of the reception apparatus 2100 converts the multiplexing scheme of the first multiplexing data among the output multiplexing data into the second multiplexing scheme, and outputs the converted data obtained by the conversion (S4002).
Thus, if the data converted into the second multiplexing scheme by the receiving apparatus 2100 is used, the conventional TS scheme processing unit can be used as it is. Therefore, in the subsequent process, the common use with the conventional installation can be easily realized, and the low cost can be realized.
Next, a reception method of the reception apparatus 2200 will be described with reference to fig. 127. Fig. 127 is an operation flow of the receiving apparatus (receiving method).
First, the reception unit 2201 of the reception device 2200 receives converted data in which first converted data composed of a first packet (MMTP packet) storing first multiplexed data constituting a first multiplexing scheme (MMT/TLV scheme) and a second packet (TS packet) used in a second multiplexing scheme (TS scheme) different from the first multiplexing scheme and second converted data composed of a second packet obtained by converting the first packet into the second multiplexing scheme are multiplexed (S4101).
Next, the inverse multiplexing unit 2202 of the reception device 2200 performs inverse multiplexing processing for inverse multiplexing the converted data received by the reception unit 2201 into first converted data and second converted data (S4102).
Next, the first decoding unit 2203 of the reception device 2200 extracts a first packet from the second packet constituting the first converted data obtained by the inverse multiplexing processing, and performs the first decoding processing for the first data constituted by the extracted first packet in the first multiplexing scheme (S4103).
The second decoding unit 2204 of the reception device 2200 performs a second decoding process on second data, which is composed of a second packet constituting second converted data obtained by the inverse multiplexing process, in a second multiplexing scheme (S4104).
Finally, the output unit 2205 of the receiving apparatus 2200 outputs the first decoded data obtained by the first decoding process and the second decoded data obtained by the second decoding process (S4105).
In this way, the receiving apparatus 2200 can use the data converted into the second multiplexing scheme, and thus, for example, the processing unit of the conventional TS scheme can be used as it is. Therefore, the receiving apparatus 2200 can be easily shared with the existing installation, and can be realized at low cost.
(other embodiments)
The above describes the transmission device, the reception device, the transmission method, and the reception method according to the embodiments, but the present disclosure is not limited to the embodiments.
Each processing unit included in the transmission device and the reception device according to the above-described embodiments is typically realized as an LSI (large scale integrated circuit) which is an integrated circuit. They may be independently integrated into a single chip, or may be partially or entirely integrated into a single chip.
The integrated circuit is not limited to an LSI, and may be realized by a dedicated circuit or a general-purpose processor. An FPGA (Field Programmable Gate Array) that can be programmed after LSI manufacture, or a reconfigurable processor that can reconfigure connection and setting of circuit cells within an LSI may be used.
In the above embodiments, each component may be configured by dedicated hardware, or may be realized by executing a software program suitable for each component. Each component may be realized by a program execution unit such as a CPU or a processor reading out and executing a software program recorded in a recording medium such as a hard disk or a semiconductor memory.
In other words, the transmission device and the reception device include a processing circuit (processing circuit) and a storage device (storage) electrically connected to the processing circuit (accessible from the control circuit). The processing circuit includes at least one of dedicated hardware and a program execution unit. In addition, in the case where the processing circuit includes a program execution unit, the storage device stores a software program executed by the program execution unit. The processing circuit uses the storage device to execute the transmission method or the reception method according to the above-described embodiments.
Further, the present disclosure may be the software program or a nonvolatile computer-readable recording medium on which the program is recorded. It is obvious that the program can be circulated via a transmission medium such as the internet.
In addition, the numbers used in the above are all numbers exemplified for specifically explaining the present disclosure, and the present disclosure is not limited to the exemplified numbers.
Note that, the division of the functional blocks in the block diagrams is an example, and a plurality of functional blocks may be implemented as 1 functional block, or 1 functional block may be divided into a plurality of functional blocks, or a part of the functions may be transferred to another functional block. Further, the functions of a plurality of functional blocks having similar functions may be processed in parallel or in time division by a single piece of hardware or software.
The order of executing the steps included in the transmission method or the reception method described above is an example order for specifically describing the present disclosure, and may be an order other than the above. Further, a part of the above steps may be executed simultaneously (in parallel) with other steps.
In the above, the transmission device, the reception device, the transmission method, and the reception method according to one or more embodiments of the present disclosure have been described based on the embodiments, but the present disclosure is not limited to the embodiments. The present invention is not limited to the embodiments described above, and various modifications and variations can be made without departing from the spirit and scope of the present invention.
Industrial applicability
The present disclosure can be applied to a device or apparatus that performs media transmission of video data, audio data, and the like.
Description of reference numerals:
15. 100, 300, 500, 700, 900, 1200, 1400 transmitting apparatus
16. 101, 301 encoding unit
17. 102 multiplexing part
18. 104 transmitting part
20. 200, 400, 600, 800, 1000, 1300, 1500, 1810, 1820, 1830, 1840, 1850, 1860, 2100, 2200 receiving apparatus
21 pack filter house
22 transmission order type discriminating part
23 random access unit
24. 212 control information acquisition unit
25 data acquisition part
26 calculating unit
27 initialization information acquisition unit
28. 206 decoding command part
29. 204A, 204B, 204C, 204D, 402 decoding unit
30 presentation part
201 tuner
202 demodulation section
203 inverse multiplexing part
205 display part
211 type discriminating section
213 slice information acquiring part
214 decoded data generating part
302 imparting part
303. 503, 702, 902, 1202, 1402 transmitter
401. 601, 801, 1001, 1301, 1501, 2201 receiving part
501 division part
502. 603 structural part
602 determination unit
701. 901, 1201, 1401 generation unit
802 first buffer
803 second buffer
804 a third buffer
805 fourth buffer
806 decoding unit
1002 reproduction unit
1302 calculating part
1303 correction unit
1502 determination unit
1811. 1821, 1831, 1841, 1851, 1861 first mode processing means a
1812. 1822, 1842, 1862 processing means B in the first mode
1823. 1853, 1864 second mode treatment means A
1824. 1833, 1844, 1854, 1865 second mode processing mechanism B
1832. 1843, 1852, 1863 mode switching mechanism
2101 first processing unit
2102 transfer unit
2202 inverse multiplexing unit
2203 first decoding unit
2204 second decoding unit
2205 and an output unit.

Claims (2)

1. A receiving apparatus includes:
a first processing unit that receives multiplexed data composed of first multiplexed data of a first multiplexing system and outputs the received multiplexed data; and
a conversion unit that converts the multiplexing scheme of the first multiplexed data among the multiplexed data to be output into a second multiplexing scheme and outputs converted data obtained by the conversion,
the conversion unit extracts first data that is a part of the first multiplexed data, performs first conversion of storing a first packet that constitutes the first data in a second packet that is used in the second multiplexing scheme, and outputs first converted data that is composed of the second packet obtained by the first conversion.
2. A method of receiving a signal having a first frequency,
receiving multiplexed data and outputting the received multiplexed data, wherein the multiplexed data is composed of first multiplexed data of a first multiplexing mode;
converting the multiplexing mode of the first multiplexing data among the multiplexed data to be output into a second multiplexing mode, outputting the converted data obtained by the conversion,
in the conversion, first data of a part of the first multiplexed data is extracted, first conversion is performed in which a first packet constituting the first data is stored in a second packet used in the second multiplexing method, and first converted data including the second packet obtained by the first conversion is output.
CN202110234610.8A 2015-10-07 2016-09-02 Receiving apparatus and receiving method Active CN112954435B (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201562238282P 2015-10-07 2015-10-07
US62/238,282 2015-10-07
JP2016109586A JP6986671B2 (en) 2015-10-07 2016-05-31 Receiver and receiving method
JP2016-109586 2016-05-31
CN201680059146.0A CN108141636B (en) 2015-10-07 2016-09-02 Receiving apparatus and receiving method
PCT/JP2016/004020 WO2017061070A1 (en) 2015-10-07 2016-09-02 Receiving device and receiving method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN201680059146.0A Division CN108141636B (en) 2015-10-07 2016-09-02 Receiving apparatus and receiving method

Publications (2)

Publication Number Publication Date
CN112954435A true CN112954435A (en) 2021-06-11
CN112954435B CN112954435B (en) 2023-02-03

Family

ID=58488315

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110234610.8A Active CN112954435B (en) 2015-10-07 2016-09-02 Receiving apparatus and receiving method

Country Status (3)

Country Link
JP (4) JP6986671B2 (en)
CN (1) CN112954435B (en)
WO (1) WO2017061070A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6605789B2 (en) 2013-06-18 2019-11-13 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Transmission method, reception method, transmission device, and reception device
JP6953238B2 (en) * 2017-08-31 2021-10-27 シャープ株式会社 Receiver, receiving method and program
JP6836971B2 (en) * 2017-08-31 2021-03-03 シャープ株式会社 Receiver, receiver method and program
JP6813933B2 (en) * 2018-07-19 2021-01-13 日本電信電話株式会社 Video / audio transmission system, transmission method, transmitter and receiver
CN114143585A (en) * 2021-11-02 2022-03-04 深圳Tcl新技术有限公司 Data processing method and device, storage medium and terminal equipment
JP2023116358A (en) * 2022-02-09 2023-08-22 株式会社東芝 Transmission system, transmission method, inserter, and data construction device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1349719A (en) * 1999-12-30 2002-05-15 皇家菲利浦电子有限公司 Method and apparatus for converting data streams
WO2003092208A1 (en) * 2002-04-26 2003-11-06 Nec Corporation Encoded packet transmission/reception method, device thereof, and program
CN1642221A (en) * 2004-01-15 2005-07-20 松下电器产业株式会社 Multiplex scheme conversion apparatus
CN101785302A (en) * 2007-08-24 2010-07-21 Lg电子株式会社 Digital broadcasting system and method of processing data in digital broadcasting system
CN101911686A (en) * 2007-12-27 2010-12-08 夏普株式会社 Information providing device, information display device, information providing system, control method, control program and storage medium
CN105247887A (en) * 2013-06-07 2016-01-13 索尼公司 Input device and transmission method; host device and reception method, and signal processing system and transmission/reception method

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005229587A (en) * 2004-01-15 2005-08-25 Matsushita Electric Ind Co Ltd Multiplex system conversion device
JP2007311917A (en) * 2006-05-16 2007-11-29 Matsushita Electric Ind Co Ltd Broadcast receiver, and measuring system, method and program for audience rating
JP6302274B2 (en) * 2014-02-07 2018-03-28 日本放送協会 Transmitting apparatus and receiving apparatus

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1349719A (en) * 1999-12-30 2002-05-15 皇家菲利浦电子有限公司 Method and apparatus for converting data streams
WO2003092208A1 (en) * 2002-04-26 2003-11-06 Nec Corporation Encoded packet transmission/reception method, device thereof, and program
CN1642221A (en) * 2004-01-15 2005-07-20 松下电器产业株式会社 Multiplex scheme conversion apparatus
CN101785302A (en) * 2007-08-24 2010-07-21 Lg电子株式会社 Digital broadcasting system and method of processing data in digital broadcasting system
CN101911686A (en) * 2007-12-27 2010-12-08 夏普株式会社 Information providing device, information display device, information providing system, control method, control program and storage medium
CN105247887A (en) * 2013-06-07 2016-01-13 索尼公司 Input device and transmission method; host device and reception method, and signal processing system and transmission/reception method

Also Published As

Publication number Publication date
JP7209234B2 (en) 2023-01-20
JP2017073761A (en) 2017-04-13
CN112954435B (en) 2023-02-03
JP2022016499A (en) 2022-01-21
JP2024028530A (en) 2024-03-04
JP6986671B2 (en) 2021-12-22
JP7429895B2 (en) 2024-02-09
JP2023029431A (en) 2023-03-03
WO2017061070A1 (en) 2017-04-13

Similar Documents

Publication Publication Date Title
CN108141636B (en) Receiving apparatus and receiving method
CN107431845B (en) Transmission method, reception method, transmission device, and reception device
CN107113462B (en) Transmission method, reception method, transmission device, and reception device
CN107925781B (en) Transmission method, reception method, transmission device, and reception device
CN106605409B (en) Transmission device, reception device, transmission method, and reception method
CN112954435B (en) Receiving apparatus and receiving method
JP7293433B2 (en) Transmission method, reception method, transmission device and reception device
JP7307237B2 (en) Receiving method, transmitting method, receiving device and transmitting device
WO2016157817A1 (en) Transmission method, reception method, transmission device and reception device
JP7036957B2 (en) Transmission method, reception method, transmission device and reception device
CN111614976B (en) Transmission device, reception device, transmission method, and reception method
JP7116137B2 (en) Transmission method, reception method, transmission device and reception device
CN113038188B (en) Transmitting method, receiving method, transmitting device and receiving device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant