WO2017022461A1 - Dispositif de réception, dispositif d'envoi, et procédé de traitement de données - Google Patents

Dispositif de réception, dispositif d'envoi, et procédé de traitement de données Download PDF

Info

Publication number
WO2017022461A1
WO2017022461A1 PCT/JP2016/071059 JP2016071059W WO2017022461A1 WO 2017022461 A1 WO2017022461 A1 WO 2017022461A1 JP 2016071059 W JP2016071059 W JP 2016071059W WO 2017022461 A1 WO2017022461 A1 WO 2017022461A1
Authority
WO
WIPO (PCT)
Prior art keywords
dialog
control information
channel
dialog control
transmitted
Prior art date
Application number
PCT/JP2016/071059
Other languages
English (en)
Japanese (ja)
Inventor
高橋 和幸
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Publication of WO2017022461A1 publication Critical patent/WO2017022461A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/28Arrangements for simultaneous broadcast of plural pieces of information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H60/00Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
    • H04H60/09Arrangements for device control with a direct linkage to broadcast information or to broadcast space-time; Arrangements for control of broadcast-related services
    • H04H60/13Arrangements for device control affected by the broadcast information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • H04S5/02Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation  of the pseudo four-channel type, e.g. in which rear channel signals are derived from two-channel stereo signals

Definitions

  • the present technology relates to a receiving device, a transmitting device, and a data processing method, and more particularly, to a receiving device, a transmitting device, and a data processing method that can perform dialog control using dialog control information.
  • multi-channel audio such as 22.2ch is being introduced as an acoustic system.
  • various descriptors related to audio streams are defined (for example, see Non-Patent Document 1).
  • dialog control By the way, in multi-channel audio, a technique called dialog control is becoming widespread. However, a technical method of dialog control has not been established, and a proposal for performing dialog control using dialog control information has been requested.
  • This technology has been made in view of such a situation, and makes it possible to perform dialog control using dialog control information.
  • a receiving device processes a receiving unit that receives a stream including a multi-channel or stereo audio component that is transmitted via a transmission path, and a signaling that is defined by a predetermined standard.
  • An acquisition unit that acquires dialog control information for controlling a dialog of multi-channel or stereo audio transmitted in the system layer, and a multi-channel transmitted through the transmission path based on the dialog control information
  • a control unit that controls a stereo audio dialog.
  • the receiving device may be an independent device, or may be an internal block constituting one device.
  • the data processing method according to the first aspect of the present technology is a data processing method corresponding to the above-described receiving device according to the first aspect of the present technology.
  • a stream including a multi-channel or stereo audio component transmitted via a transmission path is received and defined by a predetermined standard.
  • Dialog control information for controlling a multi-channel or stereo audio dialog transmitted in a system layer for processing signaling is acquired, and the multi-channel is transmitted via the transmission path based on the dialog control information Alternatively, a stereo audio dialog is controlled.
  • a transmission apparatus includes an acquisition unit that acquires a stream including a multi-channel or stereo audio component, and generation that generates dialog control information for controlling a multi-channel or stereo audio dialog. And a transmission unit that transmits the dialog control information together with the stream via a transmission path, and the dialog control information is transmitted in a system layer that processes signaling defined by a predetermined standard Device.
  • the transmission device according to the second aspect of the present technology may be an independent device, or may be an internal block constituting one device.
  • a data processing method according to the second aspect of the present technology is a data processing method corresponding to the transmission device according to the second aspect of the present technology described above.
  • a dialog control information for acquiring a stream including a multi-channel or stereo audio component and controlling a multi-channel or stereo audio dialog. And the dialog control information is transmitted along with the stream via a transmission path.
  • the dialog control information is transmitted in a system layer that processes signaling defined by a predetermined standard.
  • dialog control using dialog control information can be performed.
  • FIG. 1 is a diagram illustrating a configuration of an embodiment of a transmission system to which the present technology is applied.
  • the system refers to a logical collection of a plurality of devices.
  • the transmission system 1 includes a transmission device 10 and a reception device 20.
  • the transmission system 1 for example, data transmission conforming to a predetermined standard such as a digital broadcasting standard is performed.
  • the transmitting device 10 transmits (transmits) the content via the transmission path 30.
  • the transmission apparatus 10 transmits a stream of video and audio (components) constituting a content such as a television program and signaling and a signaling as a digital broadcast signal via the transmission path 30.
  • the receiving device 20 receives and outputs content transmitted (transmitted) from the transmitting device 10 via the transmission path 30.
  • the receiving device 20 receives a digital broadcast signal transmitted from the transmitting device 10, acquires a stream of video and audio (components thereof) and signaling that constitute the content, and obtains a video of content such as a TV program, Play audio.
  • data transmission conforming to digital broadcasting standards such as ISDB (Integrated Services Digital Broadcasting), DVB (Digital Video Broadcasting), ATSC (Advanced Television Systems Committee), etc.
  • Digital broadcasting standards such as ISDB (Integrated Services Digital Broadcasting), DVB (Digital Video Broadcasting), ATSC (Advanced Television Systems Committee), etc.
  • Data transmission conforming to the standard is performed.
  • the transmission line 30 when data transmission conforming to the digital broadcasting standard is performed, a satellite line, a cable television network (wired line), or the like can be used in addition to the ground wave.
  • a communication line such as the Internet or a telephone network can be used.
  • Dialog control refers to controlling a multi-channel audio dialog such as dialog volume level (voice level) control, dialog replacement control, or dialog localization position control.
  • dialog control information information related to dialog control is referred to as dialog control information.
  • 22.2ch multi-channel audio will be described as an example of multi-channel audio.
  • multi-channel audio refers to an acoustic system that transmits audio data of two or more channels such as 22.2ch in order to reproduce the localization of a sound image and the sense of spread of a sound field.
  • multi-channel audio will be described, but the present technology can also be applied to stereo audio.
  • FIG. 2 is a diagram showing an example of speaker arrangement in an acoustic system employing 22.2ch multi-channel audio.
  • 3 and 4 show the layer configuration of the speaker arrangement of FIG. 2 and the channel map, which will be described with reference to them as appropriate.
  • the circles indicate the position of each speaker.
  • a two-channel subwoofer is arranged for low frequency enhancement (LFE: Low Frequency Frequency).
  • LFE Low Frequency Frequency
  • the front left channel speaker FL, the front right channel speaker FR, the front center channel speaker FC, and the rear left channel are displayed in the middle layer with respect to the content video display area (TV screen).
  • the upper layer has a speaker TpFL for the upper front left channel, a speaker TpFR for the upper front right channel, a speaker TpFC for the upper front center channel, and an upper center for the content video display area (TV screen).
  • a channel speaker TpC, an upper rear left channel speaker TpBL, an upper rear right channel speaker TpBR, an upper side left channel speaker TpSiL, an upper side right channel speaker TpSiR, and an upper rear center channel speaker TpBC are arranged. Is done. That is, in the acoustic system of FIG. 2, a total of nine channels of speakers are arranged in the upper layer.
  • a lower front center channel speaker BtFC, a lower front left channel speaker BtFL, and a lower front right channel speaker BtFR are displayed in the lower layer with respect to a content video display area (TV screen). Be placed. Further, in the lower layer, a subwoofer LFE1 and a subwoofer LFE2 for low frequency enhancement (LFE) are arranged in front of the lower layer. That is, in the acoustic system of FIG. 2, a 3-channel speaker and a 2-channel subwoofer are arranged in the lower layer.
  • the receiving device 20 when the receiving device 20 has a configuration capable of realizing the 22.2 ch multi-channel audio shown in FIG. 2, the transmission device 10 via the transmission path 30.
  • the receiving apparatus 20 controls the 22.2ch multi-channel audio dialog based on the dialog control information.
  • the receiving apparatus 20 when receiving the dialog control information including an instruction for adjusting the volume of the dialog from the transmission apparatus 10, the receiving apparatus 20 adjusts the 22.2ch volume level based on the dialog control information. Also, when receiving the dialog control information including the dialog replacement instruction from the transmitting device 10, the receiving device 20 receives, for example, the front center channel speaker FC and the lower front center channel speaker BtFC based on the dialog control information. Replace the input Japanese dialog with the English or French dialog.
  • the receiving device 20 when the receiving device 20 has a configuration capable of realizing the 22.2 ch multi-channel audio shown in FIG.
  • the receiving device 20 performs a multi-channel audio dialog such as dialog volume level control, dialog replacement control, or dialog localization position control based on the dialog control information. Control.
  • MPEG-H 3D Audio Dialog control can also be used in MPEG-H 3D Audio, which defines audio compression for 3D audio, which allows multiple speakers.
  • MPEG-H 3D Audio VBAP (Vector Base Amplitude Pannning) is used as a technique for controlling the localization of a sound image by arranging a plurality of speakers. The details of VBAP are described in “Ville Pulkki,“ Virtual Sound Source Positioning Using Vector Base Amplitude Panning ”, Journal of AES, vol.45, no.6, pp.456-466, 1997”. Yes.
  • FIG. 5 is a diagram showing an example of speaker arrangement of a three-dimensional VBAP.
  • each channel is output from these speakers SP1 to SP5.
  • the speakers SP1 to SP5 are arranged on a spherical surface centering on the origin O at the position of the head of the user U11. Further, the three-dimensional vectors starting from the origin O and pointing in the direction of the positions of the speakers SP1 to SP5 are set as vectors I 1 to I 5 .
  • a triangular region surrounded by the speaker SP1, the speaker SP4, and the speaker SP5 is defined as a region TR21.
  • a triangular region surrounded by the speaker SP3, the speaker SP4, and the speaker SP5 among the region on the spherical surface centered on the origin O is defined as a region TR22, and is surrounded by the speaker SP2, the speaker SP3, and the speaker SP5.
  • a triangular region is referred to as a region TR23.
  • these regions TR21 to TR23 are one mesh (three-dimensional mesh).
  • a three-dimensional vector indicating a position where the sound image is to be localized is a vector P, in the example of FIG. 5, the vector P indicates a position on the region TR21.
  • the three-dimensional vectors facing the positions of the speaker SP1, the speaker SP4, and the speaker SP5 are the vector I 1 , the vector I 4 , and the vector I 5. As shown in (1), it can be expressed by a linear sum of a vector I 1 , a vector I 4 , and a vector I 5 .
  • the gain of the sound output from the speaker SP2 and the speaker SP3 that does not constitute the region TR21 is zero. That is, no sound is output from the speaker SP2 and the speaker SP3.
  • a sound image can be placed at an arbitrary position (sound image position) on a region composed of regions TR21 to TR23 by three-dimensional VBAP. It is possible to localize.
  • the sound image position VSP is represented by a polar coordinate system based on Azimuth ( ⁇ ), Elevation ( ⁇ ), and Radius (r).
  • the “front” represented by the arrow in the x-axis direction in the figure indicates, for example, the direction of the content video display area (for example, “TV screen” in FIG. 2) with respect to the head of the user U11. pointing.
  • FIG. 7 is a diagram illustrating a configuration example of the audio decoder 50 corresponding to MPEG-H 3D Audio.
  • an audio decoder 50 corresponding to MPEG-H 3D Audio includes a USAC-3D decoder 51, a format converter 52, a metadata decoder 53, an object renderer 54, a SAOC-3D decoder 55, a HOA renderer 56, a mixer 57, and , A binaural renderer 58.
  • the MPEG-H bit stream is input to the USAC-3D decoder 51.
  • the USAC-3D decoder 51 decodes the MPEG-H bitstream in accordance with USAC (Unified Speech and Audio Audio Coding).
  • the USAC-3D decoder 51 obtains channels (Channels), metadata (Compressed object metadata), objects (Object), SAOC transport channels (SAOC Transport Channels), and HOA coefficients (HOA (Higher order ambisonics) obtained by decoding. ) (Coefficient) is supplied to the format converter 52, the metadata decoder 53, the object renderer 54, the SAOC-3D decoder 55, and the HOA renderer 56, respectively.
  • the format converter 52 performs processing such as format conversion on the channel information of each channel supplied from the USAC-3D decoder 51, and supplies the processing result to the mixer 57.
  • the metadata decoder 53 decodes the compressed metadata supplied from the USAC-3D decoder 51 and supplies the metadata obtained thereby to the object renderer 54 and the SAOC-3D decoder 55, respectively.
  • the object renderer 54 includes object information (for example, polar coordinate information of the object sound source) regarding each object sound source from the USAC-3D decoder 51 and metadata (for example, meta information including position information of each speaker SP) from the metadata decoder 53. Data).
  • object information for example, polar coordinate information of the object sound source
  • metadata for example, meta information including position information of each speaker SP
  • the object renderer 54 is based on the object information, metadata, and the like, and has a triangular area TR (for example, FIG. 5) surrounded by three speakers SP (for example, the speakers SP1, SP4, and SP5 in FIG. 5) in the vicinity of the target object sound source. 5 region TR21), the processing is performed so that the sound image is localized at the position of the target object sound source (for example, the sound image position VSP corresponding to the vector P in FIG. 5).
  • the object renderer 54 supplies the processing result to the mixer 57.
  • the SAOC-3D decoder 55 is based on SAOC transport channel information supplied from the USAC-3D decoder 51 and metadata (metadata including position information of the speaker SP) supplied from the metadata decoder 53, etc. Processing related to the object sound source is performed, and the processing result is supplied to the mixer 57.
  • the HOA renderer 56 performs processing related to the microphone arranged on the spherical surface based on the HOA coefficient supplied from the USAC-3D decoder 51, and supplies the processing result to the mixer 57.
  • the mixer 57 mixes the processing results from the format converter 52, the object renderer 54, the SAOC-3D decoder 55, and the HOA renderer 56, and outputs the result to each speaker SP (for example, the speakers SP1 to SP5 in FIG. 5). .
  • the output from the mixer 57 is processed by the binaural renderer 58 before being output.
  • the audio decoder 50 compatible with MPEG-H 3D Audio is configured as described above.
  • dialog control is performed from the transmitting device 10 via the transmission path 30.
  • the receiving device 20 controls the dialog of the sound image that is localized at a position on the region composed of the regions TR21 to TR23, for example, based on the dialog control information.
  • the receiving device 20 when receiving the dialog control information including the instruction for adjusting the volume of the dialog from the transmitting device 10, the receiving device 20 has a position on the region composed of the regions TR21 to TR23 based on the dialog control information. Adjust the volume level for the sound image localized in In addition, when receiving the dialog control information including the dialog replacement instruction from the transmission device 10, the receiving device 20 receives, for example, the Japanese text input to the object sound source Ob1 and the object sound source Ob2 based on the dialog control information. Replace dialogs with English or French dialogs.
  • the reception device 20 when the reception device 20 has a configuration capable of realizing the three-dimensional VBAP shown in FIG. 5, the dialog from the transmission device 10 via the transmission path 30 is performed.
  • the receiving device 20 controls a dialog of multi-channel audio such as dialog volume level control, dialog replacement control, or dialog localization position control based on the dialog control information.
  • the receiving device 20 controls the dialog of the multi-channel audio based on the dialog control information from the transmitting device 10.
  • ARIB Association of Radio Industries, and Businesses: General Incorporated Association Meeting
  • dialog control information is transmitted in an audio stream.
  • FIG. 8 is a diagram illustrating the syntax of an audio stream for transmitting dialog control information. This syntax is the syntax of the bitstream added to the end of MPEG4_ancillary_data ().
  • ext_downmixing_level_status2 is set to “0” in the case of an audio mode in which a value other than “13” is set as channelConfiguration, which is a parameter set in MPEG-4 Audio, and downmix coefficients (dmix_c_idx, dmix_d_idx, dmix_e_idx, dmix_f_idx, dmix_g_idx, dmix_l_idx) will not be transmitted.
  • Ext_dialogue_status is a flag indicating whether or not dialog control information exists.
  • dialog control information exists.
  • Num_dialogue_chans is set to the number of channels dedicated to the dialog.
  • the number of bits of num_dialogue_chans is determined according to the number of channels of main audio.
  • the upper limit value of the allowable value of the gain control amount on the receiving device 20 side is set.
  • a lower limit value of an allowable value of the gain control amount on the receiving device 20 side is set.
  • the language code of the main dialog is set in 24-bit dialog_main_lang_code.
  • This language code value conforms to, for example, ISO 639-2, and a value defined in ISO / IEC 8859-1 can be used for characters.
  • 8-bit dialog_main_lang_comment_bytes the number of bytes of character string information for indicating the contents of the main dialog is set.
  • 8-bit dialog_main_lang_comment_data is arranged in a loop corresponding to the number of bytes indicated by the dialog_main_lang_comment_bytes.
  • the dialogue_main_lang_comment_data byte data of character string information for indicating the contents of the main dialog is set.
  • dialog_src_index and 4-bit dialog_gain_index are arranged in a loop according to the number of channels dedicated to the dialog indicated by num_dialogue_chans.
  • the index of the channel dedicated to the dialog is set.
  • the number of bits of dialog_src_index_ is determined according to the number of channels of main audio.
  • the gain correction value index of the additional dialog is set.
  • the number of additional dialogs is set in 4-bit num_additional_lang_chans.
  • a 24-bit dialog_additional_lang_code and an 8-bit dialog_additional_lang_comment_bytes are arranged in a loop corresponding to the number of additional dialogs indicated by the num_additional_lang_chans.
  • dialogue_additional_lang_code the language code of the additional dialog is set. This language code value conforms to, for example, ISO 639-2, and a value defined in ISO / IEC 8859-1 can be used for characters.
  • Dialog_additional_lang_comment_bytes is set with the number of bytes of the character string information for indicating the contents of the add dialog.
  • 8-bit dialog_additional_lang_comment_data is arranged in a loop corresponding to the dialog_additional_lang_comment_bytes.
  • byte data of character string information for indicating the contents of the add dialog is set.
  • Byte_alignment () is a function for adjusting the data length to a byte unit (multiple of 8 bits), and its starting point is taken from ext_dialogue_status.
  • dialog control information is transmitted with an audio stream as shown in FIG. 8, there is a possibility that the reception device 20 may have difficulty handling the dialog control information.
  • the receiving apparatus 20 can adjust the volume within the range of 0 to 100, and the dialog control information instructs the volume adjustment of the dialog within the range of 30 to 70, 30 to 70 Therefore, even if the user wants to reduce the volume to 0, the volume can only be reduced to 30.
  • the volume can be adjusted in the range of 0 to 100, but presenting a user interface indicating that the volume is currently limited to the range of 30 to 70, the volume is shown to the user. It is a general flow of processing to notify the reason why cannot be lowered.
  • dialog control information is transmitted in an audio stream, so in the system layer that is a layer that processes signaling (descriptor) and the like, dialog control information is not transmitted. It is necessary to pass dialog control information acquired from the audio stream to the system layer using API (Application Programming Interface) or the like. As a result, the system layer can present a user interface indicating that volume adjustment is limited to a range of 30 to 70, for example, based on the dialog control information obtained from the audio stream, but the API is used. It is assumed that processing will take time, such as the need to perform processing.
  • API Application Programming Interface
  • the dialog control information is transmitted in the system layer (descriptor) so that the reception apparatus 20 can easily handle the dialog control information.
  • ARIB ⁇ ⁇ STD-B60 1.3 currently being developed, information indicating whether the audio stream includes dialog control information is placed in the audio component descriptor (MH-Audio Component Descriptor) transmitted in the system layer. It is planned.
  • FIG. 9 is a diagram illustrating an example of the syntax of an audio component descriptor (MH-Audio Component Descriptor) defined in the ARIB STD-B60 version 1.3.
  • the 16-bit descriptor_tag contains a tag value that identifies each descriptor.
  • Descriptor length is set in 8-bit descriptor_length.
  • the descriptor_length is followed by a 4-bit reserved area (reserved_future_use).
  • the stream type is set in 4-bit stream_content.
  • 8-bit component_type information related to encoding for an audio component is set.
  • ⁇ Information for identifying the component stream is set in the 16-bit component_tag.
  • the audio stream format is set in the 8-bit stream_type.
  • simulcast_group_tag In 8-bit simulcast_group_tag, the same number is set for the component performing simulcast. In the 1-bit ES_multi_lingual_flag, “1” is set when bilingual multiplexing is performed in the elementary stream.
  • -1 is set in 1-bit main_component_flag when the target audio component is the main audio.
  • a predefined sound quality mode is set in the 2-bit quality_indicator.
  • ⁇ Sampling frequency is specified in 3-bit sampling_rate.
  • the sampling_rate is followed by a 1-bit reserved area (reserved_future_use).
  • the language code of the audio component is set in 24-bit ISO_639_language_code. Further, when “1” is designated as ES_multi_lingual_flag, 24-bit ISO_639_language_code_2 is arranged. ISO_639_language_code_2 is set with the language code of the second audio component in the case of the multilingual mode.
  • the audio component descriptor is configured as described above.
  • the most significant bit (b7) can indicate the presence / absence of dialog control information.
  • component_type when “0” is set in the most significant bit (b7) of component_type, it indicates that the audio stream does not include dialog control information. When 1 "is set, it indicates that the audio stream includes dialog control information.
  • the audio component descriptor (MH-Audio Component Descriptor) defined in ARIB STD-B60 1.3 contains only information indicating whether or not the audio stream includes dialog control information. Even if it is known whether or not the stream is transmitting dialog information, eventually it is necessary to pass the dialog control information to the system layer using an API or the like, so that processing takes time.
  • dialog control information may be difficult to handle in the receiving device 20, so that a proposal for making the dialog control information easy to handle and performing dialog control using the dialog control information is requested. It was.
  • a descriptor including dialog control information for responding to such a request (hereinafter also referred to as this technology descriptor) is defined and transmitted by the system layer.
  • a method for enabling the apparatus 20 to perform dialog control using dialog control information transmitted in the system layer is proposed.
  • FIG. 11 is a diagram illustrating an example of syntax and semantics of the technology descriptor.
  • ext_dialogue_status is a flag indicating that dialog control is performed.
  • dialog control is performed when “1” is set as ext_dialogue_status.
  • ext_dialogue_status is a 7-bit reserved area (reserved).
  • dialog control information information related to dialog control
  • num_of_dialog_chans The number of audio components for which dialog control is performed is set in 5-bit num_of_dialog_chans. Next to num_of_dialog_chans is a 3-bit reserved area (reserved).
  • the upper limit of the audio level at which dialog control is performed is set in the 8-bit dialog_plus_index. Also, the lower limit of the audio level at which dialog control is performed is set in the 8-bit dialog_minus_index.
  • the country code of the main dialog is set in 24-bit dialog_lang_code.
  • ISO International Organization for Standardization
  • the 8-bit dialog_main_lang_comment_bytes field contains the number of character information bytes in the main dialog contents.
  • 8-bit dialog_main_lang_comment_data is arranged in a loop corresponding to the number of bytes indicated by the dialog_main_lang_comment_bytes.
  • the content of the main dialog is set in dialogue_main_lang_comment_data.
  • a 5-bit dialog_src_index, a 3-bit reserved area (reserved), and a 4-bit dialog_gain_index are arranged in a loop corresponding to the number of audio components (dialog control is performed) indicated by num_of_dialog_chans.
  • the index of the dialog dedicated channel is set.
  • the gain correction index of the additional dialog is set.
  • num_additional_lang_chans The number of additional dialogs is set in 4-bit num_additional_lang_chans. Next to num_additional_lang_chans is a 4-bit reserved area (reserved).
  • dialog_additional_lang_code the country code of the additional dialog is set.
  • ISO International Organization for Standardization
  • dialog_additional_lang_comment_bytes the number of character information bytes of the contents of the add dialog is set.
  • 8-bit dialog_additional_lang_comment_data is arranged in a loop corresponding to the number of bytes indicated by the dialog_additional_lang_comment_bytes.
  • dialogue_additional_lang_comment_data the content of the add dialog is set.
  • This technical descriptor is configured as described above.
  • the configuration of the syntax and semantics of this technical descriptor shown in FIG. 11 is an example, and other configurations may be adopted.
  • the reception device 20 can transmit the present technology description in the system layer. Based on the child (dialog control information included therein), dialog control of multi-channel audio such as dialog volume level control, dialog replacement control, or dialog localization position control can be performed.
  • the technology descriptor (FIG. 11) is newly defined and the transmission method in the system layer has been described. However, the contents (dialog control information) described in the technology descriptor (FIG. 11) are described. ) May be adopted as long as it can be transmitted in the system layer.
  • dialog control is performed in the system layer. Audio component descriptors containing information can be transmitted. As a result, the receiving device 20 can perform dialog control such as dialog volume level control based on the technical descriptor (contained in the dialog control information) transmitted in the system layer.
  • the content of the present technology descriptor (FIG. 11) can be described after the text_char loop arranged last.
  • what is arranged after the loop of text_char is an example, and the contents of the technical descriptor (FIG. 11) can be arranged at an arbitrary position in the audio component descriptor (FIG. 9).
  • the content of this technical descriptor (FIG. 11) is not limited to the audio component descriptor (FIG. 9), but may be described in another descriptor transmitted in the system layer.
  • the description has been made on the assumption that the descriptor is standardized by ARIB, but it may be applied to other digital broadcasting standards such as DVB (Digital Video Broadcasting). That is, also in other digital broadcasting standards such as DVB, the receiving apparatus 20 controls dialog control using dialog control information by transmitting the present technology descriptor (FIG. 11) including dialog control information in the system layer. It can be performed.
  • DVB Digital Video Broadcasting
  • DE Dialogue Enhancement
  • a component descriptor (Component Descriptor) is defined, and the contents (dialog control information) of this technical descriptor in FIG. 11 are described in this component descriptor.
  • the component descriptor including the dialog control information can be transmitted in the system layer.
  • the receiving device 20 can perform dialog control such as dialog volume level control based on the technical descriptor (contained in the dialog control information) transmitted in the system layer.
  • the content of the present technical descriptor (FIG. 11) can be described after the text_char loop arranged last.
  • the arrangement following the text_char loop is an example, and the content of the present technology descriptor (FIG. 11) can be arranged at an arbitrary position in the component descriptor (FIG. 13).
  • the content of this technical descriptor (FIG. 11) is not limited to the component descriptor (FIG. 13), and may be described in another descriptor transmitted in the system layer.
  • dialog control information is standardized as being transmitted in an audio stream.
  • other types such as MPEG-H and AC-4 (Audio Code number 4) have been described.
  • dialog control information is transmitted at the bitstream level, but dialog control information is not transmitted in the system layer.
  • the technical descriptor (FIG. 11) is transmitted in the system layer, or the existing descriptor is extended.
  • dialog control information is transmitted in the system layer (descriptor).
  • the receiving device 20 can perform dialog control such as dialog volume level control based on the descriptor (dialog control information included in the descriptor) transmitted in the system layer.
  • the configuration of the transmission system 1 in FIG. 1 is as follows, for example. That is, in the transmission system 1 of FIG. 1, the transmission path 30 is a communication line such as the Internet or a telephone network, and the receiving device 20 transmits content to the transmitting device 10 installed as a server via the communication line. By requesting the distribution, the stream of the content distributed by streaming from the transmission device 10 is received and reproduced.
  • the transmission path 30 is a communication line such as the Internet or a telephone network
  • the dialog control information may be transmitted as a data broadcast application such as HTML5 (HyperText Markup Language 5), for example, in addition to transmission by descriptor.
  • the reception device 20 can perform dialog control by receiving and executing an application transmitted from the transmission device 10 via the transmission path 30.
  • the application is not limited to data broadcasting, but may be distributed from a server via communication.
  • FIG. 14 is a diagram illustrating a configuration example of the transmission device 10 of FIG.
  • the transmission apparatus 10 includes a control unit 101, a component acquisition unit 102, an encoder 103, a signaling generation unit 104, a signaling processing unit 105, a packet generation unit 106, a physical layer frame generation unit 107, and a transmission unit 108. Is done.
  • the control unit 101 controls the operation of each unit of the transmission device 10.
  • the component acquisition unit 102 acquires data such as video, audio, and subtitles (components) constituting content (for example, a television program) provided by a specific service, and supplies the acquired data to the encoder 103.
  • the encoder 103 encodes data (components) such as video and audio supplied from the component acquisition unit 102 according to a predetermined encoding method, and supplies the encoded data to the packet generation unit 106.
  • the corresponding content is acquired from the storage location of the already recorded content according to the broadcast time zone, or the live content is acquired from the studio or location location.
  • the content can be configured to include multi-channel audio components.
  • the signaling generation unit 104 acquires raw data for generating signaling from an external server, a built-in storage, or the like.
  • the signaling generation unit 104 generates signaling using the raw data of signaling and supplies it to the signaling processing unit 105.
  • the signaling for example, the present technology descriptor (FIG. 11) including the dialog control information or the existing descriptor (FIG. 9) is generated.
  • the signaling processing unit 105 processes the signaling supplied from the signaling generation unit 104 and supplies it to the packet generation unit 106.
  • the packet generator 106 processes video and audio (components) data supplied from the encoder 103 and signaling data supplied from the signaling processor 105, and converts the packet in which those data are stored. It is generated and supplied to the physical layer frame generation unit 107.
  • the physical layer frame generation unit 107 generates a physical layer frame by encapsulating a plurality of packets supplied from the packet generation unit 106 and supplies the physical layer frame to the transmission unit 108.
  • the transmission unit 108 performs, for example, OFDM (Orthogonal Frequency Division Multiplexing) modulation on the physical layer frame supplied from the physical layer frame generation unit 107, and transmits it as a digital broadcast signal via the antenna 111.
  • OFDM Orthogonal Frequency Division Multiplexing
  • the present technology descriptor (FIG. 11) or the existing descriptor (FIG. 9) including the dialog control information is transmitted in the system layer by the digital broadcast signal.
  • the transmission apparatus 10 of FIG. 14 it is not necessary that all the functional blocks are physically disposed in a single apparatus, and at least some of the functional blocks are physically independent from other functional blocks. It may be configured as a device.
  • FIG. 15 is a diagram illustrating a configuration example of the receiving device 20 of FIG.
  • the reception device 20 includes a control unit 201, a reception unit 202, a physical layer frame processing unit 203, a packet processing unit 204, a signaling processing unit 205, a decoder 206, a video output unit 207, and an audio output unit 208. Is done.
  • the video output unit 207 is connected to a display device 221, and the audio output unit 208 is connected to speakers 222-1 to 222-N (N is an integer of 1 or more).
  • the speakers 222-1 to 222-N are arranged corresponding to the speaker arrangement in the 22.2ch multi-channel audio sound system as shown in FIG. Although omitted in FIG. 15, a subwoofer may be arranged in addition to the speaker 222.
  • the control unit 201 controls the operation of each unit of the receiving device 20.
  • the reception unit 202 receives the digital broadcast signal transmitted from the transmission device 10 via the antenna 211, performs processing such as OFDM demodulation, and converts the physical layer frame obtained thereby into the physical layer frame processing unit. 203.
  • the physical layer frame processing unit 203 performs processing on the physical layer frame supplied from the receiving unit 202, extracts a packet, and supplies the packet to the packet processing unit 204.
  • the packet processing unit 204 processes the packet supplied from the physical layer frame processing unit 203 and acquires component and signaling data. Of the data acquired by the packet processing unit 204, signaling data is supplied to the signaling processing unit 205, and component data is supplied to the decoder 206.
  • the signaling processing unit 205 appropriately processes the signaling data supplied from the packet processing unit 204 and supplies it to the control unit 201.
  • the control unit 201 controls the operation of each unit based on the signaling supplied from the signaling processing unit 205. Specifically, the control unit 201 controls packet filtering performed by the packet processing unit 204 based on the analysis result of the signaling, so that data such as video and audio (components thereof) is supplied to the decoder 206. To.
  • the decoder 206 decodes (components) data such as video and audio supplied from the packet processing unit 204 according to a predetermined decoding method, and supplies the resulting video data to the video output unit 207.
  • the audio data is supplied to the audio output unit 208.
  • the video output unit 207 causes the display device 221 to display video corresponding to the video data supplied from the decoder 206 in accordance with control from the control unit 201.
  • the audio output unit 208 causes the speakers 222-1 to 222-N to output audio corresponding to the audio data supplied from the decoder 206 in accordance with control from the control unit 201.
  • control unit 201 controls the audio output unit 208 based on the dialog control information included in the technical descriptor (FIG. 11) or the existing descriptor (FIG. 9) transmitted in the system layer. It controls a dialog of multi-channel audio (for example, 22.2ch multi-channel audio shown in FIG. 2) realized by the speakers 222-1 to 222-N.
  • multi-channel audio for example, 22.2ch multi-channel audio shown in FIG. 2
  • the receiving device 20 includes, for example, a fixed receiver such as a television receiver, a recorder, and a set top box (STB), a mobile receiver such as a smartphone and a tablet terminal, and an in-vehicle television. It can be set as the apparatus mounted in motor vehicles, such as.
  • a fixed receiver such as a television receiver, a recorder, and a set top box (STB)
  • STB set top box
  • mobile receiver such as a smartphone and a tablet terminal
  • in-vehicle television in-vehicle television. It can be set as the apparatus mounted in motor vehicles, such as.
  • step S101 component signaling acquisition processing is performed.
  • components such as video and audio (multi-channel audio) are acquired by the component acquisition unit 102, and data of components such as video and audio are encoded by the encoder 103.
  • signaling is generated by the signaling generation unit 104 and the signaling is processed by the signaling processing unit 105.
  • this technical descriptor may be newly defined as a descriptor for describing the dialog control information, or an existing descriptor such as an audio component descriptor (FIG. 9) is extended. Then, dialog control information may be described in the extended area.
  • step S102 packet / frame generation processing is performed.
  • a packet is generated by the packet generation unit 106, and a physical layer frame is generated by the physical layer frame generation unit 107.
  • step S103 digital broadcast signal transmission processing is performed.
  • the transmission unit 108 performs a process on the physical layer frame and transmits it as a digital broadcast signal via the antenna 111.
  • the present technology descriptor (FIG. 11) or the audio component descriptor (FIG. 9) including the dialog control information is transmitted in the system layer by the digital broadcast signal.
  • step S201 digital broadcast signal reception processing is performed.
  • the receiver 202 receives the digital broadcast signal via the antenna 211.
  • step S202 packet / frame processing is performed.
  • the physical layer frame processing unit 203 extracts a packet from the physical layer frame, and the packet processing unit 204 processes the packet.
  • step S203 signaling component processing is performed.
  • the control unit 201 controls the operation of each unit based on the signaling, and the decoder 206 decodes component data such as video and audio.
  • the video output unit 207 displays the content video on the display device 221 in accordance with the control from the control unit 201.
  • the audio output unit 208 outputs the audio of the content from the speakers 222-1 to 222-N according to the control from the control unit 201.
  • control unit 201 controls the audio output unit 208 based on the dialog control information included in the technical descriptor (FIG. 11) or the audio component descriptor (FIG. 9) transmitted in the system layer.
  • Control of multi-channel audio dialog realized by speakers 222-1 to 222-N such as dialog volume level control, dialog replacement control, or dialog localization position control.
  • control unit 201 adjusts the volume level of the 22.2ch multi-channel audio realized by the speakers 222-1 to 222-N according to the dialog volume adjustment instruction included in the dialog control information. Do. By adjusting the volume level, for example, for a hearing impaired person, it is possible to control to increase only the level of the voice of narration.
  • dialog_plus_index Within the range of 0 dB ⁇ x ⁇ +12 dB shown, each level of 20.2 ch other than the speaker FC and speaker BtFC is lowered by xdB.
  • the speaker FC and the speaker are within the range of - ⁇ dB ⁇ -x ⁇ 0dB indicated by dialog_minus_index.
  • Each level of BtFC is lowered by xdB.
  • control unit 101 in accordance with the dialog replacement instruction included in the dialog control information, in the 22.2ch multi-channel audio realized by the speakers 222-1 to 222-N, for example, the speaker FC (for the front center channel) Replace the Japanese dialog input to the speaker BtFC (in the lower front center channel) with the English or French dialog.
  • dialog_gain_index [0] ( ⁇ 3 dB) indicating the assignment level to the speaker FC
  • dialog_gain_index [1] (0 dB) indicating the assignment level to the speaker BtFC
  • dialog control can also be performed in the same manner when 3D VBAP or the like is used.
  • dialog control information in the system layer Dialog control can be performed in the same manner by transmitting a descriptor including.
  • FIG. 18 is a diagram illustrating a configuration example of hardware of a computer that executes the above-described series of processing by a program.
  • a CPU Central Processing Unit
  • ROM Read Only Memory
  • RAM Random Access Memory
  • An input / output interface 905 is further connected to the bus 904.
  • An input unit 906, an output unit 907, a recording unit 908, a communication unit 909, and a drive 910 are connected to the input / output interface 905.
  • the input unit 906 includes a keyboard, a mouse, a microphone, and the like.
  • the output unit 907 includes a display, a speaker, and the like.
  • the recording unit 908 includes a hard disk, a nonvolatile memory, and the like.
  • the communication unit 909 includes a network interface or the like.
  • the drive 910 drives a removable medium 911 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
  • the CPU 901 loads the program recorded in the ROM 902 or the recording unit 908 to the RAM 903 via the input / output interface 905 and the bus 904, and executes the program. A series of processing is performed.
  • the program executed by the computer 900 can be provided by being recorded on a removable medium 911 as a package medium, for example.
  • the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
  • the program can be installed in the recording unit 908 via the input / output interface 905 by installing the removable medium 911 in the drive 910. Further, the program can be received by the communication unit 909 via a wired or wireless transmission medium and installed in the recording unit 908. In addition, the program can be installed in the ROM 902 or the recording unit 908 in advance.
  • the processing performed by the computer according to the program does not necessarily have to be performed in chronological order in the order described as the flowchart. That is, the processing performed by the computer according to the program includes processing executed in parallel or individually (for example, parallel processing or object processing).
  • the program may be processed by a single computer (processor) or may be distributedly processed by a plurality of computers.
  • the present technology can take the following configurations.
  • a receiver for receiving a stream including multi-channel or stereo audio components transmitted through a transmission path An acquisition unit for acquiring dialog control information for controlling a dialog of multi-channel or stereo audio transmitted in a system layer that processes signaling defined by a predetermined standard; And a control unit that controls a dialog of multi-channel or stereo audio transmitted through the transmission path based on the dialog control information.
  • the predetermined standard is a standard for digital broadcasting, The receiving unit receives a broadcast wave of the digital broadcast, The acquisition unit acquires the dialog control information transmitted in a system layer that processes signaling defined in the digital broadcasting standard, The receiving device according to (1), wherein the control unit controls a dialog of multi-channel or stereo audio transmitted through the broadcast wave.
  • the dialog control information is described in a newly defined descriptor.
  • the receiving apparatus according to (2).
  • the receiving apparatus according to (2), wherein the dialog control information is described in an extension area of a component descriptor.
  • the receiving apparatus according to (2) or (4), wherein the dialog control information is described in an extension area of an audio component descriptor.
  • the receiving apparatus according to (1), wherein the dialog control information is described as an application including an HTML5 (HyperText Markup Language 5) standard.
  • the dialog control information includes information related to dialog volume level control, dialog replacement control, or dialog localization position control.
  • the receiving device is Receiving a stream containing multi-channel or stereo audio components transmitted over a transmission line; Obtaining dialog control information for controlling a multi-channel or stereo audio dialog transmitted in a system layer that processes signaling defined by a predetermined standard; A data processing method including a step of controlling a dialog of multi-channel or stereo audio transmitted through the transmission path based on the dialog control information.
  • the predetermined standard is a standard for digital broadcasting
  • the transmission unit transmits the dialog control information together with the stream by the broadcast wave of the digital broadcast,
  • the transmission apparatus according to (9), wherein the dialog control information is transmitted in a system layer that processes signaling defined in the digital broadcasting standard.
  • the dialog control information is described in a newly defined descriptor.
  • the transmission device according to (10). (12) The transmission apparatus according to (10), wherein the dialog control information is described in an extension area of a component descriptor.
  • the transmission apparatus according to (9), wherein the dialog control information is described as an application including the HTML5 standard.
  • the transmission apparatus according to any one of (9) to (14), wherein the dialog control information includes information related to dialog volume level control, dialog replacement control, or dialog localization position control.
  • the transmitting device is Get a stream containing multi-channel or stereo audio components, Generate dialog control information to control multi-channel or stereo audio dialog, Transmitting the dialog control information together with the stream via a transmission path;
  • the dialog control information is a data processing method transmitted in a system layer that processes signaling defined by a predetermined standard.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Mathematical Physics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

La présente invention concerne un dispositif de réception, un dispositif d'envoi et un procédé de traitement de données, qui permettent d'effectuer une commande de dialogue à l'aide d'informations de commande de dialogue. Le dispositif de réception : reçoit un flux qui est transmis par l'intermédiaire d'un chemin de transmission et qui comprend une composante audio multicanal ou stéréo ; acquiert des informations de commande de dialogue qui sont transmises par une couche de système qui traite une signalisation spécifiée par une norme prescrite, et qui sont conçues pour commander le dialogue audio multicanal ou stéréo ; et commande le dialogue audio multicanal ou stéréo transmis par l'intermédiaire du chemin de transmission sur la base des informations de commande de dialogue. La présente invention est applicable, par exemple, à un récepteur TV qui est compatible avec un audio multicanal.
PCT/JP2016/071059 2015-07-31 2016-07-15 Dispositif de réception, dispositif d'envoi, et procédé de traitement de données WO2017022461A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2015-152906 2015-07-31
JP2015152906 2015-07-31

Publications (1)

Publication Number Publication Date
WO2017022461A1 true WO2017022461A1 (fr) 2017-02-09

Family

ID=57943848

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/071059 WO2017022461A1 (fr) 2015-07-31 2016-07-15 Dispositif de réception, dispositif d'envoi, et procédé de traitement de données

Country Status (1)

Country Link
WO (1) WO2017022461A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0447725A (ja) * 1990-06-15 1992-02-17 Nec Corp 音声信号多重化装置
WO2014184353A1 (fr) * 2013-05-16 2014-11-20 Koninklijke Philips N.V. Appareil de traitement audio et procédé associé
JP5713140B1 (ja) * 2014-12-05 2015-05-07 ソニー株式会社 受信装置、およびデータ処理方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0447725A (ja) * 1990-06-15 1992-02-17 Nec Corp 音声信号多重化装置
WO2014184353A1 (fr) * 2013-05-16 2014-11-20 Koninklijke Philips N.V. Appareil de traitement audio et procédé associé
JP5713140B1 (ja) * 2014-12-05 2015-05-07 ソニー株式会社 受信装置、およびデータ処理方法

Similar Documents

Publication Publication Date Title
EP3729425B1 (fr) Informations de priorité destinées à des données audio ambiophoniques d'ordre supérieur
KR101759005B1 (ko) 3d 오디오 계층적 코딩을 이용한 라우드스피커 포지션 보상
US10356545B2 (en) Method and device for processing audio signal by using metadata
JP6045696B2 (ja) オーディオ信号処理方法および装置
RU2617553C2 (ru) Система и способ для генерирования, кодирования и представления данных адаптивного звукового сигнала
Füg et al. Design, coding and processing of metadata for object-based interactive audio
US20200013426A1 (en) Synchronizing enhanced audio transports with backward compatible audio transports
US10271156B2 (en) Audio signal processing method
US20180091919A1 (en) Method and device for processing binaural audio signal
KR20140046980A (ko) 오디오 데이터 생성 장치 및 방법, 오디오 데이터 재생 장치 및 방법
US20200013414A1 (en) Embedding enhanced audio transports in backward compatible audio bitstreams
Sen et al. Efficient compression and transportation of scene-based audio for television broadcast
CN111492674B (zh) 在3d音频解码器中处理单声道信号以传送双耳内容
WO2017022461A1 (fr) Dispositif de réception, dispositif d'envoi, et procédé de traitement de données
US11270711B2 (en) Higher order ambisonic audio data
US11062713B2 (en) Spatially formatted enhanced audio data for backward compatible audio bitstreams
Herre et al. MPEG-H 3D audio: Immersive audio coding
US12035127B2 (en) Spatial audio capture, transmission and reproduction
US20230232182A1 (en) Spatial Audio Capture, Transmission and Reproduction
DeFilippis Mastering and Distributing Immersive Sound
Komori Trends in Standardization of Audio Coding Technologies
Fug et al. An Introduction to MPEG-H 3D Audio
Bleidt et al. Meeting the Requirements of Next-Generation Broadcast Television Audio
WO2016038876A1 (fr) Dispositif de codage, dispositif de décodage et dispositif de traitement de signal de parole
KR20140128563A (ko) 복호화 객체 리스트 갱신 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16832740

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: JP

122 Ep: pct application non-entry in european phase

Ref document number: 16832740

Country of ref document: EP

Kind code of ref document: A1