WO2015152661A1 - Procédé et appareil pour restituer un objet audio - Google Patents

Procédé et appareil pour restituer un objet audio Download PDF

Info

Publication number
WO2015152661A1
WO2015152661A1 PCT/KR2015/003326 KR2015003326W WO2015152661A1 WO 2015152661 A1 WO2015152661 A1 WO 2015152661A1 KR 2015003326 W KR2015003326 W KR 2015003326W WO 2015152661 A1 WO2015152661 A1 WO 2015152661A1
Authority
WO
WIPO (PCT)
Prior art keywords
type
audio object
audio
determining
spread parameter
Prior art date
Application number
PCT/KR2015/003326
Other languages
English (en)
Korean (ko)
Inventor
전상배
김선민
Original Assignee
삼성전자 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 삼성전자 주식회사 filed Critical 삼성전자 주식회사
Publication of WO2015152661A1 publication Critical patent/WO2015152661A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • H04S3/008Systems employing more than two channels, e.g. quadraphonic in which the audio signals are in digital form, i.e. employing more than two discrete digital channels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/11Positioning of individual sound objects, e.g. moving airplane, within a sound field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/03Application of parametric coding in stereophonic audio systems

Definitions

  • the present invention is directed to a method and apparatus for rendering an audio object.
  • the audio object When object rendering is performed on an audio signal including at least one audio object, such as Moving Picture Experts Group-H (MPEG-H), the audio object may be positioned at a virtual source location.
  • MPEG-H Moving Picture Experts Group-H
  • the interference between channels increases, so that the sound field and the space sense intended by the creator for each object may not be properly expressed. Can be.
  • the present invention relates to a method and apparatus for rendering an audio object for effectively expressing a color, a feeling, a distance, a feeling of space, etc. of a sound according to a creator's intention by determining a degree of spread of the audio object based on the type of the audio object.
  • the audio object may be output in an optimal state according to the producer's intention to provide to the listener.
  • FIG. 1 is a diagram illustrating an example of rendering an audio object in a two-dimensional space according to an embodiment.
  • FIG. 2 is a diagram illustrating an example of rendering an audio object in a 3D space according to an embodiment.
  • FIG. 3 is a diagram illustrating an example of rendering an audio object spread in a plurality of directions according to an embodiment.
  • FIG. 4 is a flowchart illustrating a method of determining a spread parameter based on a type of an audio object according to an exemplary embodiment.
  • FIG. 5 is a flowchart illustrating a method of determining a type of an audio object according to an exemplary embodiment.
  • FIG. 6 is a block diagram illustrating an internal structure of an apparatus for rendering an audio object according to an exemplary embodiment.
  • a method of rendering an audio object may include obtaining information related to the audio object; Determining a spread parameter representing a degree to which the audio object spreads in at least one direction based on the information related to the obtained object; Determining, according to the determined parameter, at least one direction in which the audio object is located; Based on the determined direction, rendering the audio object.
  • the information related to the object includes information about a type of an object, and the type of the object is classified according to whether the audio object provides a sense of space or reverberation to a listener.
  • the type of the audio object includes at least one of a direct type, an ambience type, an applause type, a soft decision type, and a dialog type.
  • the determining of the spread parameter includes determining the spread parameter to a value less than or equal to a reference value when the obtained object type is an ambience type or an applause type.
  • the determining of the spread parameter may include determining a spread parameter based on information indicating a degree of spatiality of the audio object when the obtained object type is a soft decision type.
  • the determining of the spread parameter includes determining the spread parameter based on at least one feature of the object and a rendering environment when the obtained object type is a direct type.
  • the determining of the spread parameter includes determining the spread parameter as a value equal to or greater than a reference value when the obtained object type is a dialog type.
  • an apparatus for rendering an audio object includes a receiver configured to receive an audio signal including at least one audio object and to extract an audio object from the audio signal; Acquiring information related to the audio object, determining a spread parameter indicating a degree to which the audio object spreads in at least one direction based on the information related to the obtained object, and according to the determined parameter, A controller which determines at least one orientation to be positioned and renders the audio object based on the determined direction; And a sound output unit configured to output the rendered audio object.
  • any part of the specification is to “include” any component, this means that it may further include other components, except to exclude other components unless otherwise stated.
  • a part is “connected” with another part, this includes not only the case where it is “directly connected” but also the case where it is “electrically connected” with another element between them.
  • part refers to a hardware component, such as software, FPGA or ASIC, and “part” plays certain roles. However, “part” is not meant to be limited to software or hardware.
  • the “unit” may be configured to be in an addressable storage medium and may be configured to play one or more processors.
  • a “part” refers to components such as software components, object-oriented software components, class components, and task components, processes, functions, properties, procedures, Subroutines, segments of program code, drivers, firmware, microcode, circuits, data, databases, data structures, tables, arrays and variables.
  • the functionality provided within the components and “parts” may be combined into a smaller number of components and “parts” or further separated into additional components and “parts”.
  • an audio object refers to each of sound components included in an audio signal.
  • One audio signal may include various audio objects.
  • the audio signal generated by recording the performance of an orchestra includes a plurality of audio objects generated from a plurality of musical instruments such as guitar, violin, and oboe.
  • the sound image means a position where the listener feels as where the sound source occurs.
  • the actual sound is output from the speaker, but the point where each sound source is virtually called a sound image.
  • the size and position of the sound image may vary according to the speaker from which sound is output. When the position of the sound of each sound source is clear and the sound of each sound source is well heard by the listener, it may be determined that the sound position is excellent.
  • the described technique is described based on the MPEG-H standard, but is not limited thereto, and may be applied to other audio coding techniques.
  • FIG. 1 is a diagram illustrating an example of rendering an audio object in a two-dimensional space according to an embodiment.
  • the audio object 100 may be output through the speakers 1 and 2 110 and 120.
  • the audio object 100 may be rendered and output to the speakers 1 and 2 110 and 120.
  • the volume of the sound constituting the audio object 100 output through the speakers 1 and 2 110 and 120 may be determined.
  • the speakers 1 and 2 110 may be positioned so that the sound image of the audio object 100 may be located at a position close to the speaker 1 110. 120 may be adjusted.
  • the acoustic size of the audio object 100 output through the speaker 1 110 may be greater than the acoustic size of the audio object 100 output through the speaker 2 120.
  • Equation 1 The sound of the audio object 100 that may be output through each speaker described above may be represented by Equation 1 below.
  • Direction vector indicating the sound image position of the audio object 100 Is a unit direction vector representing the output positions of the speakers 1 and 2 (110, 120) as It can be expressed as. Each direction vector may be determined according to the position of the sound image or the position of the speaker with respect to the position 130 of the listener.
  • g 1 and g 2 are gain factors that can be applied to the direction vectors of speakers 1 and 2, And A value that can be determined based on the value of.
  • the gain factor value corresponding to the sound volume of the audio objects 100 output to the speakers 1 and 2 may be determined according to the sound image of the audio object 100 and the positions of the speakers.
  • FIG. 2 is a diagram illustrating an example of rendering an audio object in a 3D space according to an embodiment.
  • an audio object may be output through three speakers by being rendered in channels m, n, and k. Since the position of the virtual sound source where the sound image of the audio object can be located is within a triangle formed by the channels m, n and k, the audio object may be output through the speakers of the channels m, n and k.
  • the direction vector p represents the position of the sound image on which the audio object is positioned with respect to the position 210 of the listener.
  • the vectors l m , l n , l k of the channels m, n, k indicate the positions of the channels m, n, k, respectively, with respect to the position 210 of the listener, and are unit vectors of size 1.
  • the direction vector p of the audio object may be expressed as Equation 2, similarly to Equation 1.
  • the gain factor values g m , g n , and g k corresponding to the loudness of the audio object 210 output to the speakers of channels m, n, and k are the sound of the audio object 210 and the speakers of channels m, n, and k. Can be determined according to their location.
  • the sound image of the audio object 100 may be positioned at a plurality of positions as well as one position as shown in FIGS. 1 and 2.
  • the sound image of the audio object 100 may exist in a plurality of directions spread within a predetermined range about the reference direction. This will be described in more detail with reference to FIG. 3.
  • FIG. 3 is a diagram illustrating an example of rendering an audio object spread in a plurality of directions according to an embodiment.
  • a sound image of an audio object may be positioned in a plurality of directions spreading in a range of an angle ⁇ about a reference direction 210 according to a multiple-direction amplitude panning (MDAP) method.
  • MDAP multiple-direction amplitude panning
  • the audio object When the audio object is output only in the reference direction 210, the audio object may be output through speakers 1, 2, and 5 constituting a triangle in which the reference direction 210 is located.
  • the audio objects when the audio objects are output in a plurality of directions as shown in FIG. 3, the audio objects may be output through the speakers 1, 2, 5, and 3 according to positions where sound images in each direction are located.
  • An audio object may be output through three speakers constituting a triangle at positions where sound images in each direction are positioned.
  • the plurality of directions p m in which the sound image of the audio object may be positioned may be determined according to a vector p 0 and an angle ⁇ representing the reference direction of the audio object, as shown in Equation 3 below.
  • Angle ⁇ represents a spread parameter to be described later.
  • p m ' is a direction vector value that can be determined according to a p 0 value representing a reference direction and a coordinate value corresponding to the p 0 value.
  • p m ' is a value that can be determined based on the reference position where the sound image of the audio object is located.
  • m may be determined as an integer value of 0 or more according to the number of directions in which the sound image of the audio object may be positioned.
  • a gain factor g value that can be applied to the p m value may be obtained according to Equation 1 or 2 described above.
  • the gain factor value may be determined based on the vector p m value for each direction in which the audio object is positioned and the direction vector value of the speaker from which each p m value is output.
  • the direction vector of the 330 is located in a triangle formed by the speakers 2, 3, and 5. Accordingly, the gain factor value for the direction vector of 330 may be determined based on the direction vector values of speakers 2, 3, and 5, which originate from the position 320 of the listener.
  • Equation 3 is merely an example, and the direction vector value in which the sound image of the audio object can be positioned can be obtained in various ways.
  • the degree of spreading between the direction vectors in which the sound image of the audio object may be positioned may be determined according to an ⁇ value that may be determined by the spread parameter. As the value of ⁇ increases, the maximum angle between the direction vectors in which the sound image may be positioned increases, so that the audio object may be output through more speakers.
  • the spread parameter of the audio object may be determined to provide a listener with a sense of space, reverberation, etc. according to the intention of the producer.
  • the apparatus may determine the spread parameter so that the audio object can be output with the intention or the optimal state of the producer according to the characteristics of the audio object.
  • the audio object provides the listener with reverberation or a sense of space
  • the greater the number of channels through which the audio object is output the higher the correlation between channels. Therefore, when the number of channels to which the audio object is output is large, the reverberation or space of the audio object may not be optimally provided to the listener.
  • the apparatus for rendering an audio object may determine a spread parameter according to a type of an audio object representing a characteristic of the audio object. As the spread parameter is determined according to the type of the audio object, the number of speakers to which the audio object is output may be adjusted according to the characteristics of the audio object.
  • the spread parameter may be determined according to an audio object type including a digitized value.
  • an audio object type including a digitized value may include a value according to a soft decision method.
  • the soft decision means a method of displaying data with information indicating the certainty of a data value, such as 1 close to 0 or 1 close to 1.
  • it means a data display method that displays data by using information of a real value rather than an integer or further includes other additional information.
  • the spread parameter may be determined according to the digitized data of the audio object belonging to the soft decision type.
  • FIG. 4 is a flowchart illustrating a method of determining a spread parameter based on information related to an audio object according to an exemplary embodiment.
  • the apparatus may obtain information related to an audio object to be rendered.
  • the information related to the audio object may include location information of the object, characteristic information, and the like.
  • the characteristic information of the object may include, for example, information about the type of the object.
  • the type of object may be classified according to whether the audio object provides a sense of space or reverberation to the listener. In addition, the type of the object may be classified based on whether the output performance or output characteristics of the audio object change as the audio object is output through a plurality of channels. For example, a spread parameter for a type of an audio object in which a change in output performance or characteristics of an audio object is insignificant as output to a plurality of channels may be determined as a relatively large value.
  • the type of the object is not limited thereto, and may be classified in various ways.
  • the type of the audio object may be obtained from information about the type of the audio object signaled through the bitstream, or the type of the audio object may be determined based on a result of analyzing the characteristics of the audio object.
  • the types of audio objects may be classified according to whether they provide a listener with a sense of space or reverberation.
  • the type of the audio object may be classified into a direct type, an ambience type, an appliance type, a soft decision type, a dialog type, and the like.
  • the ambience type may include an audio object that provides a sense of space by providing reverberation to the listener.
  • the applause type may include transient, transient, transient audio objects, such as claps or rain.
  • the dialog type may include an audio object including a human voice, a conversation, and the like.
  • the direct type may include an audio object from which spread parameters may be determined based on at least one feature of the object and the rendering environment.
  • the soft decision type may include an audio object whose spread parameters may be determined in accordance with artistic information of the sound associated with the spread parameters determined by the producer.
  • the information about the soft decision type audio object type may include specifically digitized information representing artistic information of a sound associated with a spread parameter.
  • Spread parameters may be determined based on the quantified information described above.
  • the digitized information may include a value indicating a degree of spatial sense of sound.
  • the digitized information may directly include spread parameter values.
  • the type of the audio object may be determined as one of a direct type and a soft decision type.
  • the device may determine a type more suitable for rendering the audio object of the two types, and render the audio object according to the determined object type.
  • One of the two types may be determined as the type of the audio object based on the characteristics of the audio object or the output environment of the audio object.
  • the apparatus may determine the spread parameter based on the information about the audio object to be rendered.
  • the spread parameter for the audio object may be determined to be a value below the reference value for the audio object of the ambience type or the applause type, which may provide a sense of space or reverberation.
  • the spread parameter may be determined to be a value less than or equal to a reference value for rendering with fewer channels for audio objects belonging to the ambience type or the applause type.
  • the spread parameter may be determined as a specific value below the reference value for each object type according to the intention of the producer.
  • the spread parameter may be determined as a specific value that allows the audio object to be optimally output among values below the reference value.
  • the audio object belonging to the dialog type is output in multiple channels due to its characteristics, the output performance of the audio object is not significantly affected. Dialog type audio objects are hardly affected by interference that may occur as they are output in multiple channels. Therefore, the spread parameter of the dialog type may be determined to be a value more than the reference value.
  • the spread parameter of an object belonging to the direct type may be determined based on at least one feature of the object and the rendering environment.
  • the spread parameter of the object belonging to the soft decision type may be determined as a value to be output as intended by the producer.
  • Spread parameters of the object may be determined based on specifically quantified information indicative of the intention of the producer.
  • the spread parameter may be determined based on not only the type of the object but also information related to the object, for example, location information of the object, characteristic information, and the like.
  • the device may determine at least one direction in which the sound image of the audio object is positioned based on the spread parameter determined in operation S420.
  • a plurality of directions that may be determined in step S430 will be referred to as a panning direction below.
  • the panning direction represents a vector value that can be determined within an angular range according to the spread parameter value about the reference direction.
  • the device may render the audio object based on the direction in which the audio object determined in operation S430 is positioned.
  • the device may obtain a gain factor for each panning direction of the audio object.
  • the gain factor may be determined based on a direction vector value indicating a position of each panning direction and channels forming a triangle in which each panning direction is located.
  • the device may render the audio object into a plurality of channels based on the panning direction and the gain factor of the audio object.
  • FIG. 5 is a flowchart illustrating a method of determining a type of an audio object according to an exemplary embodiment.
  • the device may acquire a type of an audio object.
  • the object types may be classified in various ways based on the degree to which the reverberation of the sound that may be provided to the listener, the sense of space, and the like may vary according to the extent to which the audio object is spread.
  • the type of the object may be classified based on whether the output performance or output characteristics of the audio object change as the audio object is output through a plurality of channels.
  • the type of the audio object may be obtained from information about the type of the audio object signaled through the bitstream, or the type of the audio object may be determined based on a result of analyzing the characteristics of the audio object.
  • the device may determine whether the type of the audio object acquired in operation S510 is an ambience type or an applause type.
  • step S530 if it is determined in step S520 that the type of the audio object is an ambience type or an applause type, the device may determine the spread parameter to be a value less than or equal to the reference value. Audio objects of the ambience type or the applause type may provide a listener with a sense of space or reverberation. Therefore, as the audio object belonging to the above type is rendered in a large number of channels, the interference phenomenon may increase. The apparatus may determine the spread parameter to a value below the reference value so as to minimize the interference phenomenon.
  • the apparatus may determine the spread parameter to a value capable of outputting the audio object optimally in consideration of the characteristics of the audio object, the output environment of the audio object, a user setting, and the like.
  • the device may determine whether the type of the audio object belongs to the direct type.
  • the apparatus may obtain a spread parameter based on at least one feature of the object and the rendering environment.
  • the device may determine whether the type of the audio object belongs to the dialog type.
  • the device may determine the spread parameter to be a value greater than or equal to the reference value according to the type of the audio object. Even if an audio object belonging to the dialog type is output in multiple channels, the output performance of the audio object is not significantly affected. Dialog type audio objects are hardly affected by interference that may occur as they are output in multiple channels. Therefore, the spread parameter of the dialog type may be determined to be a value greater than or equal to the reference value.
  • the device may determine the spread parameter so that the audio object is output through many channels, but may determine the spread parameter to a value less than or equal to the reference value according to the type of the audio object.
  • the device may determine that the audio object is a soft decision type.
  • the spread parameter may be determined by the spread parameter of the object based on specifically digitized information representing the intention of the producer. For example, the spread parameter may be determined based on a numerical value indicating the degree of spatiality of the object.
  • the device may render the audio object using the spread parameters determined in operations S530, S550, S570, and S580.
  • the rendered audio object may be output through at least one rendered channel.
  • FIG. 6 is a block diagram illustrating an internal structure of an apparatus for rendering an audio object according to an exemplary embodiment.
  • an apparatus 600 for rendering an audio object may be a terminal apparatus that may be used by a user.
  • the device 600 may be a smart television, ultra high definition (UHD) TV, monitor, personal computer (PC), notebook computer, mobile phone, tablet PC, navigation terminal, smart Smart phones, personal digital assistants (PDAs), portable multimedia players (PMPs), and digital broadcast receivers.
  • UHD ultra high definition
  • PC personal computer
  • notebook computer mobile phone
  • tablet PC personal digital assistants
  • PMPs portable multimedia players
  • digital broadcast receivers digital broadcast receivers.
  • the apparatus 600 for rendering an audio object may include a receiver 610, a controller 620, and a sound output unit 630.
  • the receiver 610 may receive an audio signal including an audio object for rendering from the outside. In addition, the receiver 610 may extract an audio object from the audio signal.
  • the audio signal may be received in the form of a bit stream, and the receiver 610 may extract an audio object from the bit stream including the audio signal.
  • the receiver 610 may extract information for analyzing the characteristics of the audio object or information for determining the type of the audio object from the bit stream.
  • the controller 620 may determine the spread parameter based on the information related to the audio object, and render the audio object according to the determined spread parameter.
  • the information related to the audio object may include location information of the object, characteristic information, and the like.
  • the characteristic information of the object may include, for example, information about the type of the object.
  • the spread parameter may be determined depending on whether the audio object provides a sense of space or reverberation to the listener.
  • the type of the object may be classified based on whether the output performance or output characteristics of the audio object change as the audio object is output through a plurality of channels.
  • the spread parameter of the object belonging to the soft decision type of the object type may be determined based on specifically digitized information indicating the intention of the producer.
  • the sound output unit 630 may output the audio object rendered by the controller 620 through a plurality of channels.
  • the audio object may be output in an optimal state according to the producer's intention to provide to the listener.
  • the method according to some embodiments may be embodied in the form of program instructions that may be executed by various computer means and recorded on a computer readable medium.
  • the computer readable medium may include program instructions, data files, data structures, etc. alone or in combination.
  • Program instructions recorded on the media may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well-known and available to those having skill in the computer software arts.
  • Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tape, optical media such as CD-ROMs, DVDs, and magnetic disks, such as floppy disks.
  • Examples of program instructions include not only machine code generated by a compiler, but also high-level language code that can be executed by a computer using an interpreter or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

L'invention concerne un procédé pour restituer un objet audio, comprenant les étapes consistant : à obtenir des informations associées à l'objet audio ; à déterminer, sur la base des informations obtenues associées à l'objet, un paramètre de propagation indiquant le degré de la propagation d'objet audio dans au moins une direction ; à déterminer au moins une direction dans laquelle l'objet audio est situé conformément aux paramètres déterminés ; et à restituer, sur la base de la direction déterminée, l'objet audio.
PCT/KR2015/003326 2014-04-02 2015-04-02 Procédé et appareil pour restituer un objet audio WO2015152661A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201461974052P 2014-04-02 2014-04-02
US61/974,052 2014-04-02

Publications (1)

Publication Number Publication Date
WO2015152661A1 true WO2015152661A1 (fr) 2015-10-08

Family

ID=54240883

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2015/003326 WO2015152661A1 (fr) 2014-04-02 2015-04-02 Procédé et appareil pour restituer un objet audio

Country Status (1)

Country Link
WO (1) WO2015152661A1 (fr)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9980078B2 (en) 2016-10-14 2018-05-22 Nokia Technologies Oy Audio object modification in free-viewpoint rendering
US10165386B2 (en) 2017-05-16 2018-12-25 Nokia Technologies Oy VR audio superzoom
US20190306651A1 (en) 2018-03-27 2019-10-03 Nokia Technologies Oy Audio Content Modification for Playback Audio
US10531219B2 (en) 2017-03-20 2020-01-07 Nokia Technologies Oy Smooth rendering of overlapping audio-object interactions
CN111213202A (zh) * 2017-10-20 2020-05-29 索尼公司 信号处理装置和方法以及程序
US11074036B2 (en) 2017-05-05 2021-07-27 Nokia Technologies Oy Metadata-free audio-object interactions
US11096004B2 (en) 2017-01-23 2021-08-17 Nokia Technologies Oy Spatial audio rendering point extension
US11395087B2 (en) 2017-09-29 2022-07-19 Nokia Technologies Oy Level-based audio-object interactions

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20100062773A (ko) * 2008-12-02 2010-06-10 한국전자통신연구원 오디오 컨텐츠 재생 장치
KR20100065121A (ko) * 2008-12-05 2010-06-15 엘지전자 주식회사 오디오 신호 처리 방법 및 장치
US20110013790A1 (en) * 2006-10-16 2011-01-20 Johannes Hilpert Apparatus and Method for Multi-Channel Parameter Transformation
WO2011020067A1 (fr) * 2009-08-14 2011-02-17 Srs Labs, Inc. Système pour diffuser en continu de manière adaptative des objets audio
KR20140017684A (ko) * 2011-07-01 2014-02-11 돌비 레버러토리즈 라이쎈싱 코오포레이션 향상된 3d 오디오 오서링과 렌더링을 위한 시스템 및 툴들

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110013790A1 (en) * 2006-10-16 2011-01-20 Johannes Hilpert Apparatus and Method for Multi-Channel Parameter Transformation
KR20100062773A (ko) * 2008-12-02 2010-06-10 한국전자통신연구원 오디오 컨텐츠 재생 장치
KR20100065121A (ko) * 2008-12-05 2010-06-15 엘지전자 주식회사 오디오 신호 처리 방법 및 장치
WO2011020067A1 (fr) * 2009-08-14 2011-02-17 Srs Labs, Inc. Système pour diffuser en continu de manière adaptative des objets audio
KR20140017684A (ko) * 2011-07-01 2014-02-11 돌비 레버러토리즈 라이쎈싱 코오포레이션 향상된 3d 오디오 오서링과 렌더링을 위한 시스템 및 툴들

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10433096B2 (en) 2016-10-14 2019-10-01 Nokia Technologies Oy Audio object modification in free-viewpoint rendering
US9980078B2 (en) 2016-10-14 2018-05-22 Nokia Technologies Oy Audio object modification in free-viewpoint rendering
US11096004B2 (en) 2017-01-23 2021-08-17 Nokia Technologies Oy Spatial audio rendering point extension
US10531219B2 (en) 2017-03-20 2020-01-07 Nokia Technologies Oy Smooth rendering of overlapping audio-object interactions
US11044570B2 (en) 2017-03-20 2021-06-22 Nokia Technologies Oy Overlapping audio-object interactions
US11604624B2 (en) 2017-05-05 2023-03-14 Nokia Technologies Oy Metadata-free audio-object interactions
US11442693B2 (en) 2017-05-05 2022-09-13 Nokia Technologies Oy Metadata-free audio-object interactions
US11074036B2 (en) 2017-05-05 2021-07-27 Nokia Technologies Oy Metadata-free audio-object interactions
US10165386B2 (en) 2017-05-16 2018-12-25 Nokia Technologies Oy VR audio superzoom
US11395087B2 (en) 2017-09-29 2022-07-19 Nokia Technologies Oy Level-based audio-object interactions
CN111213202A (zh) * 2017-10-20 2020-05-29 索尼公司 信号处理装置和方法以及程序
US10542368B2 (en) 2018-03-27 2020-01-21 Nokia Technologies Oy Audio content modification for playback audio
US20190306651A1 (en) 2018-03-27 2019-10-03 Nokia Technologies Oy Audio Content Modification for Playback Audio

Similar Documents

Publication Publication Date Title
WO2015152661A1 (fr) Procédé et appareil pour restituer un objet audio
US10674262B2 (en) Merging audio signals with spatial metadata
US10206030B2 (en) Microphone array system and microphone array control method
JP6039111B2 (ja) マルチチャネルオーディオデータのビデオ解析支援生成
WO2013147547A1 (fr) Dispositif audio et procédé correspondant de conversion de signal audio
WO2014088328A1 (fr) Appareil de fourniture audio et procédé de fourniture audio
WO2011115430A2 (fr) Procédé et appareil de reproduction sonore en trois dimensions
WO2012161555A2 (fr) Système de séparation de signaux utilisant un réseau de microphones directionnels et procédé permettant de mettre en œuvre ce système
WO2016089049A1 (fr) Procédé et dispositif d'émission d'un signal audio sur la base d'informations d'emplacement d'un haut-parleur
WO2017209477A1 (fr) Procédé et dispositif de traitement de signal audio
WO2019004524A1 (fr) Procédé de lecture audio et appareil de lecture audio dans un environnement à six degrés de liberté
WO2014061931A1 (fr) Dispositif et procédé de lecture de son
EP2737727A2 (fr) Procédé et appareil conçus pour le traitement d'un signal audio
EP3276982B1 (fr) Appareil de traitement d'informations, procédé de traitement d'informations et programme
WO2014175668A1 (fr) Procédé de traitement de signal audio
US9584761B2 (en) Videoconference terminal, secondary-stream data accessing method, and computer storage medium
WO2016123901A1 (fr) Terminal et procédé permettant de lire un signal audio de manière directionnelle par ledit terminal
EP2743917B1 (fr) Système d'informations, appareil de reproduction d'informations, procédé de génération d'informations et support de stockage
KR102508815B1 (ko) 오디오와 관련하여 사용자 맞춤형 현장감 실현을 위한 컴퓨터 시스템 및 그의 방법
US20190289418A1 (en) Method and apparatus for reproducing audio signal based on movement of user in virtual space
JP2022083443A (ja) オーディオと関連してユーザカスタム型臨場感を実現するためのコンピュータシステムおよびその方法
EP3494712A1 (fr) Appareil électronique, et procédé de commande associé
JP2018019294A (ja) 情報処理システム及びその制御方法、コンピュータプログラム
WO2015060696A1 (fr) Procédé et appareil de reproduction de son stéréophonique
WO2023231787A1 (fr) Procédé et appareil de traitement audio

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15773932

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase
122 Ep: pct application non-entry in european phase

Ref document number: 15773932

Country of ref document: EP

Kind code of ref document: A1