WO2024132129A1 - Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible - Google Patents

Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible Download PDF

Info

Publication number
WO2024132129A1
WO2024132129A1 PCT/EP2022/087262 EP2022087262W WO2024132129A1 WO 2024132129 A1 WO2024132129 A1 WO 2024132129A1 EP 2022087262 W EP2022087262 W EP 2022087262W WO 2024132129 A1 WO2024132129 A1 WO 2024132129A1
Authority
WO
WIPO (PCT)
Prior art keywords
room
perceptual
data processing
acoustic parameters
processing apparatus
Prior art date
Application number
PCT/EP2022/087262
Other languages
English (en)
Inventor
Liyun PANG
Shivam Saini
Stephan Werner
Lukas TREYBIG
Ulrike SLOMA
Original Assignee
Huawei Technologies Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co., Ltd. filed Critical Huawei Technologies Co., Ltd.
Priority to PCT/EP2022/087262 priority Critical patent/WO2024132129A1/fr
Publication of WO2024132129A1 publication Critical patent/WO2024132129A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • H04S7/303Tracking of listener position or orientation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • H04S7/303Tracking of listener position or orientation
    • H04S7/304For headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • H04S7/306For headphones

Definitions

  • the present disclosure relates to audio processing in general. More specifically, the disclosure relates to an apparatus and method for estimating the perceptual acoustics of a target room.
  • 3D sound can be defined as a sound arriving at a location, for example the ears of a listener, from varying directions and varying distances, which can contribute for example to a three- dimensional aural image humans hear.
  • 3D audio rendering can comprise creating a sound world by attaching a characteristic sound to virtual objects in a virtual room (also called environment or scenery) to synthesize as a 3D sound.
  • a virtual room also called environment or scenery
  • 3D audio rendering for instance, via headphones, large acoustic and also perceptual differences between a virtual, e.g. synthesized or rendered room and the real listening environment defined as a target room may degrade the spatial plausibility and immersion and, thus, lead to the so-called “Room Divergence Effect”.
  • a data processing apparatus for estimating the perceptual acoustics of a target room with respect to a perceptual attribute of the target room.
  • the data processing apparatus is configured to obtain a plurality of physical acoustic parameters for the target room and obtain the plurality of physical acoustic parameters, i.e. the same parameters as for the target room, for a virtual room, e.g. from a database.
  • the data processing apparatus is further configured to estimate a perceptual acoustic distance measure value for the selected perceptual attribute between the target room and the virtual room based on the plurality of physical acoustic parameters of the target room and on the plurality of physical acoustic parameters of the virtual room and based on a perceptual acoustic quality model.
  • the perceptual acoustic quality model defines a mapping, in particular a correlation, a) between the plurality of physical acoustic parameters and a first value or measure of the perceptual attribute of the target room and b) between the plurality of physical acoustic parameters and a second value or measure of the perceptual attribute of the virtual room.
  • the perceptual attribute of the target room and the virtual room comprises or is an envelopment attribute, a coloration attribute, a plausibility attribute, or an external ization attribute.
  • the data processing apparatus is configured to measure the plurality of physical acoustic parameters for the target room for obtaining the plurality of physical acoustic parameters for the target room.
  • the data processing apparatus is configured to obtain the plurality of physical acoustic parameters, i.e. the same parameters as for the target room, for the virtual room from a database of physical acoustic parameters for a plurality of virtual rooms.
  • the plurality of physical acoustic parameters of the target room and the virtual room comprises: an energy decay curve, EDC, parameter; a reverberation time parameter; a definition parameter; a speech transmission index, STI, parameter; a clarity index parameter; a direct-to-reverberant ratio, DRR, parameter; a centre time parameter; an inter-aural cross-correlation, IACC, parameter and/or a late lateral energy parameter.
  • the perceptual acoustic quality model defines, i.e. comprises for each perceptual attribute a correlation matrix between the pluralities of physical acoustic parameters of the target room and the virtual room and the first and second values or measures of the perceptual attribute of the target room and the virtual room for defining the mapping, in particular correlation, a) between the plurality of physical acoustic parameters and the first value or measure of the perceptual attribute of the target room and b) between the plurality of physical acoustic parameters and the second value or measure of the perceptual attribute of the virtual room.
  • the perceptual acoustic quality model defines, i.e. comprises for each perceptual attribute a correlation matrix between weighted linear combinations, i.e. the LDA discriminant functions, i.e. LDs determined by means of a linear discriminant analysis, of the plurality of physical acoustic parameters of the target room and the first value or measure of the perceptual attribute of the target room and between weighted linear combinations of the plurality of physical acoustic parameters of the virtual room and the second value or measure of the perceptual attribute of the virtual room.
  • the LDA discriminant functions i.e. LDs determined by means of a linear discriminant analysis
  • the perceptual attribute of the target room and the virtual room is a coloration attribute and wherein the weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a reverberation time parameter for an energy decay by 30 dB having the largest weight.
  • the weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a reverberation time parameter for an energy decay by 20 dB with a weight having an absolute value in the range from about 0.33 to about 0.53, in particular 0.4382, the reverberation time parameter for an energy decay by 30 dB with a weight having an absolute value in the range from about 0.9 to about 1.0, in particular 1.0, and a clarity index parameter for 80 ms with a weight having an absolute value in the range from about 0.29 to 0.49, in particular 0.3958.
  • the perceptual attribute of the target room and the virtual room is an envelopment attribute.
  • the weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room may comprise a clarity index parameter for 50 ms having the largest weight.
  • the weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a reverberation time parameter for an energy decay by 20 dB with a weight having an absolute value in the range from about 0.9 to about 1.00, in particular 0.9533, a reverberation time parameter for an energy decay by 30 dB with a weight having an absolute value in the range from about 0.85 to about 0.95, in particular 0.8834, the clarity index parameter for 50 ms having an absolute value in the range from about 0.9 to about 1.0, in particular 1 .0, and a clarity index parameter for 80 ms with a weight having an absolute value in the range from about 0.9 to 1 .00, in particular 0.9228.
  • the perceptual acoustic quality model defines, i.e. comprises for each perceptual attribute a correlation matrix between a cosine similarity of a first and a second weighted linear combination of the plurality of physical acoustic parameters of the target room and the first value or measure of the perceptual attribute of the target room and between a cosine similarity of the first and the second weighted linear combination of the plurality of physical acoustic parameters of the virtual room and the second value or measure of the perceptual attribute of the virtual room.
  • the perceptual attribute of the target room and the virtual room is a plausibility attribute.
  • the first weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 30 dB having the largest weight of the first weighted linear combination and the second weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a clarity index parameter for 50 ms having the largest weight of the second weighted linear combination.
  • the first weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a reverberation time parameter for an energy decay by 20 dB with a weight having an absolute value in the range from about 0.33 to about 0.53, in particular 0.4382, the reverberation time parameter for an energy decay by 30 dB with a weight having an absolute value in the range from about 0.9 to about 1.0, in particular 1.0, and a clarity index parameter for 80 ms with a weight having an absolute value in the range from about 0.29 to 0.49, in particular 0.3958.
  • the second weighted linear combination of the plurality of physical acoustic parameters of the target room and the virtual room comprises a reverberation time parameter for an energy decay by 20 dB with a weight having an absolute value in the range from about 0.9 to about 1.00, in particular 0.9533, a reverberation time parameter for an energy decay by 30 dB with a weight having an absolute value in the range from about 0.85 to about 0.95, in particular 0.8834, the clarity index parameter for 50 ms having an absolute value in the range from about 0.9 to about 1.0, in particular 1 .0, and a clarity index parameter for 80 ms with a weight having an absolute value in the range from about 0.9 to 1 .00, in particular 0.9228.
  • the data processing apparatus is further configured to determine a physical acoustic distance measure value between the target room and the virtual room based on the plurality of physical acoustic parameters of the target room and the plurality of physical acoustic parameters of the virtual room.
  • the data processing apparatus further comprises a display configured to display a graphical user interface configured to illustrate the perceptual acoustic distance measure value for the selected perceptual attribute between the target room and the virtual room.
  • the data processing apparatus is configured to estimate a respective perceptual acoustic distance measure value for the selected perceptual attribute between the target room and a plurality of virtual rooms and to determine a best-matching virtual room of the plurality of virtual rooms having the smallest perceptual acoustic distance measure value relative to the target room.
  • the data processing apparatus is further configured to obtain an impulse response function and/or a transfer function associated with the best-matching virtual room rooms having the smallest perceptual acoustic distance measure value relative to the target room.
  • a computer-implemented data processing method for estimating the perceptual acoustics of a target room with respect to a perceptual attribute of the target room comprises the steps of: obtaining a plurality of physical acoustic parameters for the target room; obtaining the plurality of physical acoustic parameters, i.e. the same parameters as for the target room, for a virtual room, e.g.
  • the perceptual acoustic quality model defines a mapping, in particular a correlation, a) between the plurality of physical acoustic parameters and a first value or measure of the perceptual attribute of the target room and b) between the plurality of physical acoustic parameters and a second value or measure of the perceptual attribute of the virtual room.
  • the data processing method according to the second aspect allows efficiently estimating the perceptual acoustics of a target room based on the physical acoustic parameters of the target room.
  • the method according to the second aspect can be performed by the data processing apparatus according to the first aspect.
  • further features of the method according to the second aspect result directly from the functionality of the data processing apparatus according to the first aspect as well as its different implementation forms and embodiments described above and below.
  • a computer program product comprising a computer- readable storage medium for storing program code which causes a computer or a processor to perform the method according to the second aspect, when the program code is executed by the computer or the processor.
  • Fig. 1 is a schematic diagram illustrating a data processing apparatus according to an embodiment for estimating the perceptual acoustics of a target room;
  • FIG. 2a and 2b are schematic diagrams illustrating data flows implemented by a data processing apparatus according to an embodiment
  • Fig. 3a is a schematic diagram illustrating an acoustic similarity estimation module implemented by a data processing apparatus according to an embodiment
  • Fig. 3b is a schematic diagram illustrating an acoustic parameter calculation module implemented by the data processing apparatus of figure 3a;
  • Fig. 4 is a graphical diagram illustrating weights of different linear combinations of a plurality of physical acoustic parameters implemented by a data processing apparatus according to an embodiment
  • Fig. 5 shows a matrix illustrating numerical values of the weights of figure 4.
  • Fig. 6a and 6b are schematic diagrams illustrating a classification of rooms from a database of rooms based on different LD discriminant functions
  • Fig. 7 shows processing blocks for generating a perceptual quality model implemented by a data processing apparatus according to an embodiment
  • Fig. 8 shows an exemplary correlation analysis for determining a perceptual quality model implemented by a data processing apparatus according to an embodiment
  • Fig. 9 shows a schematic diagram illustrating an implementation of a perceptual quality model with a visualization module of a data processing apparatus according to an embodiment
  • Fig. 10 shows a graphical user interface of a display of a data processing apparatus according to an embodiment
  • Fig. 11 is a computer-implemented data processing method according to an embodiment for estimating the perceptual acoustics of a target room with respect to a perceptual attribute of the target room.
  • a disclosure in connection with a described method may also hold true for a corresponding device or system configured to perform the method and vice versa.
  • a corresponding device may include one or a plurality of units, e.g. functional units, to perform the described one or plurality of method steps (e.g. one unit performing the one or plurality of steps, or a plurality of units each performing one or more of the plurality of steps), even if such one or more units are not explicitly described or illustrated in the figures.
  • a specific apparatus is described based on one or a plurality of units, e.g.
  • FIG. 1 is a schematic diagram illustrating a data processing apparatus 100 according to an embodiment.
  • the data processing apparatus 100 may comprises a processor 101.
  • the processor 101 may be implemented in hardware and/or software and may comprise digital circuitry, or both analog and digital circuitry.
  • Digital circuitry may comprise components such as application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), digital signal processors (DSPs), or general-purpose processors.
  • the data processing apparatus 100 may further comprise a memory 103, e.g. a non-transitory memory or nonvolatile memory, configured to store executable program code which, when executed by the processor 101 , causes the data processing apparatus 100 to perform the functions and methods described herein.
  • the data processing apparatus 100 may further comprise a display 105 for displaying results of the processing performed by the processor 101.
  • the display 105 may be a touchscreen.
  • a virtual room can be defined as an acoustic environment or scenery.
  • the virtual objects are provided with room acoustics which can be assumed as a room of the acoustic scenery and can be used in a Binaural Reproduction System, as further described below.
  • the acoustics of the scenery may need to be sufficiently similar to the room acoustics of the reproduction room to create a high spatial audio quality, represented, e.g. by perceptual attributes such as for example plausibility and externalization.
  • a target room can be defined as an acoustic environment or scenery.
  • the reproduction room is something like a target room for the virtual acoustics.
  • the room acoustics of the target room are known, e.g. by room acoustics measurements, virtual audio objects can be provided directly with the acoustics of the target room.
  • the acoustics of other rooms e.g. virtual rooms from a database, can be used.
  • the room acoustic similarity can be realized by evaluating the similarity of single or multiple room acoustic parameters and can be represented as distance.
  • the distance may also be used for interlinking acoustic distance with quality ratings at divergent and congruent room scenarios from exemplary perceptual evaluations to estimate the effect on spatial audio quality.
  • the representation of the distance can also be useful for cases where an audio-AR scene is created and it is considered to check how well the virtual audio objects fit into the acoustics of the target room, for example if the creator is not in the target room itself when creating the audio-AR scene.
  • Physical acoustic parameters also referred to as room acoustical parameters or more general environment acoustical parameters, can be used for acoustic distance calculation and could be extracted from measured room impulse responses or binaural room impulse responses.
  • Such physical acoustic parameters can comprise the following:
  • EDT Energy Decay Curve
  • Reverberation Time which refers to a time taken for energy to decay by ‘60’ dB (generally 60 dB).
  • Definition (D50) which refers to a ratio of energy components up to 50ms after the direct sound to the total energy of RIR.
  • Speech transmission index ST I which refers to an objective measure for predicting the intelligibility of speech. It has a value range between 0 and 1.
  • Clarity Index which refers to a logarithmic ratio between an early sound energy (until 50ms) to a later sound energy.
  • Clarity Index which refers to a logarithmic ratio between an early sound energy (until 80ms) to a later sound energy.
  • DRR Direct-to-Reverberant Ratio
  • Ts Centre Time
  • IACC Inter-aural Cross Correlation
  • Late lateral energy Lj which refers to the logarithmic ratio of the lateral sound energy from 80 ms after the direct sound and the total sound energy measured at a distance of 10m (free field).
  • Perceptual attributes also referred to as perceptual quality features or attributes, may generally depends on the context and the task. Such perceptual attributes can comprise the following:
  • Envelopment which refers to a listener envelopment impression of being surrounded by the reproduced sound field or audio signal. For example, “not surrounded at all”, “less surrounded”, “slightly less surrounded”, “medium surrounded”, “slightly more surrounded”, “more surrounded”, “completely surrounded” can be used for ratings in a subjective listening test.
  • Coloration which refers to a timbral impression which is determined by the ratio of high to low frequency components.
  • “extremely muffled”, “muffled”, “slightly muffled”, “well balanced”, “slightly bright”, “bright”, “extremely bright” can be used for ratings in a subjective listening test.
  • Plausibility which refers to a plausible auditory illusion or acoustic room congruence impression of how well the heard audio signal fits into the current listening environment, considering room acoustic characteristics. For example, “extremely bad fit”, “bad fit”, “poor fit”, “fair fit”, “good fit”, “excellent fit”, “ideal fit” can be used for ratings in a subjective listening test.
  • Externalization which refers to a perception of the audio signal being placed outside the head of a receiver or listener within the surrounding environment or being placed inside the head of a receiver or listener, including the ability to localize the direction of the incoming sound.
  • “inside the head, but diffuse”, “inside the head, localizable”, “very near the head, localizable”, “outside the head, localizable”, “outside the head, but diffuse” can be used for ratings in a subjective listening test.
  • the general idea of embodiments disclosed herein is to provide an efficient and intuitive data processing apparatus 100 for calculation and representation of acoustic and perceptual distances between two or more spatial audio signals using room acoustic parameters.
  • the spatial audio signals may at least be one virtual sound source embedded in a room acoustics, i.e. target room, and at least one further room acoustics, i.e. virtual room, to be compared against. Further spatial audio signals can be further room acoustics, i.e. further virtual rooms, against which a further comparison can be performed.
  • the perceptual distances (or similarities) can be related to the perceptual and intuitive attributes such as envelopment, externalization, plausibility, coloration and others.
  • the data processing apparatus 100 is configured for estimating the perceptual acoustics of a target room with respect to a perceptual attribute of the target room.
  • the data processing apparatus 100 is configured to obtain a plurality of physical acoustic parameters 111 , 113 for the target room and obtain the plurality of physical acoustic parameters 111 , 113, i.e. the same parameters as for the target room for a virtual room, e.g. from a database 201a (illustrated in figure 2).
  • the data processing apparatus 100 is further configured to estimate a perceptual acoustic distance measure value 115 for the selected perceptual attribute between the target room and the virtual room based on the plurality of physical acoustic parameters 111 of the target room and on the plurality of physical acoustic parameters 113 of the virtual room and based on a perceptual acoustic quality model 150.
  • the perceptual acoustic quality model 150 may be stored in the memory 103 of the data processing apparatus 100.
  • the perceptual acoustic quality model 150 defines a mapping, in particular a correlation a) between the plurality of physical acoustic parameters 111 , 113 and a first value or measure of the perceptual attribute of the target room and b) between the plurality of physical acoustic parameters 111 , 113 and a second value or measure of the perceptual attribute of the virtual room.
  • the perceptual attribute of the target room and the virtual room may comprise or is an envelopment attribute, a coloration attribute, a plausibility attribute, or an externalization attribute, in particular as described above.
  • the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise: an energy decay curve, EDC, parameter; a reverberation time parameter; a definition parameter; a speech transmission index, STI, parameter; a clarity index parameter; a direct-to-reverberant ratio, DRR, parameter; a centre time parameter; an inter-aural crosscorrelation, IACC, parameter and/or a late lateral energy parameter, in particular as described above.
  • Figures 2a and 2b are schematic diagrams illustrating data flows in the data processing apparatus 100 according to an embodiment. More specifically, figure 2b shows a first spatial audio signal 220 or binaural audio signal 220 characterized by room acoustics 221 of the virtual room and a second spatial audio signal 230 or binaural audio signal 230 characterized by room acoustics 231 of the target room.
  • the room acoustics 221 , 231 may comprise a direct sound, early acoustic room reflections, and late acoustic room reflections.
  • the data processing apparatus 100 may be configured to assess perceptual/auditory distances (or similarities) between two or more spatial audio signals 220, 230 using the physical acoustic parameters 111 , 113 of the rooms.
  • the data processing apparatus 100 may perform a) a comparison of room acoustic parameters using statistical data analysis and calculation of acoustic distance between different room acoustics, and b) an estimation of perceptual distance using a perceptual quality model which describes the effect of acoustic room divergence on perceived spatial audio quality.
  • the room acoustics 221 , 231 may be directionally weighted representations (binaural and/or monaural) and/or omnidirectional representations.
  • the acoustic representations may be transfer functions 203a-c in the form of BRI Rs, SRI Rs, and/or RIRs, recordings 201b or spatial audio signals which have been created by simulations 201c.
  • the recordings 201b map comprise spatial audio signals recorded by microphones and/or a piece of audio signal recorded in a room (or environment/scenery), for instance, with a mono microphone or a pair of binaural microphones.
  • the room acoustics 221 , 231 may be calculated by the data processing apparatus 100 from the acoustic representations and/or they originate from other sources, for example from the database 201a.
  • the room acoustic parameters, i.e. the plurality of physical acoustic parameters 111 , 113 may be calculated by the data processing apparatus 100 from the transfer functions 203a-c or from microphone recordings 201 b.
  • the room acoustic parameters i.e. the plurality of physical acoustic parameters 111 , 113 may be previously documented parameters for the acoustic description of rooms (e.g. Teo, DRR, Cso, Cso, IACC) and/or new or adapted parameters for the description of audio-AR scenes.
  • the room acoustics parameters i.e. the plurality of physical acoustic parameters 111 , 113 may come directly from measurements or recordings 210b, come from the database 201a where room acoustics parameters for different rooms are stored, come from simulations 201c of room acoustics, be set as values elsewhere (e.g. by manual specification of Teo, DRR, etc.) or come from other sources such as a remote server.
  • the data processing apparatus 100 may comprise a module 240 for determining similarity and comparison of the physical acoustic parameters 111 , 113 of the rooms, which may be realized by statistical data analysis and for calculating an acoustic distance between different room acoustics.
  • Data analysis may be performed by statistical ratios and considering single or combinations of the physical acoustic parameters 11 , 113 of the rooms. Additionally or alternatively, data analysis may be performed by applying methods of multivariate statistics, e.g. principal component analysis, cluster analysis, multidimensional scaling and others.
  • multivariate statistics e.g. principal component analysis, cluster analysis, multidimensional scaling and others.
  • the similarity between the groups/cl usters may be calculated, e.g. cosine similarity, geometric distance, etc.
  • the similarity may be a measure of multivariate acoustic distance.
  • the groups/clusters may be called spatial classes.
  • the perceptual acoustic quality model 150 may be designed based on the correlation matrices between the plurality of physical acoustic parameters 111 , 113 and the perceptual attributes. To determine the correlation, subjective evaluation/listening tests may be designed and organized for accessing the overall subjective audio quality as well as some describing attributes.
  • binaural audio signals auralized with spatial room acoustics of different rooms i.e. the virtual room illustrated in figure 2b may be used as stimuli for the listening tests.
  • the listening tests may be conducted for different Quality Features, i.e. the perceptual attributes such as plausibility, coloration, envelopment, externalization, etc., separately.
  • An evaluation room may be used for the listing tests, where the listening tests were conducted, and which may be the target room.
  • the data processing apparatus 100 may further comprise a module 250 configured for estimation of the perceptual distance 115, which may be realized by the perceptual acoustic quality model 150 describing the effect of acoustic room divergence on perceived spatial audio quality.
  • a room database which may be the same database as the database 201a ora different database, may be used from which combinations of convergent and divergent audio scenes may be selected for perceptual evaluation.
  • the evaluation may measure the perceived quality for the features, i.e. perceptual attributes: Plausibility, Externalization, Coloration, and others.
  • the data processing apparatus 100 may interlink the rated quality and the acoustic similarity to estimate the correlation between acoustics and auditory perception.
  • the data processing apparatus 100 can be suited for auditory AR use cases as well as for position-dynamic binaural synthesis.
  • FIG 2a an exemplary use case as a binaural auralization system used in spatial audio rendering for a headphone 209a and loudspeaker 209b is illustrated. This may involve using the transfer functions 203a-c for processing spatial room acoustics, in particular RIR, BRIR and SRIR as described above and a following spatial audio rendering 205. Resulting spatial audio signals 220, 230 may then be processed by a binaural playback 207 in order to achieve binaural audio signals for the headphone 209a or the loudspeakers 209b.
  • a virtual room based on certain spatial room acoustics may be used to enhance overall spatial experience.
  • the playback device may be the headphone 209a or the loudspeakers 209b.
  • loudspeakers 209b additional processing such as crosstalk cancellation processing may be required.
  • AR applications it may be important to use the perceptual “similar” virtual room as the real listening room (target room) to result in a plausible illusion.
  • the data processing apparatus 100 may be used for AR applications where it is important to have a perceptually congruent virtual-target acoustics.
  • Fixed rooms are standard and currently being implemented in the spatial audio/3D audio features of more and more products.
  • a perceptually convergent, i.e. the opposite to congruent, virtual-target acoustics will destroy the plausibility and also the overall user experience.
  • a best matching virtual room can be selected from a pre-collected database based on estimated perceptual virtual-target distance in order to give the best coherent virtual-target acoustics. If the physical acoustic parameters 111 , 113 of the target room are unknown, a piece of audio signals (speech, music or noise etc.) can be recorded with a mono or a pair of binaural microphones and the physical acoustic parameters 111 , 113 are calculated from the recorded signals.
  • the data processing apparatus 100 can also be used for content creation and spatial audio mixing by visualizing acoustic distance and perceptual distance of different virtual rooms.
  • Figure 3a is a schematic diagram illustrating an acoustic similarity estimation module implemented by the data processing apparatus 100 according to an embodiment.
  • a first step may be to compare the room acoustic parameters using statistical data analysis and calculate acoustic distance between different room acoustics.
  • a second step may involve an estimation of perceptual distance using the perceptual acoustic quality model 150 which describes the effect of acoustic room divergence on perceived spatial audio quality.
  • the acoustic similarity estimation module shown in figure 3a may be configured to estimate acoustic similarity using single physical acoustic parameters 111 , 113 and/or multi-variate statistics.
  • the room acoustic database 201a which may comprise the measurement of multiple rooms and the analysis of relevant physical acoustic parameters 111 , 113 of the rooms may be coupled to the acoustic similarity estimation module.
  • the transfer functions 203a-c in particular RIR, BRIR or SRIR, are measured or simulated for one or multiple DoA and DoV in certain rooms, one or more of the room acoustic parameters, i.e. the plurality of physical acoustic parameters 111 , 113, as described above may be extracted by an acoustic parameter calculation 301.
  • an acoustic parameter calculation 301 Based on the transfer functions 203a-c and/or an audio recording 201b as an input 301a, an acoustic parameter calculation 301 , which is illustrated in figure 3b in more detail, is performed.
  • the parameter calculation 301 the physical acoustic parameters 111 , 113 for broadband signals and/or for several frequency bands based on a broadband or frequency-band decomposition 301b may be calculated.
  • the corresponding output 301c may be the room acoustic parameters, i.e. the plurality of physical acoustic parameters 111 , 113.
  • Multivariate statistics 307 such as PCA and/or by applying LDA 303
  • the physical acoustic parameters 111 , 113 of the rooms can be described by a few linear combinations (principal component in PCA or discriminant function/component in LDA), i.e. by providing a parameter difference 305 or a similarity index 309.
  • different rooms can be classified on the basis of the physical acoustic parameters 111 , 113 of the rooms by describing them by a few meaningful combinations of them.
  • this may involve supplying at least some of the plurality of physical acoustic parameters 111 , 113 by the room acoustic database 201a.
  • observations by means of LDA are considered by example, where the room separate is illustrated in a more comprehensible manner in the LDA than in the PCA.
  • Figure 4 is a graphical diagram illustrating the weights of specific parameters of the plurality of physical acoustic parameters 111 ,113 in regard to linear combinations of the LDA.
  • a first linear combination to an eight linear combination is illustrated (also referred to as LD1 to LD8 in the following).
  • the specific parameters of the plurality of physical acoustic parameters 111 ,113 are illustrated, namely from top to bottom, EDT, T20, T30, D50, C50, Cso, DRR and T s .
  • the matrix W may comprise the following weights:
  • the perceptual acoustic quality model 150 may define, i.e. comprises for each perceptual attribute a correlation matrix, such as the matrix W for example, between the pluralities of physical acoustic parameters 111 , 113 of the target room and the virtual room and the first and second values or measures of the perceptual attribute of the target room and the virtual room for defining the mapping, i.e.
  • Figures 6a and 6b are schematic diagrams illustrating a classification of the rooms from the room database 201a using different LD discriminant functions/components.
  • Figure 6a shows the first and second linear combination, i.e. LD1 and LD2 and figure 6b shows the first to the third linear combination, i.e. LD1 to LD3.
  • the data processing apparatus 100 may determine, how the measured rooms differentiate and could be classified, and which combinations of the plurality of physical acoustic parameters 111 , 113 can explain these differentiations.
  • distance measures may be further used. Possible distance measure comprises to (i) calculate a mean value of each point cloud, (ii) calculate a distance between reference room to all rooms, and (iii) calculate a cosine similarity between reference room and all other rooms.
  • the data processing apparatus 100 may determine which of the LD1 to LD8 are sufficient to describe most of the data.
  • the observations of the first three LD functions/components, i.e. LD1 to LD3 are considered to describe most of the data set. This is illustrated in figures 6a and 6c, where LD1 to LD3 describe up to 99% of the data combined.
  • the distance and the cosine similarity may be calculated as illustrated in equation 1 and equation 2 below.
  • Equation 1 may be as follows:
  • Equation 2 may be as follows:
  • LD n describes the mean value of the observations of the n-th discriminant function for the target/reference room R re f and the virtual room R, used in binaural rendering.
  • R re f may be the room in which the evaluation/listening test was performed, and R, may be the virtual room used to generate the spatial audio signals with room acoustic parameters for listening.
  • the target of the perceptual acoustic quality model 150 may be to make statements about the perception attributes such as envelopment, plausibility, coloration, externalization only looking at the plurality of physical acoustic parameters 111 , 113.
  • the perceptual acoustic quality model 150 may be designed based on the correlation matrices between the plurality of physical acoustic parameters 111 , 113 and the perception attributes.
  • Figure 7 shows components of the perceptual acoustic quality model 150 implemented by the data processing apparatus 100 according to an embodiment.
  • the task is to interlink the acoustic similarity index with several quality features (regarding spatial audio quality), i.e. the perceptual attributes.
  • subjective evaluation/listening tests may be designed, as already described above, and organized to create a perceptual evaluation database 701 , which may be stored in the memory 103 of the data processing apparatus 100.
  • the data processing apparatus 100 may be further configured to implement a module 703 for supplying the quality features, i.e. the perceptual attributes, and a perceived quality analysis module 705 which may process the quality features, i.e. the perceptual attributes, based on the data from the perceptual evaluation database 701 .
  • the data processing apparatus may perform single parameter and multivariate analysis 303, 307, as described above.
  • SUBSTITUTE SHEET (RULE 26)
  • the corresponding results of the perceived quality analysis module 705 and the single parameter and multivariate analysis 303, 307, which results may comprise the plurality of physical acoustic parameters 111 , 113, the LD components and the determined distances, may be forwarded to an interlinking and correlation module 707.
  • the perceptual acoustic quality model 150 can be aimed to access the overall subjective audio quality as well as some describing attributes. This can be archived, by a combined analysis being performed by the interlinking and correlation module 707 to find the correlations between perceptive evaluations results and physical/acoustic, comprising the physical room acoustic parameters 111 , 113, the LD components and the similarity/distances of the single parameter and multivariate analysis 303, 307.
  • a parameter or distance may have a significant effect with regard to the asked perception in the perceptual evaluation test, if it correlates in all evaluation conditions (evaluated rooms) in a similar, significant way. Coloration and envelopment may be perceived similar for the different BRI Rs in all evaluation rooms. Plausibility perception of the different BRI Rs may differ strongly between the evaluation rooms, so the evaluation depends on if the render room/target room is congruent or divergent, i.e. similar or un-similar. Therefore, it is difficult to design the perceptual acoustic quality model 150 based on fixed acoustic parameters to estimate different quality features.
  • the perceptual acoustic quality model 150 may be designed separately for each relevant quality feature, i.e. perceptual attribute.
  • Figure 8 shows an exemplary correlation analysis for determining the perceptual acoustic quality model 150 implemented by the data processing apparatus 100 according to an embodiment.
  • the data processing apparatus may determine the perceptual acoustic quality model 150 by comparing a plurality of combinations for each of the perceptual attributes.
  • each first row of each of the diagrams 801-803, 811-813, 821-823 may indicate a same first room, each second row a same second room and each third row a same third room.
  • Each column of the diagrams 801 , 811 , 821 may indicate specific parameters of the plurality of physical acoustic parameters 111 , 113, namely from left to right the parameters EDT, T20, T30, Teo, D50, C50, Oso, DRR and T s .
  • Each column of the diagrams 802, 812, 822 may indicate from left to right the linear combinations LD1 to LD8 based on the LDA.
  • Each column of the diagrams 803, 813, 823 may indicate from left to right (i) a distance of LD1 , LD2 and LD3, (ii) a cosine similarity over LD 1 and L2, for example according to the equation 2, and (iii) a cosine similarity over LD 1 and L3, for example according to the equation 2.
  • the data processing apparatus 100 may determine the parameters used in the perceptual acoustic quality model 150, in particular those parameters of the plurality of physical acoustic parameters 111 , 113 showing a high correlation between acoustic evaluation and perceptual evaluation.
  • the data processing apparatus 100 may determine for the parameters used in the perceptual acoustic quality model 150 the cosine similarity over LD1 and LD3 for the perceptual attribute “plausibility”, LD3 for the perceptual attribute “envelopment” and LD1 for the perceptual attribute “coloration”.
  • the room perceptual similarity may then be estimated based on the pre-designed perceptual acoustic quality model 150 comprising the parameters as chosen above.
  • the data processing apparatus 100 can be configured according to one or more of the following modes based on the matrix W and the comparison performed as described above:
  • the perceptual acoustic quality model 150 may define, i.e. comprises for each perceptual attribute a correlation matrix between weighted linear combinations, i.e. the LDA discriminant functions, i.e. LDs determined by means of the linear discriminant analysis, of the plurality of physical acoustic parameters 111 of the target room and the first value or measure of the perceptual attribute of the target room and between weighted linear combinations of the plurality of physical acoustic parameters 113 of the virtual room and the second value or measure of the perceptual attribute of the virtual room.
  • weighted linear combinations i.e. the LDA discriminant functions, i.e. LDs determined by means of the linear discriminant analysis
  • the perceptual attribute of the target room and the virtual room may be a coloration attribute.
  • the weighted linear combination of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 30 dB, i.e. T30, having the largest weight.
  • the weighted linear combination of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 20 dB, i.e.
  • T20 with a weight having an absolute value in the range from about 0.33 to about 0.53, in particular 0.4382, the reverberation time parameter for an energy decay by 30 dB, i.e. T30, with a weight having an absolute value in the range from about 0.9 to about 1.0, in particular 1.0, and a clarity index parameter for 80 ms, i.e. Cso, with a weight having an absolute value in the range from about 0.29 to 0.49, in particular 0.3958.
  • the perceptual attribute of the target room and the virtual room may be an envelopment attribute.
  • the weighted linear combination of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprises a clarity index parameter for 50 ms, i.e. C50, having the largest weight.
  • the weighted linear combination of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 20 dB, i.e.
  • T20 with a weight having an absolute value in the range from about 0.9 to about 1 .00, in particular 0.9533, a reverberation time parameter for an energy decay by 30 dB, i.e. T30, with a weight having an absolute value in the range from about 0.85 to about 0.95, in particular 0.8834, the clarity index parameter for 50 ms, i.e. C50, having an absolute value in the range from about 0.9 to about 1 .0, in particular 1.0, and a clarity index parameter for 80 ms, i.e. Cso, with a weight having an absolute value in the range from about 0.9 to 1.00, in particular 0.9228.
  • the perceptual acoustic quality model 150 may defines, i.e. comprise for each perceptual attribute a correlation matrix between a cosine similarity of a first and a second weighted linear combination of the plurality of physical acoustic parameters 111 of the target room and the first value or measure of the perceptual attribute of the target room and between a cosine similarity of the first and the second weighted linear combination of the plurality of physical acoustic parameters 113 of the virtual room and the second value or measure of the perceptual attribute of the virtual room.
  • the perceptual attribute of the target room and the virtual room may be a plausibility attribute.
  • the first weighted linear combination i.e.
  • LD1 of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 30 dB, i.e. T30, having the largest weight of the first weighted linear combination.
  • the second weighted linear combination, i.e. LD3, of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a clarity index parameter for 50 ms, i.e. C50, having the largest weight of the second weighted linear combination.
  • the first weighted linear combination, i.e. LD1 , of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 20 dB, i.e. T20, with a weight having an absolute value in the range from about 0.33 to about 0.53, in particular 0.4382, the reverberation time parameter for an energy decay by 30 dB, i.e. T30, with a weight having an absolute value in the range from about 0.9 to about 1 .0, in particular 1 .0, and a clarity index parameter for 80 ms, i.e. Cso, with a weight having an absolute value in the range from about 0.29 to 0.49, in particular 0.3958.
  • the second weighted linear combination, i.e. LD3, of the plurality of physical acoustic parameters 111 , 113 of the target room and the virtual room may comprise a reverberation time parameter for an energy decay by 20 dB, i.e. T20, with a weight having an absolute value in the range from about 0.9 to about 1.00, in particular 0.9533, a reverberation time parameter for an energy decay by 30 dB, i.e. T30, with a weight having an absolute value in the range from about 0.85 to about 0.95, in particular 0.8834, the clarity index parameter for 50 ms, i.e.
  • C50 having an absolute value in the range from about 0.9 to about 1.0, in particular 1.0, and a clarity index parameter for 80 ms, i.e. Cso, with a weight having an absolute value in the range from about 0.9 to 1.00, in particular 0.9228.
  • Figure 9 shows a schematic diagram illustrating an implementation of the perceptual acoustic quality model 150 with a visualization module 901 of the data processing apparatus 100 according to an embodiment.
  • the data processing apparatus 100 may be configured to implement the module 240 for determining similarity and comparison of the physical acoustic parameters 111 , 113 of the rooms.
  • the visualization module 901 may receive the acoustic distance values as described above from the module 240 and the perceptual distance 115 from the perceptual acoustic quality model 150.
  • the visualization module 901 may allow the data processing apparatus 100 to present or visualize acoustic and perceptual distances (or similarities) 115 between two or more spatial audio signals using the physical acoustic parameters 111 , 113 of the rooms by the display 105. This can be useful for content creation or spatial audio mixing.
  • Figure 10 shows a graphical user interface 1000 of the display 105 of the data processing apparatus 100 according to an embodiment.
  • the graphical user interface 1000 may be configured to illustrate the perceptual acoustic distance measure value 115 for the selected perceptual attribute between the target room and the virtual room.
  • the perceptual acoustic distance measure value 115 may be illustrated in the form of an arrow in relation to a plurality of further distances 1001a-c to predefined distances for predefined room classes, which may be stored in the memory 103 of the data processing apparatus 100.
  • the plurality of further distances 1001a-c may be graphically illustrated different to the perceptual acoustic distance measure value 115, for example by a circle.
  • the perceptual acoustic distance measure value 115 and the plurality of further distances 1001a-c may be illustrated on different classification quadrants of a visualization diagram 1003 of the graphical user interface 1000.
  • the graphical user interface 1000 may further comprise a display section 1005 for illustrating one or more of the perceptual attributes, i.e. for the estimated perceived spatial audio quality.
  • the one or more of the perceptual attributes may be illustrated in forms of bars, which represent the ratings of the perceptual attributes described above.
  • the graphical user interface 1000 may further comprise a menu 1007 for configuring visualization settings of the rooms, in particular the target room and the virtual room, the room classes, selection of the physical acoustic parameters 111 , 113 and/or selection of the perceptual attributes.
  • the menu 1007 there may be selectable different rooms for the target room and different rooms for the virtual room which can be used in the rendering system.
  • a user of the display 105 may first select the target room which could be a good mixing studio or the real listening room, then click on different virtual rooms. When doing that, the acoustic distance between two rooms is shown based on two components of the single parameter and multivariate analysis, for example two LDA components. Consequently, the estimated ratings for different quality features may be shown on the right side of the graphical user interface 1000 in the display section 1005.
  • Figure 11 is a flow diagram illustrating a computer-implemented data processing method 1100 according to an embodiment for estimating the perceptual acoustics of a target room with respect to a perceptual attribute of the target room.
  • the data processing method 1100 comprises a step 1101 of obtaining a plurality of physical acoustic parameters 111 , 113 for the target room.
  • the data processing method 1100 further comprises a step 1103 of obtaining the plurality of physical acoustic parameters 111 , 113, i.e. the same parameters as for the target room, for a virtual room, e.g. from database 201a.
  • the data processing method 1100 further comprises a step 1105 of estimating a perceptual acoustic distance measure value 115 for the selected perceptual attribute between the target room and the virtual room based on the plurality of physical acoustic parameters 111 of the target room and on the plurality of physical acoustic parameters 113 of the virtual room and based on the perceptual acoustic quality model 150, wherein the perceptual acoustic quality model 150 defines a mapping, in particular correlation, a) between the plurality of physical acoustic parameters 111 , 113 and a first value or measure of the perceptual attribute of the target room and b) between the plurality of physical acoustic parameters 111 , 113 and a second value or measure of the perceptual attribute of the virtual room.
  • the method data processing 1100 can be performed by the data processing apparatus 100 according to an embodiment.
  • further features of the data processing method 1100 result directly from the functionality of the data processing apparatus 100 as well as its different embodiments described above and below.
  • perceptually meaningful parameters i.e. the perceptual acoustic distance measure value 115
  • perceptually meaningful parameters i.e. the perceptual acoustic distance measure value 115
  • Visualizing acoustic distance and perceptual distances (or similarities) between two or more spatial audio signals using the plurality of physical acoustic parameters 111 , 113 can generate knowledge by which it may achieved to adapt the plurality of physical acoustic parameters 111 , 113 and parts of the BRIRs to achieve an externalized and very plausible perceptual impression.
  • the data processing apparatus 100 and the data processing method 1100 can use perceptually meaningful parameters, i.e. the perceptual acoustic distance measure value 115, instead of physical and/or acoustic parameters which can improve the user experience by providing perceptually more similar results.
  • the disclosed system, apparatus, and method may be implemented in other manners.
  • the described embodiment of an apparatus is merely exemplary.
  • the unit division is merely logical function division and may be another division in an actual implementation.
  • a plurality of units or components may be combined or integrated into another system, or some features may be ignored or not performed.
  • the displayed or discussed mutual couplings or direct couplings or communication connections may be implemented by using some interfaces.
  • the indirect couplings or communication connections between the apparatuses or units may be implemented in electronic, mechanical, or other forms.
  • the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
  • functional units in the embodiments of the invention may be integrated into one processing unit, or each of the units may exist alone physically, or two or more units are integrated into one unit.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

L'invention concerne un appareil de traitement de données (100) destiné à estimer l'acoustique perceptuelle d'une salle cible par rapport à un attribut perceptuel de la salle cible. L'appareil de traitement de données (100) est conçu pour obtenir une pluralité de paramètres acoustiques physiques (111) pour la salle cible et obtenir la pluralité de paramètres acoustiques physiques (113) pour une salle virtuelle. L'appareil de traitement de données (100) est en outre conçu pour estimer une distance d'acoustique perceptuelle (115) entre la salle cible et la salle virtuelle sur la base de la pluralité de paramètres acoustiques physiques (111) de la salle cible et sur la pluralité de paramètres acoustiques physiques (113) de la salle virtuelle et sur la base d'un modèle de qualité d'acoustique perceptuelle (150). Le modèle de qualité d'acoustique perceptuelle (150) définit une mise en correspondance entre la pluralité de paramètres acoustiques physiques (111, 113) et une valeur de l'attribut perceptuel de la salle cible et entre la pluralité de paramètres acoustiques physiques (111, 113) et une valeur de l'attribut perceptuel de la salle virtuelle.
PCT/EP2022/087262 2022-12-21 2022-12-21 Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible WO2024132129A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/EP2022/087262 WO2024132129A1 (fr) 2022-12-21 2022-12-21 Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/EP2022/087262 WO2024132129A1 (fr) 2022-12-21 2022-12-21 Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible

Publications (1)

Publication Number Publication Date
WO2024132129A1 true WO2024132129A1 (fr) 2024-06-27

Family

ID=84829623

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2022/087262 WO2024132129A1 (fr) 2022-12-21 2022-12-21 Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible

Country Status (1)

Country Link
WO (1) WO2024132129A1 (fr)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210231488A1 (en) * 2018-09-18 2021-07-29 Huawei Technologies Co., Ltd. Device and method for adaptation of virtual 3d audio to a real room

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210231488A1 (en) * 2018-09-18 2021-07-29 Huawei Technologies Co., Ltd. Device and method for adaptation of virtual 3d audio to a real room

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SPORER THOMAS ET AL: "Adjustment of the direct-to-Reverberant-Energy-Ratio to Reach Externalization within a Binaural Synthesis System", CONFERENCE: 2016 AES INTERNATIONAL CONFERENCE ON AUDIO FOR VIRTUAL AND AUGMENTED REALITY; SEPTEMBER 2016, AES, 60 EAST 42ND STREET, ROOM 2520 NEW YORK 10165-2520, USA, 21 September 2016 (2016-09-21), XP040681043 *
ZAHORIK PAVEL: "Perceptually relevant parameters for virtual listening simulation of small room acoustics", THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, AMERICAN INSTITUTE OF PHYSICS, 2 HUNTINGTON QUADRANGLE, MELVILLE, NY 11747, vol. 126, no. 2, 1 August 2009 (2009-08-01), pages 776 - 791, XP012128619, ISSN: 0001-4966, DOI: 10.1121/1.3167842 *

Similar Documents

Publication Publication Date Title
Brinkmann et al. A round robin on room acoustical simulation and auralization
CN109644314B (zh) 渲染声音程序的方法、音频回放系统和制造制品
CN111107482B (zh) 修改房间特性以通过耳机进行空间音频呈现的系统和方法
US20060274901A1 (en) Audio image control device and design tool and audio image control device
US20080137870A1 (en) Method And Device For Individualizing Hrtfs By Modeling
CN111294724B (zh) 多个音频流的空间重新定位
CN112740324B (zh) 用于使虚拟3d音频适应现实房间的设备和方法
Postma et al. The influence of visual distance on the room-acoustic experience of auralizations
CN104240695A (zh) 一种优化的基于耳机重放的虚拟声合成方法
Engel et al. Perceptual implications of different Ambisonics-based methods for binaural reverberation
US11418903B2 (en) Spatial repositioning of multiple audio streams
US7116788B1 (en) Efficient head related transfer function filter generation
JP2022515266A (ja) 深層学習画像解析を用いた室内音響シミュレーション
Treybig et al. Room acoustic analysis and BRIR matching based on room acoustic measurements
US11678111B1 (en) Deep-learning based beam forming synthesis for spatial audio
Grimm et al. Virtual acoustic environments for comprehensive evaluation of model-based hearing devices
Pike et al. Descriptive analysis of binaural rendering with virtual loudspeakers using a rate-all-that-apply approach
WO2024132129A1 (fr) Appareil et procédé d'estimation de l'acoustique perceptuelle d'une salle cible
GB2612173A (en) Determining a virtual listening environment
de Taillez et al. Acoustic and perceptual effects of magnifying interaural difference cues in a simulated “binaural” hearing aid
Alonso-Martınez Improving Binaural Audio Techniques for Augmented Reality
Michaud et al. Objective characterization of perceptual dimensions underlying the sound reproduction of 37 single loudspeakers in a room
Li et al. The influence of acoustic cues in early reflections on source localization
Surdu et al. A. LI. EN: An Audiovisual Dataset of different Acoustical Impulse Responses Measured in a Living Room Environment
JP2019184933A (ja) マルチチャンネル客観評価装置及びプログラム