NL2024434B1 - Generating an audio signal associated with a virtual sound source - Google Patents

Generating an audio signal associated with a virtual sound source Download PDF

Info

Publication number
NL2024434B1
NL2024434B1 NL2024434A NL2024434A NL2024434B1 NL 2024434 B1 NL2024434 B1 NL 2024434B1 NL 2024434 A NL2024434 A NL 2024434A NL 2024434 A NL2024434 A NL 2024434A NL 2024434 B1 NL2024434 B1 NL 2024434B1
Authority
NL
Netherlands
Prior art keywords
audio signal
sound source
signal
virtual
virtual sound
Prior art date
Application number
NL2024434A
Other languages
Dutch (nl)
Inventor
Oomen Paulus
Original Assignee
Liquid Oxigen Lox B V
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Liquid Oxigen Lox B V filed Critical Liquid Oxigen Lox B V
Priority to NL2024434A priority Critical patent/NL2024434B1/en
Priority to US17/784,466 priority patent/US20230017323A1/en
Priority to CN202080093387.3A priority patent/CN114946199A/en
Priority to JP2022536511A priority patent/JP2023506240A/en
Priority to CA3164476A priority patent/CA3164476A1/en
Priority to PCT/NL2020/050774 priority patent/WO2021118352A1/en
Priority to EP20829377.9A priority patent/EP4074078A1/en
Application granted granted Critical
Publication of NL2024434B1 publication Critical patent/NL2024434B1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/08Arrangements for producing a reverberation or echo sound
    • G10K15/12Arrangements for producing a reverberation or echo sound using electronic time-delay networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/11Positioning of individual sound objects, e.g. moving airplane, within a sound field

Abstract

A method for generating an audio signal associated with a virtual sound source is disclosed. The method comprises 5 obtaining an input audio signal X(t) and modifying the input audio signal X(t) to obtain a modified audio signal. The latter step comprises performing a signal delay operation. Optionally, modifying the input audio signal comprises a signal inverting operation and/or a signal amplification or 10 attenuation and/or a signal feedback operation. The method further comprises generating the audio signal y(t) based on a combination, e.g. a summation, of the input audio signal X(t) and the modified audio signal.

Description

NL28743-LD/av Generating an audio signal associated with a virtual sound source
FIELD OF THE INVENTION This disclosure relates to a method and system for generating an audio signal associated with a virtual sound source. In particular to such method and system wherein an input audio signal x(t) is modified to obtain a modified audio signal and wherein the modification comprises performing a signal delay operation. The audio signal y(t) is generated based on a combination, e.g. a summation, of the input audio signal x(t) and the modified audio signal.
BACKGROUND In the playback of sound through audio transmitters, i.e. loudspeakers, much of the inherent spatial information of the (recorded) sound is lost. Therefore, the experience of sound through speakers is often felt to lack depth (it sounds ‘flat’) and dimensionality (it sounds ‘in-the-box’). The active perception of height is altogether missing from the sound experience across the speakers. These conditions create an inherent detachment between the listener and sound in the environment. This creates an obstacle for the observer to fully identify physically and emotionally with the sound environment and in general this makes sound experiences more passive and less engaging. A classical demonstration of this problem is described by Von Bekésy’s (Experiments in Hearing, 1960): the ‘in-the-box’ sound effect seems to increase with the decrease of the loudspeaker’s dimensions. In an experimental research on the relation between acoustic power, spectral balance and perceived spatial dimensions and loudness, Von Bekésy’'s test subjects were unable to correctly indicate the relative dimensional shape of a reproduced sound source as soon as the source’s dimensions exceeded the actual shape of the reproducing loudspeaker box. One may conclude that the loudspeaker’ s spatio-spectral properties introduce a message- media conflict when transmitting sound information. We cannot recognize the spatial dimensions of the sound source in the reproduced sound. Instead, we listen to the properties of the loudspeaker.
In the prior art there is no satisfying approach to record or compute dimensional information of sound sources. The near- field information of sound producing objects cannot be accurately captured by microphones, or would theoretically require an infinite grid of pressure and particle velocity transducers to capture the dimensional information of the object.
For a computational simulation of dimensional information, solutions to the wave equation are only applicable to a limited amount of basic geometrical shapes and for a limited frequency range. Given the lack of an analytical solution to the problem, simulation models have to resort to finite computation methods to attempt to reproduce the desired data. The data gathered in this way and reproduced by means of techniques involving FFT (Fast Fourier Transform), such as convolution or additive synthesis, require complex calculations and very large amounts of data processing and are thus inherently very intensive for computer processing. This limits the application of such methods and poses a problem for the audio playback system that can accurately reproduce the information.
Hence, there is a need in the art for a method for generating audio signals asscciated with a virtual sound source that are less computationally expensive.
SUMMARY To that end, a method for generating an audio signal associated with a virtual sound source is disclosed. The method comprises obtaining an input audio signal x(t) and modifying the input audio signal x(t) to obtain a modified audio signal. The latter step comprises performing a signal delay operation. Optionally, modifying the input audio signal comprises a signal inverting operation and/or a signal amplification or attenuation and/or a signal feedback operation. The method further comprises generating the audio signal y(t) based on a combination, e.g. a summation, of the input audio signal x(t) and the modified audio signal.
When a virtual sound source is said to have a particular size and shape and/or to be positioned at a particular distance and/or to be positioned at a particular height or depth it may be understood as that an observer, when hearing the generated audio signal, perceives the audio signal as originating from a sound source having that particular size and shape and/or being positioned at said particular distance and/or at said particular height or depth. The human hearing is very sensitive, as also illustrated by the Von Bekésy experiment described above, to spectral information that correlates with the dimensions of the object producing the sound. The human hearing recognizes the features of a sounding object primarily by its resonance, i.e. the amplification of one or several fundamental frequencies and their correlating higher harmonics, such amplification resulting from standing waves that occur inside the object or space due to its particular size and shape. By adding and subtracting spectral information from the audio signal in such a way that its resulting spectrum will closely resemble the resonance of the intended object or space, one can at least partially overrule the spatio-spectral properties of the loudspeaker (s) and create a coherent spatial projection of the sound signal by means of its size and shape.
The applicant has realized that such spatial information, related to the dimensions of a sound source and its virtual distance, height and depth in relation to an observer, can be added to an audio signal by performing relatively simple operations onto an input audio signal.
In particular, the applicant has found that these simple operations are sufficient for generating an audio signal having properties such that the physiology of the human hearing apparatus causes an observer to perceive the audio signal as coming from a sound source having a certain position and dimensions, other than the position and dimensions of the loudspeakers that produce the sound.
The above-described method does not require filtering or synthesizing individual (bands of) frequencies and amplitudes to add this spatial information to the input audio signal.
The method thus bypasses the need for FFT synthesis techniques for such purpose, in this way simplifying the process and considerably reducing the processing power required.
Optionally, the method comprises playing back the generated audio signal, e.g. by providing the generated audio signal to one or more loudspeakers in order to have the generated audio signal played back by the one or more loudspeakers.
The generated audio signal, once played out by a loudspeaker system, causes the desired perception by an observer irrespective of how many loudspeakers are used and irrespective of the position of the observer relative to the loudspeakers.
A signal that is said to have been generated based on a combination of two or more signals may be the combination, e.g. the summation, of these two or more signals.
In an example, the generated audio signal is stored onto a 5 computer readable medium so that it can be played out at a later time by a loudspeaker system.
The audio signal can be generated in real-time, which may be understood as that the audio signal is generated immediately as the input audio signal comes in and/or may be understood as that any variation in the input audio signal at a particular time is reflected in the generated audio signal within three seconds, preferably within 0.5 seconds, more preferably within 50 ms, most preferably within 10 ms. The relatively simple operations for generating the audio signal allows for such real-time processing. Optionally, the generated audio signal is played back in real-time, which may be understood as that the audio signal, once generated, is played back without substantial delay.
In an embodiment, the virtual sound source has a shape.
Such embodiment comprises generating audio signal components associated with respective virtual points on the virtual sound source's shape. This step comprises generating a first audio signal component associated with a first virtual point on the virtual sound source’s shape and a second audio signal component associated with a second virtual point on the virtual sound source’s shape. Herein, generating the first audio signal component comprises modifying the input audio signal to obtain a modified first audio signal component using a first signal delay operation introducing a first time delay and comprises generating the first audio signal component based on a combination, e.g. a summation, of the input audio signal and the modified first audio signal component. Further, generating the second audio signal component comprises modifying the input audio signal to obtain a modified second audio signal component using a second signal delay operation introducing a second time delay different from the first time delay and comprises generating the second audio signal component based on a combination, e.g. a summation, of the input audio signal and the modified second audio signal component.
The applicant has found out that this embodiment allows to add the dimensional information of the virtual sound source to the input audio signal x(t) in a simple manner, without requiring complex algorithms, such as FFT algorithms, additive synthesis of individual frequency bands or multitudes of bandpass filters to obtain the desired result, as has been the case in the prior art.
Preferably, many more than two virtual points may be defined on the virtual sound source’s shape. An arbitrary number of virtual points may be defined on the shape of the virtual sound source. For each of these virtual points, an audio signal component may be determined. Each determination of audio signal component may then comprise determining a modified audio signal component using a signal delay operation introducing a respective time delay. Each audio signal component may then be determined based on a combination, e.g. a summation, of its modified audio signal component and the input audio signal.
Each determination of a modified audio signal component may further comprise performing a signal inverting operation and/or a signal amplification or attenuation and/or a signal feedback operation. Herein, preferably, the signal feedback operation is performed last. In principle, the signal inverting operation, amplification/attenuation and signal delay operation may be performed in any order.
The virtual points may be positioned equidistant from each other on the shape of the virtual sound source. Further, the virtual sound source may have any shape, such as a one- dimensional shape, e.g. a 1D string, a two-dimensional shape, e.g. a 2D plate shape, or a three-dimensional shape, e.g. a 3D cube.
The time period with which an audio signal is delayed may be zero for some audio signal components. To illustrate, if the virtual sound source is a string, the time delay for the two virtual points at the respective ends of the string where its vibration is restricted, may be zero. This will be illustrated below with reference to the figures.
In an embodiment, the method comprises obtaining shape data representing the virtual positions of the respective virtual points on the virtual sound source’s shape and determining the first resp. second time delay based on the virtual position of the first resp. second virtual point. Thus, the respective time delays for determining the respective audio signal components for the different virtual points may be determined based on the respective virtual positions of these virtual points.
The applicant has found out that this embodiment enables to take into account how sound waves propagates through a dimensional shape, which enables to accurately generate audio signals that are perceived by an observer to originate from a sound source having that particular shape. When generated audio signal components associated with the virtual points are played back through a loudspeaker, or distributed across multiple loudspeakers, the result is perceived as one coherent sound source in space because the signal components strengthen their coherence at corresponding wavelengths in harmonic ratios according to the fundamental resonance frequencies of the virtual shape. This at least partially overrules the mechanism of the ear to detect its actual output components, i.e. the loudspeaker (s).
Preferably, the time period for each time delayed version of the audio input signal is determined following a relationship between spatial dimensions and time, examples of which are given below in the figure descriptions.
In an embodiment, the to be generated audio signal y(t) is associated with a virtual sound source having a distance from an observer. This embodiment comprises (i) modifying the input audio signal using a time delay operation introducing a time delay and a signal feedback operation to obtain a first modified audio signal, and (ii) generating a second modified audio signal based on a combination of the input audio signal x(t) and the first modified audio signal; and (iii) generating the audio signal y(t) based on the second modified audio signal, this step comprising attenuating the second modified audio signal and optionally comprising performing a time delay operation introducing a second time delay.
The human hearing recognizes a sound source distance detecting primarily the changes in the overall intensity of the auditory stimulus and the proportionally faster dissipation of energy from the high to the lower frequencies. The applicant has found out that this embodiment allows to add such distance information to the input audio signal in a very simple and computationally inexpensive manner.
The second introduced time delay may be used to cause a Doppler effect for the observer. This embodiment further allows to control a Q-factor, which narrows or widens the bandwidth of the resonant frequencies in the signal. In this case, since the perceived resonant frequency is infinitely low at the furthest possible virtual distance, the Q-factor influences the steepness of a curve covering the entire audible frequency range from high to the low frequencies,
resulting in the intended gradual increase of high-frequency dissipation in the signal.
Preferably, the time delay introduced by the time delay operation that is performed to obtain the first modified audio signal is shorter than 0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds, most preferably approximately 0.00001 seconds.
The second modified audio signal may be attenuated in dependence of the distance of the virtual sound source. For the signal feedback operation that is performed in order to determine the first modified audio signal, in which an attenuated version of a signal is recursively added to itself, the signal attenuation is preferably also performed in dependence of said distance. Optionally, such embodiment comprises obtaining distance data representing the distance of the virtual sound source so that the attenuation can be automatically appropriately controlled. This embodiment allows to “move” the virtual sound source towards and away from an observer by simply adjusting a few values.
In the above embodiment, the signal feedback operation comprises attenuating a signal, e.g. the signal as obtained after performing the time delay operation introducing said time delay, and recursively adding the attenuated signal to the signal itself. Such embodiment may further comprise controlling the degree of attenuation in the signal feedback operation and the degree of attenuation of the second modified audio signal in dependence of said distance, such that the larger the distance is, the lower the degree of attenuation in the signal feedback operation and the higher the degree of attenuation of the second modified audio signal.
In an embodiment, the to be generated audio signal y(t) associated with a virtual sound source is positioned at a virtual height above an observer. In such embodiment, the method comprises (i) modifying the input audio signal x{t) using a signal inverting operation, a signal attenuation operation and a time delay operation introducing a time delay in order to obtain a third modified audio signal, and (ii) generating the audio signal based on a combination, e.g. a summation, of the input audio signal and the third modified audio signal.
The applicant has found out that this embodiment allows to, in a simple manner, generate audio signals that come from a virtual sound source positioned at a certain height.
In this embodiment, the introduced time delay is preferably shorter than 0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds, most preferably approximately 0.00001 seconds.
In the above embodiment, modifying the input audio signal to obtain the third modified audio signal optionally comprises performing a signal feedback operation. In a particular example, this step comprises recursively adding an attenuated version of a signal, e.g. the signal resulting from the time delay operation, signal attenuation operation and signal inverting operation that are performed to eventually obtain the third modified audio signal, to itself.
In an embodiment, the to be generated audio signal is associated with a virtual sound source that is positioned at a virtual depth below an observer. Such embodiment comprises modifying the input audio signal x(f) using a time delay operation introducing a time delay, a signal attenuation operation and a signal feedback operation in order to obtain a sixth modified audio signal. Performing the signal feedback operation e.g. comprises recursively adding an attenuated version of a signal, e.g. the signal resulting from the time delay operation and signal attenuation operation that are performed to eventually obtain the sixth modified audio signal, to itself.
This embodiment further comprises generating the audio signal based on a combination of the input audio signal and the sixth modified audic signal.
In this embodiment, the introduced time delay is preferably shorter than 0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds, most preferably approximately 0.00001 seconds.
In an embodiment, the method comprises receiving a user input indicative of the virtual sound source’s shape and/or indicative of respective virtual positions of virtual points on the virtual sound source’s shape and/or indicative of the distance between the virtual sound source and the observer and/or indicative of the height at which the virtual sound source is positioned above the observer and/or indicative of the depth at which the virtual sound source is positioned below the observer.
This embodiment allows a user to input parameters relating to the virtual sound source, which allows to generate the audio signal in accordance with these parameters.
This embodiment may comprise determining values of parameters as described herein and using these determined parameters to generated the audio signal.
In an embodiment, the method comprises generating a user interface enabling a user to input at least one of: -the virtual sound source’s shape, -respective virtual positions of virtual points on the virtual sound source’s shape, -the distance between the virtual sound source and the observer, -the height at which the virtual sound source is positioned above the observer, -the depth at which the virtual sound source is positioned below the observer.
This allows a user to easily input parameters relating to the virtual sound source and as such allows a user to easily control virtual sound source.
The methods as described herein may be computer-implemented methods.
One aspect of this disclosure relates to a computer comprising a computer readable storage medium having computer readable program code embodied therewith, and a processor, preferably a microprocessor, coupled to the computer readable storage medium, wherein responsive to executing the computer readable program code, the processor is configured to perform one or more of the method steps as described herein for generating an audio signal associated with a virtual sound source.
Cne aspect of this disclosure relates to a computer program or suite of computer programs comprising at least one software code portion or a computer program product storing at least one software code portion, the software code portion, when run on a computer system, being configured for executing one or more of the method steps as described herein for generating an audio signal associated with a virtual sound source.
One aspect of this disclosure relates to a computer non- transitory computer-readable storage medium storing at least one software code portion, the software code portion, when executed or processed by a computer, is configured to perform one or more of the method steps as described herein for generating an audio signal associated with a virtual sound source.
One aspect of this disclosure relates to a user interface as described herein.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a "circuit," "module" or "system”. Functions described in this disclosure may be implemented as an algorithm executed by a microprocessor of a computer. Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable mediums) having computer readable program code embodied, e.g., stored, thereon.
Any combination of one or more computer readable mediums) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non- exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber, cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including a functional or an object oriented programming language such as Java(TM), Scala, C++, Python or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages.
The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer, server or virtualized server.
In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention.
It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor, in particular a microprocessor or central processing unit (CPU), or graphics processing unit (GPU), of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer, other programmable data processing apparatus, or other devices create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to varicus embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The invention will be further illustrated with reference to the attached drawings, which schematically will show embodiments according to the invention. It will be understood that the invention is not in any way restricted to these specific embodiments.
BRIEF DESCRIPTION OF THE DRAWINGS Aspects of the invention will be explained in greater detail by reference to exemplary embodiments shown in the drawings, in which: FIG. 1 illustrates a method and system according to an embodiment; FIG. 2 shows spectrograms of audio signals generated using a method and/or system according to an embodiment; FIG. 3A shows an virtual sound source according to an embodiment, in particular a virtual sound source shape as a string;
FIG. 3B schematically shows the input audio signal and signal inverted, time-delayed versions of the input audio signal that may be involved in embodiments;
FIG. 4 illustrates a method for adding dimensional information to the audio signal, the dimensional information relating to a shape of the virtual sound source;
FIG. 5 illustrates a panning system that may be used in an embodiment;
FIG. 6A illustrates two-dimensional and three-dimensional virtual sound sources;
FIG. 6B shows an input signal and time-delayed version of this signal which may be involved in embodiments;
FIG. 7A illustrates a method for generating an audio signal associated with a two-dimensional virtual sound source, such as a plate;
FIG. 7B schematically shows how several parameters may be determined that are used in an embodiment;
FIG. 8A and 8B show spectrograms of respective audio signal components associated with respective virtual points on a virtual sound source;
FIG. 9A and 9B illustrate the generation of a virtual sound source that is positioned at a distance from an observer according to an embodiment;
FIG. 10 shows spectrograms associated with a virtual sound source that is positioned at respective distances;
FIG. 11A and 11B illustrate the generation of a virtual sound source that is positioned at a height above the observer according to an embodiment;
FIG. 12 shows spectrograms associated with a virtual sound source that is positioned at respective heights;
FIG. 13A and 13B illustrate the generation of a virtual sound source that is positioned at a depth below the observer according to an embodiment;
FIG. 14 illustrates the generation of an audio signal associated with a virtual sound source having a certain shape, positioned at a certain position.
FIG. 15 illustrates a user interface according to an embodiment; FIG. 16 illustrates a data processing system according to an embodiment.
DETAILED DESCRIPTION OF THE DRAWINGS Sound waves inherently carry detailed information about the environment, and about the observer of sound within the environment. This disclosure describes a soundwave transformation (spatial wave transform, or SWT), a method for generating an audio signal, that is perceived to have spatially coherent properties with regards to the dimensional size and shape of the reproduced sound source, its relative distance towards the observer, its height or depth above or below the observer and its directionality if the source is moving towards or away from the observer.
Typically, the spatial wave transform is an algorithm executed by a computer with as input a digital audio signal (e.g. a digital recording) and as output one or multiple modified audio signal (s) which can be played back on conventional audio playback systems. Alternatively, the transform could also apply to analogue (non-digital) means of generating and/or processing audio signal(s). Playing back the modified sound signal(s) will give the observer an improved perception of dimensional size and shape of the reproduced sound source (f.i. a recorded signal of a violin will sound as if the violin is physically present) and the sound source’s spatial distance, height and depth in relation to the observer (£.i. the violin sounds at distinctive distance from the listener, and height above or depth below), while masking the physical properties of the sound output medium, i.e. the loudspeaker (s) (that is, the violin does not sound as if it is coming from a speaker). Fig. 1 is a flow chart depicting a method and/or system according to an embodiment.
An input audio signal x(t) is obtained.
The input audio signal x(t) may be analog or digital.
Thus, the operations that are shown in figure 1, i.e. each of the operations 4, 6, 8, 10, 12, 14, may be performed by an analog circuit component or a digital circuit component.
The flow chart of figure 1 may also be understood to depict method steps that can be performed by a computer executing appropriate software code.
The input audio signal x{t) may have been output by a recording process in which sounds have been recorded and optionally converted into a digital signal.
In an example, a musical instrument, such as a violin, has been recorded in a studio to obtain the audio signal that is input for the method for generating the audio signal as described herein.
The input audio signal x(t) is subsequently modified to obtain a modified audio signal.
The signal modification comprises a signal delay operation 4 and/or a signal inverting operation 6 and/or a signal amplification or attenuation 8 and/or a signal feedback operation 10,12. The signal delay operation 4 may be performed using well- known components, such as a delay line.
The signal inverting operation 6 may be understood as inverting a signal such that an input signal x(t) is converted into -x(t). The amplification or attenuation 8 may be a linear amplification or attenuation, which may be understood as amplifying or attenuating a signal by a constant factor a, such that a signal x{t) is converted into a * x(t). The signal feedback operation may be understood to comprise recursively combining a signal with an attenuated version of itself. This is schematically depicted by the attenuation operation 12 that sits in the feedback loop and the combining operation 10. Decreasing the attenuation, i.e. enlarging constant b in figure 1, may increase the peak intensity and narrow the bandwidth of resonance frequencies in the spectrum of the sound, the so-called Q-factor. Herewith, the response of different materials to vibrations can be simulated based on their density and stiffness. For instance, the response of a metal object will generate a higher Q-factor than an object of the same size and shape made out of wood.
The combining operations 10 and 14 may be understood to combine two or more signals {xi(t), ..., Xr(t)}. The input signals may be converted into a signal y(t) as follows.
In figure 1, the audio signal y(t) is generated based on a combination, e.g. a summation, of the input audio signal x(t) and the modified audio signal. In an example, the audio signal y(t) is the result of combining, e.g. summing, the input audio signal x(t) and the modified audio signal.
The transformation of the input audio signal x(t) to the audio signal y(t) may be referred to hereinafter as the Spatial Wave Transform (SWT).
The method for generating the audio signal y(t} does not require finite computational methods, such as methods involving Fast Fourier Transforms, which may limit the achievable resolution of the generated audio signal. Thus, the method disclosed herein enables to form high-resolution audio signals. Herein, high-resolution may be understood as a signal with spectral modifications for an infinite amount of frequency components. The virtually infinite resolution is achieved because the desired spectral information does not need to be computed and modified for each individual frequency component, as would be the case in convolution or simulation models, but the desired spectral modification of frequency components results from the simple summation, i.e. wave interference of two identical audio signals with a specific time delay, amplitude and/or phase difference.
This operation results in phase and amplitude differences for each frequency component in harmonic ratios, i.e. corresponding to the spectral patterns caused by rescnance.
The time delays relevant to the method are typically between 0.00001 - 0.02 seconds, but not excluding longer times.
The generated audio signal y(t) may be presented to an observer through a conventional audic output medium, e.g. one or more loudspeakers.
The generated audio signal may be delayed in time and/or attenuated before being output to the audio output medium.
Figure 2 (top) shows the spectrogram of the generated audio signal when the input audio signal x(t) is white noise, the introduced time delay by the time delay operation 4 is ~0.00001 sec, the signal inverting operation 6 is performed and the signal feedback operation 10,12 is not performed.
Figure 2 (middle) shows the spectrogram of the generated audio signal when the input audio signal x(t) is white noise, the introduced time delay by the time delay operation 4 is ~0.00036 sec, the signal inverting operation 6 is performed and the signal feedback operation 10,12 is not performed.
Figure 2 (bottom) shows the spectrogram of the generated audio signal when the input audio signal x(t) is white noise, the introduced time delay by the time delay operation 4 is ~0.00073 sec, the signal inverting operation 6 is performed and the signal feedback operation 10,12 is not performed.
These figures show that the spectrum of an audio signal can be modified precisely according to harmonic ratios, using a very simple operation.
Figure 3A illustrates a virtual sound source in the form of a string.
A number of virtual points n have been defined on the string’s shape, in this example 17 virtual points. The points may be equidistant from each other as shown. The regular distance chosen between each two particles determines the resolution with which the virtual sound source is defined.
Figures 4 and 7 illustrate embodiments of the method and/or system that may be used to generate an audio signal that is perceived to originate from a sound source having a particular shape, e.g. the string shape as shown in figure 3A, the plate- shaped source or cubic source illustrated in figure 6. In these embodiment, the method comprises generating audio signal components vy,(t) associated with respective virtual points on the virtual sound source’s shape. Generating each audio signal component y(t) comprises modifying the input audio signal to obtain a modified audio signal component using a signal delay operation introducing a time delay At,, Then, each audio signal component y(t} is generated based on a combination, e.g. a summation, of the input audio signal and its modified audio signal component. Preferably, the amplitude of each signal component resulting from said combination is attenuated, e.d.
with -6 dB, by signal attenuating elements 19; - 19,. At least two of the time delays that are introduced differ from each other. The audio signal components y.(ft) together may be understood to constitute the generated audio signal y(t). In an example, the audio signal components are combined to generate the audio signal. However, in another example, these audio signal components are individually fed to a panning system that distributes each component individually to a plurality of loudspeakers. When the audio signal components are played back simultaneously through an audio output medium, e.g. through one or more loudspeakers, the resulting audio signal will be perceived by an observer as originating from a sound source having the particular shape.
Figure 4 in particular illustrates an embodiment for generating an audio signal that is perceived to originate from a sound source that is shaped as a string, e.g. the string shown in figure 3A. Thus, referring to figure 3A, generated audio signal component y;(t) is associated with point n=1, audio signal component y:{t) with point n=2, et cetera. In this embodiment, each modification to the input audio signal not only comprises the introduction of a time delay At, but also inverting the audio input signal as indicated by signal inverting operations 16; - 16,, in order to obtain a modified audio signal component. The modified audio signal components are inverted with respect to the input audio signal, in the case of a sounding object that cannot freely vibrate on its edges, such as is the case with a string under tension, or the skin of a drum. In case of a sounding object that freely vibrates on all its edges, none of the modified audio signal components are inverted, and preferably a high-pass filter is added to the resulting signal component vy,(t) to attenuate the low frequencies of the audio signal as will be explained with reference to figure 7.
Optionally, the modification also comprises a signal feedback operation 18; - 18,, but this is not required for adding the dimensional information of the virtual sound source to the audio signal. The depicted embodiment shows that each audio signal component y,(ft) may be the result of a summation of the input audio signal x(t) and the inverted, time-delayed input audio signal. While figure 4 shows that the time delay operation is performed prior to the signal inverting operation 16, this may be the other way around.
For a string shaped virtual sound source of 1 meter long, the time differences for 17 equidistant positioned virtual points on the string may be as follows:
At (s)
0.00000
0.00036
0.00073
0.00109
0.00146 ee
0.00219
EE ee 0.002%
0.00255
0.00219
0.00182
0.00146
0.00105
0.00075
0.00036
0.00000 These values for the introduced time delays are in accordance with At,=Lx,/v, wherein L indicates the length of the string, wherein x, denotes for virtual point n a multiplication factor and v relates to the speed of sound through a medium. For the values in the table, a value of 343 m/s was used, which is the velocity of soundwaves moving through air at 20 degrees Celsius. A virtual point may be understood to be positioned on a line segment that runs from the center of the virtual sound source, e.g. the center of a string, plate or cube to an edge of the virtual sound source. As such, the virtual point may be understood to divide the line segment in two parts, namely a first part of the line segment that runs between an end of the virtual sound source and the virtual point and a second part of the line segment that runs between the virtual point and the center of the virtual sound source. The multiplication factor may be equal to the ratio between the length of the line segment’s first part and the length of the line segment’s second part. Accordingly, if the virtual point is positioned at an end of the sound source, the multiplication factor is zero and if the virtual point is positioned at the center of the virtual sound source, the multiplication factor is one. Thus, with these values, a user will perceive the generated audio signal as originating from a string-shaped sound source that is one meter in length, whereas the loudspeakers need not be spatially arranged in a particular manner.
In an embodiment, the method comprises obtaining shape data representing the virtual positions of the respective virtual points on the virtual sound source’s shape and determining the time delays that are to be introduced by the respective time delay operations based on the virtual positions of the respective virtual points, preferably in accordance with the above described formula.
Figure 3B schematically shows modified audio signal components 22, 223 and 224 for points n=1, 2, 3 respectively. These audio signal components have been inverted with respect to the audio input signal 20 and time delayed by At;, Ats3, Ata respectively.
Figure 5 shows that the generated audio signal, or the generated audio signal components together forming the generated audio signal can be panned to one or more loudspeakers. This panning step may be performed using methods known in the art. In principle, with the method disclosed herein, the spatial information regarding dimensions, distance, height and depth of the virtual sound source can be added to an audio signal irrespective of the panning method and irrespective of how many loudspeakers are used to playback the audio signal.
In an embodiment, each of the generated audio signal components may in principle be fed to all loudspeakers that are present. However, depending on the panning method that is used, some of the audio signal components may be fed to a loudspeaker with zero amplification. Herewith, effectively,
such loudspeaker does not receive such audio signal component. This is depicted in figure 5 for yl in relation to loudspeaker C and D, for y2 in relation to loudspeakers A and D, and for y3 in relation to loudspeaker A. Typically, a panning system will provide the audio signal components to the loudspeakers with a discrete amplification of each audio signal component to each loudspeaker between zero and one.
Fig. 6A depicts further example of virtual sound sources in order to illustrate that the method may be used for virtual sound sources having a more complex shape. The generated audio signal y(t) may for example be perceived as originating from a plate-shaped sound source 24 or a cubic-shaped sound source
26. Virtual points are defined onto the shape of the virtual sound source. A total of twenty-five virtual points have been defined on the plate shape of source 24 in the depicted example.
The virtual sound source may be shaped as a set of regular polygons; as well as shapes that are non-symmetrical, irregular or organically formed.
Figure 6B illustrates a number of modified audio signal components that may be used when the virtual sound source has a two-dimensional or three-dimensional shape. The figure shows that all modified audio signal components may be time delayed, and none of the modified audio signal components are inverted with respect to the input audio signal, in accordance with a virtual sound source that freely vibrates on all its edges.
Figure 7A is a flowchart illustrating an embodiment in which the generated audio signal y(t) is perceived by an observer to originate from a sound source that is shaped as a plate. Again, a plurality of audio signal components vy,(t) is determined respectively associated with virtual points that are defined on the shape. In this embodiment, each determination of an audio signal component vy,(t) comprises modifying the input audio signal using a signal delay operation introducing a time delay At,.; optionally using a signal feedback operation 30 in order to obtain a modified audio signal component. Subsequently, a second modified audio signal component is generated based on a combination 32 of the input audio signal and the modified audio signal component. The second modified audio signal component may be attenuated, e.g. with approximately -6 dB (see attenuating elements 34). The second modified audio signal component may be modified using a signal delay operation At,.; introducing a second time delay and optionally a signal feedback operation 36 to obtain a third modified audio signal component. Then, the audio signal component v(t) may be generated based on a combination 38 of the second and third modified audio signal component.
Optionally, this step of generating the audio signal component vn (t) comprises performing an attenuation operation 40, e.d. with -6dB, and/or a high pass filter operation 42 that applies a cut off frequency of fx, which may be understood to attenuate frequencies below the lowest fundamental frequency occurring in the plate.
In this embodiment, determining an audio signal component comprises determining a first modified audio signal component and a third modified audio signal component. Determining the first resp. third modified audio signal component may comprise using a first resp. second time delay operation and a signal inverting operation and, optionally, a first resp. second signal feedback operation.
In this example, two combinations 32 and 38 are performed per audio signal component, however, for more complex shaped virtual sound sources, such as three dimensionally shaped sources, three or even more combination operations are performed per audio signal component. An examples of this is shown in figure 14.
Figure 7B illustrates how for each virtual point on a virtual sound source 50 that is shaped as a square plate, the associated time delays and cut-off frequency can be calculated.
As an example, figure 7B illustrates how the time delays and cut-off frequency is calculated for point n=7 on the virtual sound source 50 shaped as a plate.
A first step comprises determining, for each virtual point, three values for the above mentioned multiplication factor x, viz.
Xa, Xp, Xc in accordance with the following formulas: 12 x= (1-224) /3; xp = 22, xc = (1-2) /6 for ME < 0.5; xc = (1-12€) /2 for me >0.5. Herein R denotes the radius of a circle 52 passing through the vertices where two or more edges of the virtual sound source 50 meet.
In this example, R is the radius of the circumscribed circle 52 of the square plate 50. Further, T‚.a denotes (see left illustration in figure 7B) the radius of a circle 56 passing through the vertices of a square 54, wherein the square 54 is a square having a mid point that coincides with the mid point of the virtual sound source 50 and has point n, point 7 in this example, at one of its sides.
The sides of square 54 are parallel to the edges of the plate 50. r..g denotes {see middle illustration in figure 7B) the radius of a circle 60 passing through the vertices of a square 58, wherein the square 58 has a mid point that coincides with vertex that is nearest to point n and has sides that are parallel to the edges of the virtual plate sound source 50. Fre denotes {see right hand side illustration in figure 7B) the smallest distance between the mid point of the plate 50 and an edge of square 62, wherein square 62 has a mid point that coincides with the mid point of the virtual sound source 50 and has point n on one of its sides. Further, square 62 has a side that is perpendicular to at least one diagonal of the plate A. Since the virtual sound source in this example is square, square 62 1s tilted 45 degrees with respect to the plate 50.
In a next step, the associated time delays Ata, Ats, Atc are determined in accordance with At=Ax/v, wherein Atg is only determined if xz is equal to or smaller than 0.25. Accordingly, for a square plate having 25 cm long edges and 25 virtual points as shown in figures 6A and 7B, and v=500m/s, the values for Xa, Xp, Xc and Ata, Atty, Atc are as follows. Ee lo Jo | zes [0.125 Jo [0.003125 [0.00156 5 [0 [1 Jo.0e33 Jo [- TET Tio [0.25 [0.125 [0 [0.003125 | 0.00156 so lo lo To To To | 6 [0 [0.25 [0.125 [0 [as
0.0833 [0.003125] 0.003125 | 0.00104 Ee [1 [0.125 [0.003125] - | 0.00156 49 Jo.25 0.25 [0.0833 [0.003125] 0.003125 [0.00104 EC [0.25 [0.125 [0 [0.005125 [0.00156 1 Jo [1 [0.0653 [0 [- [0.00104
0.125 [0.003125] [0.00156
0.167 [0.004167] ~ | 0.00208
0.125 [0.003125] ~ [0.00156 Jo [1 [0.0833 Jo |- [0.00104 Te [0 [0.25 [6.128 [0 [0.00313E [0.00166
0.0833 [0.003125] 0.003125 | 0.00104
0.125 ]0.003125]- [0.00156
0.0833 | 0.003125] 0.003125 | 0.00104 [0 [0.25 [0.125 [0 [0.005125 [0.0015 21 Jo Jo Jo To Jo Jo | 22 [0 [0.25 [0.125 [0 [0.003125 | 0.0015 33 Jo [1 [0.0833 [0 [- [0.00104 24 Jo [0.25 [0.125 [0 | 0.003125 [0.00156 eee]
As shown, some values of Ata, Ata, Atc are zero, or not determined because xp > 0.25. As a result, for each virtual point n, one or two different nonzero values are present for Ata, Ata, Ate. These values are then determined to be At; and Ats. (See below table). The cut-off frequency for the high pass filter for each virtual point n may be determined as =v na fe = mam for — < 0.5 and =r na fe = mem for — > 0.5.
Thus, for a virtual sound source having a plate shape with a total surface area A of 625 cm? which vibrates freely on its edges and is homogenous in its material structure, the following values for At and f. may be used. io oa
0.003125 0.00156 53.33
0.00104 | 0 | 80 |
0.003125 0.00156 53.33 Ee 6 0.003125 0.00156 53.33
0.003125] 0.00104] 80 | IEE 0.003125 0.00156 53.33 5 | 0.003125] 0.00104] 80
0.003125 0.00156 53.33
0.00104 [B | 80 |
0.003125 0.00156 53.33
0.004167] 0.00208] 40 |
0.003125 0.00156 53.33
0.00104 | 0 | #0 |
0.003125 0.00156 53.33
0.003125] 0.00104] 80 |
0.003125 0.00156 53.33
0.005125] 0.00104] 80 |
0.003125 0.00156 53.33 Ta
0.003125 0.00156 53.33
0.00l0s | 0 | so |
0.003125 0.00156 53.33 ese Ta
Thus, with these values, a user will perceive the generated audio signal as originating from a plate-shaped sound source of homogeneous substance and of particular size, whereas the loudspeakers need not be spatially arranged in a particular manner.
In an embodiment, the method comprises obtaining shape data representing the virtual positions of the respective virtual points on the virtual sound source’s shape and determining the time delays that are to be introduced by the respective time delay operations based on the virtual positions of the respective virtual points. If the virtual sound source is shaped as a square plate, then the time delays may be determined using the formula described above.
Similarly as for 2D shapes, for a 3D shape two or more modified audio signal components are determined for some or each of the generated audio signal components vy,(t) associated with virtual points that are defined on the shape. The values for the to be introduced time delays for each virtual point are in accordance with At=Vx/v, wherein V being the volume of the shape, wherein x denotes for virtual point n a multiplication factor according to the radial length r, from the centre and/or the edges of the shape to point n, and v relates to the speed of sound through a medium.
For each geometrical shape and/or different materials of heterogenous substance or material conditions, different variations of the algorithm may apply in accordance with the relationship between spatial dimensions of the shape and the time difference value at each virtual point.
For shapes that are not regular polygons and/or irregularly shaped, more than two or many modified audio signal components may be obtained for some or each of the generated audio signal components yn (t).
Figure 8 shows (top) the spectrogram of the audio signal component yi (tL) and (second from top) the spectrogram of the audio signal component ye{t) and (middle) the spectrogram of the audio signal component yy (ft) and (second from bottom) the spectrogram of the audio signal component yi; (ft) and (bottom) the spectrogram of the audio signal component yi3(t) indicated in figure 6A. The values for the time delays and the value of the frequency cut-off f. may be found in the above table. Figure 9A shows a flow chart according to an embodiment of the method wherein the generated audio signal will be perceived by an observer O as originating from a sound source S that is positioned at a distance, such as a horizontal distance away from him. The horizontal distance may be understood as the distance between the perceived virtual sound source and observer, wherein the virtual sound source is positioned in front of the observer.
In this embodiment, the input audio signal x(t) is modified using a time delay operation introducing a time delay and a signal feedback operation to obtain a first modified audio signal. Then, a second modified audio signal is generated based on a combination of the input audio signal x(t) and the first modified audio signal. The audio signal y(t} is generated by attenuating the second modified audio signal and optionally by performing a time delay operation as shown.
Preferably, the time delay that is introduced by the time delay operation performed for obtaining the first modified audio signal is as short as possible, e.g. shorter than
0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds. Most preferably, approximately 0.0001 seconds. In case of a digital sample rate of 96 kHz, the time delay may be 0.00001 seconds.
In dependence of the value of c together with value d, an observer will perceive different distances between himself and the virtual sound source. Herein, values in the triangles, i.e. in the attenuation or amplification operations may be understood to indicate a constant with which a signal is multiplied. Thus, if such value is larger than 1, then a signal amplification is performed. If such value is smaller than 1, then a signal attenuation is performed. When c=0 and d=1 no distance will be perceived and when c=1 and d=0 a maximum distance will be perceived corresponding a relative distance where the sound source has become imperceptible, and thus the output of the resulting sum audio signal will be 0 (- inf db). For performing the signal feedback operation to determine the first modified audio signal, the value for d may relate to the value for c as d=1-cx where the value for x is a multiplication factor equal to or smaller than 1 applied to the amount of signal feedback that influences the steepness of a high-frequency dissipation curve.
In an example, the method comprises obtaining distance data representing the distance of the virtual sound source. Then, the input audio signal is attenuated in dependence of the distance of the virtual sound source in order to obtain the modified audio signal.
The optional time delay indicated by At: can create a Doppler effect associated with movement of the virtual sound source. At: may be determined as At: = L / v, wherein L is a distance between the sound source S and the observer O and v is the speed of sound through a medium.
Fig. 10 (top) shows the spectrogram of sum audio signal after applying c=0, The input audio signal is white noise. Here, if c=0 then no modification is visible in the sum audio signal.
Fig. 10 (middle) shows the spectrogram of sum audio signal after applying c=0.5, The input audio signal is white noise. The observable result is a decrease of loudness of -12 db and a gradual damping of higher frequencies, as the perceived distance between the observer and the sound on length L increases, i.e. the higher frequencies of the sound dissipate proportionally faster than the lower frequencies. The curvature of the high-frequency dissipation will increase or decrease by varying the value x that is smaller than 1 and that multiplies the signal feedback amplitude.
Fig. 10 (bottom) shows the spectrogram of sum audio signal after applying c=0.99, The input audio signal is white noise.
The overall loudness has decreased -32 db and the steepness of the high-frequency dissipation curve has increased, rendering the output audio signal close to inaudible, the perceived effect being as if the sound has dissipated in the distance almost entirely.
Figure 11A shows a flow chart illustrating an embodiment of the method when the virtual sound source S is positioned at a virtual height H above an observer O (see figure 11B as well). Herein, the input audio signal x(t) is modified using a signal inverting operation, a signal attenuation operation and a time delay operation introducing a time delay in order to obtain a third modified audio signal. Then, the audio signal is generated based on a combination, e.g. summation, of the input audio signal and the third modified audio signal.
The input audio signal x(t) may be attenuated in dependence of the height to obtain the third modified audio signal, preferably such that the higher the virtual sound source is positioned above the observer, the lower the degree of attenuation is. This is shown in figure 11 in that the value for e increases with increasing height of the sound source S.
The introduced time delays as depicted in figure 11A are preferably as short as possible, e.g. shorter than 0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds. Most preferably in case of a digital sample rate of 96 kHz, the time delay may be
0.00001 seconds In case the virtual sound source is positioned above a listener modifying the input audio signal to obtain the third modified audio signal optionally comprises performing a signal feedback operation. In a particular example, this step comprises recursively adding an attenuated version of a signal, e.g. the signal resulting from the time delay operation, signal attenuation operation and signal inverting operation that are performed to eventually obtain the third modified audio signal, to itself. If the signal feedback operation is performed, value f may be equal to f=e*x where the value for x is a multiplication factor smaller than 1 applied to the amount of signal feedback that influences the steepness of a low-frequency dissipation curve. By varying value e, preferably between 0-1, a perception of height can be added to an audio signal, optionally with value f simultaneously. Herein, e=0 and f=0 correspond to no perceived height and e=1 and f<1 to a maximum perceived height, i.e. a distance above the observer where the sound source has become close to imperceptible. Fig. 12A-12C depicts the spectra of audio signals according to an embodiment of the invention. Fig. 12A shows the spectrogram of sum audio signal after applying e=0. The input audio signal is white noise. Here, if e=0, then no modification is visible in the sum audio signal. Fig. 12B shows the spectrogram of sum audio signal after applying e=0.5. The input audio signal is white noise. The observable result is a gradual damping of lower frequencies, as the perceived height H of the sound source S above the observer O increases, i.e. the lower frequencies of the sound dissipate with proportional increase of the value e. The steepness of the curve of the low-frequency dissipation increases or decreases by varying the value x that is smaller than 1 and that multiplies the signal feedback amplitude f.
Fig. 12C shows the spectrogram of sum audio signal after applying e=0.99, The input audio signal is white noise.
The steepness of the high-frequency dissipation curve has increased, rendering the output audio signal close to inaudible for f<12 kHz, the perceived effect being as if the sound is at far distance above the head of the perceiver.
Figure 13A shows a flow chart illustrating an embodiment of the method wherein the virtual sound source S is positioned at a virtual depth D below an observer O. (See figure 13B as well). This embodiment comprises modifying the input audio signal x(t) using a time delay operation introducing a time delay, a signal attenuation and a signal feedback operation in order to obtain a sixth modified audio signal.
In the depicted embodiment, performing the signal feedback operation comprises recursively adding an attenuated version of a signal, e.g. the signal resulting from the time delay operation that is performed to eventually obtain the sixth modified audio signal, to itself.
For the depicted embodiment this means that the value for h is nonzero.
Preferably, the signal that is recursively added is attenuated in dependence of the depth below the observer, e.g. such that the lower the virtual sound source 1s positioned below the observer, the lower this attenuation is (corresponding to higher values for h in figure 13). The attenuation of the input audio signal before the feedback operation may be performed such that the lower the virtual sound source is positioned below the observer, the lower the attenuation (corresponding to higher values for g in figure 13). Then, the audio signal y(t) is generated based on a combination of the input audio signal and the sixth modified audio signal.
The introduced time delay as depicted in figure 13A is preferably as short as possible, e.g. shorter than 0.00007 seconds, preferably shorter than 0.00005 seconds, more preferably shorter than 0.00002 seconds. Most preferably in case of a digital sample rate of 96 kHz, the time delay may be
0.00001 seconds.
When g=0 and h=0 no depth will be perceived and when g=1 and h=1 a maximum depth will be perceived between the sound source S and the observer O. For performing the signal feedback operation to determine the third modified audio signal, the value for h may relate to the value for g as h=g*x where the value for x is a multiplication factor equal to or smaller than 1 applied to the amount of signal feedback, which influences the steepness of a high-frequency dissipation curve.
Fig. 14 depicts a method and system for generating an audio signal according to an embodiment of the invention. In particular, Fig. 14 describes a complex flowchart of a spatial wave transform. Based on input signal x(t) several audio signal components y(t) are determined, e.g. one for each virtual point on the virtual sound source’s shape. Each audio signal component y(t) is determined by performing steps that are indicated in the boxes 70. Audio signal component yi{t}) is determined by performing the steps as shown in box 701. In each box 70, similar steps may be performed, yet while using other valued parameters.
Figure 14 in particular illustrates an example combination of several embodiment as described herein. Box 74 comprises the same method steps as illustrated in figure 9A, box 76 comprises the same method steps as illustrated in figure 114, box 78 comprises the same method steps as illustrated in figure 13A. Accordingly, the time delays that are introduced by the time delay operations of box 72 may be determined in accordance with methods described herein with reference to figures 7A and 7B.
As described above, the signal inverting operations in box 72 may only be performed if the virtual sound source cannot freely vibrate on its edges.
In such case, the high-pass filter 73 is inactive.
If the virtual sound source can freely vibrate on its edges, the signal inverting operation in box 72 are not performed.
In such case, preferably, the high-pass filter is active.
The value for the cut-off frequency may be determined in accordance with methods described with reference to figures 7A and 7B.
Further, the parameters c and d and the time delay in box 74 may be valued and/or varied and/or determined as described with reference to figures 9A and 9B.
The parameters e and f may be valued and/or varied and/or determined as described with reference to figures 11A and 11B.
The parameters g and h may be valued and/or varied and/or determined as described with reference to figures 13A and 13B.
In the depicted embodiment, generating an audio signal component thus comprises adding dimensional information to the input audio signal, which may be performed by the steps indicated by box 72, adding distance information, which may be performed by steps indicated by box 74, and adding height information, which may be performed by steps indicated by box 76, or depth information, which may be performed by steps indicated by box 78. Further, a doppler effect may be added to the input audio signal, for example by adding an additional time delay as shown in box 80. Preferably, because a virtual sound source is either positioned above or below an observer, only one of the modules 76 or 78 is performed.
Module 76 can be set as inactive by setting e=0 and module 78 can be set inactive by setting g=0. Figure 15 depicts a user interface 90 according to an embodiment of the invention.
An embodiment of the method comprises generating a user interface 90 as described herein. This user interface 90 enables a user to input the virtual sound source’s shape, -respective virtual positions of virtual points on the virtual sound source’s shape, -the distance between the virtual sound source and the observer, -the height at which the virtual sound source is positioned above the observer, -the depth at which the virtual sound source is positioned below the observer.
All functional operations of a spatial wave transform are translated to front-end user properties, i.e. audible manipulations of sound in a virtual space. The application of the invention is in no way limited to the lay-out and of this particular interface example and can be the subject of numerous approaches in system design and involve numerous levels of control for shaping and positioning sound sources in a virtual space, nor is it limited to any particular platform, medium or visual design and layout.
The depicted user interface 90 comprises an input module that enables a user to control the input audio signal of a chain using input receives. The input receives may comprise of multiple audio channels, either receiving from other chains or external audio sources, together combined as the audio input signal of a chain. The user interface enables a user to control the amplification of each input channel, e.g. by using gain knobs 92.
The user interface 90 may further comprise an output module that enables a user to route the summed audio output signal of the chain as an audio input signal to other chains.
The user interface 90 may further comprise a virtual sound source definition section that enables a user to input parameters relating to the virtual sound source, such as its shape, e.g. by means of a drop-down menu 96, and/or whether the virtual sound source is hollow or solid and/or the scale of the virtual sound source and/or its dimensions, e.g. its Cartesian dimensions and/or a rotation and/or a resolution. The latter indicates how many virtual points are determined per unit of virtual surface area. This allows a user to control the amount of required calculations.
The input means for inputting parameters relating to rotation may be presented as endless rotational knobs for dimensions x, y and =z The user interface 90 may further comprise a position sector that enables a user to input parameters relating to the position of the virtual sound source. the position of the shape in 3-dimensional space may be expressed in Cartesian coordinates +/- x, vy, z wherein the virtual center of the space is denoted as 0,0,0; and which may be presented as a visual 3-dimensional field that one can place and move a virtual object within. This 3-dimensional control field may be scaled in size by adjusting the radius of the field.
The user interface 90 may further comprise an attributes section 100 that enables a user to control various parameters, such as the bandwidth and peak level of the resonance, perceived distance, perceived elevation, doppler effect.
The user interface 90 may further comprise an output section 102 that enables a user to control the output. For example, the discrete amplification of each audio signal component that is distributed to a configured amount of audio output channels may be controlled. The gain of each loudspeaker may be automatically controlled by i) the modelling of the virtual sound source’s shape, ii) the rotation of the shape in 3-dimensional space and iii) the position of the shape in 3-dimensional space. The method for distribution of the audio signal components to the audio output channels may depend on the type of loudspeaker configuration and may be achieved by any such methods known in the art. The output section 102 may comprise a master level fader
104.
The user input that is received through the user interface may be used to determine appropriate values for the parameters according to methods described herein.
Fig. 16 depicts a block diagram illustrating a data processing system according to an embodiment. As shown in Fig. 16, the data processing system 1100 may include at least one processor 1102 coupled to memory elements 1104 through a system bus 1106. As such, the data processing system may store program code within memory elements 1104. Further, the processor 1102 may execute the program code accessed from the memory elements 1104 via a system bus 1106. In one aspect, the data processing system may be implemented as a computer that is suitable for storing and/or executing program code. It should be appreciated, however, that the data processing system 1100 may be implemented in the form of any system including a processor and a memory that is capable of performing the functions described within this specification.
The memory elements 1104 may include one or more physical memory devices such as, for example, local memory 1108 and one or more bulk storage devices 1110. The local memory may refer to random access memory or other non-persistent memory device(s) generally used during actual execution of the program code. A bulk storage device may be implemented as a hard drive or other persistent data storage device. The processing system 1100 may also include one or more cache memories (not shown) that provide temporary storage of at least some program code in order to reduce the number of times program code must be retrieved from the bulk storage device 1110 during execution.
Input/output (I/0) devices depicted as an input device 1112 and an output device 1114 optionally can be coupled to the data processing system. Examples of input devices may include, but are not limited to, a keyboard, a pointing device such as a mouse, or the like. Examples of output devices may include, but are not limited to, a monitor or a display, speakers, or the like. Input and/or output devices may be coupled to the data processing system either directly or through intervening I/O controllers.
In an embodiment, the input and the output devices may be implemented as a combined input/output device (illustrated in Fig. 16 with a dashed line surrounding the input device 1112 and the output device 1114). An example of such a combined device is a touch sensitive display, also sometimes referred to as a “touch screen display” or simply “touch screen”. In such an embodiment, input to the device may be provided by a movement of a physical object, such as e.g. a stylus or a finger of a user, on or near the touch screen display.
A network adapter 1116 may also be coupled to the data processing system to enable it to become coupled to other systems, computer systems, remote network devices, and/or remote storage devices through intervening private or public networks. The network adapter may comprise a data receiver for receiving data that is transmitted by said systems, devices and/or networks to the data processing system 1100, and a data transmitter for transmitting data from the data processing system 1100 to said systems, devices and/or networks. Modems, cable modems, and Ethernet cards are examples of different types of network adapter that may be used with the data processing system 1100.
As pictured in Fig. 16, the memory elements 1104 may store an application 1118. In various embodiments, the application 1118 may be stored in the local memory 1108, the one or more bulk storage devices 1110, or apart from the local memory and the bulk storage devices.
It should be appreciated that the data processing system 1100 may further execute an operating system (not shown in Fig. 11) that can facilitate execution of the application 1118. The application 1118, being implemented in the form of executable program code, can be executed by the data processing system 1100, e.g., by the processor 1102. Responsive to executing the application, the data processing system 1100 may be configured to perform one or more operations or method steps described herein.
In one aspect of the present invention, the data processing system 1100 may represent an audio signal processing system.
Various embodiments of the invention may be implemented as a program product for use with a computer system, where the program(s) of the program product define functions of the embodiments (including the methods described herein). In one embodiment, the program(s) can be contained on a variety of non-transitory computer-readable storage media, where, as used herein, the expression “non-transitory computer readable storage media” comprises all computer-readable media, with the sole exception being a transitory, propagating signal.
In another embodiment, the program(s) can be contained on a variety of transitory computer-readable storage media.
Illustrative computer-readable storage media include, but are not limited to: (i) non-writable storage media (e.g., read- only memory devices within a computer such as CD-ROM disks readable by a CD-ROM drive, ROM chips or any type of solid- state non-volatile semiconductor memory) on which information is permanently stored; and (ii) writable storage media (e.g., flash memory, floppy disks within a diskette drive or hard-
disk drive or any type of solid-state random-access semiconductor memory) on which alterable information is stored. The computer program may be run on the processor 1102 described herein.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
It will be further understood that the terms "comprises" and/or "comprising," when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of embodiments of the present invention has been presented for purposes of illustration, but is not intended to be exhaustive or limited to the implementations in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the present invention. The embodiments were chosen and described in order to best explain the principles and some practical applications of the present invention, and to enable others of ordinary skill in the art to understand the present invention for various embodiments with various modifications as are suited to the particular use contemplated.

Claims (24)

CONCLUSIESCONCLUSIONS 1. Werkwijze voor het genereren van een met een virtuele geluidsbron geassocieerd audiosignaal y(t), de werkwijze omvattende het verkrijgen van een invoer-audiosignaal x(t), en het modificeren van het invoer-audiosignaal x(t) om een gemodificeerd audiosignaal te verkrijgen onder gebruikmaking van een tijdsvertraging introducerende signaalvertragingsbewerking; en het genereren van het audiosignaal y(t) op basis van een combinatie, bijv. een sommatie, van het invoer-audiosignaal x(t) en het gemodificeerde audiosignaal.A method of generating an audio signal y(t) associated with a virtual sound source, the method comprising obtaining an input audio signal x(t), and modifying the input audio signal x(t) to form a modified audio signal obtainable using a time delay introducing signal delay operation; and generating the audio signal y(t) based on a combination, e.g. a summation, of the input audio signal x(t) and the modified audio signal. 2. De werkwijze volgens conclusie 1, waarbij de virtuele geluidbron een vorm heeft, de werkwijze omvattende het genereren van met respectievelijke virtuele punten op de vorm van de virtuele geluidsbron geassocieerde audiosignaalcomponenten, deze stap omvattende het genereren van een eerste audiosignaalcomponent geassocieerd met een eerste virtueel punt op de vorm van de virtuele geluidsbron en een tweede audiosignaalcomponent geassocieerd met een tweede virtueel punt op de vorm van de virtuele geluidsbron, waarbij het genereren van de eerste audiosignaalcomponent het modificeren van het invoer-audiosignaal omvat om een gemodificeerd eerste audiosignaalcomponent te verkrijgen onder gebruikmaking van een eerste signaalvertragingsbewerking die een eerste tijdsverlenging introduceert en het genereren omvat van de eerste audiosignaalcomponent op basis van een combinatie , bijvoorbeeld een sommatie, van het invoer- audiosignaal en de gemodificeerde eerste audiosignaalcomponent, en waarbij het genereren van de tweede audiosignaalcomponent het modificeren van het invoer-audiosignaal omvat om een gemodificeerd tweede audiosignaalcomponent te verkrijgen onder gebruikmaking van een tweede signaalvertragingsbewerking die een tweede tijdsverlenging introduceert verschillend van de eerste tijdsvertraging en het genereren omvat van de tweede audiosignaalcomponent op basis van een combinatie, bijvoorbeeld een sommatie, van het invoer-audiosignaal en de gemodificeerde tweede audiosignaalcomponent.The method of claim 1, wherein the virtual sound source has a shape, the method comprising generating audio signal components associated with respective virtual points on the shape of the virtual sound source, the step comprising generating a first audio signal component associated with a first virtual point on the shape of the virtual sound source and a second audio signal component associated with a second virtual point on the shape of the virtual sound source, wherein generating the first audio signal component comprises modifying the input audio signal to obtain a modified first audio signal component using of a first signal delay operation introducing a first time extension and comprising generating the first audio signal component based on a combination, e.g., a summation, of the input audio signal and the modified first audio signal component, and wherein the generating the second audio signal component comprises modifying the input audio signal to obtain a modified second audio signal component using a second signal delay operation introducing a second time extension different from the first time delay and comprising generating the second audio signal component based on a combination e.g. a summation of the input audio signal and the modified second audio signal component. 3. De werkwijze volgens conclusie 2, omvattende het verkrijgen van vormgegevens die de virtuele posities van de respectievelijke virtuele punten op de vorm van de virtuele geluidsbron representeren, en het bepalen van de eerste resp. tweede tijdsvertraging op basis van de virtuele positie van het eerste resp. tweede virtuele punt.The method of claim 2, comprising obtaining shape data representing the virtual positions of the respective virtual points on the shape of the virtual sound source, and determining the first resp. second time delay based on the virtual position of the first resp. second virtual point. 4. De werkwijze volgens en of meer van de voorgaande conclusies, waarbij de virtuele geluidsbron een afstand van een observator heeft, de werkwijze omvattende het modificeren van het invoer-audiosignaal onder gebruikmaking van een tijdsvertragingsbewerking die een tijdsvertraging introduceert en een signaalterugkoppelbewerking om een eerste gemodificeerd audiosignaal te verkrijgen, en het genereren van een tweede gemodificeerd audiosignaal op basis van een combinatie van het invoer-audiosignaal x(t) en het eerste gemodificeerde audiosignaal; het genereren van het audiosignaal y{t) op basis van het tweede gemodificeerde audiosignaal, deze stap omvattende het dempen van het tweede gemodificeerde audiosignaal en optioneel omvattende het uitvoeren van een tijdsvertragingsbewerking die een tweede tijdsvertraging introduceert.The method of any one of the preceding claims, wherein the virtual sound source is remote from an observer, the method comprising modifying the input audio signal using a time delay operation introducing a time delay and a signal feedback operation to produce a first modified obtain an audio signal, and generate a second modified audio signal based on a combination of the input audio signal x(t) and the first modified audio signal; generating the audio signal y{t) based on the second modified audio signal, this step comprising muting the second modified audio signal and optionally performing a time delay operation introducing a second time delay. 5. De werkwijze volgens conclusie 4, waarbij de geïntroduceerde tijdsvertraging korter is dan 0,00007 seconden, bij voorkeur korter dan 0,00005 seconden, bij hogere voorkeur korter dan 0,00002 seconden, bij hoogste voorkeur ongeveer 0,00001 seconden.The method of claim 4, wherein the time delay introduced is less than 0.00007 seconds, preferably less than 0.00005 seconds, more preferably less than 0.00002 seconds, most preferably about 0.00001 seconds. 6. De werkwijze volgens conclusies 4 of 5, omvattende het dempen van het tweede gemodificeerde audiosignaal in afhankelijkheid van afstand van de virtuele geluidsbron.The method of claims 4 or 5, comprising attenuating the second modified audio signal in dependence on distance from the virtual sound source. 7. De werkwijze volgens conclusie 6, waarbij de signaalterugkoppelbewerking het dempen van een signaal omvat, bijvoorbeeld het signaal zoals verkregen na het uitvoeren van de tijdsvertragingsbewerking die genoemde tijdsvertraging introduceert, en het recursief toevoegen van het gedempte signaal aan het signaal zelf, de werkwijze verder omvattende het besturen van een mate van demping in de signaalterugkoppelbewerking en de mate van demping van het tweede gemodificeerde audiosignaal in afhankelijkheid van genoemde afstand, zodat hoe groter de afstand is, hoe lager de mate van demping in de signaalterugkoppelbewerking en hoe hoger de mate van demping van het tweede gemodificeerde audiosignaal.The method of claim 6, wherein the signal feedback operation comprises muting a signal, e.g., the signal obtained after performing the time delay processing introducing said time delay, and adding the attenuated signal recursively to the signal itself, the method further comprising controlling a degree of attenuation in the signal feedback processing and the degree of attenuation of the second modified audio signal in dependence on said distance, so that the greater the distance, the lower the degree of attenuation in the signal feedback processing and the higher the degree of attenuation of the second modified audio signal. 8. De werkwijze volgens een of meer van de voorgaande conclusies, waarbij de virtuele geluidsbron gepositioneerd is op een virtuele hoogte boven een observator, de werkwijze omvattende het modificeren van het invoer-audiosignaal x(t) onder gebruikmaking van een signaalinverteringsbewerking, een signaaldempingsbewerking en een tijdsvertragingsbewerking die een tijdsvertraging introduceert, om een derde gemodificeerd audiosignaal te verkrijgen, en het genereren van het audiosignaal op basis van een combinatie van het invoer-audiosignaal en het derde gemodificeerde audiosignaal.The method of any preceding claim, wherein the virtual sound source is positioned at a virtual height above an observer, the method comprising modifying the input audio signal x(t) using a signal inverting operation, a signal attenuating operation and a time delay operation introducing a time delay to obtain a third modified audio signal, and generating the audio signal based on a combination of the input audio signal and the third modified audio signal. 9. De werkwijze volgens conclusie 8, waarbij het modificeren van het invoer-audiosignaal om het derde gemodificeerde audiosignaal te verkrijgen het uitvoeren omvat en een signaalterugkoppelbewerking.The method of claim 8, wherein modifying the input audio signal to obtain the third modified audio signal comprises outputting and signal feedback processing. 10. De werkwijze volgens conclusie 8 of 9, waarbij genoemde signaaldempingsbewerking voor het verkrijgen van het derde gemodificeerde audiosignaal uitgevoerd wordt in afhankelijkheid van de hoogte van de virtuele geluidsbron.The method according to claim 8 or 9, wherein said signal attenuation operation for obtaining the third modified audio signal is performed depending on the height of the virtual sound source. 11. De werkwijze volgens conclusie 10, waarbij genoemde signaaldempingsbewerking zodanig uitgevoerd wordt dat hoe hoger de virtuele geluidsbron is gepositioneerd boven de observator, hoe lager de mate van demping is.The method of claim 10, wherein said signal muting operation is performed such that the higher the virtual sound source is positioned above the observer, the lower the degree of muting. 12. De werkwijze volgens een of meer van de voorgaande conclusies 8-11, waarbij de tijdsvertraging die wordt geïntroduceerd voor het verkrijgen van het derde gemodificeerde audiosignaal korter is dan 0,00007 seconden, bij voorkeur korter dan 0,00005 seconden, bij hogere voorkeur korter dan 0,00002 seconden, bij hoogste voorkeur ongeveer 0,00001 seconden.The method according to any one of the preceding claims 8-11, wherein the time delay introduced for obtaining the third modified audio signal is less than 0.00007 seconds, preferably less than 0.00005 seconds, more preferably less than 0.00002 seconds, most preferably about 0.00001 seconds. 13. De werkwijze volgens een of meer van de voorgaande conclusies 1-7, waarbij de virtuele geluidsbron is gepositioneerd op een virtuele diepte onder een observator, de werkwijze omvattende het modificeren van het invoer-audiosignaal x(t) onder gebruikmaking van een tijdvertragingsbewerking die een tijdsvertraging introduceert, een eerste signaaldempingsbewerking en een signaalterugkoppelbewerking om een zesde gemodificeerd audiosignaal te verkrijgen, en het genereren van het audiosignaal op basis van een combinatie van het invoer-audiosignaal en het zesde gemodificeerde audiosignaal.The method of any of the preceding claims 1-7, wherein the virtual sound source is positioned at a virtual depth below an observer, the method comprising modifying the input audio signal x(t) using a time delay operation that introduces a time delay, a first signal attenuation operation and a signal feedback operation to obtain a sixth modified audio signal, and generating the audio signal based on a combination of the input audio signal and the sixth modified audio signal. 14. De werkwijze volgens conclusie 13, waarbij de geïntroduceerde tijdsvertraging voor het verkrijgen van het zesde gemodificeerde audiosignaal korter is dan 0,0007% seconden, bij voorkeur korter dan 0,00005 seconden, bij hogere voorkeur korter dan 0,00002 seconden, bij hoogste voorkeur ongeveer 0,00001 seconden.The method of claim 13, wherein the time delay introduced for obtaining the sixth modified audio signal is less than 0.0007% seconds, preferably less than 0.00005 seconds, more preferably less than 0.00002 seconds, at most preferably about 0.00001 seconds. 15. De werkwijze volgens conclusie 13 of 14, waarbij het uitvoeren van de signaalterugkoppelbewerking het recursief toevoegen omvat van een gedempte versie van een signaal, bijvoorbeeld van het signaal dat resulteert uit de tijdsvertragingsbewerking en -dempingsbewerking die worden uitgevoerd om uiteindelijk het zesde gemodificeerde audiosignaal te verkrijgen, aan zichzelf.The method of claim 13 or 14, wherein performing the signal feedback processing comprises recursively adding a muted version of a signal, e.g., the signal resulting from the time delay processing and muting processing performed to finally produce the sixth modified audio signal. obtain to itself. 16. De werkwijze volgens conclusie 15, waarbij het signaal dat recursief wordt toegevoegd, gedempt wordt in afhankelijkheid van de diepte onder de observator.The method of claim 15, wherein the signal added recursively is attenuated depending on the depth below the observer. 17. De werkwijze volgens conclusie 16, waarbij het signaal dat recursief wordt toegevoegd, zodanig gedempt wordt dat hoe lager de virtuele geluidsbron is gepositioneerd onder de observator, hoe lager de demping is.The method of claim 16, wherein the signal added recursively is attenuated such that the lower the virtual sound source is positioned below the observer, the lower the attenuation. 18. De werkwijze volgens conclusie 16 of 17, waarbij de eerste signaaldempingsbewerking is uitgevoerd in afhankelijkheid van de diepte van de virtuele geluidsbron onder de observator.The method of claim 16 or 17, wherein the first signal attenuation operation is performed in dependence on the depth of the virtual sound source below the observer. 19. De werkwijze volgens conclusie 18, waarbij genoemde eerste signaaldempingsbewerking wordt uitgevoerd zodanig dat hoe lager de virtuele geluidsbron is gepositioneerd onder de observator, hoe lager de demping is.The method of claim 18, wherein said first signal attenuation operation is performed such that the lower the virtual sound source is positioned below the observer, the lower the attenuation is. 20. De werkwijze volgens een of meer van de voorgaande conclusies, verder omvattende het ontvangen een gebruikersinvoer die indicatief is voor -de vorm van de virtuele geluidsbron, en/of indicatief voor -respectievelijke virtuele posities van virtuele punten op de vorm van de virtuele geluidsbron, en/of indicatief voor -de afstand tussen de virtuele geluidsbron en de observator, en/of indicatief voor -de hoogte waarop de virtuele geluidsbron is gepositioneerd boven de observator, en/of indicatief voor -de diepte waarop de virtuele geluidsbron is gepositioneerd onder de observator.The method of any preceding claim, further comprising receiving a user input indicative of the shape of the virtual sound source, and/or indicative of respective virtual positions of virtual points on the shape of the virtual sound source , and/or indicative of -the distance between the virtual sound source and the observer, and/or indicative of -the height at which the virtual sound source is positioned above the observer, and/or indicative of -the depth at which the virtual sound source is positioned below the observer. 21. De werkwijze volgens een of meer van de voorgaande conclusies, verder omvattende het genereren van een gebruikersinterface die een gebruiker in staat stelt om ten minste één van: -de vorm van de virtuele geluidsbron, en/of indicatief voor -respectievelijke virtuele posities van virtuele punten op de vorm van de virtuele geluidsbron, en/of indicatief voor -de afstand tussen de virtuele geluidsbron en de observator, en/of indicatief voor -de hoogte waarop de virtuele geluidsbron is gepositioneerd boven de observator, en/of indicatief voor -de diepte waarop de virtuele geluidsbron is gepositioneerd onder de observator, in te voeren.The method according to one or more of the preceding claims, further comprising generating a user interface that enables a user to select at least one of: - the shape of the virtual sound source, and/or indicative of - respective virtual positions of virtual points on the shape of the virtual sound source, and/or indicative of -the distance between the virtual sound source and the observer, and/or indicative of -the height at which the virtual sound source is positioned above the observer, and/or indicative of - enter the depth at which the virtual sound source is positioned below the observer. 22. Een computer omvattende een een voor een computer leesbaar opslagmedium waarop voor een computer leesbare programmacode is uitgevoerd, en een processor, bij voorkeur een microprocessor, gekoppeld aan het voor een computer leesbare opslagmedium, waarbij in antwoord op het uitvoeren van de voor een computer leesbare programmacode, de processor ingericht is om de werkwijze volgens een of meer van de voorgaande conclusies -21 uit te voeren.A computer comprising a computer-readable storage medium on which computer-readable program code is executed, and a processor, preferably a microprocessor, coupled to the computer-readable storage medium, wherein in response to the execution of the computer readable program code, the processor is arranged to perform the method according to one or more of the preceding claims -21. 23. Een computerprogramma of suite van computerprogramma’s omvattende ten minste één softwarecodedeel of een computerprogrammaproduct dat ten minste één softwarecodedeel heeft opgeslagen, waarbij het softwarecodedeel, wanneer uitgevoerd op een computersysteem, ingericht is voor het uitvoeren van de werkwijze volgens een of meer van de conclusies 1-21.A computer program or suite of computer programs comprising at least one software code portion or a computer program product having stored at least one software code portion, wherein the software code portion, when executed on a computer system, is configured to perform the method of any of claims 1 -21. 24. Een niet-vluchtig voor een computer leesbaar opslagmedium dat ten minste één softwarecodedeel heeft opgeslagen, waarbij het softwarecodedeel, wanneer uitgevoerd of verwerkt door een computer, is ingericht om de werkwijze volgens een of meer van de voorgaande conclusies 1-21 uit te voeren.A non-volatile computer readable storage medium that has stored at least one software code portion, wherein the software code portion, when executed or processed by a computer, is configured to perform the method of any one of the preceding claims 1 to 21 .
NL2024434A 2019-12-12 2019-12-12 Generating an audio signal associated with a virtual sound source NL2024434B1 (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
NL2024434A NL2024434B1 (en) 2019-12-12 2019-12-12 Generating an audio signal associated with a virtual sound source
US17/784,466 US20230017323A1 (en) 2019-12-12 2020-12-10 Generating an audio signal associated with a virtual sound source
CN202080093387.3A CN114946199A (en) 2019-12-12 2020-12-10 Generating audio signals associated with virtual sound sources
JP2022536511A JP2023506240A (en) 2019-12-12 2020-12-10 Generating an audio signal associated with a virtual sound source
CA3164476A CA3164476A1 (en) 2019-12-12 2020-12-10 Generating an audio signal associated with a virtual sound source
PCT/NL2020/050774 WO2021118352A1 (en) 2019-12-12 2020-12-10 Generating an audio signal associated with a virtual sound source
EP20829377.9A EP4074078A1 (en) 2019-12-12 2020-12-10 Generating an audio signal associated with a virtual sound source

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
NL2024434A NL2024434B1 (en) 2019-12-12 2019-12-12 Generating an audio signal associated with a virtual sound source

Publications (1)

Publication Number Publication Date
NL2024434B1 true NL2024434B1 (en) 2021-09-01

Family

ID=70155240

Family Applications (1)

Application Number Title Priority Date Filing Date
NL2024434A NL2024434B1 (en) 2019-12-12 2019-12-12 Generating an audio signal associated with a virtual sound source

Country Status (1)

Country Link
NL (1) NL2024434B1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030007648A1 (en) * 2001-04-27 2003-01-09 Christopher Currell Virtual audio system and techniques
US20060120534A1 (en) * 2002-10-15 2006-06-08 Jeong-Il Seo Method for generating and consuming 3d audio scene with extended spatiality of sound source
US20190020968A1 (en) * 2016-03-23 2019-01-17 Yamaha Corporation Audio processing method and audio processing apparatus

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030007648A1 (en) * 2001-04-27 2003-01-09 Christopher Currell Virtual audio system and techniques
US20060120534A1 (en) * 2002-10-15 2006-06-08 Jeong-Il Seo Method for generating and consuming 3d audio scene with extended spatiality of sound source
US20190020968A1 (en) * 2016-03-23 2019-01-17 Yamaha Corporation Audio processing method and audio processing apparatus

Similar Documents

Publication Publication Date Title
KR102370896B1 (en) Acoustic holographic recording and playback system using metamaterial layer
JP6933215B2 (en) Sound field forming device and method, and program
KR100813272B1 (en) Apparatus and method for bass enhancement using stereo speaker
JP5342521B2 (en) Local reproduction method, local reproduction device and program thereof
US20230306953A1 (en) Method for generating a reverberation audio signal
US20230017323A1 (en) Generating an audio signal associated with a virtual sound source
JP6329679B1 (en) Audio controller, ultrasonic speaker, audio system, and program
NL2024434B1 (en) Generating an audio signal associated with a virtual sound source
JP5596632B2 (en) Filter coefficient determination device, local reproduction device, filter coefficient determination method, and program
JP6970366B2 (en) Sound image reproduction device, sound image reproduction method and sound image reproduction program
JP7010231B2 (en) Signal processing equipment and methods, as well as programs
Kronland-Martinet et al. Real-time perceptual simulation of moving sources: application to the Leslie cabinet and 3D sound immersion
JP2002159097A (en) System and method for simulating sound field
JP5705162B2 (en) Filter coefficient determination device, local reproduction device, filter coefficient determination method, and program
US20170099557A1 (en) Systems and Methods for Playing a Venue-Specific Object-Based Audio
Ziemer et al. Psychoacoustic Sound Field Synthesis
CN116320899B (en) Sounding method, device and equipment
EP3613043A1 (en) Ambience generation for spatial audio mixing featuring use of original and extended signal
JP2019032446A (en) Acoustic simulation method, device, and program
Simionato Numerical Simulation of a Tube-Delay Audio Effect
JPH09244663A (en) Transient response signal generating method, and method and device for sound reproduction
CN117528378A (en) Acoustic test box construction method and system for loudspeaker detection
van Dorp Schuitman Wave field synthesis using multi actuator panel loudspeakers: design and application of various digital filtering algorithms
Ando et al. Temporal and spatial aspects of sounds and sound fields
SMurzyNSki Acoustic foundations of signal enhancement and room acoustics