WO2007096792A1 - Device for and a method of processing audio data - Google Patents

Device for and a method of processing audio data Download PDF

Info

Publication number
WO2007096792A1
WO2007096792A1 PCT/IB2007/050375 IB2007050375W WO2007096792A1 WO 2007096792 A1 WO2007096792 A1 WO 2007096792A1 IB 2007050375 W IB2007050375 W IB 2007050375W WO 2007096792 A1 WO2007096792 A1 WO 2007096792A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
audio data
reproduction
unit
class
Prior art date
Application number
PCT/IB2007/050375
Other languages
French (fr)
Inventor
Daniel W. E. Schobben
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Publication of WO2007096792A1 publication Critical patent/WO2007096792A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • H04S3/008Systems employing more than two channels, e.g. quadraphonic in which the audio signals are in digital form, i.e. employing more than two discrete digital channels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control

Definitions

  • the invention relates to a device for processing audio data.
  • the invention further relates to a method of processing audio data.
  • the invention relates to a program element.
  • the invention relates to a computer-readable medium.
  • Audio playback devices become more and more important. Particularly, an increasing number of users buy audio players comprising multiple loudspeakers and other entertainment equipment.
  • US 5,812,674 discloses a method for simulating the acoustical quality produced by a virtual sound source and for localizing this source with respect to one or more listeners, and one or more original sound sources. This method performs the steps: 1) fixing values of perceptual parameters defining the acoustical quality to be simulated and values of parameters defining the localization of a virtual source, 2) converting these values into a pulse response described by its energy distribution as a function of the time and the frequency, 3) carrying out a context compensation so as to take account of an existing room effect, 4) obtaining an artificial reverberation from elementary signals so as to achieve a virtual acoustic environment in real time and control the localizing of the virtual source.
  • US 5,784,467 discloses a method of and an apparatus for obtaining acoustic characteristics of sound of a broad frequency range of from 0 to 20 kHz in a relatively short time with high accuracy even though an inexpensive (i.e. a relatively low processing power) computer is used.
  • Multiple sound ray vectors are defined, a virtual space is defined by a polygonal boundary, the propagation history data of the vector is calculated and stored, the vector being reflected at the boundary, and, based on the data, as for each of the vectors, a transient response thereof at an observation point is added to a time-series numerical array and stored, the response being determined on the basis of the reflected vector and a velocity potential determined at the observation point by a micro-area element of the vector defined on the boundary.
  • a device for processing audio data comprising a classification unit adapted to classify the audio data to belong to an audio class, a reproduction configuration unit adapted to configure a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and an audio data processing unit adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit.
  • a method of processing audio data comprising classifying the audio data to belong to an audio class, configuring a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit.
  • a program element is provided, which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.
  • a computer- readable medium in which a computer program is stored which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.
  • the audio processing according to embodiments of the invention can be realized by a computer program, which is by software, or by using one or more special electronic optimization circuits, that is in hardware, or in hybrid form, that is by means of software components and hardware components.
  • an audio data processing system in which, in accordance with a class, genre or sound source having generated audio data, playback characteristics of audio content by reproduction units may be set.
  • the characteristics of both the audio data to be reproduced and audio reproduction units for generating acoustic waves for playing back the audio data are adjusted in accordance with this audio class.
  • loudspeakers and audio playback parameters may be adjusted so as to match with characteristics of the sound source. For instance, when the music is piano music, the piano music may be reproduced so as to give a human user the impression that the piano is located in the room in which the user listens to the music, for instance in a corner of the room.
  • a virtual audio source simulation system may be provided.
  • a lifelike 3D audio rendering system may be provided.
  • Such a three-dimensional audio rendering system may classify the incoming audio (for instance piano sound, speech or similar) which can be stereo, multi-channel recording or separate sound recordings for the various sound sources such as an in MPEG7 (formally named "Multimedia Content Description Interface", which is a standard for describing the multimedia content data that supports some degree of interpretation of the information meaning, which can be passed onto, or accessed by, a device or a computer code; further information available on http://www.chiariglione.org/mpeg/standards/mpeg-7/mpeg-7.htm) or MPEG21.
  • the audio rendering system may determine the loudspeaker configuration (for instance factory preset, based on user input or using automatic loudspeaker configuration) including parameters like position, orientation and/or radiation of the loudspeakers.
  • the system may obtain, from a look-up table, the desired radiation pattern of the identified source (for instance that of a piano).
  • the system may then process the incoming audio (for instance amplifying, delaying and/or decorrelating the signals), and may play the processed audio over the loudspeaker setup such that the radiation pattern of the sound source may be matched.
  • the feature of processing the incoming audio data and play back the processed audio data to achieve a matching with a radiation pattern of a sound source may also be denoted or implemented in the context of a wave field synthesis (WFS) system, that is the radiation pattern of a sound using loudspeakers may be reconstructed using loudspeakers.
  • WFS wave field synthesis
  • Embodiments of the invention may be implemented in the context of a system that determines on the reproduction side what the radiation characteristics of the sound sources should be.
  • image recognition systems may be implemented which derives audio-related characteristics from image data. For instance, a speaking direction of a person may be detected. In another situation, when a piano is detected on an image, this information may be processed so as to simulate, emulate, recreate or pattern the emission characteristics of a piano with the audio reproduction system.
  • the sound- rendering unit together with the audio may receive some additional data.
  • Embodiments of the invention may provide a significantly improved, more realistic audio playback as compared to current home entertainment systems which produce a reasonable sound stage when the loudspeakers are positioned in the way they are assumed to when the audio is mixed and when the listener is exactly in the middle (i.e. sweet spot) of the loudspeakers. Embodiments of the invention may simulate such a situation in a more realistic manner.
  • a method of processing multichannel audio signals of a surround sound system comprising classifying an incoming signal yielding classification information.
  • a loudspeaker configuration of the surround sound system may be determined including parameters like position/orientation/radiation of the loudspeakers. From a database or lookup table, desired radiation pattern information of a sound source may be obtained based on the classification information.
  • the incoming audio signal may be processed based on the radiation pattern information, and the processed audio may be played back over the loudspeaker setup such that the radiation pattern of the sound sources is matched.
  • Exemplary embodiments of the invention may have the advantage that the sound reproduction may take into account instrument "directivity".
  • a virtual piano in the corner of the living room may sound essentially like a real piano, for example, no matter the position of the listener.
  • Exemplary fields of application of embodiments of the invention are home entertainment systems, and systems allowing for adding many compact wireless loudspeaker units can provide a compelling lifelike sound reproduction when implementing embodiments of the invention.
  • a lifelike 3D audio rendering system may provide a sound reproduction system that restores the radiation pattern of the sound source that is reproduced.
  • a virtual sound source like a piano or a speaking person
  • the real audio source for example like a real piano
  • Embodiments of the invention may overcome the limitation that current home entertainment systems are mostly limited to emphasize loudspeaker boxes.
  • the evolution has mainly been in increasing the number of channels and the performance is determined by the multi-channel recording/mixing.
  • One important aspect for audio playback which may be taken into account according to exemplary embodiments of the invention, is the instrument's directivity that may be quite different from that of a loudspeaker.
  • models may be used derived for sound reduction caption and reproduction.
  • a piano may be recorded with several microphones, and the sound may be reproduced by a set of loudspeakers carefully arranged to reproduce the piano sound as good as possible.
  • the loudspeaker setup is not known at the time of recording.
  • the classification unit may be adapted to classify the audio data to belong to an audio class based on audio data classification data attached to the audio data.
  • the device may be capable to read the audio data having some kind of metadata as an extension or attachment, the metadata providing information concerning the class.
  • the classification unit may be adapted to classify the audio data to belong to an audio class based on evaluating characteristics of the audio data itself.
  • algorithms may be performed which may allow deriving information concerning the audio class from the audio data itself.
  • the occurrence of beats may be an indication that the sound is rock music.
  • Other characteristics are typical for classical music, jazz, etc.
  • Speech has a completely different sound characteristic than music.
  • the kind of detected instruments (having a specific audio emission characteristics) may allow to conclude on a corresponding audio class.
  • the classification unit may be adapted to classify the audio data to belong to an audio class based on a comparison of the audio data with pre-stored characteristics of audio data related to (a particular one of) a plurality of audio classes.
  • predetermined audio class information may be stored in the device. For instance, characteristics for different instruments, vocalist, genres, etc. may be recorded, and corresponding abstract or generic audio class information may be stored in a corresponding database. Then, characteristics of a present audio item may be compared to the pre-stored class profiles, and the profile with the best match may be assumed to represent the class of the present audio item.
  • the classification unit may be adapted to classify the audio data to belong to an audio class in accordance with an audio source being an origin of the audio data.
  • an audio source being an origin of the audio data.
  • the characteristics of the sound source that has generated the audio for instance a piano, a human singer, etc.
  • the characteristics of the sound source that has generated the audio may determine the class like "instrumental”, “piano music”, etc.
  • the classification unit may be adapted to classify the audio data to belong to an audio class in accordance with a recording mode of recording the audio data. Examples for such recording modes are stereo, multi-channel, plugged/unplugged, or the like. Thus, the recording mode may form some kind of fingerprint of the audio class or may at least provide hints to which class an audio piece might belong.
  • the reproduction configuration unit may be adapted to configure reproduction properties of a plurality of audio reproduction units for reproducing the audio data in accordance with the assigned audio class.
  • embodiments of the invention may implement several audio production units like loudspeakers so that the three-dimensional impression of a three-dimensional sound source may be simulated in a realistic manner.
  • the configuration of the loudspeakers may include a correlation between different loudspeakers, and may also be adjusted based on the relative position of the loudspeakers to one another. It is possible that the loudspeakers automatically detect their positions with respect to one another, for instance by exchanging audio signals. By evaluating the time characteristics of the emission and detection of the audio signals may then allow the system to automatically determine the relative geometrical arrangement of the loudspeakers with respect to one another.
  • a multiple reproduction unit system may be particularly suitable to realistically simulate the characteristics of a physical sound source.
  • Embodiments of the invention may be used in an audio surround sound system.
  • the reproduction configuration unit may be adapted to configure, as the reproduction characteristics of the audio reproduction unit, at least one of the group consisting of a spatial position of the audio reproduction unit, a spatial orientation of the audio reproduction unit, and a radiation profile (for instance frequency and/or amplitude characteristics) of the audio reproduction unit.
  • the positions of the audio reproduction units may be modified during reproduction so as to simulate a motion of the audio source during recording.
  • the emission direction may be modified, for instance to simulate that a person giving a speech walks along a lecture hall during speaking and/or changes direction of speech.
  • the radiation profile may be adjusted, for instance by adjusting equalizing and/or reverberation properties.
  • the reproduction configuration unit may be adapted to configure the reproduction characteristics of the audio reproduction unit based on pre-stored configuration information and/or user-defined configuration information.
  • the configuration of the loudspeakers may be performed based on user commands (for instance a user may input the command "I wish that the content is played back with the acoustic of the Royal Albert Hall” or "I wish that the content is played back to sound like a piano”).
  • Class information may be linked in a database to loudspeaker configuration information. For instance, to simulate a piano, corresponding three-dimensional configuration information may be pre-stored.
  • the audio data processing unit may be adapted to process the audio data by at least one of the group consisting of amplifying the audio data, delaying the audio data, adjusting a correlation of the audio data, equalizing the audio data, adding reverberation to the audio data, and filtering the audio data.
  • the loudness of the audio data may be adjusted, or the signals may be delayed.
  • the signals may be decorrelated, a frequency distribution of the audio data may be adjusted, artificial effects like reverberation or the like may be added so as to simulate a specific audible impression.
  • the audio data may be filtered, for instance to remove very low frequency components.
  • the audio data processing unit may be adapted to process the audio data for generating reproducible audio data for reproduction so as to match a radiation pattern of a sound source being the origin of the audio data.
  • the acoustic wave emission characteristics of a three-dimensional physical object may be matched.
  • the device may comprise a database unit (for instance a storage device like a harddisk, an EEPROM, a flash memory, etc.) for storing audio reproduction characteristics data assigned to classes of audio data, wherein the database may be accessed by the reproduction configuration unit and/or by the audio data processing unit.
  • a database unit for instance a storage device like a harddisk, an EEPROM, a flash memory, etc.
  • the device may be adapted for a dynamic processing of the audio data.
  • the term "dynamic" may particularly denote that, during reproduction of an audio piece, the reproduction unit configuration and/or the audio data configuration may be continuously updated. For instance, when the sound source changes direction during recording, this direction modification may be mapped on the virtual audio simulation system.
  • the device may comprise the audio reproduction unit(s) for reproducing the reproducible audio data in accordance with the configured reproduction characteristics and in accordance with the assigned audio class.
  • Such one or more audio reproduction units may be loudspeakers, headphones, or earpieces.
  • the reproduction unit or units may be supplied with control information with respect to loudspeaker configuration and audio data modification.
  • the communication between audio processing components of the audio device and reproduction units may be carried out in a wired manner (for instance using a cable) or in a wireless manner (for instance via a WLAN, infrared communication or Bluetooth).
  • the audio device may be a realized as a gaming device, a laptop, a portable audio player, a DVD player, a CD player, a harddisk-based media player, an internet radio device, a public entertainment device, an MP3 player, a hi-fi system, a vehicle entertainment device, a car entertainment device, a portable video player, a medical communication system, a body-worn device, an audio conference system, a video conference system, or a hearing aid device.
  • a "car entertainment device” may be a hi-fi system for an automobile.
  • system primarily intends to facilitate the playback of sound or audio data
  • system for a combination of audio data and visual data
  • an embodiment of the invention may be implemented in audiovisual applications like a video player in which a loudspeaker is used, or a home cinema system.
  • Fig. 1 shows an audio data processing device according to an exemplary embodiment of the invention.
  • Fig. 2 shows an audio data processing device according to another exemplary embodiment of the invention.
  • the system 100 shows an audio surround system comprising a device 100a for processing audio data and a sound reproduction unit 100b and a user interface 101.
  • the user interface 101 in the present case is a graphical user interface (GUI) and includes a display device for displaying information to a human operator, which display device is a liquid crystal display in this case. Alternative, the display device may be a plasma display device or the like).
  • the user interface 101 comprises an input device allowing the user to input data (like data specifying the audio reproduction) or to provide the system with the control commands.
  • Such an input device may include buttons, a keypad, a joystick, a trackball, a touch screen, or may even be a microphone of a voice recognition system.
  • the user interface 101 may allow a human user to communicate in a bi-directional manner with the system 100.
  • the sound reproduction unit 100b comprises loudspeakers 107 to 109 for reproducing audio data.
  • audio data stored in an audio data storage device 102 may be processed for reproduction.
  • the audio data storage device 102 may be a CD, a DVD, or a harddisk.
  • Audio data AD stored in the audio data content storage unit 102 may be provided to a classification unit 103 for classifying the audio data AD to belong to an audio class (for instance a particular genre or a particular sound source).
  • An output of the classification unit 103 provides classification information CLI and is coupled to an input of a reproduction configuration unit 104, which configuration unit 104 is adapted for processing the classification information CLI to provide configuration instructions COI for configuring reproduction characteristics of the loudspeakers 107 to 109 for reproducing the audio data in accordance with the assigned audio class.
  • an output of the classification unit 103 is coupled to an input of an audio data processing unit 105 to provide the audio data AD, which audio data processing unit 105 is adapted for processing the audio data AD in accordance with the assigned audio class for generating reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109.
  • a database unit 106 which stores audio reproduction characteristics data assigned to classes of audio data. As can be taken from Fig. 1, the database unit 106 may be accessed by the reproduction configuration unit 104 and by the audio data processing unit 105 so as to provide these units 104, 105 with information being valuable for these units 104, 105 to correspondingly adapt loudspeaker configuration and audio data properties so as to achieve a proper correspondence between the audio classification and the audio playback.
  • An output of the reproduction configuration unit 104 is adapted to supply loudspeaker configuration instructions COI to the loudspeakers 107 to 109.
  • Each of the loudspeakers 107 to 109 receives individual reproduction configuration information.
  • an output of the audio data processing unit 105 is coupled to the loudspeakers 107 to 109 for providing the loudspeakers 107 to 109 individually with reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109. Therefore, configuration unit 104 controls loudspeaker configuration, and processing unit 105 provides audio content.
  • the classification unit 103 identifies the class or genre of this audio data to be "piano music" and/or "classical music". Such a determination may be based on appropriate algorithms. For instance, it is possible to distinguish piano music from hardrock by the acoustical properties of such music items. It is also possible that the classification unit 103 evaluates a class identifier which may be attached to the audio data and which may be indicative of the class of the audio item.
  • the class information CLI and the audio data AD are supplied to the configuration unit 104 and processing unit 105.
  • the reproduction configuration unit 104 adjusts reproduction parameters of the loudspeakers 107 to 109 like an adjustment of the position of the loudspeakers 107 to 109 (which may be modified automatically by means of electric motors or the like), an acoustic wave emission direction of the loudspeakers 107 to 109 (for instance the loudspeakers 107 to 109 may be turned), or the like.
  • the loudspeakers 107 to 109 are configured to match with an audio pattern of the source of the audio data, in the present example match with the 3D emission characteristics of a piano.
  • the classification is supplied to the audio data processing unit 105 that correspondingly adjusts the audio data so as to match to the characteristics of the sound source.
  • the audio data is equalized, filtered, amplified or modified in another manner so as to be in proper accordance with characteristics of the original sound source.
  • the processed audio data may sound, in the present example, like audio emitted from the piano.
  • loudspeaker configuration and audio data characteristics are adjusted to match to characteristics of the audio source to give a human user a realistic impression of a virtual piano.
  • an audio data processing system 200 according to an exemplary embodiment of the invention is shown.
  • the system 200 according to figure 2 comprises a audio data interface 201 that is adapted to receive audio data from a remote server via wired internet connection.
  • the interface 201 may be adapted to receive audio data via a wireless link, for instance WLAN, Bluetooth, ZigBee etc.
  • the remote server may comprise the classification unit.
  • the audio data processing system 200 may comprise an adapted interface 201a that is adapted to receive the class information CLI and the audio data AD for supplying this information to the configuration unit 104 and processing unit 105.
  • the term “comprising” does not exclude other elements or features and the "a” or “an” does not exclude a plurality. Also elements described in association with different embodiments may be combined.

Abstract

A device (100a) for processing audio data, the device (100a) comprising a classification unit (103) adapted to classify the audio data to belong to an audio class, a reproduction configuration unit (104) adapted to configure a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class, and an audio data processing unit (105) adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).

Description

Device for and a method of processing audio data
FIELD OF THE INVENTION
The invention relates to a device for processing audio data. The invention further relates to a method of processing audio data. Moreover, the invention relates to a program element. Further, the invention relates to a computer-readable medium.
BACKGROUND OF THE INVENTION
Audio playback devices become more and more important. Particularly, an increasing number of users buy audio players comprising multiple loudspeakers and other entertainment equipment.
US 5,812,674 discloses a method for simulating the acoustical quality produced by a virtual sound source and for localizing this source with respect to one or more listeners, and one or more original sound sources. This method performs the steps: 1) fixing values of perceptual parameters defining the acoustical quality to be simulated and values of parameters defining the localization of a virtual source, 2) converting these values into a pulse response described by its energy distribution as a function of the time and the frequency, 3) carrying out a context compensation so as to take account of an existing room effect, 4) obtaining an artificial reverberation from elementary signals so as to achieve a virtual acoustic environment in real time and control the localizing of the virtual source. US 5,784,467 discloses a method of and an apparatus for obtaining acoustic characteristics of sound of a broad frequency range of from 0 to 20 kHz in a relatively short time with high accuracy even though an inexpensive (i.e. a relatively low processing power) computer is used. Multiple sound ray vectors are defined, a virtual space is defined by a polygonal boundary, the propagation history data of the vector is calculated and stored, the vector being reflected at the boundary, and, based on the data, as for each of the vectors, a transient response thereof at an observation point is added to a time-series numerical array and stored, the response being determined on the basis of the reflected vector and a velocity potential determined at the observation point by a micro-area element of the vector defined on the boundary. OBJECT AND SUMMARY OF THE INVENTION
It is an object of the invention to enable efficient audio data processing. In order to achieve the object defined above, a device for processing audio data, a method of processing audio data, a program element, and a computer-readable medium according to the independent claims are provided.
According to an exemplary embodiment of the invention, a device for processing audio data is provided, the device comprising a classification unit adapted to classify the audio data to belong to an audio class, a reproduction configuration unit adapted to configure a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and an audio data processing unit adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit.
According to another exemplary embodiment of the invention, a method of processing audio data is provided, the method comprising classifying the audio data to belong to an audio class, configuring a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit. According to still another exemplary embodiment of the invention, a program element is provided, which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.
According to yet another exemplary embodiment of the invention, a computer- readable medium is provided, in which a computer program is stored which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.
The audio processing according to embodiments of the invention can be realized by a computer program, which is by software, or by using one or more special electronic optimization circuits, that is in hardware, or in hybrid form, that is by means of software components and hardware components.
According to an exemplary embodiment of the invention, an audio data processing system is provided in which, in accordance with a class, genre or sound source having generated audio data, playback characteristics of audio content by reproduction units may be set. In this context, the characteristics of both the audio data to be reproduced and audio reproduction units for generating acoustic waves for playing back the audio data are adjusted in accordance with this audio class. In other words, loudspeakers and audio playback parameters may be adjusted so as to match with characteristics of the sound source. For instance, when the music is piano music, the piano music may be reproduced so as to give a human user the impression that the piano is located in the room in which the user listens to the music, for instance in a corner of the room. Thus, a virtual audio source simulation system may be provided.
According to an exemplary embodiment of the invention, a lifelike 3D audio rendering system may be provided. Such a three-dimensional audio rendering system may classify the incoming audio (for instance piano sound, speech or similar) which can be stereo, multi-channel recording or separate sound recordings for the various sound sources such as an in MPEG7 (formally named "Multimedia Content Description Interface", which is a standard for describing the multimedia content data that supports some degree of interpretation of the information meaning, which can be passed onto, or accessed by, a device or a computer code; further information available on http://www.chiariglione.org/mpeg/standards/mpeg-7/mpeg-7.htm) or MPEG21. The audio rendering system may determine the loudspeaker configuration (for instance factory preset, based on user input or using automatic loudspeaker configuration) including parameters like position, orientation and/or radiation of the loudspeakers. The system may obtain, from a look-up table, the desired radiation pattern of the identified source (for instance that of a piano). The system may then process the incoming audio (for instance amplifying, delaying and/or decorrelating the signals), and may play the processed audio over the loudspeaker setup such that the radiation pattern of the sound source may be matched.
The feature of processing the incoming audio data and play back the processed audio data to achieve a matching with a radiation pattern of a sound source may also be denoted or implemented in the context of a wave field synthesis (WFS) system, that is the radiation pattern of a sound using loudspeakers may be reconstructed using loudspeakers.
Embodiments of the invention may be implemented in the context of a system that determines on the reproduction side what the radiation characteristics of the sound sources should be.
According to an exemplary embodiment, it is possible to process a video stream for a videoconference, and the orientation of the sound source (for instance a person speaking) may be changed in case of detecting from the video that a person on the far end turns around. In this context, explicit reference is made to paragraphs [0015], [0027] to [0031] of US 2003/0053680 Al explaining a similar system that may be implemented in the context of an exemplary embodiment of the invention.
Thus, image recognition systems may be implemented which derives audio- related characteristics from image data. For instance, a speaking direction of a person may be detected. In another situation, when a piano is detected on an image, this information may be processed so as to simulate, emulate, recreate or pattern the emission characteristics of a piano with the audio reproduction system.
According to another exemplary embodiment of the invention, the sound- rendering unit together with the audio, such as orientation information, radiation profile, etc, may receive some additional data.
Embodiments of the invention may provide a significantly improved, more realistic audio playback as compared to current home entertainment systems which produce a reasonable sound stage when the loudspeakers are positioned in the way they are assumed to when the audio is mixed and when the listener is exactly in the middle (i.e. sweet spot) of the loudspeakers. Embodiments of the invention may simulate such a situation in a more realistic manner.
According to an exemplary embodiment, a method of processing multichannel audio signals of a surround sound system may be provided, the method comprising classifying an incoming signal yielding classification information. A loudspeaker configuration of the surround sound system may be determined including parameters like position/orientation/radiation of the loudspeakers. From a database or lookup table, desired radiation pattern information of a sound source may be obtained based on the classification information. The incoming audio signal may be processed based on the radiation pattern information, and the processed audio may be played back over the loudspeaker setup such that the radiation pattern of the sound sources is matched.
Exemplary embodiments of the invention may have the advantage that the sound reproduction may take into account instrument "directivity". A virtual piano in the corner of the living room may sound essentially like a real piano, for example, no matter the position of the listener. Exemplary fields of application of embodiments of the invention are home entertainment systems, and systems allowing for adding many compact wireless loudspeaker units can provide a compelling lifelike sound reproduction when implementing embodiments of the invention.
Thus, a lifelike 3D audio rendering system according to an exemplary embodiment may provide a sound reproduction system that restores the radiation pattern of the sound source that is reproduced. A virtual sound source (like a piano or a speaking person) at a particular position of a room (for instance in the corner of a living room) may sound like the real audio source (for example like a real piano), no matter the position of the listener.
Embodiments of the invention may overcome the limitation that current home entertainment systems are mostly limited to emphasize loudspeaker boxes. The evolution has mainly been in increasing the number of channels and the performance is determined by the multi-channel recording/mixing. One important aspect for audio playback, which may be taken into account according to exemplary embodiments of the invention, is the instrument's directivity that may be quite different from that of a loudspeaker. To take into account the instrument's directivity, models may be used derived for sound reduction caption and reproduction. As an example, a piano may be recorded with several microphones, and the sound may be reproduced by a set of loudspeakers carefully arranged to reproduce the piano sound as good as possible. However, in many cases, such a recording is not available and the loudspeaker setup is not known at the time of recording.
It is possible to measure, in a laboratory, the radiation patterns of the sound sources in all classes used in a particular application. Such information may be used for assigning classes to audio content. According to an exemplary embodiment, scaled and delayed versions of the audio signal are played out over the loudspeakers. In a more sophisticated embodiment, such features may be preceded by upmixing the audio into more channels using decorrelation techniques (for instance Lauridsen decorrelation).
Next, further exemplary embodiments of the invention will be explained. In the following, further exemplary embodiments of the device for processing audio data will be described. However, these embodiments also apply for the method of processing audio data, for the computer-readable medium and for the program element.
The classification unit may be adapted to classify the audio data to belong to an audio class based on audio data classification data attached to the audio data. Thus, the device may be capable to read the audio data having some kind of metadata as an extension or attachment, the metadata providing information concerning the class. For instance, such a class may be the genre of the music (jazz, rock, classic music, easy listening, etc.), may provide information concerning the origin of the sound (information concerning the instrument or instruments, about an artist, etc.). However, additionally or alternatively, the classification unit may be adapted to classify the audio data to belong to an audio class based on evaluating characteristics of the audio data itself. Thus, algorithms may be performed which may allow deriving information concerning the audio class from the audio data itself. For instance, the occurrence of beats may be an indication that the sound is rock music. Other characteristics are typical for classical music, jazz, etc. Speech has a completely different sound characteristic than music. The kind of detected instruments (having a specific audio emission characteristics) may allow to conclude on a corresponding audio class.
The classification unit may be adapted to classify the audio data to belong to an audio class based on a comparison of the audio data with pre-stored characteristics of audio data related to (a particular one of) a plurality of audio classes. Thus, predetermined audio class information may be stored in the device. For instance, characteristics for different instruments, vocalist, genres, etc. may be recorded, and corresponding abstract or generic audio class information may be stored in a corresponding database. Then, characteristics of a present audio item may be compared to the pre-stored class profiles, and the profile with the best match may be assumed to represent the class of the present audio item.
The classification unit may be adapted to classify the audio data to belong to an audio class in accordance with an audio source being an origin of the audio data. Thus, the characteristics of the sound source that has generated the audio (for instance a piano, a human singer, etc.) may determine the class like "instrumental", "piano music", etc.
The classification unit may be adapted to classify the audio data to belong to an audio class in accordance with a recording mode of recording the audio data. Examples for such recording modes are stereo, multi-channel, plugged/unplugged, or the like. Thus, the recording mode may form some kind of fingerprint of the audio class or may at least provide hints to which class an audio piece might belong.
The reproduction configuration unit may be adapted to configure reproduction properties of a plurality of audio reproduction units for reproducing the audio data in accordance with the assigned audio class. Thus, embodiments of the invention may implement several audio production units like loudspeakers so that the three-dimensional impression of a three-dimensional sound source may be simulated in a realistic manner. The configuration of the loudspeakers may include a correlation between different loudspeakers, and may also be adjusted based on the relative position of the loudspeakers to one another. It is possible that the loudspeakers automatically detect their positions with respect to one another, for instance by exchanging audio signals. By evaluating the time characteristics of the emission and detection of the audio signals may then allow the system to automatically determine the relative geometrical arrangement of the loudspeakers with respect to one another. Thus, a multiple reproduction unit system may be particularly suitable to realistically simulate the characteristics of a physical sound source. Embodiments of the invention may be used in an audio surround sound system.
The reproduction configuration unit may be adapted to configure, as the reproduction characteristics of the audio reproduction unit, at least one of the group consisting of a spatial position of the audio reproduction unit, a spatial orientation of the audio reproduction unit, and a radiation profile (for instance frequency and/or amplitude characteristics) of the audio reproduction unit. Thus, the positions of the audio reproduction units may be modified during reproduction so as to simulate a motion of the audio source during recording. Also the emission direction may be modified, for instance to simulate that a person giving a speech walks along a lecture hall during speaking and/or changes direction of speech. Also the radiation profile may be adjusted, for instance by adjusting equalizing and/or reverberation properties.
The reproduction configuration unit may be adapted to configure the reproduction characteristics of the audio reproduction unit based on pre-stored configuration information and/or user-defined configuration information. Thus, the configuration of the loudspeakers may be performed based on user commands (for instance a user may input the command "I wish that the content is played back with the acoustic of the Royal Albert Hall" or "I wish that the content is played back to sound like a piano"). Class information may be linked in a database to loudspeaker configuration information. For instance, to simulate a piano, corresponding three-dimensional configuration information may be pre-stored.
The audio data processing unit may be adapted to process the audio data by at least one of the group consisting of amplifying the audio data, delaying the audio data, adjusting a correlation of the audio data, equalizing the audio data, adding reverberation to the audio data, and filtering the audio data. Thus, in order to give a human user a realistic perception or impression of the reproduced audio, the loudness of the audio data may be adjusted, or the signals may be delayed. The signals may be decorrelated, a frequency distribution of the audio data may be adjusted, artificial effects like reverberation or the like may be added so as to simulate a specific audible impression. Moreover, the audio data may be filtered, for instance to remove very low frequency components.
The audio data processing unit may be adapted to process the audio data for generating reproducible audio data for reproduction so as to match a radiation pattern of a sound source being the origin of the audio data. By taking this measure, the acoustic wave emission characteristics of a three-dimensional physical object may be matched.
The device may comprise a database unit (for instance a storage device like a harddisk, an EEPROM, a flash memory, etc.) for storing audio reproduction characteristics data assigned to classes of audio data, wherein the database may be accessed by the reproduction configuration unit and/or by the audio data processing unit. By implementing a database that has pre-stored empiric characteristics of sound sources, the classification and modification of the audio data and the loudspeaker configuration may be significantly improved. The device may be adapted for a dynamic processing of the audio data. In this context, the term "dynamic" may particularly denote that, during reproduction of an audio piece, the reproduction unit configuration and/or the audio data configuration may be continuously updated. For instance, when the sound source changes direction during recording, this direction modification may be mapped on the virtual audio simulation system. The device may comprise the audio reproduction unit(s) for reproducing the reproducible audio data in accordance with the configured reproduction characteristics and in accordance with the assigned audio class. Such one or more audio reproduction units may be loudspeakers, headphones, or earpieces. The reproduction unit or units may be supplied with control information with respect to loudspeaker configuration and audio data modification. The communication between audio processing components of the audio device and reproduction units may be carried out in a wired manner (for instance using a cable) or in a wireless manner (for instance via a WLAN, infrared communication or Bluetooth).
The audio device may be a realized as a gaming device, a laptop, a portable audio player, a DVD player, a CD player, a harddisk-based media player, an internet radio device, a public entertainment device, an MP3 player, a hi-fi system, a vehicle entertainment device, a car entertainment device, a portable video player, a medical communication system, a body-worn device, an audio conference system, a video conference system, or a hearing aid device. A "car entertainment device" may be a hi-fi system for an automobile.
However, although the system according to embodiments of the invention primarily intends to facilitate the playback of sound or audio data, it is also possible to apply the system for a combination of audio data and visual data. For instance, an embodiment of the invention may be implemented in audiovisual applications like a video player in which a loudspeaker is used, or a home cinema system.
The aspects defined above and further aspects of the invention are apparent from the examples of embodiment to be described hereinafter and are explained with reference to these examples of embodiment.
BRIEF DESCRIPTION OF THE DRAWINGS The invention will be described in more detail hereinafter with reference to examples of embodiment but to which the invention is not limited.
Fig. 1 shows an audio data processing device according to an exemplary embodiment of the invention. Fig. 2 shows an audio data processing device according to another exemplary embodiment of the invention.
DESCRIPTION OF EMBODIMENTS
The illustration in the drawing is schematically. In different drawings, similar or identical elements are provided with the same reference signs.
In the following, referring to Fig. 1, an audio data processing system 100 according to an exemplary embodiment of the invention will be explained.
The system 100 shows an audio surround system comprising a device 100a for processing audio data and a sound reproduction unit 100b and a user interface 101. The user interface 101 in the present case is a graphical user interface (GUI) and includes a display device for displaying information to a human operator, which display device is a liquid crystal display in this case. Alternative, the display device may be a plasma display device or the like). Furthermore, the user interface 101 comprises an input device allowing the user to input data (like data specifying the audio reproduction) or to provide the system with the control commands. Such an input device may include buttons, a keypad, a joystick, a trackball, a touch screen, or may even be a microphone of a voice recognition system. The user interface 101 may allow a human user to communicate in a bi-directional manner with the system 100.
The sound reproduction unit 100b comprises loudspeakers 107 to 109 for reproducing audio data. By a user command inputted by a user via the user interface 101, audio data stored in an audio data storage device 102 may be processed for reproduction. The audio data storage device 102 may be a CD, a DVD, or a harddisk.
Audio data AD stored in the audio data content storage unit 102 may be provided to a classification unit 103 for classifying the audio data AD to belong to an audio class (for instance a particular genre or a particular sound source). An output of the classification unit 103 provides classification information CLI and is coupled to an input of a reproduction configuration unit 104, which configuration unit 104 is adapted for processing the classification information CLI to provide configuration instructions COI for configuring reproduction characteristics of the loudspeakers 107 to 109 for reproducing the audio data in accordance with the assigned audio class.
Furthermore, an output of the classification unit 103 is coupled to an input of an audio data processing unit 105 to provide the audio data AD, which audio data processing unit 105 is adapted for processing the audio data AD in accordance with the assigned audio class for generating reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109.
Beyond this, a database unit 106 is provided which stores audio reproduction characteristics data assigned to classes of audio data. As can be taken from Fig. 1, the database unit 106 may be accessed by the reproduction configuration unit 104 and by the audio data processing unit 105 so as to provide these units 104, 105 with information being valuable for these units 104, 105 to correspondingly adapt loudspeaker configuration and audio data properties so as to achieve a proper correspondence between the audio classification and the audio playback.
An output of the reproduction configuration unit 104 is adapted to supply loudspeaker configuration instructions COI to the loudspeakers 107 to 109. Each of the loudspeakers 107 to 109 receives individual reproduction configuration information.
Furthermore, an output of the audio data processing unit 105 is coupled to the loudspeakers 107 to 109 for providing the loudspeakers 107 to 109 individually with reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109. Therefore, configuration unit 104 controls loudspeaker configuration, and processing unit 105 provides audio content.
Thus, when an audio data item related to piano music is to be played back, the classification unit 103 identifies the class or genre of this audio data to be "piano music" and/or "classical music". Such a determination may be based on appropriate algorithms. For instance, it is possible to distinguish piano music from hardrock by the acoustical properties of such music items. It is also possible that the classification unit 103 evaluates a class identifier which may be attached to the audio data and which may be indicative of the class of the audio item.
The class information CLI and the audio data AD are supplied to the configuration unit 104 and processing unit 105. The reproduction configuration unit 104 adjusts reproduction parameters of the loudspeakers 107 to 109 like an adjustment of the position of the loudspeakers 107 to 109 (which may be modified automatically by means of electric motors or the like), an acoustic wave emission direction of the loudspeakers 107 to 109 (for instance the loudspeakers 107 to 109 may be turned), or the like. Thus, the loudspeakers 107 to 109 are configured to match with an audio pattern of the source of the audio data, in the present example match with the 3D emission characteristics of a piano. Furthermore, the classification is supplied to the audio data processing unit 105 that correspondingly adjusts the audio data so as to match to the characteristics of the sound source. For instance, the audio data is equalized, filtered, amplified or modified in another manner so as to be in proper accordance with characteristics of the original sound source. Thus, the processed audio data may sound, in the present example, like audio emitted from the piano.
Thus, loudspeaker configuration and audio data characteristics are adjusted to match to characteristics of the audio source to give a human user a realistic impression of a virtual piano.
Referring to Fig. 2, an audio data processing system 200 according to an exemplary embodiment of the invention is shown. As a difference of the system 100 according to figure 1 the system 200 according to figure 2 comprises a audio data interface 201 that is adapted to receive audio data from a remote server via wired internet connection. Alternatively the interface 201 may be adapted to receive audio data via a wireless link, for instance WLAN, Bluetooth, ZigBee etc. By a user command inputted by a user via the user interface 101, audio data received by the interface 201 is provided to the classification unit 103. In a further embodiment, instead of audio data processing system 200 the remote server may comprise the classification unit. In this case the audio data processing system 200 may comprise an adapted interface 201a that is adapted to receive the class information CLI and the audio data AD for supplying this information to the configuration unit 104 and processing unit 105. It should be noted that the term "comprising" does not exclude other elements or features and the "a" or "an" does not exclude a plurality. Also elements described in association with different embodiments may be combined.
It should also be noted that reference signs in the claims shall not be construed as limiting the scope of the claims.

Claims

CLAIMS:
1. A device (100a) for processing audio data, the device (100a) comprising a classification unit (103) adapted to classify the audio data to belong to an audio class; a reproduction configuration unit (104) adapted to configure a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; an audio data processing unit (105) adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
2. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on audio data classification data attached to the audio data.
3. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on evaluating characteristics of the audio data.
4. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on a comparison of the audio data with pre-stored characteristics of audio data related to a plurality of audio classes.
5. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with an audio source being an origin of the audio data.
6. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with a music instrument and/or a human being being an origin of the audio data.
7. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with a recording mode of recording the audio data.
8. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure reproduction properties of a plurality of audio reproduction units for reproducing the audio data in accordance with the assigned audio class.
9. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure, as the reproduction characteristics of the audio reproduction unit (107 to 109), at least one of the group consisting of a position of the audio reproduction unit (107 to 109), an orientation of the audio reproduction unit (107 to 109), and a radiation profile of the audio reproduction unit (107 to 109).
10. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure the reproduction characteristics of the audio reproduction unit (107 to 109) based on pre-stored configuration information and/or user-defined configuration information.
11. The device (100a) according to claim 1, wherein the audio data processing unit (105) is adapted to process the audio data by at least one of the group consisting of amplifying the audio data, delaying the audio data, adjusting a correlation of the audio data, equalizing the audio data, adding reverberation to the audio data, and filtering the audio data.
12. The device (100a) according to claim 1, wherein the audio data processing unit (105) is adapted to process the audio data for generating reproducible audio data for reproduction so as to match a radiation pattern of a sound source being the origin of the audio data.
13. The device (100a) according to claim 1, comprising a database unit (106) for storing audio reproduction characteristics data assigned to classes of audio data, the database unit (106) being accessible by the reproduction configuration unit (104) and/or by the audio data processing unit (105).
14. The device (100a) according to claim 1, adapted for dynamically processing the audio data.
15. The device (100a) according to claim 1, comprising the audio reproduction unit (107 to 109) for reproducing the reproducible audio data in accordance with the configured reproduction characteristics and in accordance with the assigned audio class.
16. The device (100a) according to claim 15, wherein the audio reproduction unit (107 to 109) comprises at least one of the group consisting of one or more loudspeakers, one or more headphones, and one or more earpieces.
17. The device (100a) according to claim 1, realized as at least one of the group consisting of a gaming device, a laptop, a portable audio player, a DVD player, a CD player, a harddisk-based media player, an internet radio device, a public entertainment device, an MP3 player, a hi-fi system, a vehicle entertainment device, a car entertainment device, a video player, a medical communication system, a body-worn device, an audio conference system, a video conference system, and a hearing aid device.
18. A method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
19. A program element, which, when being executed by a processor (103 to 105), is adapted to control or carry out a method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
20. A computer-readable medium, in which a computer program is stored which, when being executed by a processor (103 to 105), is adapted to control or carry out a method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
PCT/IB2007/050375 2006-02-22 2007-02-05 Device for and a method of processing audio data WO2007096792A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP06110307 2006-02-22
EP06110307.3 2006-02-22

Publications (1)

Publication Number Publication Date
WO2007096792A1 true WO2007096792A1 (en) 2007-08-30

Family

ID=38055641

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2007/050375 WO2007096792A1 (en) 2006-02-22 2007-02-05 Device for and a method of processing audio data

Country Status (1)

Country Link
WO (1) WO2007096792A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102007048973A1 (en) * 2007-10-12 2009-04-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal with voice signal processing
CN103262578A (en) * 2010-12-20 2013-08-21 峰力公司 Method for operating a hearing device and a hearing device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1226853A2 (en) * 2001-01-19 2002-07-31 Sony Computer Entertainment Inc. Sound control device and sound control method
JP2005049773A (en) * 2003-07-31 2005-02-24 Denso Corp Music reproducing device
WO2005106843A1 (en) * 2004-04-30 2005-11-10 Axeon Limited Reproduction control of an audio signal based on musical genre classification

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1226853A2 (en) * 2001-01-19 2002-07-31 Sony Computer Entertainment Inc. Sound control device and sound control method
JP2005049773A (en) * 2003-07-31 2005-02-24 Denso Corp Music reproducing device
WO2005106843A1 (en) * 2004-04-30 2005-11-10 Axeon Limited Reproduction control of an audio signal based on musical genre classification

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HAN K-P ET AL: "GENRE CLASSIFICATION SYSTEM OF TV SOUND SIGNALS BASED ON A SPECTROGRAM ANALYSIS", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, IEEE SERVICE CENTER, NEW YORK, NY, US, vol. 44, no. 1, February 1998 (1998-02-01), pages 33 - 42, XP000779248, ISSN: 0098-3063 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102007048973A1 (en) * 2007-10-12 2009-04-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal with voice signal processing
DE102007048973B4 (en) * 2007-10-12 2010-11-18 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal with voice signal processing
US8731209B2 (en) 2007-10-12 2014-05-20 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Device and method for generating a multi-channel signal including speech signal processing
CN103262578A (en) * 2010-12-20 2013-08-21 峰力公司 Method for operating a hearing device and a hearing device
US20130272553A1 (en) * 2010-12-20 2013-10-17 Phonak Ag Method for operating a hearing device and a hearing device
US9363612B2 (en) * 2010-12-20 2016-06-07 Sonova Ag Method for operating a hearing device and a hearing device

Similar Documents

Publication Publication Date Title
JP7389081B2 (en) System and method for providing three-dimensional enhanced audio
US7379552B2 (en) Smart speakers
USRE44611E1 (en) System and method for integral transference of acoustical events
CN101536609A (en) Control of light in response to an audio signal
CN107889033A (en) Space audio for Wave beam forming loudspeaker array is presented
KR20000053152A (en) Multi-channel audio enhancement system for use in recording and playback and methods for providing same
Bartlett et al. On Location Recording Techniques
Francombe et al. Qualitative evaluation of media device orchestration for immersive spatial audio reproduction
US20060050908A1 (en) Personalized surround sound headphone system
JP2013504837A (en) Phase layering apparatus and method for complete audio signal
Bartlett et al. Recording Music on Location: Capturing the Live Performance
Bartlett Stereo microphone techniques
WO2007096792A1 (en) Device for and a method of processing audio data
CN109923877A (en) The device and method that stereo audio signal is weighted
WO2021124906A1 (en) Control device, signal processing method and speaker device
Kim et al. A study on the implementation of immersive sound using multiple speaker systems according to the location of sound sources in live performance
TWI262738B (en) Expansion method of multi-channel panoramic audio effect
JP2024512493A (en) Electronic equipment, methods and computer programs
Brandenburg et al. Audio Codecs: Listening pleasure from the digital world
CN115119133A (en) Sound field assisting method, sound field assisting apparatus, and storage medium
KR20080018409A (en) Web-based 3d sound editing system for 2 channels output
US20010031051A1 (en) Stereo to enhanced spatialisation in stereo sound HI-FI decoding process method and apparatus
Sousa The development of a'Virtual Studio'for monitoring Ambisonic based multichannel loudspeaker arrays through headphones

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 07705792

Country of ref document: EP

Kind code of ref document: A1