WO2007096792A1 - Device for and a method of processing audio data - Google Patents

Device for and a method of processing audio data

Info

Publication number
WO2007096792A1
WO2007096792A1 PCT/IB2007/050375 IB2007050375W WO2007096792A1 WO 2007096792 A1 WO2007096792 A1 WO 2007096792A1 IB 2007050375 W IB2007050375 W IB 2007050375W WO 2007096792 A1 WO2007096792 A1 WO 2007096792A1
Authority
WO
Grant status
Application
Patent type
Prior art keywords
audio
data
reproduction
unit
device
Prior art date
Application number
PCT/IB2007/050375
Other languages
French (fr)
Inventor
Daniel W. E. Schobben
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • H04S3/008Systems employing more than two channels, e.g. quadraphonic in which the audio signals are in digital form, i.e. employing more than two discrete digital channels, e.g. Dolby Digital, Digital Theatre Systems [DTS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control

Abstract

A device (100a) for processing audio data, the device (100a) comprising a classification unit (103) adapted to classify the audio data to belong to an audio class, a reproduction configuration unit (104) adapted to configure a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class, and an audio data processing unit (105) adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).

Description

Device for and a method of processing audio data

FIELD OF THE INVENTION

The invention relates to a device for processing audio data. The invention further relates to a method of processing audio data. Moreover, the invention relates to a program element. Further, the invention relates to a computer-readable medium.

BACKGROUND OF THE INVENTION

Audio playback devices become more and more important. Particularly, an increasing number of users buy audio players comprising multiple loudspeakers and other entertainment equipment.

US 5,812,674 discloses a method for simulating the acoustical quality produced by a virtual sound source and for localizing this source with respect to one or more listeners, and one or more original sound sources. This method performs the steps: 1) fixing values of perceptual parameters defining the acoustical quality to be simulated and values of parameters defining the localization of a virtual source, 2) converting these values into a pulse response described by its energy distribution as a function of the time and the frequency, 3) carrying out a context compensation so as to take account of an existing room effect, 4) obtaining an artificial reverberation from elementary signals so as to achieve a virtual acoustic environment in real time and control the localizing of the virtual source. US 5,784,467 discloses a method of and an apparatus for obtaining acoustic characteristics of sound of a broad frequency range of from 0 to 20 kHz in a relatively short time with high accuracy even though an inexpensive (i.e. a relatively low processing power) computer is used. Multiple sound ray vectors are defined, a virtual space is defined by a polygonal boundary, the propagation history data of the vector is calculated and stored, the vector being reflected at the boundary, and, based on the data, as for each of the vectors, a transient response thereof at an observation point is added to a time-series numerical array and stored, the response being determined on the basis of the reflected vector and a velocity potential determined at the observation point by a micro-area element of the vector defined on the boundary. OBJECT AND SUMMARY OF THE INVENTION

It is an object of the invention to enable efficient audio data processing. In order to achieve the object defined above, a device for processing audio data, a method of processing audio data, a program element, and a computer-readable medium according to the independent claims are provided.

According to an exemplary embodiment of the invention, a device for processing audio data is provided, the device comprising a classification unit adapted to classify the audio data to belong to an audio class, a reproduction configuration unit adapted to configure a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and an audio data processing unit adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit.

According to another exemplary embodiment of the invention, a method of processing audio data is provided, the method comprising classifying the audio data to belong to an audio class, configuring a reproduction characteristics of an audio reproduction unit for reproducing the audio data in accordance with the assigned audio class, and processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit. According to still another exemplary embodiment of the invention, a program element is provided, which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.

According to yet another exemplary embodiment of the invention, a computer- readable medium is provided, in which a computer program is stored which, when being executed by a processor, is adapted to control or carry out a method of processing audio data having the above mentioned features.

The audio processing according to embodiments of the invention can be realized by a computer program, which is by software, or by using one or more special electronic optimization circuits, that is in hardware, or in hybrid form, that is by means of software components and hardware components.

According to an exemplary embodiment of the invention, an audio data processing system is provided in which, in accordance with a class, genre or sound source having generated audio data, playback characteristics of audio content by reproduction units may be set. In this context, the characteristics of both the audio data to be reproduced and audio reproduction units for generating acoustic waves for playing back the audio data are adjusted in accordance with this audio class. In other words, loudspeakers and audio playback parameters may be adjusted so as to match with characteristics of the sound source. For instance, when the music is piano music, the piano music may be reproduced so as to give a human user the impression that the piano is located in the room in which the user listens to the music, for instance in a corner of the room. Thus, a virtual audio source simulation system may be provided.

According to an exemplary embodiment of the invention, a lifelike 3D audio rendering system may be provided. Such a three-dimensional audio rendering system may classify the incoming audio (for instance piano sound, speech or similar) which can be stereo, multi-channel recording or separate sound recordings for the various sound sources such as an in MPEG7 (formally named "Multimedia Content Description Interface", which is a standard for describing the multimedia content data that supports some degree of interpretation of the information meaning, which can be passed onto, or accessed by, a device or a computer code; further information available on http://www.chiariglione.org/mpeg/standards/mpeg-7/mpeg-7.htm) or MPEG21. The audio rendering system may determine the loudspeaker configuration (for instance factory preset, based on user input or using automatic loudspeaker configuration) including parameters like position, orientation and/or radiation of the loudspeakers. The system may obtain, from a look-up table, the desired radiation pattern of the identified source (for instance that of a piano). The system may then process the incoming audio (for instance amplifying, delaying and/or decorrelating the signals), and may play the processed audio over the loudspeaker setup such that the radiation pattern of the sound source may be matched.

The feature of processing the incoming audio data and play back the processed audio data to achieve a matching with a radiation pattern of a sound source may also be denoted or implemented in the context of a wave field synthesis (WFS) system, that is the radiation pattern of a sound using loudspeakers may be reconstructed using loudspeakers.

Embodiments of the invention may be implemented in the context of a system that determines on the reproduction side what the radiation characteristics of the sound sources should be.

According to an exemplary embodiment, it is possible to process a video stream for a videoconference, and the orientation of the sound source (for instance a person speaking) may be changed in case of detecting from the video that a person on the far end turns around. In this context, explicit reference is made to paragraphs [0015], [0027] to [0031] of US 2003/0053680 Al explaining a similar system that may be implemented in the context of an exemplary embodiment of the invention.

Thus, image recognition systems may be implemented which derives audio- related characteristics from image data. For instance, a speaking direction of a person may be detected. In another situation, when a piano is detected on an image, this information may be processed so as to simulate, emulate, recreate or pattern the emission characteristics of a piano with the audio reproduction system.

According to another exemplary embodiment of the invention, the sound- rendering unit together with the audio, such as orientation information, radiation profile, etc, may receive some additional data.

Embodiments of the invention may provide a significantly improved, more realistic audio playback as compared to current home entertainment systems which produce a reasonable sound stage when the loudspeakers are positioned in the way they are assumed to when the audio is mixed and when the listener is exactly in the middle (i.e. sweet spot) of the loudspeakers. Embodiments of the invention may simulate such a situation in a more realistic manner.

According to an exemplary embodiment, a method of processing multichannel audio signals of a surround sound system may be provided, the method comprising classifying an incoming signal yielding classification information. A loudspeaker configuration of the surround sound system may be determined including parameters like position/orientation/radiation of the loudspeakers. From a database or lookup table, desired radiation pattern information of a sound source may be obtained based on the classification information. The incoming audio signal may be processed based on the radiation pattern information, and the processed audio may be played back over the loudspeaker setup such that the radiation pattern of the sound sources is matched.

Exemplary embodiments of the invention may have the advantage that the sound reproduction may take into account instrument "directivity". A virtual piano in the corner of the living room may sound essentially like a real piano, for example, no matter the position of the listener. Exemplary fields of application of embodiments of the invention are home entertainment systems, and systems allowing for adding many compact wireless loudspeaker units can provide a compelling lifelike sound reproduction when implementing embodiments of the invention.

Thus, a lifelike 3D audio rendering system according to an exemplary embodiment may provide a sound reproduction system that restores the radiation pattern of the sound source that is reproduced. A virtual sound source (like a piano or a speaking person) at a particular position of a room (for instance in the corner of a living room) may sound like the real audio source (for example like a real piano), no matter the position of the listener.

Embodiments of the invention may overcome the limitation that current home entertainment systems are mostly limited to emphasize loudspeaker boxes. The evolution has mainly been in increasing the number of channels and the performance is determined by the multi-channel recording/mixing. One important aspect for audio playback, which may be taken into account according to exemplary embodiments of the invention, is the instrument's directivity that may be quite different from that of a loudspeaker. To take into account the instrument's directivity, models may be used derived for sound reduction caption and reproduction. As an example, a piano may be recorded with several microphones, and the sound may be reproduced by a set of loudspeakers carefully arranged to reproduce the piano sound as good as possible. However, in many cases, such a recording is not available and the loudspeaker setup is not known at the time of recording.

It is possible to measure, in a laboratory, the radiation patterns of the sound sources in all classes used in a particular application. Such information may be used for assigning classes to audio content. According to an exemplary embodiment, scaled and delayed versions of the audio signal are played out over the loudspeakers. In a more sophisticated embodiment, such features may be preceded by upmixing the audio into more channels using decorrelation techniques (for instance Lauridsen decorrelation).

Next, further exemplary embodiments of the invention will be explained. In the following, further exemplary embodiments of the device for processing audio data will be described. However, these embodiments also apply for the method of processing audio data, for the computer-readable medium and for the program element.

The classification unit may be adapted to classify the audio data to belong to an audio class based on audio data classification data attached to the audio data. Thus, the device may be capable to read the audio data having some kind of metadata as an extension or attachment, the metadata providing information concerning the class. For instance, such a class may be the genre of the music (jazz, rock, classic music, easy listening, etc.), may provide information concerning the origin of the sound (information concerning the instrument or instruments, about an artist, etc.). However, additionally or alternatively, the classification unit may be adapted to classify the audio data to belong to an audio class based on evaluating characteristics of the audio data itself. Thus, algorithms may be performed which may allow deriving information concerning the audio class from the audio data itself. For instance, the occurrence of beats may be an indication that the sound is rock music. Other characteristics are typical for classical music, jazz, etc. Speech has a completely different sound characteristic than music. The kind of detected instruments (having a specific audio emission characteristics) may allow to conclude on a corresponding audio class.

The classification unit may be adapted to classify the audio data to belong to an audio class based on a comparison of the audio data with pre-stored characteristics of audio data related to (a particular one of) a plurality of audio classes. Thus, predetermined audio class information may be stored in the device. For instance, characteristics for different instruments, vocalist, genres, etc. may be recorded, and corresponding abstract or generic audio class information may be stored in a corresponding database. Then, characteristics of a present audio item may be compared to the pre-stored class profiles, and the profile with the best match may be assumed to represent the class of the present audio item.

The classification unit may be adapted to classify the audio data to belong to an audio class in accordance with an audio source being an origin of the audio data. Thus, the characteristics of the sound source that has generated the audio (for instance a piano, a human singer, etc.) may determine the class like "instrumental", "piano music", etc.

The classification unit may be adapted to classify the audio data to belong to an audio class in accordance with a recording mode of recording the audio data. Examples for such recording modes are stereo, multi-channel, plugged/unplugged, or the like. Thus, the recording mode may form some kind of fingerprint of the audio class or may at least provide hints to which class an audio piece might belong.

The reproduction configuration unit may be adapted to configure reproduction properties of a plurality of audio reproduction units for reproducing the audio data in accordance with the assigned audio class. Thus, embodiments of the invention may implement several audio production units like loudspeakers so that the three-dimensional impression of a three-dimensional sound source may be simulated in a realistic manner. The configuration of the loudspeakers may include a correlation between different loudspeakers, and may also be adjusted based on the relative position of the loudspeakers to one another. It is possible that the loudspeakers automatically detect their positions with respect to one another, for instance by exchanging audio signals. By evaluating the time characteristics of the emission and detection of the audio signals may then allow the system to automatically determine the relative geometrical arrangement of the loudspeakers with respect to one another. Thus, a multiple reproduction unit system may be particularly suitable to realistically simulate the characteristics of a physical sound source. Embodiments of the invention may be used in an audio surround sound system.

The reproduction configuration unit may be adapted to configure, as the reproduction characteristics of the audio reproduction unit, at least one of the group consisting of a spatial position of the audio reproduction unit, a spatial orientation of the audio reproduction unit, and a radiation profile (for instance frequency and/or amplitude characteristics) of the audio reproduction unit. Thus, the positions of the audio reproduction units may be modified during reproduction so as to simulate a motion of the audio source during recording. Also the emission direction may be modified, for instance to simulate that a person giving a speech walks along a lecture hall during speaking and/or changes direction of speech. Also the radiation profile may be adjusted, for instance by adjusting equalizing and/or reverberation properties.

The reproduction configuration unit may be adapted to configure the reproduction characteristics of the audio reproduction unit based on pre-stored configuration information and/or user-defined configuration information. Thus, the configuration of the loudspeakers may be performed based on user commands (for instance a user may input the command "I wish that the content is played back with the acoustic of the Royal Albert Hall" or "I wish that the content is played back to sound like a piano"). Class information may be linked in a database to loudspeaker configuration information. For instance, to simulate a piano, corresponding three-dimensional configuration information may be pre-stored.

The audio data processing unit may be adapted to process the audio data by at least one of the group consisting of amplifying the audio data, delaying the audio data, adjusting a correlation of the audio data, equalizing the audio data, adding reverberation to the audio data, and filtering the audio data. Thus, in order to give a human user a realistic perception or impression of the reproduced audio, the loudness of the audio data may be adjusted, or the signals may be delayed. The signals may be decorrelated, a frequency distribution of the audio data may be adjusted, artificial effects like reverberation or the like may be added so as to simulate a specific audible impression. Moreover, the audio data may be filtered, for instance to remove very low frequency components.

The audio data processing unit may be adapted to process the audio data for generating reproducible audio data for reproduction so as to match a radiation pattern of a sound source being the origin of the audio data. By taking this measure, the acoustic wave emission characteristics of a three-dimensional physical object may be matched.

The device may comprise a database unit (for instance a storage device like a harddisk, an EEPROM, a flash memory, etc.) for storing audio reproduction characteristics data assigned to classes of audio data, wherein the database may be accessed by the reproduction configuration unit and/or by the audio data processing unit. By implementing a database that has pre-stored empiric characteristics of sound sources, the classification and modification of the audio data and the loudspeaker configuration may be significantly improved. The device may be adapted for a dynamic processing of the audio data. In this context, the term "dynamic" may particularly denote that, during reproduction of an audio piece, the reproduction unit configuration and/or the audio data configuration may be continuously updated. For instance, when the sound source changes direction during recording, this direction modification may be mapped on the virtual audio simulation system. The device may comprise the audio reproduction unit(s) for reproducing the reproducible audio data in accordance with the configured reproduction characteristics and in accordance with the assigned audio class. Such one or more audio reproduction units may be loudspeakers, headphones, or earpieces. The reproduction unit or units may be supplied with control information with respect to loudspeaker configuration and audio data modification. The communication between audio processing components of the audio device and reproduction units may be carried out in a wired manner (for instance using a cable) or in a wireless manner (for instance via a WLAN, infrared communication or Bluetooth).

The audio device may be a realized as a gaming device, a laptop, a portable audio player, a DVD player, a CD player, a harddisk-based media player, an internet radio device, a public entertainment device, an MP3 player, a hi-fi system, a vehicle entertainment device, a car entertainment device, a portable video player, a medical communication system, a body-worn device, an audio conference system, a video conference system, or a hearing aid device. A "car entertainment device" may be a hi-fi system for an automobile.

However, although the system according to embodiments of the invention primarily intends to facilitate the playback of sound or audio data, it is also possible to apply the system for a combination of audio data and visual data. For instance, an embodiment of the invention may be implemented in audiovisual applications like a video player in which a loudspeaker is used, or a home cinema system.

The aspects defined above and further aspects of the invention are apparent from the examples of embodiment to be described hereinafter and are explained with reference to these examples of embodiment.

BRIEF DESCRIPTION OF THE DRAWINGS The invention will be described in more detail hereinafter with reference to examples of embodiment but to which the invention is not limited.

Fig. 1 shows an audio data processing device according to an exemplary embodiment of the invention. Fig. 2 shows an audio data processing device according to another exemplary embodiment of the invention.

DESCRIPTION OF EMBODIMENTS

The illustration in the drawing is schematically. In different drawings, similar or identical elements are provided with the same reference signs.

In the following, referring to Fig. 1, an audio data processing system 100 according to an exemplary embodiment of the invention will be explained.

The system 100 shows an audio surround system comprising a device 100a for processing audio data and a sound reproduction unit 100b and a user interface 101. The user interface 101 in the present case is a graphical user interface (GUI) and includes a display device for displaying information to a human operator, which display device is a liquid crystal display in this case. Alternative, the display device may be a plasma display device or the like). Furthermore, the user interface 101 comprises an input device allowing the user to input data (like data specifying the audio reproduction) or to provide the system with the control commands. Such an input device may include buttons, a keypad, a joystick, a trackball, a touch screen, or may even be a microphone of a voice recognition system. The user interface 101 may allow a human user to communicate in a bi-directional manner with the system 100.

The sound reproduction unit 100b comprises loudspeakers 107 to 109 for reproducing audio data. By a user command inputted by a user via the user interface 101, audio data stored in an audio data storage device 102 may be processed for reproduction. The audio data storage device 102 may be a CD, a DVD, or a harddisk.

Audio data AD stored in the audio data content storage unit 102 may be provided to a classification unit 103 for classifying the audio data AD to belong to an audio class (for instance a particular genre or a particular sound source). An output of the classification unit 103 provides classification information CLI and is coupled to an input of a reproduction configuration unit 104, which configuration unit 104 is adapted for processing the classification information CLI to provide configuration instructions COI for configuring reproduction characteristics of the loudspeakers 107 to 109 for reproducing the audio data in accordance with the assigned audio class.

Furthermore, an output of the classification unit 103 is coupled to an input of an audio data processing unit 105 to provide the audio data AD, which audio data processing unit 105 is adapted for processing the audio data AD in accordance with the assigned audio class for generating reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109.

Beyond this, a database unit 106 is provided which stores audio reproduction characteristics data assigned to classes of audio data. As can be taken from Fig. 1, the database unit 106 may be accessed by the reproduction configuration unit 104 and by the audio data processing unit 105 so as to provide these units 104, 105 with information being valuable for these units 104, 105 to correspondingly adapt loudspeaker configuration and audio data properties so as to achieve a proper correspondence between the audio classification and the audio playback.

An output of the reproduction configuration unit 104 is adapted to supply loudspeaker configuration instructions COI to the loudspeakers 107 to 109. Each of the loudspeakers 107 to 109 receives individual reproduction configuration information.

Furthermore, an output of the audio data processing unit 105 is coupled to the loudspeakers 107 to 109 for providing the loudspeakers 107 to 109 individually with reproducible processed audio data PAD for reproduction by the loudspeakers 107 to 109. Therefore, configuration unit 104 controls loudspeaker configuration, and processing unit 105 provides audio content.

Thus, when an audio data item related to piano music is to be played back, the classification unit 103 identifies the class or genre of this audio data to be "piano music" and/or "classical music". Such a determination may be based on appropriate algorithms. For instance, it is possible to distinguish piano music from hardrock by the acoustical properties of such music items. It is also possible that the classification unit 103 evaluates a class identifier which may be attached to the audio data and which may be indicative of the class of the audio item.

The class information CLI and the audio data AD are supplied to the configuration unit 104 and processing unit 105. The reproduction configuration unit 104 adjusts reproduction parameters of the loudspeakers 107 to 109 like an adjustment of the position of the loudspeakers 107 to 109 (which may be modified automatically by means of electric motors or the like), an acoustic wave emission direction of the loudspeakers 107 to 109 (for instance the loudspeakers 107 to 109 may be turned), or the like. Thus, the loudspeakers 107 to 109 are configured to match with an audio pattern of the source of the audio data, in the present example match with the 3D emission characteristics of a piano. Furthermore, the classification is supplied to the audio data processing unit 105 that correspondingly adjusts the audio data so as to match to the characteristics of the sound source. For instance, the audio data is equalized, filtered, amplified or modified in another manner so as to be in proper accordance with characteristics of the original sound source. Thus, the processed audio data may sound, in the present example, like audio emitted from the piano.

Thus, loudspeaker configuration and audio data characteristics are adjusted to match to characteristics of the audio source to give a human user a realistic impression of a virtual piano.

Referring to Fig. 2, an audio data processing system 200 according to an exemplary embodiment of the invention is shown. As a difference of the system 100 according to figure 1 the system 200 according to figure 2 comprises a audio data interface 201 that is adapted to receive audio data from a remote server via wired internet connection. Alternatively the interface 201 may be adapted to receive audio data via a wireless link, for instance WLAN, Bluetooth, ZigBee etc. By a user command inputted by a user via the user interface 101, audio data received by the interface 201 is provided to the classification unit 103. In a further embodiment, instead of audio data processing system 200 the remote server may comprise the classification unit. In this case the audio data processing system 200 may comprise an adapted interface 201a that is adapted to receive the class information CLI and the audio data AD for supplying this information to the configuration unit 104 and processing unit 105. It should be noted that the term "comprising" does not exclude other elements or features and the "a" or "an" does not exclude a plurality. Also elements described in association with different embodiments may be combined.

It should also be noted that reference signs in the claims shall not be construed as limiting the scope of the claims.

Claims

CLAIMS:
1. A device (100a) for processing audio data, the device (100a) comprising a classification unit (103) adapted to classify the audio data to belong to an audio class; a reproduction configuration unit (104) adapted to configure a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; an audio data processing unit (105) adapted to process the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
2. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on audio data classification data attached to the audio data.
3. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on evaluating characteristics of the audio data.
4. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class based on a comparison of the audio data with pre-stored characteristics of audio data related to a plurality of audio classes.
5. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with an audio source being an origin of the audio data.
6. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with a music instrument and/or a human being being an origin of the audio data.
7. The device (100a) according to claim 1, wherein the classification unit (103) is adapted to classify the audio data to belong to an audio class in accordance with a recording mode of recording the audio data.
8. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure reproduction properties of a plurality of audio reproduction units for reproducing the audio data in accordance with the assigned audio class.
9. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure, as the reproduction characteristics of the audio reproduction unit (107 to 109), at least one of the group consisting of a position of the audio reproduction unit (107 to 109), an orientation of the audio reproduction unit (107 to 109), and a radiation profile of the audio reproduction unit (107 to 109).
10. The device (100a) according to claim 1, wherein the reproduction configuration unit (104) is adapted to configure the reproduction characteristics of the audio reproduction unit (107 to 109) based on pre-stored configuration information and/or user-defined configuration information.
11. The device (100a) according to claim 1, wherein the audio data processing unit (105) is adapted to process the audio data by at least one of the group consisting of amplifying the audio data, delaying the audio data, adjusting a correlation of the audio data, equalizing the audio data, adding reverberation to the audio data, and filtering the audio data.
12. The device (100a) according to claim 1, wherein the audio data processing unit (105) is adapted to process the audio data for generating reproducible audio data for reproduction so as to match a radiation pattern of a sound source being the origin of the audio data.
13. The device (100a) according to claim 1, comprising a database unit (106) for storing audio reproduction characteristics data assigned to classes of audio data, the database unit (106) being accessible by the reproduction configuration unit (104) and/or by the audio data processing unit (105).
14. The device (100a) according to claim 1, adapted for dynamically processing the audio data.
15. The device (100a) according to claim 1, comprising the audio reproduction unit (107 to 109) for reproducing the reproducible audio data in accordance with the configured reproduction characteristics and in accordance with the assigned audio class.
16. The device (100a) according to claim 15, wherein the audio reproduction unit (107 to 109) comprises at least one of the group consisting of one or more loudspeakers, one or more headphones, and one or more earpieces.
17. The device (100a) according to claim 1, realized as at least one of the group consisting of a gaming device, a laptop, a portable audio player, a DVD player, a CD player, a harddisk-based media player, an internet radio device, a public entertainment device, an MP3 player, a hi-fi system, a vehicle entertainment device, a car entertainment device, a video player, a medical communication system, a body-worn device, an audio conference system, a video conference system, and a hearing aid device.
18. A method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
19. A program element, which, when being executed by a processor (103 to 105), is adapted to control or carry out a method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
20. A computer-readable medium, in which a computer program is stored which, when being executed by a processor (103 to 105), is adapted to control or carry out a method of processing audio data, the method comprising: classifying the audio data to belong to an audio class; configuring a reproduction characteristics of an audio reproduction unit (107 to 109) for reproducing the audio data in accordance with the assigned audio class; processing the audio data in accordance with the assigned audio class for generating reproducible audio data for reproduction by the audio reproduction unit (107 to 109).
PCT/IB2007/050375 2006-02-22 2007-02-05 Device for and a method of processing audio data WO2007096792A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP06110307.3 2006-02-22
EP06110307 2006-02-22

Publications (1)

Publication Number Publication Date
WO2007096792A1 true true WO2007096792A1 (en) 2007-08-30

Family

ID=38055641

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2007/050375 WO2007096792A1 (en) 2006-02-22 2007-02-05 Device for and a method of processing audio data

Country Status (1)

Country Link
WO (1) WO2007096792A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102007048973A1 (en) * 2007-10-12 2009-04-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal having a speech signal processing
CN103262578A (en) * 2010-12-20 2013-08-21 峰力公司 Method for operating a hearing device and a hearing device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1226853A2 (en) * 2001-01-19 2002-07-31 Sony Computer Entertainment Inc. Sound control device and sound control method
JP2005049773A (en) * 2003-07-31 2005-02-24 Denso Corp Music reproducing device
WO2005106843A1 (en) * 2004-04-30 2005-11-10 Axeon Limited Reproduction control of an audio signal based on musical genre classification

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1226853A2 (en) * 2001-01-19 2002-07-31 Sony Computer Entertainment Inc. Sound control device and sound control method
JP2005049773A (en) * 2003-07-31 2005-02-24 Denso Corp Music reproducing device
WO2005106843A1 (en) * 2004-04-30 2005-11-10 Axeon Limited Reproduction control of an audio signal based on musical genre classification

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HAN K-P ET AL: "GENRE CLASSIFICATION SYSTEM OF TV SOUND SIGNALS BASED ON A SPECTROGRAM ANALYSIS" IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, IEEE SERVICE CENTER, NEW YORK, NY, US, vol. 44, no. 1, February 1998 (1998-02), pages 33-42, XP000779248 ISSN: 0098-3063 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102007048973A1 (en) * 2007-10-12 2009-04-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal having a speech signal processing
DE102007048973B4 (en) * 2007-10-12 2010-11-18 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for generating a multi-channel signal having a speech signal processing
US8731209B2 (en) 2007-10-12 2014-05-20 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Device and method for generating a multi-channel signal including speech signal processing
CN103262578A (en) * 2010-12-20 2013-08-21 峰力公司 Method for operating a hearing device and a hearing device
US20130272553A1 (en) * 2010-12-20 2013-10-17 Phonak Ag Method for operating a hearing device and a hearing device
US9363612B2 (en) * 2010-12-20 2016-06-07 Sonova Ag Method for operating a hearing device and a hearing device

Similar Documents

Publication Publication Date Title
Toole Sound reproduction: The acoustics and psychoacoustics of loudspeakers and rooms
Blauert Communication acoustics
US5764777A (en) Four dimensional acoustical audio system
US6075868A (en) Apparatus for the creation of a desirable acoustical virtual reality
US6590983B1 (en) Apparatus and method for synthesizing pseudo-stereophonic outputs from a monophonic input
US7680288B2 (en) Apparatus and method for generating, storing, or editing an audio representation of an audio scene
US20030007648A1 (en) Virtual audio system and techniques
US20100014692A1 (en) Apparatus and method for generating audio output signals using object based metadata
US20110081024A1 (en) System for spatial extraction of audio signals
US20160330562A1 (en) Calibration of virtual height speakers using programmable portable devices
US20030028273A1 (en) Recording and playback control system
US5784468A (en) Spatial enhancement speaker systems and methods for spatially enhanced sound reproduction
US6931134B1 (en) Multi-dimensional processor and multi-dimensional audio processor system
Spors et al. Spatial sound with loudspeakers and its perception: A review of the current state
US7492907B2 (en) Multi-channel audio enhancement system for use in recording and playback and methods for providing same
WO1999014983A1 (en) Utilisation of filtering effects in stereo headphone devices to enhance spatialization of source around a listener
US20140358567A1 (en) Spatial audio rendering and encoding
US20070025560A1 (en) Audio processing method and sound field reproducing system
US20060023898A1 (en) Apparatus and method for producing sound
US20080212788A1 (en) Recording, Synthesis And Reproduction Of Sound Fields In An Enclosure
US20040131192A1 (en) System and method for integral transference of acoustical events
JP2000354300A (en) Multi-channel audio reproducing device
US20060062401A1 (en) Smart speakers
Avendano et al. Frequency domain techniques for stereo to multichannel upmix
US20150271620A1 (en) Reflected and direct rendering of upmixed content to individually addressable drivers

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct app. not ent. europ. phase

Ref document number: 07705792

Country of ref document: EP

Kind code of ref document: A1