GB2599742A - Personalised audio output - Google Patents

Personalised audio output Download PDF

Info

Publication number
GB2599742A
GB2599742A GB2020147.1A GB202020147A GB2599742A GB 2599742 A GB2599742 A GB 2599742A GB 202020147 A GB202020147 A GB 202020147A GB 2599742 A GB2599742 A GB 2599742A
Authority
GB
United Kingdom
Prior art keywords
audio
processing algorithm
user
audio processing
transducer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
GB2020147.1A
Other versions
GB202020147D0 (en
Inventor
Morrissey Brendan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hears Technology Ltd
Original Assignee
Hears Technology Ltd
Hears Tech Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hears Technology Ltd, Hears Tech Ltd filed Critical Hears Technology Ltd
Priority to GB2020147.1A priority Critical patent/GB2599742A/en
Publication of GB202020147D0 publication Critical patent/GB202020147D0/en
Priority to PCT/EP2021/086433 priority patent/WO2022129487A1/en
Publication of GB2599742A publication Critical patent/GB2599742A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1091Details not provided for in groups H04R1/1008 - H04R1/1083
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/04Circuits for transducers, loudspeakers or microphones for correcting frequency response
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/033Headphones for stereophonic communication
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/04Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments

Abstract

The method involves determining an audio processing algorithm characterising a user’s hearing based on results of a hearing test taken by the user 102, 104, 106. The hearing test may be administered by the user’s playback device or separately administered. The algorithm may be transmitted to either a media service provider or to the user’s playback device 108, 110, 118, 120. Processed audio data is generated by applying the determined audio processing algorithm to audio data from a pre-recorded media file 114, 116, 124. The method further involves outputting, by an audio transducer, audio corresponding to the processed audio data. The processed audio data is independent of a frequency response characteristic of an audio transducer. Systems for outputting audio are also described.

Description

I
PERSONALISED AUDIO OUTPUT
Technical Field
The present invention relates to systems and methods of outputting audio.
Background
Reproduction of audio from recorded data involves many elements in the chain from the recording artist to the consumer. At each stage of the recording process, care is taken to ensure high quality, producing a recording which closely reflects the intention of the recording artist, whether reproducing the sound of a live performance or perhaps including some additional processing to alter the sound at the mastering stage.
Similarly, when recorded music is reproduced, a wide choice of equipment is available to a user, with many equipment purporting to provide a high quality of sound.
Playback devices can have specific physical characteristics to enhance the sound, such as a physical configuration of an audio driver. Playback devices can also apply signal processing to output music, such as selectively boosting particular frequencies known to enhance a particular genre of music.
One area of the reproduction chain for recorded music which is often overlooked, is how a listener's hearing impacts the sound It would be desirable to enhance the quality of audio output from pre-recorded data.
Summary
According to a first aspect of the present invention, there is provided a method of outputting audio. The method comprises: determining an audio processing algorithm characterising a user's hearing based on results of a hearing test taken by the user; generating processed audio data by applying the determined audio processing algorithm to audio data from a pre-recorded media file; and outputting, by an audio transducer, audio corresponding to the processed audio data. The processed audio data is independent of a frequency response characteristic of an audio transducer.
Such a method may enhance the audio quality because the processed audio data is generated using the results of a hearing test taken by the user. The determined audio processing algorithm can therefore adjust the output audio for the particular characteristics of the user's hearing. It has been found that an improvement in sound quality results regardless of the characteristics of audio transducer. This contrasts with a view that the frequency response characteristic of the audio transducer itself should be taken into account in the output processing.
In the present method, the audio processing algorithm characterises a user' s hearing and so audio data is processed to compensate for the user's hearing. For example, frequencies that the hearing test indicates a user shows less sensitivity to may be boosted to provide a flatter equalisation profile to the audio as heard by the user. However, users also have individual preferences for audio reproduction, so might not wish to have a "pure" flat audio experience. As the processing is independent of the response of the transducer, the reproduction can still be influenced by the choice of transducer. A user might choose a pair of headphones to emphasise bass, or which aim to provide as "neutral-reproduction as they wish. A fiirther benefit is that the method can improve perceived audio quality without requiring use of a specific, pre-calibrated audio transducer for the audio reproduction.
In contrast to a hearing aid, for example, the method is applied to pre-recorded audio, not environmental audio. In addition, while a hearing aid will generally focus on restoring hearing response which has been lost, with a focus on amplifying sound at the ear drum, the present invention can improve audio quality for all users, by taking into account the unique characteristics of a user's hearing even without what is conventionally considered to be hearing loss.
The audio transducer is any audio transducer capable of converting electrical signals from the processed audio data into output sound. Example audio transducers include, but are not limited to, loudspeakers and headphones. The personal nature of the processing lends itself to use of headphones, but it can also apply to loudspeakers, especially when the loudspeakers output sound for a single listener.
The audio processing algorithm can be determined by, for example, selectively adjusting individual frequency bands based on the hearing test data. It can be unique to a user, such as determined to give an overall flat, or substantially similar response, across all frequencies as perceived by the user.
The hearing test can be carried out using the same or different audio transducer than the output audio transducer used to output the audio during media playback. In one example, the audio transducer for outputting audio during media playback is a first audio transducer and the hearing test taken by the user uses sound output by a second audio transducer, the second audio transducer having a different frequency response from that of the first audio transducer. In this case, the audio processing algorithm is determined independently of the frequency response of the first audio transducer, and will be based on the frequency response of the second transducer to reproduce audio. As described above, this may nevertheless still enhance the quality of the audio output by the first audio transducer. Further, in this case, the second audio transducer may be specifically for performing the hearing test so that its frequency response is known. This ensures that the audio processing algorithm better characterises the user's hearing. In this example, the second audio transducer may be part of specialist hearing test apparatus.
In another example, the hearing test taken by the user comprises listening to sound output by the same audio transducer that will output the sound. In this case, the hearing test data will be influenced by the frequency response of the audio transducer, so that determined audio processing algorithm may inherently reflect the frequency response characteristic of the audio transducer. In that case, the overall result of the audio processing algorithm can be based on both a user's hearing and the particular output device, which may compensate to at least some extent the inherent frequency response characteristic of the audio transducer, potentially changing the characteristic sound of the audio transducer as well as compensating for user hearing.
The method may further comprise compressing the processed audio data using lossy compression. Applying the audio processing algorithm early in the processing chain can maximise the quality of the delivered processed audio. For instance, applying the audio processing algorithm before compressing the audio data using lossy compression may allow it to operate on higher quality data. Audio data may be compressed with lossy compression for transmission from a media service provider to a playback device and/or for transmission to wireless headphones. In both cases a higher quality version (such as a file without lossy compression, or recorded at a higher bit rate and/or sampling rate) may be available, stored on a media provider server, on a playback device, or other source.
The method may be for use with headphones comprising a left audio transducer and a right audio transducer. In that case, the determined audio processing algorithm comprises a first audio processing algorithm characterising a user's hearing in a left ear and a second audio processing algorithm, different from the first audio processing algorithm, characterising the user's hearing in a right ear based on the results of the hearing test.
Generating processed audio data comprises applying the first audio processing algorithm to audio for output by the left audio transducer and applying the second audio processing algorithm to audio for output by the right audio transducer. For example, a left sound channel is processed differently from a right sound channel. Applying respective audio processing algorithms to audio to be output to each ear further enhances the quality of the audio. It is very likely that a user has a different hearing response between their left and right ears. Thus, applying audio processing algorithms characterising hearing in each ear to the audio data ensures that the processed audio provided to each ear is suited to maximise audio quality.
Some examples allow a user to adjust the extent to which the processing is applied.
In this respect, an input representing an amount by which the audio processing algorithm is to be applied to the audio data may be received. For example, the input may be received from a user interface, such as a graphical user interface, a physical control, and/or a voice input interface. Applying the audio processing algorithm then further comprises scaling the effect of the audio processing algorithm on the audio data in accordance with the amount.
The scaling determines how much of an effect the audio processing algorithm has on the processed audio and allows the user to customise the amount of processing to be applied to audio to be output by the playback device. In an example, a user may wish to dial down the effect of the audio processing algorithm on the audio data, and so selects an amount to reduce its effect.
The scaling may be a factor greater than 1. That is, scaling by a factor of greater than 1 may overcompensate the effect of the audio processing algorithm. This may improve the quality of audio output by particular transducers, for example if the response of the transducer is poor in the frequencies adjusted by the determined audio processing algorithm then it can be useful to apply it to boost the response beyond a nominally flat response. The scaling may not be evenly applied across all frequencies. For example, the scaling may apply to particular bands of frequencies within the audio data. The particular bands of frequencies may be a predetermined selection of bands or may be determined by the user. This, for example, allows a user to increase the effect of mid-range frequencies without overcompensating for the low, bass frequencies. In this tone or equaliser controls can be provided with a response specific to a user's hearing.
Determining the audio processing algorithm may involve determining an adjustment to component frequencies in audio to give a substantially flat hearing response. For example, determining the audio processing algorithm could include analysing the hearing test data and calculating an algorithm to give a desired overall response. Alternatively, determining an audio processing algorithm may involve at least one of selecting from a library of pre-determined algorithms based on the hearing test results. The determining could be carried out locally or remotely. The determining may be done by a computing device coupled to apparatus used to administer the hearing test. In this case, a playback device may receive the audio processing algorithm from the computing device. Alternatively, the determining may be done by the playback device itself This may be the case when the hearing test is administered on the playback device, or where hearing test data is received directly by the playback device.
In some examples, the determining the audio processing algorithm can comprise receiving a predetermined audio processing algorithm from another device, such as a remote computing device which stores it along with a user identifier to identify the user.
Other examples may retrieve an algorithm from local storage.
A flat hearing response may correspond to what was intended when the audio data was mastered in a recording studio. This might be as close to a live performance as possible, or reflect the desire of the recording artist to emphasise a bass line, for example. In the present method, the processed audio is generated independently of the frequency response of the audio transducers. Although it might seem that quality of the audio output is reduced because the effect of the audio transducer on the audio corresponding to the processed audio data is unknown, this is not necessarily the case. For example, a user may choose a particular pair of headphones because they like their inherent sound, whether it is overall "warm', -neutral" or "clear". When output by the audio transducer, the processed audio data will be influenced by the user's choice in headphones but the overall quality can still increase because of the adjustment to the user's personal hearing.
The method may further comprise applying an adjustment to the processed audio before output by the audio transducer. For example, the adjustment may reflect a user's preference in audio output (more bass, more treble etc.). Combining the user's preferred adjustment on top of the audio processing algorithm determined from the hearing test may enhance the overall perceived output quality, the adjustment is applied to audio specific to a user so takes into account both personal preference and the user's personal hearing response.
Generating the processed audio data may comprise applying the determined audio processing algorithm to audio data by an operating system of a playback device coupled to the audio transducer. Applying the audio processing algorithm at the level of the operating system ensures that all audio output by the audio transducer will be processed according to the audio processing algorithm. Audio output associated with media applications on the playback device will thus all benefit from the audio processing algorithm, independent of support in any particular application. In addition, if a media application allows a user to adjust the processing of the audio data, for example by adjusting the audio data in a range of frequencies, then the output audio may be further enhanced because it is based on a hearing test taken by the user and further adjusted according to the user's preference. Likewise, alternatively or additionally, the operating system itself can offer further equalisation options for a user, applied in addition to the determined audio processing algorithm. For example, the operating system might offer an option for "personalised sound" which corresponds to the audio processing algorithm and further options, including pre-set equalisations, such as "Bass Boost", "Classical", "Spoken Word", etc -and/or tone control or graphic equaliser type adjustments.
Generating the processed audio data may comprise applying the audio processing algorithm to audio data by a media service provider. The method further comprises receiving the processed audio data from the media service provider by a playback device coupled to the audio transducer. In this case, the media service provider can deliver personalised audio independently of the playback device. This ensures that the user has access to enhanced audio across all their devices accessing the media service provider, and further does not require a change of hardware or software to any of the devices in order to play the processed audio. Another benefit is a reduction in the amount of processing required by the playback device because the processing is done by the media service provider before it is received at the playback device. Quality may also be enhanced when the media service provider has access to a high-quality file to apply the processing to before applying lossy compression.
According to a second aspect of the present invention, there is provided a playback system. The system comprises a playback device and an audio transducer coupled to the playback device. The playback device is configured to perform the method according to the first aspect of the present invention, with or without the optional features also discussed. Such as system may provide enhanced quality of audio output as discussed above.
According to a third aspect of the present invention, there is provided a media service provider server configured to retrieve an audio processing algorithm characterising a user's hearing, wherein the audio processing algorithm is based on results of a hearing test taken by the user. The server is further configured to generate processed audio data by applying the determined audio processing algorithm to audio data from a pre-recorded media file, and to transmit the processed audio data to a playback device coupled to an audio transducer for output of audio.
The audio processing algorithm characterising a user's hearing may be received from a hearing test apparatus and stored for later retrieval. For example, a user may take a hearing test to generate hearing test results, from which an audio processing algorithm is determined. The audio processing algorithm may then be transmitted to the media service provider server. The audio processing algorithm can be stored by the media service provider server and applied to media files on the server-side, before transmission to a user. The user then has access to audio content processed by the algorithm across all their playback devices, without requiring specific software or hardware on the device itself. This also reduces the amount of processing required by the playback devices because the processing is done by the media service provider before it is received at the playback device.
In another example, the audio processing algorithm may be retrieved from the playback device by the media service provider. This may enhance the privacy of the hearing test data and/or the determined audio processing algorithm. For example, the user can then determine whether the media service provider service has permission to access the hearing test data and/or the determined audio processing algorithm. If permission is granted, the playback device may transmit the determined audio processing algorithm to the media service provider server, wherein it can be applied to audio data from a prerecorded media file to generate the processed audio data. The processed audio data can then be transmitted back to the user's playback device for output After use, the determined audio processing algorithm can be discarded or deleted by the media service provider server.
According to a fourth aspect of the present invention, there is provided a system comprising a hearing test apparatus and a computing device. The hearing test apparatus is configured to generate hearing test data from a hearing test taken by a user The computing device is coupled to the apparatus and configured to receive the hearing test data. The computing device determines, based on the received hearing test data, an audio processing algorithm characterising the user's hearing, and cause the determined audio processing algorithm to be stored for later application to pre-recorded media files.
Such a system is useful in a retail location. Users may visit the retail location to take a hearing test and cause a determined audio processing algorithm to be stored.
For example, the hearing test can be offered as a service to load a determined audio processing algorithm into the playback device. In this context, the computing device may be a playback device belonging to the user, such as the playback device according to the second aspect. Alternatively, the computing device may be configured to transmit the determined audio processing algorithm to a playback device, and the playback device is configured to store the determined audio processing algorithm.
In another example, the service could be offered by media service provider to enable a hearing test to be taken at a convenient location. In this case, the computing device may be a media service provider server, such as the media service provider server of the third aspect. Alternatively, the computing device may be configured to transmit the determined audio processing algorithm to the media service provider server together with a user identifier, and the media service provider server is configured to store the determined audio processing algorithm in association with the user identifier. Example user identifiers include usemames, credentials, a globally unique identifier, an IMEI number of a mobile device, a telephone number and so on.
In a further example, the computing device may offer a choice of whether to transmit the determined audio processing algorithm or one or both a playback device and a media provider server.
The hearing test apparatus may comprise an audio transducer configured to output predefined audio to the user and receive input from the user on audible tones. The user's input in response to the audio may then be used to generate the hearing test data. The hearing test may involve playing multiple test sounds at specific frequencies and asking if the user heard the sounds. The frequency content of the sounds used in the hearing test may then be utilised to determine at least one characteristic of the audio processing algorithm. The user may wish to listen to audio output using a first audio transducer coupled to the playback device. The audio transducer of the hearing test apparatus may then be a second, different audio transducer comprising a different frequency response characteristic from that of the first audio transducer. In this way, the audio processing algorithm is determined independently of the frequency response of the first audio transducer. Applying the determined audio processing algorithm to audio data from a pre-recorded media file improves the perceived quality of reproduction by the first audio transducer even though the frequency response of the first audio transducer was not considered when the audio processing algorithm was determined.
In another example, the hearing test apparatus comprises the same audio transducer that is coupled to the playback device, and used to output audio by the user. This could be the same audio transducer, or another audio transducer of the same make and model. In the shop example discussed above, this may involve a user bringing their preferred pair of headphones with them to the shop and coupling the headphones to the system. When the hearing test apparatus comprises the same audio transducer as coupled to a playback device for subsequent audio output, the audio processing algorithm will indirectly take into account the frequency response characteristic of the headphone, via the hearing test, even though it is not measured directly or taken into account explicitly when the audio processing algorithm is determined. In this example, a user may take a different hearing test with each audio transducer, such as for each pair of headphones owned by a user, or when a new pair of headphones is obtained.
According to a fifth aspect of the present invention, there is provided a computer readable medium comprising instructions for execution by a processor. The instructions, when executed, cause the processor to determine an audio processing algorithm characterising a user's hearing based on results of a hearing test taken by the user, generate processed audio data for output by an audio transducer by applying the determined audio processing algorithm to audio data from a pre-recorded media file without considering a frequency response characteristic of the audio transducer; and cause audio corresponding to the processed audio data to be output by the audio transducer.
The instructions may be associated with a software application on a playback device. In this case, the playback device comprises the processor. The software application may be associated with a media service provider, such as Spotify®. The user may interact with the software application via a user interface of the playback device. The user interface can be utilised to take inputs from the user. The inputs may correspond to a playback control command (play, pause, volume change etc.) or may represent an amount by which the audio processing algorithm is to be applied to the audio data.
In other examples, the instructions may be associated with an operating system of a playback device, so that the determined audio processing algorithm can be applied to output from the playback device irrespective of the source of the audio data for output, so that there is no need to use a particular client application.
The pre-recorded media file may be stored in memory of the playback device.
Alternatively, the pre-recorded media file may be received from a media service provider server of the media service provider associated with the software application. In either case, the processed audio data is generated within the software application associated with the computer readable medium.
The computer readable medium can be a tangible, non-transitory computer-readable medium, such as solid-state memory, optical disc or magnetic disc.
Throughout the above-described aspects, reference to elements "coupled" to each other includes both direct and indirect connections, for example a point-to-point or a connection via another device. Similarly, the coupling may be via a wired or wireless connection, including connection via IEEE 802.11 ("Wi-Fi") and/or Bluetooth protocols.
Further features and advantages of the invention will become apparent from the following description of preferred embodiments of the invention, given by way of example only, which is made with reference to the accompanying drawings.
Brief Description of the Drawings
Figure 1 shows a flow diagram of a method for outputting audio according to an embodiment; Figure 2 shows a schematic diagram of a system for outputting audio according to an embodiment; Figure 3 shows a schematic diagram of a system for outputting audio according to a further embodiment; Figure 4 shows a schematic diagram of a computing device according to an embodiment; Figure 5 shows a schematic diagram of a media service provider server according to an embodiment; Figure 6 shows a example user interface according to an embodiment; and Figure 7 shows another example user interface.
Detailed Description
Figure 1 is a flow diagram of a method 100 for outputting audio. The audio is to be output by an audio transducer coupled to a playback device. The audio transducer is any audio transducer capable of converting electrical signals into audio corresponding to the audio data in the electrical signals. Examples of audio transducers include, but are not limited to, loudspeakers and headphones.
At block 102, the method 100 involves administering a hearing test. The hearing test is administered using an audio transducer. The audio transducer used to administer the hearing test may differ from the audio transducer coupled to the playback device and used to output subsequent audio. A different audio transducer is one that is not the same audio transducer used for subsequent reproduction or one of different make and model from the audio transducer used for subsequent reproduction. For example, the audio transducer used to administer the hearing test may be part of hearing test apparatus provided at a retail location.
Alternatively, the hearing test at block 102 may be administered by the playback device itself, in which case it may be more likely that the audio transducer used to administer the hearing test is the same as the audio transducer used to output subsequent audio. However, a user may own, and use, more than one audio transducer, so it may be possible that the audio transducer used to administer the hearing test differs from the audio transducer used to output subsequent audio.
In one example, the hearing test involves playing multiple test sounds at specific frequencies and asking if the user heard the sounds. It will be appreciated that the precise nature of the hearing test may vary and any form of hearing test may be used. For example, a hearing test apparatus may measure an electrical response in the brain to stimulus sounds.
At block 104, the method 100 involves generating hearing test data based on the hearing test administered at block 102. For example, hearing test data may be expressed as a sensitivity to sounds of particular frequencies or frequency bands. It will be appreciated that other types of hearing test data can be generated as the result of the user taking a hearing test.
At block 106, at least one audio processing algorithm based on the hearing test data is determined based on the hearing test data. This may comprise determining an adjustment to frequencies in order to give a substantially flat hearing response, where different sensitivities between frequencies and/or frequency bands are compensated so that the overall sensitivity across the frequencies is substantially the same. Such a flat hearing response enables audio to be perceived as intended when the audio data was generated, from a live performance and/or the mastering process.
For example, if the hearing test data indicates that a user is not sensitive to frequencies in a particular range of frequencies, then the audio processing algorithm may be determined to boost audio data in those particular frequencies when applied to the audio data. The determining may involve receiving or calculating the audio processing algorithm.
For example, determining the audio processing algorithm could include analysing the hearing test data and calculating an algorithm to give a desired overall response. Alternatively, determining an audio processing algorithm may involve selecting from a library of pre-determined algorithms based on the hearing test results, receiving a predetermined algorithm from another device, or retrieving an algorithm from storage.
In one example, the hearing test may comprise data specific to a left ear and a right ear. In that case, a first audio processing algorithm characterising a user's hearing in a left ear and a second audio processing algorithm, different from the first audio processing algorithm, characterising a user's hearing in a right ear can be determined.
At block 108, it is determined whether to transmit the determined audio processing algorithm to a media service provider. The determination may be based on a user preference, for example, whether the user wishes to share their audio processing algorithm with the media service provider. In another example, the determination may be done automatically by the computing system. This may occur when the hearing test was administered on the playback device, in which case, the audio processing algorithm may be stored on the playback device only. This could also occur when there is no option to store the audio processing algorithm on the playback device, in which case the determined audio processing algorithm is always transmitted to the media service provider.
If it is determined that the audio processing algorithm is to be transmitted to the media service provider, the method 100 proceeds to block 110, wherein the audio processing algorithm is transmitted to a server associated with the media service provider.
The audio processing algorithm may be transmitted via a wired or wireless communication interface, such as Ethernet, Wi-Fi and/or cellular data including 3G, 4G and SG cellular standards.
At block 112, the method 100 involves adjusting an amount by which the at least one audio processing algorithm is to be applied to the audio data. For example, a user interface may be implemented in a display of the playback device. The user interface allows a user to input the amount by which the audio processing algorithm is applied to the data. This could be through a slide level, toggles or radio buttons. The numerical choice may be between 1-10 or 1-100, although it should be appreciated that other examples are possible. Based on received input from the playback device of the amount of adjustment, the effect of the audio processing algorithm on the audio data is scaled in accordance with the amount.
The scaling determines how much of an effect the audio processing algorithm has on the processed audio and allows the user to customise the amount of processing to be applied to audio to be output by the playback device. In an example, a user may wish to dial down the effect of the audio processing algorithm on the audio data, and so selects an amount to reduce its effect.
The scaling may be a factor greater than 1. For example, an audio processing algorithm may be determined to boost a power of audio data at particular frequencies by a particular percentage when applied without scaling. Taking the example of a 10% boost at 262 Hz, corresponding to the note middle C, applying the audio processing algorithm with a scaling by a factor of greater than 1 in this case means that the power of the audio data at 262 Hz is boosted by more than 10%. That is, scaling by a factor of greater than 1 may overcompensate for the effect of a user's hearing response. This may improve perceived quality by giving a more "dynamic sound" and/or help to overcome limitations in the frequency response of a particular output audio transducer. The scaling may not be evenly applied across all frequencies. For example, the scaling may apply to particular bands of frequencies within the audio data. The particular bands of frequencies may be a predetermined selection of bands, or may be determined by the user. This, for example, allows a user to increase the effect of mid-range frequencies without overcompensating for the low, bass frequencies.
At block 114, the method 100 involves generating, by the media service provider server, processed audio data. The generating comprises applying the determined audio processing algorithm to audio data from a pre-recorded media file corresponding to audio to be played by an audio transducer. The pre-recorded media file may be in any audio file format, but preferably is provided in an uncompressed or lossless audio format, such as Waveform Audio File Format (WAV), Au format, or Free Lossless Audio Codec (FLAC). The pre-recorded media file may be an audio stream embedded within or associated with a corresponding video file.
This method of generating processed audio data means that the processed audio is generated without consideration of the effect that the audio transducer has on the processed audio data when it is output. Some playback systems seek to improve audio quality by attempting to make output of an audio transducer as neutral as possible, such as by compensating or otherwise removing a frequency response characteristic of the audio transducer from the output audio. However, this requires first determining the frequency response characteristic of the audio transducer, and so typically can only be done on playback systems that have been sufficiently calibrated. A limited number of supported audio transducers can be used. Surprisingly, it has been found that an increase in perceived audio quality is achieved by applying an audio processing algorithm based on a hearing test taken by the user without considering the response of the audio transducer reproducing he audio.
At block 116, the method 100 involves transmitting the audio data to the playback device. As discussed with regards to block 110, the transmitting may be via a wired or wireless network communication interface. For example, both the media service provider server and the playback device may comprise a network interface thorough which they can exchange data. The transmitting may first involve compressing the processed audio data using a lossy compression scheme and /or applying further processing procedures on the audio data. Applying the audio processing algorithm earlier in the processing chain (at block 114) can maximise the quality of the delivered processed audio. For instance, applying the audio processing algorithm before compressing the audio data using lossy compression allows the audio processing algorithm to operate on as much data as possible, rather than adjusting areas which may already have lost detail due to the lossy compression process.
If, at block 108, it is determined that the audio processing algorithm is not to be transmitted to a media service provider, the method 100 proceeds to block 118, wherein it is determined whether the hearing test was administered on the playback device. If it is determined that the hearing test was not administered on the playback device, the method proceeds to block 120, wherein the at least one audio processing algorithm is transmitted to the playback device. The transmitting may be via a wired or wireless communication network interface. The method 100 then proceeds to block 122.
If, on the other hand, it is determined that the hearing test was administered on the playback device, then the method 100 proceeds straight to block 122. M this case, the hearing test data is generated on the playback device, and the at least one audio processing algorithm is determined by the playback device. Thus, the audio processing algorithm does not need to be transmitted to the playback device.
At block 122, the method 100 involves adjusting an amount by which the at least one audio processing algorithm is to be applied to the audio data. The adjusting is done in the same way as discussed above at block 112. However, at block 122, the adjusting is performed at the playback device.
At block 124, the method 100 involves generating, by the playback device, processed audio data independently of a frequency response of the audio transducer coupled to the playback device. The generating at block 124 is done in the same wat as discussed above at block 114. However, at block 124, the generating is done by the playback device rather than the media service provider server.
At block 126, the method 100 comprises outputting audio corresponding to the processed audio by the audio transducer. After the processed audio is received by the playback device after the transmission at block 116, or the processed audio is generated on the playback device at block 124, the audio is output by an audio transducer coupled to the playback device. The audio transducer is configured to convert electrical signals corresponding to the processed data into mechanical vibrations perceivable as sound.
The method 100 may include one or more operations, functions, or actions as illustrated by one or more of blocks 102-126. Although the series of blocks 102-126 are illustrated in their respective sequential orders, the respective series of blocks may also be performed in parallel, and/or in a different order than those described herein. Also, the various blocks may be combined into fewer blocks, divided into additional blocks, and/or removed based upon the desired implementation. For example, the method may begin by determining at least one audio processing algorithm based on hearing data. The hearing data, in this case, is predetermined and provided to the playback device. This may be the case if the user switches to a new device, but has already undertaken a hearing test. In this case, blocks 102 and 104 may not be part of the method 100. In another example, adjusting an amount by which the audio processing algorithm is to be applied may not be required. In this case, the blocks 112 and 122 may not be present in the method 100. In some cases, block 118 may be omitted and either block 120 or block 122 included, depending on where the method is executed.
Figure 2 shows a schematic diagram of an example system 200 for outputting audio according to an embodiment. The system 200 comprises hearing test apparatus 202. The hearing test apparatus 202 is configured to administer a hearing test. The hearing test apparatus 202 is coupled to an audio transducer 203. The audio transducer 203 is configured to output predetermined audio according to the hearing test. The hearing test apparatus 202 may also comprise an interface to measure a user's responses to the hearing test. The responses result in hearing test data. Examples of hearing tests and associated hearing test data are discussed further with regards to block 102 and 104 of the method 100 shown in Figure 1.
The system 200 further comprises a computing device 204, coupled to the hearing test apparatus 202. The computing device 204 receives hearing test data generated by the hearing test apparatus 202 and is configured to determine at least one audio processing algorithm characterising a user's hearing based on results of a hearing test taken by the user in accordance with block 106 of the method 100 shown in Figure 1.
The system 200 further comprises a playback device 206 coupled to the computing device 204. The playback device 206 is configured to receive the determined audio processing algorithm from the computing device 204 via the coupling. The coupling may be a wired connection, such as via a Universal Serial Bus (USB) connection or a Thunderbolt interface. Alternatively, the coupling may be wireless, such as through wireless interfaces on the computing device 204 and the playback device 206.
The playback device 206 is configured to generate processed audio data by applying the audio processing algorithm to audio data from a pre-recorded media file. Generating the processed audio data is performed independently of a frequency response characteristic of the audio transducer 208 yet still improves the perceived quality because the output sound is personalised to a user.
Figure 3 shows an example system 300 for outputting audio according to a further embodiment. The system 300 comprises hearing test apparatus 302. The hearing test apparatus is coupled to an audio transducer 303. The hearing test apparatus 302 and audio transducer 303 are the same as the hearing test apparatus 202 and audio transducer 203 in the system 200 discussed with reference to Figure 2.
The system 300 further comprises a computing device 304. Unlike the computing device 204 in the system 200, the computing device is coupled to a media service provider server 306. The media service provider server 306 is configured to receive the at least one audio processing algorithm from the computing device 304.
The system 300 further comprises a playback device 308, coupled to the media service provider server 306. Through the coupling, the media service provider server 306 communicates with the playback device 308. The user may interact with the playback device 308, via a user interface provided in an application of the media service provider to select a piece of music to play. This causes the playback device 308 to send a request to the media service provider server 306. On receiving the request for the piece of music, the media service provider server retrieves the audio data corresponding to the piece of music and generates processed audio data by applying the determined audio processing algorithm. For example, this can be done as described above with reference to block 114 of the method 100 shown in Figure 1. The media service provider server 306 then transmits the processed audio data to the playback device 308 to be output by an audio transducer 310, coupled to the playback device 308.
Figure 4 shows an example playback device 400 which may be used in embodiments. The playback device 400 comprises an operating system 402 and a client application 404. The operating system 402 can interact with the client application 404 via an Application Program Interface (API) 406. The client application 404 may be associated with a media service provider. For example, the client application 404 may be Apple® Music which is associated with Apple® Inc. Example playback devices in mobile devices running the Android or iOS operating systems, and computing devices running the Windows, Mac OS or Linux operating systems.
In one example, the audio processing is applied by an operating system of the playback device 400 independent of the client application. A user of the computing device 400 interacts with the client application 404 using a user interface (not shown). For example, the user selects a piece of music for playback via the client application 404. When the piece of music is selected, the client application 404, via the API 406, sends a request to the operating system 402 to output the audio, resulting in the operating system carrying out the method 408. At block 410, the method 408 involves receiving an audio processing algorithm based on results of a hearing test taken by the user. The audio processing algorithm may be received from a remote device or retrieved from memory of the playback device 400. At block 412, the operating system 402 generates processed audio data. The generating may be done in accordance with block 124 in the method 100 shown in Figure 1. At block 414, the operating system 402 provides the processed audio data for playback. The processed audio data may then be output by the audio transducer coupled to the playback device 400.
In another example, the processing of method 408 may be carried out by the client application, so that it is limited to audio reproduced via client application and is not applied to all audio output from the playback device.
Figure 5 shows an example media service provider sewer 500 according to an embodiment. The media service provider server 500 comprises a processor configured to perform a method 502. At block 504, an audio processing algorithm based on results of a hearing test taken by a user is retrieved. At block 506, the method 506 involves generating processed audio data. The generating may be done in accordance with block 114 of the method 100 shown in Figure 1. At block 508, the method 502 involves transmitting the processed audio data to the playback device for playback in accordance with block 116 of the method 100 shown in Figure 1. For example, the processed audio data may be transmitted to a client application on the playback device.
Figure 6 shows an example a user interface 600 for displaying results of a hearing test taken by a user. The hearing test may be done in accordance with block 102 of the method 100 shown in Figure 1. The user interface 600 may be associated with the client application 402 of the computing device 400 shown in Figure 4 or with an operating system.
The user interface 600 shows hearing test data indicating a user's sensitivity to audio at particular frequencies. The hearing test data may be generated in accordance with block 104 of the method 100 shown in Figure 1. In particular, the hearing test data comprises hearing test data corresponding to a user's left ear 602 and hearing test data corresponding to a user's right ear 604. In this example, the hearing test data is represented as an audiogram that indicates a level of hearing loss expressed in decibels.
The hearing test data for each ear 602, 604 is averaged to determine a level of hearing loss of the user in each ear. This information is displayed in a textbox 606 allowing the user to see an approximation of their hearing loss in each ear. An audio processing algorithm can be determined based on the hearing test data 602, 604. In one example, the audio processing algorithm may be determined to boost frequencies that the user shows lower sensitivity to (expressed as a hearing loss in the audiogram), so that applying the audio processing algorithm to audio data results in a flatter equalisation profile, as perceived by the user.
Figure 7 shows an example a user interface 700 for controlling an amount by which a determined audio processing algorithm is to be applied to audio data. The user interface 700 includes a mood wheel or circle 702. The amount by which the audio processing algorithm is to be applied to the audio data is represented by a location of a pointer 704 within the mood wheel 702. Movement of the pointer 704 within the radial direction represents a magnitude of the amount by which the determined audio processing unit is to be applied, while rotational movement of the pointer 704 about the centre of the mood wheel 702 represents a change in the frequency content of audio that is to be adjusted. For example, a position of the pointer 704 towards the top of the mood wheel 702 may be associated with applying an adjustment of the audio processing algorithm to so that higher frequency bands in the audio processing algorithm are adjusted by a greater proportion than lower frequency bands in the audio processing algorithm. The position of the pointer 704 towards the bottom of the mood wheel 702 may be associated with an adjustment of the audio processing algorithm so that lower frequency bands in the audio processing algorithm are adjusted by a greater proportion than higher frequency bands in the audio processing algorithm. A position of the pointer on a horizontal line passing through the centred of the circle 702 may correspond to a uniform scaling of the audio processing algorithm across all frequency bands The amount by which the audio processing algorithm is to be applied to the audio data does not need to linearly scale with the position of the mood wheel 702 along the radial direction. For example, there may be a logarithmic mapping between the position of the pointer 702 and the amount by which the audio processing algorithm is applied to the audio data.
In certain examples, the pointer 704 can be positioned beyond the radius of the mood wheel 702, providing an overcompensation of the audio processing algorithm (i.e. a scaling by a factor greater than one), which can emphasise the effect. In other examples, a position of the pointer beyond a predetermined radial distance from the centre of the circle 702 can be interpreted as overcompensation, scaling the audio processing algorithm by a factor greater than one.
The user interface 700 also comprises a set of radio buttons 706, representing preset adjustments of the amount by which the audio processing algorithm is to be applied. Each radio button of the set of radio buttons corresponds to a predetermined position of the pointer 704 within the mood wheel 702. For example, as depicted in Figure 7 selection of a "Folk" present can set the audio processing algorithm to levels chosen to work with Folk music The pointer 704 is moved to an appropriate position on the mood wheel 702. The above embodiments are to be understood as illustrative examples of the invention. Further embodiments of the invention are envisaged. It is to be understood that any feature described in relation to any one embodiment may be used alone, or in combination with other features described, and may also be used in combination with one or more features of any other of the embodiments, or any combination of any other of the embodiments. Furthermore, equivalents and modifications not described above may also be employed without departing from the scope of the invention, which is defined in the accompanying claims.

Claims (18)

  1. CLAIMSA method of outputting audio comprising: determining an audio processing algorithm characterising a user's hearing based on results of a hearing test taken by the user; generating processed audio data by applying the determined audio processing algorithm to audio data from a pre-recorded media file; and outputting, by an audio transducer, audio corresponding to the processed audio data; wherein the processed audio data is independent of a frequency response characteristic of an audio transducer.
  2. 2. The method according to claim 1, wherein the audio transducer is a first audio transducer, and wherein the hearing test taken by the user uses sound output by a second audio transducer, the second audio transducer having a different frequency response characteristic from that of the first audio transducer.
  3. 3. The method according to claim 1, wherein the hearing test taken by the user uses sound output by the same audio transducer that will output the sound
  4. 4. The method according to any preceding claim, further comprising compressing the processed audio data using lossy compression.
  5. 5. The method according to any preceding claim for use with headphones comprising a left audio transducer and a right audio transducer, wherein: the determined audio processing algorithm comprises a first audio processing algorithm characterising a user's hearing in a left ear and a second audio processing algorithm, different from the first audio processing algorithm, characterising the user's hearing in a right ear based on the results of the hearing test; and the generating processed audio data comprises applying the first audio processing algorithm to audio for output by the left audio transducer and applying the second audio processing algorithm to audio for output by the right audio transducer.
  6. 6 The method according to any preceding claim, comprising: receiving input representing an amount by which the determined audio processing algorithm is to be applied to the audio data; and wherein applying the audio processing algorithm further comprises scaling the effect of the audio processing algorithm on the audio data in accordance with the amount.
  7. 7 The method according to claim 6, wherein the scaling is by a factor greater than 1.
  8. 8. The method according to any preceding claim, wherein determining the audio processing algorithm comprises determining an adjustment to component frequencies in audio to give a substantially flat hearing response.
  9. 9. The method according to any preceding claim, comprising applying an adjustment to the processed audio before output by the audio transducer.
  10. 10. The method according to any preceding claim, wherein generating the processed audio data comprises applying the determined audio processing algorithm to audio data by an operating system of a playback device coupled to the audio transducer.
  11. 1 1. The method according to any of claims Ito 9, wherein: generating the processed audio data comprises applying the determined audio processing algorithm to audio data by a media service provider; and the method comprises receiving the processed audio data from the media service provider by a playback device coupled to the audio transducer.
  12. 12. A playback system comprising: a playback device; and an audio transducer coupled to the playback device, wherein the playback device is configured to perform the method of any of claims I to 10.
  13. 13. A media service provider server configured to: retrieve an audio processing algorithm characterising a user's hearing, wherein the audio processing algorithm is based on results of a hearing test taken by the user; generate processed audio data by applying the determined audio processing algorithm to audio data from a pre-recorded media file; and transmit the processed audio data to a playback device coupled to an audio transducer for output of audio.
  14. 14. The media service provider server of claim 13, wherein the audio processing algorithm is retrieved from the playback device. 10
  15. 15. A system comprising: a hearing test apparatus configured to generate hearing test data from a hearing test taken by a user; and a computing device coupled to the hearing test apparatus and configured to: receive the hearing test data; determine, based on the received hearing test data, an audio processing algorithm characterising the user's hearing; and cause the determined audio processing algorithm to be stored for later application to pre-recorded media files.
  16. 16. The system of claim 15, further comprising a playback device according to claim 12, wherein the computing device is configured to transmit the determined audio processing algorithm to the playback device, and the playback device is configured to store the determined audio processing algorithm.
  17. 17. The system of claim 15 or 16, further comprising a media service provider server according to claim 13, wherein the computing device is configured to transmit the determined audio processing algorithm to the media service provider server together with a user identifier, and the media service provider server is configured to store the determined audio processing algorithm in association with the user identifier.
  18. 18. A computer readable medium comprising instructions for execution by a processor, wherein the instructions, when executed, cause the processor to: determine an audio processing algorithm characterising a user's hearing based on results of a hearing test taken by the user; generate processed audio data for output by an audio transducer by applying the determined audio processing algorithm to audio data from a pre-recorded media file without considering a frequency response characteristic of the audio transducer; and cause audio corresponding to the processed audio data to be output by the audio transducer.
GB2020147.1A 2020-12-18 2020-12-18 Personalised audio output Pending GB2599742A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
GB2020147.1A GB2599742A (en) 2020-12-18 2020-12-18 Personalised audio output
PCT/EP2021/086433 WO2022129487A1 (en) 2020-12-18 2021-12-17 Personalised audio output

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB2020147.1A GB2599742A (en) 2020-12-18 2020-12-18 Personalised audio output

Publications (2)

Publication Number Publication Date
GB202020147D0 GB202020147D0 (en) 2021-02-03
GB2599742A true GB2599742A (en) 2022-04-13

Family

ID=74221080

Family Applications (1)

Application Number Title Priority Date Filing Date
GB2020147.1A Pending GB2599742A (en) 2020-12-18 2020-12-18 Personalised audio output

Country Status (2)

Country Link
GB (1) GB2599742A (en)
WO (1) WO2022129487A1 (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020068986A1 (en) * 1999-12-01 2002-06-06 Ali Mouline Adaptation of audio data files based on personal hearing profiles
WO2003026349A1 (en) * 2001-09-20 2003-03-27 Sound Id Sound enhancement for mobile phones and other products producing personalized audio for users
GB2380917A (en) * 2001-06-30 2003-04-16 Hewlett Packard Co An audio player with personalised audio profiles
US20130182855A1 (en) * 2012-01-13 2013-07-18 Samsung Electronics Co., Ltd. Multimedia playing apparatus and method for outputting modulated sound according to hearing characteristic of user
US20140194775A1 (en) * 2010-08-05 2014-07-10 Ace Communications Limited Method and System for Self-Managed Sound Enhancement
US20190045302A1 (en) * 2017-08-07 2019-02-07 Leigh M. Rothschild System and Method For Adjusting Audio Parameters for a User
US20200027467A1 (en) * 2018-07-20 2020-01-23 Mimi Hearing Technologies GmbH Systems and methods for encoding an audio signal using custom psychoacoustic models
WO2020019020A1 (en) * 2018-07-23 2020-01-30 Audeara Pty Ltd Calibration method for customizable personal sound delivery systems

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150281853A1 (en) * 2011-07-11 2015-10-01 SoundFest, Inc. Systems and methods for enhancing targeted audibility
JP2015177392A (en) * 2014-03-17 2015-10-05 ヤマハ株式会社 Signal processing apparatus and equalizer module
NL2020909B1 (en) * 2018-05-09 2019-11-18 Audus B V Method for personalizing the audio signal of an audio or video stream

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020068986A1 (en) * 1999-12-01 2002-06-06 Ali Mouline Adaptation of audio data files based on personal hearing profiles
GB2380917A (en) * 2001-06-30 2003-04-16 Hewlett Packard Co An audio player with personalised audio profiles
WO2003026349A1 (en) * 2001-09-20 2003-03-27 Sound Id Sound enhancement for mobile phones and other products producing personalized audio for users
US20140194775A1 (en) * 2010-08-05 2014-07-10 Ace Communications Limited Method and System for Self-Managed Sound Enhancement
US20130182855A1 (en) * 2012-01-13 2013-07-18 Samsung Electronics Co., Ltd. Multimedia playing apparatus and method for outputting modulated sound according to hearing characteristic of user
US20190045302A1 (en) * 2017-08-07 2019-02-07 Leigh M. Rothschild System and Method For Adjusting Audio Parameters for a User
US20200027467A1 (en) * 2018-07-20 2020-01-23 Mimi Hearing Technologies GmbH Systems and methods for encoding an audio signal using custom psychoacoustic models
WO2020019020A1 (en) * 2018-07-23 2020-01-30 Audeara Pty Ltd Calibration method for customizable personal sound delivery systems

Also Published As

Publication number Publication date
WO2022129487A1 (en) 2022-06-23
GB202020147D0 (en) 2021-02-03

Similar Documents

Publication Publication Date Title
US11075609B2 (en) Transforming audio content for subjective fidelity
US10231074B2 (en) Cloud hosted audio rendering based upon device and environment profiles
US10070245B2 (en) Method and apparatus for personalized audio virtualization
US8977376B1 (en) Reproducing audio signals with a haptic apparatus on acoustic headphones and their calibration and measurement
KR102502465B1 (en) Digital audio filters for variable sample rates
US20110002471A1 (en) Systems and methods for transducer calibration and tuning
KR102302683B1 (en) Sound output apparatus and signal processing method thereof
US11134353B2 (en) Customized audio processing based on user-specific and hardware-specific audio information
AU2010291203A1 (en) An auditory test and compensation method
US10827264B2 (en) Audio preferences for media content players
JP2023123514A (en) Method for setting parameter for individual adaptation of audio signal
US20200081681A1 (en) Mulitple master music playback
GB2599742A (en) Personalised audio output
JP5598722B2 (en) Audio reproduction device and reproduction sound adjustment method in audio reproduction device
US11601752B2 (en) Sound quality enhancement and personalization
CN116208908A (en) Recording file playing method and device, electronic equipment and storage medium
KR101060546B1 (en) Method and apparatus for converting audio play file according to auditory capacity of individual user

Legal Events

Date Code Title Description
COOA Change in applicant's name or ownership of the application

Owner name: HEARS TECHNOLOGY LIMITED

Free format text: FORMER OWNER: BRENDAN MORRISSEY