EP3255904A1 - Distributed audio mixing - Google Patents

Distributed audio mixing Download PDF

Info

Publication number
EP3255904A1
EP3255904A1 EP16173264.9A EP16173264A EP3255904A1 EP 3255904 A1 EP3255904 A1 EP 3255904A1 EP 16173264 A EP16173264 A EP 16173264A EP 3255904 A1 EP3255904 A1 EP 3255904A1
Authority
EP
European Patent Office
Prior art keywords
audio
spatial
sources
signal
subset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP16173264.9A
Other languages
German (de)
French (fr)
Inventor
Antti Eronen
Arto Lehtiniemi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nokia Technologies Oy
Original Assignee
Nokia Technologies Oy
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nokia Technologies Oy filed Critical Nokia Technologies Oy
Priority to EP16173264.9A priority Critical patent/EP3255904A1/en
Publication of EP3255904A1 publication Critical patent/EP3255904A1/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/041Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal based on mfcc [mel -frequency spectral coefficients]
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/076Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of timing, tempo; Beat detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/155Musical effects
    • G10H2210/265Acoustic effect simulation, i.e. volume, spatial, resonance or reverberation effects added to a musical sound, usually by appropriate filtering or delays
    • G10H2210/295Spatial effects, musical uses of multiple audio channels, e.g. stereo
    • G10H2210/305Source positioning in a soundscape, e.g. instrument positioning on a virtual soundstage, stereo panning or related delay or reverberation changes; Changing the stereo width of a musical source
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/155User input interfaces for electrophonic musical instruments
    • G10H2220/201User input interfaces for electrophonic musical instruments for movement interpretation, i.e. capturing and recognizing a gesture or a specific kind of movement, e.g. to control a musical instrument
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/075Musical metadata derived from musical analysis or for use in electrophonic musical instruments
    • G10H2240/081Genre classification, i.e. descriptive metadata for classification or selection of musical pieces according to style
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/03Aspects of down-mixing multi-channel audio to configurations with lower numbers of playback channels, e.g. 7.1 -> 5.1
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/11Positioning of individual sound objects, e.g. moving airplane, within a sound field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/15Aspects of sound capture and related signal processing for recording or reproduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • H04S7/303Tracking of listener position or orientation
    • H04S7/304For headphones

Definitions

  • This relates to distributed audio mixing.
  • the further relates to, but it not limited to, methods and apparatus for distributed audio capture, mixing and rendering of spatial audio signals to enable spatial reproduction of audio signals.
  • Spatial audio refers to playable audio data that exploits sound localisation.
  • audio sources for example the different members of an orchestra or band, located at different locations on the stage.
  • the location and movement of the sound sources is a parameter of the captured audio.
  • rendering the audio as spatial audio for playback such parameters are incorporated in the data using processing algorithms so that the listener is provided with an immersive and spatially oriented experience.
  • SPAC Spatial Audio Capture
  • An example application of spatial audio is in virtual reality (VR) whereby both video and audio data is captured within a real world space.
  • VR virtual reality
  • the user through a VR headset, can view and listen to the captured video and audio which has a spatial percept.
  • Rendering of captured audio can be complex and the quality of spatial audio produced in terms of listener experience can be degraded in current systems, for example in situations where one or more audio sources change position or move during capture.
  • a method comprises: receiving a plurality of audio signals representing captured audio from respective audio sources in a space; processing the received audio signals to generate a spatial audio signal, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • the method may further comprise: defining a group of audio sources from the plurality of audio signals; and wherein the change in the spatial position in the generated spatial audio signal depends on the position of one or more of the other audio sources in said group of audio sources.
  • the method may further comprise receiving a spatial audio signal associated with a microphone array configured to provide spatial audio capture, and wherein the processing comprises generating an updated version of the spatial audio signal associated with the microphone array.
  • the processing may further comprise identifying a subset of the audio sources, and the detecting and generating steps are performed only in relation to the subset.
  • the method may further comprise identifying the subset manually through a user interface. Identifying the subset may be performed automatically based on attributes of the audio sources. Identifying the subset may be based on the position of one or more sources in the space. Identifying the subset may be based on the position of the one or more sources relative to a primary source whose position is permitted to change in the signal. Identifying the subset may be based on the position of the one or more sources relative to the position of the microphone array.
  • the method may further comprise receiving for each audio source positional data received from, or derived from, a positioning tag carried by the audio source. Identifying the subset may be determined based on the type of audio source. The type of audio source may be determined by signal analysis of the audio signals from each audio source. Analysis of the audio signals may identify a type of vocal performance.
  • Analysis of the audio signals may identify a type of instrument.
  • Determination of the audio source type may be performed by receiving identification data from a tag associated with the audio source.
  • Generating the spatial audio signal may comprise controlling the change in spatial position in accordance with one or more position modification rules.
  • Selection of a position modification rule may be received from a user through a user interface.
  • the user interface may display a spatial representation of each audio source together with the one or more selectable position modification rules for each.
  • The, or one of said position modification rules may prevent a change in spatial position.
  • The, or one of said position modification rules may comprise calculating an average detected change in spatial position over a predetermined time interval and wherein the change in spatial position in the signal may be based on the calculated average.
  • Identifying the subset may be determined by first identifying a group of associated audio sources based on spatial position, identifying a substantially common or average change in position for members of the group, and responsive to detecting a deviation of one or more members of the group from the average change in position, generating the spatial audio signal such that said deviation is not present therein.
  • Generating the spatial audio signal may comprise changing the position of the deviating members to substantially track the common or average change in position.
  • Generating the spatial audio signal may comprise setting the change in position at a single location, substantially at the centre of the group.
  • Generating the spatial audio signal may comprise setting the change in position for all members of the group to be equidistant from the centre of the group.
  • an apparatus comprises at least one processor and at least one memory including computer program code, the at least one memory and the computer program code configured with the processor to cause the apparatus to: receive a plurality of audio signals representing captured audio from respective audio sources in a space; process the received audio signals to generate a spatial audio signal, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generate the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • the processor may be configured to define a group of audio sources from the plurality of audio signals, and wherein the change in the spatial position in the generated spatial audio signal depends on the position of one or more of the other audio sources in said group of audio sources.
  • the processor may further cause the apparatus to receive a spatial audio signal associated with a microphone array configured to provide spatial audio capture, and to generate an augmented version of the spatial audio signal associated with the microphone array.
  • the apparatus may identify the subset is performed by receiving identification data entered manually through a user interface. Identifying the subset may be performed automatically based on attributes of the audio sources. Identifying the subset may be based on the position of one or more sources in the space. Identifying the subset may be based on the position of the one or more sources relative to a primary source whose position is permitted to change in the signal. Identifying the subset may be based on the position of the one or more sources relative to the position of the microphone array.
  • the processor may be further configured to receive for each audio source positional data received from, or derived from, a positioning tag carried by the audio source.
  • Identifying the subset may be determined based on the type of audio source.
  • the type of audio source may be determined by means of the processor performing signal analysis of the audio signals from each audio source.
  • the analysis of the audio signals may identify a type of vocal performance.
  • the processor may be arranged to generate the spatial audio signal by controlling the change in spatial position in accordance with one or more position modification rules.
  • Selection of the position modification rule may be received from a user through a user interface provided by the processor.
  • the processor may be configured to display a spatial representation of each audio source together with the one or more selectable position modification rules for each.
  • The, or one of said position modification rules may prevent a change in spatial position.
  • The, or one of said position modification rules may comprise calculating an average detected change in spatial position over a predetermined time interval and wherein the change in spatial position in the signal is based on the calculated average.
  • the processor may be configured to identify the subset by first identifying a group of associated audio sources based on spatial position, to identify a substantially common or average change in position for members of the group, and responsive to detecting a deviation of one or more members of the group from the average change in position, to generate the spatial audio signal such that said deviation is not present therein.
  • the processor may be configured to generate the spatial audio signal by changing the position of the deviating members to substantially track the common or average change in position.
  • the processor may be configured to generate the spatial audio signal by setting the change in position at a single location, substantially at the centre of the group.
  • the processor may be configured to generate the spatial audio signal by setting the change in position for all members of the group to be equidistant from the centre of the group.
  • the apparatus may be further configured to receive spatial video signals in association with the audio signals.
  • the processor may be configured to identify the subset of audio sources based on one or both of the audio signals and the video signals.
  • an apparatus comprises at least one processor and at least one memory including computer program code, the at least one memory and the computer program code configured with the processor to cause the apparatus to: receive a plurality of audio signals representing captured audio from respective audio sources in a space; process the received audio signals to generate a spatial audio signal for playback, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • a computer program comprises instructions that when executed by a computer apparatus control it to perform the method of any preceding definition
  • a non-transitory computer-readable storage medium having stored thereon computer-readable code, which, when executed by at least one processor, causes the at least one processor to perform a method, comprising: receiving a plurality of audio signals representing captured audio from respective audio sources in a space; processing the received audio signals to generate a spatial audio signal for playback, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • apparatus having at least one processor and at least one memory having computer-readable code stored thereon which when executed controls the at least one processor: to receive a plurality of audio signals representing captured audio from respective audio sources in a space; to process the received audio signals to generate a spatial audio signal for playback, in which the processing comprises:
  • Embodiments herein relate to systems and methods relating to the capture and rendering of spatial audio data for playback.
  • the embodiments relate to distributed spatial and rendering methods in which there are multiple audio sources which may move within a virtual space over time. Each audio source generates respective audio signals and, in some embodiments, positioning information for use by the system.
  • An example application is in a VR capture and rendering system in which video is also captured and rendered to provide an immersive user experience.
  • Nokia's OZO (RTM) VR camera is used as an example of a VR capture device which comprises a microphone array to provide a spatial audio signal, but it will be appreciated that the embodiments are not limited to VR applications nor the use of microphone arrays at the capture point.
  • FIG. 1 an overview of a VR capture scenario 1 is shown together with a first embodiment capture and rendering system (CRS) 15 with associated user interface 16.
  • CRS capture and rendering system
  • the Figure shows in plan-view a real world space 3 which may be for example a concert hall or other music venue.
  • the CRS 15 is applicable to any real world space, however.
  • a VR device 6 for video and spatial audio capture is supported on a floor 5 of the space 3 in front of multiple audio sources, in this case a band; the position of the VR device 6 is known, e.g. through predetermined positional data or signals derived from a positioning tag on the VR device (not shown).
  • the VR device 6 comprises a microphone array configured to provide spatial audio capture.
  • the band may be comprised of multiple members each of which has an associated external microphone or (in the case of guitarists) a pick-up feed providing audio signals. Each may therefore be termed an audio source for convenience. In other embodiments, other types of audio source can be used.
  • the audio sources in this case comprise a lead vocalist 7, a drummer 8, lead guitarist 9, bass guitarist 10, and three members of a choir or backing singers 11, 12, 13 which members are spatially close together in a group.
  • the audio sources 7 - 13 carry a positioning tag which can be any module capable of indicating through data its respective spatial position to the CRS 15.
  • the positioning tag may be a high accuracy indoor positioning (HAIP) tag which works in association with one or more HAIP locators 20 within the space 3.
  • HAIP systems use Bluetooth Low Energy (BLE) communication between the tags and the one or more locators 20.
  • BLE Bluetooth Low Energy
  • a respective HAIP locator may be to the front, left, back and right of the VR device 6.
  • Each tag sends BLE signals from which the HAIP locators derive the tag, and therefore, audio source location.
  • such direction of arrival (DoA) positioning systems are based on (i) a known location and orientation of the or each locator, and (ii) measurement of the DoA angle of the signal from the respective tag towards the locators in the locators' local co-ordinate system. Based on the location and angle information from one or more locators, the position of the tag can be calculated using geometry.
  • the CRS 15 is a processing system having an associated user interface (UI) 16 which will explained in further detail below. As shown in Figure 1 , it receives as input from the capture device 6 spatial audio and video data, and positioning data, through a signal line 17. Alternatively, the positioning data can be received from the HAIP locator 20. The CRS 15 also receives as input from each of the audio sources 7 - 13 audio data and positioning data from the respective positioning tags, or the HAIP locator 20, through separate signal lines 18. The CRS 15 generates spatial audio data for output to a user device 19, such as a VR headset with video and audio output.
  • a user device 19 such as a VR headset with video and audio output.
  • the input audio data can be multichannel audio in loudspeaker format, e.g. stereo signals, 4.0 signals, 5.1 signals, Dolby Atmos (RTM) signals or the like.
  • loudspeaker format audio the input can be in the multi microphone signal format, such as the raw eight signal input from the OZO VR camera, if used for the capture device 6.
  • FIG. 2 shows an example schematic diagram of components of the CRS 15.
  • the CRS 15 has a controller 22, a touch sensitive display 24 comprised of a display part 26 and a tactile interface part 28, hardware keys 30, a memory 32, RAM 34 and an input interface 36.
  • the controller 22 is connected to each of the other components in order to control operation thereof.
  • the touch sensitive display 24 is optional, and as an alternative a conventional display may be used with the hardware keys 30 and/or a mouse peripheral used to control the CRS 15 by conventional means.
  • the memory 32 may be a non-volatile memory such as read only memory (ROM) a hard disk drive (HDD) or a solid state drive (SSD).
  • the memory 32 stores, amongst other things, an operating system 38 and software applications 40.
  • the RAM 34 is used by the controller 22 for the temporary storage of data.
  • the operating system 38 may contain code which, when executed by the controller 22 in conjunction with RAM 34, controls operation of each of hardware components of the terminal.
  • the controller 22 may take any suitable form. For instance, it may be a microcontroller, plural microcontrollers, a processor, or plural processors.
  • the software application 40 is configured to provide video and distributed spatial audio capture, mixing and rendering to generate a VR environment, or virtual space, including the rendered spatial audio.
  • the software application 40 also provides the UI 16 shown in Figure 1 , through its output to the display 24 and receives user input through the tactile interface 28 or other input peripherals such as the hardware keys 30 or a mouse (not shown).
  • the mixing stage which forms a sub-part of the rendering stage in this case, may be performed manually through the UI 16 or all or part of said mixing stage may be performed automatically.
  • the software application 40 may render the virtual space, including the spatial audio, using known signal processing techniques and algorithms based on the mixing stage.
  • the input interface 36 receives video and audio data from the capture device 6, such as Nokia's OZO (RTM) device, and audio data from each of the audio sources 7 - 13.
  • the input interface 36 also receives the positioning data from (or derived from) the positioning tags on each of the capture device 6 and the audio sources 7 - 13, from which can be made an accurate determination of their respective positions in the real world space 3.
  • the software application 40 may be configured to operate in any of real-time, near real-time or even offline using pre-stored captured data.
  • One example aspect of the mixing stage of software application 40 is controlling how audio sources move, or change position, in the rendered virtual space responsive to detected movement in the captured real world space 3.
  • audio sources move For example, in the Figure 1 situation, any one of the audio sources 7 - 13 may move over time, as therefore will their respective audio position with respect to the capture device 6 and also to each other.
  • users may be used to static sources where the audio source is generally central. When audio sources move, the rendered result may be overwhelming and distracting.
  • the software application 40 may be configured to restrict or filter movement of a subset of the audio sources 7 - 13 when rendering the spatial audio part of the virtual space. This restriction / filtering is performed based on a number of attributes, to be described below, some of which can be set manually through the UI 16 and some of which can be handled automatically by predetermined algorithms.
  • FIG. 3 shows an overview flow diagram of the capture, mixing and rendering stages of software application 40.
  • the mixing and rendering stages may be combined.
  • Mixing (step 3.2) may be dependent on a manual or automatic control step 3.4 which may be based on attributes of the captured video and/or audio. Other attributes may be used.
  • a determination may be made as to the one or most important audio sources 7 - 13 at each point in time.
  • the remaining audio sources are assigned to a subset subject to filtering or restriction, meaning that their position in virtual space does not track real-world movement.
  • the most important audio sources are permitted to move, meaning that the spatial position of their corresponding audio signals also move, typically to closely track the real world movement.
  • Filtering or restriction of the remaining subset may mean preventing their movement completely (freezing / no movement) or limiting movement by filtering, e.g. using time-based averaging.
  • FIG 4a which is similar to Figure 1 , in this real world scenario the CRS 15 may determine that the lead vocalist 7 is the most important and therefore its captured sound is allowed to move in the rendered virtual space 42, represented graphically in Figure 4b .
  • the remaining audio sources 8 - 13 are in this case frozen or restricted in terms of rendered movement.
  • Figure 4b shows the result of the lead vocalist 7 and the lead guitarist 9 moving during capture, which is indicated in Figure 4a by the dotted arrows.
  • the other audio sources 8, 10, 11, 12, 13 are also frozen from movement.
  • FIG. 5 shows the processing steps performed by the CRS 15 through software application 40.
  • the capturing step 5.1 is followed by the step 5.2 of identifying the most important source (MIS).
  • Step 5.3 mixes the audio so as to allow movement of the MIS and to prevent movement of the other audio sources. Rendering takes place in step 5.4 in accordance with said mixing step 5.3.
  • the feedback arrow indicates that the steps 5.1 - 5.4 may be repeated over time.
  • Determination of the MIS can be done using various methods.
  • One method is for the CRS 15 to analyse the position of each audio source 7 - 13 by means of the positioning tags, and/or using the captured video data from capture device 6.
  • the front and centre positioning of the audio source 7 in this case is indicative of its importance relative to the others.
  • the spatial audio signal from the capture device 6 and/or audio signals from each of the audio sources 7 - 13 can be analysed by the CRS 15 to determine the most important sound. For example, the sound indicative of the lead melody or vocals may be indicative of the most important source.
  • the MIS may also be identified manually through the UI 16.
  • Figure 6 shows an example screenshot of the UI 16 which shows the detected positions of the audio sources within the space 3, the UI permitting manual selection of whether a source is movable (M) or fixed/frozen (F). Clicking or touching on a source may toggle between M and F or a drop down menu can provide the options. Further filtering options can be provided.
  • M movable
  • F fixed/frozen
  • one or more anchor sources is or are determined and their position is fixed or filtered in a manner similar to the first example.
  • audio sources corresponding to drums 8, lead guitar 9 and bass guitar 10 are identified and restricted. Restriction here may mean freezing or fixing the position.
  • the respective positions of these audio sources 8 - 10 may still be obtained and tracked using the tag-based positioning, and rather than fixing or freezing their position, their position may be filtered using a long term average position of each tag.
  • long term may mean one minute, but shorter or longer periods may be used.
  • the average position after one minute, to take this example, may indicate an allowed degree of movement in the rendered version.
  • Figure 8 shows the main processing steps, comprised of capture (step 8.1), identifying anchor sources (step 8.2), mixing (step 8.3) and rendering (step 8.4).
  • setting the anchor sources may be performed automatically by position, audio analysis or by other means, or manually through the UI 16.
  • Filtering parameters may also be set through the UI 16, e.g. to set the averaging period.
  • source movement may depend on the input signal type over time.
  • the lead guitarist 9 may be playing a guitar solo at a point in time and so is considered an important source and allowed to move, as is the lead vocalist 7.
  • the source may become frozen or filtered as before.
  • the vocalisation type for the lead vocalist 7 may change over time; speaking or whistling can be identified by signal analysis and at this time the source position is frozen or time-filtered as explained above.
  • vocalisation type changes to singing normal movement may be enabled.
  • This may be used to identify a speech-type monologue from singing, which may be the basis of MIS determination.
  • a generalised method may comprise training a model, for example a Gaussian mixture model or a neural network for the classes to be classified, e.g. singing, speaking, shouting etc.
  • a model for example a Gaussian mixture model or a neural network for the classes to be classified, e.g. singing, speaking, shouting etc.
  • a set of audio files is obtained where the vocal type is known.
  • suitable features may be, for example, mel-frequency cepstral coefficients and the fundamental frequency (fo).
  • other features related to, for example, voice characteristics can be used.
  • the models are then trained to represent typical feature values of different vocalisation types, in the case of a Gaussian mixture model, or a neural network is trained to discriminate between the different vocalisation types given the features.
  • Online classification is then performed by extracting the same features from the input signal, and then evaluating which model has the highest likelihood of having generated the features (Gaussian mixture model) or which category gives the largest output value (neural network).
  • the actions to be taken responsive to signal type can be set using the mixing UI 16.
  • the UI 16 can enable the operator to identify input signal types and associate each with different restrictions or filtered movements, for example.
  • the manner in which the mixing operator sets the restrictions can cater for a variety of situations.
  • the system may have a set of presets which can be applied.
  • the behaviour indicated in Figure 9 may be one example preset.
  • Figure 10 shows the main processing steps, comprised of capture (step 10.1), identification of signal type (step 10.2), mixing based on presets or rules (step 10.3) and rendering (step 10.4).
  • the rules in step 10.3 assign certain types of signal to be moved and other types to be restricted; this represents a first preset, which can itself be modified over time to apply different rules.
  • the CRS 15 is configured to filter or restrict movement within a group of associated audio sources.
  • the choir members 11 - 13 are spatially close to one another and identified to the CRS 15 manually or automatically as a group.
  • Each of the choir members 11 - 13 has a positioning tag, but the CRS 15 may be configured to receive, or generate, a single source position which may be the middle or mean position of the group members, notwithstanding any movement or change in position. This may be desirable if we wish to have a single sound source that represents the whole group rather than using individual spatial positions. Alternatively, the individual sounds from each group member 11 - 13 may be adjusted so that they are equidistant from the group centre.
  • the CRS 15 is again configured to filter or restrict movement with a group of associated audio sources, and in this case the CRS selectively filters the group movement so that coherent movement (all members moving in generally the same direction) is allowed in the rendered output, but movement in incoherent (different) directions is filtered, frozen and/or adjusted. This may be performed by determining a reference movement direction and subsequently filtering movement based on deviation of one or more group members. As an example, the CRS 15 may measure the most common movement direction and identifies if one or more group members deviate from the common direction.
  • the CRS 15 when detecting coherent movement allows that movement in the rendered version as shown.
  • the CRS 15 when detecting that one or more members (in this case the choir source 13) deviates away from the common movement moves that member in the direction of common movement in contrast to their actual movement.
  • the CRS 15 may be user adjustable. For example, when there is a sufficient amount of coherent movement, all movement from the selected group is rendered. Other movement functions may be provided.
  • the locator tags as before can be visualised and grouped.
  • the mixer operator may select tags to form a group and then select an appropriate function for that group.
  • Another example is to draw a rectangle or other shape to select tags within the area and then select an appropriate function for that group.
  • the CRS 15 can filter movement of audio sources 7 - 13 based on the listener location within the virtual space.
  • the CRS 15 can in real time determine the movement direction and/or velocity with reference to the listener location and through audio processing determine if it is audibly acceptable or pleasing, or not. For example, movement of a particular source when the listener is positioned directly in front of that source may be acceptable. However, if the listener were to move closer to, and to one side of, the same source, this may result in significant sideways movement producing unacceptable results.
  • filtering can be applied, for example to reposition the audio source front and centre of the listener's position.
  • FIG. 13 an example schematic representation of capture and rendering parts of the Figure 1 scenario is shown, some parts of which are provided by the CRS 15, in terms of functional components for mixing and rendering.
  • the Figure 13 shows separate capture and rendering parts 101, 103 with mixing being performed as part of the rendering part.
  • the audio source is in this example a Lavalier microphone 111 associated with a movable person, e.g. a singer.
  • the Lavalier microphone 111 is an example of a 'close' audio source capture apparatus and may in some embodiments be a boom microphone or similar neighbouring microphone capture system.
  • a Lavalier microphone and thus a Lavalier audio signal the concept may be extended to any microphone external or separate to e.g. the microphones or array of microphones of the capture device 5 configured to capture the spatial audio signal.
  • the concept is applicable to any external/additional microphones in addition to a spatial audio capture (SPAC) microphone array, used in the OZO device, be they Lavalier microphones, hand held microphones, mounted mics, etc.
  • SPAC spatial audio capture
  • the external microphones can be worn/carried by persons or mounted as close-up microphones for instruments or a microphone in some relevant location which the designer wishes to capture accurately.
  • the Lavalier microphone 111 may in some embodiments be a microphone array.
  • the Lavalier microphone typically comprises a small microphone worn around the ear or otherwise close to the mouth.
  • the audio signal may be provided either by a Lavalier microphone or by an internal microphone system of the instrument (e.g., pick-up microphones in the case of an electric guitar).
  • the Lavalier microphone 111 may be configured to output the captured audio signals to a variable delay compensator 117.
  • the Lavalier microphone may be connected to a transmitter unit (not shown), which wirelessly transmits the audio signal to a receiver unit (not shown).
  • the capture part 101 comprises a Lavalier (or close source) microphone position tag 112.
  • the Lavalier microphone position tag 112 may be configured to determine information identifying the position or location of the Lavalier microphone 111 or other close microphone. It is important to note that microphones worn by people may be freely moved in the acoustic space and the system supporting location sensing of wearable microphone may support continuous sensing of user or microphone location.
  • the Lavalier microphone position tag 112 may be configured to output this determination of the position of the Lavalier microphone to a position tracker 115.
  • the capture part 101 comprises a spatial audio capture device 113 which corresponds to the capture device 5 in Figure 2 .
  • the spatial audio capture device 113 is an example of an 'audio field' capture apparatus and may in some embodiments be a directional or omnidirectional microphone array.
  • the spatial audio capture device 113 may be configured to output the captured audio signals to a variable delay compensator 117.
  • the capture part 101 may comprise a spatial capture position tag 114.
  • the spatial capture position tag 114 may be configured to determine information identifying the position or location of the spatial audio capture device 113.
  • the spatial capture position tag 114 may be configured to output this determination of the position of the spatial capture microphone to the position tracker 115.
  • the capture apparatus does not need to comprise a position tag.
  • the spatial audio capture device 113 is implemented within a mobile device.
  • the spatial audio capture device is thus configured to capture spatial audio, which, when rendered to a listener, enables the listener to experience the sound field as if they were present in the location of the spatial audio capture device.
  • the Lavalier microphone 111 in such embodiments may be configured to capture high quality close-up audio signals (for example from a key person's voice, or a musical instrument).
  • the attributes of the key source such as gain, timbre and spatial position may be adjusted in order to provide the listener with a much more realistic immersive experience.
  • it is possible to produce more point-like auditory objects thus increasing the engagement and intelligibility.
  • the capture part 101 furthermore may comprise a position tracker 115.
  • the position tracker 115 may be configured to receive the positional tag information identifying positions of the Lavalier microphone 111 and the spatial audio capture device 113 and generate a suitable output identifying the relative position of the Lavalier microphone 111 relative to the spatial audio capture device 113 and output this to the render apparatus 103 and specifically in this example to a location date filtering model 120.
  • the position tracker 115 may be configured to output the tracked position information to a variable delay compensator 117.
  • the locations of the Lavalier microphones (or the persons carrying them) with respect to the spatial audio capture device 113 may be tracked and used for mixing the sources to correct spatial positions.
  • the position tags namely, the microphone position tag 112 and the spatial capture position tag 114 may be implemented using High Accuracy Indoor Positioning (HAIP) or another suitable indoor positioning technology.
  • HAIP High Accuracy Indoor Positioning
  • the position tracker may use video content analysis and/or sound source localization, as will be explained below.
  • the capture part 101 furthermore may comprise a variable delay compensator 117 configured to receive the outputs of the Lavalier microphone 111 and the spatial audio capture device 113. Furthermore in some embodiments the variable delay compensator 117 may be configured to receive source position and tracking information from the position tracker 115. The variable delay compensator 117 may be configured to determine any timing mismatch or lack of synchronisation between the close audio source signals and the spatial capture audio signals and determine the timing delay which would be required to restore synchronisation between the signals. In some embodiments the variable delay compensator 117 may be configured to apply the delay to one of the signals before outputting the signals to the render apparatus 103 and specifically in this example to the audio renderer 121.
  • the timing delay may be referred as being a positive time delay or a negative time delay with respect to an audio signal.
  • a first (spatial) audio signal by x
  • another (Lavalier) audio signal by y.
  • the delay T can be either positive or negative.
  • the capture stage 101 furthermore may comprise a camera 118 configured to capture video signals; the camera 118 may be a single camera or array of cameras, as in the OZO device example, in which spatial video signals are produced. In some embodiments, the camera 118 may form part of the same physical unit as the spatial audio capture device 113.
  • the capture stage 101 further may comprise a content analysis module 119 which receives both the video signals from the camera 118 and audio from the variable time delay compensator 117, which corrects for any timing mismatch or lack of synchronisation between the close audio signals, the spatial capture audio signals and the video signals.
  • the output from the content analysis module 119 is provided to the rendering stage 103.
  • the render part 103 comprises a head tracker 123.
  • the head tracker 123 may be any suitable means for generating a positional input, for example a sensor attached to a set of headphones configured to monitor the orientation of the listener, with respect to a defined or reference orientation and provide a value or input which can be used by the audio renderer 121.
  • the head tracker 123 may in some embodiments be implemented by at least one gyroscope and/or digital compass.
  • the render part 103 comprises a location data filtering module 120 which takes as input the unfiltered HAIP location data from the position tracker 115, and the audio-visual content analysis data from the content analysis module 119.
  • the render part 103 comprises an audio renderer 121.
  • the audio renderer 121 is configured to receive output for the location data filtering module 120.
  • the audio renderer 121 can furthermore be configured to receive an input from the head tracker 123.
  • the audio renderer 121 can be configured to receive other user inputs for example entering manually through the UI 16.
  • the audio renderer 121 as described herein in further detail later, can be configured to mix together the audio signals, namely the Lavalier microphone audio signals and the spatial audio signals based on the positional information and/or the head tracker inputs in order to generate a mixed audio signal in accordance with the methods and rules described above to cater for detected movement over time.
  • the mixed audio signal can for example be passed to headphones 125.
  • the output mixed audio signal can be passed to any other suitable audio system for playback (for example a 5.1 channel audio amplifier).
  • the audio renderer 121 may be configured to perform spatial audio processing on the audio signals from the microphone array and from the close microphone.
  • the Lavalier audio signal is from the Lavalier microphones and the spatial audio captured by the microphone array and processed with the spatial analysis may in some embodiments be combined by the audio renderer to a single binaural output which can be listened through headphones.
  • the spatial audio signal may be converted into a multichannel signal.
  • the multichannel output may then be binaurally rendered, and summed with binaurally rendered Lavalier source signals.
  • the rendering may be described initially with respect to a single (mono) channel, which can be one of the multichannel signals from the spatial audio signal or one of the Lavalier sources.
  • a single (mono) channel which can be one of the multichannel signals from the spatial audio signal or one of the Lavalier sources.
  • Each channel in the multichannel signal set may be processed in a similar manner, with the treatment for Lavalier audio signals and multichannel signals having the following differences:
  • the render part 103 in some embodiments may comprise headphones 125.
  • the headphones can be used by the listener to generate the audio experience using the output from the audio renderer 121.
  • the Lavalier microphone signals can be mixed to suitable spatial positions in the spatial audio field in accordance with predetermined rules or user input.
  • the rendering can be done by rendering the spatial audio signal using virtual loudspeakers with fixed positions, and the captured Lavalier source is rendered from a time varying position.
  • the audio renderer 121 is configured to control the azimuth, elevation, and distance of the Lavalier or close source based on the tracked position data and the filtering rules.
  • the user may be allowed to adjust the gain and/or spatial position of the Lavalier source using the output from the head-tracker 123.
  • the head-tracker input may affect the mix of the Lavalier source relative to the spatial sound. This may be changing the 'spatial position' of the Lavalier source based on the head-tracker or by changing the gain of the Lavalier source where the head-tracker input is indicating that the listener's head is 'towards' or 'focussing' on a specific source.
  • the mixing/rendering may be dependent on the relative position/orientation of the Lavalier source and the spatial microphones but also be dependent on the orientation of the head as measured by the head-tracker.
  • the user input may be any suitable user interface input, such as an input from a touchscreen indicating the listening direction or orientation.
  • a spatial downmix into a 5.1 channel format or other format could be employed.
  • the Lavalier or close source can in some embodiments mixed to its 'proper' spatial position using known amplitude panning techniques.
  • Any one of the classifications can be used alongside selection rules to determine the MIS.
  • the first classifier and the third classifier may be a non-probabilistic classifier, such as a support vector machine (SVM) classifier.
  • the second classifier may be a probabilistic classifier, for example based on one or more Gaussian Mixture Models (GMMs).
  • GMMs Gaussian Mixture Models
  • the classification data (or tags, as they are described in the reference) may indicate one or more of the following characteristics for determining the MIS using the predefined rules: a musical instrument, the presence or absence of vocals and/or a vocalist gender, presence or absence of music and a musical genre. This assumes that the audio data is music. In other embodiments, the audio data may comprise spoken word elements or a combination of music and spoken word elements.
  • Figure 14 is an overview of a method of determination of tag information for the audio track.
  • Acoustic features 131 of the audio are extracted and input to first level classifiers 132 to generate first level classifications for the audio track.
  • first classifiers 133 and second classifiers 134 are used to generate first and second classifications respectively.
  • the first classifiers 133 are non-probabilistic classifiers
  • the second classifiers 134 are probabilistic classifiers.
  • the first and second classifications generated by the first level classifiers 132 are provided as inputs to a second level classifier 135.
  • One or more second level classifications are generated by the second level classifier 135, based at least in part on the first and second classifications.
  • the second level classifiers 135 include a third classifier 136, which outputs a third classification.
  • One or more tags 137 are generated, based on the second level classifications. Such tags 137 may be stored by a tagging module 138 for determining the MIS.
  • the input signal is decoded into pulse code modulation (PCM) data (step s15.1).
  • PCM pulse code modulation
  • the samples for decoding are taken at a rate of 44.1 kHz and have a resolution of 16 bits.
  • acoustic features 131 or descriptors are extracted which indicate characteristics of the audio track (step s13.2).
  • the features 131 are based on mel-frequency cepstral coefficients (MFCCs).
  • MFCCs mel-frequency cepstral coefficients
  • other features such as fluctuation pattern and danceability features, beats per minute (BPM) and related features, chorus features and other features may be used instead of, or as well as MFCCs.
  • An example method for extracting acoustic features 131 from the input signal at step s15.2 is described in detail in PCT/FI2014/051036 .
  • the audio features 131 produced at step 15.2 may include one or more of:
  • one or more accent signals are derived from the input signal, for detection of events and/or changes in the audio track.
  • a first one of the accent signals may be a chroma accent signal based on fundamental frequency Fo salience estimation, while a second one of the accent signals may be based on a multi-rate filter bank decomposition of the input signal.
  • a BPM estimate may be obtained based on a periodicity analysis for extraction of a sequence of periodicity vectors on the basis of the accent signals, where each periodicity vector includes a plurality of periodicity values, each periodicity value describing the strength of periodicity for a respective period length, or "lag".
  • a point-wise mean or median of the periodicity vectors over time may be used to indicate a single representative periodicity vector over a time period of the audio track. For example, the time period may be over the whole duration of the audio track.
  • an analysis can be performed on the periodicity vector to determine a most likely tempo for the audio track.
  • One example approach comprises performing k-nearest neighbours regression to determine the tempo. In this case, the system is trained with representative music tracks with known tempo.
  • the k-nearest neighbours regression is then used to predict the tempo value of the audio track based on the tempi of k-nearest representative tracks. More details of such an approach have been described in Eronen, Klapuri, "Music Tempo Estimation With k -NN Regression", IEEE Transactions on Audio, Speech, and Language Processing, Vol. 18, Issue 1, pages 50-57 , the disclosure of which is incorporated herein by reference.
  • Chorus related features that may be extracted at step s15.2 include: a chorus start time; and a chorus end time.
  • Example systems and methods that can be used to detect chorus related features are disclosed in US 2008/236371 A1 , the disclosure of which is hereby incorporated by reference in its entirety.
  • Figure 16 is an overview of a process of extracting multiple acoustic features 131, some or all of which may be obtained in step s15.2.
  • Figure 16 shows how some input features are derived, at least in part, from computations of other input features.
  • the features 131 shown in Figure 16 include the MFCCs, delta MFCCs and mel-band energies discussed above, indicated using bold text and solid lines.
  • steps s15.3 to s15.10 the method produces the first level classifications, that is the first classifications and the second classifications, based on the features 131 extracted in step s15.2.
  • Figure 15 shows steps s15.3 to s15.10 being performed sequentially, in other embodiments, steps s15.3 to s15.7 may be performed after, or in parallel with, steps s15.8 to s15.10.
  • the first and second classifications are generated using the first classifiers 133 and the second classifiers 134 respectively, where the first and second classifiers 133, 134 are different from one another.
  • the first classifiers 133 may be non-probabilistic and the second classifiers 134 may be probabilistic classifiers, or vice versa.
  • the first classifiers 133 are support vector machine (SVM) classifiers, which are non-probabilistic.
  • the second classifiers 134 are based on one or more Gaussian Mixture Models (GMMs).
  • GMMs Gaussian Mixture Models
  • step s15.3 one, some or all of the features 131 or descriptors extracted in step s15.2, to be used to produce the first classifications 133, are selected and, optionally, normalised.
  • a look up table or database may be stored for each of the first classifications to be produced, that provides a list of features to be used to generate each first classifier and statistics, such as mean and variance of the selected features, that can be used in normalisation of the extracted features 131.
  • the list of features is received from the look up table, and accordingly the method selects and normalises the listed features for each of the first classifications to be generated.
  • the normalisation statistics for each first classification in the database may be determined during training of the first classifiers 133.
  • the first classifiers 133 are SVM classifiers.
  • the SVM classifiers 133 are trained using a database of audio tracks for which information regarding musical instruments and genre is already available.
  • the database may include tens of thousands of tracks for each particular musical instrument that might be tagged.
  • Examples of musical instruments for which information may be provided in the database include:
  • the training database includes indications of genres that the audio tracks belong to, as well as indications of genres that the audio tracks do not belong to.
  • a SVM classifier 133 can be trained to determine whether or not an audio track includes a particular instrument, for example, an electric guitar.
  • another SVM classifier 133 can be trained to determine whether or not the audio track belongs to a particular genre.
  • the training process can include determining a selection of one or more features 131 to be used as a basis for particular first classifications and statistics for normalising those features 131.
  • the number of features available for selection, M may be much greater than the number of features selected for determining a particular first classification, N; that is, M >> N.
  • the selection of features 131 to be used is determined iteratively, based on a development set of audio tracks for which the relevant instrument or genre information is available.
  • the normalisation of the selected features 131 at step s15.3 is optional. Where provided, the normalization of the selected features 131 in step s15.3 may potentially improve the accuracy of the first classifications.
  • the method defines a single "feature vector" for each set of selected features 131 or selected combination of features 131.
  • the feature vectors may then be normalized to have a zero mean and a variance of 1, based on statistics determined and stored during the training process.
  • the method generates one or more first probabilities that the audio track has a certain characteristic, corresponding to a potential tag 137, based on the normalized transformed feature vector or vectors.
  • a first classifier 133 is used to calculate a respective probability for each feature vector defined in step s15.4. In this manner, the number of SVM classifiers 133 corresponds to the number of characteristics or tags 137 to be predicted for the audio track.
  • a probability is generated for each instrument tag and for each musical genre tag to be predicted for the audio track, based on mean MFCCs and a MFCC covariance matrix.
  • a probability may be generated based on whether the audio track is likely to be an instrumental track or a vocal track.
  • another first classification may be generated based on whether the vocals are provided by a male or female vocalist.
  • the controller may generate only one or some of these probabilities and/or calculate additional probabilities at step 15.5.
  • the different classifications may be based on respective selections of features from the available features 131 extracted in step s15.2.
  • a logarithmic transformation may be applied to the probabilities output by the SVM classifiers 133 (step s15.6), so that the probabilities of all the first classifications are on the same scale and the optimal predicated probability threshold may correspond to a predetermined value, such as 0.5.
  • the first classifications are then output (step s15.7).
  • the first classifications correspond to the normalized probability p norm that a respective one of the tags 137 to be considered applies to the audio track.
  • the first classifications may include probabilities p inst1 that a particular instrument is included in the audio track and probabilities p gen1 that the audio track belongs to a particular genre.
  • second classifications for the input signal are based on the MFCCs and other parameters produced in step s15.2, using the second classifiers 134.
  • the features 131 on which the second classifications are based are the MFCC matrix for the audio track and the first time derivatives of the MFCCs.
  • the probabilities of the audio track including a particular instrument or belonging to a particular genre are assessed using probabilistic models that have been trained to represent the distribution of features extracted from audio signals captured from each instrument or genre.
  • the probabilistic models are GMMs.
  • Such models can be trained using an expectation maximisation algorithm that iteratively adjusts the model parameters to maximise the likelihood of the model for a particular instrument or genre generating features matching one or more input features in the captured audio signals for that instrument or genre.
  • the parameters of the trained probabilistic models may be stored in a database or in remote storage.
  • At least one likelihood is evaluated that the respective probabilistic model could have generated the selected or transformed features from the input signal.
  • the second classifications correspond to the models which have the largest likelihood of having generated the features of the input signal.
  • probabilities are generated for each instrument tag at step s15.8 and for each musical genre tag at step s15.9, as well as a probability whether the audio track is likely to be an instrumental track or a vocal track may also be generated. Also, for vocal tracks, another probability may be generated based on whether the vocals are provided by a male or female vocalist. In other embodiments, the method may generate only one or some of these second classifications and/or calculate additional second classifications at steps s15.8 and s15.9.
  • probabilities p inst2 that the instrument tags will apply, or not apply are produced by the second classifiers 134 using first and second Gaussian Mixture Models (GMMs), based on the MFCCs and their first time derivatives calculated in step s15.2.
  • GMMs Gaussian Mixture Models
  • probabilities p gen2 that the audio track belongs to a particular musical genre are produced by the second classifiers 134 using third GMMs.
  • the first and second GMMs used to compute the instrument-based probabilities p inst2 may be trained and used slightly differently from third GMMs used to compute the genre-based probabilities gen2 , as will now be explained.
  • step s15.8 precedes step s15.9.
  • step s15.9 may be performed before, or in parallel with, step s15.8.
  • first and second GMMs are used to generate the instrument-based probabilities p inst2 (step s15.8),based on MFCC features 131 obtained in step s15.2.
  • a likelihood L is computed for the audio track belonging to that genre, based on the likelihood of each of the third GMMs being capable of outputting the MFCC feature vector of the audio track. For example, to determine which of the eighteen genres in the list hereinabove might apply to the audio track, eighteen likelihoods would be produced.
  • the second classifications which correspond to the probabilities p inst2 and p gen2 , are then output (step s15.10).
  • the first classifications p inst1 and p gen1 and the second classifications p inst2 and p gen2 for the audio track are normalized to have a mean of zero and a variance of 1 (step s15.11) and collected to form a feature vector for input to one or more second level classifiers 135 (step s15.12).
  • the second level classifiers 135 include third classifiers 136.
  • the third classifiers 136 may be non-probabilistic classifiers, such as SVM classifiers.
  • the third classifiers 136 may be trained in a similar manner to that described above in relation to the first classifiers 133.
  • the first classifiers 133 and the second classifiers 134 may be used to output probabilities for the training sets of example audio tracks from the database.
  • the outputs from the first and second classifiers 133, 134 are then used as input data to train the third classifier 135.
  • the third classifier 136 generates determine probabilities p inst3 for whether the audio track contains a particular instrument and/or probabilities p gen3 for whether the audio track belongs to a particular genre (step s15.13).
  • the probabilities p inst3 , p gen3 are then log normalized (step s15.14), as described above in relation to the first classifications, so that a threshold of 0.5 may be applied to generate the third classifications, which are output at step s15.15.
  • the method determines whether each instrument tag and each genre tag 137 applies to the audio track based on the third classifications (step s15.16).
  • the tag 137 is associated with the track (step s15.17). The process ends at s15.18.

Abstract

An apparatus comprises at least one processor and at least one memory including computer program code, the at least one memory and the computer program code configured with the processor to cause the apparatus to perform the following steps. In one step, the apparatus receives a plurality of audio signals representing captured audio from respective audio sources in a space. In another step, the apparatus processes the received audio signals to generate a spatial audio signal for playback. The processing comprises identifying a subset of the audio sources, and responsive to a detected change in spatial position of one or more of said subset of audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.

Description

    Field
  • This relates to distributed audio mixing. The further relates to, but it not limited to, methods and apparatus for distributed audio capture, mixing and rendering of spatial audio signals to enable spatial reproduction of audio signals.
  • Background
  • Spatial audio refers to playable audio data that exploits sound localisation. In a real world space, for example in a concert hall, there will be multiple audio sources, for example the different members of an orchestra or band, located at different locations on the stage. The location and movement of the sound sources is a parameter of the captured audio. In rendering the audio as spatial audio for playback such parameters are incorporated in the data using processing algorithms so that the listener is provided with an immersive and spatially oriented experience.
  • Nokia's Spatial Audio Capture (SPAC) is an example technology for processing audio captured via a microphone array into spatial audio; that is audio with a spatial percept. The intention is to capture audio so that when it is rendered to a user the user will experience the sound field as if they are present at the location of the capture device.
  • An example application of spatial audio is in virtual reality (VR) whereby both video and audio data is captured within a real world space. In the rendered version of the space, i.e. the virtual space, the user, through a VR headset, can view and listen to the captured video and audio which has a spatial percept.
  • Rendering of captured audio can be complex and the quality of spatial audio produced in terms of listener experience can be degraded in current systems, for example in situations where one or more audio sources change position or move during capture.
  • Summary
  • According to one aspect, a method comprises: receiving a plurality of audio signals representing captured audio from respective audio sources in a space; processing the received audio signals to generate a spatial audio signal, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • The method may further comprise: defining a group of audio sources from the plurality of audio signals; and wherein the change in the spatial position in the generated spatial audio signal depends on the position of one or more of the other audio sources in said group of audio sources.
  • The method may further comprise receiving a spatial audio signal associated with a microphone array configured to provide spatial audio capture, and wherein the processing comprises generating an updated version of the spatial audio signal associated with the microphone array.
  • The processing may further comprise identifying a subset of the audio sources, and the detecting and generating steps are performed only in relation to the subset. The method may further comprise identifying the subset manually through a user interface. Identifying the subset may be performed automatically based on attributes of the audio sources. Identifying the subset may be based on the position of one or more sources in the space. Identifying the subset may be based on the position of the one or more sources relative to a primary source whose position is permitted to change in the signal. Identifying the subset may be based on the position of the one or more sources relative to the position of the microphone array. The method may further comprise receiving for each audio source positional data received from, or derived from, a positioning tag carried by the audio source. Identifying the subset may be determined based on the type of audio source. The type of audio source may be determined by signal analysis of the audio signals from each audio source. Analysis of the audio signals may identify a type of vocal performance.
  • Analysis of the audio signals may identify a type of instrument.
  • Determination of the audio source type may be performed by receiving identification data from a tag associated with the audio source.
  • Generating the spatial audio signal may comprise controlling the change in spatial position in accordance with one or more position modification rules. Selection of a position modification rule may be received from a user through a user interface. The user interface may display a spatial representation of each audio source together with the one or more selectable position modification rules for each. The, or one of said position modification rules may prevent a change in spatial position. The, or one of said position modification rules may comprise calculating an average detected change in spatial position over a predetermined time interval and wherein the change in spatial position in the signal may be based on the calculated average.
  • Identifying the subset may be determined by first identifying a group of associated audio sources based on spatial position, identifying a substantially common or average change in position for members of the group, and responsive to detecting a deviation of one or more members of the group from the average change in position, generating the spatial audio signal such that said deviation is not present therein. Generating the spatial audio signal may comprise changing the position of the deviating members to substantially track the common or average change in position. Generating the spatial audio signal may comprise setting the change in position at a single location, substantially at the centre of the group. Generating the spatial audio signal may comprise setting the change in position for all members of the group to be equidistant from the centre of the group.
  • According to a second aspect, an apparatus comprises at least one processor and at least one memory including computer program code, the at least one memory and the computer program code configured with the processor to cause the apparatus to: receive a plurality of audio signals representing captured audio from respective audio sources in a space; process the received audio signals to generate a spatial audio signal, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generate the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • The processor may be configured to define a group of audio sources from the plurality of audio signals, and wherein the change in the spatial position in the generated spatial audio signal depends on the position of one or more of the other audio sources in said group of audio sources.
  • The processor may further cause the apparatus to receive a spatial audio signal associated with a microphone array configured to provide spatial audio capture, and to generate an augmented version of the spatial audio signal associated with the microphone array.
  • The apparatus may identify the subset is performed by receiving identification data entered manually through a user interface. Identifying the subset may be performed automatically based on attributes of the audio sources. Identifying the subset may be based on the position of one or more sources in the space. Identifying the subset may be based on the position of the one or more sources relative to a primary source whose position is permitted to change in the signal. Identifying the subset may be based on the position of the one or more sources relative to the position of the microphone array.
  • The processor may be further configured to receive for each audio source positional data received from, or derived from, a positioning tag carried by the audio source.
  • Identifying the subset may be determined based on the type of audio source. The type of audio source may be determined by means of the processor performing signal analysis of the audio signals from each audio source. The analysis of the audio signals may identify a type of vocal performance. The analysis of the audio signals may identify a type of instrument. Determination of the audio source type may be performed by receiving identification data from a tag associated with the audio source.
  • The processor may be arranged to generate the spatial audio signal by controlling the change in spatial position in accordance with one or more position modification rules.
  • Selection of the position modification rule may be received from a user through a user interface provided by the processor.
  • The processor may be configured to display a spatial representation of each audio source together with the one or more selectable position modification rules for each.
  • The, or one of said position modification rules may prevent a change in spatial position.
  • The, or one of said position modification rules may comprise calculating an average detected change in spatial position over a predetermined time interval and wherein the change in spatial position in the signal is based on the calculated average.
  • The processor may be configured to identify the subset by first identifying a group of associated audio sources based on spatial position, to identify a substantially common or average change in position for members of the group, and responsive to detecting a deviation of one or more members of the group from the average change in position, to generate the spatial audio signal such that said deviation is not present therein.
  • The processor may be configured to generate the spatial audio signal by changing the position of the deviating members to substantially track the common or average change in position.
  • The processor may be configured to generate the spatial audio signal by setting the change in position at a single location, substantially at the centre of the group.
  • The processor may be configured to generate the spatial audio signal by setting the change in position for all members of the group to be equidistant from the centre of the group.
  • The apparatus may be further configured to receive spatial video signals in association with the audio signals.
  • The processor may be configured to identify the subset of audio sources based on one or both of the audio signals and the video signals.
  • According to a third aspect, an apparatus comprises at least one processor and at least one memory including computer program code, the at least one memory and the computer program code configured with the processor to cause the apparatus to: receive a plurality of audio signals representing captured audio from respective audio sources in a space; process the received audio signals to generate a spatial audio signal for playback, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • According to a fourth aspect, a computer program comprises instructions that when executed by a computer apparatus control it to perform the method of any preceding definition
  • According to a fifth aspect, a non-transitory computer-readable storage medium is provided having stored thereon computer-readable code, which, when executed by at least one processor, causes the at least one processor to perform a method, comprising: receiving a plurality of audio signals representing captured audio from respective audio sources in a space; processing the received audio signals to generate a spatial audio signal for playback, in which the processing comprises: detecting spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  • According to a sixth aspect, apparatus is provided, the apparatus having at least one processor and at least one memory having computer-readable code stored thereon which when executed controls the at least one processor: to receive a plurality of audio signals representing captured audio from respective audio sources in a space; to process the received audio signals to generate a spatial audio signal for playback, in which the processing comprises:
    • to detect spatial positions of the audio sources; and responsive to a detected change in spatial position of one or more of said audio sources, to generate the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
    Brief Description of the Drawings
  • Embodiments will now be described, by way of non-limiting example, with reference to the accompanying drawings, in which:
    • Figure 1 is a schematic representation of a distributed audio capture scenario, including use of a rendering apparatus according to embodiments;
    • Figure 2 is a schematic diagram illustrating components of the Figure 1 rendering apparatus;
    • Figure 3 is a flow diagram showing method steps of audio capture and rendering according to an embodiment;
    • Figure 4a is a graphical representation of a real-world space in which audio sources may move;
    • Figure 4b is a screenshot of a User Interface (UI) indicating resulting filtered movement in the spatial sound field resulting from use of embodiments herein;
    • Figure 5 is a flow diagram showing method steps in one example filtering operation;
    • Figure 6 is a screenshot of a UI indicating how a user can select the Figure 5 filtering;
    • Figure 7 is a screenshot of a UI indicating another example filtering operation;
    • Figure 8 is a flow diagram showing method steps in the Figure 7 filtering operation;
    • Figure 9 is a screenshot of a UI indicating another example filtering operation;
    • Figure 10 is a flow diagram showing method steps in the Figure 9 filtering operation;
    • Figure 11 is a screenshot of a UI indicating another example filtering operation;
    • Figures 12a and 12b are schematic representations indicating different methods of filtering movement in virtual space for associated groups of audio sources;
    • Figure 13 is a schematic diagram indicating functional components used in capture and rendering stages of an apparatus according to embodiments;
    • Figure 14 is a an overview of an example method of determining characteristics of an audio signal, which may be useful in the method steps of any embodiment herein;
    • Figure 15 is a flowchart of a method according to Figure 14; and
    • Figure 16 is an overview of a process for obtaining multiple types of acoustic features in the method of claim Figure 15.
    Detailed Description of Embodiments
  • Embodiments herein relate to systems and methods relating to the capture and rendering of spatial audio data for playback. In particular, the embodiments relate to distributed spatial and rendering methods in which there are multiple audio sources which may move within a virtual space over time. Each audio source generates respective audio signals and, in some embodiments, positioning information for use by the system. An example application is in a VR capture and rendering system in which video is also captured and rendered to provide an immersive user experience. Nokia's OZO (RTM) VR camera is used as an example of a VR capture device which comprises a microphone array to provide a spatial audio signal, but it will be appreciated that the embodiments are not limited to VR applications nor the use of microphone arrays at the capture point.
  • Referring to Figure 1, an overview of a VR capture scenario 1 is shown together with a first embodiment capture and rendering system (CRS) 15 with associated user interface 16. The Figure shows in plan-view a real world space 3 which may be for example a concert hall or other music venue. The CRS 15 is applicable to any real world space, however. A VR device 6 for video and spatial audio capture is supported on a floor 5 of the space 3 in front of multiple audio sources, in this case a band; the position of the VR device 6 is known, e.g. through predetermined positional data or signals derived from a positioning tag on the VR device (not shown). The VR device 6 comprises a microphone array configured to provide spatial audio capture.
  • The band may be comprised of multiple members each of which has an associated external microphone or (in the case of guitarists) a pick-up feed providing audio signals. Each may therefore be termed an audio source for convenience. In other embodiments, other types of audio source can be used. The audio sources in this case comprise a lead vocalist 7, a drummer 8, lead guitarist 9, bass guitarist 10, and three members of a choir or backing singers 11, 12, 13 which members are spatially close together in a group.
  • As well as having an associated microphone or audio feed, the audio sources 7 - 13 carry a positioning tag which can be any module capable of indicating through data its respective spatial position to the CRS 15. For example the positioning tag may be a high accuracy indoor positioning (HAIP) tag which works in association with one or more HAIP locators 20 within the space 3. HAIP systems use Bluetooth Low Energy (BLE) communication between the tags and the one or more locators 20. For example, there may be four HAIP locators mounted on, or placed relative to, the VR device 6. A respective HAIP locator may be to the front, left, back and right of the VR device 6. Each tag sends BLE signals from which the HAIP locators derive the tag, and therefore, audio source location.
  • In general, such direction of arrival (DoA) positioning systems are based on (i) a known location and orientation of the or each locator, and (ii) measurement of the DoA angle of the signal from the respective tag towards the locators in the locators' local co-ordinate system. Based on the location and angle information from one or more locators, the position of the tag can be calculated using geometry.
  • The CRS 15 is a processing system having an associated user interface (UI) 16 which will explained in further detail below. As shown in Figure 1, it receives as input from the capture device 6 spatial audio and video data, and positioning data, through a signal line 17. Alternatively, the positioning data can be received from the HAIP locator 20. The CRS 15 also receives as input from each of the audio sources 7 - 13 audio data and positioning data from the respective positioning tags, or the HAIP locator 20, through separate signal lines 18. The CRS 15 generates spatial audio data for output to a user device 19, such as a VR headset with video and audio output.
  • The input audio data can be multichannel audio in loudspeaker format, e.g. stereo signals, 4.0 signals, 5.1 signals, Dolby Atmos (RTM) signals or the like. Instead of loudspeaker format audio, the input can be in the multi microphone signal format, such as the raw eight signal input from the OZO VR camera, if used for the capture device 6.
  • Figure 2 shows an example schematic diagram of components of the CRS 15. The CRS 15 has a controller 22, a touch sensitive display 24 comprised of a display part 26 and a tactile interface part 28, hardware keys 30, a memory 32, RAM 34 and an input interface 36. The controller 22 is connected to each of the other components in order to control operation thereof. The touch sensitive display 24 is optional, and as an alternative a conventional display may be used with the hardware keys 30 and/or a mouse peripheral used to control the CRS 15 by conventional means.
  • The memory 32 may be a non-volatile memory such as read only memory (ROM) a hard disk drive (HDD) or a solid state drive (SSD). The memory 32 stores, amongst other things, an operating system 38 and software applications 40. The RAM 34 is used by the controller 22 for the temporary storage of data. The operating system 38 may contain code which, when executed by the controller 22 in conjunction with RAM 34, controls operation of each of hardware components of the terminal.
  • The controller 22 may take any suitable form. For instance, it may be a microcontroller, plural microcontrollers, a processor, or plural processors.
  • In embodiments herein, the software application 40 is configured to provide video and distributed spatial audio capture, mixing and rendering to generate a VR environment, or virtual space, including the rendered spatial audio. The software application 40 also provides the UI 16 shown in Figure 1, through its output to the display 24 and receives user input through the tactile interface 28 or other input peripherals such as the hardware keys 30 or a mouse (not shown). The mixing stage, which forms a sub-part of the rendering stage in this case, may be performed manually through the UI 16 or all or part of said mixing stage may be performed automatically. The software application 40 may render the virtual space, including the spatial audio, using known signal processing techniques and algorithms based on the mixing stage.
  • The input interface 36 receives video and audio data from the capture device 6, such as Nokia's OZO (RTM) device, and audio data from each of the audio sources 7 - 13. The input interface 36 also receives the positioning data from (or derived from) the positioning tags on each of the capture device 6 and the audio sources 7 - 13, from which can be made an accurate determination of their respective positions in the real world space 3.
  • The software application 40 may be configured to operate in any of real-time, near real-time or even offline using pre-stored captured data.
  • One example aspect of the mixing stage of software application 40 is controlling how audio sources move, or change position, in the rendered virtual space responsive to detected movement in the captured real world space 3. In this respect, during capture it is sometimes the case that audio sources move. For example, in the Figure 1 situation, any one of the audio sources 7 - 13 may move over time, as therefore will their respective audio position with respect to the capture device 6 and also to each other. Commonly, users may be used to static sources where the audio source is generally central. When audio sources move, the rendered result may be overwhelming and distracting.
  • To handle real world movement, the software application 40 may be configured to restrict or filter movement of a subset of the audio sources 7 - 13 when rendering the spatial audio part of the virtual space. This restriction / filtering is performed based on a number of attributes, to be described below, some of which can be set manually through the UI 16 and some of which can be handled automatically by predetermined algorithms.
  • Figure 3 shows an overview flow diagram of the capture, mixing and rendering stages of software application 40. As mentioned, the mixing and rendering stages may be combined. First, video and audio capture is performed in step 3.1; next mixing is performed in step 3.2, followed by rendering in step 3.3. Mixing (step 3.2) may be dependent on a manual or automatic control step 3.4 which may be based on attributes of the captured video and/or audio. Other attributes may be used.
  • Examples of how the mixing step 3.2 may be performed will now be described.
  • In a first example, a determination may be made as to the one or most important audio sources 7 - 13 at each point in time. The remaining audio sources are assigned to a subset subject to filtering or restriction, meaning that their position in virtual space does not track real-world movement. The most important audio sources are permitted to move, meaning that the spatial position of their corresponding audio signals also move, typically to closely track the real world movement. Filtering or restriction of the remaining subset may mean preventing their movement completely (freezing / no movement) or limiting movement by filtering, e.g. using time-based averaging.
  • Referring to Figure 4a, which is similar to Figure 1, in this real world scenario the CRS 15 may determine that the lead vocalist 7 is the most important and therefore its captured sound is allowed to move in the rendered virtual space 42, represented graphically in Figure 4b. The remaining audio sources 8 - 13 are in this case frozen or restricted in terms of rendered movement. Figure 4b shows the result of the lead vocalist 7 and the lead guitarist 9 moving during capture, which is indicated in Figure 4a by the dotted arrows. The other audio sources 8, 10, 11, 12, 13 are also frozen from movement.
  • Figure 5 shows the processing steps performed by the CRS 15 through software application 40. The capturing step 5.1 is followed by the step 5.2 of identifying the most important source (MIS). Step 5.3 mixes the audio so as to allow movement of the MIS and to prevent movement of the other audio sources. Rendering takes place in step 5.4 in accordance with said mixing step 5.3. The feedback arrow indicates that the steps 5.1 - 5.4 may be repeated over time.
  • Determination of the MIS can be done using various methods. One method is for the CRS 15 to analyse the position of each audio source 7 - 13 by means of the positioning tags, and/or using the captured video data from capture device 6. The front and centre positioning of the audio source 7 in this case is indicative of its importance relative to the others. Alternatively, or additionally, the spatial audio signal from the capture device 6 and/or audio signals from each of the audio sources 7 - 13 can be analysed by the CRS 15 to determine the most important sound. For example, the sound indicative of the lead melody or vocals may be indicative of the most important source.
  • The MIS may also be identified manually through the UI 16. Figure 6 shows an example screenshot of the UI 16 which shows the detected positions of the audio sources within the space 3, the UI permitting manual selection of whether a source is movable (M) or fixed/frozen (F). Clicking or touching on a source may toggle between M and F or a drop down menu can provide the options. Further filtering options can be provided.
  • In a second example, one or more anchor sources is or are determined and their position is fixed or filtered in a manner similar to the first example. For example, referring to the UI 16 representation of Figure 7, audio sources corresponding to drums 8, lead guitar 9 and bass guitar 10 are identified and restricted. Restriction here may mean freezing or fixing the position. The respective positions of these audio sources 8 - 10 may still be obtained and tracked using the tag-based positioning, and rather than fixing or freezing their position, their position may be filtered using a long term average position of each tag. In this context, long term may mean one minute, but shorter or longer periods may be used. The average position after one minute, to take this example, may indicate an allowed degree of movement in the rendered version. Figure 8 shows the main processing steps, comprised of capture (step 8.1), identifying anchor sources (step 8.2), mixing (step 8.3) and rendering (step 8.4).
  • As before, setting the anchor sources may be performed automatically by position, audio analysis or by other means, or manually through the UI 16. Filtering parameters may also be set through the UI 16, e.g. to set the averaging period.
  • In a third example, source movement may depend on the input signal type over time. Referring to Figure 9, for example, the lead guitarist 9 may be playing a guitar solo at a point in time and so is considered an important source and allowed to move, as is the lead vocalist 7. When the guitar solo ends, and reverts back to an accompanied performance, the source may become frozen or filtered as before.
  • As another example, the vocalisation type for the lead vocalist 7 may change over time; speaking or whistling can be identified by signal analysis and at this time the source position is frozen or time-filtered as explained above. When the vocalisation type changes to singing, normal movement may be enabled.
  • A method for detecting singing from non-singing type vocalisation is described in Berenzweig, A.L. and Ellis, D.P.W. "Locating Singing Voice Segments within Music Signals" Applications of Signal Processing to Audio and Acoustics, 2001, IEEE Workshop on the, 2001,119-122.
  • This may be used to identify a speech-type monologue from singing, which may be the basis of MIS determination.
  • A further method of identifying audio signal type, e.g. for identifying singing, speaking, a change in vocalisation or the presence of a particular instrument or musical genre, is described in PCT/F12014/051036 the contents of which are incorporated herein by reference. An overview of the system and method described therein is described later on.
  • A generalised method may comprise training a model, for example a Gaussian mixture model or a neural network for the classes to be classified, e.g. singing, speaking, shouting etc. For the training stage, a set of audio files is obtained where the vocal type is known. From the audio files, a set of representative features are extracted; suitable features may be, for example, mel-frequency cepstral coefficients and the fundamental frequency (fo). Also, other features related to, for example, voice characteristics can be used.
  • Using the features, the models are then trained to represent typical feature values of different vocalisation types, in the case of a Gaussian mixture model, or a neural network is trained to discriminate between the different vocalisation types given the features. Online classification is then performed by extracting the same features from the input signal, and then evaluating which model has the highest likelihood of having generated the features (Gaussian mixture model) or which category gives the largest output value (neural network).
  • The actions to be taken responsive to signal type can be set using the mixing UI 16. The UI 16 can enable the operator to identify input signal types and associate each with different restrictions or filtered movements, for example. The manner in which the mixing operator sets the restrictions can cater for a variety of situations. For creating fast mixes, e.g. for live situations, the system may have a set of presets which can be applied. The behaviour indicated in Figure 9 may be one example preset. Figure 10 shows the main processing steps, comprised of capture (step 10.1), identification of signal type (step 10.2), mixing based on presets or rules (step 10.3) and rendering (step 10.4). In this case, the rules in step 10.3 assign certain types of signal to be moved and other types to be restricted; this represents a first preset, which can itself be modified over time to apply different rules.
  • In a fourth example, the CRS 15 is configured to filter or restrict movement within a group of associated audio sources. Referring to Figure 11, the choir members 11 - 13 are spatially close to one another and identified to the CRS 15 manually or automatically as a group. Each of the choir members 11 - 13 has a positioning tag, but the CRS 15 may be configured to receive, or generate, a single source position which may be the middle or mean position of the group members, notwithstanding any movement or change in position. This may be desirable if we wish to have a single sound source that represents the whole group rather than using individual spatial positions. Alternatively, the individual sounds from each group member 11 - 13 may be adjusted so that they are equidistant from the group centre.
  • In a fifth example, the CRS 15 is again configured to filter or restrict movement with a group of associated audio sources, and in this case the CRS selectively filters the group movement so that coherent movement (all members moving in generally the same direction) is allowed in the rendered output, but movement in incoherent (different) directions is filtered, frozen and/or adjusted. This may be performed by determining a reference movement direction and subsequently filtering movement based on deviation of one or more group members. As an example, the CRS 15 may measure the most common movement direction and identifies if one or more group members deviate from the common direction.
  • Referring to Figure 12a, which shows only the group of choir audio sources 11 - 13 from Figure 1, the CRS 15 when detecting coherent movement allows that movement in the rendered version as shown. Referring to Figure 12b, the CRS 15 when detecting that one or more members (in this case the choir source 13) deviates away from the common movement, moves that member in the direction of common movement in contrast to their actual movement.
  • The CRS 15 may be user adjustable. For example, when there is a sufficient amount of coherent movement, all movement from the selected group is rendered. Other movement functions may be provided.
  • In the UI 16, the locator tags as before can be visualised and grouped. The mixer operator may select tags to form a group and then select an appropriate function for that group. Another example is to draw a rectangle or other shape to select tags within the area and then select an appropriate function for that group.
  • In a sixth example, the CRS 15 can filter movement of audio sources 7 - 13 based on the listener location within the virtual space. The CRS 15 can in real time determine the movement direction and/or velocity with reference to the listener location and through audio processing determine if it is audibly acceptable or pleasing, or not. For example, movement of a particular source when the listener is positioned directly in front of that source may be acceptable. However, if the listener were to move closer to, and to one side of, the same source, this may result in significant sideways movement producing unacceptable results. In this case, filtering can be applied, for example to reposition the audio source front and centre of the listener's position.
  • Referring now to Figure 13, an example schematic representation of capture and rendering parts of the Figure 1 scenario is shown, some parts of which are provided by the CRS 15, in terms of functional components for mixing and rendering. For ease of explanation, the Figure 13 shows separate capture and rendering parts 101, 103 with mixing being performed as part of the rendering part.
  • In this example, there is shown only one audio signal source, indicated by reference numeral 111; however in practice there will be multiple such sources each of which may move in real world space. The audio source is in this example a Lavalier microphone 111 associated with a movable person, e.g. a singer.
  • The Lavalier microphone 111 is an example of a 'close' audio source capture apparatus and may in some embodiments be a boom microphone or similar neighbouring microphone capture system. Although the following examples are described with respect to a Lavalier microphone and thus a Lavalier audio signal the concept may be extended to any microphone external or separate to e.g. the microphones or array of microphones of the capture device 5 configured to capture the spatial audio signal. Thus the concept is applicable to any external/additional microphones in addition to a spatial audio capture (SPAC) microphone array, used in the OZO device, be they Lavalier microphones, hand held microphones, mounted mics, etc. The external microphones can be worn/carried by persons or mounted as close-up microphones for instruments or a microphone in some relevant location which the designer wishes to capture accurately. The Lavalier microphone 111 may in some embodiments be a microphone array. The Lavalier microphone typically comprises a small microphone worn around the ear or otherwise close to the mouth. For other sound sources, such as musical instruments, the audio signal may be provided either by a Lavalier microphone or by an internal microphone system of the instrument (e.g., pick-up microphones in the case of an electric guitar).
  • The Lavalier microphone 111 may be configured to output the captured audio signals to a variable delay compensator 117. The Lavalier microphone may be connected to a transmitter unit (not shown), which wirelessly transmits the audio signal to a receiver unit (not shown).
  • Furthermore the capture part 101 comprises a Lavalier (or close source) microphone position tag 112. The Lavalier microphone position tag 112 may be configured to determine information identifying the position or location of the Lavalier microphone 111 or other close microphone. It is important to note that microphones worn by people may be freely moved in the acoustic space and the system supporting location sensing of wearable microphone may support continuous sensing of user or microphone location. The Lavalier microphone position tag 112 may be configured to output this determination of the position of the Lavalier microphone to a position tracker 115.
  • The capture part 101 comprises a spatial audio capture device 113 which corresponds to the capture device 5 in Figure 2. The spatial audio capture device 113 is an example of an 'audio field' capture apparatus and may in some embodiments be a directional or omnidirectional microphone array. The spatial audio capture device 113 may be configured to output the captured audio signals to a variable delay compensator 117.
  • Furthermore the capture part 101 may comprise a spatial capture position tag 114. The spatial capture position tag 114 may be configured to determine information identifying the position or location of the spatial audio capture device 113. The spatial capture position tag 114 may be configured to output this determination of the position of the spatial capture microphone to the position tracker 115. In the case the position tracker is co-located with the capture apparatus or the position of the capture apparatus with respect to the position tracker is otherwise known, and location data is obtained in relation to the capture apparatus, the capture apparatus does not need to comprise a position tag.
  • In some embodiments the spatial audio capture device 113 is implemented within a mobile device. The spatial audio capture device is thus configured to capture spatial audio, which, when rendered to a listener, enables the listener to experience the sound field as if they were present in the location of the spatial audio capture device. The Lavalier microphone 111 in such embodiments may be configured to capture high quality close-up audio signals (for example from a key person's voice, or a musical instrument). When mixed to the spatial audio field, the attributes of the key source such as gain, timbre and spatial position may be adjusted in order to provide the listener with a much more realistic immersive experience. In addition, it is possible to produce more point-like auditory objects, thus increasing the engagement and intelligibility.
  • The capture part 101 furthermore may comprise a position tracker 115. The position tracker 115 may be configured to receive the positional tag information identifying positions of the Lavalier microphone 111 and the spatial audio capture device 113 and generate a suitable output identifying the relative position of the Lavalier microphone 111 relative to the spatial audio capture device 113 and output this to the render apparatus 103 and specifically in this example to a location date filtering model 120. Furthermore in some embodiments the position tracker 115 may be configured to output the tracked position information to a variable delay compensator 117.
  • Thus in some embodiments the locations of the Lavalier microphones (or the persons carrying them) with respect to the spatial audio capture device 113 may be tracked and used for mixing the sources to correct spatial positions. In some embodiments the position tags namely, the microphone position tag 112 and the spatial capture position tag 114 may be implemented using High Accuracy Indoor Positioning (HAIP) or another suitable indoor positioning technology. In some embodiments, in addition to or instead of HAIP, the position tracker may use video content analysis and/or sound source localization, as will be explained below.
  • The capture part 101 furthermore may comprise a variable delay compensator 117 configured to receive the outputs of the Lavalier microphone 111 and the spatial audio capture device 113. Furthermore in some embodiments the variable delay compensator 117 may be configured to receive source position and tracking information from the position tracker 115. The variable delay compensator 117 may be configured to determine any timing mismatch or lack of synchronisation between the close audio source signals and the spatial capture audio signals and determine the timing delay which would be required to restore synchronisation between the signals. In some embodiments the variable delay compensator 117 may be configured to apply the delay to one of the signals before outputting the signals to the render apparatus 103 and specifically in this example to the audio renderer 121. The timing delay may be referred as being a positive time delay or a negative time delay with respect to an audio signal. For example, denote a first (spatial) audio signal by x, and another (Lavalier) audio signal by y. The variable delay compensator 117 is configured to try to find a delay T, such that x(n)=y(n-T). Here, the delay T can be either positive or negative.
  • The capture stage 101 furthermore may comprise a camera 118 configured to capture video signals; the camera 118 may be a single camera or array of cameras, as in the OZO device example, in which spatial video signals are produced. In some embodiments, the camera 118 may form part of the same physical unit as the spatial audio capture device 113.
  • The capture stage 101 further may comprise a content analysis module 119 which receives both the video signals from the camera 118 and audio from the variable time delay compensator 117, which corrects for any timing mismatch or lack of synchronisation between the close audio signals, the spatial capture audio signals and the video signals. The output from the content analysis module 119 is provided to the rendering stage 103.
  • Also provided to the rendering stage 103 are signals from the position tracker 115 and the signals from the time delay compensator 117.
  • In some embodiments, the render part 103 comprises a head tracker 123. The head tracker 123 may be any suitable means for generating a positional input, for example a sensor attached to a set of headphones configured to monitor the orientation of the listener, with respect to a defined or reference orientation and provide a value or input which can be used by the audio renderer 121. The head tracker 123 may in some embodiments be implemented by at least one gyroscope and/or digital compass.
  • In some embodiments, the render part 103 comprises a location data filtering module 120 which takes as input the unfiltered HAIP location data from the position tracker 115, and the audio-visual content analysis data from the content analysis module 119.
  • The render part 103 comprises an audio renderer 121. The audio renderer 121 is configured to receive output for the location data filtering module 120. The audio renderer 121 can furthermore be configured to receive an input from the head tracker 123. Furthermore the audio renderer 121 can be configured to receive other user inputs for example entering manually through the UI 16. The audio renderer 121, as described herein in further detail later, can be configured to mix together the audio signals, namely the Lavalier microphone audio signals and the spatial audio signals based on the positional information and/or the head tracker inputs in order to generate a mixed audio signal in accordance with the methods and rules described above to cater for detected movement over time. The mixed audio signal can for example be passed to headphones 125. However the output mixed audio signal can be passed to any other suitable audio system for playback (for example a 5.1 channel audio amplifier).
  • In some embodiments the audio renderer 121 may be configured to perform spatial audio processing on the audio signals from the microphone array and from the close microphone.
  • The Lavalier audio signal is from the Lavalier microphones and the spatial audio captured by the microphone array and processed with the spatial analysis may in some embodiments be combined by the audio renderer to a single binaural output which can be listened through headphones.
  • The spatial audio signal may be converted into a multichannel signal. The multichannel output may then be binaurally rendered, and summed with binaurally rendered Lavalier source signals.
  • The rendering may be described initially with respect to a single (mono) channel, which can be one of the multichannel signals from the spatial audio signal or one of the Lavalier sources. Each channel in the multichannel signal set may be processed in a similar manner, with the treatment for Lavalier audio signals and multichannel signals having the following differences:
    1. 1) The Lavalier audio signals have time-varying location data (direction of arrival and distance) whereas the multichannel signals are rendered from a fixed location.
    2. 2) The ratio between synthesized "direct" and "ambient" components may be used to control the distance perception for Lavalier sources, whereas the multichannel signals are rendered with a fixed ratio.
    3. 3) The gain of Lavalier signals may be adjusted by the user whereas the gain for multichannel signals is kept constant.
  • The render part 103 in some embodiments may comprise headphones 125. The headphones can be used by the listener to generate the audio experience using the output from the audio renderer 121.
  • Thus based on the location tracking, the Lavalier microphone signals can be mixed to suitable spatial positions in the spatial audio field in accordance with predetermined rules or user input. The rendering can be done by rendering the spatial audio signal using virtual loudspeakers with fixed positions, and the captured Lavalier source is rendered from a time varying position. Thus, the audio renderer 121 is configured to control the azimuth, elevation, and distance of the Lavalier or close source based on the tracked position data and the filtering rules.
  • Moreover, the user may be allowed to adjust the gain and/or spatial position of the Lavalier source using the output from the head-tracker 123. For example by moving the listeners head the head-tracker input may affect the mix of the Lavalier source relative to the spatial sound. This may be changing the 'spatial position' of the Lavalier source based on the head-tracker or by changing the gain of the Lavalier source where the head-tracker input is indicating that the listener's head is 'towards' or 'focussing' on a specific source. Thus the mixing/rendering may be dependent on the relative position/orientation of the Lavalier source and the spatial microphones but also be dependent on the orientation of the head as measured by the head-tracker. In some embodiments the user input may be any suitable user interface input, such as an input from a touchscreen indicating the listening direction or orientation.
  • Alternatively to a binaural rendering (for headphones), a spatial downmix into a 5.1 channel format or other format could be employed. In this case, the Lavalier or close source can in some embodiments mixed to its 'proper' spatial position using known amplitude panning techniques.
  • For completeness, an overview of one method for identifying audio signal type is now given with reference to Figures 14 to 16. Further details are provided in in PCT/FI2014/051036 . The method may be used in the MIS identification step previously described. This method generally involves:
    1. (i) determining one or more acoustic features of audio data,
    2. (ii) generating at least one first classification based on the one or more determined acoustic features using a first classifier,
    3. (iii) generating at least one second classification based on the one or more determined acoustic features using at least one second classifier, which is different from the first,
    4. (iv) generating a third classification based on said first and second classifications using a third classifier, and
    5. (v) providing classification data, or tags, for the audio data based on said third classification.
  • Any one of the classifications can be used alongside selection rules to determine the MIS.
  • The first classifier and the third classifier may be a non-probabilistic classifier, such as a support vector machine (SVM) classifier. The second classifier may be a probabilistic classifier, for example based on one or more Gaussian Mixture Models (GMMs). The classification data (or tags, as they are described in the reference) may indicate one or more of the following characteristics for determining the MIS using the predefined rules: a musical instrument, the presence or absence of vocals and/or a vocalist gender, presence or absence of music and a musical genre. This assumes that the audio data is music. In other embodiments, the audio data may comprise spoken word elements or a combination of music and spoken word elements.
  • Figure 14 is an overview of a method of determination of tag information for the audio track. Acoustic features 131 of the audio are extracted and input to first level classifiers 132 to generate first level classifications for the audio track. In this example, first classifiers 133 and second classifiers 134 are used to generate first and second classifications respectively. In the embodiments to be described below, the first classifiers 133 are non-probabilistic classifiers, while the second classifiers 134 are probabilistic classifiers. The first and second classifications generated by the first level classifiers 132 are provided as inputs to a second level classifier 135. One or more second level classifications are generated by the second level classifier 135, based at least in part on the first and second classifications. In the embodiments to be described below, the second level classifiers 135 include a third classifier 136, which outputs a third classification. One or more tags 137 are generated, based on the second level classifications. Such tags 137 may be stored by a tagging module 138 for determining the MIS.
  • The method will now be described in more detail, with reference to Figures 15 and 16.
  • Beginning at step s15.0 of Figure 15, if the received input signal is in a compressed format, such as MPEG-1 Audio Layer 3 (MP3), Advanced Audio Coding (AAC) and so on, the input signal is decoded into pulse code modulation (PCM) data (step s15.1). In this particular example, the samples for decoding are taken at a rate of 44.1 kHz and have a resolution of 16 bits.
  • Next, acoustic features 131 or descriptors are extracted which indicate characteristics of the audio track (step s13.2). In this particular embodiment, the features 131 are based on mel-frequency cepstral coefficients (MFCCs). In other embodiments, other features such as fluctuation pattern and danceability features, beats per minute (BPM) and related features, chorus features and other features may be used instead of, or as well as MFCCs. An example method for extracting acoustic features 131 from the input signal at step s15.2 is described in detail in PCT/FI2014/051036 . As described therein, the audio features 131 produced at step 15.2 may include one or more of:
    • a MFCC matrix for the audio track;
    • first and, optionally, second time derivatives of the MFCCs, also referred to as "delta MFCCs";
    • a mean of the MFCCs of the audio track;
    • a covariance matrix of the MFCCs of the audio track;
    • an average of mel-band energies over the audio track, based on output from the channels of the filter bank obtained;
    • a standard deviation of the mel-band energies over the audio track;
    • an average logarithmic energy over the frames of the audio track, obtained as an average of cmel(O) over a period of time obtained;
    • a standard deviation of the logarithmic energy;
    • a median fluctuation pattern over the song;
    • a fluctuation pattern bass feature;
    • a fluctuation pattern gravity feature;
    • a fluctuation pattern focus feature;
    • a fluctuation pattern maximum feature;
    • a fluctuation pattern sum feature;
    • a fluctuation pattern aggressiveness feature;
    • a fluctuation pattern low-frequency domination feature;
    • a danceability feature (detrended fluctuation analysis exponent for at least one predetermined time scale);and
    • a club-likeness value.
    • an average of an accent signal in a low, or lowest, frequency band;
    • a standard deviation of said accent signal;
    • a maximum value of a median or mean of periodicity vectors;
    • a sum of values of the median or mean of the periodicity vectors;
      tempo indicator for indicating whether a tempo identified for the input signal is considered constant, or essentially constant, or is considered non-constant, or ambiguous;
    • a first BPM estimate and its confidence;
    • a second BPM estimate and its confidence;
    • a tracked BPM estimate over the audio track and its variation;
    • a BPM estimate from a lightweight tempo estimator.
  • Example techniques for beat tracking, using accent information, are disclosed in US published patent application no. 2007/240558 A1 , US patent application no. 14/302,057 , International (PCT) published patent application nos. WO2013/164661 A1 and WO2014/001849 A1 , the disclosures of which are hereby incorporated by reference. In one example beat tracking method, described in GB 1401626.5 , one or more accent signals are derived from the input signal, for detection of events and/or changes in the audio track. A first one of the accent signals may be a chroma accent signal based on fundamental frequency Fo salience estimation, while a second one of the accent signals may be based on a multi-rate filter bank decomposition of the input signal. A BPM estimate may be obtained based on a periodicity analysis for extraction of a sequence of periodicity vectors on the basis of the accent signals, where each periodicity vector includes a plurality of periodicity values, each periodicity value describing the strength of periodicity for a respective period length, or "lag". A point-wise mean or median of the periodicity vectors over time may be used to indicate a single representative periodicity vector over a time period of the audio track. For example, the time period may be over the whole duration of the audio track. Then, an analysis can be performed on the periodicity vector to determine a most likely tempo for the audio track. One example approach comprises performing k-nearest neighbours regression to determine the tempo. In this case, the system is trained with representative music tracks with known tempo. The k-nearest neighbours regression is then used to predict the tempo value of the audio track based on the tempi of k-nearest representative tracks. More details of such an approach have been described in Eronen, Klapuri, "Music Tempo Estimation With k -NN Regression", IEEE Transactions on Audio, Speech, and Language Processing, Vol. 18, , the disclosure of which is incorporated herein by reference.
  • Chorus related features that may be extracted at step s15.2 include: a chorus start time; and a chorus end time.
  • Example systems and methods that can be used to detect chorus related features are disclosed in US 2008/236371 A1 , the disclosure of which is hereby incorporated by reference in its entirety.
  • Other features that may be used as additional input include:
    • a duration of the audio track in seconds,
    • an A-weighted sound pressure level (SPL);
    • a standard deviation of the SPL;
    • an average brightness, or spectral centroid (SC), of the audio track, calculated as a spectral balancing point of a windowed FFT signal magnitude in frames of, for example, 40 ms in length;
    • a standard deviation of the brightness;
    • an average low frequency ratio (LFR), calculated as a ratio of energy of the input signal below 100Hz to total energy of the input signal, using a windowed FFT signal magnitude in 40 ms frames; and
    • a standard deviation of the low frequency ratio.
  • Figure 16 is an overview of a process of extracting multiple acoustic features 131, some or all of which may be obtained in step s15.2. Figure 16 shows how some input features are derived, at least in part, from computations of other input features. The features 131 shown in Figure 16 include the MFCCs, delta MFCCs and mel-band energies discussed above, indicated using bold text and solid lines.
  • Returning to Figure 15, in steps s15.3 to s15.10, the method produces the first level classifications, that is the first classifications and the second classifications, based on the features 131 extracted in step s15.2. Although Figure 15 shows steps s15.3 to s15.10 being performed sequentially, in other embodiments, steps s15.3 to s15.7 may be performed after, or in parallel with, steps s15.8 to s15.10.
  • The first and second classifications are generated using the first classifiers 133 and the second classifiers 134 respectively, where the first and second classifiers 133, 134 are different from one another. For instance, the first classifiers 133 may be non-probabilistic and the second classifiers 134 may be probabilistic classifiers, or vice versa. In this particular embodiment, the first classifiers 133 are support vector machine (SVM) classifiers, which are non-probabilistic. Meanwhile, the second classifiers 134 are based on one or more Gaussian Mixture Models (GMMs).
  • In step s15.3, one, some or all of the features 131 or descriptors extracted in step s15.2, to be used to produce the first classifications 133, are selected and, optionally, normalised. For example, a look up table or database may be stored for each of the first classifications to be produced, that provides a list of features to be used to generate each first classifier and statistics, such as mean and variance of the selected features, that can be used in normalisation of the extracted features 131. In such an example, the list of features is received from the look up table, and accordingly the method selects and normalises the listed features for each of the first classifications to be generated. The normalisation statistics for each first classification in the database may be determined during training of the first classifiers 133.
  • As noted above, in this example, the first classifiers 133 are SVM classifiers. The SVM classifiers 133 are trained using a database of audio tracks for which information regarding musical instruments and genre is already available. The database may include tens of thousands of tracks for each particular musical instrument that might be tagged.
  • Examples of musical instruments for which information may be provided in the database include:
    • Accordion;
    • Acoustic guitar;
    • Backing vocals;
    • Banjo;
    • Bass synthesizer;
    • Brass instruments;
    • Glockenspiel;
    • Drums;
    • Eggs;
    • Electric guitar;
    • Electric piano;
    • Guitar synthesizer;
    • Keyboards;
    • Lead vocals;
    • Organ;
    • Percussion;
    • Piano;
    • Saxophone;
    • Stringed instruments;
    • Synthesizer; and
    • Woodwind instruments.
  • The training database includes indications of genres that the audio tracks belong to, as well as indications of genres that the audio tracks do not belong to. By analysing features 131 extracted from the audio tracks in the training database, for which instruments and/or genre are known, a SVM classifier 133 can be trained to determine whether or not an audio track includes a particular instrument, for example, an electric guitar. Similarly, another SVM classifier 133 can be trained to determine whether or not the audio track belongs to a particular genre.
  • As mentioned above, the training process can include determining a selection of one or more features 131 to be used as a basis for particular first classifications and statistics for normalising those features 131. The number of features available for selection, M, may be much greater than the number of features selected for determining a particular first classification, N; that is, M >> N. The selection of features 131 to be used is determined iteratively, based on a development set of audio tracks for which the relevant instrument or genre information is available. The normalisation of the selected features 131 at step s15.3 is optional. Where provided, the normalization of the selected features 131 in step s15.3 may potentially improve the accuracy of the first classifications.
  • At step s15.4, the method defines a single "feature vector" for each set of selected features 131 or selected combination of features 131. The feature vectors may then be normalized to have a zero mean and a variance of 1, based on statistics determined and stored during the training process.
  • At step s15.5, the method generates one or more first probabilities that the audio track has a certain characteristic, corresponding to a potential tag 137, based on the normalized transformed feature vector or vectors. A first classifier 133 is used to calculate a respective probability for each feature vector defined in step s15.4. In this manner, the number of SVM classifiers 133 corresponds to the number of characteristics or tags 137 to be predicted for the audio track.
  • In this particular example, a probability is generated for each instrument tag and for each musical genre tag to be predicted for the audio track, based on mean MFCCs and a MFCC covariance matrix. In addition, a probability may be generated based on whether the audio track is likely to be an instrumental track or a vocal track. Also, for vocal tracks, another first classification may be generated based on whether the vocals are provided by a male or female vocalist. In other embodiments, the controller may generate only one or some of these probabilities and/or calculate additional probabilities at step 15.5. The different classifications may be based on respective selections of features from the available features 131 extracted in step s15.2.
  • Optionally, a logarithmic transformation may be applied to the probabilities output by the SVM classifiers 133 (step s15.6), so that the probabilities of all the first classifications are on the same scale and the optimal predicated probability threshold may correspond to a predetermined value, such as 0.5.
  • The first classifications are then output (step s15.7). The first classifications correspond to the normalized probability pnorm that a respective one of the tags 137 to be considered applies to the audio track. The first classifications may include probabilities pinst1 that a particular instrument is included in the audio track and probabilities pgen1 that the audio track belongs to a particular genre.
  • Returning to Figure 15, in steps s15.8 to s15.10, second classifications for the input signal are based on the MFCCs and other parameters produced in step s15.2, using the second classifiers 134. In this particular example, the features 131 on which the second classifications are based are the MFCC matrix for the audio track and the first time derivatives of the MFCCs.
  • In steps s15.8 to s15.10, the probabilities of the audio track including a particular instrument or belonging to a particular genre are assessed using probabilistic models that have been trained to represent the distribution of features extracted from audio signals captured from each instrument or genre. As noted above, in this example the probabilistic models are GMMs. Such models can be trained using an expectation maximisation algorithm that iteratively adjusts the model parameters to maximise the likelihood of the model for a particular instrument or genre generating features matching one or more input features in the captured audio signals for that instrument or genre. The parameters of the trained probabilistic models may be stored in a database or in remote storage.
  • For each instrument or genre, at least one likelihood is evaluated that the respective probabilistic model could have generated the selected or transformed features from the input signal. The second classifications correspond to the models which have the largest likelihood of having generated the features of the input signal.
  • In this example, probabilities are generated for each instrument tag at step s15.8 and for each musical genre tag at step s15.9, as well as a probability whether the audio track is likely to be an instrumental track or a vocal track may also be generated. Also, for vocal tracks, another probability may be generated based on whether the vocals are provided by a male or female vocalist. In other embodiments, the method may generate only one or some of these second classifications and/or calculate additional second classifications at steps s15.8 and s15.9.
  • In this embodiment, in steps s15.8 and s15.9, probabilities pinst2 that the instrument tags will apply, or not apply, are produced by the second classifiers 134 using first and second Gaussian Mixture Models (GMMs), based on the MFCCs and their first time derivatives calculated in step s15.2. Meanwhile, probabilities pgen2 that the audio track belongs to a particular musical genre are produced by the second classifiers 134 using third GMMs. However, the first and second GMMs used to compute the instrument-based probabilities pinst2 may be trained and used slightly differently from third GMMs used to compute the genre-based probabilities gen2 , as will now be explained.
  • In the following, step s15.8 precedes step s15.9. However, in other embodiments, step s15.9 may be performed before, or in parallel with, step s15.8.
  • In this particular example, first and second GMMs are used to generate the instrument-based probabilities p inst2 (step s15.8),based on MFCC features 131 obtained in step s15.2.
  • In step s15.9, for each of the genres that may be tagged, a likelihood L is computed for the audio track belonging to that genre, based on the likelihood of each of the third GMMs being capable of outputting the MFCC feature vector of the audio track. For example, to determine which of the eighteen genres in the list hereinabove might apply to the audio track, eighteen likelihoods would be produced.
  • The genre likelihoods are then mapped to probabilities pgen2 , as follows: p gen 2 i = L i j = 1 m L j
    Figure imgb0001
    where m is the number of genre tags to be considered.
  • The second classifications, which correspond to the probabilities pinst2 and pgen2 , are then output (step s15.10).
  • The first classifications pinst1 and pgen1 and the second classifications pinst2 and pgen2 for the audio track are normalized to have a mean of zero and a variance of 1 (step s15.11) and collected to form a feature vector for input to one or more second level classifiers 135 (step s15.12). In this particular example, the second level classifiers 135 include third classifiers 136. The third classifiers 136 may be non-probabilistic classifiers, such as SVM classifiers.
  • The third classifiers 136 may be trained in a similar manner to that described above in relation to the first classifiers 133. At the training stage, the first classifiers 133 and the second classifiers 134 may be used to output probabilities for the training sets of example audio tracks from the database. The outputs from the first and second classifiers 133, 134 are then used as input data to train the third classifier 135.
  • The third classifier 136 generates determine probabilities pinst3 for whether the audio track contains a particular instrument and/or probabilities pgen3 for whether the audio track belongs to a particular genre (step s15.13). The probabilities pinst3, pgen3 are then log normalized (step s15.14), as described above in relation to the first classifications, so that a threshold of 0.5 may be applied to generate the third classifications, which are output at step s15.15.
  • The method then determines whether each instrument tag and each genre tag 137 applies to the audio track based on the third classifications (step s15.16).
  • Where it is determined that an instrument or genre tag 137 applies to the audio track (step s15.16), the tag 137 is associated with the track (step s15.17). The process ends at s15.18.
  • Further details are provided in PCT/FI2014/051036 . In the present case, the classification data, or tag 137 generated at s15.17 is used by the rules to identify the MIS. It will be appreciated however that the overview given with reference to Figures 14 to 16 is merely one of a number of possible methods used in identifying the MIS.
  • It will be appreciated that the above described embodiments are purely illustrative and are not limiting on the scope of the invention. Other variations and modifications will be apparent to persons skilled in the art upon reading the present application.
  • Moreover, the disclosure of the present application should be understood to include any novel features or any novel combination of features either explicitly or implicitly disclosed herein or any generalization thereof and during the prosecution of the present application or of any application derived therefrom, new claims may be formulated to cover any such features and/or combination of such features.

Claims (15)

  1. A method comprising:
    receiving a plurality of audio signals representing captured audio from respective audio sources in a space;
    processing the received audio signals to generate a spatial audio signal, in which the processing comprises:
    detecting spatial positions of the audio sources; and
    responsive to a detected change in spatial position of one or more of said audio sources, generating the spatial audio signal so that said change in spatial position in the generated signal is different from the detected change in spatial position for the respective audio sources.
  2. The method of claim 1, further comprising:
    defining a group of audio sources from the plurality of audio signals; and
    wherein the change in the spatial position in the generated spatial audio signal depends on the position of one or more of the other audio sources in said group of audio sources.
  3. The method of claim 1 or claim 2, further comprising receiving a spatial audio signal associated with a microphone array configured to provide spatial audio capture, and wherein the processing comprises generating an updated version of the spatial audio signal associated with the microphone array.
  4. The method of any preceding claim, wherein the processing further comprises identifying a subset of the audio sources, and the detecting and generating steps are performed only in relation to the subset.
  5. The method of claim 4, wherein identifying the subset is performed manually through a user interface.
  6. The method of claims 4, wherein identifying the subset is performed automatically based on attributes of the audio sources.
  7. The method of any of claims 4 to 6, further wherein identifying the subset is based on the position of one or more sources in the space.
  8. The method of claim 7, wherein identifying the subset is based on the position of the one or more sources relative to a primary source whose position is permitted to change in the signal.
  9. The method of claim 7 or claim 8 when dependent on claim 3, wherein identifying the subset is based on the position of the one or more sources relative to the position of the microphone array.
  10. The method of any of claims 4 to 6, wherein identifying the subset is determined based on the type of audio source.
  11. The method of claim 10, wherein the type of audio source is determined by signal analysis of the audio signals from each audio source.
  12. The method of claim 11, wherein the analysis of the audio signals identifies a type of vocal performance or instrument.
  13. The method of any preceding claim, wherein generating the spatial audio signal comprises controlling the change in spatial position in accordance with one or more position modification rules.
  14. Apparatus configured to perform the method of any preceding claim.
  15. A computer program comprising instructions that when executed by a computer apparatus control it to perform the method of any of claims 1 to 13.
EP16173264.9A 2016-06-07 2016-06-07 Distributed audio mixing Withdrawn EP3255904A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP16173264.9A EP3255904A1 (en) 2016-06-07 2016-06-07 Distributed audio mixing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
EP16173264.9A EP3255904A1 (en) 2016-06-07 2016-06-07 Distributed audio mixing

Publications (1)

Publication Number Publication Date
EP3255904A1 true EP3255904A1 (en) 2017-12-13

Family

ID=56119344

Family Applications (1)

Application Number Title Priority Date Filing Date
EP16173264.9A Withdrawn EP3255904A1 (en) 2016-06-07 2016-06-07 Distributed audio mixing

Country Status (1)

Country Link
EP (1) EP3255904A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019121864A1 (en) * 2017-12-19 2019-06-27 Koninklijke Kpn N.V. Enhanced audiovisual multiuser communication
EP3506661A1 (en) * 2017-12-29 2019-07-03 Nokia Technologies Oy An apparatus, method and computer program for providing notifications
US10448186B2 (en) 2016-12-14 2019-10-15 Nokia Technologies Oy Distributed audio mixing
WO2023285732A1 (en) * 2021-07-14 2023-01-19 Nokia Technologies Oy A method and apparatus for ar rendering adaptation

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070240558A1 (en) 2006-04-18 2007-10-18 Nokia Corporation Method, apparatus and computer program product for providing rhythm information from an audio signal
US20080236371A1 (en) 2007-03-28 2008-10-02 Nokia Corporation System and method for music data repetition functionality
US20090116652A1 (en) * 2007-11-01 2009-05-07 Nokia Corporation Focusing on a Portion of an Audio Scene for an Audio Signal
US20100223552A1 (en) * 2009-03-02 2010-09-02 Metcalf Randall B Playback Device For Generating Sound Events
WO2011020065A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. Object-oriented audio streaming system
US20110081024A1 (en) * 2009-10-05 2011-04-07 Harman International Industries, Incorporated System for spatial extraction of audio signals
WO2013164661A1 (en) 2012-04-30 2013-11-07 Nokia Corporation Evaluation of beats, chords and downbeats from a musical audio signal
WO2014001849A1 (en) 2012-06-29 2014-01-03 Nokia Corporation Audio signal analysis

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070240558A1 (en) 2006-04-18 2007-10-18 Nokia Corporation Method, apparatus and computer program product for providing rhythm information from an audio signal
US20080236371A1 (en) 2007-03-28 2008-10-02 Nokia Corporation System and method for music data repetition functionality
US20090116652A1 (en) * 2007-11-01 2009-05-07 Nokia Corporation Focusing on a Portion of an Audio Scene for an Audio Signal
US20100223552A1 (en) * 2009-03-02 2010-09-02 Metcalf Randall B Playback Device For Generating Sound Events
WO2011020065A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. Object-oriented audio streaming system
US20110081024A1 (en) * 2009-10-05 2011-04-07 Harman International Industries, Incorporated System for spatial extraction of audio signals
WO2013164661A1 (en) 2012-04-30 2013-11-07 Nokia Corporation Evaluation of beats, chords and downbeats from a musical audio signal
WO2014001849A1 (en) 2012-06-29 2014-01-03 Nokia Corporation Audio signal analysis

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BERENZWEIG, A.L.; ELLIS, D.P.W: "Locating Singing Voice Segments within Music Signals'' Applications of Signal Processing to Audio and Acoustics", IEEE WORKSHOP, 2001, pages 119 - 122
ERONEN, KLAPURI: "Music Tempo Estimation With k -NN Regression", IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, vol. 18, no. 1, pages 50 - 57, XP011329110, DOI: doi:10.1109/TASL.2009.2023165

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10448186B2 (en) 2016-12-14 2019-10-15 Nokia Technologies Oy Distributed audio mixing
WO2019121864A1 (en) * 2017-12-19 2019-06-27 Koninklijke Kpn N.V. Enhanced audiovisual multiuser communication
CN111466124A (en) * 2017-12-19 2020-07-28 皇家Kpn公司 Enhanced audiovisual multi-user communication
US11082662B2 (en) * 2017-12-19 2021-08-03 Koninklijke Kpn N.V. Enhanced audiovisual multiuser communication
CN111466124B (en) * 2017-12-19 2022-04-15 皇家Kpn公司 Method, processor system and computer readable medium for rendering an audiovisual recording of a user
EP3506661A1 (en) * 2017-12-29 2019-07-03 Nokia Technologies Oy An apparatus, method and computer program for providing notifications
WO2019130151A1 (en) * 2017-12-29 2019-07-04 Nokia Technologies Oy An apparatus, method and computer program for providing notifications
US11696085B2 (en) 2017-12-29 2023-07-04 Nokia Technologies Oy Apparatus, method and computer program for providing notifications
WO2023285732A1 (en) * 2021-07-14 2023-01-19 Nokia Technologies Oy A method and apparatus for ar rendering adaptation

Similar Documents

Publication Publication Date Title
CN108369811B (en) Distributed audio capture and mixing
Mesaros et al. Detection and classification of acoustic scenes and events: Outcome of the DCASE 2016 challenge
JP6814146B2 (en) Systems and methods for capturing and interpreting audio
Yamada et al. A rhythm practice support system with annotation-free real-time onset detection
US9532136B2 (en) Semantic audio track mixer
US20150373455A1 (en) Presenting and creating audiolinks
EP3255904A1 (en) Distributed audio mixing
US11644528B2 (en) Sound source distance estimation
Wu et al. Polyphonic pitch estimation and instrument identification by joint modeling of sustained and attack sounds
EP3255905A1 (en) Distributed audio mixing
JP4906776B2 (en) Voice control device
US20230186782A1 (en) Electronic device, method and computer program
US20200279544A1 (en) Techniques for controlling the expressive behavior of virtual instruments and related systems and methods
EP2660815B1 (en) Methods and apparatus for audio processing
CN108369800B (en) Sound processing device
WO2022070639A1 (en) Information processing device, information processing method, and program
Pimpale et al. A survey on: Sound source separation methods
CN114631142A (en) Electronic device, method, and computer program
WO2021124919A1 (en) Information processing device and method, and program
EP3613043A1 (en) Ambience generation for spatial audio mixing featuring use of original and extended signal
US20230135778A1 (en) Systems and methods for generating a mixed audio file in a digital audio workstation
Einbond et al. Embodying Spatial Sound Synthesis with AI in Two Compositions for Instruments and 3-D Electronics
Ramires Automatic Transcription of Drums and Vocalised percussion
Musick et al. Timbral hauntings: An interactive system re-interpreting the present in echoes of the past
WO2023062865A1 (en) Information processing apparatus, method, and program

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

17P Request for examination filed

Effective date: 20180612

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

17Q First examination report despatched

Effective date: 20190617

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: NOKIA TECHNOLOGIES OY

18W Application withdrawn

Effective date: 20190816