WO2023070778A1 - Audio output control method and system, and related assemblies - Google Patents

Audio output control method and system, and related assemblies Download PDF

Info

Publication number
WO2023070778A1
WO2023070778A1 PCT/CN2021/131709 CN2021131709W WO2023070778A1 WO 2023070778 A1 WO2023070778 A1 WO 2023070778A1 CN 2021131709 W CN2021131709 W CN 2021131709W WO 2023070778 A1 WO2023070778 A1 WO 2023070778A1
Authority
WO
WIPO (PCT)
Prior art keywords
head
sound
mounted device
modulation parameter
audio signal
Prior art date
Application number
PCT/CN2021/131709
Other languages
French (fr)
Chinese (zh)
Inventor
于心月
Original Assignee
歌尔股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 歌尔股份有限公司 filed Critical 歌尔股份有限公司
Publication of WO2023070778A1 publication Critical patent/WO2023070778A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/04Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/307Frequency adjustment, e.g. tone control

Definitions

  • the present application relates to the field of head-mounted devices, in particular to an audio output control method, system and related components.
  • VR Virtual Reality, virtual reality
  • VR products use computer technology to simulate a three-dimensional virtual world to create an immersive experience for users.
  • VR products have three core features: immersion, interactivity, and imagination.
  • Immersion is the most basic feature of VR products, and sound effects are one of the important indicators for perfect immersion in VR products.
  • most of the VR products on the market are dual-speaker.
  • the two speakers sound at the left and right ears.
  • the amplitude translation of the two speakers is distributed, so that the two speakers can sound according to the allocated amplitude, so as to achieve the purpose of sound image reconstruction.
  • this simple amplitude translation scheme makes the user's perception of the direction of the sound source relatively simple when using it, and it is easy to produce a localization effect in the sound source.
  • the sense of space in the sound field is poor, and the acoustic immersion effect of the product is poor.
  • the purpose of this application is to provide an audio output control method and system, a head-mounted device and a computer-readable storage medium, which uses amplitude translation to reconstruct the sense of space on the horizontal plane, and uses phase modulation to enhance the sense of space in the vertical direction. It enables the wearer to better perceive the real orientation of the sound source, avoids the localization effect of the sound source, and improves the sound field space and user experience of the headset.
  • an audio output control method including:
  • phase modulation parameters based on current head profile parameters of the wearer of the head-mounted device
  • the phase modulation parameter includes a head-related transfer function.
  • the head-mounted device includes n sound emitting devices, where n is an integer greater than 2;
  • the audio output control method also includes:
  • a sound image control interval is reconstructed for every two adjacent sound emitting devices
  • the process of determining the sound emitting device to be operated among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device includes:
  • the n sound emitting devices are located on the same horizontal plane, and the line connecting the n sound emitting devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
  • the process of determining the target sound image control interval where the incident direction is located in all the sound image control intervals includes:
  • the sound image control interval where the projection is located is determined as a target sound image control interval.
  • the process of selecting a sounding device to be worked among the sounding devices in the reconstruction of the target sound image control interval includes:
  • the sound emitting device corresponding to the line where the projection is located is used as the sound emitting device to be operated;
  • the two sound emitting devices reconstructing the target sound image control interval are both used as the waiting sound emitting devices.
  • the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
  • g1 is the amplitude modulation parameter of the first sounding device
  • g2 is the amplitude modulation parameter of the second sounding device.
  • an audio output control system including:
  • An acquisition module configured to acquire an initial audio signal of a virtual sound source in the current scene where the head-mounted device is located;
  • a first calculation module configured to calculate a phase modulation parameter based on the current head profile parameters of the wearer of the head-mounted device
  • a second calculation module configured to calculate an amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device
  • the output control module is used to control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
  • the present application also provides a head-mounted device, including:
  • a processor configured to implement the steps of the audio output control method described in any one of the above when executing the computer program.
  • the present application also provides a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, it realizes any of the above-mentioned The steps of the audio output control method.
  • This application provides an audio output control method, which calculates the amplitude modulation parameters of the sound emitting device according to the position of the virtual sound source relative to the head-mounted device, and at the same time obtains the current head contour parameters of the wearer of the head-mounted device, and calculates the sound based on the head contour parameters The phase modulation parameters of the device, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it.
  • This application has considered the size of the sound amplitude output by the sounding device.
  • the audio signal output by the device can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the head-mounted device.
  • the present application also provides an audio output control system, a head-mounted device, and a computer-readable storage medium, which have the same beneficial effect as the above-mentioned audio output control method.
  • Fig. 1 is a flow chart of the steps of an audio output control method provided by the present application
  • FIG. 2 is a schematic diagram of a phase modulation effect provided by the present application.
  • FIG. 3 is a schematic structural diagram of a head-mounted device provided by the present application.
  • Fig. 4 is a schematic layout diagram of a sounding device provided by the present application.
  • FIG. 5 is a schematic diagram of an audio-image control interval provided by the present application.
  • FIG. 6 is a schematic diagram of a tangent law amplitude translation provided by the present application.
  • Fig. 7 is a comparison diagram of the effect of amplitude translation binaural sound intensity difference provided by the present application.
  • FIG. 8 is a schematic structural diagram of an audio output control system provided by the present application.
  • the core of this application is to provide an audio output control method and system, a head-mounted device and a computer-readable storage medium, which uses amplitude translation to reconstruct the sense of space on the horizontal plane, and uses phase modulation to enhance the sense of space in the vertical direction. It enables the wearer to better perceive the real orientation of the sound source, avoids the localization effect of the sound source, and improves the sound field space and user experience of the headset.
  • Figure 1 is a flow chart of the steps of an audio output control method provided by the present application, the audio output control method includes:
  • the head-mounted device can specifically be a head-mounted VR device.
  • the VR device can use a computer to simulate a three-dimensional virtual scene. Users can experience VR movies or VR games through the VR device. In different VR movies or VR games, VR will be required.
  • the device provides corresponding sound effects for the user, so that the user can be immersed in the virtual world.
  • the head-mounted device in this example includes a sound generating device, which is used to output the audio signal of the virtual sound source in the current three-dimensional virtual scene, so as to provide the user with the sound effect of a certain VR movie or a certain VR game in the current scene
  • the sound generating device may be a loudspeaker. After obtaining the initial audio signal of the virtual sound source, the sound generating device will modulate the initial audio signal according to the set modulation parameters before outputting, where the initial audio signal is the unprocessed audio signal of the virtual sound source.
  • the virtual sound source that needs to be sounded can be determined currently, and the number of virtual sound sources in a scene can be one or more indivual.
  • S102 Calculate the phase modulation parameter based on the current head profile parameter of the wearer of the head-mounted device
  • phase modulation parameter here specifically refers to the head-related transfer function of the wearer, and the phase modulation of the audio signal is supplemented by the head-related transfer function to further enhance the sense of use of spatial audio.
  • the effect of phase modulation is shown in Figure 2. It can be seen that the phase spectrum of the audio signal after phase modulation is closer to the phase spectrum of the original sound source than the phase spectrum of the audio signal originally emitted by the sounding device.
  • the head-mounted device is also equipped with an infrared scanner for collecting the wearer's current head contour parameters.
  • Two infrared scanners can be set up, arranged on both sides of the head respectively.
  • the structure diagram of the head-mounted device can be referred to in Figure 3 Figure 3 only shows the layout of the infrared scanner 1 and the sounding device 2 on one side of the head.
  • the current head contour parameters include head contour parameters and ear contour parameters. It can be understood that, considering that the wearer of the head-mounted device is not fixed, the current head profile parameters of the current wearer can be acquired each time the head-mounted device is used, so as to obtain the head shape corresponding to the current wearer. related transfer functions.
  • H L and H R represent the head-related transfer function corresponding to the left ear of the current wearer and the head-related transfer function corresponding to the right ear respectively
  • p L and p R represent the audio signal received by the left ear and the audio signal received by the right ear respectively
  • p 0 represents the audio signal at the center of the head when the head is not present.
  • HL and HR are functions of the horizontal azimuth angle ⁇ , the elevation angle ⁇ of the virtual sound source, the distance r from the virtual sound source to the center of the head, and the sound wave parameter f.
  • S103 Calculate an amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device
  • S104 Control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter, and output the modulated audio signal.
  • the amplitude modulation parameter of the sound generating device After obtaining the position of the virtual sound source relative to the head-mounted device, calculate the amplitude modulation parameter of the sound generating device, control the sound generating device to modulate the amplitude of the initial audio signal according to the amplitude modulation parameter calculated in S103, and control the sound generating device
  • the phase of the initial audio signal is modulated according to the phase modulation parameter calculated in S102, and then the modulated audio signal is output, so that the user can better perceive the orientation of the virtual sound source and improve the acoustic immersion of the sound emitting device.
  • the amplitude of the original audio signal may be modulated first, and then the phase of the amplitude-modulated audio signal may be modulated. It can be understood that when phase modulation is performed, the modulated audio signal can be obtained by inputting the corresponding audio signal into the head-related transfer function.
  • the execution sequence of S101-S103 is not limited.
  • the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it.
  • This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source
  • the sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application
  • the signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
  • the head-mounted device includes n sound emitting devices, where n is an integer greater than 2;
  • the audio output control method also includes:
  • a sound emitting device in order to construct the spatial sense of the sound field, a sound emitting device is generally installed on the head-mounted device near the two ears, and after the audio signal corresponding to the virtual sound source is obtained, the two sound generating devices Set the modulation parameters so that the two sound-generating devices respectively modulate and output the received audio signals to achieve the purpose of reconstructing the sound image.
  • this scheme of setting up two sound emitting devices will make the user's perception of the direction and position of the sound source relatively simple when using the head-mounted device, and it is easy to produce a localization effect in the sound source, making the sound field spatial sense felt by the user relatively small. Difference.
  • the head-mounted device provided in this embodiment is provided with at least three sound generating devices for reconstructing the sound image.
  • this embodiment first selects the sound emitting device in the head-mounted device according to the position of the virtual sound source relative to the head-mounted device, and the sound-emitting device to be operated is modulated according to the calculated phase modulation parameters and amplitude
  • the parameters modulate the stereo sound effect on the initial audio signal of the virtual sound source, and output the modulated audio signal.
  • the sound image reconstructed by the working sound generating device can improve the user's perception of the direction of the virtual sound source.
  • there may be one or two sound emitting devices to be operated which can be selected according to the position of the virtual sound source relative to the head-mounted device, which is not limited in this application.
  • every two adjacent sounding devices reconstruct a sound image control interval
  • the process of determining the sound emitting device to work among all the sound emitting devices based on the position of the virtual sound source relative to the headset includes:
  • the incident direction of the virtual sound source is obtained based on the position of the virtual sound source relative to the head-mounted device;
  • the n sounding devices are located on the same horizontal plane, and the line connecting the n sounding devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
  • the process of determining the target panning interval in which the incident direction is located among all panning intervals includes:
  • the sound image control interval where the projection is located is determined as the target sound image control interval.
  • this embodiment will be described by taking a head-mounted device provided with six sound emitting devices as an example.
  • the sounding points of the six sounding devices are distributed symmetrically on both sides of the human head.
  • two adjacent sounding devices are selected to obtain the control interval for reconstructing the sound image, that is, the sound image like control intervals.
  • the 6 sounding devices can be located on the same plane or on different planes.
  • the 6 sounding devices can be set on the same plane. level. A coordinate system with the midpoint of the human head as the origin is established on this horizontal plane.
  • the lines connecting the six sound emitting devices with the origin of the coordinate system can divide the horizontal plane into six sound-image control intervals, as shown in Figure 4 , the 6 sound generating devices are respectively marked as a, b, c, d, e, f, and the 6 sound-image control intervals are respectively marked as I, II, III, IV, V, VI.
  • the incident direction of the virtual sound source can be determined. Assuming that there is a virtual sound source in the current scene of a VR movie or VR game, the projection position of the incident direction of the virtual sound source on the horizontal plane is shown in Figure 5.
  • the projection is located in the sound-image control interval VI, and the sound-image The control interval VI is the target audio image control interval, and the sound emitting device to be worked is determined from the two sound emitting devices responsible for the sound image control interval VI, that is, the sound emitting device to be worked is selected from the sound emitting device a and the sound emitting device d, so as to improve the The accuracy of sound image construction improves the sense of space in the sound field of the head-mounted device.
  • the process of selecting a sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval includes:
  • the sound emitting device corresponding to the line where the projection is located is taken as the sound emitting device to be worked;
  • the two sound emitting devices in the reconstructed target sound image control interval are both regarded as standby sound emitting devices.
  • the orientation of the virtual sound source when selecting a sounding device to be operated, it is necessary to determine the orientation of the virtual sound source first, and determine whether the virtual sound source is between two sounding devices or overlaps with any sounding device.
  • the sounding device that overlaps the virtual sound source can produce sound. Specifically, it can be judged whether the projection is on any connection that divides the audio-image control interval, and if the projection is on any connection that divides the audio-image control interval, then use the sounding device corresponding to the connection where it is located as the waiting to work Sounding device, if not, it means that the orientation of the virtual sound source is between the two sounding devices constituting the target sound image control section, and these two sounding devices are both regarded as the waiting sounding devices at this time.
  • the angle between each sound emitting device and the target direction can be determined in advance.
  • the target direction here can be the positive direction of the x-axis or the negative direction of the x-axis or the positive direction of the y-axis or the negative direction of the y-axis of the coordinate axis.
  • the sounding device a After determining the incident direction of the virtual sound source, determine the angle ⁇ r between the projection of the incident direction of the virtual sound source on the horizontal plane in the clockwise direction and the positive direction of the y-axis, and determine whether there is the same angle as the ⁇ r in ⁇ a ⁇ f , assuming that ⁇ r and ⁇ a are the same, then the sounding device a is the waiting sounding device, if ⁇ d ⁇ r ⁇ a, then the sounding device d and the sounding device a are the waiting sounding devices.
  • the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
  • g1 is the amplitude modulation parameter of the first sounding device
  • g2 is the amplitude modulation parameter of the second sounding device.
  • the amplitude modulation parameters of the two to-be-operated sound-emitting devices can be allocated by means of amplitude translation.
  • the present application uses the first relational expression to calculate the amplitude modulation parameters of the two to-be-operated sound emitting devices.
  • FIG. 6 establish a coordinate system with the human head as the origin, and the two sounding devices to be worked are located on the circle centered at the listening point (the midpoint of the human head), and the angles of the two sounding devices to be worked are respectively and The angle between the virtual sound source and the bisector of the two sounding devices is
  • a sound-generating device closer to the virtual sound source is used as the first sound-generating device, and the other sound-generating device is used as the second sound-generating device.
  • the first sounding device and the second sounding device are respectively based on g1, g2 and the head-related transfer function.
  • the initial audio signal of the virtual sound source is subjected to stereo sound modulation, and the modulated audio signals are respectively output.
  • the real sound source sends out the binaural amplitude difference received by the human ear from the direction of the arrow, and the binaural amplitude difference received by the human ear when the sound played by traditional VR is played by the ear, and the binaural amplitude difference received by the output sound according to the multi-speaker distribution amplitude.
  • the comparison results of the ear amplitude difference are shown in Figure 7. It is not difficult to see that the binaural amplitude difference of the sound allocated by the amplitude translation is closer to the binaural amplitude difference of the original sound. In the case of VR playback, it is closer to the real position.
  • this application increases the number of sounding devices on the VR device, and distributes the amplitude of the sound of different sounding devices with the amplitude translation of the tangent law, and uses infrared scanning to calculate the head transfer function to supplement the phase modulation of the speaker to achieve Control the human ear's perception of the position of the sound source, and improve the sense of space in the sound field of VR products.
  • FIG. 8 is a schematic structural diagram of an audio output control system provided by the present application.
  • the audio output control system includes:
  • An acquisition module 11 configured to acquire an initial audio signal of a virtual sound source in the current scene where the head-mounted device is located;
  • the first calculation module 12 is used to calculate the phase modulation parameter based on the current head profile parameters of the wearer of the head-mounted device;
  • the second calculation module 13 is used to calculate the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device;
  • the output control module 14 is used to control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
  • the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it.
  • This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source
  • the sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application
  • the signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
  • the phase modulation parameter includes a head-related transfer function.
  • the head-mounted device includes n sound emitting devices, where n is an integer greater than 12;
  • the audio output control system also includes:
  • a determining module configured to determine a sound emitting device to be operated among all sound emitting devices based on the position of the virtual sound source relative to the head-mounted device;
  • every two adjacent sounding devices reconstruct a sound image control interval
  • the process of determining the sound emitting device to work among all the sound emitting devices based on the position of the virtual sound source relative to the headset includes:
  • the incident direction of the virtual sound source is obtained based on the position of the virtual sound source relative to the head-mounted device;
  • the n sounding devices are located on the same horizontal plane, and the line connecting the n sounding devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
  • the process of determining the target panning interval in which the incident direction is located among all panning intervals includes:
  • the sound image control interval where the projection is located is determined as the target sound image control interval.
  • the process of selecting a sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval includes:
  • the sound emitting device corresponding to the line where the projection is located is taken as the sound emitting device to be worked;
  • the two sound emitting devices in the reconstructed target sound image control interval are both regarded as standby sound emitting devices.
  • the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
  • g1 is the amplitude modulation parameter of the first sounding device
  • g2 is the amplitude modulation parameter of the second sounding device.
  • a head-mounted device including:
  • the processor is configured to implement the steps of the audio output control method described in any one of the above embodiments when executing the computer program.
  • the memory includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system and computer-readable instructions
  • the internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the processor provides calculation and control capabilities for the head-mounted device.
  • the following steps can be implemented: obtaining the initial audio signal of the virtual sound source in the current scene where the head-mounted device is located; Calculate the phase modulation parameter based on the current head contour parameters of the user; calculate the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device; control the sounding device of the head-mounted device to perform stereo sound effects on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter Modulate and output the modulated audio signal.
  • the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it.
  • This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source
  • the sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application
  • the signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
  • the processor executes the computer subroutine stored in the memory, the following steps can be implemented: calculate the head-related transfer function based on the current head contour parameters of the wearer of the head-mounted device, and use the head-related transfer function as Phase modulation parameters.
  • the processor executes the computer subroutine stored in the memory
  • the following steps can be implemented: determine the sound emitting device to be worked among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device; control the head
  • the sounding device of the wearable device performs stereo modulation on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and outputs the modulated audio signal.
  • the parameter performs stereo modulation on the original audio signal and outputs the modulated audio signal.
  • the processor executes the computer subroutine stored in the memory
  • the following steps can be implemented: obtain the incident direction of the virtual sound source based on the position of the virtual sound source relative to the head-mounted device; In the interval, determine the target sound image control interval where the incident direction is located, and reconstruct a sound image control interval for every two adjacent sound emitting devices; select the sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval.
  • the processor executes the computer subroutine stored in the memory, the following steps can be implemented: obtain the projection of the incident direction on the horizontal plane; determine the sound image control interval where the projection is located as the target sound image control interval .
  • the processor executes the computer subroutine stored in the memory, the following steps can be implemented: determine whether the projection is on any connection line dividing the sound image control interval; if so, reconstruct the target sound image Among the two sounding devices in the control section, the sounding device corresponding to the line where the projection is located is taken as the waiting sounding device; if not, the two sounding devices in the control section of the reconstructed target sound image are both taken as the waiting sounding device.
  • the processor executes the computer subroutine stored in the memory, the following steps may be implemented: determining and reconstructing the two sounding devices of the target sound image control interval based on the midpoint position of the human head The bisector of the included angle;
  • the first relational expression is in, is the angle between any one of the sounding device and the midpoint of the human head and the bisector, is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
  • Figure X is a structural diagram of another head-mounted device provided by the embodiment of the present application.
  • the head-mounted device also includes:
  • the input interface is connected with the processor, and is used to obtain the computer program, parameters and instructions imported from the outside, and store them in the memory under the control of the processor.
  • the input interface can be connected with an input device to receive parameters or instructions manually input by the user.
  • the input device may be a touch layer covered on the display screen, or a button, a trackball or a touch pad provided on the terminal shell.
  • the display unit is connected with the processor and used for displaying the data sent by the processor.
  • the display unit may be a liquid crystal display or an electronic ink display.
  • the network port is connected with the processor and is used for communication connection with various external terminal devices.
  • the communication technology used in the communication connection can be wired communication technology or wireless communication technology, such as mobile high-definition link technology (MHL), universal serial bus (USB), high-definition multimedia interface (HDMI), wireless fidelity technology (WiFi), Bluetooth communication technology, low-power Bluetooth communication technology, communication technology based on IEEE802.11s, etc.
  • the present application also provides a computer-readable storage medium, on which a computer program is stored.
  • a computer program is stored.
  • the audio output control method as described in any one of the above embodiments is implemented. step.
  • the computer storage medium may include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk, etc., which can store program codes. medium.
  • a computer program is stored on the storage medium, and when the computer program is executed by the processor, the following steps are implemented: obtaining the initial audio signal of the virtual sound source in the current scene where the head-mounted device is located; based on the current human head contour parameters of the wearer of the head-mounted device Calculate the phase modulation parameters; calculate the amplitude modulation parameters based on the position of the virtual sound source relative to the head-mounted device; control the sounding device of the head-mounted device to modulate the stereo sound effect of the initial audio signal according to the amplitude modulation parameters and phase modulation parameters and output the modulated audio signal.
  • the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it.
  • This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source
  • the sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application
  • the signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
  • the following steps are implemented: calculating the head-related transfer function based on the current head profile parameters of the wearer of the head-mounted device, and using the head-related transfer function as a phase modulation parameter.
  • the following steps are implemented: based on the position of the virtual sound source relative to the head-mounted device, among all the sound-emitting devices, the sound-emitting devices to be worked are determined; the sound-emitting device of the head-mounted device is controlled to press The amplitude modulation parameter and phase modulation parameter modulate the initial audio signal for stereo sound effects and output the modulated audio signal. Modulate the stereo effect and output the modulated audio signal.
  • the following steps are implemented: obtaining the incident direction of the virtual sound source based on the position of the virtual sound source relative to the head-mounted device; determining where the incident direction is located in all sound image control intervals The target sound image control interval of each two adjacent sound emitting devices reconstructs a sound image control interval; select the sound emitting device to be operated among the sound emitting devices of the reconstructed target sound image control interval.
  • the following steps are implemented: obtaining the projection of the incident direction on the horizontal plane; determining the sound image control interval where the projection is located as the target sound image control interval.
  • the following steps are implemented: judging whether the projection is on any connection line dividing the sound image control interval; if so, reconstructing the two vocalizations of the target sound image control interval
  • the sounding device corresponding to the line where the projection is located is regarded as the sounding device to be operated; if not, the two sounding devices in the control area of the reconstructed target sound image are both regarded as the sounding device to be operated.
  • the following steps are implemented: determining the bisector of the angle between the two to-be-operated sounding devices based on the midpoint position of the human head; determining the reconstruction based on the midpoint position of the human head
  • the bisector of the angle between the two sounding devices in the target sound image control interval; the respective amplitude modulation parameters of the two sounding devices are calculated according to the first relational expression, and the first relational expression is in, is the angle between any one of the sounding device and the midpoint of the human head and the bisector, is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

Disclosed in the present application are an audio output control method and system, and a head-mounted device and a computer-readable storage medium. The audio output control method comprises: acquiring an initial audio signal of a virtual sound source in the current environment where a head-mounted device is located; calculating a phase modulation parameter on the basis of the current head contour parameter of a wearer of the head-mounted device; calculating an amplitude modulation parameter on the basis of the position of the virtual sound source relative to the head-mounted device; and controlling a sound production apparatus of the head-mounted device to perform stereoscopic sound effect modulation on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter, and outputting the modulated audio signal. In the present application, the sense of space on a horizontal plane is reconstructed by using amplitude translation, and the sense of space in a vertical direction is enhanced by using phase modulation, so that a wearer can sense the real orientation of a sound source better, and production of a positioning effect in the sound source is avoided, thereby improving the sound field sense of space and the user experience of a head-mounted device.

Description

一种音频输出控制方法、系统及相关组件An audio output control method, system and related components
本申请要求于2021年10月29日提交中国专利局、申请号为202111276794.0、发明名称为“一种音频输出控制方法、系统及相关组件”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202111276794.0 and the title of the invention "An Audio Output Control Method, System and Related Components" filed with the China Patent Office on October 29, 2021, the entire contents of which are incorporated by reference in this application.
技术领域technical field
本申请涉及头戴设备领域,特别涉及一种音频输出控制方法、系统及相关组件。The present application relates to the field of head-mounted devices, in particular to an audio output control method, system and related components.
背景技术Background technique
VR(Virtual Reality,虚拟现实)产品是利用计算机技术模拟三维虚拟世界,为使用者营造身临其境的体验感。VR产品有三个核心特征:沉浸感、交互性及想象力,沉浸感是VR产品最基本的特征,而音效是令VR产品具有完美沉浸感重要指标之一。目前市场上的VR产品多为双喇叭,两个喇叭在左、右耳边发声,对两个喇叭进行幅度平移分配,使两个喇叭按照分配的幅度发声,从而达到声像重构的目的,但是这种简单的幅度平移方案使得用户在使用时对声源方位的感知比较单一,容易产生声源头中定位效应,感受到的声场空间感较差,产品的声学沉浸感效果较差。VR (Virtual Reality, virtual reality) products use computer technology to simulate a three-dimensional virtual world to create an immersive experience for users. VR products have three core features: immersion, interactivity, and imagination. Immersion is the most basic feature of VR products, and sound effects are one of the important indicators for perfect immersion in VR products. At present, most of the VR products on the market are dual-speaker. The two speakers sound at the left and right ears. The amplitude translation of the two speakers is distributed, so that the two speakers can sound according to the allocated amplitude, so as to achieve the purpose of sound image reconstruction. However, this simple amplitude translation scheme makes the user's perception of the direction of the sound source relatively simple when using it, and it is easy to produce a localization effect in the sound source. The sense of space in the sound field is poor, and the acoustic immersion effect of the product is poor.
因此,如何提供一种解决上述技术问题的方案是本领域技术人员目前需要解决的问题。Therefore, how to provide a solution to the above technical problems is a problem that those skilled in the art need to solve at present.
发明内容Contents of the invention
本申请的目的是提供一种音频输出控制方法、系统,头戴设备及计算机可读存储介质,采用幅度平移重构了水平面上的空间感,采用相位调制加强了竖直方向上的空间感,使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。The purpose of this application is to provide an audio output control method and system, a head-mounted device and a computer-readable storage medium, which uses amplitude translation to reconstruct the sense of space on the horizontal plane, and uses phase modulation to enhance the sense of space in the vertical direction. It enables the wearer to better perceive the real orientation of the sound source, avoids the localization effect of the sound source, and improves the sound field space and user experience of the headset.
为解决上述技术问题,本申请提供了一种音频输出控制方法,包括:In order to solve the above technical problems, the application provides an audio output control method, including:
获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;Obtain the initial audio signal of the virtual sound source in the current scene where the headset is located;
基于所述头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;calculating phase modulation parameters based on current head profile parameters of the wearer of the head-mounted device;
基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数;calculating an amplitude modulation parameter based on a position of the virtual sound source relative to the head-mounted device;
控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
可选的,所述相位调制参数包括头相关传输函数。Optionally, the phase modulation parameter includes a head-related transfer function.
可选的,所述头戴设备包括n个所述发声装置,n为大于2的整数;Optionally, the head-mounted device includes n sound emitting devices, where n is an integer greater than 2;
该音频输出控制方法还包括:The audio output control method also includes:
基于所述虚拟声源相对于所述头戴设备的位置在所有所述发声装置中确定待工作发声装置;determining a to-be-operated sound emitting device among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device;
所述控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:The process of controlling the sounding device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal includes:
控制头戴设备中的所述待工作发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the standby sound generating device in the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
可选的,每两个相邻的所述发声装置重构一个声像控制区间;Optionally, a sound image control interval is reconstructed for every two adjacent sound emitting devices;
所述基于所述虚拟声源相对于所述头戴设备的位置在所有所述发声装置中确定待工作发声装置的过程包括:The process of determining the sound emitting device to be operated among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device includes:
基于所述虚拟声源相对于所述头戴设备的位置得到所述虚拟声源的入射方向;obtaining an incident direction of the virtual sound source based on a position of the virtual sound source relative to the head-mounted device;
在所有所述声像控制区间中确定所述入射方向所在的目标声像控制区间;determining the target sound image control interval where the incident direction is located in all the sound image control intervals;
在重构所述目标声像控制区间的所述发声装置中选择待工作发声装置。Selecting a to-be-worked sound emitting device among the sound emitting devices reconstructing the target sound image control interval.
可选的,n个所述发声装置位于同一水平面,n个所述发声装置与所述水平面上的人头中点位置的连线将所述水平面划分为n个所述声像控制区间;Optionally, the n sound emitting devices are located on the same horizontal plane, and the line connecting the n sound emitting devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
所述在所有所述声像控制区间中确定所述入射方向所在的目标声像控制区间的过程包括:The process of determining the target sound image control interval where the incident direction is located in all the sound image control intervals includes:
获取所述入射方向在所述水平面上的投影;Obtain the projection of the incident direction on the horizontal plane;
将所述投影所在的所述声像控制区间确定为目标声像控制区间。The sound image control interval where the projection is located is determined as a target sound image control interval.
可选的,所述在重构所述目标声像控制区间的所述发声装置中选择待工 作发声装置的过程包括:Optionally, the process of selecting a sounding device to be worked among the sounding devices in the reconstruction of the target sound image control interval includes:
判断所述投影是否处于划分所述声像控制区间的任一所述连线上;judging whether the projection is on any of the lines dividing the audio-image control interval;
若是,将重构所述目标声像控制区间的两个所述发声装置中,所述投影所处的所述连线对应的所述发声装置作为待工作发声装置;If yes, among the two sound emitting devices reconstructing the target sound image control interval, the sound emitting device corresponding to the line where the projection is located is used as the sound emitting device to be operated;
若否,将重构所述目标声像控制区间的两个所述发声装置均作为所述待工作发声装置。If not, the two sound emitting devices reconstructing the target sound image control interval are both used as the waiting sound emitting devices.
可选的,所述基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数的过程包括:Optionally, the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;determining and reconstructing the bisector of the angle between the two sounding devices in the target sound image control interval based on the position of the midpoint of the human head;
根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
Figure PCTCN2021131709-appb-000001
Calculate the respective amplitude modulation parameters of the two sounding devices according to the first relational expression, the first relational expression is
Figure PCTCN2021131709-appb-000001
其中,
Figure PCTCN2021131709-appb-000002
为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
Figure PCTCN2021131709-appb-000003
为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
in,
Figure PCTCN2021131709-appb-000002
is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
Figure PCTCN2021131709-appb-000003
is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
为解决上述技术问题,本申请还提供了一种音频输出控制系统,包括:In order to solve the above technical problems, the present application also provides an audio output control system, including:
获取模块,用于获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;An acquisition module, configured to acquire an initial audio signal of a virtual sound source in the current scene where the head-mounted device is located;
第一计算模块,用于基于所述头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;A first calculation module, configured to calculate a phase modulation parameter based on the current head profile parameters of the wearer of the head-mounted device;
第二计算模块,用于基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数;A second calculation module, configured to calculate an amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device;
输出控制模块,用于控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。The output control module is used to control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
为解决上述技术问题,本申请还提供了一种头戴设备,包括:In order to solve the above technical problems, the present application also provides a head-mounted device, including:
存储器,用于存储计算机程序;memory for storing computer programs;
处理器,用于执行所述计算机程序时实现如上文任意一项所述的音频输出控制方法的步骤。A processor, configured to implement the steps of the audio output control method described in any one of the above when executing the computer program.
为解决上述技术问题,本申请还提供了一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如上文任意一项所述的音频输出控制方法的步骤。本申请提供了一种音频输出控制方法,根据虚拟声源相对于头戴设备的位置计算发声装置的幅度调制参数,同时获取头戴设备的佩戴者的当前人头轮廓参数,基于人头轮廓参数计算发声装置的相位调制参数,然后控制发声装置按照幅度调制参数和相位调制参数对虚拟声源的初始音频信号进行立体音效的调制然后再输出,本申请既考虑了发声装置输出的声音幅度的大小,重构虚拟声源在水平面上的空间感,还根据当前人头轮廓参数补充校正幅度平移的虚拟声源的相位,以此重构虚拟声源在竖直方向上的空间感,因此,本申请的发声装置输出的音频信号能够使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。本申请还提供了一种音频输出控制系统,头戴设备及计算机可读存储介质,具有和上述音频输出控制方法相同的有益效果。In order to solve the above-mentioned technical problems, the present application also provides a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, it realizes any of the above-mentioned The steps of the audio output control method. This application provides an audio output control method, which calculates the amplitude modulation parameters of the sound emitting device according to the position of the virtual sound source relative to the head-mounted device, and at the same time obtains the current head contour parameters of the wearer of the head-mounted device, and calculates the sound based on the head contour parameters The phase modulation parameters of the device, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it. This application has considered the size of the sound amplitude output by the sounding device. Construct the sense of space of the virtual sound source on the horizontal plane, and also supplement and correct the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. The audio signal output by the device can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the head-mounted device. The present application also provides an audio output control system, a head-mounted device, and a computer-readable storage medium, which have the same beneficial effect as the above-mentioned audio output control method.
附图说明Description of drawings
为了更清楚地说明本申请实施例,下面将对实施例中所需要使用的附图做简单的介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the embodiments of the present application more clearly, the following will briefly introduce the accompanying drawings used in the embodiments. Obviously, the accompanying drawings in the following description are only some embodiments of the present application. As far as people are concerned, other drawings can also be obtained based on these drawings on the premise of not paying creative work.
图1为本申请所提供的一种音频输出控制方法的步骤流程图;Fig. 1 is a flow chart of the steps of an audio output control method provided by the present application;
图2为本申请所提供的一种相位调制效果示意图;FIG. 2 is a schematic diagram of a phase modulation effect provided by the present application;
图3为本申请所提供的一种头戴设备的结构示意图;FIG. 3 is a schematic structural diagram of a head-mounted device provided by the present application;
图4为本申请所提供的一种发声装置的布置示意图;Fig. 4 is a schematic layout diagram of a sounding device provided by the present application;
图5为本申请所提供的一种声像控制区间的示意图;FIG. 5 is a schematic diagram of an audio-image control interval provided by the present application;
图6为本申请所提供的一种正切法则幅度平移原理图;FIG. 6 is a schematic diagram of a tangent law amplitude translation provided by the present application;
图7为本申请所提供的一种幅度平移双耳声强差效果对比图;Fig. 7 is a comparison diagram of the effect of amplitude translation binaural sound intensity difference provided by the present application;
图8为本申请所提供的一种音频输出控制系统的结构示意图。FIG. 8 is a schematic structural diagram of an audio output control system provided by the present application.
具体实施方式Detailed ways
本申请的核心是提供一种音频输出控制方法、系统,头戴设备及计算机可读存储介质,采用幅度平移重构了水平面上的空间感,采用相位调制加强了竖直方向上的空间感,使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。The core of this application is to provide an audio output control method and system, a head-mounted device and a computer-readable storage medium, which uses amplitude translation to reconstruct the sense of space on the horizontal plane, and uses phase modulation to enhance the sense of space in the vertical direction. It enables the wearer to better perceive the real orientation of the sound source, avoids the localization effect of the sound source, and improves the sound field space and user experience of the headset.
为使本申请实施例的目的、技术方案和优点更加清楚,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the purposes, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the drawings in the embodiments of the present application. Obviously, the described embodiments It is a part of the embodiments of this application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.
请参照图1,图1为本申请所提供的一种音频输出控制方法的步骤流程图,该音频输出控制方法包括:Please refer to Figure 1, Figure 1 is a flow chart of the steps of an audio output control method provided by the present application, the audio output control method includes:
S101:获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;S101: Obtain an initial audio signal of a virtual sound source in the current scene where the headset is located;
其中,头戴设备具体可以为头戴式VR设备,VR设备可以利用计算机模拟三维虚拟场景,用户可以通过该VR设备体验VR影片或VR游戏,在不同的VR影片或VR游戏中,都会需要VR设备为用户提供对应的音效,以使用户能够沉浸在虚拟世界中。Among them, the head-mounted device can specifically be a head-mounted VR device. The VR device can use a computer to simulate a three-dimensional virtual scene. Users can experience VR movies or VR games through the VR device. In different VR movies or VR games, VR will be required. The device provides corresponding sound effects for the user, so that the user can be immersed in the virtual world.
具体的,本实例中的头戴设备中包括发声装置,用于输出当前三维虚拟场景中的虚拟声源的音频信号,以便为用户提供某一VR电影或某一VR游戏在当前场景的音效,发声装置具体可以为扬声器。发声装置在获取到虚拟声源的初始音频信号后,会根据设置的调制参数对该初始音频信号进行调制后再输出,这里的初始音频信号即未经处理的虚拟声源的音频信号。Specifically, the head-mounted device in this example includes a sound generating device, which is used to output the audio signal of the virtual sound source in the current three-dimensional virtual scene, so as to provide the user with the sound effect of a certain VR movie or a certain VR game in the current scene, Specifically, the sound generating device may be a loudspeaker. After obtaining the initial audio signal of the virtual sound source, the sound generating device will modulate the initial audio signal according to the set modulation parameters before outputting, where the initial audio signal is the unprocessed audio signal of the virtual sound source.
可以理解的是,根据佩戴者在VR电影或VR游戏的当前场景中所处的位置,可以确定当前需要发声的虚拟声源,在一个场景中虚拟声源的个数可以为一个也可以为多个。It can be understood that according to the position of the wearer in the current scene of the VR movie or VR game, the virtual sound source that needs to be sounded can be determined currently, and the number of virtual sound sources in a scene can be one or more indivual.
S102:基于头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;S102: Calculate the phase modulation parameter based on the current head profile parameter of the wearer of the head-mounted device;
考虑到目前为了使头戴设备带给用户的音效能够具有空间感,为用户提供相对较高的听觉体验,通常是对头戴设备中的两个发声装置采用幅度平移的方法进行声像重构,但是仅通过幅度平移不能准确重构声源在竖直方向上 的空间感,因此本步骤还采集了头戴设备的佩戴者的当前人头轮廓参数,基于当前人头轮廓参数可以计算相位调制参数,这里的相位调制参数具体指佩戴者的头相关传输函数,通过头相关传输函数补充音频信号的相位调制,进一步增强空间音频的使用感。相位调制的效果如图2所示,可以看出经过相位调制的音频信号相位谱与发声装置原本发出的音频信号相位谱相比更接近原始声源的相位谱。Considering that in order to make the sound effect brought by the head-mounted device to the user to have a sense of space and provide the user with a relatively high auditory experience, it is usually used to reconstruct the sound image of the two sounding devices in the head-mounted device by means of amplitude translation , but the spatial sense of the sound source in the vertical direction cannot be accurately reconstructed only by amplitude translation, so this step also collects the current head contour parameters of the wearer of the head-mounted device, and the phase modulation parameters can be calculated based on the current head contour parameters, The phase modulation parameter here specifically refers to the head-related transfer function of the wearer, and the phase modulation of the audio signal is supplemented by the head-related transfer function to further enhance the sense of use of spatial audio. The effect of phase modulation is shown in Figure 2. It can be seen that the phase spectrum of the audio signal after phase modulation is closer to the phase spectrum of the original sound source than the phase spectrum of the audio signal originally emitted by the sounding device.
具体的,头戴设备上还设有用于采集佩戴者的当前人头轮廓参数的红外线扫描仪,红外线扫描仪可以设置两个,在人头两侧分别布置,头戴设备的结构示意图可以参照图3所示,图3仅示出人头一侧的红外线扫描仪①和发声装置②的布放图。当前人头轮廓参数包括头部轮廓参数和耳部轮廓参数。可以理解的是,考虑到头戴设备的佩戴者并不是固定的,因此,可在头戴设备每次使用时,对当前佩戴者的当前人头轮廓参数进行获取,从而得到当前佩戴者对应的头相关传输函数。Specifically, the head-mounted device is also equipped with an infrared scanner for collecting the wearer's current head contour parameters. Two infrared scanners can be set up, arranged on both sides of the head respectively. The structure diagram of the head-mounted device can be referred to in Figure 3 Figure 3 only shows the layout of the infrared scanner ① and the sounding device ② on one side of the head. The current head contour parameters include head contour parameters and ear contour parameters. It can be understood that, considering that the wearer of the head-mounted device is not fixed, the current head profile parameters of the current wearer can be acquired each time the head-mounted device is used, so as to obtain the head shape corresponding to the current wearer. related transfer functions.
其中,头相关传输函数的关系式为
Figure PCTCN2021131709-appb-000004
Among them, the relational expression of the head-related transfer function is
Figure PCTCN2021131709-appb-000004
H L和H R分别代表当前佩戴者的左耳对应的头相关传输函数和右耳对应的头相关传输函数,p L和p R分别代表左耳接收到的音频信号和右耳接收到音频信号,p 0表示人头不在时,人头中心位置的音频信号。一般情况下H L和H R是虚拟声源的水平方位角θ、仰角φ、虚拟声源到头中心的距离r以及声波参数f的函数。 H L and H R represent the head-related transfer function corresponding to the left ear of the current wearer and the head-related transfer function corresponding to the right ear respectively, p L and p R represent the audio signal received by the left ear and the audio signal received by the right ear respectively , p 0 represents the audio signal at the center of the head when the head is not present. In general, HL and HR are functions of the horizontal azimuth angle θ, the elevation angle φ of the virtual sound source, the distance r from the virtual sound source to the center of the head, and the sound wave parameter f.
S103:基于虚拟声源相对于头戴设备的位置计算幅度调制参数;S103: Calculate an amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device;
S104:控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。S104: Control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter, and output the modulated audio signal.
具体的,在获取到虚拟声源相对于头戴设备的位置后,计算发声装置的幅度调制参数,控制发声装置按照S103计算得到的幅度调制参数对初始音频信号的幅度进行调制,并控制发声装置按照S102计算得到的相位调制参数对初始音频信号的相位进行调制,然后输出调制后的音频信号,使用户可以更好的感知虚拟声源的方位,提高发声设备的声学沉浸感。作为一种可选的实 施例,可以先对初始音频信号的幅度进行调制,再对幅度调制后的音频信号的相位进行调制。可以理解的是,进行相位调制时,将对应的音频信号输入该头相关传输函数即可得到调制后的音频信号。本实施例在此对S101-S103的执行顺序不作限定。Specifically, after obtaining the position of the virtual sound source relative to the head-mounted device, calculate the amplitude modulation parameter of the sound generating device, control the sound generating device to modulate the amplitude of the initial audio signal according to the amplitude modulation parameter calculated in S103, and control the sound generating device The phase of the initial audio signal is modulated according to the phase modulation parameter calculated in S102, and then the modulated audio signal is output, so that the user can better perceive the orientation of the virtual sound source and improve the acoustic immersion of the sound emitting device. As an optional embodiment, the amplitude of the original audio signal may be modulated first, and then the phase of the amplitude-modulated audio signal may be modulated. It can be understood that when phase modulation is performed, the modulated audio signal can be obtained by inputting the corresponding audio signal into the head-related transfer function. In this embodiment, the execution sequence of S101-S103 is not limited.
可见,本实施例中,根据虚拟声源相对于头戴设备的位置计算发声装置的幅度调制参数,同时获取头戴设备的佩戴者的当前人头轮廓参数,基于人头轮廓参数计算发声装置的相位调制参数,然后控制发声装置按照幅度调制参数和相位调制参数对虚拟声源的初始音频信号进行立体音效的调制然后再输出,本申请既考虑了发声装置输出的声音幅度的大小,重构虚拟声源在水平面上的空间感,还根据当前人头轮廓参数补充校正幅度平移的虚拟声源的相位,以此重构虚拟声源在竖直方向上的空间感,因此,本申请的发声装置输出的音频信号能够使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。It can be seen that in this embodiment, the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it. This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source The sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application The signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
在上述实施例的基础上:On the basis of above-mentioned embodiment:
作为一种可选的实施例,头戴设备包括n个发声装置,n为大于2的整数;As an optional embodiment, the head-mounted device includes n sound emitting devices, where n is an integer greater than 2;
该音频输出控制方法还包括:The audio output control method also includes:
基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置;determining the sound emitting device to be operated among all sound emitting devices based on the position of the virtual sound source relative to the head-mounted device;
控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:The process of controlling the sounding device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal includes:
控制头戴设备中的待工作发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the ready-to-work sound emitting device in the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter, and output the modulated audio signal.
现有技术中,为了构造声场空间感,一般是在头戴设备上靠近两耳侧分别设置一个发声装置,在获取到虚拟声源对应的音频信号后,通过幅度平移的方式为两个发声装置设置调制参数,以便两个发声装置分别对各自接收到的音频信号进行调制并输出,以达到重构声像的目的。但是这种设置两个发声装置的方案,会使得用户在使用该头戴设备时,对于声音来源方向与位置的感知比较单一,容易产生声源头中定位效应,使得用户感受到的声场空间 感较差。基于此,本实施例所提供的头戴设备中设置有至少3个用于重构声像的发声装置。为了进一步增强声场空间感,本实施例首先根据虚拟声源相对于头戴设备的位置对头戴设备中的待工作发声装置进行选择,由待工作发声装置按计算到的相位调制参数和幅度调制参数对虚拟声源的初始音频信号进行立体音效的调制,并输出调制后的音频信号,待工作发声装置重构的声像可以提高用户对虚拟声源方位的感知能力。其中,待工作发声装置可以为一个也可以为两个,根据虚拟声源相对于头戴设备的位置进行选择即可,本申请在此不作限定。In the prior art, in order to construct the spatial sense of the sound field, a sound emitting device is generally installed on the head-mounted device near the two ears, and after the audio signal corresponding to the virtual sound source is obtained, the two sound generating devices Set the modulation parameters so that the two sound-generating devices respectively modulate and output the received audio signals to achieve the purpose of reconstructing the sound image. However, this scheme of setting up two sound emitting devices will make the user's perception of the direction and position of the sound source relatively simple when using the head-mounted device, and it is easy to produce a localization effect in the sound source, making the sound field spatial sense felt by the user relatively small. Difference. Based on this, the head-mounted device provided in this embodiment is provided with at least three sound generating devices for reconstructing the sound image. In order to further enhance the spatial sense of the sound field, this embodiment first selects the sound emitting device in the head-mounted device according to the position of the virtual sound source relative to the head-mounted device, and the sound-emitting device to be operated is modulated according to the calculated phase modulation parameters and amplitude The parameters modulate the stereo sound effect on the initial audio signal of the virtual sound source, and output the modulated audio signal. The sound image reconstructed by the working sound generating device can improve the user's perception of the direction of the virtual sound source. Wherein, there may be one or two sound emitting devices to be operated, which can be selected according to the position of the virtual sound source relative to the head-mounted device, which is not limited in this application.
作为一种可选的实施例,每两个相邻的发声装置重构一个声像控制区间;As an optional embodiment, every two adjacent sounding devices reconstruct a sound image control interval;
基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置的过程包括:The process of determining the sound emitting device to work among all the sound emitting devices based on the position of the virtual sound source relative to the headset includes:
基于虚拟声源相对于头戴设备的位置得到虚拟声源的入射方向;The incident direction of the virtual sound source is obtained based on the position of the virtual sound source relative to the head-mounted device;
在所有声像控制区间中确定入射方向所在的目标声像控制区间;Determining the target sound image control interval where the incident direction is located in all sound image control intervals;
在重构目标声像控制区间的发声装置中选择待工作发声装置。Select the to-be-worked sounding device among the sounding devices in the reconstruction target sound image control section.
作为一种可选的实施例,n个发声装置位于同一水平面,n个发声装置与水平面上的人头中点位置的连线将水平面划分为n个声像控制区间;As an optional embodiment, the n sounding devices are located on the same horizontal plane, and the line connecting the n sounding devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
在所有声像控制区间中确定入射方向所在的目标声像控制区间的过程包括:The process of determining the target panning interval in which the incident direction is located among all panning intervals includes:
获取入射方向在水平面上的投影;Obtain the projection of the incident direction on the horizontal plane;
将投影所在的声像控制区间确定为目标声像控制区间。The sound image control interval where the projection is located is determined as the target sound image control interval.
具体的,以设有6个发声装置的头戴设备为例,对本实施例进行说明。6个发声装置的发声点在人头两侧对称分布,为了进一步提高声像重构的准确性,本实施例中选择相邻的两个发声装置得到用于重构声像的控制区间,即声像控制区间。6个发声装置可以位于同一平面,也可以位于不同平面,为便于确定虚拟声源处于哪两个发声装置负责的控制区间,作为一种优选的实施例,可以将6个发声装置设置在同一个水平面上。在这个水平面上建立以人头中点位置为原点的坐标系,可以理解的是,6个发声装置与该坐标系原点的连线可以将水平面划分为6个声像控制区间,如图4所示,6个发声装置分别记为a、b、c、d、e、f,6个声像控制区间分别为标记为Ⅰ、Ⅱ、Ⅲ、Ⅳ、Ⅴ、Ⅵ。Specifically, this embodiment will be described by taking a head-mounted device provided with six sound emitting devices as an example. The sounding points of the six sounding devices are distributed symmetrically on both sides of the human head. In order to further improve the accuracy of the sound image reconstruction, in this embodiment, two adjacent sounding devices are selected to obtain the control interval for reconstructing the sound image, that is, the sound image like control intervals. The 6 sounding devices can be located on the same plane or on different planes. In order to facilitate the determination of which two sounding devices are responsible for the control interval of the virtual sound source, as a preferred embodiment, the 6 sounding devices can be set on the same plane. level. A coordinate system with the midpoint of the human head as the origin is established on this horizontal plane. It is understandable that the lines connecting the six sound emitting devices with the origin of the coordinate system can divide the horizontal plane into six sound-image control intervals, as shown in Figure 4 , the 6 sound generating devices are respectively marked as a, b, c, d, e, f, and the 6 sound-image control intervals are respectively marked as Ⅰ, Ⅱ, Ⅲ, Ⅳ, Ⅴ, Ⅵ.
可以理解的是,基于虚拟声源相对于头戴设备的位置,可以确定虚拟声源的入射方向。假设在VR影片或VR游戏的当前场景中,有个虚拟声源,该虚拟声源的入射方向在水平面上的投影的位置如图5所示,该投影位于声像控制区间Ⅵ内,声像控制区间Ⅵ即为目标声像控制区间,从负责该声像控制区间Ⅵ的两个发声装置中确定待工作发声装置,即从发声装置a及发声装置d中选择待工作发声装置,以提高重构声像的准确性,从而提高头戴设备的声场空间感。It can be understood that, based on the position of the virtual sound source relative to the head-mounted device, the incident direction of the virtual sound source can be determined. Assuming that there is a virtual sound source in the current scene of a VR movie or VR game, the projection position of the incident direction of the virtual sound source on the horizontal plane is shown in Figure 5. The projection is located in the sound-image control interval VI, and the sound-image The control interval VI is the target audio image control interval, and the sound emitting device to be worked is determined from the two sound emitting devices responsible for the sound image control interval VI, that is, the sound emitting device to be worked is selected from the sound emitting device a and the sound emitting device d, so as to improve the The accuracy of sound image construction improves the sense of space in the sound field of the head-mounted device.
作为一种可选的实施例,在重构目标声像控制区间的发声装置中选择待工作发声装置的过程包括:As an optional embodiment, the process of selecting a sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval includes:
判断投影是否处于划分声像控制区间的任一连线上;Judging whether the projection is on any connection line dividing the audio-image control interval;
若是,将重构目标声像控制区间的两个发声装置中,投影所处的连线对应的发声装置作为待工作发声装置;If so, among the two sound emitting devices in the reconstruction target sound image control section, the sound emitting device corresponding to the line where the projection is located is taken as the sound emitting device to be worked;
若否,将重构目标声像控制区间的两个发声装置均作为待工作发声装置。If not, the two sound emitting devices in the reconstructed target sound image control interval are both regarded as standby sound emitting devices.
具体的,在选择待工作发声装置时,需要先确定虚拟声源的方位,判断虚拟声源是在两个发声装置之间,还是和任一发声装置有重合,如果有重合,那么仅需要与虚拟声源重合的发声装置发声即可。具体的,可以对投影是否处于划分声像控制区间的任一连线上进行判断,如果投影处于划分声像控制区间的任一连线上,则将其所在连线对应的发声装置作为待工作发声装置,如果否,说明虚拟声源的方位在构成目标声像控制区间的两个发声装置之间,此时将这两个发声装置均作为待工作发声装置。Specifically, when selecting a sounding device to be operated, it is necessary to determine the orientation of the virtual sound source first, and determine whether the virtual sound source is between two sounding devices or overlaps with any sounding device. The sounding device that overlaps the virtual sound source can produce sound. Specifically, it can be judged whether the projection is on any connection that divides the audio-image control interval, and if the projection is on any connection that divides the audio-image control interval, then use the sounding device corresponding to the connection where it is located as the waiting to work Sounding device, if not, it means that the orientation of the virtual sound source is between the two sounding devices constituting the target sound image control section, and these two sounding devices are both regarded as the waiting sounding devices at this time.
具体的,可以预先确定各发声装置与目标方向之间的角度,这里的目标方向可以为坐标轴的x轴正方向或x轴负方向或y轴正方向或y轴负方向等,在确定角度时所有发声装置均按顺时针方向或均按逆时针方向确定,假设均按顺时针方向确定各发声装置与y轴正方向之间的角度,即θa~θf。当确定虚拟声源的入射方向后,判断虚拟声源的入射方向在水平面上的投影按顺时针方向与y轴正方向之间的角度θr,判断θa~θf中是否存在与该θr相同的角度,假设θr和θa相同,那么发声装置a即为待工作发声装置,如果θd<θr<θa,那么发声装置d和发声装置a为待工作发声装置。Specifically, the angle between each sound emitting device and the target direction can be determined in advance. The target direction here can be the positive direction of the x-axis or the negative direction of the x-axis or the positive direction of the y-axis or the negative direction of the y-axis of the coordinate axis. When all the sounding devices are determined clockwise or counterclockwise, it is assumed that the angle between each sounding device and the positive direction of the y-axis is determined clockwise, that is, θa~θf. After determining the incident direction of the virtual sound source, determine the angle θr between the projection of the incident direction of the virtual sound source on the horizontal plane in the clockwise direction and the positive direction of the y-axis, and determine whether there is the same angle as the θr in θa~θf , assuming that θr and θa are the same, then the sounding device a is the waiting sounding device, if θd<θr<θa, then the sounding device d and the sounding device a are the waiting sounding devices.
作为一种可选的实施例,所述基于所述虚拟声源相对于所述头戴设备的 位置计算幅度调制参数的过程包括:As an optional embodiment, the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;determining and reconstructing the bisector of the angle between the two sounding devices in the target sound image control interval based on the position of the midpoint of the human head;
根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
Figure PCTCN2021131709-appb-000005
Calculate the respective amplitude modulation parameters of the two sounding devices according to the first relational expression, the first relational expression is
Figure PCTCN2021131709-appb-000005
其中,
Figure PCTCN2021131709-appb-000006
为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
Figure PCTCN2021131709-appb-000007
为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
in,
Figure PCTCN2021131709-appb-000006
is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
Figure PCTCN2021131709-appb-000007
is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
具体的,在确定目标声像区间对应的两个待工作发声装置后,可以通过幅度平移的方式对两个待工作发声装置的幅度调制参数进行分配,为了剔除正弦法则计算幅度平移时包含的头部转动的影响,本申请采用第一关系式计算两个待工作发声装置的幅度调制参数。参照图6所示,建立以人头为原点的坐标系,两个待工作发声装置均位于听音点(人头中点位置)为圆心的圆上,两个待工作发声装置的角度分别为
Figure PCTCN2021131709-appb-000008
Figure PCTCN2021131709-appb-000009
虚拟声源与两个发声装置的等分线的夹角为
Figure PCTCN2021131709-appb-000010
本实施例中,将与虚拟声源距离更近的一个发声装置作为第一个发声装置,另一个发声装置作为第二发声装置。进一步的,分别计算出第一个发声装置的幅度调制参数g1和第二个发声装置的幅度调制参数g2后,第一个发声装置和第二发声装置分别根据g1、g2及头相关传输函数对虚拟声源的初始音频信号进行立体音效调制,并分别输出各自调制后的音频信号。
Specifically, after determining the two to-be-operated sound-emitting devices corresponding to the target sound image interval, the amplitude modulation parameters of the two to-be-operated sound-emitting devices can be allocated by means of amplitude translation. In order to avoid the influence of head rotation, the present application uses the first relational expression to calculate the amplitude modulation parameters of the two to-be-operated sound emitting devices. Referring to Figure 6, establish a coordinate system with the human head as the origin, and the two sounding devices to be worked are located on the circle centered at the listening point (the midpoint of the human head), and the angles of the two sounding devices to be worked are respectively
Figure PCTCN2021131709-appb-000008
and
Figure PCTCN2021131709-appb-000009
The angle between the virtual sound source and the bisector of the two sounding devices is
Figure PCTCN2021131709-appb-000010
In this embodiment, a sound-generating device closer to the virtual sound source is used as the first sound-generating device, and the other sound-generating device is used as the second sound-generating device. Further, after calculating the amplitude modulation parameter g1 of the first sounding device and the amplitude modulation parameter g2 of the second sounding device respectively, the first sounding device and the second sounding device are respectively based on g1, g2 and the head-related transfer function. The initial audio signal of the virtual sound source is subjected to stereo sound modulation, and the modulated audio signals are respectively output.
真实的声源从箭头方向发出被人耳接收到的双耳幅度差,传统VR在耳边播放声音被人耳接收到的双耳幅度差,以及根据多扬声器分配幅度输出声 音被接收到的双耳幅度差对比结果如图7所示,不难看出经过幅度平移分配的声音双耳幅度差与原始声音的双耳幅度差更接近,用户人耳判断重构的虚拟声像的位置会比传统VR播放的情况下更接近真实方位。The real sound source sends out the binaural amplitude difference received by the human ear from the direction of the arrow, and the binaural amplitude difference received by the human ear when the sound played by traditional VR is played by the ear, and the binaural amplitude difference received by the output sound according to the multi-speaker distribution amplitude. The comparison results of the ear amplitude difference are shown in Figure 7. It is not difficult to see that the binaural amplitude difference of the sound allocated by the amplitude translation is closer to the binaural amplitude difference of the original sound. In the case of VR playback, it is closer to the real position.
综上所述,本申请在VR设备上增加发声装置的数量,以正切法则的幅度平移分配不同发声装置声音的幅度大小,并利用红外线扫描人头轮廓计算头传函数补充对扬声器的相位调制,来控制人耳对声源位置的感知,提升VR产品的声场空间感。To sum up, this application increases the number of sounding devices on the VR device, and distributes the amplitude of the sound of different sounding devices with the amplitude translation of the tangent law, and uses infrared scanning to calculate the head transfer function to supplement the phase modulation of the speaker to achieve Control the human ear's perception of the position of the sound source, and improve the sense of space in the sound field of VR products.
请参照图8,图8为本申请所提供的一种音频输出控制系统的结构示意图,该音频输出控制系统包括:Please refer to FIG. 8. FIG. 8 is a schematic structural diagram of an audio output control system provided by the present application. The audio output control system includes:
获取模块11,用于获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;An acquisition module 11, configured to acquire an initial audio signal of a virtual sound source in the current scene where the head-mounted device is located;
第一计算模块12,用于基于头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;The first calculation module 12 is used to calculate the phase modulation parameter based on the current head profile parameters of the wearer of the head-mounted device;
第二计算模块13,用于基于虚拟声源相对于头戴设备的位置计算幅度调制参数;The second calculation module 13 is used to calculate the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device;
输出控制模块14,用于控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。The output control module 14 is used to control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
可见,本实施例中,根据虚拟声源相对于头戴设备的位置计算发声装置的幅度调制参数,同时获取头戴设备的佩戴者的当前人头轮廓参数,基于人头轮廓参数计算发声装置的相位调制参数,然后控制发声装置按照幅度调制参数和相位调制参数对虚拟声源的初始音频信号进行立体音效的调制然后再输出,本申请既考虑了发声装置输出的声音幅度的大小,重构虚拟声源在水平面上的空间感,还根据当前人头轮廓参数补充校正幅度平移的虚拟声源的相位,以此重构虚拟声源在竖直方向上的空间感,因此,本申请的发声装置输出的音频信号能够使佩戴者更好的感知声源的真实方位,避免产生声源头 中定位效应,提升了头戴设备的声场空间感和用户体验。It can be seen that in this embodiment, the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it. This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source The sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application The signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
作为一种可选的实施例,相位调制参数包括头相关传输函数。As an optional embodiment, the phase modulation parameter includes a head-related transfer function.
作为一种可选的实施例,头戴设备包括n个发声装置,n为大于12的整数;As an optional embodiment, the head-mounted device includes n sound emitting devices, where n is an integer greater than 12;
该音频输出控制系统还包括:The audio output control system also includes:
确定模块,用于基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置;A determining module, configured to determine a sound emitting device to be operated among all sound emitting devices based on the position of the virtual sound source relative to the head-mounted device;
控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:The process of controlling the sounding device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal includes:
控制头戴设备中的待工作发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the ready-to-work sound emitting device in the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter, and output the modulated audio signal.
作为一种可选的实施例,每两个相邻的发声装置重构一个声像控制区间;As an optional embodiment, every two adjacent sounding devices reconstruct a sound image control interval;
基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置的过程包括:The process of determining the sound emitting device to work among all the sound emitting devices based on the position of the virtual sound source relative to the headset includes:
基于虚拟声源相对于头戴设备的位置得到虚拟声源的入射方向;The incident direction of the virtual sound source is obtained based on the position of the virtual sound source relative to the head-mounted device;
在所有声像控制区间中确定入射方向所在的目标声像控制区间;Determining the target sound image control interval where the incident direction is located in all sound image control intervals;
在重构目标声像控制区间的发声装置中选择待工作发声装置。Select the to-be-worked sounding device among the sounding devices in the reconstruction target sound image control section.
作为一种可选的实施例,n个发声装置位于同一水平面,n个发声装置与水平面上的人头中点位置的连线将水平面划分为n个声像控制区间;As an optional embodiment, the n sounding devices are located on the same horizontal plane, and the line connecting the n sounding devices and the midpoint of the human head on the horizontal plane divides the horizontal plane into n sound image control intervals;
在所有声像控制区间中确定入射方向所在的目标声像控制区间的过程包括:The process of determining the target panning interval in which the incident direction is located among all panning intervals includes:
获取入射方向在水平面上的投影;Obtain the projection of the incident direction on the horizontal plane;
将投影所在的声像控制区间确定为目标声像控制区间。The sound image control interval where the projection is located is determined as the target sound image control interval.
作为一种可选的实施例,在重构目标声像控制区间的发声装置中选择待工作发声装置的过程包括:As an optional embodiment, the process of selecting a sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval includes:
判断投影是否处于划分声像控制区间的任一连线上;Judging whether the projection is on any connection line dividing the audio-image control interval;
若是,将重构目标声像控制区间的两个发声装置中,投影所处的连线对应的发声装置作为待工作发声装置;If so, among the two sound emitting devices in the reconstruction target sound image control section, the sound emitting device corresponding to the line where the projection is located is taken as the sound emitting device to be worked;
若否,将重构目标声像控制区间的两个发声装置均作为待工作发声装置。If not, the two sound emitting devices in the reconstructed target sound image control interval are both regarded as standby sound emitting devices.
作为一种可选的实施例,所述基于所述虚拟声源相对于所述头戴设备的 位置计算幅度调制参数的过程包括:As an optional embodiment, the process of calculating the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device includes:
基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;determining and reconstructing the bisector of the angle between the two sounding devices in the target sound image control interval based on the position of the midpoint of the human head;
根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
Figure PCTCN2021131709-appb-000011
Calculate the respective amplitude modulation parameters of the two sounding devices according to the first relational expression, the first relational expression is
Figure PCTCN2021131709-appb-000011
其中,
Figure PCTCN2021131709-appb-000012
为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
Figure PCTCN2021131709-appb-000013
为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
in,
Figure PCTCN2021131709-appb-000012
is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
Figure PCTCN2021131709-appb-000013
is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
另外,本申请提供了一种头戴设备,包括:In addition, the present application provides a head-mounted device, including:
存储器,用于存储计算机程序;memory for storing computer programs;
处理器,用于执行计算机程序时实现如上文任意一个实施例所描述的音频输出控制方法的步骤。The processor is configured to implement the steps of the audio output control method described in any one of the above embodiments when executing the computer program.
具体的,存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统和计算机可读指令,该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。处理器为头戴设备提供计算和控制能力,执行存储器中保存的计算机程序时,可以实现以下步骤:获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;基于头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;基于虚拟声源相对于头戴设备的位置计算幅度调制参数;控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。Specifically, the memory includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and computer-readable instructions, and the internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium. The processor provides calculation and control capabilities for the head-mounted device. When executing the computer program stored in the memory, the following steps can be implemented: obtaining the initial audio signal of the virtual sound source in the current scene where the head-mounted device is located; Calculate the phase modulation parameter based on the current head contour parameters of the user; calculate the amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device; control the sounding device of the head-mounted device to perform stereo sound effects on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter Modulate and output the modulated audio signal.
可见,本实施例中,根据虚拟声源相对于头戴设备的位置计算发声装置的幅度调制参数,同时获取头戴设备的佩戴者的当前人头轮廓参数,基于人头轮廓参数计算发声装置的相位调制参数,然后控制发声装置按照幅度调制参数和相位调制参数对虚拟声源的初始音频信号进行立体音效的调制然后再输出,本申请既考虑了发声装置输出的声音幅度的大小,重构虚拟声源在水平面上的空间感,还根据当前人头轮廓参数补充校正幅度平移的虚拟声源的相位,以此重构虚拟声源在竖直方向上的空间感,因此,本申请的发声装置 输出的音频信号能够使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。It can be seen that in this embodiment, the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it. This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source The sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application The signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:基于头戴设备的佩戴者的当前人头轮廓参数计算头相关传输函数,将头相关传输函数作为相位调制参数。As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps can be implemented: calculate the head-related transfer function based on the current head contour parameters of the wearer of the head-mounted device, and use the head-related transfer function as Phase modulation parameters.
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置;控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:控制头戴设备中的待工作发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps can be implemented: determine the sound emitting device to be worked among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device; control the head The sounding device of the wearable device performs stereo modulation on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and outputs the modulated audio signal. The parameter performs stereo modulation on the original audio signal and outputs the modulated audio signal.
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:基于虚拟声源相对于头戴设备的位置得到虚拟声源的入射方向;在所有声像控制区间中确定入射方向所在的目标声像控制区间,每两个相邻的发声装置重构一个声像控制区间;在重构目标声像控制区间的发声装置中选择待工作发声装置。As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps can be implemented: obtain the incident direction of the virtual sound source based on the position of the virtual sound source relative to the head-mounted device; In the interval, determine the target sound image control interval where the incident direction is located, and reconstruct a sound image control interval for every two adjacent sound emitting devices; select the sound emitting device to be operated among the sound emitting devices in the reconstructed target sound image control interval.
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:获取入射方向在水平面上的投影;将投影所在的声像控制区间确定为目标声像控制区间。As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps can be implemented: obtain the projection of the incident direction on the horizontal plane; determine the sound image control interval where the projection is located as the target sound image control interval .
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:判断投影是否处于划分声像控制区间的任一连线上;若是,将重构目标声像控制区间的两个发声装置中,投影所处的连线对应的发声装置作为待工作发声装置;若否,将重构目标声像控制区间的两个发声装置均作为待工作发声装置。As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps can be implemented: determine whether the projection is on any connection line dividing the sound image control interval; if so, reconstruct the target sound image Among the two sounding devices in the control section, the sounding device corresponding to the line where the projection is located is taken as the waiting sounding device; if not, the two sounding devices in the control section of the reconstructed target sound image are both taken as the waiting sounding device.
作为一种可选的实施例,处理器执行存储器中保存的计算机子程序时,可以实现以下步骤:基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;As an optional embodiment, when the processor executes the computer subroutine stored in the memory, the following steps may be implemented: determining and reconstructing the two sounding devices of the target sound image control interval based on the midpoint position of the human head The bisector of the included angle;
根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
Figure PCTCN2021131709-appb-000014
其中,
Figure PCTCN2021131709-appb-000015
为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
Figure PCTCN2021131709-appb-000016
为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
Calculate the respective amplitude modulation parameters of the two sounding devices according to the first relational expression, the first relational expression is
Figure PCTCN2021131709-appb-000014
in,
Figure PCTCN2021131709-appb-000015
is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
Figure PCTCN2021131709-appb-000016
is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
在上述实施例的基础上,作为优选实施方式,参见图X,图x为本申请实施例提供的另一种头戴设备的结构图,该头戴设备还包括:On the basis of the above-mentioned embodiments, as a preferred implementation mode, refer to Figure X, which is a structural diagram of another head-mounted device provided by the embodiment of the present application. The head-mounted device also includes:
输入接口,与处理器相连,用于获取外部导入的计算机程序、参数和指令,经处理器控制保存至存储器中。该输入接口可以与输入装置相连,接收用户手动输入的参数或指令。该输入装置可以是显示屏上覆盖的触摸层,也可以是终端外壳上设置的按键、轨迹球或触控板。The input interface is connected with the processor, and is used to obtain the computer program, parameters and instructions imported from the outside, and store them in the memory under the control of the processor. The input interface can be connected with an input device to receive parameters or instructions manually input by the user. The input device may be a touch layer covered on the display screen, or a button, a trackball or a touch pad provided on the terminal shell.
显示单元,与处理器相连,用于显示处理器发送的数据。该显示单元可以为液晶显示屏或者电子墨水显示屏等。The display unit is connected with the processor and used for displaying the data sent by the processor. The display unit may be a liquid crystal display or an electronic ink display.
网络端口,与处理器相连,用于与外部各终端设备进行通信连接。该通信连接所采用的通信技术可以为有线通信技术或无线通信技术,如移动高清链接技术(MHL)、通用串行总线(USB)、高清多媒体接口(HDMI)、无线保真技术(WiFi)、蓝牙通信技术、低功耗蓝牙通信技术、基于IEEE802.11s的通信技术等。The network port is connected with the processor and is used for communication connection with various external terminal devices. The communication technology used in the communication connection can be wired communication technology or wireless communication technology, such as mobile high-definition link technology (MHL), universal serial bus (USB), high-definition multimedia interface (HDMI), wireless fidelity technology (WiFi), Bluetooth communication technology, low-power Bluetooth communication technology, communication technology based on IEEE802.11s, etc.
另一方面,本申请还提供了一种计算机可读存储介质,计算机可读存储介质上存储有计算机程序,计算机程序被处理器执行时实现如上文任意一个实施例所描述的音频输出控制方法的步骤。On the other hand, the present application also provides a computer-readable storage medium, on which a computer program is stored. When the computer program is executed by a processor, the audio output control method as described in any one of the above embodiments is implemented. step.
具体的,计算机存储介质可以包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。该存储介质上存储有计算机程序,计算机程序被处理器执行时实现以下步骤:获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;基于头戴设备的佩戴者的当前 人头轮廓参数计算相位调制参数;基于虚拟声源相对于头戴设备的位置计算幅度调制参数;控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。Specifically, the computer storage medium may include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk, etc., which can store program codes. medium. A computer program is stored on the storage medium, and when the computer program is executed by the processor, the following steps are implemented: obtaining the initial audio signal of the virtual sound source in the current scene where the head-mounted device is located; based on the current human head contour parameters of the wearer of the head-mounted device Calculate the phase modulation parameters; calculate the amplitude modulation parameters based on the position of the virtual sound source relative to the head-mounted device; control the sounding device of the head-mounted device to modulate the stereo sound effect of the initial audio signal according to the amplitude modulation parameters and phase modulation parameters and output the modulated audio signal.
可见,本实施例中,根据虚拟声源相对于头戴设备的位置计算发声装置的幅度调制参数,同时获取头戴设备的佩戴者的当前人头轮廓参数,基于人头轮廓参数计算发声装置的相位调制参数,然后控制发声装置按照幅度调制参数和相位调制参数对虚拟声源的初始音频信号进行立体音效的调制然后再输出,本申请既考虑了发声装置输出的声音幅度的大小,重构虚拟声源在水平面上的空间感,还根据当前人头轮廓参数补充校正幅度平移的虚拟声源的相位,以此重构虚拟声源在竖直方向上的空间感,因此,本申请的发声装置输出的音频信号能够使佩戴者更好的感知声源的真实方位,避免产生声源头中定位效应,提升了头戴设备的声场空间感和用户体验。It can be seen that in this embodiment, the amplitude modulation parameters of the sounding device are calculated according to the position of the virtual sound source relative to the head-mounted device, and at the same time, the current head contour parameters of the wearer of the head-mounted device are obtained, and the phase modulation of the sounding device is calculated based on the head contour parameters parameter, and then control the sounding device to modulate the stereo sound effect on the initial audio signal of the virtual sound source according to the amplitude modulation parameter and the phase modulation parameter, and then output it. This application not only considers the magnitude of the sound output by the sounding device, but also reconstructs the virtual sound source The sense of space on the horizontal plane also supplements and corrects the phase of the virtual sound source with amplitude translation according to the current head contour parameters, so as to reconstruct the sense of space of the virtual sound source in the vertical direction. Therefore, the audio output by the sounding device of the present application The signal can enable the wearer to better perceive the real orientation of the sound source, avoid the localization effect of the sound source, and improve the spatial sense of the sound field and user experience of the headset.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:基于头戴设备的佩戴者的当前人头轮廓参数计算头相关传输函数,将头相关传输函数作为相位调制参数。As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: calculating the head-related transfer function based on the current head profile parameters of the wearer of the head-mounted device, and using the head-related transfer function as a phase modulation parameter.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:基于虚拟声源相对于头戴设备的位置在所有发声装置中确定待工作发声装置;控制头戴设备的发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:控制头戴设备中的待工作发声装置按幅度调制参数及相位调制参数对初始音频信号进行立体音效的调制并输出调制后的音频信号。As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: based on the position of the virtual sound source relative to the head-mounted device, among all the sound-emitting devices, the sound-emitting devices to be worked are determined; the sound-emitting device of the head-mounted device is controlled to press The amplitude modulation parameter and phase modulation parameter modulate the initial audio signal for stereo sound effects and output the modulated audio signal. Modulate the stereo effect and output the modulated audio signal.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:基于虚拟声源相对于头戴设备的位置得到虚拟声源的入射方向;在所有声像控制区间中确定入射方向所在的目标声像控制区间,每两个相邻的发声装置重构一个声像控制区间;在重构目标声像控制区间的发声装置中选择待工作发声装置。As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: obtaining the incident direction of the virtual sound source based on the position of the virtual sound source relative to the head-mounted device; determining where the incident direction is located in all sound image control intervals The target sound image control interval of each two adjacent sound emitting devices reconstructs a sound image control interval; select the sound emitting device to be operated among the sound emitting devices of the reconstructed target sound image control interval.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:获取入射方向在水平面上的投影;将投影所在的声像控制区间确定为目标声像控制区间。As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: obtaining the projection of the incident direction on the horizontal plane; determining the sound image control interval where the projection is located as the target sound image control interval.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:判 断投影是否处于划分声像控制区间的任一连线上;若是,将重构目标声像控制区间的两个发声装置中,投影所处的连线对应的发声装置作为待工作发声装置;若否,将重构目标声像控制区间的两个发声装置均作为待工作发声装置。As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: judging whether the projection is on any connection line dividing the sound image control interval; if so, reconstructing the two vocalizations of the target sound image control interval In the device, the sounding device corresponding to the line where the projection is located is regarded as the sounding device to be operated; if not, the two sounding devices in the control area of the reconstructed target sound image are both regarded as the sounding device to be operated.
作为一种可选的实施例,计算机程序被处理器执行时实现以下步骤:基于人头中点位置确定两个待工作发声装置的夹角的等分线;基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
Figure PCTCN2021131709-appb-000017
其中,
Figure PCTCN2021131709-appb-000018
为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
Figure PCTCN2021131709-appb-000019
为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
As an optional embodiment, when the computer program is executed by the processor, the following steps are implemented: determining the bisector of the angle between the two to-be-operated sounding devices based on the midpoint position of the human head; determining the reconstruction based on the midpoint position of the human head The bisector of the angle between the two sounding devices in the target sound image control interval; the respective amplitude modulation parameters of the two sounding devices are calculated according to the first relational expression, and the first relational expression is
Figure PCTCN2021131709-appb-000017
in,
Figure PCTCN2021131709-appb-000018
is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
Figure PCTCN2021131709-appb-000019
is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
还需要说明的是,在本说明书中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的状况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should also be noted that in this specification, relative terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply that these entities or operations There is no such actual relationship or order between the operations. Furthermore, the term "comprises", "comprises" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, but also includes elements not expressly listed. other elements of or also include elements inherent in such a process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not exclude the presence of additional identical elements in the process, method, article or apparatus comprising said element.
对所公开的实施例的上述说明,使本领域专业技术人员能够实现或使用本申请。对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本申请的精神或范围的情况下,在其他实施例中实现。因此,本申请将不会被限制于本文所示的这些实施例,而是要符合与本文所公开的原理和新颖特点相一致的最宽的范围。The above description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be implemented in other embodiments without departing from the spirit or scope of the application. Therefore, the present application will not be limited to the embodiments shown herein, but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

  1. 一种音频输出控制方法,其特征在于,包括:An audio output control method, characterized in that, comprising:
    获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;Obtain the initial audio signal of the virtual sound source in the current scene where the headset is located;
    基于所述头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;calculating phase modulation parameters based on current head profile parameters of the wearer of the head-mounted device;
    基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数;calculating an amplitude modulation parameter based on a position of the virtual sound source relative to the head-mounted device;
    控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
  2. 根据权利要求1所述的音频输出控制方法,其特征在于,所述相位调制参数包括头相关传输函数。The audio output control method according to claim 1, wherein the phase modulation parameter includes a head-related transfer function.
  3. 根据权利要求1所述的音频输出控制方法,其特征在于,所述头戴设备包括n个所述发声装置,n为大于2的整数;The audio output control method according to claim 1, wherein the head-mounted device includes n sound emitting devices, and n is an integer greater than 2;
    该音频输出控制方法还包括:The audio output control method also includes:
    基于所述虚拟声源相对于所述头戴设备的位置在所有所述发声装置中确定待工作发声装置;determining a to-be-operated sound emitting device among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device;
    所述控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号的过程包括:The process of controlling the sounding device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal includes:
    控制头戴设备中的所述待工作发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。Controlling the standby sound generating device in the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
  4. 根据权利要求3所述的音频输出控制方法,其特征在于,每两个相邻的所述发声装置重构一个声像控制区间;The audio output control method according to claim 3, wherein each two adjacent sound emitting devices reconstruct a sound image control interval;
    所述基于所述虚拟声源相对于所述头戴设备的位置在所有所述发声装置中确定待工作发声装置的过程包括:The process of determining the sound emitting device to be operated among all the sound emitting devices based on the position of the virtual sound source relative to the head-mounted device includes:
    基于所述虚拟声源相对于所述头戴设备的位置得到所述虚拟声源的入射方向;obtaining an incident direction of the virtual sound source based on a position of the virtual sound source relative to the head-mounted device;
    在所有所述声像控制区间中确定所述入射方向所在的目标声像控制区间;determining the target sound image control interval where the incident direction is located in all the sound image control intervals;
    在重构所述目标声像控制区间的所述发声装置中选择待工作发声装置。Selecting a to-be-worked sound emitting device among the sound emitting devices reconstructing the target sound image control interval.
  5. 根据权利要求4所述的音频输出控制方法,其特征在于,n个所述发声装置位于同一水平面,n个所述发声装置与所述水平面上的人头中点位置的连线将所述水平面划分为n个所述声像控制区间;The audio output control method according to claim 4, wherein the n sound emitting devices are located on the same horizontal plane, and the line connecting the n sound emitting devices and the midpoint of the human head on the horizontal plane divides the horizontal plane being n said sound and image control intervals;
    所述在所有所述声像控制区间中确定所述入射方向所在的目标声像控制区间的过程包括:The process of determining the target sound image control interval where the incident direction is located in all the sound image control intervals includes:
    获取所述入射方向在所述水平面上的投影;Obtain the projection of the incident direction on the horizontal plane;
    将所述投影所在的所述声像控制区间确定为目标声像控制区间。The sound image control interval where the projection is located is determined as a target sound image control interval.
  6. 根据权利要求5所述的音频输出控制方法,其特征在于,所述在重构所述目标声像控制区间的所述发声装置中选择待工作发声装置的过程包括:The audio output control method according to claim 5, wherein the process of selecting a sound emitting device to be operated from among the sound emitting devices reconstructing the target sound image control interval comprises:
    判断所述投影是否处于划分所述声像控制区间的任一所述连线上;judging whether the projection is on any of the lines dividing the audio-image control interval;
    若是,将重构所述目标声像控制区间的两个所述发声装置中,所述投影所处的所述连线对应的所述发声装置作为待工作发声装置;If yes, among the two sound emitting devices reconstructing the target sound image control interval, the sound emitting device corresponding to the line where the projection is located is used as the sound emitting device to be operated;
    若否,将重构所述目标声像控制区间的两个所述发声装置均作为所述待工作发声装置。If not, the two sound emitting devices reconstructing the target sound image control interval are both used as the waiting sound emitting devices.
  7. 根据权利要求1-6任意一项所述的音频输出控制方法,其特征在于,所述基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数的过程包括:The audio output control method according to any one of claims 1-6, wherein the process of calculating amplitude modulation parameters based on the position of the virtual sound source relative to the head-mounted device comprises:
    基于所述人头中点位置确定重构所述目标声像控制区间的两个所述发声装置的夹角的等分线;determining and reconstructing the bisector of the angle between the two sounding devices in the target sound image control interval based on the position of the midpoint of the human head;
    根据第一关系式计算两个所述发声装置各自的幅度调制参数,所述第一关系式为
    Figure PCTCN2021131709-appb-100001
    Calculate the respective amplitude modulation parameters of the two sounding devices according to the first relational expression, the first relational expression is
    Figure PCTCN2021131709-appb-100001
    其中,
    Figure PCTCN2021131709-appb-100002
    为任意一个所述发声装置与所述人头中点位置的连线与所述等分线的夹角,
    Figure PCTCN2021131709-appb-100003
    为所述投影与所述等分线的夹角,g1为第一个所述发声装置的幅度调制参数,g2为第二个所述发声装置的幅度调制参数。
    in,
    Figure PCTCN2021131709-appb-100002
    is the angle between any one of the sounding device and the midpoint of the human head and the bisector,
    Figure PCTCN2021131709-appb-100003
    is the angle between the projection and the bisector, g1 is the amplitude modulation parameter of the first sounding device, and g2 is the amplitude modulation parameter of the second sounding device.
  8. 一种音频输出控制系统,其特征在于,包括:An audio output control system, characterized in that it comprises:
    获取模块,用于获取头戴设备所在的当前场景中的虚拟声源的初始音频信号;An acquisition module, configured to acquire an initial audio signal of a virtual sound source in the current scene where the head-mounted device is located;
    第一计算模块,用于基于所述头戴设备的佩戴者的当前人头轮廓参数计算相位调制参数;A first calculation module, configured to calculate a phase modulation parameter based on the current head profile parameters of the wearer of the head-mounted device;
    第二计算模块,用于基于所述虚拟声源相对于所述头戴设备的位置计算幅度调制参数;A second calculation module, configured to calculate an amplitude modulation parameter based on the position of the virtual sound source relative to the head-mounted device;
    输出控制模块,用于控制所述头戴设备的发声装置按所述幅度调制参数及所述相位调制参数对所述初始音频信号进行立体音效的调制并输出调制后的音频信号。The output control module is used to control the sound generating device of the head-mounted device to modulate the stereo sound effect on the initial audio signal according to the amplitude modulation parameter and the phase modulation parameter and output the modulated audio signal.
  9. 一种头戴设备,其特征在于,包括:A head-mounted device, characterized in that it includes:
    存储器,用于存储计算机程序;memory for storing computer programs;
    处理器,用于执行所述计算机程序时实现如权利要求1-7任意一项所述的音频输出控制方法的步骤。A processor, configured to implement the steps of the audio output control method according to any one of claims 1-7 when executing the computer program.
  10. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1-7任意一项所述的音频输出控制方法的步骤。A computer-readable storage medium, characterized in that a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the audio output control according to any one of claims 1-7 is realized method steps.
PCT/CN2021/131709 2021-10-29 2021-11-19 Audio output control method and system, and related assemblies WO2023070778A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111276794.0 2021-10-29
CN202111276794.0A CN114025287B (en) 2021-10-29 2021-10-29 Audio output control method, system and related components

Publications (1)

Publication Number Publication Date
WO2023070778A1 true WO2023070778A1 (en) 2023-05-04

Family

ID=80059006

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/131709 WO2023070778A1 (en) 2021-10-29 2021-11-19 Audio output control method and system, and related assemblies

Country Status (2)

Country Link
CN (1) CN114025287B (en)
WO (1) WO2023070778A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107168518A (en) * 2017-04-05 2017-09-15 北京小鸟看看科技有限公司 A kind of synchronous method, device and head-mounted display for head-mounted display
CN107995583A (en) * 2016-10-26 2018-05-04 宏达国际电子股份有限公司 Sound playing method, system and its non-transient computer medium capable of reading record
JP2018125784A (en) * 2017-02-03 2018-08-09 Hoyaサービス株式会社 Sound output device
WO2019045622A1 (en) * 2017-08-31 2019-03-07 Terranet Ab Headset and method of operating headset
CN110267166A (en) * 2019-07-16 2019-09-20 上海艺瓣文化传播有限公司 A kind of virtual sound field real-time interaction system based on binaural effect
CN113382350A (en) * 2016-09-23 2021-09-10 苹果公司 Method, system, and medium for coordinated tracking of binaural audio rendering

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05336599A (en) * 1992-06-03 1993-12-17 Fujitsu Ltd Sound image localization headphone device and virtual reality audio-visual equipment using it
CN103716748A (en) * 2007-03-01 2014-04-09 杰里·马哈布比 Audio spatialization and environment simulation
CN103618986B (en) * 2013-11-19 2015-09-30 深圳市新一代信息技术研究院有限公司 The extracting method of source of sound acoustic image body and device in a kind of 3d space
CN105910702B (en) * 2016-04-18 2019-01-25 北京大学 A kind of asynchronous head-position difficult labor measurement method based on phase compensation
EP3346726A1 (en) * 2017-01-04 2018-07-11 Harman Becker Automotive Systems GmbH Arrangements and methods for active noise cancelling
JP2018110366A (en) * 2017-01-04 2018-07-12 伊藤 和幸 3d sound video audio apparatus
CN108574925A (en) * 2017-03-14 2018-09-25 株式会社东芝 The method and apparatus that audio signal output is controlled in virtual auditory environment
CN108966110B (en) * 2017-05-19 2020-02-14 华为技术有限公司 Sound signal processing method, device and system, terminal and storage medium
CN107480100B (en) * 2017-07-04 2020-02-28 中国科学院自动化研究所 Head-related transfer function modeling system based on deep neural network intermediate layer characteristics
CN112292872A (en) * 2018-06-26 2021-01-29 索尼公司 Sound signal processing device, mobile device, method, and program

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113382350A (en) * 2016-09-23 2021-09-10 苹果公司 Method, system, and medium for coordinated tracking of binaural audio rendering
CN107995583A (en) * 2016-10-26 2018-05-04 宏达国际电子股份有限公司 Sound playing method, system and its non-transient computer medium capable of reading record
JP2018125784A (en) * 2017-02-03 2018-08-09 Hoyaサービス株式会社 Sound output device
CN107168518A (en) * 2017-04-05 2017-09-15 北京小鸟看看科技有限公司 A kind of synchronous method, device and head-mounted display for head-mounted display
WO2019045622A1 (en) * 2017-08-31 2019-03-07 Terranet Ab Headset and method of operating headset
CN110267166A (en) * 2019-07-16 2019-09-20 上海艺瓣文化传播有限公司 A kind of virtual sound field real-time interaction system based on binaural effect

Also Published As

Publication number Publication date
CN114025287A (en) 2022-02-08
CN114025287B (en) 2023-02-17

Similar Documents

Publication Publication Date Title
US11557055B2 (en) Arrangement for producing head related transfer function filters
US9767618B2 (en) Adaptive ambisonic binaural rendering
CN111615834B (en) Method, system and apparatus for sweet spot adaptation of virtualized audio
Bronkhorst Localization of real and virtual sound sources
US20200128347A1 (en) Head-Related Impulse Responses for Area Sound Sources Located in the Near Field
JP7165215B2 (en) Virtual Reality, Augmented Reality, and Mixed Reality Systems with Spatialized Audio
JP7194271B2 (en) Near-field audio rendering
JP2008506140A (en) Horizontal perspective display
US20200097251A1 (en) Emoji to Select How or Where Sound Will Localize to a Listener
JP7210602B2 (en) Method and apparatus for processing audio signals
US11356795B2 (en) Spatialized audio relative to a peripheral device
US11297456B2 (en) Moving an emoji to move a location of binaural sound
US11943602B1 (en) Systems and methods for spatial update latency compensation for head-tracked audio
CN111696513A (en) Audio signal processing method and device, electronic equipment and storage medium
WO2021143574A1 (en) Augmented reality glasses, augmented reality glasses-based ktv implementation method and medium
CN111818441B (en) Sound effect realization method and device, storage medium and electronic equipment
JP2022500917A (en) Equipment and methods for processing audiovisual data
WO2023070778A1 (en) Audio output control method and system, and related assemblies
CN111246345B (en) Method and device for real-time virtual reproduction of remote sound field
JP6756777B2 (en) Information processing device and sound generation method
WO2019174442A1 (en) Adapterization equipment, voice output method, device, storage medium and electronic device
TW202031058A (en) Method and system for correcting energy distributions of audio signal
US11589184B1 (en) Differential spatial rendering of audio sources
CN115623156B (en) Audio processing method and related device
CN116193196A (en) Virtual surround sound rendering method, device, equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21962122

Country of ref document: EP

Kind code of ref document: A1