WO2023169367A1 - Procédé de lecture audio et dispositif électronique - Google Patents

Procédé de lecture audio et dispositif électronique Download PDF

Info

Publication number
WO2023169367A1
WO2023169367A1 PCT/CN2023/079874 CN2023079874W WO2023169367A1 WO 2023169367 A1 WO2023169367 A1 WO 2023169367A1 CN 2023079874 W CN2023079874 W CN 2023079874W WO 2023169367 A1 WO2023169367 A1 WO 2023169367A1
Authority
WO
WIPO (PCT)
Prior art keywords
electronic device
spatial
target
audio data
audio
Prior art date
Application number
PCT/CN2023/079874
Other languages
English (en)
Chinese (zh)
Inventor
文梁宇
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2023169367A1 publication Critical patent/WO2023169367A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/162Interface to dedicated audio devices, e.g. audio drivers, interface to CODECs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Definitions

  • This application belongs to the field of electronic technology, and specifically relates to an audio playback method and electronic equipment.
  • the purpose of the embodiments of the present application is to provide an audio playback and electronic device that can solve the problem of how to avoid the user's feeling of disorientation when listening to multiple different audio data.
  • embodiments of the present application provide an audio playback method, applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playback device,
  • the audio playback method includes:
  • spatial mixing processing is performed on the audio data played by the first electronic device and the audio data played by each of the second electronic devices to obtain the first target audio data;
  • embodiments of the present application provide an audio playback device, applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playback device,
  • the audio playback device includes:
  • Acquisition module used to obtain control instructions for space status information
  • a determination module configured to determine the spatial status information of the first electronic device and each of the second electronic devices according to the control instruction
  • a processing module configured to perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each of the second electronic devices according to each of the spatial state information, and obtain the first target audio data.
  • a sending module configured to send the first target audio data to the audio playing device for playing.
  • inventions of the present application provide an electronic device.
  • the electronic device includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor.
  • the program or instructions are When executed by the processor, the steps of the audio playback method described in the first aspect are implemented.
  • embodiments of the present application provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the audio playback method as described in the first aspect is implemented. A step of.
  • inventions of the present application provide a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the first aspect.
  • the audio playback method is used to run programs or instructions to implement the first aspect.
  • a control instruction for spatial status information is obtained; according to the control instruction, the spatial status information of the first electronic device and each second electronic device is determined; and according to each spatial status information, the video played by the first electronic device is
  • the audio data is spatially mixed with the audio data played by each second electronic device to obtain the first target audio data; and the first target audio data is sent to the audio playback device for playback.
  • Figure 1 is a first flow diagram of an audio playback method provided by an embodiment of the present application
  • Figure 2 is a schematic diagram of the connection relationship between the first electronic device, the second electronic device and the audio playback device provided by an embodiment of the present application;
  • Figure 3 is an interface diagram for setting the spatial audio status in the audio playback method provided by an embodiment of the present application
  • Figure 4A is a schematic diagram of the first scenario of an audio playback method provided by an embodiment of the present application.
  • Figure 4B is a schematic diagram of the second scenario of the audio playback method provided by an embodiment of the present application.
  • Figure 5 is a schematic diagram of the third scenario of the audio playback method provided by an embodiment of the present application.
  • Figure 6 is a second flow diagram of an audio playback method provided by an embodiment of the present application.
  • Figure 7 is a schematic structural diagram of an audio playback device provided by an embodiment of the present application.
  • Figure 8 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of the hardware structure of an electronic device provided by an embodiment of the present application.
  • first, second, etc. in the description and claims of this application are used to distinguish similar objects and are not used to describe a specific order or sequence. It is to be understood that the figures so used are interchangeable under appropriate circumstances so that the embodiments of the present application can be practiced in orders other than those illustrated or described herein, and that "first,” “second,” etc. are distinguished Objects are usually of one type, and the number of objects is not limited. For example, the first object can be one or multiple.
  • “and/or” in the description and claims indicates at least one of the connected objects, and the character “/" generally indicates that the related objects are in an "or” relationship.
  • FIG. 1 is a schematic flowchart of a first audio playback method provided by an embodiment of the present application.
  • the audio playing method is applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playing device.
  • the first electronic device may be an electronic device with an audio data processing function and an audio playback function, For example, computers, mobile phones, tablets, etc.
  • the second electronic device may be an electronic device with an audio data playback function, such as a computer, a mobile phone, a tablet, etc.
  • the at least one second electronic device may be one second electronic device or multiple second electronic devices.
  • the audio playback device can be headphones or other audio electronic devices.
  • FIG. 2 is a schematic diagram of the connection relationship between a first electronic device, a second electronic device and an audio playback device according to an embodiment of the present application.
  • a first electronic device 201 is connected to a second electronic device 202
  • the first electronic device 201 is connected to an earphone 203 .
  • the user wears the earphone 203 and listens to the audio data played by the first electronic device 201 and the audio data played by the second electronic device 202 at the same time.
  • the user faces the second electronic device 202 and can also watch the video screen played by the second electronic device 202 .
  • Step 102 Obtain control instructions for space status information.
  • the spatial status information may be spatial position information, spatial position information and audio parameters, or spatial audio status.
  • Audio parameters can be volume information, timbre information, and other parameters that can affect the auditory effect.
  • the spatial position information may be the spatial position information of the virtual sound source corresponding to the first electronic device in the preset sphere, or it may be the spatial position information of the virtual sound source corresponding to any second electronic device in the preset sphere.
  • the volume information may be the volume of the first electronic device or the volume of any second electronic device.
  • the control instruction can be a setting instruction, a modification instruction, a detection instruction, etc.
  • volume information can be adjusted independently or simultaneously with the spatial position information.
  • the volume information can also be adjusted to the volume information corresponding to the adjusted spatial position information after adjusting the spatial position information.
  • the spatial status information includes spatial position information and audio parameters.
  • Obtaining the control instruction of the spatial status information includes: receiving a first control instruction for the spatial position information; and/or receiving a second control instruction for the audio parameters.
  • the first control instruction may be a setting instruction or a modification instruction for the spatial location information of the first electronic device, or may be a setting instruction or modification instruction for the spatial location information of at least one second electronic device.
  • receiving the first control instruction for the spatial location information includes: in the first electronic device and at least one second electronic device, for any one electronic device, on the user interaction interface, Determine the position information of the virtual sound source corresponding to the electronic device in the preset sphere as the spatial position information of the electronic device; receive the position adjustment instruction of the virtual sound source; the position adjustment instruction is used to adjust the position of the virtual sound source in the preset sphere information.
  • FIG. 3 is an interface diagram for setting the spatial audio status in the audio playback method provided by an embodiment of the present application.
  • the spatial audio state setting interface diagram shows spatial audio state one and spatial audio state two.
  • the spatial position information of the second electronic device 1, the second electronic device 2, ... the second electronic device n can be referred to as shown in Figure 3.
  • the setting interface diagram of the spatial audio state also shows the parameter setting interface of the volume of the first electronic device, the volume of the second electronic device 1, the second electronic device 2,...the volume of the second electronic device n in the spatial audio state, and Parameter setting interface for the volume of the first electronic device, the volume of the second electronic device 1, the second electronic device 2,...the volume of the second electronic device n in spatial audio state 2.
  • the circle in Figure 3 represents a sphere.
  • spatial audio state 1 the position indicated by the arrow corresponding to the first electronic device is the spatial position of the first electronic device on the sphere.
  • each second electronic device The positions indicated by the corresponding arrows are the spatial positions of each second electronic device on the sphere.
  • Spatial audio state two is similar to spatial audio state one and will not be described again here.
  • the position adjustment instruction can be to add a new virtual sound source in the preset sphere and set the position information of the virtual sound source, or it can be to adjust a virtual sound source in the preset sphere from the preset initial position information to
  • the target position information that meets the user's needs can also be to adjust a virtual sound source in a preset sphere from the position information that meets the user's old hearing needs to the position information that meets the user's new hearing needs.
  • the second control instruction may be a setting instruction or modification instruction for the audio parameters of the first electronic device, or may be a setting instruction or modification instruction for the audio parameters of at least one second electronic device.
  • Audio parameters can be volume information, timbre information, and other parameters that can affect the auditory effect.
  • volume information takes the audio parameter as volume information as an example to illustrate:
  • the user can pre-set the volume information in different scenarios. Specifically, multiple volume states and the volume information of the first electronic device and the volume information of each second electronic device in each volume state can be configured.
  • the second control instruction may also be a switching instruction for the volume state.
  • the audio data played by the first electronic device facing directly in front of the user has a larger volume
  • the volume of the audio data played by the second electronic device on both sides of the user is smaller.
  • the audio parameters include volume information
  • receiving the second control instruction for the audio parameters includes: receiving, for any one electronic device among the first electronic device and at least one second electronic device, on the user interaction interface. Volume adjustment instructions for electronic devices.
  • the user interaction interface can be referred to Figure 3.
  • the volume information of each electronic device in the first electronic device and at least one second electronic device can be set separately, and the volume information can also be increased or decreased.
  • the first electronic device is a primary device located directly in front of the user; each second electronic device is a consonant device located not directly in front of the user;
  • the control instructions for obtaining the spatial state information include: obtaining the first primary and auxiliary devices Switching instruction; the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's direct front, and the first electronic device The spatial status information of the device corresponds to the non-direct front of the user; or, at least one second electronic device includes a target second electronic device; the target second electronic device is a primary device, and the first electronic device is a consonant device; obtaining the spatial status information
  • the control instructions include: obtaining a second main-auxiliary switching instruction; the second switching instruction is used to change the spatial status information of the first electronic device and the target second electronic device, so that the spatial status information of the first electronic device corresponds to what is directly in front of the user , the spatial state information of the
  • the main sound device is located directly in front of the user, and the user's attention is mainly focused on the audio data played by the main sound device; the consonant device is located not directly in front of the user, which can be understood as the user's attention to the audio data played by the consonant device.
  • the priority is after the audio data played by the main device.
  • the computer located directly in front of the user is playing an online class.
  • the computer is the main sound device.
  • the mobile phone located to the left of the user is playing a shopping live broadcast.
  • the mobile phone is a consonant device.
  • the user's main focus is on the online class. Attention to audio data of live broadcasts is lower than that of audio data of online classes.
  • At least one second electronic device includes the target second electronic device
  • the first electronic device is the main sound device and the target second electronic device is the consonant device.
  • the second electronic device is switched to the main sound device and the first electronic device is switched to the consonant device.
  • the first electronic device is a consonant device
  • the target second electronic device is a main sound device.
  • the first electronic device is switched to the main sound device
  • the second electronic device is switched to a consonant device.
  • the first electronic device is the primary device and is located directly in front of the user
  • the target second electronic device is the consonant device and is located on the left side of the user.
  • the first primary and secondary switching instructions are changed. Spatial state information of the first electronic device and the target second electronic device.
  • the target second electronic device serves as the main device and is located directly in front of the user
  • the first electronic device serves as the consonant device and is located on the right side of the user.
  • At least one second electronic device includes at least two second electronic devices
  • at least one second electronic device including second electronic device 1 and second electronic device 2 taking at least one second electronic device including second electronic device 1 and second electronic device 2 as an example, the tonic device and the consonant device in this case will be described. Switch between:
  • the first electronic device is the main sound device, and the second electronic device 1 and the second electronic device 2 are both consonant devices.
  • the selected second electronic device 1 is switched to the main sound device, and the second electronic device 1 is switched to the main sound device.
  • the first electronic device switches to a consonant device;
  • the first electronic device is the main sound device, and the second electronic device 1 and the second electronic device 2 are both consonant devices.
  • the selected second electronic device 2 is switched to the main sound device, and the second electronic device 2 is switched to the main sound device.
  • the first electronic device switches to a consonant device;
  • the second electronic device 2 is the main sound device, and the first electronic device and the second electronic device 1 are both consonant devices. Through the third main and auxiliary switching command, the selected first electronic device is switched to the main sound device, and the third electronic device is switched to the main sound device. The second electronic device 2 switches to a consonant device;
  • the second electronic device 2 is the main sound device, and the first electronic device and the second electronic device 1 are both consonant devices. Through the third main and auxiliary switching command, the selected second electronic device 1 is switched to the main sound device, and the second electronic device 1 is switched to the main sound device. The second electronic device 2 switches to a consonant device;
  • the second electronic device 1 is the main sound device, and the first electronic device and the second electronic device 2 are both consonant devices.
  • the selected first electronic device is switched to the main sound device, and the third electronic device is switched to the main sound device.
  • the second electronic device 1 is switched to a consonant device;
  • the second electronic device 1 is the main sound device, and the first electronic device and the second electronic device 2 are both consonant devices. Through the third main and auxiliary switching command, the selected second electronic device 2 is switched to the main sound device, and the second electronic device 2 is switched to the main sound device. The second electronic device 1 is switched to a consonant device.
  • the first primary-auxiliary switching instruction at least includes the following implementation methods:
  • the preset sphere is rotated so that the position information of the virtual sound source corresponding to the electronic device selected as the main sound device in the preset sphere reaches the preset position corresponding to the main sound device after rotation.
  • the position information of the virtual sound source corresponding to the electronic device selected as the main sound device is reached after the rotation.
  • the position information of the virtual sound source corresponding to the electronic device selected as the main sound device is The virtual sound source moves to other locations in the sphere after rotation.
  • each electronic device corresponds to a virtual sound source.
  • Drag the virtual sound source in the preset sphere to change the position information of the virtual sound source in the preset sphere.
  • the user can drag the virtual sound source corresponding to the first electronic device so that the virtual sound source leaves the preset position corresponding to the main sound device; the user can also drag the virtual sound source corresponding to the second electronic device to the preset position.
  • the user interaction interface displays multiple preset spatial audio states.
  • spatial audio state 1 the first electronic device is the main sound device and the target second electronic device is the consonant device; in spatial audio state 2, the target second electronic device is It is the main sound device and the first electronic device is the consonant sound device.
  • spatial audio state 2 it will switch to spatial audio state 2 according to user operations.
  • a user initially looks at the computer directly in front of them, then turns to look at the phone to the left.
  • the second main-auxiliary switching instruction and the third main-auxiliary switching instruction are similar to the first main-auxiliary switching instruction, and will not be described again here.
  • the spatial state information of the second electronic device corresponds to what is directly in front of the user. It may be that the position information of the virtual sound source corresponding to the second electronic device in the preset sphere is located at the preset position of the main sound device in the preset sphere, or It may be that the pupil gaze point is located on the second electronic device.
  • the audio data played by the second electronic device sounds like it is coming from directly in front.
  • the spatial state information of the first electronic device corresponds to the position not directly in front of the user, which may be that the position information of the virtual sound source corresponding to the first electronic device in the preset sphere is not at the preset position of the main sound device in the preset sphere, It may also be that the pupil gaze point is located outside the first electronic device.
  • the audio data played by the first electronic device sounds like it is coming from other directions than directly in front.
  • the spatial state information of the first electronic device corresponds to the direct front of the user
  • the spatial state information of the target second electronic device corresponds to the non-direct front of the user.
  • FIG. 4A is a schematic diagram of the first scene of the audio playback method provided by an embodiment of the present application
  • FIG. 4B is a schematic diagram of the second scene of the audio playback method provided by an embodiment of the present application.
  • the first electronic device 401 is located directly in front of the user, and the second electronic device 402 is located on the right hand side of the user.
  • the first electronic device 401 is located on the left side of the user, and the second electronic device 402 is located directly in front of the user.
  • the user is looking at the first electronic device 401 directly in front, while listening to the audio data played by the first electronic device 401 and the audio data played by the second electronic device on the right. Then, in the second scene, the user turns to look at the second electronic device 402 while listening to the audio data played by the first electronic device 401 and the audio data played by the second electronic device on the right.
  • the actual positions of the first electronic device 401 and the second electronic device 402 do not change. What changes is the position of the user's pupil gaze point. Then when the second electronic device 402 is located directly in front of the user, the first electronic device 401 Located on the user's left.
  • control instruction for obtaining the spatial state information includes: obtaining the first detection result of the first electronic device for the pupil gaze point and the second detection result of each second electronic device for the pupil gaze point.
  • the first electronic device and each second electronic device may be provided with a sensor having a detection capability of the pupil gaze point. After detecting the pupil gaze point through the sensor, each second electronic device may send the obtained second detection result to the first electronic device.
  • the pupil fixation point may exist in the first electronic device or in a second electronic device.
  • the first detection result may include whether the pupil gaze point is detected, and may also include position information of the pupil gaze point on the first electronic device.
  • the second detection result may include whether the pupil gaze point is detected, and may also include position information of the pupil gaze point on the second electronic device.
  • Step 104 Determine the spatial status information of the first electronic device and each second electronic device according to the control instruction.
  • determining the spatial status information of the first electronic device and each second electronic device according to the control instruction includes: determining the target position watched by the pupil according to the first detection result and at least one second detection result; The location determines the spatial status information of the first electronic device and each second electronic device.
  • the first detection result can indicate that the pupil gaze point and the position information of the pupil gaze point are detected on the first electronic device
  • the second detection result can indicate that the pupil gaze point is not detected on the second electronic device
  • the second detection result can indicate that the pupil gaze point and the position information of the pupil gaze point are detected on the second electronic device
  • the first detection result can indicate that the pupil gaze point is not detected on the first electronic device
  • the spatial state information of the first electronic device and each second electronic device can be determined.
  • determining the spatial state information of the first electronic device and each second electronic device according to the target position includes: determining the target gazed by the pupil in the first electronic device and at least one second electronic device according to the target position.
  • Electronic equipment and at least one non-target electronic equipment that is not being watched determine the spatial status information of the target electronic equipment based on the target position; determine each non-target electronic equipment based on a preset spatial status information set and the spatial status information of the target electronic equipment Space status information of the device.
  • a first electronic device is connected to two second electronic devices, and the two second electronic devices are second electronic device 1 and second electronic device 2 respectively. If the target position is on the first electronic device, the first electronic device is determined as the target electronic device that is gazed at by the pupil, and the second electronic device 1 and the second electronic device 2 are determined as non-target electronic devices that are not gazed upon. If the target position is on the second electronic device 2, the second electronic device 2 is determined as the target electronic device that is gazed by the pupil, and the second electronic device 1 and the first electronic device are determined as non-target electronic devices that are not gazed upon.
  • Determining the spatial state information of the target electronic device according to the target position can be understood as making the position information of the pupil gaze point correspond to the position corresponding to the user's direct front in the preset sphere.
  • the preset spatial state information set may correspond to the preset spatial audio state.
  • the space state information of each non-target electronic device is determined, which can be understood as the first electronic device and each second electronic device in the preset sphere.
  • the relative position is fixed, so the spatial state information of each non-target electronic device can be determined when the position information of the pupil gaze point corresponds to the position corresponding to the user's direct front in the preset sphere.
  • Figure 5 is a schematic diagram of a third scenario of an audio playback method provided by an embodiment of the present application.
  • the sensor detects that the user's pupil gaze point 503 is located on the first electronic device 501, and the user focuses on the first electronic device 501.
  • the first electronic device 501 is located directly in front of the user, and the first electronic device
  • the spatial position information of 501 in the preset sphere corresponds to the user's front
  • the volume of the first electronic device 501 is a preset value.
  • the second electronic device 502 is located on the right side of the user.
  • the spatial position information of the second electronic device 502 in the preset sphere corresponds to the right side of the user.
  • the volume of the second electronic device 502 is slightly smaller than the volume of the first electronic device 501 .
  • the audio data played by the first electronic device 501 is mainly used, and the audio data played by the second electronic device 502 is used as a supplement.
  • the user's pupil gaze point 503 moves away from the edge of the screen of the first electronic device.
  • the volume of the first electronic device 501 gradually becomes smaller as the pupil gaze point 503 moves away from the screen.
  • the spatial position information of the first electronic device 501 in the preset sphere gradually moves to the left as the pupil gaze point 503 moves, and the direction tends to the left hand of the user.
  • the user's pupil gaze point 503 begins to enter the edge of the screen of the second electronic device 502 .
  • the volume of the second electronic device 502 gradually becomes larger as the pupil gaze point 503 enters the screen.
  • the spatial position information of the second electronic device in the preset sphere gradually moves to the left as the pupil gaze point 503 moves, and the direction tends to be directly in front of the user.
  • the pupil fixation point 503 is located on the second electronic device 502 .
  • the volume of the second electronic device 502 gradually increases until it reaches a preset value, and the spatial position information of the second electronic device 502 in the preset sphere corresponds to what is directly in front of the user.
  • the volume of the first electronic device 501 gradually decreases, and the spatial position information of the first electronic device in the preset sphere corresponds to the user's left hand side.
  • the audio data played by the second electronic device 502 is mainly the audio data played by the first electronic device 501 .
  • determining the spatial state information of the first electronic device and each second electronic device according to the target position includes: determining the target gazed by the pupil in the first electronic device and at least one second electronic device according to the target position. electronic device and at least one non-target electronic device that is not being watched; according to the target electronic device, a corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes The spatial state information of the target electronic device and the spatial state information of each non-target electronic device.
  • the multiple preset alternative spatial state information combinations may be spatial state information combinations corresponding to multiple preset spatial audio states.
  • the alternative spatial state information combination may include the spatial state information combination 1 corresponding to the first electronic device directly in front of the user and the spatial state information combination 2 corresponding to the second electronic device directly in front of the user. Then, when the target electronic device is the first In the case of an electronic device, the space state information combination 1 can be determined based on the first electronic device, and each space state information corresponding to the space state combination 1 is determined as the space state information of the first electronic device and each second electronic device.
  • Step 106 According to each spatial state information, perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device to obtain the first target audio number. according to.
  • the spatial mixing process may be, in the presence of at least two sound sources at different locations, mixing audio data from each sound source, so that when the mixed audio data is played through the audio playback device, each audio data It sounds like it's coming from the corresponding sound source, not from the same direction.
  • the mixing operation may be to produce multiple types of audio data into one audio file. When the audio file is played, each type of audio data still sounds independent. And each comes from the corresponding sound source, and multiple sound sources are located in different directions of the user.
  • the spatial position information of the first electronic device corresponds to the front of the user
  • the value of the volume information of the first electronic device is x
  • the spatial position information of the second electronic device corresponds to the left side of the user
  • the value of the volume information of the second electronic device is x.
  • the value is y, x>y.
  • Each sound source can be located in different directions, and various audio parameters such as the volume of each audio data can also be obtained. It can be flexibly set separately, so that only one audio playback device can play the audio data after spatial mixing processing, and multiple audio playback devices located in different directions of the user can be used as sound sources to simultaneously play the corresponding audio data.
  • Step 108 Send the first target audio data to the audio playback device for playback.
  • the first electronic device may send the first target audio data to the earphone, thereby playing the first target audio data through the earphone. For example, if the spatial state information of the first electronic device corresponds to the front of the user, and the spatial state information of the second electronic device corresponds to the left side of the user, then the user can listen to the first target audio data through the earphones.
  • the first audio data comes from the front and has a louder volume, while the second audio data can be heard from the left and has a lower volume.
  • the audio playback method also includes: in the case where the audio data played by the first electronic device is replaced, according to each spatial state information, the replaced audio data played by the first electronic device and each second electronic device are replaced.
  • the audio data played by the device is subjected to spatial mixing processing to obtain the second target audio data and sent to the audio playback device for playback; or, in the case where the audio data played by each second electronic device is replaced, according to each For spatial status information, perform spatial mixing processing on the audio data played by the first electronic device and the replaced audio data played by each second electronic device, obtain the third target audio data, and send it to the audio playback device for playback.
  • the control instruction for obtaining spatial state information Order determines the spatial state information of the first electronic device and each second electronic device; according to each spatial state information, determine the audio data played by the first electronic device and the audio data played by each second electronic device Perform spatial mixing processing to obtain the first target audio data; send the first target audio data to the audio playback device for playback.
  • spatial mixing processing can be performed according to the spatial state information corresponding to the first electronic device and at least one second electronic device, thereby preventing the user from listening to multiple electronic devices located in different locations playing different music at the same time.
  • the audio data has a sense of disorientation.
  • FIG. 6 is a second flow diagram of an audio playback method provided by an embodiment of the present application.
  • the audio playback device in this embodiment may be headphones.
  • step 602 the first electronic device and the second electronic device are connected to each other, and the earphone is connected to the first electronic device.
  • Step 604 The second electronic device transmits audio data to the first electronic device.
  • Step 606 The user sets spatial location information according to the actual location status of the device or preferences.
  • Step 608 The user sets the volume for different scenes, and the user can configure multiple volume states.
  • Step 610 Perform spatial mixing processing on the audio data played by the first electronic device and the second electronic device, and output the audio data to the headphones.
  • step 610 After step 610, at least one of step 612, step 618, and step 620 may be performed.
  • Step 612 Detect the pupil fixation point.
  • Step 614 Whether the target electronic device is switched.
  • step 616 If yes, return to step 608; if not, execute step 616.
  • Step 616 Maintain the original output state.
  • Step 618 Determine whether the audio data has been replaced.
  • step 616 If yes, return to step 606; if not, execute step 616.
  • Step 620 Determine whether the volume status is switched.
  • step 616 If yes, return to step 608; if not, execute step 616.
  • the audio playback method provided by the embodiment shown in FIG. 6 can implement various processes implemented by the foregoing audio playback method embodiment. To avoid repetition, details will not be described here.
  • the execution subject may be an audio playback device, or a control module in the audio playback device for executing the audio playback method.
  • the method of performing audio playback by the audio playback device is taken as an example to describe the audio playback device provided by the embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an audio playback device provided by an embodiment of the present application.
  • the audio playback device is applied to a first electronic device.
  • the first electronic device is connected to at least one second electronic device.
  • the first electronic device is connected to the audio playback device.
  • the audio playback device includes:
  • Acquisition module 701 used to obtain control instructions for space status information
  • Determining module 702 configured to determine the space status information of the first electronic device and each second electronic device according to the control instruction
  • the processing module 703 is configured to perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device according to each spatial state information, and obtain the first target audio data;
  • the sending module 704 is used to send the first target audio data to the audio playback device for playback.
  • the first electronic device is a main sound device; each second electronic device is a consonant device; the acquisition module is specifically used for:
  • the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's normal Ahead, the spatial state information of the first electronic device corresponds to the non-direct front of the user;
  • At least one second electronic device includes a target second electronic device; the target second electronic device is a main sound device, and the first electronic device is a consonant device; and the acquisition module is specifically used for:
  • the second switching instruction is used to change the spatial state information of the first electronic device and the target second electronic device, so that the spatial state information of the first electronic device corresponds to the user's direct front, and the target second electronic device
  • the device's spatial status information corresponds to what is not directly in front of the user.
  • the acquisition module 701 is specifically used for:
  • a first detection result of the first electronic device for the pupil gaze point and a second detection result of each second electronic device for the pupil gaze point are obtained.
  • the determining module 702 includes:
  • a first determination unit configured to determine the position of the target gazed by the pupil based on the first detection result and at least one second detection result
  • a second determination unit configured to determine the first electronic device and each second electronic device according to the target position. Space status information of the device.
  • the second determination unit is specifically used for:
  • the target location determine the spatial status information of the target electronic device
  • the space state information of each non-target electronic device is determined.
  • the second determination unit is specifically used for:
  • the corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes the space state information of the target electronic device and each non-target electronic device space status information.
  • the audio playback device also includes:
  • a mixing module configured to combine the replaced audio data played by the first electronic device and the audio played by each second electronic device according to each spatial state information when the audio data played by the first electronic device is replaced.
  • the data is subjected to spatial mixing processing to obtain the second target audio data and sent to the audio playback device for playback; or, in the case where the audio data played by each second electronic device is replaced, according to each spatial status information, Perform spatial mixing processing on the audio data played by the first electronic device and the replaced audio data played by each second electronic device to obtain third target audio data, and send it to the audio playback device for playback.
  • the spatial state information includes spatial position information and audio parameters.
  • the acquisition module 701 includes:
  • a first receiving unit configured to receive a first control instruction for spatial position information
  • the second receiving unit is used to receive the second control instruction for the audio parameters.
  • the first receiving unit is specifically used for:
  • the position information of the virtual sound source corresponding to the electronic device in the preset sphere is determined as the spatial position information of the electronic device ;
  • the position adjustment instruction is used to adjust the virtual sound source in the preset Position information in the sphere.
  • the audio parameters include volume information; the second receiving unit is specifically used for:
  • a volume adjustment instruction for the electronic device is received on the user interaction interface.
  • the audio playback device obtaineds the control instruction of the spatial state information; determines the spatial state information of the first electronic device and each second electronic device according to the control instruction; and determines the first electronic device according to each spatial state information.
  • the audio data played by the electronic device and the audio data played by each second electronic device are spatially mixed to obtain the first target audio data; the first target audio data is sent to the audio playback device for playback.
  • spatial mixing processing can be performed according to the spatial state information corresponding to the first electronic device and at least one second electronic device, thereby preventing the user from listening to multiple electronic devices located in different locations playing different music at the same time.
  • the audio data has a sense of disorientation.
  • the audio playback device in the embodiment of the present application may be a device, or may be a component, integrated circuit, or chip in a terminal.
  • the device may be a mobile electronic device or a non-mobile electronic device.
  • the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a personal digital assistant (personal digital assistant).
  • UMPC ultra-mobile personal computer
  • PDA personal digital assistant
  • non-mobile electronic devices can be servers, network attached storage (Network Attached Storage, NAS), personal computers (personal computers, PC), televisions (television, TV), teller machines or self-service machines, etc., this application The examples are not specifically limited.
  • the audio playback device in the embodiment of the present application may be a device with an operating system.
  • the operating system can be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiments of this application.
  • the audio playback device provided by the embodiments of the present application can implement various processes implemented by the foregoing audio playback method embodiments. To avoid repetition, they will not be described again here.
  • this embodiment of the present application also provides an electronic device 800, including a processor 801, a memory 802, and a program or instructions stored on the memory 802 and executable on the processor 801.
  • the program Or when the instruction is executed by the processor 801, each process of the above audio playback method embodiment is implemented, and the same technical effect can be achieved. To avoid repetition, the details will not be described here.
  • the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
  • Figure 9 is a schematic diagram of the hardware structure of an electronic device that implements an embodiment of the present application.
  • the electronic device 900 includes but is not limited to: radio frequency unit 901, network module 902, audio output unit 903, input unit 904, sensor 905, display unit 906, user input unit 907, interface unit 908, memory 909, processor 910, etc. part.
  • the electronic device 900 may also include a power supply (such as a battery) that supplies power to various components.
  • the power supply may be logically connected to the processor 910 through a power management system, thereby managing charging, discharging, and function through the power management system. Consumption management and other functions.
  • the structure of the electronic device shown in Figure 9 does not constitute a limitation on the electronic device.
  • the electronic device may include more or less components than shown in the figure, or combine certain components, or arrange different components, which will not be described again here. .
  • the processor 910 is used to obtain control instructions for spatial state information
  • each spatial state information perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device to obtain the first target audio data;
  • a control instruction for spatial status information is obtained; according to the control instruction, the spatial status information of the first electronic device and each second electronic device is determined; and according to each spatial status information, the video played by the first electronic device is
  • the audio data is spatially mixed with the audio data played by each second electronic device to obtain the first target audio data; and the first target audio data is sent to the audio playback device for playback.
  • the first electronic device is a tonic device located directly in front of the user; each second electronic device is a consonant device located not directly in front of the user; the processor 910 is configured to:
  • Control instructions for obtaining space status information including:
  • the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's normal Ahead, the spatial state information of the first electronic device corresponds to the non-direct front of the user;
  • At least one second electronic device includes a target second electronic device; the target second electronic device is a main sound device, and the first electronic device is a consonant device; and a control instruction for obtaining spatial state information, include:
  • the second switching instruction is used to change the spatial state information of the first electronic device and the target second electronic device, so that the spatial state information of the first electronic device corresponds to the user's direct front, and the target second electronic device
  • the device's spatial status information corresponds to what is not directly in front of the user.
  • processor 910 is also used to:
  • Control instructions for obtaining space status information including:
  • a first detection result of the first electronic device for the pupil gaze point and a second detection result of each second electronic device for the pupil gaze point are obtained.
  • processor 910 is also used to:
  • determining the space status information of the first electronic device and each second electronic device includes:
  • spatial status information of the first electronic device and each second electronic device is determined.
  • processor 910 is also used to:
  • determine the spatial status information of the first electronic device and each second electronic device including:
  • the target location determine the spatial status information of the target electronic device
  • the space state information of each non-target electronic device is determined.
  • processor 910 is also used to:
  • determine the spatial status information of the first electronic device and each second electronic device including:
  • the corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes the space state information of the target electronic device and each non-target electronic device space status information.
  • processor 910 is also used to:
  • spatial mixing processing is performed on the replaced audio data played by the first electronic device and the audio data played by each second electronic device according to each spatial state information. , obtain the second target audio data and send it to the audio playback device for playback;
  • spatial mixing is performed on the audio data played by the first electronic device and the replaced audio data played by each second electronic device according to each spatial state information. Audio processing, obtaining the third target audio data, and sending it to the audio playback device for playback.
  • processor 910 is also used to:
  • Spatial status information includes spatial position information and audio parameters.
  • Control instructions for obtaining spatial status information include:
  • processor 910 is also used to:
  • Receiving the first control instruction for spatial location information includes:
  • the position information of the virtual sound source corresponding to the electronic device in the preset sphere is determined as the spatial position information of the electronic device ;
  • the position adjustment instruction is used to adjust the position information of the virtual sound source in the preset sphere.
  • the audio parameters include volume information; processor 910 is also used to:
  • Receive second control instructions for audio parameters including:
  • a volume adjustment instruction for the electronic device is received on the user interaction interface.
  • the main device and the consonant device can be flexibly switched between the first electronic device and at least one second electronic device; according to the detected pupil Gaze point can freely and flexibly control spatial position information and spatial volume information, providing users with the most suitable audio playback effect during the process of users naturally shifting their gaze; by receiving position adjustment instructions and targeting of virtual sound sources on the user interaction interface electronic device volume Adjustment instructions can flexibly set the spatial status information of each electronic device and enrich the auditory effect.
  • the input unit 904 may include a graphics processor (Graphics Processing Unit, GPU) 9041 and a microphone 9042.
  • the graphics processor 9041 is responsible for the image capture device (GPU) in the video capture mode or the image capture mode. Process the image data of still pictures or videos obtained by cameras (such as cameras).
  • the display unit 906 may include a display panel 9061, which may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the user input unit 907 includes a touch panel 9071 and other input devices 9072. Touch panel 9071, also known as touch screen.
  • the touch panel 9071 may include two parts: a touch detection device and a touch controller.
  • Other input devices 9072 may include but are not limited to physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which will not be described again here.
  • Memory 909 may be used to store software programs as well as various data, including but not limited to application programs and operating systems.
  • the processor 910 can integrate an application processor and a modem processor, where the application processor mainly processes the operating system, user interface, application programs, etc., and the modem processor mainly processes wireless communications. It can be understood that the above modem processor may not be integrated into the processor 910.
  • Embodiments of the present application also provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the program or instructions are executed by a processor, each process of the audio playback method embodiment is implemented, and the same can be achieved. The technical effects will not be repeated here to avoid repetition.
  • the processor is the processor in the electronic device described in the above embodiment.
  • the readable storage media includes computer-readable storage media, such as computer read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks, etc.
  • An embodiment of the present application further provides a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the above audio playback method embodiment. Each process can achieve the same technical effect. To avoid duplication, it will not be described again here.
  • chips mentioned in the embodiments of this application may also be called system-on-chip, system-on-a-chip, system-on-a-chip or system-on-chip, etc.
  • the methods of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. implementation.
  • the technical solution of the present application can be embodied in the form of a computer software product that is essentially or contributes to the existing technology.
  • the computer software product is stored in a storage medium (such as ROM/RAM, disk , optical disk), including several instructions to cause a terminal (which can be a mobile phone, computer, server, or network device, etc.) to execute the methods described in various embodiments of this application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Stereophonic System (AREA)

Abstract

La présente demande, qui appartient au domaine technique de l'électronique, concerne un procédé de lecture audio et un dispositif électronique. Le procédé de lecture audio comprend : l'acquisition d'une instruction de commande pour des informations d'état spatiales ; selon l'instruction de commande, la détermination d'informations d'état spatiales d'un premier dispositif électronique et de chaque second dispositif électronique ; selon chaque élément d'informations d'état spatiales, la réalisation d'un traitement de mélange sonore spatial sur des données audio lues par le premier dispositif électronique et sur des données audio lues par chaque second dispositif électronique de façon à obtenir des premières données audio cibles ; et l'envoi des premières données audio cibles à un dispositif de lecture audio aux fins de lecture.
PCT/CN2023/079874 2022-03-07 2023-03-06 Procédé de lecture audio et dispositif électronique WO2023169367A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210225832.8 2022-03-07
CN202210225832.8A CN114650496A (zh) 2022-03-07 2022-03-07 音频播放方法和电子设备

Publications (1)

Publication Number Publication Date
WO2023169367A1 true WO2023169367A1 (fr) 2023-09-14

Family

ID=81993315

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/079874 WO2023169367A1 (fr) 2022-03-07 2023-03-06 Procédé de lecture audio et dispositif électronique

Country Status (2)

Country Link
CN (1) CN114650496A (fr)
WO (1) WO2023169367A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080008342A1 (en) * 2006-07-07 2008-01-10 Harris Corporation Method and apparatus for creating a multi-dimensional communication space for use in a binaural audio system
US20140285312A1 (en) * 2013-03-19 2014-09-25 Nokia Corporation Audio Mixing Based Upon Playing Device Location
CN112581932A (zh) * 2020-11-26 2021-03-30 交通运输部南海航海保障中心广州通信中心 一种基于dsp的有线无线混音系统
US20210176585A1 (en) * 2019-12-04 2021-06-10 Roland Corporation Headphone
CN113823250A (zh) * 2021-11-25 2021-12-21 广州酷狗计算机科技有限公司 音频播放方法、装置、终端及存储介质
CN113890932A (zh) * 2020-07-02 2022-01-04 华为技术有限公司 一种音频控制方法、系统及电子设备
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080008342A1 (en) * 2006-07-07 2008-01-10 Harris Corporation Method and apparatus for creating a multi-dimensional communication space for use in a binaural audio system
US20140285312A1 (en) * 2013-03-19 2014-09-25 Nokia Corporation Audio Mixing Based Upon Playing Device Location
US20210176585A1 (en) * 2019-12-04 2021-06-10 Roland Corporation Headphone
CN113890932A (zh) * 2020-07-02 2022-01-04 华为技术有限公司 一种音频控制方法、系统及电子设备
CN112581932A (zh) * 2020-11-26 2021-03-30 交通运输部南海航海保障中心广州通信中心 一种基于dsp的有线无线混音系统
CN113823250A (zh) * 2021-11-25 2021-12-21 广州酷狗计算机科技有限公司 音频播放方法、装置、终端及存储介质
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Also Published As

Publication number Publication date
CN114650496A (zh) 2022-06-21

Similar Documents

Publication Publication Date Title
RU2666966C2 (ru) Способ и прибор управления для воспроизведения аудио
CN109660817B (zh) 视频直播方法、装置及系统
WO2017173793A1 (fr) Procédé et appareil pour une projection sur écran d'une vidéo
CN112237012B (zh) 用于控制多视点全方位内容中的音频的装置及方法
US20180349088A1 (en) Apparatus and Method for Controlling Audio Mixing in Virtual Reality Environments
WO2017193540A1 (fr) Procédé, dispositif et système pour lire un commentaire de recouvrement
JP2023529868A (ja) 共有方法、装置及び電子機器
CN111866539A (zh) 直播界面的切换方法、装置、终端及存储介质
US20220070599A1 (en) Method for projecting screen, display device, screen projection terminal, and storage medium
WO2023011473A1 (fr) Procédé et appareil de commutation de canal, et dispositif électronique
WO2022156662A1 (fr) Procédé et appareil de commutation de mode de lecture audio, dispositif électronique et support de stockage
WO2023169367A1 (fr) Procédé de lecture audio et dispositif électronique
CN107147957A (zh) 视频播放方法和装置
CN112291615A (zh) 音频输出方法、音频输出装置
CN112673651A (zh) 多视点多用户音频用户体验
KR20150136314A (ko) 디스플레이 장치, 사용자 단말 장치, 서버 및 그 제어 방법
WO2023246166A1 (fr) Procédé et appareil de réglage de déroulement de vidéo, dispositif informatique et support d'enregistrement
WO2023030292A1 (fr) Procédé et appareil de lecture de fichier multimédia
US20230024761A1 (en) Method for playing videos and electronic device
WO2022247784A1 (fr) Procédé de commande, appareil, dispositif électronique et support de stockage lisible
US20230046440A1 (en) Video playback method and device
CA2985731A1 (fr) Systeme multi-interface, multi-support audio/video active vocalement/manuellement et integre
US10110943B2 (en) Flexible output of streaming media
CN113992786A (zh) 音频播放方法及装置
KR20230120668A (ko) 영상통화 방법 및 장치

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23765945

Country of ref document: EP

Kind code of ref document: A1