WO2023169367A1 - 音频播放方法和电子设备 - Google Patents

音频播放方法和电子设备 Download PDF

Info

Publication number
WO2023169367A1
WO2023169367A1 PCT/CN2023/079874 CN2023079874W WO2023169367A1 WO 2023169367 A1 WO2023169367 A1 WO 2023169367A1 CN 2023079874 W CN2023079874 W CN 2023079874W WO 2023169367 A1 WO2023169367 A1 WO 2023169367A1
Authority
WO
WIPO (PCT)
Prior art keywords
electronic device
spatial
target
audio data
audio
Prior art date
Application number
PCT/CN2023/079874
Other languages
English (en)
French (fr)
Inventor
文梁宇
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2023169367A1 publication Critical patent/WO2023169367A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/162Interface to dedicated audio devices, e.g. audio drivers, interface to CODECs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Definitions

  • This application belongs to the field of electronic technology, and specifically relates to an audio playback method and electronic equipment.
  • the purpose of the embodiments of the present application is to provide an audio playback and electronic device that can solve the problem of how to avoid the user's feeling of disorientation when listening to multiple different audio data.
  • embodiments of the present application provide an audio playback method, applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playback device,
  • the audio playback method includes:
  • spatial mixing processing is performed on the audio data played by the first electronic device and the audio data played by each of the second electronic devices to obtain the first target audio data;
  • embodiments of the present application provide an audio playback device, applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playback device,
  • the audio playback device includes:
  • Acquisition module used to obtain control instructions for space status information
  • a determination module configured to determine the spatial status information of the first electronic device and each of the second electronic devices according to the control instruction
  • a processing module configured to perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each of the second electronic devices according to each of the spatial state information, and obtain the first target audio data.
  • a sending module configured to send the first target audio data to the audio playing device for playing.
  • inventions of the present application provide an electronic device.
  • the electronic device includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor.
  • the program or instructions are When executed by the processor, the steps of the audio playback method described in the first aspect are implemented.
  • embodiments of the present application provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the audio playback method as described in the first aspect is implemented. A step of.
  • inventions of the present application provide a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the first aspect.
  • the audio playback method is used to run programs or instructions to implement the first aspect.
  • a control instruction for spatial status information is obtained; according to the control instruction, the spatial status information of the first electronic device and each second electronic device is determined; and according to each spatial status information, the video played by the first electronic device is
  • the audio data is spatially mixed with the audio data played by each second electronic device to obtain the first target audio data; and the first target audio data is sent to the audio playback device for playback.
  • Figure 1 is a first flow diagram of an audio playback method provided by an embodiment of the present application
  • Figure 2 is a schematic diagram of the connection relationship between the first electronic device, the second electronic device and the audio playback device provided by an embodiment of the present application;
  • Figure 3 is an interface diagram for setting the spatial audio status in the audio playback method provided by an embodiment of the present application
  • Figure 4A is a schematic diagram of the first scenario of an audio playback method provided by an embodiment of the present application.
  • Figure 4B is a schematic diagram of the second scenario of the audio playback method provided by an embodiment of the present application.
  • Figure 5 is a schematic diagram of the third scenario of the audio playback method provided by an embodiment of the present application.
  • Figure 6 is a second flow diagram of an audio playback method provided by an embodiment of the present application.
  • Figure 7 is a schematic structural diagram of an audio playback device provided by an embodiment of the present application.
  • Figure 8 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of the hardware structure of an electronic device provided by an embodiment of the present application.
  • first, second, etc. in the description and claims of this application are used to distinguish similar objects and are not used to describe a specific order or sequence. It is to be understood that the figures so used are interchangeable under appropriate circumstances so that the embodiments of the present application can be practiced in orders other than those illustrated or described herein, and that "first,” “second,” etc. are distinguished Objects are usually of one type, and the number of objects is not limited. For example, the first object can be one or multiple.
  • “and/or” in the description and claims indicates at least one of the connected objects, and the character “/" generally indicates that the related objects are in an "or” relationship.
  • FIG. 1 is a schematic flowchart of a first audio playback method provided by an embodiment of the present application.
  • the audio playing method is applied to a first electronic device, the first electronic device is connected to at least one second electronic device, and the first electronic device is connected to an audio playing device.
  • the first electronic device may be an electronic device with an audio data processing function and an audio playback function, For example, computers, mobile phones, tablets, etc.
  • the second electronic device may be an electronic device with an audio data playback function, such as a computer, a mobile phone, a tablet, etc.
  • the at least one second electronic device may be one second electronic device or multiple second electronic devices.
  • the audio playback device can be headphones or other audio electronic devices.
  • FIG. 2 is a schematic diagram of the connection relationship between a first electronic device, a second electronic device and an audio playback device according to an embodiment of the present application.
  • a first electronic device 201 is connected to a second electronic device 202
  • the first electronic device 201 is connected to an earphone 203 .
  • the user wears the earphone 203 and listens to the audio data played by the first electronic device 201 and the audio data played by the second electronic device 202 at the same time.
  • the user faces the second electronic device 202 and can also watch the video screen played by the second electronic device 202 .
  • Step 102 Obtain control instructions for space status information.
  • the spatial status information may be spatial position information, spatial position information and audio parameters, or spatial audio status.
  • Audio parameters can be volume information, timbre information, and other parameters that can affect the auditory effect.
  • the spatial position information may be the spatial position information of the virtual sound source corresponding to the first electronic device in the preset sphere, or it may be the spatial position information of the virtual sound source corresponding to any second electronic device in the preset sphere.
  • the volume information may be the volume of the first electronic device or the volume of any second electronic device.
  • the control instruction can be a setting instruction, a modification instruction, a detection instruction, etc.
  • volume information can be adjusted independently or simultaneously with the spatial position information.
  • the volume information can also be adjusted to the volume information corresponding to the adjusted spatial position information after adjusting the spatial position information.
  • the spatial status information includes spatial position information and audio parameters.
  • Obtaining the control instruction of the spatial status information includes: receiving a first control instruction for the spatial position information; and/or receiving a second control instruction for the audio parameters.
  • the first control instruction may be a setting instruction or a modification instruction for the spatial location information of the first electronic device, or may be a setting instruction or modification instruction for the spatial location information of at least one second electronic device.
  • receiving the first control instruction for the spatial location information includes: in the first electronic device and at least one second electronic device, for any one electronic device, on the user interaction interface, Determine the position information of the virtual sound source corresponding to the electronic device in the preset sphere as the spatial position information of the electronic device; receive the position adjustment instruction of the virtual sound source; the position adjustment instruction is used to adjust the position of the virtual sound source in the preset sphere information.
  • FIG. 3 is an interface diagram for setting the spatial audio status in the audio playback method provided by an embodiment of the present application.
  • the spatial audio state setting interface diagram shows spatial audio state one and spatial audio state two.
  • the spatial position information of the second electronic device 1, the second electronic device 2, ... the second electronic device n can be referred to as shown in Figure 3.
  • the setting interface diagram of the spatial audio state also shows the parameter setting interface of the volume of the first electronic device, the volume of the second electronic device 1, the second electronic device 2,...the volume of the second electronic device n in the spatial audio state, and Parameter setting interface for the volume of the first electronic device, the volume of the second electronic device 1, the second electronic device 2,...the volume of the second electronic device n in spatial audio state 2.
  • the circle in Figure 3 represents a sphere.
  • spatial audio state 1 the position indicated by the arrow corresponding to the first electronic device is the spatial position of the first electronic device on the sphere.
  • each second electronic device The positions indicated by the corresponding arrows are the spatial positions of each second electronic device on the sphere.
  • Spatial audio state two is similar to spatial audio state one and will not be described again here.
  • the position adjustment instruction can be to add a new virtual sound source in the preset sphere and set the position information of the virtual sound source, or it can be to adjust a virtual sound source in the preset sphere from the preset initial position information to
  • the target position information that meets the user's needs can also be to adjust a virtual sound source in a preset sphere from the position information that meets the user's old hearing needs to the position information that meets the user's new hearing needs.
  • the second control instruction may be a setting instruction or modification instruction for the audio parameters of the first electronic device, or may be a setting instruction or modification instruction for the audio parameters of at least one second electronic device.
  • Audio parameters can be volume information, timbre information, and other parameters that can affect the auditory effect.
  • volume information takes the audio parameter as volume information as an example to illustrate:
  • the user can pre-set the volume information in different scenarios. Specifically, multiple volume states and the volume information of the first electronic device and the volume information of each second electronic device in each volume state can be configured.
  • the second control instruction may also be a switching instruction for the volume state.
  • the audio data played by the first electronic device facing directly in front of the user has a larger volume
  • the volume of the audio data played by the second electronic device on both sides of the user is smaller.
  • the audio parameters include volume information
  • receiving the second control instruction for the audio parameters includes: receiving, for any one electronic device among the first electronic device and at least one second electronic device, on the user interaction interface. Volume adjustment instructions for electronic devices.
  • the user interaction interface can be referred to Figure 3.
  • the volume information of each electronic device in the first electronic device and at least one second electronic device can be set separately, and the volume information can also be increased or decreased.
  • the first electronic device is a primary device located directly in front of the user; each second electronic device is a consonant device located not directly in front of the user;
  • the control instructions for obtaining the spatial state information include: obtaining the first primary and auxiliary devices Switching instruction; the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's direct front, and the first electronic device The spatial status information of the device corresponds to the non-direct front of the user; or, at least one second electronic device includes a target second electronic device; the target second electronic device is a primary device, and the first electronic device is a consonant device; obtaining the spatial status information
  • the control instructions include: obtaining a second main-auxiliary switching instruction; the second switching instruction is used to change the spatial status information of the first electronic device and the target second electronic device, so that the spatial status information of the first electronic device corresponds to what is directly in front of the user , the spatial state information of the
  • the main sound device is located directly in front of the user, and the user's attention is mainly focused on the audio data played by the main sound device; the consonant device is located not directly in front of the user, which can be understood as the user's attention to the audio data played by the consonant device.
  • the priority is after the audio data played by the main device.
  • the computer located directly in front of the user is playing an online class.
  • the computer is the main sound device.
  • the mobile phone located to the left of the user is playing a shopping live broadcast.
  • the mobile phone is a consonant device.
  • the user's main focus is on the online class. Attention to audio data of live broadcasts is lower than that of audio data of online classes.
  • At least one second electronic device includes the target second electronic device
  • the first electronic device is the main sound device and the target second electronic device is the consonant device.
  • the second electronic device is switched to the main sound device and the first electronic device is switched to the consonant device.
  • the first electronic device is a consonant device
  • the target second electronic device is a main sound device.
  • the first electronic device is switched to the main sound device
  • the second electronic device is switched to a consonant device.
  • the first electronic device is the primary device and is located directly in front of the user
  • the target second electronic device is the consonant device and is located on the left side of the user.
  • the first primary and secondary switching instructions are changed. Spatial state information of the first electronic device and the target second electronic device.
  • the target second electronic device serves as the main device and is located directly in front of the user
  • the first electronic device serves as the consonant device and is located on the right side of the user.
  • At least one second electronic device includes at least two second electronic devices
  • at least one second electronic device including second electronic device 1 and second electronic device 2 taking at least one second electronic device including second electronic device 1 and second electronic device 2 as an example, the tonic device and the consonant device in this case will be described. Switch between:
  • the first electronic device is the main sound device, and the second electronic device 1 and the second electronic device 2 are both consonant devices.
  • the selected second electronic device 1 is switched to the main sound device, and the second electronic device 1 is switched to the main sound device.
  • the first electronic device switches to a consonant device;
  • the first electronic device is the main sound device, and the second electronic device 1 and the second electronic device 2 are both consonant devices.
  • the selected second electronic device 2 is switched to the main sound device, and the second electronic device 2 is switched to the main sound device.
  • the first electronic device switches to a consonant device;
  • the second electronic device 2 is the main sound device, and the first electronic device and the second electronic device 1 are both consonant devices. Through the third main and auxiliary switching command, the selected first electronic device is switched to the main sound device, and the third electronic device is switched to the main sound device. The second electronic device 2 switches to a consonant device;
  • the second electronic device 2 is the main sound device, and the first electronic device and the second electronic device 1 are both consonant devices. Through the third main and auxiliary switching command, the selected second electronic device 1 is switched to the main sound device, and the second electronic device 1 is switched to the main sound device. The second electronic device 2 switches to a consonant device;
  • the second electronic device 1 is the main sound device, and the first electronic device and the second electronic device 2 are both consonant devices.
  • the selected first electronic device is switched to the main sound device, and the third electronic device is switched to the main sound device.
  • the second electronic device 1 is switched to a consonant device;
  • the second electronic device 1 is the main sound device, and the first electronic device and the second electronic device 2 are both consonant devices. Through the third main and auxiliary switching command, the selected second electronic device 2 is switched to the main sound device, and the second electronic device 2 is switched to the main sound device. The second electronic device 1 is switched to a consonant device.
  • the first primary-auxiliary switching instruction at least includes the following implementation methods:
  • the preset sphere is rotated so that the position information of the virtual sound source corresponding to the electronic device selected as the main sound device in the preset sphere reaches the preset position corresponding to the main sound device after rotation.
  • the position information of the virtual sound source corresponding to the electronic device selected as the main sound device is reached after the rotation.
  • the position information of the virtual sound source corresponding to the electronic device selected as the main sound device is The virtual sound source moves to other locations in the sphere after rotation.
  • each electronic device corresponds to a virtual sound source.
  • Drag the virtual sound source in the preset sphere to change the position information of the virtual sound source in the preset sphere.
  • the user can drag the virtual sound source corresponding to the first electronic device so that the virtual sound source leaves the preset position corresponding to the main sound device; the user can also drag the virtual sound source corresponding to the second electronic device to the preset position.
  • the user interaction interface displays multiple preset spatial audio states.
  • spatial audio state 1 the first electronic device is the main sound device and the target second electronic device is the consonant device; in spatial audio state 2, the target second electronic device is It is the main sound device and the first electronic device is the consonant sound device.
  • spatial audio state 2 it will switch to spatial audio state 2 according to user operations.
  • a user initially looks at the computer directly in front of them, then turns to look at the phone to the left.
  • the second main-auxiliary switching instruction and the third main-auxiliary switching instruction are similar to the first main-auxiliary switching instruction, and will not be described again here.
  • the spatial state information of the second electronic device corresponds to what is directly in front of the user. It may be that the position information of the virtual sound source corresponding to the second electronic device in the preset sphere is located at the preset position of the main sound device in the preset sphere, or It may be that the pupil gaze point is located on the second electronic device.
  • the audio data played by the second electronic device sounds like it is coming from directly in front.
  • the spatial state information of the first electronic device corresponds to the position not directly in front of the user, which may be that the position information of the virtual sound source corresponding to the first electronic device in the preset sphere is not at the preset position of the main sound device in the preset sphere, It may also be that the pupil gaze point is located outside the first electronic device.
  • the audio data played by the first electronic device sounds like it is coming from other directions than directly in front.
  • the spatial state information of the first electronic device corresponds to the direct front of the user
  • the spatial state information of the target second electronic device corresponds to the non-direct front of the user.
  • FIG. 4A is a schematic diagram of the first scene of the audio playback method provided by an embodiment of the present application
  • FIG. 4B is a schematic diagram of the second scene of the audio playback method provided by an embodiment of the present application.
  • the first electronic device 401 is located directly in front of the user, and the second electronic device 402 is located on the right hand side of the user.
  • the first electronic device 401 is located on the left side of the user, and the second electronic device 402 is located directly in front of the user.
  • the user is looking at the first electronic device 401 directly in front, while listening to the audio data played by the first electronic device 401 and the audio data played by the second electronic device on the right. Then, in the second scene, the user turns to look at the second electronic device 402 while listening to the audio data played by the first electronic device 401 and the audio data played by the second electronic device on the right.
  • the actual positions of the first electronic device 401 and the second electronic device 402 do not change. What changes is the position of the user's pupil gaze point. Then when the second electronic device 402 is located directly in front of the user, the first electronic device 401 Located on the user's left.
  • control instruction for obtaining the spatial state information includes: obtaining the first detection result of the first electronic device for the pupil gaze point and the second detection result of each second electronic device for the pupil gaze point.
  • the first electronic device and each second electronic device may be provided with a sensor having a detection capability of the pupil gaze point. After detecting the pupil gaze point through the sensor, each second electronic device may send the obtained second detection result to the first electronic device.
  • the pupil fixation point may exist in the first electronic device or in a second electronic device.
  • the first detection result may include whether the pupil gaze point is detected, and may also include position information of the pupil gaze point on the first electronic device.
  • the second detection result may include whether the pupil gaze point is detected, and may also include position information of the pupil gaze point on the second electronic device.
  • Step 104 Determine the spatial status information of the first electronic device and each second electronic device according to the control instruction.
  • determining the spatial status information of the first electronic device and each second electronic device according to the control instruction includes: determining the target position watched by the pupil according to the first detection result and at least one second detection result; The location determines the spatial status information of the first electronic device and each second electronic device.
  • the first detection result can indicate that the pupil gaze point and the position information of the pupil gaze point are detected on the first electronic device
  • the second detection result can indicate that the pupil gaze point is not detected on the second electronic device
  • the second detection result can indicate that the pupil gaze point and the position information of the pupil gaze point are detected on the second electronic device
  • the first detection result can indicate that the pupil gaze point is not detected on the first electronic device
  • the spatial state information of the first electronic device and each second electronic device can be determined.
  • determining the spatial state information of the first electronic device and each second electronic device according to the target position includes: determining the target gazed by the pupil in the first electronic device and at least one second electronic device according to the target position.
  • Electronic equipment and at least one non-target electronic equipment that is not being watched determine the spatial status information of the target electronic equipment based on the target position; determine each non-target electronic equipment based on a preset spatial status information set and the spatial status information of the target electronic equipment Space status information of the device.
  • a first electronic device is connected to two second electronic devices, and the two second electronic devices are second electronic device 1 and second electronic device 2 respectively. If the target position is on the first electronic device, the first electronic device is determined as the target electronic device that is gazed at by the pupil, and the second electronic device 1 and the second electronic device 2 are determined as non-target electronic devices that are not gazed upon. If the target position is on the second electronic device 2, the second electronic device 2 is determined as the target electronic device that is gazed by the pupil, and the second electronic device 1 and the first electronic device are determined as non-target electronic devices that are not gazed upon.
  • Determining the spatial state information of the target electronic device according to the target position can be understood as making the position information of the pupil gaze point correspond to the position corresponding to the user's direct front in the preset sphere.
  • the preset spatial state information set may correspond to the preset spatial audio state.
  • the space state information of each non-target electronic device is determined, which can be understood as the first electronic device and each second electronic device in the preset sphere.
  • the relative position is fixed, so the spatial state information of each non-target electronic device can be determined when the position information of the pupil gaze point corresponds to the position corresponding to the user's direct front in the preset sphere.
  • Figure 5 is a schematic diagram of a third scenario of an audio playback method provided by an embodiment of the present application.
  • the sensor detects that the user's pupil gaze point 503 is located on the first electronic device 501, and the user focuses on the first electronic device 501.
  • the first electronic device 501 is located directly in front of the user, and the first electronic device
  • the spatial position information of 501 in the preset sphere corresponds to the user's front
  • the volume of the first electronic device 501 is a preset value.
  • the second electronic device 502 is located on the right side of the user.
  • the spatial position information of the second electronic device 502 in the preset sphere corresponds to the right side of the user.
  • the volume of the second electronic device 502 is slightly smaller than the volume of the first electronic device 501 .
  • the audio data played by the first electronic device 501 is mainly used, and the audio data played by the second electronic device 502 is used as a supplement.
  • the user's pupil gaze point 503 moves away from the edge of the screen of the first electronic device.
  • the volume of the first electronic device 501 gradually becomes smaller as the pupil gaze point 503 moves away from the screen.
  • the spatial position information of the first electronic device 501 in the preset sphere gradually moves to the left as the pupil gaze point 503 moves, and the direction tends to the left hand of the user.
  • the user's pupil gaze point 503 begins to enter the edge of the screen of the second electronic device 502 .
  • the volume of the second electronic device 502 gradually becomes larger as the pupil gaze point 503 enters the screen.
  • the spatial position information of the second electronic device in the preset sphere gradually moves to the left as the pupil gaze point 503 moves, and the direction tends to be directly in front of the user.
  • the pupil fixation point 503 is located on the second electronic device 502 .
  • the volume of the second electronic device 502 gradually increases until it reaches a preset value, and the spatial position information of the second electronic device 502 in the preset sphere corresponds to what is directly in front of the user.
  • the volume of the first electronic device 501 gradually decreases, and the spatial position information of the first electronic device in the preset sphere corresponds to the user's left hand side.
  • the audio data played by the second electronic device 502 is mainly the audio data played by the first electronic device 501 .
  • determining the spatial state information of the first electronic device and each second electronic device according to the target position includes: determining the target gazed by the pupil in the first electronic device and at least one second electronic device according to the target position. electronic device and at least one non-target electronic device that is not being watched; according to the target electronic device, a corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes The spatial state information of the target electronic device and the spatial state information of each non-target electronic device.
  • the multiple preset alternative spatial state information combinations may be spatial state information combinations corresponding to multiple preset spatial audio states.
  • the alternative spatial state information combination may include the spatial state information combination 1 corresponding to the first electronic device directly in front of the user and the spatial state information combination 2 corresponding to the second electronic device directly in front of the user. Then, when the target electronic device is the first In the case of an electronic device, the space state information combination 1 can be determined based on the first electronic device, and each space state information corresponding to the space state combination 1 is determined as the space state information of the first electronic device and each second electronic device.
  • Step 106 According to each spatial state information, perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device to obtain the first target audio number. according to.
  • the spatial mixing process may be, in the presence of at least two sound sources at different locations, mixing audio data from each sound source, so that when the mixed audio data is played through the audio playback device, each audio data It sounds like it's coming from the corresponding sound source, not from the same direction.
  • the mixing operation may be to produce multiple types of audio data into one audio file. When the audio file is played, each type of audio data still sounds independent. And each comes from the corresponding sound source, and multiple sound sources are located in different directions of the user.
  • the spatial position information of the first electronic device corresponds to the front of the user
  • the value of the volume information of the first electronic device is x
  • the spatial position information of the second electronic device corresponds to the left side of the user
  • the value of the volume information of the second electronic device is x.
  • the value is y, x>y.
  • Each sound source can be located in different directions, and various audio parameters such as the volume of each audio data can also be obtained. It can be flexibly set separately, so that only one audio playback device can play the audio data after spatial mixing processing, and multiple audio playback devices located in different directions of the user can be used as sound sources to simultaneously play the corresponding audio data.
  • Step 108 Send the first target audio data to the audio playback device for playback.
  • the first electronic device may send the first target audio data to the earphone, thereby playing the first target audio data through the earphone. For example, if the spatial state information of the first electronic device corresponds to the front of the user, and the spatial state information of the second electronic device corresponds to the left side of the user, then the user can listen to the first target audio data through the earphones.
  • the first audio data comes from the front and has a louder volume, while the second audio data can be heard from the left and has a lower volume.
  • the audio playback method also includes: in the case where the audio data played by the first electronic device is replaced, according to each spatial state information, the replaced audio data played by the first electronic device and each second electronic device are replaced.
  • the audio data played by the device is subjected to spatial mixing processing to obtain the second target audio data and sent to the audio playback device for playback; or, in the case where the audio data played by each second electronic device is replaced, according to each For spatial status information, perform spatial mixing processing on the audio data played by the first electronic device and the replaced audio data played by each second electronic device, obtain the third target audio data, and send it to the audio playback device for playback.
  • the control instruction for obtaining spatial state information Order determines the spatial state information of the first electronic device and each second electronic device; according to each spatial state information, determine the audio data played by the first electronic device and the audio data played by each second electronic device Perform spatial mixing processing to obtain the first target audio data; send the first target audio data to the audio playback device for playback.
  • spatial mixing processing can be performed according to the spatial state information corresponding to the first electronic device and at least one second electronic device, thereby preventing the user from listening to multiple electronic devices located in different locations playing different music at the same time.
  • the audio data has a sense of disorientation.
  • FIG. 6 is a second flow diagram of an audio playback method provided by an embodiment of the present application.
  • the audio playback device in this embodiment may be headphones.
  • step 602 the first electronic device and the second electronic device are connected to each other, and the earphone is connected to the first electronic device.
  • Step 604 The second electronic device transmits audio data to the first electronic device.
  • Step 606 The user sets spatial location information according to the actual location status of the device or preferences.
  • Step 608 The user sets the volume for different scenes, and the user can configure multiple volume states.
  • Step 610 Perform spatial mixing processing on the audio data played by the first electronic device and the second electronic device, and output the audio data to the headphones.
  • step 610 After step 610, at least one of step 612, step 618, and step 620 may be performed.
  • Step 612 Detect the pupil fixation point.
  • Step 614 Whether the target electronic device is switched.
  • step 616 If yes, return to step 608; if not, execute step 616.
  • Step 616 Maintain the original output state.
  • Step 618 Determine whether the audio data has been replaced.
  • step 616 If yes, return to step 606; if not, execute step 616.
  • Step 620 Determine whether the volume status is switched.
  • step 616 If yes, return to step 608; if not, execute step 616.
  • the audio playback method provided by the embodiment shown in FIG. 6 can implement various processes implemented by the foregoing audio playback method embodiment. To avoid repetition, details will not be described here.
  • the execution subject may be an audio playback device, or a control module in the audio playback device for executing the audio playback method.
  • the method of performing audio playback by the audio playback device is taken as an example to describe the audio playback device provided by the embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an audio playback device provided by an embodiment of the present application.
  • the audio playback device is applied to a first electronic device.
  • the first electronic device is connected to at least one second electronic device.
  • the first electronic device is connected to the audio playback device.
  • the audio playback device includes:
  • Acquisition module 701 used to obtain control instructions for space status information
  • Determining module 702 configured to determine the space status information of the first electronic device and each second electronic device according to the control instruction
  • the processing module 703 is configured to perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device according to each spatial state information, and obtain the first target audio data;
  • the sending module 704 is used to send the first target audio data to the audio playback device for playback.
  • the first electronic device is a main sound device; each second electronic device is a consonant device; the acquisition module is specifically used for:
  • the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's normal Ahead, the spatial state information of the first electronic device corresponds to the non-direct front of the user;
  • At least one second electronic device includes a target second electronic device; the target second electronic device is a main sound device, and the first electronic device is a consonant device; and the acquisition module is specifically used for:
  • the second switching instruction is used to change the spatial state information of the first electronic device and the target second electronic device, so that the spatial state information of the first electronic device corresponds to the user's direct front, and the target second electronic device
  • the device's spatial status information corresponds to what is not directly in front of the user.
  • the acquisition module 701 is specifically used for:
  • a first detection result of the first electronic device for the pupil gaze point and a second detection result of each second electronic device for the pupil gaze point are obtained.
  • the determining module 702 includes:
  • a first determination unit configured to determine the position of the target gazed by the pupil based on the first detection result and at least one second detection result
  • a second determination unit configured to determine the first electronic device and each second electronic device according to the target position. Space status information of the device.
  • the second determination unit is specifically used for:
  • the target location determine the spatial status information of the target electronic device
  • the space state information of each non-target electronic device is determined.
  • the second determination unit is specifically used for:
  • the corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes the space state information of the target electronic device and each non-target electronic device space status information.
  • the audio playback device also includes:
  • a mixing module configured to combine the replaced audio data played by the first electronic device and the audio played by each second electronic device according to each spatial state information when the audio data played by the first electronic device is replaced.
  • the data is subjected to spatial mixing processing to obtain the second target audio data and sent to the audio playback device for playback; or, in the case where the audio data played by each second electronic device is replaced, according to each spatial status information, Perform spatial mixing processing on the audio data played by the first electronic device and the replaced audio data played by each second electronic device to obtain third target audio data, and send it to the audio playback device for playback.
  • the spatial state information includes spatial position information and audio parameters.
  • the acquisition module 701 includes:
  • a first receiving unit configured to receive a first control instruction for spatial position information
  • the second receiving unit is used to receive the second control instruction for the audio parameters.
  • the first receiving unit is specifically used for:
  • the position information of the virtual sound source corresponding to the electronic device in the preset sphere is determined as the spatial position information of the electronic device ;
  • the position adjustment instruction is used to adjust the virtual sound source in the preset Position information in the sphere.
  • the audio parameters include volume information; the second receiving unit is specifically used for:
  • a volume adjustment instruction for the electronic device is received on the user interaction interface.
  • the audio playback device obtaineds the control instruction of the spatial state information; determines the spatial state information of the first electronic device and each second electronic device according to the control instruction; and determines the first electronic device according to each spatial state information.
  • the audio data played by the electronic device and the audio data played by each second electronic device are spatially mixed to obtain the first target audio data; the first target audio data is sent to the audio playback device for playback.
  • spatial mixing processing can be performed according to the spatial state information corresponding to the first electronic device and at least one second electronic device, thereby preventing the user from listening to multiple electronic devices located in different locations playing different music at the same time.
  • the audio data has a sense of disorientation.
  • the audio playback device in the embodiment of the present application may be a device, or may be a component, integrated circuit, or chip in a terminal.
  • the device may be a mobile electronic device or a non-mobile electronic device.
  • the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a personal digital assistant (personal digital assistant).
  • UMPC ultra-mobile personal computer
  • PDA personal digital assistant
  • non-mobile electronic devices can be servers, network attached storage (Network Attached Storage, NAS), personal computers (personal computers, PC), televisions (television, TV), teller machines or self-service machines, etc., this application The examples are not specifically limited.
  • the audio playback device in the embodiment of the present application may be a device with an operating system.
  • the operating system can be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiments of this application.
  • the audio playback device provided by the embodiments of the present application can implement various processes implemented by the foregoing audio playback method embodiments. To avoid repetition, they will not be described again here.
  • this embodiment of the present application also provides an electronic device 800, including a processor 801, a memory 802, and a program or instructions stored on the memory 802 and executable on the processor 801.
  • the program Or when the instruction is executed by the processor 801, each process of the above audio playback method embodiment is implemented, and the same technical effect can be achieved. To avoid repetition, the details will not be described here.
  • the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
  • Figure 9 is a schematic diagram of the hardware structure of an electronic device that implements an embodiment of the present application.
  • the electronic device 900 includes but is not limited to: radio frequency unit 901, network module 902, audio output unit 903, input unit 904, sensor 905, display unit 906, user input unit 907, interface unit 908, memory 909, processor 910, etc. part.
  • the electronic device 900 may also include a power supply (such as a battery) that supplies power to various components.
  • the power supply may be logically connected to the processor 910 through a power management system, thereby managing charging, discharging, and function through the power management system. Consumption management and other functions.
  • the structure of the electronic device shown in Figure 9 does not constitute a limitation on the electronic device.
  • the electronic device may include more or less components than shown in the figure, or combine certain components, or arrange different components, which will not be described again here. .
  • the processor 910 is used to obtain control instructions for spatial state information
  • each spatial state information perform spatial mixing processing on the audio data played by the first electronic device and the audio data played by each second electronic device to obtain the first target audio data;
  • a control instruction for spatial status information is obtained; according to the control instruction, the spatial status information of the first electronic device and each second electronic device is determined; and according to each spatial status information, the video played by the first electronic device is
  • the audio data is spatially mixed with the audio data played by each second electronic device to obtain the first target audio data; and the first target audio data is sent to the audio playback device for playback.
  • the first electronic device is a tonic device located directly in front of the user; each second electronic device is a consonant device located not directly in front of the user; the processor 910 is configured to:
  • Control instructions for obtaining space status information including:
  • the first main-auxiliary switching instruction is used to change the spatial status information of the first electronic device and each second electronic device, so that the spatial status information of the selected second electronic device corresponds to the user's normal Ahead, the spatial state information of the first electronic device corresponds to the non-direct front of the user;
  • At least one second electronic device includes a target second electronic device; the target second electronic device is a main sound device, and the first electronic device is a consonant device; and a control instruction for obtaining spatial state information, include:
  • the second switching instruction is used to change the spatial state information of the first electronic device and the target second electronic device, so that the spatial state information of the first electronic device corresponds to the user's direct front, and the target second electronic device
  • the device's spatial status information corresponds to what is not directly in front of the user.
  • processor 910 is also used to:
  • Control instructions for obtaining space status information including:
  • a first detection result of the first electronic device for the pupil gaze point and a second detection result of each second electronic device for the pupil gaze point are obtained.
  • processor 910 is also used to:
  • determining the space status information of the first electronic device and each second electronic device includes:
  • spatial status information of the first electronic device and each second electronic device is determined.
  • processor 910 is also used to:
  • determine the spatial status information of the first electronic device and each second electronic device including:
  • the target location determine the spatial status information of the target electronic device
  • the space state information of each non-target electronic device is determined.
  • processor 910 is also used to:
  • determine the spatial status information of the first electronic device and each second electronic device including:
  • the corresponding space state information combination is determined from a plurality of preset alternative space state information combinations; the space state information combination corresponding to the target electronic device includes the space state information of the target electronic device and each non-target electronic device space status information.
  • processor 910 is also used to:
  • spatial mixing processing is performed on the replaced audio data played by the first electronic device and the audio data played by each second electronic device according to each spatial state information. , obtain the second target audio data and send it to the audio playback device for playback;
  • spatial mixing is performed on the audio data played by the first electronic device and the replaced audio data played by each second electronic device according to each spatial state information. Audio processing, obtaining the third target audio data, and sending it to the audio playback device for playback.
  • processor 910 is also used to:
  • Spatial status information includes spatial position information and audio parameters.
  • Control instructions for obtaining spatial status information include:
  • processor 910 is also used to:
  • Receiving the first control instruction for spatial location information includes:
  • the position information of the virtual sound source corresponding to the electronic device in the preset sphere is determined as the spatial position information of the electronic device ;
  • the position adjustment instruction is used to adjust the position information of the virtual sound source in the preset sphere.
  • the audio parameters include volume information; processor 910 is also used to:
  • Receive second control instructions for audio parameters including:
  • a volume adjustment instruction for the electronic device is received on the user interaction interface.
  • the main device and the consonant device can be flexibly switched between the first electronic device and at least one second electronic device; according to the detected pupil Gaze point can freely and flexibly control spatial position information and spatial volume information, providing users with the most suitable audio playback effect during the process of users naturally shifting their gaze; by receiving position adjustment instructions and targeting of virtual sound sources on the user interaction interface electronic device volume Adjustment instructions can flexibly set the spatial status information of each electronic device and enrich the auditory effect.
  • the input unit 904 may include a graphics processor (Graphics Processing Unit, GPU) 9041 and a microphone 9042.
  • the graphics processor 9041 is responsible for the image capture device (GPU) in the video capture mode or the image capture mode. Process the image data of still pictures or videos obtained by cameras (such as cameras).
  • the display unit 906 may include a display panel 9061, which may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the user input unit 907 includes a touch panel 9071 and other input devices 9072. Touch panel 9071, also known as touch screen.
  • the touch panel 9071 may include two parts: a touch detection device and a touch controller.
  • Other input devices 9072 may include but are not limited to physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which will not be described again here.
  • Memory 909 may be used to store software programs as well as various data, including but not limited to application programs and operating systems.
  • the processor 910 can integrate an application processor and a modem processor, where the application processor mainly processes the operating system, user interface, application programs, etc., and the modem processor mainly processes wireless communications. It can be understood that the above modem processor may not be integrated into the processor 910.
  • Embodiments of the present application also provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the program or instructions are executed by a processor, each process of the audio playback method embodiment is implemented, and the same can be achieved. The technical effects will not be repeated here to avoid repetition.
  • the processor is the processor in the electronic device described in the above embodiment.
  • the readable storage media includes computer-readable storage media, such as computer read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks, etc.
  • An embodiment of the present application further provides a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the above audio playback method embodiment. Each process can achieve the same technical effect. To avoid duplication, it will not be described again here.
  • chips mentioned in the embodiments of this application may also be called system-on-chip, system-on-a-chip, system-on-a-chip or system-on-chip, etc.
  • the methods of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. implementation.
  • the technical solution of the present application can be embodied in the form of a computer software product that is essentially or contributes to the existing technology.
  • the computer software product is stored in a storage medium (such as ROM/RAM, disk , optical disk), including several instructions to cause a terminal (which can be a mobile phone, computer, server, or network device, etc.) to execute the methods described in various embodiments of this application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Acoustics & Sound (AREA)
  • Stereophonic System (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

本申请公开了一种音频播放方法和电子设备,属于电子技术领域。该音频播放方法包括:获取空间状态信息的控制指令;根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;向音频播放设备发送第一目标音频数据以进行播放。

Description

音频播放方法和电子设备
交叉引用
本发明要求在2022年03月07日提交中国专利局、申请号为202210225832.8、发明名称为“音频播放方法和电子设备”的中国专利申请的优先权,该申请的全部内容通过引用结合在本发明中。
技术领域
本申请属于电子技术领域,具体涉及一种音频播放方法和电子设备。
背景技术
随着电子技术的发展,电子设备的使用越来越普遍。在实际的应用场景中,用户有时会使用多个电子设备同时播放不同的音频数据,例如,一边在电脑上打游戏,一边在手机上挂着直播间。
若采用公放的音频播放方式,可能会干扰到其他人,且隐私性较差。若通过耳机同时播放多种音频数据,无论电子设备相对于用户在哪个方向,用户都会感知声音来自正前方,且多种不同的音频数据到达耳朵时不存在时间差。因此,可能导致用户难以分辨音频数据来自哪个电子设备,产生方位错乱的感觉。
发明内容
本申请实施例的目的是提供一种音频播放和电子设备,能够解决如何避免用户在听多种不同的音频数据时产生方位错乱的感觉的问题。
第一方面,本申请实施例提供了一种音频播放方法,应用于第一电子设备,所述第一电子设备与至少一个第二电子设备连接,所述第一电子设备与音频播放设备连接,该音频播放方法包括:
获取空间状态信息的控制指令;
根据所述控制指令,确定所述第一电子设备和每个所述第二电子设备的空间状态信息;
根据每个所述空间状态信息,对所述第一电子设备播放的音频数据和每个所述第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频 数据;
向所述音频播放设备发送第一目标音频数据以进行播放。
第二方面,本申请实施例提供了一种音频播放装置,应用于第一电子设备,所述第一电子设备与至少一个第二电子设备连接,所述第一电子设备与音频播放设备连接,该音频播放装置包括:
获取模块,用于获取空间状态信息的控制指令;
确定模块,用于根据所述控制指令,确定所述第一电子设备和每个所述第二电子设备的空间状态信息;
处理模块,用于根据每个所述空间状态信息,对所述第一电子设备播放的音频数据和每个所述第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;
发送模块,用于向所述音频播放设备发送第一目标音频数据以进行播放。
第三方面,本申请实施例提供了一种电子设备,该电子设备包括处理器、存储器及存储在所述存储器上并可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如第一方面所述的音频播放方法的步骤。
第四方面,本申请实施例提供了一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如第一方面所述的音频播放方法的步骤。
第五方面,本申请实施例提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如第一方面所述的音频播放方法。
在本申请实施例中,获取空间状态信息的控制指令;根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;向音频播放设备发送第一目标音频数据以进行播放。通过本申请实施例的技术方案,能够按照第一电子设备和至少一个第二电子设备分别对应的空间状态信息进行空间混音处理,避免用户在同时听多个位于不同位置的电子设备播放不同的音频数据时产生方位错乱的感觉。
附图说明
图1是本申请一实施例提供的音频播放方法的第一种流程示意图;
图2是本申请一实施例提供的第一电子设备、第二电子设备以及音频播放设备的连接关系示意图;
图3是本申请一实施例提供的音频播放方法中空间音频状态的设置界面图;
图4A是本申请一实施例提供的音频播放方法的第一种场景示意图;
图4B是本申请一实施例提供的音频播放方法的第二种场景示意图;
图5是本申请一实施例提供的音频播放方法的第三种场景示意图;
图6是本申请一实施例提供的音频播放方法的第二种流程示意图;
图7是本申请一实施例提供的音频播放装置的结构示意图;
图8是本申请一实施例提供的一种电子设备的结构示意图;
图9是本申请一实施例提供的一种电子设备的硬件结构示意图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员获得的所有其他实施例,都属于本申请保护的范围。
本申请的说明书和权利要求书中的术语“第一”、“第二”等是用于区别类似的对象,而不用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便本申请的实施例能够以除了在这里图示或描述的那些以外的顺序实施,且“第一”、“第二”等所区分的对象通常为一类,并不限定对象的个数,例如第一对象可以是一个,也可以是多个。此外,说明书以及权利要求中“和/或”表示所连接对象的至少其中之一,字符“/”,一般表示前后关联对象是一种“或”的关系。
下面结合附图,通过具体的实施例及其应用场景对本申请实施例提供的音频播放方法进行详细地说明。
图1是本申请一实施例提供的音频播放方法的第一种流程示意图。
该音频播放方法应用于第一电子设备,第一电子设备与至少一个第二电子设备连接,第一电子设备与音频播放设备连接。
第一电子设备可以是具有音频数据处理功能和音频播放功能的电子设备, 例如,电脑、手机、平板等。第二电子设备可以是具有音频数据播放功能的电子设备,例如,电脑、手机、平板等。
至少一个第二电子设备可以是一个第二电子设备,也可以是多个第二电子设备。
音频播放设备可以是耳机,也可以是其他音频电子设备。
图2是本申请一实施例提供的第一电子设备、第二电子设备以及音频播放设备的连接关系示意图。如图2所示,第一电子设备201与一个第二电子设备202连接,且第一电子设备201与耳机203连接。用户戴着耳机203同时听第一电子设备201播放的音频数据和第二电子设备202播放的音频数据。并且,在图2中,用户面向第二电子设备202,还可以观看第二电子设备202所播放的视频画面。
步骤102,获取空间状态信息的控制指令。
空间状态信息可以为空间位置信息,也可以为空间位置信息和音频参数,还可以为空间音频状态。音频参数可以是音量信息、音色信息等多种可影响听觉效果的参数。
空间位置信息可以是第一电子设备对应的虚拟声源在预设球体中的空间位置信息,也可以是任意一个第二电子设备对应的虚拟声源在预设球体中的空间位置信息。音量信息可以是第一电子设备的音量大小,也可以是任意一个第二电子设备的音量大小。
控制指令可以是设置指令,也可以是修改指令,还可以是检测指令,等。
需要注意的是,音量信息可以独立调节,也可以与空间位置信息同时调节,还可以在调节空间位置信息之后,将音量信息调节为与调节后的空间位置信息对应的音量信息。
可选地,空间状态信息包括空间位置信息和音频参数,获取空间状态信息的控制指令,包括:接收针对空间位置信息的第一控制指令;和/或,接收针对音频参数的第二控制指令。
第一控制指令可以是针对第一电子设备的空间位置信息的设置指令或修改指令,也可以是针对至少一个第二电子设备的空间位置信息的设置指令或修改指令。
可选地,接收针对空间位置信息的第一控制指令,包括:在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上, 将电子设备对应的虚拟声源在预设球体中的位置信息确定为电子设备的空间位置信息;接收虚拟声源的位置调节指令;位置调节指令用于调节虚拟声源在预设球体中的位置信息。
用户交互界面可以参照图3。图3是本申请一实施例提供的音频播放方法中空间音频状态的设置界面图。该空间音频状态的设置界面图中示出了空间音频状态一和空间音频状态二。
空间音频状态一所对应的第一电子设备、第二电子设备1、第二电子设备2、…、第二电子设备n的空间位置信息,以及空间音频状态二所对应的第一电子设备、第二电子设备1、第二电子设备2、…第二电子设备n的空间位置信息可以参照图3所示。
该空间音频状态的设置界面图还示出了在空间音频状态一下的第一电子设备音量、第二电子设备1音量、第二电子设备2、…第二电子设备n音量的参数设置接口,以及在空间音频状态二下的第一电子设备音量、第二电子设备1音量、第二电子设备2、…第二电子设备n音量的参数设置接口。
图3中的圆形代表的是一个球体,在空间音频状态一,第一电子设备对应的箭头所指示的位置是第一电子设备在该球体上的空间位置,同理,各个第二电子设备对应的箭头所指示的位置是各个第二电子设备在该球体上的空间位置。空间音频状态二与空间音频状态一相似,此处不再赘述。
位置调节指令可以是在预设球体中添加一个新的虚拟声源并设置该虚拟声源的位置信息,也可以是在预设球体中,将一个虚拟声源从预先设置的初始位置信息调节为符合用户需求的目标位置信息,也可以是在预设球体中,将一个虚拟声源从符合用户旧的听觉需求的位置信息调节为符合用户新的听觉需求的位置信息。
第二控制指令可以是针对第一电子设备的音频参数的设置指令或修改指令,也可以是针对至少一个第二电子设备的音频参数的设置指令或修改指令。
音频参数可以是音量信息、音色信息等多种可影响听觉效果的参数。下面以音频参数为音量信息为例进行说明:
用户可以预先设置不同场景下的音量信息,具体地,可以配置多种音量状态以及每种音量状态下第一电子设备的音量信息和每个第二电子设备的音量信息。第二控制指令还可以是针对音量状态的切换指令。
例如,在用户正前方面对的第一电子设备播放的音频数据的音量较大, 在用户两侧的第二电子设备播放的音频数据的音量较小。
可选地,音频参数包括音量信息;接收针对音频参数的第二控制指令,包括:在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,接收针对电子设备的音量调节指令。
用户交互界面可以参照图3。在如图3所示的用户交互界面中,可以分别设置第一电子设备和至少一个第二电子设备中的每个电子设备的音量信息,也可以将音量信息增大或减小。
可选地,第一电子设备为位于用户的正前方的主音设备;每个第二电子设备为位于用户的非正前方的辅音设备;获取空间状态信息的控制指令,包括:获取第一主辅切换指令;第一主辅切换指令用于改变第一电子设备和每个第二电子设备的空间状态信息,使得被选中的第二电子设备的空间状态信息对应于用户的正前方,第一电子设备的空间状态信息对应于用户的非正前方;或者,至少一个第二电子设备包括目标第二电子设备;目标第二电子设备为主音设备,第一电子设备为辅音设备;获取空间状态信息的控制指令,包括:获取第二主辅切换指令;第二切换指令用于改变第一电子设备和目标第二电子设备的空间状态信息,使得第一电子设备的空间状态信息对应于用户的正前方,目标第二电子设备的空间状态信息对应于用户的非正前方。
主音设备位于用户的正前方,用户的注意力主要集中在该主音设备所播放的音频数据上;辅音设备位于用户的非正前方,可以理解为,用户对该辅音设备所播放的音频数据的关注优先级在对主音设备所播放的音频数据之后。
例如,位于用户的正前方的计算机正在播放网课,该计算机为主音设备,位于用户的左侧的手机在播放一场购物直播,该手机为辅音设备,用户的主要精力集中在网课上,对直播的音频数据的关注度低于网课的音频数据。
在至少一个第二电子设备包括目标第二电子设备的情况下,主音设备和辅音设备之间的切换存在如下可能:
(a1)第一电子设备为主音设备,目标第二电子设备为辅音设备,通过第一主辅切换指令,将第二电子设备切换为主音设备,将第一电子设备切换为辅音设备。
(a2)第一电子设备为辅音设备,目标第二电子设备为主音设备,通过第二主辅切换指令,将第一电子设备切换为主音设备,将第二电子设备切换为辅音设备。
例如,第一电子设备为主音设备,位于用户的正前方,目标第二电子设备为辅音设备,位于用户的左侧,当用户转身面向目标第二电子设备时,通过第一主辅切换指令改变第一电子设备和目标第二电子设备的空间状态信息,该情况下,目标第二电子设备作为主音设备,位于用户的正前方,该第一电子设备作为辅音设备,位于用户的右侧。
在至少一个第二电子设备包括至少两个第二电子设备的情况下,以至少一个第二电子设备包括第二电子设备1和第二电子设备2为例,说明该情况下主音设备和辅音设备之间的切换:
(b1)第一电子设备为主音设备,第二电子设备1和第二电子设备2均为辅音设备,通过第一主辅切换指令,将被选中的第二电子设备1切换为主音设备,将第一电子设备切换为辅音设备;
(b2)第一电子设备为主音设备,第二电子设备1和第二电子设备2均为辅音设备,通过第一主辅切换指令,将被选中的第二电子设备2切换为主音设备,将第一电子设备切换为辅音设备;
(b3)第二电子设备2为主音设备,第一电子设备和第二电子设备1均为辅音设备,通过第三主辅切换指令,将被选中的第一电子设备切换为主音设备,将第二电子设备2切换为辅音设备;
(b4)第二电子设备2为主音设备,第一电子设备和第二电子设备1均为辅音设备,通过第三主辅切换指令,将被选中的第二电子设备1切换为主音设备,将第二电子设备2切换为辅音设备;
(b5)第二电子设备1为主音设备,第一电子设备和第二电子设备2均为辅音设备,通过第三主辅切换指令,将被选中的第一电子设备切换为主音设备,将第二电子设备1切换为辅音设备;
(b6)第二电子设备1为主音设备,第一电子设备和第二电子设备2均为辅音设备,通过第三主辅切换指令,将被选中的第二电子设备2切换为主音设备,将第二电子设备1切换为辅音设备。
第一主辅切换指令至少包括如下实施方式:
(c1)在预先设置的用户交互界面上针对空间状态信息执行的控制操作。
例如,转动预设球体,使被选中作为主音设备的电子设备对应的虚拟声源在预设球体中的位置信息在转动后到达主音设备对应的预设位置,该情况下,原本在预设位置的虚拟声源在转动后移动到球体中的其他位置。
又例如,在第一电子设备和至少一个第二电子设备中,每个电子设备对应于一个虚拟声源。在预设球体中拖动虚拟声源,以改变虚拟声源在该预设球体中的位置信息。用户可以拖动第一电子设备对应的虚拟声源,使得该虚拟声源离开主音设备对应的预设位置;用户还可以拖动第二电子设备对应的虚拟声源至该预设位置。
还例如,用户交互界面上显示有预先设置的多种空间音频状态,空间音频状态1中第一电子设备为主音设备,目标第二电子设备为辅音设备;空间音频状态2中目标第二电子设备为主音设备,第一电子设备为辅音设备。当前处于空间音频状态1,根据用户操作切换至空间音频状态2。
(c2)移动瞳孔注视点。
例如,用户最初注视着正前方的计算机,然后转头看向左侧的手机。
第二主辅切换指令和第三主辅切换指令与第一主辅切换指令类似,此处不再赘述。
第二电子设备的空间状态信息对应于用户的正前方,可以是,第二电子设备对应的虚拟声源在预设球体中的位置信息位于主音设备在预设球体中的预设位置处,也可以是,瞳孔注视点位于第二电子设备。
该情况下,第二电子设备所播放的音频数据听起来是从正前方传来的。
第一电子设备的空间状态信息对应于用户的非正前方,可以是,第一电子设备对应的虚拟声源在预设球体中的位置信息不在主音设备在预设球体中的预设位置处,也可以是,瞳孔注视点位于第一电子设备之外。
该情况下,第一电子设备所播放的音频数据听起来是从正前方之外的其他方向传来的。
第一电子设备的空间状态信息对应于所述用户的正前方,以及,目标第二电子设备的空间状态信息对应于所述用户的非正前方,与上述特征类似,此处不再赘述。
需要注意的是,在第一电子设备与第二电子设备的实际位置并未发生改变的情况下,当第一电子设备在预设球体中的空间位置信息发生改变时,每个第二电子设备在预设球体中的空间位置信息随之改变,可以参照图4A和图4B。
图4A是本申请一实施例提供的音频播放方法的第一种场景示意图;图4B是本申请一实施例提供的音频播放方法的第二种场景示意图。
如图4A所示,在第一个场景中,第一电子设备401位于用户的正前方,第二电子设备402位于用户的右手侧。如图4B所示,在第二个场景中,第一电子设备401位于用户的左侧,第二电子设备402位于用户的正前方。
例如,在第一个场景中,用户在看正前方的第一电子设备401,同时听第一电子设备401播放的音频数据和右侧的第二电子设备播放的音频数据。接着,在第二个场景中,用户转头看向第二电子设备402,同时听第一电子设备401播放的音频数据和右侧的第二电子设备播放的音频数据。第一电子设备401与第二电子设备402的实际位置没有发生变化,改变的是用户的瞳孔注视点的位置,则在第二电子设备402位于用户的正前方的情况下,第一电子设备401位于用户左侧。
可选地,获取空间状态信息的控制指令,包括:获取第一电子设备针对瞳孔注视点的第一检测结果和每个第二电子设备针对瞳孔注视点的第二检测结果。
第一电子设备和每个第二电子设备上可以设置有具有瞳孔注视点的检测能力的传感器。每个第二电子设备在通过传感器进行瞳孔注视点检测以后,可以将获得的第二检测结果发送到第一电子设备。
在同一时间点,瞳孔注视点可能存在于第一电子设备,也可能存在于一个第二电子设备。
第一检测结果可以包括是否检测到瞳孔注视点,也可以包括瞳孔注视点在第一电子设备上的位置信息。
第二检测结果可以包括是否检测到瞳孔注视点,也可以包括瞳孔注视点在第二电子设备上的位置信息。
步骤104,根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息。
可选地,根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息,包括:根据第一检测结果和至少一个第二检测结果,确定被瞳孔注视的目标位置;根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息。
若第一检测结果可以表示在第一电子设备上检测到瞳孔注视点以及该瞳孔注视点的位置信息,且第二检测结果可以表示在第二电子设备上未检测到瞳孔注视点,则可以确定被瞳孔注视的目标位置在第一电子设备,该目标位 置的位置信息即该瞳孔注视点的位置信息。
若第二检测结果可以表示在第二电子设备上检测到瞳孔注视点以及该瞳孔注视点的位置信息,且第一检测结果可以表示在第一电子设备上未检测到瞳孔注视点,则可以确定被瞳孔注视的目标位置在第二电子设备,该目标位置的位置信息即该瞳孔注视点的位置信息。
根据瞳孔注视点的位置信息,可以确定第一电子设备和每个第二电子设备的空间状态信息。
可选地,根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息,包括:根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;根据目标位置,确定目标电子设备的空间状态信息;根据预设的空间状态信息集合和目标电子设备的空间状态信息,确定每个非目标电子设备的空间状态信息。
例如,第一电子设备与两个第二电子设备连接,两个第二电子设备分别为第二电子设备1和第二电子设备2。若目标位置在第一电子设备上,则将第一电子设备确定为被瞳孔注视的目标电子设备,将第二电子设备1和第二电子设备2确定为未被注视的非目标电子设备。若目标位置在第二电子设备2上,则将第二电子设备2确定为被瞳孔注视的目标电子设备,将第二电子设备1和第一电子设备确定为未被注视的非目标电子设备。
根据目标位置,确定目标电子设备的空间状态信息,可以理解为,使瞳孔注视点的位置信息对应于预设球体中与用户的正前方对应的位置。
预设的空间状态信息集合可以对应于预设的空间音频状态。根据预设的空间状态信息集合和目标电子设备的空间状态信息,确定每个非目标电子设备的空间状态信息,可以理解为,第一电子设备和每个第二电子设备在预设球体中的相对位置是固定的,故可以在瞳孔注视点的位置信息对应于预设球体中与用户的正前方对应的位置的情况下,确定各个非目标电子设备的空间状态信息。
图5是本申请一实施例提供的音频播放方法的第三种场景示意图。
如图5所示,首先,传感器检测到用户的瞳孔注视点503位于第一电子设备501,用户聚焦于第一电子设备501,此时第一电子设备501位于用户的正前方,第一电子设备501在预设球体中的空间位置信息对应于用户的正前 方,第一电子设备501的音量大小为预设数值。第二电子设备502位于用户右手边,第二电子设备502在预设球体中的空间位置信息对应于用户的右侧,第二电子设备502的音量与第一电子设备501的音量相比稍小。此时以第一电子设备501播放的音频数据为主,第二电子设备502播放的音频数据为辅。
其次,用户的瞳孔注视点503从第一电子设备的屏幕边缘移出。第一电子设备501的音量大小随着瞳孔注视点503移出屏幕而逐渐变小。第一电子设备501在预设球体中的空间位置信息随着瞳孔注视点503移动而逐渐向左移动,方向趋于用户左手方。
接着,用户的瞳孔注视点503开始进入第二电子设备502的屏幕边缘。第二电子设备502的音量大小随着瞳孔注视点503进入屏幕而逐渐变大。第二电子设备在预设球体中的空间位置信息随着瞳孔注视点503移动而逐渐向左移动,方向趋于用户正前方。
最后,瞳孔注视点503位于第二电子设备502。第二电子设备502的音量大小逐渐变大,直至增大到预设数值,第二电子设备502在预设球体中的空间位置信息对应于用户的正前方。第一电子设备501的音量大小逐渐降低,第一电子设备在预设球体中的空间位置信息对应于用户的左手边。此时以第二电子设备502播放的音频数据为主,第一电子设备501播放的音频数据。
可选地,根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息,包括:根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;根据目标电子设备,从预设的多个备选空间状态信息组合中确定对应的空间状态信息组合;目标电子设备对应的空间状态信息组合包括目标电子设备的空间状态信息和每个非目标电子设备的空间状态信息。
预设的多个备选空间状态信息组合可以是多个预先设置的空间音频状态对应的空间状态信息组合。例如,备选空间状态信息组合可以包括第一电子设备在用户正前方对应的空间状态信息组合1和第二电子设备在用户正前方对应的空间状态信息组合2,则在目标电子设备为第一电子设备的情况下,可以根据第一电子设备确定空间状态信息组合1,将空间状态组合1对应的各个空间状态信息确定为第一电子设备和每个第二电子设备的空间状态信息。
步骤106,根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数 据。
空间混音处理,可以是,在存在至少两个不同位置的声源的情况下,混合来自各个声源的音频数据,使得通过音频播放设备播放混合后的多种音频数据时,每种音频数据听起来是从对应的声源传过来的,而不是从相同方向传来的。该过程中,尽管对多种音频数据进行了混合操作,但该混合操作可以是将多种音频数据制作为一个音频文件,该音频文件被播放时,每种音频数据听起来依然是独立的,且分别来自对应的声源,多个声源位于用户不同的方向。例如,第一电子设备的空间位置信息对应于用户正前方,第一电子设备的音量信息的数值为x,第二电子设备的空间位置信息对应于用户左侧,第二电子设备的音量信息的数值为y,x>y。则对第一音频数据和第二音频数据进行空间混音处理后,第一音频数据听起来来自用户正前方,音量为x,第二音频数据听起来来自用户左侧,音量为y。
通过空间混音处理,可以获得方位明确的音频感受,使得每种音频数据听起来来自于对应的声源,各个声源可以位于不同的方向,且每种音频数据的音量等多种音频参数也可以分别灵活设置,使得仅通过一个音频播放设备播放空间混音处理后的音频数据就可以实现位于用户不同的方向的多个音频播放设备作为声源同时播放对应的音频数据的效果。
步骤108,向音频播放设备发送第一目标音频数据以进行播放。
第一电子设备可以向耳机发送第一目标音频数据,从而通过耳机播放该第一目标音频数据。例如,第一电子设备的空间状态信息对应于用户的正前方,第二电子设备的空间状态信息对应于用户的左侧,则用户在通过耳机听该第一目标音频数据的情况下,可以听到第一音频数据来自于正前方且音量较大,同时可以听到第二音频数据来自于左侧且音量较小。
可选地,音频播放方法还包括:在第一电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的更换后的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第二目标音频数据,并向音频播放设备发送以进行播放;或者,在每个第二电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的更换后的音频数据进行空间混音处理,获得第三目标音频数据,并向音频播放设备发送以进行播放。
在如图1所示的音频播放方法的实施例中,获取空间状态信息的控制指 令;根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;向音频播放设备发送第一目标音频数据以进行播放。通过本申请实施例的技术方案,能够按照第一电子设备和至少一个第二电子设备分别对应的空间状态信息进行空间混音处理,避免用户在同时听多个位于不同位置的电子设备播放不同的音频数据时产生方位错乱的感觉。
基于相同的技术构思,本申请还提供一种音频播放方法的实施例,如图6所示。图6是本申请一实施例提供的音频播放方法的第二种流程示意图。本实施例中的音频播放设备可以是耳机。
参照图6,步骤602,第一电子设备与第二电子设备相互连接,耳机连接第一电子设备。
步骤604,第二电子设备传输音频数据到第一电子设备。
步骤606,用户根据设备实际位置状态或者喜好设置空间位置信息。
步骤608,用户设置不同场景的音量大小,用户可配置多种音量状态。
步骤610,对第一电子设备与第二电子设备播放的音频数据进行空间混音处理,输出到耳机。
在步骤610之后,可以执行步骤612、步骤618以及步骤620中的至少一项。
步骤612,检测瞳孔注视点。
步骤614,目标电子设备是否切换。
若是,则返回执行步骤608;若否,则执行步骤616。
步骤616,维持原输出状态。
步骤618,判断音频数据是否更换。
若是,则返回执行步骤606;若否,则执行步骤616。
步骤620,判断音量状态是否切换。
若是,则返回执行步骤608;若否,则执行步骤616。
如图6所示的实施例提供的音频播放方法能够实现前述的音频播放方法实施例实现的各个过程,为避免重复,这里不再赘述。
需要说明的是,本申请实施例提供的音频播放方法,执行主体可以为音频播放装置,或者该音频播放装置中的用于执行音频播放的方法的控制模块。 本申请实施例中以音频播放装置执行音频播放的方法为例,说明本申请实施例提供的音频播放装置。
图7是本申请一实施例提供的音频播放装置的结构示意图。
参照图7所示,音频播放装置,应用于第一电子设备,第一电子设备与至少一个第二电子设备连接,第一电子设备与音频播放设备连接,该音频播放装置包括:
获取模块701,用于获取空间状态信息的控制指令;
确定模块702,用于根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;
处理模块703,用于根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;
发送模块704,用于向音频播放设备发送第一目标音频数据以进行播放。
可选地,第一电子设备为主音设备;每个第二电子设备为辅音设备;获取模块,具体用于:
获取第一主辅切换指令;第一主辅切换指令用于改变第一电子设备和每个第二电子设备的空间状态信息,使得被选中的第二电子设备的空间状态信息对应于用户的正前方,第一电子设备的空间状态信息对应于用户的非正前方;
或者,至少一个第二电子设备包括目标第二电子设备;目标第二电子设备为主音设备,第一电子设备为辅音设备;获取模块,具体用于:
获取第二主辅切换指令;第二切换指令用于改变第一电子设备和目标第二电子设备的空间状态信息,使得第一电子设备的空间状态信息对应于用户的正前方,目标第二电子设备的空间状态信息对应于用户的非正前方。
可选地,获取模块701,具体用于:
获取第一电子设备针对瞳孔注视点的第一检测结果和每个第二电子设备针对瞳孔注视点的第二检测结果。
可选地,确定模块702,包括:
第一确定单元,用于根据第一检测结果和至少一个第二检测结果,确定被瞳孔注视的目标位置;
第二确定单元,用于根据目标位置,确定第一电子设备和每个第二电子 设备的空间状态信息。
可选地,第二确定单元,具体用于:
根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
根据目标位置,确定目标电子设备的空间状态信息;
根据预设的空间状态信息集合和目标电子设备的空间状态信息,确定每个非目标电子设备的空间状态信息。
可选地,第二确定单元,具体用于:
根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
根据目标电子设备,从预设的多个备选空间状态信息组合中确定对应的空间状态信息组合;目标电子设备对应的空间状态信息组合包括目标电子设备的空间状态信息和每个非目标电子设备的空间状态信息。
可选地,音频播放装置还包括:
混音模块,用于在第一电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的更换后的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第二目标音频数据,并向音频播放设备发送以进行播放;或者,在每个第二电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的更换后的音频数据进行空间混音处理,获得第三目标音频数据,并向音频播放设备发送以进行播放。
可选地,空间状态信息包括空间位置信息和音频参数,获取模块701,包括:
第一接收单元,用于接收针对空间位置信息的第一控制指令;
和/或,
第二接收单元,用于接收针对音频参数的第二控制指令。
可选地,第一接收单元,具体用于:
在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,将电子设备对应的虚拟声源在预设球体中的位置信息确定为电子设备的空间位置信息;
接收虚拟声源的位置调节指令;位置调节指令用于调节虚拟声源在预设 球体中的位置信息。
可选地,音频参数包括音量信息;第二接收单元,具体用于:
在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,接收针对电子设备的音量调节指令。
本申请实施例所提供的音频播放装置,获取空间状态信息的控制指令;根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;向音频播放设备发送第一目标音频数据以进行播放。通过本申请实施例的技术方案,能够按照第一电子设备和至少一个第二电子设备分别对应的空间状态信息进行空间混音处理,避免用户在同时听多个位于不同位置的电子设备播放不同的音频数据时产生方位错乱的感觉。
本申请实施例中的音频播放装置可以是装置,也可以是终端中的部件、集成电路、或芯片。该装置可以是移动电子设备,也可以为非移动电子设备。示例性的,移动电子设备可以为手机、平板电脑、笔记本电脑、掌上电脑、车载电子设备、可穿戴设备、超级移动个人计算机(ultra-mobile personal computer,UMPC)、上网本或者个人数字助理(personal digital assistant,PDA)等,非移动电子设备可以为服务器、网络附属存储器(Network Attached Storage,NAS)、个人计算机(personal computer,PC)、电视机(television,TV)、柜员机或者自助机等,本申请实施例不作具体限定。
本申请实施例中的音频播放装置可以为具有操作系统的装置。该操作系统可以为安卓(Android)操作系统,可以为ios操作系统,还可以为其他可能的操作系统,本申请实施例不作具体限定。
本申请实施例提供的音频播放装置能够实现前述的音频播放方法实施例实现的各个过程,为避免重复,这里不再赘述。
可选地,如图8所示,本申请实施例还提供一种电子设备800,包括处理器801,存储器802,存储在存储器802上并可在处理器801上运行的程序或指令,该程序或指令被处理器801执行时实现上述音频播放方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。
需要说明的是,本申请实施例中的电子设备包括上述的移动电子设备和非移动电子设备。
图9是实现本申请实施例的一种电子设备的硬件结构示意图。
该电子设备900包括但不限于:射频单元901、网络模块902、音频输出单元903、输入单元904、传感器905、显示单元906、用户输入单元907、接口单元908、存储器909、以及处理器910等部件。
本领域技术人员可以理解,电子设备900还可以包括给各个部件供电的电源(比如电池),电源可以通过电源管理系统与处理器910逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。图9中示出的电子设备结构并不构成对电子设备的限定,电子设备可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置,在此不再赘述。
其中,处理器910,用于获取空间状态信息的控制指令;
根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;
根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;
向音频播放设备发送第一目标音频数据以进行播放。
在本申请实施例中,获取空间状态信息的控制指令;根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息;根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;向音频播放设备发送第一目标音频数据以进行播放。通过本申请实施例的技术方案,能够按照第一电子设备和至少一个第二电子设备分别对应的空间状态信息进行空间混音处理,避免用户在同时听多个位于不同位置的电子设备播放不同的音频数据时产生方位错乱的感觉。
可选地,第一电子设备为位于用户的正前方的主音设备;每个第二电子设备为位于用户的非正前方的辅音设备;处理器910,用于:
获取空间状态信息的控制指令,包括:
获取第一主辅切换指令;第一主辅切换指令用于改变第一电子设备和每个第二电子设备的空间状态信息,使得被选中的第二电子设备的空间状态信息对应于用户的正前方,第一电子设备的空间状态信息对应于用户的非正前方;
或者,至少一个第二电子设备包括目标第二电子设备;目标第二电子设备为主音设备,第一电子设备为辅音设备;获取空间状态信息的控制指令, 包括:
获取第二主辅切换指令;第二切换指令用于改变第一电子设备和目标第二电子设备的空间状态信息,使得第一电子设备的空间状态信息对应于用户的正前方,目标第二电子设备的空间状态信息对应于用户的非正前方。
可选地,处理器910,还用于:
获取空间状态信息的控制指令,包括:
获取第一电子设备针对瞳孔注视点的第一检测结果和每个第二电子设备针对瞳孔注视点的第二检测结果。
可选地,处理器910,还用于:
根据控制指令,确定第一电子设备和每个第二电子设备的空间状态信息,包括:
根据第一检测结果和至少一个第二检测结果,确定被瞳孔注视的目标位置;
根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息。
可选地,处理器910,还用于:
根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息,包括:
根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
根据目标位置,确定目标电子设备的空间状态信息;
根据预设的空间状态信息集合和目标电子设备的空间状态信息,确定每个非目标电子设备的空间状态信息。
可选地,处理器910,还用于:
根据目标位置,确定第一电子设备和每个第二电子设备的空间状态信息,包括:
根据目标位置,在第一电子设备和至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
根据目标电子设备,从预设的多个备选空间状态信息组合中确定对应的空间状态信息组合;目标电子设备对应的空间状态信息组合包括目标电子设备的空间状态信息和每个非目标电子设备的空间状态信息。
可选地,处理器910,还用于:
在第一电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的更换后的音频数据和每个第二电子设备播放的音频数据进行空间混音处理,获得第二目标音频数据,并向音频播放设备发送以进行播放;
或者,
在每个第二电子设备播放的音频数据被更换的情况下,根据每个空间状态信息,对第一电子设备播放的音频数据和每个第二电子设备播放的更换后的音频数据进行空间混音处理,获得第三目标音频数据,并向音频播放设备发送以进行播放。
可选地,处理器910,还用于:
空间状态信息包括空间位置信息和音频参数,获取空间状态信息的控制指令,包括:
接收针对空间位置信息的第一控制指令;
和/或,
接收针对音频参数的第二控制指令。
可选地,处理器910,还用于:
接收针对空间位置信息的第一控制指令,包括:
在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,将电子设备对应的虚拟声源在预设球体中的位置信息确定为电子设备的空间位置信息;
接收虚拟声源的位置调节指令;位置调节指令用于调节虚拟声源在预设球体中的位置信息。
可选地,音频参数包括音量信息;处理器910,还用于:
接收针对音频参数的第二控制指令,包括:
在第一电子设备和至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,接收针对电子设备的音量调节指令。
在本申请实施例中,通过第一主辅切换指令和第二主辅切换指令,可以灵活地在第一电子设备和至少一个第二电子设备中进行主音设备和辅音设备的切换;根据检测瞳孔注视点,可以自由灵活地控制空间位置信息和空间音量信息,在用户自然转移视线过程中为用户提供最适合用户的音频播放效果;通过在用户交互界面上接收虚拟声源的位置调节指令和针对电子设备的音量 调节指令,可以灵活地设置各个电子设备的空间状态信息,丰富听觉效果。
应理解的是,本申请实施例中,输入单元904可以包括图形处理器(Graphics Processing Unit,GPU)9041和麦克风9042,图形处理器9041对在视频捕获模式或图像捕获模式中由图像捕获装置(如摄像头)获得的静态图片或视频的图像数据进行处理。显示单元906可包括显示面板9061,可以采用液晶显示器、有机发光二极管等形式来配置显示面板9061。用户输入单元907包括触控面板9071以及其他输入设备9072。触控面板9071,也称为触摸屏。触控面板9071可包括触摸检测装置和触摸控制器两个部分。其他输入设备9072可以包括但不限于物理键盘、功能键(比如音量控制按键、开关按键等)、轨迹球、鼠标、操作杆,在此不再赘述。存储器909可用于存储软件程序以及各种数据,包括但不限于应用程序和操作系统。处理器910可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器910中。
本申请实施例还提供一种可读存储介质,所述可读存储介质上存储有程序或指令,该程序或指令被处理器执行时实现上述音频播放方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。
其中,所述处理器为上述实施例中所述的电子设备中的处理器。所述可读存储介质,包括计算机可读存储介质,如计算机只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等。
本申请实施例另提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现上述音频播放方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。
应理解,本申请实施例提到的芯片还可以称为系统级芯片、系统芯片、芯片系统或片上系统芯片等。
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情 况下,由语句“包括一个、第二电子设备2、…”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。此外,需要指出的是,本申请实施方式中的方法和装置的范围不限按示出或讨论的顺序来执行功能,还可包括根据所涉及的功能按基本同时的方式或按相反的顺序来执行功能,例如,可以按不同于所描述的次序来执行所描述的方法,并且还可以添加、省去、或组合各种步骤。另外,参照某些示例所描述的特征可在其他示例中被组合。
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以计算机软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端(可以是手机,计算机,服务器,或者网络设备等)执行本申请各个实施例所述的方法。
上面结合附图对本申请的实施例进行了描述,但是本申请并不局限于上述的具体实施方式,上述的具体实施方式仅仅是示意性的,而不是限制性的,本领域的普通技术人员在本申请的启示下,在不脱离本申请宗旨和权利要求所保护的范围情况下,还可做出很多形式,均属于本申请的保护之内。

Claims (14)

  1. 一种音频播放方法,应用于第一电子设备,所述第一电子设备与至少一个第二电子设备连接,所述第一电子设备与音频播放设备连接,包括:
    获取空间状态信息的控制指令;
    根据所述控制指令,确定所述第一电子设备和每个所述第二电子设备的空间状态信息;
    根据每个所述空间状态信息,对所述第一电子设备播放的音频数据和每个所述第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;
    向所述音频播放设备发送第一目标音频数据以进行播放。
  2. 根据权利要求1所述的方法,其中,所述第一电子设备为位于用户的正前方的主音设备;每个所述第二电子设备为位于所述用户的非正前方的辅音设备;所述获取空间状态信息的控制指令,包括:
    获取第一主辅切换指令;所述第一主辅切换指令用于改变所述第一电子设备和每个所述第二电子设备的空间状态信息,使得被选中的第二电子设备的空间状态信息对应于所述用户的正前方,所述第一电子设备的空间状态信息对应于所述用户的非正前方;
    或者,所述至少一个第二电子设备包括目标第二电子设备;所述目标第二电子设备为所述主音设备,所述第一电子设备为所述辅音设备;所述获取空间状态信息的控制指令,包括:
    获取第二主辅切换指令;所述第二主辅切换指令用于改变所述第一电子设备和所述目标第二电子设备的空间状态信息,使得所述第一电子设备的空间状态信息对应于所述用户的正前方,所述目标第二电子设备的空间状态信息对应于所述用户的非正前方。
  3. 根据权利要求1所述的方法,其中,所述获取空间状态信息的控制指令,包括:
    获取所述第一电子设备针对瞳孔注视点的第一检测结果和每个所述第二电子设备针对所述瞳孔注视点的第二检测结果。
  4. 根据权利要求3所述的方法,其中,所述根据所述控制指令,确定所述第一电子设备和每个所述第二电子设备的空间状态信息,包括:
    根据所述第一检测结果和至少一个所述第二检测结果,确定被瞳孔注视 的目标位置;
    根据所述目标位置,确定所述第一电子设备和每个所述第二电子设备的空间状态信息。
  5. 根据权利要求4所述的方法,其中,所述根据所述目标位置,确定所述第一电子设备和每个所述第二电子设备的空间状态信息,包括:
    根据所述目标位置,在所述第一电子设备和所述至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
    根据所述目标位置,确定所述目标电子设备的空间状态信息;
    根据预设的空间状态信息集合和所述目标电子设备的空间状态信息,确定每个所述非目标电子设备的空间状态信息。
  6. 根据权利要求4所述的方法,其中,所述根据所述目标位置,确定所述第一电子设备和每个所述第二电子设备的空间状态信息,包括:
    根据所述目标位置,在所述第一电子设备和所述至少一个第二电子设备中确定被瞳孔注视的目标电子设备和至少一个未被注视的非目标电子设备;
    根据所述目标电子设备,从预设的多个备选空间状态信息组合中确定对应的空间状态信息组合;所述目标电子设备对应的空间状态信息组合包括所述目标电子设备的空间状态信息和每个所述非目标电子设备的空间状态信息。
  7. 根据权利要求1所述的方法,其中,还包括:
    在所述第一电子设备播放的音频数据被更换的情况下,根据每个所述空间状态信息,对所述第一电子设备播放的更换后的音频数据和每个所述第二电子设备播放的音频数据进行空间混音处理,获得第二目标音频数据,并向所述音频播放设备发送以进行播放;
    或者,
    在每个所述第二电子设备播放的音频数据被更换的情况下,根据每个所述空间状态信息,对所述第一电子设备播放的音频数据和每个所述第二电子设备播放的更换后的音频数据进行空间混音处理,获得第三目标音频数据,并向所述音频播放设备发送以进行播放。
  8. 根据权利要求1所述的方法,其中,所述空间状态信息包括所述空间位置信息和音频参数,所述获取空间状态信息的控制指令,包括:
    接收针对所述空间位置信息的第一控制指令;
    和/或,
    接收针对所述音频参数的第二控制指令。
  9. 根据权利要求8所述的方法,其中,所述接收针对所述空间位置信息的第一控制指令,包括:
    在所述第一电子设备和所述至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,将所述电子设备对应的虚拟声源在预设球体中的位置信息确定为所述电子设备的空间位置信息;
    接收所述虚拟声源的位置调节指令;所述位置调节指令用于调节所述虚拟声源在所述预设球体中的位置信息。
  10. 根据权利要求8所述的方法,其中,所述音频参数包括音量信息;所述接收针对所述音频参数的第二控制指令,包括:
    在所述第一电子设备和所述至少一个第二电子设备中,针对任意一个电子设备,在用户交互界面上,接收针对所述电子设备的音量调节指令。
  11. 一种音频播放装置,应用于第一电子设备,所述第一电子设备与至少一个第二电子设备连接,所述第一电子设备与音频播放设备连接,包括:
    获取模块,用于获取空间状态信息的控制指令;
    确定模块,用于根据所述控制指令,确定所述第一电子设备和每个所述第二电子设备的空间状态信息;
    处理模块,用于根据每个所述空间状态信息,对所述第一电子设备播放的音频数据和每个所述第二电子设备播放的音频数据进行空间混音处理,获得第一目标音频数据;
    发送模块,用于向所述音频播放设备发送第一目标音频数据以进行播放。
  12. 一种电子设备,包括处理器,存储器及存储在所述存储器上并可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如权利要求1-10任一项所述的音频播放方法的步骤。
  13. 一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如权利要求1-10任一项所述的音频播放方法的步骤。
  14. 一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如权利要求1-10任一项所述的音频播放方法。
PCT/CN2023/079874 2022-03-07 2023-03-06 音频播放方法和电子设备 WO2023169367A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210225832.8 2022-03-07
CN202210225832.8A CN114650496A (zh) 2022-03-07 2022-03-07 音频播放方法和电子设备

Publications (1)

Publication Number Publication Date
WO2023169367A1 true WO2023169367A1 (zh) 2023-09-14

Family

ID=81993315

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/079874 WO2023169367A1 (zh) 2022-03-07 2023-03-06 音频播放方法和电子设备

Country Status (2)

Country Link
CN (1) CN114650496A (zh)
WO (1) WO2023169367A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080008342A1 (en) * 2006-07-07 2008-01-10 Harris Corporation Method and apparatus for creating a multi-dimensional communication space for use in a binaural audio system
US20140285312A1 (en) * 2013-03-19 2014-09-25 Nokia Corporation Audio Mixing Based Upon Playing Device Location
CN112581932A (zh) * 2020-11-26 2021-03-30 交通运输部南海航海保障中心广州通信中心 一种基于dsp的有线无线混音系统
US20210176585A1 (en) * 2019-12-04 2021-06-10 Roland Corporation Headphone
CN113823250A (zh) * 2021-11-25 2021-12-21 广州酷狗计算机科技有限公司 音频播放方法、装置、终端及存储介质
CN113890932A (zh) * 2020-07-02 2022-01-04 华为技术有限公司 一种音频控制方法、系统及电子设备
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080008342A1 (en) * 2006-07-07 2008-01-10 Harris Corporation Method and apparatus for creating a multi-dimensional communication space for use in a binaural audio system
US20140285312A1 (en) * 2013-03-19 2014-09-25 Nokia Corporation Audio Mixing Based Upon Playing Device Location
US20210176585A1 (en) * 2019-12-04 2021-06-10 Roland Corporation Headphone
CN113890932A (zh) * 2020-07-02 2022-01-04 华为技术有限公司 一种音频控制方法、系统及电子设备
CN112581932A (zh) * 2020-11-26 2021-03-30 交通运输部南海航海保障中心广州通信中心 一种基于dsp的有线无线混音系统
CN113823250A (zh) * 2021-11-25 2021-12-21 广州酷狗计算机科技有限公司 音频播放方法、装置、终端及存储介质
CN114650496A (zh) * 2022-03-07 2022-06-21 维沃移动通信有限公司 音频播放方法和电子设备

Also Published As

Publication number Publication date
CN114650496A (zh) 2022-06-21

Similar Documents

Publication Publication Date Title
RU2666966C2 (ru) Способ и прибор управления для воспроизведения аудио
CN109660817B (zh) 视频直播方法、装置及系统
WO2017173793A1 (zh) 一种视频投屏方法及装置
CN112237012B (zh) 用于控制多视点全方位内容中的音频的装置及方法
WO2017193540A1 (zh) 弹幕播放方法、播放装置及播放系统
JP2023529868A (ja) 共有方法、装置及び電子機器
US20220070599A1 (en) Method for projecting screen, display device, screen projection terminal, and storage medium
WO2022156662A1 (zh) 音频播放方式的切换方法、装置、电子设备和存储介质
WO2023011473A1 (zh) 声道切换方法、装置和电子设备
CN111866539A (zh) 直播界面的切换方法、装置、终端及存储介质
CN107896337B (zh) 信息推广方法、装置及存储介质
WO2023169367A1 (zh) 音频播放方法和电子设备
KR20130056404A (ko) 휴대 단말기의 이어폰 인식 방법 및 장치
CN107147957A (zh) 视频播放方法和装置
TW201448604A (zh) 利用影音程式切換顯示畫面上前景區與後景區播音之方法
CN112291615A (zh) 音频输出方法、音频输出装置
CN112673651A (zh) 多视点多用户音频用户体验
KR20150136314A (ko) 디스플레이 장치, 사용자 단말 장치, 서버 및 그 제어 방법
WO2023246166A1 (zh) 一种视频进度的调节方法、装置、计算机设备及存储介质
WO2023030292A1 (zh) 多媒体文件的播放方法和装置
US20230024761A1 (en) Method for playing videos and electronic device
WO2022247784A1 (zh) 控制方法、装置、电子设备和可读存储介质
US20230046440A1 (en) Video playback method and device
CA2985731A1 (en) Voice/manual activated and integrated audio/video multi-media, multi-in terface system
US10110943B2 (en) Flexible output of streaming media

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23765945

Country of ref document: EP

Kind code of ref document: A1