CN113068056A - Audio playing method and device, electronic equipment and computer readable storage medium - Google Patents

Audio playing method and device, electronic equipment and computer readable storage medium Download PDF

Info

Publication number
CN113068056A
CN113068056A CN202110291949.1A CN202110291949A CN113068056A CN 113068056 A CN113068056 A CN 113068056A CN 202110291949 A CN202110291949 A CN 202110291949A CN 113068056 A CN113068056 A CN 113068056A
Authority
CN
China
Prior art keywords
data
output
audio
playback
audio data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110291949.1A
Other languages
Chinese (zh)
Other versions
CN113068056B (en
Inventor
杨钧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Huya Technology Co Ltd
Original Assignee
Guangzhou Huya Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Huya Technology Co Ltd filed Critical Guangzhou Huya Technology Co Ltd
Priority to CN202110291949.1A priority Critical patent/CN113068056B/en
Publication of CN113068056A publication Critical patent/CN113068056A/en
Application granted granted Critical
Publication of CN113068056B publication Critical patent/CN113068056B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4398Processing of audio elementary streams involving reformatting operations of audio signals

Abstract

The embodiment of the invention provides an audio playing method, an audio playing device, electronic equipment and a computer readable storage medium, and relates to the technical field of Internet, wherein the method comprises the following steps: the method comprises the steps of receiving at least one piece of audio data, dividing the at least one piece of audio data into more than two pieces of data to be output, and further respectively transmitting the more than two pieces of data to be output to different playback devices for playback, so that the flexibility and the effect of audio playing are improved.

Description

Audio playing method and device, electronic equipment and computer readable storage medium
Technical Field
The invention relates to the technical field of internet, in particular to an audio playing method, an audio playing device, electronic equipment and a computer readable storage medium.
Background
Audio playback is required in many scenarios, for example, in the live broadcast field, audio data on the anchor side is acquired and audio playback is required on the viewer side watching the live broadcast. Nowadays, an audio signal stream is received mainly via a network or a bus and reproduced at a loudspeaker of an audio playback device. Research shows that the audio playback mode has the advantages that the audio playing flexibility and the audio playing effect are improved.
Disclosure of Invention
One of the objects of the present invention includes, for example, providing an audio playing method, apparatus, electronic device and computer readable storage medium to at least partially enhance the flexibility and effect of audio playing.
Embodiments of the invention may be implemented as follows:
in a first aspect, the present invention provides an audio playing method, including:
receiving at least one audio data;
dividing at least one audio data into more than two data to be output;
and respectively transmitting more than two kinds of data to be output to different playback devices for playback.
In an optional embodiment, the step of dividing at least one of the audio data into two or more data to be output includes:
under the condition that the audio data comprise background sound and human sound, taking the background sound in the audio data as data to be output, and taking the human sound in the audio data as another data to be output; alternatively, the first and second electrodes may be,
and under the condition that the audio data come from more than two channels, dividing the audio data into at least two kinds of data to be output according to the channels.
In an optional implementation manner, the received audio data is two or more, and the step of dividing at least one audio data into two or more data to be output includes:
finding out objects corresponding to the audio data, and dividing all the audio data corresponding to the same object into the same data to be output; alternatively, the first and second electrodes may be,
and searching the sound source position of each audio data, and dividing all the audio data with the same sound source position into the same data to be output.
In an alternative embodiment, the step of transferring two or more types of the data to be output to different playback devices for playback respectively includes:
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling a loudspeaker with a set direction in the playback devices to play back; alternatively, the first and second electrodes may be,
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling the playback devices to synchronously play back; alternatively, the first and second electrodes may be,
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling each playback device to play back according to set delay time; alternatively, the first and second electrodes may be,
and respectively transmitting more than two kinds of data to be output to the playback devices at the positions for playback based on the pre-stored positions of the playback devices.
In an optional embodiment, the audio data carries a tag, where the tag is used to identify an object, a sound source type, a sound source channel, and a sound source position corresponding to the audio data;
the step of transferring two or more kinds of the data to be output to different playback devices, respectively, includes:
and respectively transmitting the data to be output to different playback devices according to a set rule based on the label.
In an optional embodiment, the audio data carries a calibrated acquisition time stamp; the step of transferring more than two kinds of the data to be output to different playback devices respectively for playback comprises:
and determining the playback time corresponding to each audio data based on the acquisition time stamp of each audio data, and controlling each playback device to play the corresponding data to be output at the corresponding playback time.
In an alternative embodiment, the received audio data is from at least one group of audio collecting devices, and the step of respectively transferring two or more types of the data to be output to different playback devices for playback includes:
determining a target scene corresponding to the data to be output according to the audio acquisition equipment from which the data to be output comes;
finding out all target playback devices corresponding to the target scene;
respectively transmitting each data to be output corresponding to the target scene to different target playback devices for playback;
wherein each of the target playback devices is a heterogeneous device.
In a second aspect, an embodiment of the present invention provides an audio playing apparatus, including:
the information receiving module is used for receiving at least one piece of audio data;
and the information processing module is used for dividing at least one piece of audio data into more than two pieces of data to be output and respectively transmitting the more than two pieces of data to be output to different playback devices for playback.
In a third aspect, the present invention provides an electronic device comprising: a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the audio playing method according to any of the preceding embodiments when executing the program.
In a fourth aspect, the present invention provides a computer-readable storage medium, where the computer-readable storage medium includes a computer program, and the computer program controls, when running, an electronic device where the computer-readable storage medium is located to execute the audio playing method according to any of the foregoing embodiments.
The beneficial effects of the embodiment of the invention include, for example: compared with the prior art that received audio data are played back in one audio playing device, the received audio data are divided into more than two data to be output in the embodiment of the invention, and the more than two data to be output are transmitted to different playing devices for playing back, so that the possibility of audio playing is increased, and the flexibility and the effect of audio playing are improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and for those skilled in the art, other related drawings can be obtained according to the drawings without inventive efforts.
Fig. 1 shows an exemplary structural block diagram of an electronic device according to an embodiment of the present invention.
Fig. 2 is a schematic flowchart illustrating an audio playing method according to an embodiment of the present invention.
Fig. 3 is a schematic diagram illustrating an exemplary application scenario provided by an embodiment of the present invention.
Fig. 4 shows an exemplary structural block diagram of an audio playing apparatus according to an embodiment of the present invention.
Icon: 100-an electronic device; 110-a memory; 120-a processor; 130-a communication module; 140-an audio playing device; 141-an information receiving module; 142-information processing module.
Detailed Description
In many cases, audio playback is required, for example, in a scene, one or more users are making voices, and in order for listeners to listen to the voices in the scene, audio data in the scene needs to be collected and played back on the listener side.
Nowadays, after a listener initiates a playback request through a terminal such as a mobile phone, the collected audio data is generally transmitted to the mobile phone of the listener for playing, so as to realize audio playback, and the flexibility is limited, and the actual voice effect of the scene where the sound source is located cannot be 'stereoscopically' restored.
Based on the above research, embodiments of the present invention provide an audio playing scheme, in which one or more pieces of received audio data are divided into more than two pieces of data to be output, and the more than two pieces of data to be output are transmitted to different playback devices on the listener side for playback, so that the more than two playback devices cooperatively reproduce the actual voice of the scene where the sound source is located, thereby increasing the possibility of audio playing and improving the flexibility and effect of audio playing.
The defects existing in the above solutions are the results obtained after the inventor has practiced and studied carefully, so the discovery process of the above problems and the solutions proposed by the embodiments of the present invention below to the above problems should be the contributions of the inventor in the invention process.
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations.
Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It is to be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures.
It should be noted that the features of the embodiments of the present invention may be combined with each other without conflict.
Referring to fig. 1, a block diagram of an electronic device 100 provided in this embodiment is shown, where the electronic device 100 in this embodiment may be a server, a processing device, a processing platform, and the like capable of performing data interaction and processing. For example, the electronic device 100 may be a server in an audio playback system, and may also be a playback device in the audio playback system, which is not limited in this embodiment.
The electronic device 100 includes a memory 110, a processor 120, and a communication module 130. The memory 110, the processor 120 and the communication module 130 are electrically connected to each other directly or indirectly to realize data transmission or interaction. For example, the components may be electrically connected to each other via one or more communication buses or signal lines.
The memory 110 is used to store programs or data. The Memory 110 may be, but is not limited to, a Random Access Memory (RAM), a Read Only Memory (ROM), a Programmable Read-Only Memory (PROM), an Erasable Read-Only Memory (EPROM), an electrically Erasable Read-Only Memory (EEPROM), and the like.
The processor 120 is used to read/write data or programs stored in the memory 110 and perform corresponding functions.
The communication module 130 is configured to establish a communication connection between the electronic device 100 and another communication terminal through the network, and to transmit and receive data through the network.
It should be understood that the structure shown in fig. 1 is only a schematic structural diagram of the electronic device 100, and the electronic device 100 may also include more or fewer components than shown in fig. 1, or have a different configuration than shown in fig. 1. The components shown in fig. 1 may be implemented in hardware, software, or a combination thereof. Exemplarily, in case that the electronic device 100 is a playback device in an audio playback system, the electronic device 100 may further include a plurality of speakers.
Referring to fig. 2, a flowchart of an audio playing method according to an embodiment of the present invention can be executed by the electronic device 100 shown in fig. 1, for example, can be executed by the processor 120 in the electronic device 100. The audio playing method includes S110, S120, and S130.
S110, at least one piece of audio data is received.
And S120, dividing at least one audio data into more than two data to be output.
And S130, respectively transmitting more than two kinds of data to be output to different playback devices for playback.
In this embodiment, the audio data received in S110 may be various. For example, it may be one or more audio data from a target scene. The sound source in the target scene may be various, for example, may include one or more objects, which may be one or a combination of two or more of a person, an animal, a musical instrument, an environment, and the like.
The audio data in the target scene may be captured by an audio capture device. In order to improve comprehensiveness and reliability of the audio data acquired in the target scene, the audio data can be acquired by acquiring sounds in the target scene through one or more groups of audio acquisition devices. For example, if the target scene includes three objects, which are an object a, an object B, and an object C, three groups of audio acquisition devices may be configured, such as a group a audio acquisition device, a group B audio acquisition device, and a group C audio acquisition device, to acquire sounds of the three objects respectively. According to the respective characteristics of the object a, the object B, and the object C, the number of the audio capturing devices included in the three groups of configured audio capturing devices may be the same or different, and the types of the included audio capturing devices may be the same or different. For example, three audio capture devices may be included in group a of audio capture devices, one audio capture device may be included in group B of audio capture devices, and two audio capture devices may be included in group C of audio capture devices.
Based on the diversity of the audio data, S120 may be implemented in various ways, and in one implementation, the single audio data may be divided according to a set rule. For example, in the case where a certain piece of audio data includes a background sound and a human voice, the background sound in the audio data may be regarded as one piece of data to be output, and the human voice in the audio data may be regarded as another piece of data to be output. Illustratively, in a case where a target scene is a certain anchor live broadcast, sound emitted by the anchor is human voice, and sound generated by other people other than the anchor, animals, musical instruments, environment, and the like is Background sound (also called Background Music, BGM). And under the condition that the sound of the target scene is collected to obtain the audio data, the human voice in the audio data is used as one kind of data to be output, and the background sound is used as the other kind of data to be output.
For another example, in the case where the audio data is from two or more channels, the audio data may be divided into at least two kinds of data to be output by channels. For example, the audio data from each channel may be divided into one kind of data to be output, or the audio data from some two or more channels may be divided into one kind of data to be output.
For another example, when the number of the received audio data is two or more, the object corresponding to each audio data may be found, and all the audio data corresponding to the same object may be divided into the same kind of data to be output. For example, in the case where the target scene is a debate scene, the objects in the target scene may include a square, an inverse, a referee, a viewer, and the like. The audio data of the square, the audio data of the opposite side, the audio data of the referee, and the audio data of the audience can be divided into one kind of data to be output, respectively. For example, in a case where each audio capture device is configured to capture audio data of each object, the type of data to be output may be determined based on the audio capture device from which the audio data comes. For another example, each audio data may carry a tag, and the tags carried by the audio data of the same object are the same, so that the type of data to be output may be determined based on the tags.
For another example, the classification of the kind of data to be output may be performed according to the sound source direction. For example, directional microphones may be used to collect sound sources located in directions such as front, left, right, and back of the directional microphone, and determine the type of data to be output according to the directions, for example, audio data of sound sources located in the same direction may be divided into the same type of data to be output, or audio data of sound sources located in more than two directions may be divided into the same type of data to be output.
The above is merely an illustration of S120, and the classification of the types of data to be output may be other. For example, the classification of the kind of data to be output may be performed in a combination of two or more of the manners listed above. Illustratively, the nested classification of categories may be performed in such a way that a parent group contains child groups. For example, for a certain debate scene, the class grouping of the first layer (parent grouping) can be performed according to the object of the square, the inverse, the referee, the audience, etc., and the class grouping of the second layer (child grouping) can be performed according to the sound source type such as the background sound and the human voice. For another example, the category subdivision may be performed again on the basis of the sub-packets, for example, after the category grouping of the second layer is performed according to the sound source type such as background sound and human voice, the category grouping of the third layer may be further performed according to the sound channel.
After more than two kinds of data to be output are obtained in the above manner, according to the actual application scenario, it can be flexibly determined to which playback devices the more than two kinds of data to be output are transmitted.
For example, it is possible to determine to which playback devices two or more types of data to be output are delivered, according to an object to which audio data corresponds, a sound source type, a sound source channel, a sound source orientation, and the like.
In one implementation, the audio data may carry tags, and the tags carried in the audio data of different objects, different sound source types, different sound source channels, different sound source orientations, and the like are different, so that the object, the sound source type, the sound source channel, and the sound source orientation corresponding to the audio data can be obtained based on the tags. The rule that the audio data of each object, the sound source type, the sound source channel, and the sound source azimuth are assigned to each playback device may be configured in advance to obtain a set rule, so that each data to be output may be transmitted to different playback devices according to the pre-configured set rule based on the tag carried in the audio data.
In another implementation, two or more types of data to be output may be delivered to the playback devices located at the respective positions for playback, respectively, based on the pre-stored positions of the respective playback devices.
The setting rule may be flexibly set, and for example, the setting rule may be flexibly configured based on an object corresponding to the audio data, a sound source type, a sound source channel, a sound source direction, and the like. For example, in the case of setting based on the object corresponding to the audio data, it may be set that the audio data of different objects is delivered to different playback devices, or it may be set that the audio data of two or more objects is delivered to the same playback device, and the like, which is not limited in this embodiment.
After determining to which playback devices more than two types of data to be output are transmitted, how to perform playback can be flexibly determined according to actual application scenarios.
For example, after two or more kinds of data to be output are transferred to different playback devices, respectively, the playback devices may be controlled to perform playback in synchronization. Illustratively, data A to be output is delivered to a playback device A, data B to be output is delivered to a playback device B, and the playback device A and the playback device B are controlled to synchronously play back the data A to be output and the data B to be output.
For another example, after two or more types of data to be output are transferred to different playback devices, the playback devices may be controlled to perform playback at a set delay time. Illustratively, data a to be output is delivered to a playback device a, data B to be output is delivered to a playback device B, the playback device B is controlled to play back the data B to be output, and the playback device a is controlled to play back the data a to be output after the playback device B sets a delay time for the playback of the data B to be output.
For another example, after two or more kinds of data to be output are transferred to different playback devices, respectively, a speaker whose orientation is set in the playback device may be controlled to play back. Illustratively, data a to be output may be delivered to the playback apparatus a, data a to be output may be controlled to be played back from a left speaker of the playback apparatus a, data B to be output may be delivered to the playback apparatus B, and data B to be output may be controlled to be played back from a right speaker of the playback apparatus B.
For another example, the audio data may carry calibrated acquisition timestamps, and the playback time corresponding to each audio data is determined based on the acquisition timestamp of each audio data, so as to control each playback device to play corresponding data to be output at the corresponding playback time. Illustratively, data a to be output may be transferred to a playback device a, data B to be output may be transferred to a playback device B, if the interval time between the collection time stamp of the data a to be output and the collection time stamp of the data B to be output is T, then the playback time of the data a to be output may be T1, and the playback time of the data B to be output may be T2, and the playback device a may be controlled to play back the data a to be output at time T1, and the playback device B may be controlled to play back the data B to be output at time T2.
The collecting timestamp may be calibrated by an ntp (network Time protocol) server.
For another example, an audio capture device and a target playback device corresponding to each target scene may be configured in advance, and accordingly, the target scene corresponding to the data to be output may be determined according to the audio capture device from which the data to be output comes, all the target playback devices corresponding to the target scene may be found, and each data to be output corresponding to the target scene may be respectively transferred to different target playback devices for playback.
Where each target playback device may be a heterogeneous device, each heterogeneous device may have multiple speakers. Based on the design, the audio data acquired by the audio acquisition device from the same target scene can be played back from a plurality of target playback devices (for example, speakers based on android, windows and ios, such as mobile phones, tablets and computers, form a surround sound playback system), and the immersive audio hearing experience fitting the target scene is provided.
In order to more clearly illustrate the implementation of the embodiment of the present invention, the following scenario is illustrated as an example.
If a certain target scene is a live debate scene, the objects in the target scene comprise a square, a reverse side, a judge and a spectator. Referring to fig. 3, three sets of audio acquisition devices are disposed in the target scene, wherein a set of audio acquisition devices is used to acquire audio data of a square, B set of audio acquisition devices is used to acquire audio data of a referee and a viewer, and C set of audio acquisition devices is used to acquire audio data of an opposite side. Three sets of audio acquisition devices use the NTP server for calibration of audio data acquisition timestamps.
All electronic devices such as mobile phones, flat panels and sound equipment with loudspeakers in the environment where audiences who watch live broadcast of the target scene are located are used as target playback devices, and the target playback devices build heterogeneous networks through NTP servers, so that the target playback devices form heterogeneous devices.
After audio data (carrying calibrated acquisition timestamps) of the square, the referee, the audience and the other party are acquired by the three groups of audio acquisition devices in real time, the audio data are pushed to the server, and the server groups the audio data from the three groups of audio acquisition devices and sends the audio data to one target playback device.
The target playback device receives the audio data sent by the server, performs processing such as jitter buffering, and obtains the target playback device to which each audio data is to be sent according to a set rule (such as a rule defined by audiences), for example, the audio data from the group a of audio acquisition devices is sent to a target playback device mobile phone, the audio data from the group B of audio acquisition devices is sent to a target playback device tablet, and the audio data from the group C of audio acquisition devices is sent to a target playback device sound.
And controlling the target playback devices to respectively play back according to the acquisition time stamps of the audio data. For example, if the collecting time stamp of the audio data from the group a audio collecting device is t1 time point, the collecting time stamp of the audio data from the group B audio collecting device is t1+ x time point, and the collecting time stamp of the audio data from the group C audio collecting device is t1+ x + y time point, then the target playback device handset is controlled to play the received audio data at t2 time point, the target playback device tablet is controlled to play the received audio data at t2+ x time point, and the target playback device audio is controlled to play the received audio data at t2+ x + y time point. For another example, if the capture time stamp of the audio data from the group a audio capture device is t1 time point, the capture time stamp of the audio data from the group B audio capture device is t1+ x time point, and the capture time stamp of the audio data from the group C audio capture device is t1+ x time point, then the control target playback device handset plays the received audio data at t2 time point, and the control target playback device tablet and the target playback device audio play the received audio data synchronously at t2+ x time point.
Wherein, each target playback device can be arranged back and forth, left and right, around and the like relative to the audience according to the human-computer interaction instruction.
While the above describes an implementation of a target playback device as an electronic device, other implementations are possible. For example, a server may be used as the electronic device, and the server executes a process to determine each target playback device to which audio data is to be delivered, and controls playback of each target playback device. For another example, the electronic device may be independent of the server and the target playback device, which is not limited in this embodiment.
Based on the setting, the time for playing the audio data by each target playback device can be determined according to the acquisition time stamp of each audio data, so that each target playback device can play the audio data at the same time or according to the relative delay time, and the three-dimensional effect of audio playback is improved by skillfully setting the position of each target playback device, and the audio playing requirements of various scenes are flexibly met. For example, after the audio data of the debate scene is replayed according to the mode, the antagonism and the debate rhythm of the debate scene can be effectively restored, and the audience experience is improved.
In order to execute the corresponding steps in the above embodiments and various possible manners, an implementation manner of the audio playing apparatus is given below. Referring to fig. 4, fig. 4 is a functional block diagram of an audio playing apparatus 140 according to an embodiment of the present invention, where the audio playing apparatus 140 can be applied to the electronic device 100 shown in fig. 1. It should be noted that the basic principle and the generated technical effect of the audio playing device 140 provided in the present embodiment are the same as those of the above embodiments, and for the sake of brief description, no part of the present embodiment is mentioned, and reference may be made to the corresponding contents in the above embodiments. The audio playing device 140 includes an information receiving module 141 and an information processing module 142.
The information receiving module 141 is configured to receive at least one piece of audio data.
The information processing module 142 is configured to divide at least one piece of audio data into two or more pieces of data to be output, and transmit the two or more pieces of data to be output to different playback devices for playback.
On the basis, an embodiment of the present invention further provides a computer-readable storage medium, where the computer-readable storage medium includes a computer program, and the computer program controls, when running, the electronic device where the computer-readable storage medium is located to execute the audio playing method.
By adopting the scheme in the embodiment of the invention, the immersive surround sound playback system can be built based on a plurality of existing heterogeneous devices in the environment where the user listening to the audio is located, so that the scene sounds of the target scene, such as grouping countermeasure, game explanation, role interaction and the like in live broadcast, can be played back more truly, and the auditory experience is improved.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus and method can be implemented in other ways. The apparatus embodiments described above are merely illustrative, and for example, the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
In addition, the functional modules in the embodiments of the present invention may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1. An audio playing method, comprising:
receiving at least one audio data;
dividing at least one audio data into more than two data to be output;
and respectively transmitting more than two kinds of data to be output to different playback devices for playback.
2. The audio playing method according to claim 1, wherein the step of dividing at least one of the audio data into two or more data to be outputted comprises:
under the condition that the audio data comprise background sound and human sound, taking the background sound in the audio data as data to be output, and taking the human sound in the audio data as another data to be output; alternatively, the first and second electrodes may be,
and under the condition that the audio data come from more than two channels, dividing the audio data into at least two kinds of data to be output according to the channels.
3. The audio playing method of claim 1, wherein the number of the received audio data is two or more, and the step of dividing at least one of the audio data into two or more data to be output comprises:
finding out objects corresponding to the audio data, and dividing all the audio data corresponding to the same object into the same data to be output; alternatively, the first and second electrodes may be,
and searching the sound source position of each audio data, and dividing all the audio data with the same sound source position into the same data to be output.
4. The audio playing method according to any one of claims 1 to 3, wherein the step of transferring two or more types of the data to be output to different playback devices for playback respectively comprises:
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling a loudspeaker with a set direction in the playback devices to play back; alternatively, the first and second electrodes may be,
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling the playback devices to synchronously play back; alternatively, the first and second electrodes may be,
respectively transmitting more than two kinds of data to be output to different playback devices, and controlling each playback device to play back according to set delay time; alternatively, the first and second electrodes may be,
and respectively transmitting more than two kinds of data to be output to the playback devices at the positions for playback based on the pre-stored positions of the playback devices.
5. The audio playing method according to claim 4, wherein the audio data carries a label, and the label is used to identify an object, a sound source type, a sound source channel, and a sound source direction corresponding to the audio data;
the step of transferring two or more kinds of the data to be output to different playback devices, respectively, includes:
and respectively transmitting the data to be output to different playback devices according to a set rule based on the label.
6. The audio playing method according to claim 1, wherein the audio data carries a calibrated acquisition time stamp; the step of transferring more than two kinds of the data to be output to different playback devices respectively for playback comprises:
and determining the playback time corresponding to each audio data based on the acquisition time stamp of each audio data, and controlling each playback device to play the corresponding data to be output at the corresponding playback time.
7. The audio playing method according to claim 1, wherein the received audio data is from at least one group of audio capturing devices, and the step of respectively transferring two or more types of the data to be output to different playback devices for playback comprises:
determining a target scene corresponding to the data to be output according to the audio acquisition equipment from which the data to be output comes;
finding out all target playback devices corresponding to the target scene;
respectively transmitting each data to be output corresponding to the target scene to different target playback devices for playback;
wherein each of the target playback devices is a heterogeneous device.
8. An audio playback apparatus, comprising:
the information receiving module is used for receiving at least one piece of audio data;
and the information processing module is used for dividing at least one piece of audio data into more than two pieces of data to be output and respectively transmitting the more than two pieces of data to be output to different playback devices for playback.
9. An electronic device, comprising: memory, processor and computer program stored on the memory and executable on the processor, which when executed by the processor implements the audio playback method of any of claims 1 to 7.
10. A computer-readable storage medium, comprising a computer program, which when executed controls an electronic device in which the computer-readable storage medium is located to perform the audio playing method according to any one of claims 1 to 7.
CN202110291949.1A 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium Active CN113068056B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110291949.1A CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110291949.1A CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN113068056A true CN113068056A (en) 2021-07-02
CN113068056B CN113068056B (en) 2023-08-22

Family

ID=76561997

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110291949.1A Active CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN113068056B (en)

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2214062Y (en) * 1994-03-01 1995-11-29 唐徵 Recording and videocording two-purpose machine with eight-track, four-track and two-track stereo
CN101001485A (en) * 2006-10-23 2007-07-18 中国传媒大学 Finite sound source multi-channel sound field system and sound field analogy method
US20110077755A1 (en) * 2009-09-30 2011-03-31 Nortel Networks Limited Method and system for replaying a portion of a multi-party audio interaction
CN104125534A (en) * 2013-07-18 2014-10-29 中国传媒大学 Synchronous multi-channel audio recording and playing method and system
CN104919821A (en) * 2012-09-27 2015-09-16 声摩逊实验室 Method and system for playing back an audio signal
CN109391896A (en) * 2018-10-29 2019-02-26 中国传媒大学 A kind of audio generation method and device
CN110377265A (en) * 2019-06-24 2019-10-25 贵安新区新特电动汽车工业有限公司 Sound playing method and device
CN110827843A (en) * 2018-08-14 2020-02-21 Oppo广东移动通信有限公司 Audio processing method and device, storage medium and electronic equipment
CN111370018A (en) * 2020-02-28 2020-07-03 维沃移动通信有限公司 Audio data processing method, electronic device and medium
CN111757171A (en) * 2020-07-03 2020-10-09 海信视像科技股份有限公司 Display device and audio playing method
CN112312298A (en) * 2020-11-19 2021-02-02 北京小米松果电子有限公司 Audio playing method and device, electronic equipment and storage medium
CN112492506A (en) * 2019-09-11 2021-03-12 深圳市优必选科技股份有限公司 Audio playing method and device, computer readable storage medium and robot

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2214062Y (en) * 1994-03-01 1995-11-29 唐徵 Recording and videocording two-purpose machine with eight-track, four-track and two-track stereo
CN101001485A (en) * 2006-10-23 2007-07-18 中国传媒大学 Finite sound source multi-channel sound field system and sound field analogy method
US20110077755A1 (en) * 2009-09-30 2011-03-31 Nortel Networks Limited Method and system for replaying a portion of a multi-party audio interaction
CN104919821A (en) * 2012-09-27 2015-09-16 声摩逊实验室 Method and system for playing back an audio signal
CN104125534A (en) * 2013-07-18 2014-10-29 中国传媒大学 Synchronous multi-channel audio recording and playing method and system
CN110827843A (en) * 2018-08-14 2020-02-21 Oppo广东移动通信有限公司 Audio processing method and device, storage medium and electronic equipment
CN109391896A (en) * 2018-10-29 2019-02-26 中国传媒大学 A kind of audio generation method and device
CN110377265A (en) * 2019-06-24 2019-10-25 贵安新区新特电动汽车工业有限公司 Sound playing method and device
CN112492506A (en) * 2019-09-11 2021-03-12 深圳市优必选科技股份有限公司 Audio playing method and device, computer readable storage medium and robot
CN111370018A (en) * 2020-02-28 2020-07-03 维沃移动通信有限公司 Audio data processing method, electronic device and medium
CN111757171A (en) * 2020-07-03 2020-10-09 海信视像科技股份有限公司 Display device and audio playing method
CN112312298A (en) * 2020-11-19 2021-02-02 北京小米松果电子有限公司 Audio playing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN113068056B (en) 2023-08-22

Similar Documents

Publication Publication Date Title
US11758329B2 (en) Audio mixing based upon playing device location
CN110264986B (en) Online K song device, method and computer readable storage medium
US20230171458A1 (en) System and Method for Real-Time Synchronization of Media Content via Multiple Devices and Speaker Systems
US20090080632A1 (en) Spatial audio conferencing
CN106790940B (en) Recording method, recording playing method, device and terminal
US9967437B1 (en) Dynamic audio synchronization
CN107301028B (en) Audio data processing method and device based on multi-person remote call
US20170195817A1 (en) Simultaneous Binaural Presentation of Multiple Audio Streams
CN113050916A (en) Audio playing method, device and storage medium
US11044569B2 (en) Audio content format selection
WO2013083133A1 (en) System for multimedia broadcasting
EP1784049A1 (en) A method and system for sound reproduction, and a program product
CN113068056B (en) Audio playing method, device, electronic equipment and computer readable storage medium
US20230362571A1 (en) Information processing device, information processing terminal, information processing method, and program
CN108141693B (en) Signal processing apparatus, signal processing method, and computer-readable storage medium
EP3497939A1 (en) A method and a system for presenting media
CN113905321A (en) Object-based audio channel metadata and generation method, device and storage medium
WO2022208609A1 (en) Distribution system, distribution method, and program
US20230370801A1 (en) Information processing device, information processing terminal, information processing method, and program
CN105260173A (en) File processing method and electronic equipment
CN116057927A (en) Information processing device, information processing terminal, information processing method, and program
KR20240044403A (en) Participational contents processing system and control method thereof
CN117640987A (en) Offline online real-time chorus method, device and medium
WO2023192770A1 (en) Generating media queues by multiple participants
CN115278857A (en) Anti-interference audio data transmission method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant