CN113068056B - Audio playing method, device, electronic equipment and computer readable storage medium - Google Patents

Audio playing method, device, electronic equipment and computer readable storage medium Download PDF

Info

Publication number
CN113068056B
CN113068056B CN202110291949.1A CN202110291949A CN113068056B CN 113068056 B CN113068056 B CN 113068056B CN 202110291949 A CN202110291949 A CN 202110291949A CN 113068056 B CN113068056 B CN 113068056B
Authority
CN
China
Prior art keywords
data
output
audio
replay
audio data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110291949.1A
Other languages
Chinese (zh)
Other versions
CN113068056A (en
Inventor
杨钧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Huya Technology Co Ltd
Original Assignee
Guangzhou Huya Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Huya Technology Co Ltd filed Critical Guangzhou Huya Technology Co Ltd
Priority to CN202110291949.1A priority Critical patent/CN113068056B/en
Publication of CN113068056A publication Critical patent/CN113068056A/en
Application granted granted Critical
Publication of CN113068056B publication Critical patent/CN113068056B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4398Processing of audio elementary streams involving reformatting operations of audio signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Stereophonic System (AREA)

Abstract

The embodiment of the invention provides an audio playing method, an audio playing device, electronic equipment and a computer readable storage medium, relating to the technical field of Internet, wherein the method comprises the following steps: and receiving at least one piece of audio data, dividing the at least one piece of audio data into more than two pieces of data to be output, and further respectively transmitting the more than two pieces of data to be output to different replay equipment for replay, so that the flexibility and the effect of audio playing are improved.

Description

Audio playing method, device, electronic equipment and computer readable storage medium
Technical Field
The present invention relates to the field of internet technologies, and in particular, to an audio playing method, an audio playing device, an electronic device, and a computer readable storage medium.
Background
Audio playback is required in many scenes, for example, in the live field, it is required to collect audio data on the main broadcasting side and to perform audio playback on the viewer side viewing the live. Today, a stream of audio signals is received mainly via a network or bus and played back at the loudspeakers of an audio playback device. According to research, the audio playback mode has the advantage that the audio playback flexibility and effect are required to be improved.
Disclosure of Invention
One of the objects of the present invention includes, for example, providing an audio playback method, apparatus, electronic device, and computer-readable storage medium to at least partially enhance the flexibility and effect of audio playback.
Embodiments of the invention may be implemented as follows:
in a first aspect, the present invention provides an audio playing method, including:
receiving at least one audio data;
dividing at least one audio data into more than two types of data to be output;
and respectively transmitting more than two types of data to be output to different replay devices for replay.
In an alternative embodiment, the step of dividing at least one of the audio data into two or more data to be output includes:
in the case that the audio data comprises background sound and human voice, taking the background sound in the audio data as one data to be output and taking the human voice in the audio data as the other data to be output; or alternatively, the process may be performed,
in case the audio data comes from more than two channels, the audio data is divided per channel into at least two data to be output.
In an alternative embodiment, the received audio data is more than two, and the step of dividing at least one audio data into more than two data to be output includes:
searching out objects corresponding to the audio data, and dividing all the audio data corresponding to the same object into the same data to be output; or alternatively, the process may be performed,
and searching the sound source position of each piece of audio data, and dividing all pieces of audio data with the same sound source position into the same piece of data to be output.
In an alternative embodiment, the step of transferring the two or more types of data to be output to different playback devices for playback includes:
respectively transmitting more than two kinds of data to be output to different replay devices, and controlling a loudspeaker with a set azimuth in the replay devices to replay; or alternatively, the process may be performed,
respectively transmitting more than two kinds of data to be output to different replay devices, and controlling each replay device to replay synchronously; or alternatively, the process may be performed,
respectively transmitting more than two kinds of data to be output to different replay devices, and controlling each replay device to replay according to a set delay time; or alternatively, the process may be performed,
based on the pre-stored positions of the playback devices, more than two types of data to be output are respectively transmitted to the playback devices at the positions for playback.
In an optional embodiment, the audio data carries a tag, and the tag is used for identifying an object, a sound source type, a sound source channel and a sound source azimuth corresponding to the audio data;
the step of transferring the two or more kinds of data to be output to different playback devices respectively includes:
and respectively transmitting the data to be output to different replay devices according to a set rule based on the label.
In an alternative embodiment, the audio data carries a calibrated acquisition time stamp; the step of transferring the two or more data to be output to different replay devices for replay respectively includes:
and determining the corresponding replay time of each audio data based on the acquisition time stamp of each audio data, and controlling each replay device to play the corresponding data to be output at the corresponding replay time.
In an alternative embodiment, the received audio data is from at least one group of audio collection devices, and the step of transferring two or more types of data to be output to different playback devices for playback includes:
determining a target scene corresponding to the data to be output according to audio acquisition equipment from which the data to be output is derived;
searching all target replay devices corresponding to the target scene;
transmitting each data to be output corresponding to the target scene to different target replay devices for replay;
wherein each of the target playback devices is a heterogeneous device.
In a second aspect, an embodiment of the present invention provides an audio playing device, including:
the information receiving module is used for receiving at least one audio data;
and the information processing module is used for dividing at least one piece of audio data into more than two pieces of data to be output, and respectively transmitting the more than two pieces of data to be output to different replay equipment for replay.
In a third aspect, the present invention provides an electronic device comprising: a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the audio playing method of any of the preceding embodiments when the program is executed.
In a fourth aspect, the present invention provides a computer readable storage medium, where the computer readable storage medium includes a computer program, where the computer program controls an electronic device where the computer readable storage medium is located to execute the audio playing method according to any one of the foregoing embodiments.
The beneficial effects of the embodiment of the invention include, for example: compared with the prior art that received audio data is replayed in one audio playing device, in the embodiment of the invention, the received one or more pieces of audio data are divided into more than two pieces of data to be output, and the more than two pieces of data to be output are transmitted to different replay devices for replay, so that the possibility of audio playing is increased, and the flexibility and effect of audio playing are improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the embodiments will be briefly described below, it being understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and other related drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 shows an exemplary block diagram of an electronic device according to an embodiment of the present invention.
Fig. 2 is a schematic flow chart of an audio playing method according to an embodiment of the present invention.
Fig. 3 shows an exemplary application scenario provided by the embodiment of the present invention.
Fig. 4 is a block diagram illustrating an exemplary structure of an audio playing device according to an embodiment of the present invention.
Icon: 100-an electronic device; 110-memory; a 120-processor; 130-a communication module; 140-an audio playing device; 141-an information receiving module; 142-information processing module.
Detailed Description
In many cases, audio playback is required, for example, in a certain scene, where there are one or more users who utter voices, in order for a listener to hear voices in the scene, audio data in the scene needs to be collected and played back on the listener side.
Nowadays, after a listener initiates a playback request through a terminal such as a mobile phone, collected audio data is generally transmitted to the mobile phone of the listener for playing, so that audio playback is realized, flexibility is limited, and an actual voice effect of a scene where a sound source is located cannot be restored in a three-dimensional manner.
Based on the above-mentioned research, the embodiment of the invention provides an audio playing scheme, which divides one or more pieces of received audio data into more than two pieces of data to be output, and transmits the more than two pieces of data to be output to different playback devices at the audience side for playback, so that the more than two playback devices cooperatively restore actual voices of scenes where sound sources are located, further the possibility of audio playing is increased, and the flexibility and effect of audio playing are improved.
The present invention is directed to a method for manufacturing a semiconductor device, and a semiconductor device manufactured by the method.
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the described embodiments are some embodiments of the present invention, but not all embodiments of the present invention. The components of the embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations.
Thus, the following detailed description of the embodiments of the invention, as presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It should be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
It should be noted that: like reference numerals and letters denote like items in the following figures, and thus once an item is defined in one figure, no further definition or explanation thereof is necessary in the following figures.
It should be noted that the features of the embodiments of the present invention may be combined with each other without conflict.
Fig. 1 is a block diagram of an electronic device 100 according to the present embodiment, where the electronic device 100 in the present embodiment may be a server, a processing device, a processing platform, etc. capable of performing data interaction and processing. The electronic device 100 may be a server in an audio playback system or a playback device in an audio playback system, for example, and the present embodiment is not limited thereto.
The electronic device 100 includes a memory 110, a processor 120, and a communication module 130. The memory 110, the processor 120, and the communication module 130 are electrically connected directly or indirectly to each other to realize data transmission or interaction. For example, the components may be electrically connected to each other via one or more communication buses or signal lines.
Wherein the memory 110 is used for storing programs or data. The Memory 110 may be, but is not limited to, random access Memory (Random Access Memory, RAM), read Only Memory (ROM), programmable Read Only Memory (Programmable Read-Only Memory, PROM), erasable Read Only Memory (Erasable Programmable Read-Only Memory, EPROM), electrically erasable Read Only Memory (Electric Erasable Programmable Read-Only Memory, EEPROM), etc.
The processor 120 is used to read/write data or programs stored in the memory 110 and perform corresponding functions.
The communication module 130 is configured to establish a communication connection between the electronic device 100 and other communication terminals through the network, and is configured to transmit and receive data through the network.
It should be understood that the structure shown in fig. 1 is merely a schematic diagram of the structure of the electronic device 100, and that the electronic device 100 may further include more or fewer components than those shown in fig. 1, or have a different configuration than that shown in fig. 1. The components shown in fig. 1 may be implemented in hardware, software, or a combination thereof. For example, where the electronic device 100 is a playback device in an audio playback system, the electronic device 100 may also include multiple speakers.
Referring to fig. 2 in combination, a flowchart of an audio playing method according to an embodiment of the invention may be executed by the electronic device 100 shown in fig. 1, for example, may be executed by the processor 120 in the electronic device 100. The audio playing method includes S110, S120, and S130.
S110, at least one audio data is received.
S120, dividing at least one audio data into more than two data to be output.
And S130, respectively transmitting more than two types of data to be output to different replay devices for replay.
In this embodiment, the audio data received in S110 may be various. For example, one or more audio data from a certain target scene. The sound sources in the target scene may be various, for example, may include one or more objects, which may be one or a combination of two or more of a person, an animal, a musical instrument, an environment, and the like.
The audio data in the target scene may be acquired by an audio acquisition device. In order to improve the comprehensiveness and reliability of the audio data collected in the target scene, the audio data can be obtained by collecting the sound in the target scene through one or more groups of audio collection devices. For example, if the target scene includes three objects, namely, object a, object B, and object C, three sets of audio capturing devices may be configured, for example, a set of audio capturing devices B, and a set of audio capturing devices C, to capture sounds of the three objects, respectively. The number of audio collection devices included in the three configured sets of audio collection devices may be the same or different, and the types of the included audio collection devices may be the same or different, according to the characteristics of each of the object a, the object B, and the object C. For example, three audio capture devices may be included in a group a audio capture device, one audio capture device may be included in a group B audio capture device, and two audio capture devices may be included in a group C audio capture device.
Based on the diversity of the audio data, S120 may be implemented in various ways, and in one implementation, the individual audio data may be divided according to a set rule. For example, in the case where certain audio data includes a background sound and a human voice, the background sound in the audio data may be taken as one type of data to be output, and the human voice in the audio data may be taken as another type of data to be output. Illustratively, in the case where a target scene is live for a certain anchor, the sound made by the anchor is a human voice, and the sound made by other people, animals, musical instruments, environments, and the like other than the anchor is a Background sound (also called Background Music). Under the condition that the sound of the target scene is collected to obtain the audio data, the human voice in the audio data is used as one type of data to be output, and the background sound is used as the other type of data to be output.
For another example, in the case where the audio data is from more than two channels, the audio data may be divided into at least two data to be output per channel. For example, audio data from each channel may be divided into one type of data to be output, or audio data from some two or more channels may be divided into one type of data to be output.
For another example, in the case that the received audio data is more than two, the object corresponding to each audio data may be found, and all the audio data corresponding to the same object are divided into the same data to be output. For example, in the case where the target scene is a dialect scene, the objects in the target scene may include squares, opponents, referees, spectators, and the like. The square audio data, the opposite audio data, the referee audio data, and the audience audio data may be divided into one type of data to be output, respectively. Wherein the identification of the audio data about each object can be flexibly realized, for example, in the case that each audio acquisition device is configured to acquire the audio data of each object, the type of the data to be output can be determined based on the audio acquisition device from which the audio data is derived. For another example, each audio data may carry a tag, and the tags carried by the audio data of the same object are the same, so that the type of data to be output may be determined based on the tags.
For another example, the classification of the data to be output may be performed according to the sound source orientation. For example, directional microphones may be used to collect sound sources located in front of, left of, right of, behind, and the like of the directional microphones, and the type of data to be output may be determined according to the directions, for example, audio data in which the sound sources are located in the same direction may be divided into the same data to be output, or audio data in which the sound sources are located in some two or more directions may be divided into the same data to be output.
The above is merely an illustration of S120, and the classification of the data to be output may be other. For example, the classification of data to be output may be performed in combination of two or more of the above-listed ways. Illustratively, the nested category classification may be performed in such a way that the parent packet contains child packets. For example, for a certain dialectical scene, the category classification of the first layer (parent group) may be performed according to the objects of square, opposite, referee, audience, etc., and the category classification of the second layer (child group) may be performed according to the sound source type such as background sound and human sound. For another example, the classification may be performed again on the basis of the sub-packets, such as the classification of the second layer according to the type of the sound source such as the background sound and the human voice, and then the classification of the third layer according to the sound channel.
After the above-mentioned method is adopted to obtain more than two kinds of data to be output, according to the practical application scene, it can flexibly determine which playback devices the more than two kinds of data to be output are transferred to.
For example, it is possible to determine to which playback devices two or more kinds of data to be output are delivered, based on the object, the sound source type, the sound source channel, the sound source azimuth, and the like to which the audio data corresponds.
In one implementation, the audio data may carry a tag, where the tag carried in the audio data of different objects, different sound source types, different sound source channels, different sound source orientations, etc. is different, so that the object, the sound source type, the sound source channel, and the sound source orientation corresponding to the audio data may be obtained based on the tag. The rule of distributing the audio data of each object, the sound source type, the sound source channel and the sound source direction to each playback device can be configured in advance to obtain a set rule, so that each piece of data to be output can be respectively transmitted to different playback devices according to the preset set rule based on the labels carried in the audio data.
In another implementation, based on the pre-stored locations of the playback devices, two or more types of data to be output may be transferred to the playback devices located at the locations, respectively, for playback.
The setting rule may be flexibly set, for example, may be flexibly configured based on an object, a sound source type, a sound source channel, and a sound source azimuth corresponding to the audio data. For example, in the case of setting based on the object to which the audio data corresponds, the audio data set as different objects may be transferred to different playback apparatuses, or the audio data set as some two or more objects may be transferred to the same playback apparatus, or the like, which is not limited in this embodiment.
After determining to which playback devices two or more kinds of data to be output are transferred, how to perform playback can be flexibly determined according to actual application scenes.
For example, after two or more kinds of data to be output are transferred to different playback apparatuses, respectively, the playback apparatuses may be controlled to perform playback in synchronization. Illustratively, the data a to be output is transferred to the playback device a, the data B to be output is transferred to the playback device B, and the playback device a and the playback device B are controlled to synchronously play back the data a to be output and the data B to be output.
For another example, after two or more kinds of data to be output are transferred to different playback apparatuses, respectively, each playback apparatus may be controlled to perform playback with a set delay time. Illustratively, the data a to be output is transferred to the playback device a, the data B to be output is transferred to the playback device B, the playback device B is controlled to play back the data B to be output, and the playback device a is controlled to play back the data a to be output after the playback device B plays back the data B to be output for a set delay time.
For another example, it is possible to control a speaker set with an azimuth in a playback apparatus to perform playback after two or more kinds of data to be output are transferred to different playback apparatuses, respectively. For example, the data a to be output may be transferred to the playback device a, the data a to be output may be controlled to be played back from the left speaker of the playback device a, the data B to be output may be transferred to the playback device B, and the data B to be output may be controlled to be played back from the right speaker of the playback device B.
For another example, the audio data may carry a calibrated acquisition time stamp, and the playback time corresponding to each audio data is determined based on the acquisition time stamp of each audio data, so as to control each playback device to play the corresponding data to be output at the corresponding playback time. For example, the data a to be output may be transferred to the playback device a, the data B to be output may be transferred to the playback device B, and if the interval time between the collection timestamp of the data a to be output and the collection timestamp of the data B to be output is T, the playback time of the data a to be output may be determined to be T1 according to the collection timestamp of the data a to be output, the collection timestamp of the data B to be output, and the interval time T, and the playback time of the data B to be output is T2, so that the playback device a is controlled to play back the data a to be output at the time point of T1, and the playback device B is controlled to play back the data B to be output at the time point of T2.
Wherein calibration of the acquisition time stamp may be performed by a NTP (Network Time Protocol) server.
For another example, the audio acquisition device and the target playback device corresponding to each target scene may be preconfigured, and accordingly, the target scene corresponding to the data to be output may be determined according to the audio acquisition device from which the data to be output is derived, all the target playback devices corresponding to the target scene are found, and each data to be output corresponding to the target scene is transferred to different target playback devices for playback.
Wherein each target playback device may be a heterogeneous device, each heterogeneous device may have a plurality of speakers. Based on the design, the audio data collected by the audio collection device from the same target scene can be replayed from a plurality of target replay devices (such as a surround sound replay system formed by mobile phones, tablets, computers and other loudspeakers based on android, windows and ios), so that an immersive audio hearing experience fitting the target scene is provided.
In order to more clearly illustrate the implementation of the embodiments of the present invention, the following scenario will be taken as an example.
If a target scene is a live debate scene, the objects in the target scene comprise square, opposite, referees and audiences. Referring to fig. 3, three sets of audio collection devices are set in the target scene, wherein the set a of audio collection devices are used for collecting square audio data, the set B of audio collection devices are used for collecting audio data of referees and audiences, and the set C of audio collection devices are used for collecting opposite audio data. Three sets of audio acquisition devices use the NTP server for calibration of audio data acquisition time stamps.
All electronic devices such as mobile phones, tablets and audios which are positioned in the environment where the audience watching live broadcasting of the target scene is positioned are taken as target replay devices, and each target replay device builds a heterogeneous network through an NTP server, so that each target replay device forms heterogeneous devices.
After the three groups of audio acquisition devices acquire the audio data of the square, referee, audience and opposite party in real time (carrying the calibrated acquisition time stamp), the audio data are pushed to a server, and the server distributes the audio data from the three groups of audio acquisition devices to one of target replay devices in a grouping way.
The target playback device receives the audio data sent by the server, performs processes such as jitter buffering, and obtains the target playback device to which each audio data is to be sent according to a set rule (such as a rule customized by a viewer), for example, sends the audio data from the audio acquisition device of group A to a mobile phone of the target playback device, sends the audio data from the audio acquisition device of group B to a tablet of the target playback device, and sends the audio data from the audio acquisition device of group C to a sound of the target playback device.
And controlling each target replay device to replay respectively according to the acquisition time stamp of each audio data. For example, if the collection time stamp of the audio data from the group a audio collection device is t1 time point, the collection time stamp of the audio data from the group B audio collection device is t1+x time point, and the collection time stamp of the audio data from the group C audio collection device is t1+x+y time point, then the target playback device handset is controlled to play the received audio data at the time point t2, the target playback device tablet is controlled to play the received audio data at the time point t2+x, and the target playback device stereo is controlled to play the received audio data at the time point t2+x+y. For another example, if the collection time stamp of the audio data from the group a audio collection device is t1 time point, the collection time stamp of the audio data from the group B audio collection device is t1+x time point, and the collection time stamp of the audio data from the group C audio collection device is t1+x time point, then the target playback device mobile phone is controlled to play the received audio data at the time point t2, and the target playback device flat panel and the target playback device sound are controlled to synchronously play the received audio data at the time point t2+x.
Wherein each target playback device may be arranged back and forth, left and right, surrounding, etc. with respect to the viewer in accordance with the human-machine interaction indication.
The above has enumerated an implementation in which a certain target playback device is the electronic device, and in other implementations, other implementations are possible. For example, a server may be used as the electronic device, a processing procedure may be performed by the server, each target playback device to which audio data is to be delivered may be determined, and playback by each target playback device may be controlled. For another example, the electronic device may also be independent of the server and the target playback device, which is not limited by the present embodiment.
Based on the setting, the time for playing the audio data by each target playback device can be determined according to the acquisition time stamp of each audio data, so that each target playback device plays the audio data at the same time or according to the relative delay time, and the three-dimensional effect of audio playback is improved by skillfully setting the positions of each target playback device, and the audio playing requirements of various scenes are flexibly met. For example, after the audio data of the debate scene is replayed in the above manner, the antagonism and debate rhythm of the debate scene can be effectively restored, and the audience experience is improved.
In order to perform the respective steps of the above embodiments and the various possible ways, an implementation of an audio playback device is given below. Referring to fig. 4, fig. 4 is a functional block diagram of an audio playing device 140 according to an embodiment of the invention, where the audio playing device 140 can be applied to the electronic apparatus 100 shown in fig. 1. It should be noted that, the basic principle and the technical effects of the audio playing device 140 provided in this embodiment are the same as those of the above embodiment, and for brevity, reference should be made to the corresponding contents of the above embodiment. The audio playing device 140 includes an information receiving module 141 and an information processing module 142.
Wherein the information receiving module 141 is configured to receive at least one audio data.
The information processing module 142 is configured to divide at least one of the audio data into two or more data to be output, and respectively transfer the two or more data to be output to different playback devices for playback.
On the basis of the above, the embodiment of the invention also provides a computer readable storage medium, which comprises a computer program, and when the computer program runs, the electronic equipment where the computer readable storage medium is located is controlled to execute the audio playing method.
By adopting the scheme in the embodiment of the invention, the immersive surround sound playback system can be built based on the existing heterogeneous devices in the environment where the user listening to the audio is located, so that scene sounds of target scenes such as group countermeasure, competition explanation, role interaction and the like in live broadcasting can be played back more truly, and the hearing experience is improved.
In the several embodiments provided in the present invention, it should be understood that the disclosed apparatus and method may be implemented in other manners. The apparatus embodiments described above are merely illustrative, for example, of the flowcharts and block diagrams in the figures that illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
In addition, functional modules in the embodiments of the present invention may be integrated together to form a single part, or each module may exist alone, or two or more modules may be integrated to form a single part.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored in a computer-readable storage medium. Based on this understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art or in a part of the technical solution, in the form of a software product stored in a storage medium, comprising several instructions for causing a computer device (which may be a personal computer, a server, a network device, etc.) to perform all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a magnetic disk, or an optical disk, or other various media capable of storing program codes.
The above description is only of the preferred embodiments of the present invention and is not intended to limit the present invention, but various modifications and variations can be made to the present invention by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (8)

1. An audio playing method, comprising:
receiving at least one audio data;
dividing at least one audio data into more than two types of data to be output;
respectively transmitting more than two kinds of data to be output to different replay devices for replay;
the received audio data come from at least one group of audio acquisition equipment, and the audio data carry calibrated acquisition time stamps; the step of transferring the two or more data to be output to different replay devices for replay respectively includes:
determining a target scene corresponding to data to be output according to audio acquisition equipment from which the data to be output come;
searching all target replay devices corresponding to the target scene, and determining replay time corresponding to each audio data respectively based on the acquisition time stamp of each audio data;
transmitting each data to be output corresponding to the target scene to different target replay devices for replay;
wherein each of the target playback devices is a heterogeneous device; the delivering each of the data to be output corresponding to the target scene to different target playback devices for playback includes: transmitting each piece of data to be output corresponding to the target scene to different target playback devices respectively, and controlling each target playback device to play the corresponding data to be output at the corresponding playback time; the step of respectively transmitting each piece of data to be output corresponding to the target scene to different target playback devices, and controlling each target playback device to play the corresponding data to be output at the corresponding playback time, including: and respectively transmitting more than two types of data to be output to different replay devices, and controlling each replay device to replay according to the set delay time.
2. The audio playing method according to claim 1, wherein the step of dividing at least one of the audio data into two or more data to be output comprises:
in the case that the audio data comprises background sound and human voice, taking the background sound in the audio data as one data to be output and taking the human voice in the audio data as the other data to be output; or alternatively, the process may be performed,
in case the audio data comes from more than two channels, the audio data is divided per channel into at least two data to be output.
3. The audio playing method according to claim 1, wherein the received audio data is two or more, and the step of dividing at least one of the audio data into two or more data to be outputted comprises:
searching out objects corresponding to the audio data, and dividing all the audio data corresponding to the same object into the same data to be output; or alternatively, the process may be performed,
and searching the sound source position of each piece of audio data, and dividing all pieces of audio data with the same sound source position into the same piece of data to be output.
4. The audio playing method according to any one of claims 1 to 3, wherein two or more of said data to be output are respectively transferred to different playback apparatuses, and each of said playback apparatuses is controlled to play back with a set delay time, comprising:
based on the pre-stored positions of the playback devices, more than two types of data to be output are respectively transmitted to the playback devices at the positions for playback.
5. The audio playing method according to claim 4, wherein the audio data carries a tag, and the tag is used for identifying an object, a sound source type, a sound source channel and a sound source azimuth corresponding to the audio data;
the step of transferring the two or more kinds of data to be output to different playback devices respectively includes:
and respectively transmitting the data to be output to different replay devices according to a set rule based on the label.
6. An audio playback apparatus, comprising:
the information receiving module is used for receiving at least one audio data;
the information processing module is used for dividing at least one audio data into more than two types of data to be output, and respectively transmitting the more than two types of data to be output to different replay equipment for replay;
the received audio data come from at least one group of audio acquisition equipment, and the audio data carry calibrated acquisition time stamps; the information processing module is used for respectively transmitting more than two types of data to be output to different replay devices for replay by the following steps:
determining a target scene corresponding to the data to be output according to audio acquisition equipment from which the data to be output is derived;
searching all target replay devices corresponding to the target scene, and determining replay time corresponding to each audio data respectively based on the acquisition time stamp of each audio data;
transmitting each data to be output corresponding to the target scene to different target replay devices for replay;
wherein each of the target playback devices is a heterogeneous device; the delivering each of the data to be output corresponding to the target scene to different target playback devices for playback includes: transmitting each piece of data to be output corresponding to the target scene to different target playback devices respectively, and controlling each target playback device to play the corresponding data to be output at the corresponding playback time; the step of respectively transmitting each piece of data to be output corresponding to the target scene to different target playback devices, and controlling each target playback device to play the corresponding data to be output at the corresponding playback time, including: and respectively transmitting more than two types of data to be output to different replay devices, and controlling each replay device to replay according to the set delay time.
7. An electronic device, comprising: a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the audio playing method of any one of claims 1 to 5 when the program is executed.
8. A computer readable storage medium, characterized in that the computer readable storage medium comprises a computer program, which when run controls an electronic device in which the computer readable storage medium is located to perform the audio playing method according to any one of claims 1 to 5.
CN202110291949.1A 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium Active CN113068056B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110291949.1A CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110291949.1A CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN113068056A CN113068056A (en) 2021-07-02
CN113068056B true CN113068056B (en) 2023-08-22

Family

ID=76561997

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110291949.1A Active CN113068056B (en) 2021-03-18 2021-03-18 Audio playing method, device, electronic equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN113068056B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2214062Y (en) * 1994-03-01 1995-11-29 唐徵 Recording and videocording two-purpose machine with eight-track, four-track and two-track stereo
CN101001485A (en) * 2006-10-23 2007-07-18 中国传媒大学 Finite sound source multi-channel sound field system and sound field analogy method
CN104125534A (en) * 2013-07-18 2014-10-29 中国传媒大学 Synchronous multi-channel audio recording and playing method and system
CN110377265A (en) * 2019-06-24 2019-10-25 贵安新区新特电动汽车工业有限公司 Sound playing method and device
CN110827843A (en) * 2018-08-14 2020-02-21 Oppo广东移动通信有限公司 Audio processing method and device, storage medium and electronic equipment
CN111370018A (en) * 2020-02-28 2020-07-03 维沃移动通信有限公司 Audio data processing method, electronic device and medium
CN111757171A (en) * 2020-07-03 2020-10-09 海信视像科技股份有限公司 Display device and audio playing method
CN112312298A (en) * 2020-11-19 2021-02-02 北京小米松果电子有限公司 Audio playing method and device, electronic equipment and storage medium
CN112492506A (en) * 2019-09-11 2021-03-12 深圳市优必选科技股份有限公司 Audio playing method and device, computer readable storage medium and robot

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8547880B2 (en) * 2009-09-30 2013-10-01 Avaya Inc. Method and system for replaying a portion of a multi-party audio interaction
FR2996094B1 (en) * 2012-09-27 2014-10-17 Sonic Emotion Labs METHOD AND SYSTEM FOR RECOVERING AN AUDIO SIGNAL
CN109391896B (en) * 2018-10-29 2021-05-18 中国传媒大学 Sound effect generation method and device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2214062Y (en) * 1994-03-01 1995-11-29 唐徵 Recording and videocording two-purpose machine with eight-track, four-track and two-track stereo
CN101001485A (en) * 2006-10-23 2007-07-18 中国传媒大学 Finite sound source multi-channel sound field system and sound field analogy method
CN104125534A (en) * 2013-07-18 2014-10-29 中国传媒大学 Synchronous multi-channel audio recording and playing method and system
CN110827843A (en) * 2018-08-14 2020-02-21 Oppo广东移动通信有限公司 Audio processing method and device, storage medium and electronic equipment
CN110377265A (en) * 2019-06-24 2019-10-25 贵安新区新特电动汽车工业有限公司 Sound playing method and device
CN112492506A (en) * 2019-09-11 2021-03-12 深圳市优必选科技股份有限公司 Audio playing method and device, computer readable storage medium and robot
CN111370018A (en) * 2020-02-28 2020-07-03 维沃移动通信有限公司 Audio data processing method, electronic device and medium
CN111757171A (en) * 2020-07-03 2020-10-09 海信视像科技股份有限公司 Display device and audio playing method
CN112312298A (en) * 2020-11-19 2021-02-02 北京小米松果电子有限公司 Audio playing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN113068056A (en) 2021-07-02

Similar Documents

Publication Publication Date Title
CN106464953B (en) Two-channel audio system and method
CN1984310B (en) Method and communication apparatus for reproducing a moving picture
US20230171458A1 (en) System and Method for Real-Time Synchronization of Media Content via Multiple Devices and Speaker Systems
CN110264986B (en) Online K song device, method and computer readable storage medium
CN108616800B (en) Audio playing method and device, storage medium and electronic device
CA2918732C (en) Multiple contact and/or sense promotional systems and methods
US8428236B2 (en) Conference audio system, process for distributing audio signals and computer program
EP1479231A1 (en) Internet broadcasting system and method thereof for personal telecommunication terminal
KR20220077132A (en) Method and system for generating binaural immersive audio for audiovisual content
WO2014053875A1 (en) An apparatus and method for reproducing recorded audio with correct spatial directionality
US11683654B2 (en) Audio content format selection
WO2021028683A1 (en) Media system and method of generating media content
CN108174264B (en) Synchronous lyric display method, system, device, medium and equipment
WO2013083133A1 (en) System for multimedia broadcasting
CN113068056B (en) Audio playing method, device, electronic equipment and computer readable storage medium
CN107147946A (en) A kind of method for processing video frequency and device
EP1784049A1 (en) A method and system for sound reproduction, and a program product
CN107195308B (en) Audio mixing method, device and system of audio and video conference system
US20230370801A1 (en) Information processing device, information processing terminal, information processing method, and program
CN112083379B (en) Audio playing method and device based on sound source localization, projection equipment and medium
US20230362571A1 (en) Information processing device, information processing terminal, information processing method, and program
CN106535060A (en) Pickup control method, audio frequency playing method, and pickup control device
US10728296B1 (en) System, method and apparatus for distributed content dissemination
WO2022208609A1 (en) Distribution system, distribution method, and program
GB2526154A (en) Broadcast media synchronisation system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant