WO2020233263A1 - 音频处理方法及电子设备 - Google Patents

音频处理方法及电子设备 Download PDF

Info

Publication number
WO2020233263A1
WO2020233263A1 PCT/CN2020/083613 CN2020083613W WO2020233263A1 WO 2020233263 A1 WO2020233263 A1 WO 2020233263A1 CN 2020083613 W CN2020083613 W CN 2020083613W WO 2020233263 A1 WO2020233263 A1 WO 2020233263A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
decoding result
output
earphone
speaker
Prior art date
Application number
PCT/CN2020/083613
Other languages
English (en)
French (fr)
Inventor
魏建德
王振菊
Original Assignee
青岛海信传媒网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 青岛海信传媒网络技术有限公司 filed Critical 青岛海信传媒网络技术有限公司
Publication of WO2020233263A1 publication Critical patent/WO2020233263A1/zh
Priority to US17/347,372 priority Critical patent/US11956497B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/60Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
    • H04N5/607Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals for more than one sound signal, e.g. stereo, multilanguages
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute
    • G09B21/001Teaching or communicating with blind persons
    • G09B21/006Teaching or communicating with blind persons using audible presentation of the information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/438Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving encoded video stream packets from an IP network
    • H04N21/4382Demodulation or channel decoding, e.g. QPSK demodulation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/60Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
    • H04N5/602Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals for digital sound signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/04Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/01Input selection or mixing for amplifiers or loudspeakers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/03Connection circuits to selectively connect loudspeakers or headphones to amplifiers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/05Detection of connection of loudspeakers or headphones to amplifiers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/07Applications of wireless loudspeakers or wireless microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2499/00Aspects covered by H04R or H04S not otherwise provided for in their subgroups
    • H04R2499/10General applications
    • H04R2499/15Transducers incorporated in visual displaying devices, e.g. televisions, computer displays, laptops
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/033Headphones for stereophonic communication

Definitions

  • Some embodiments of the present application relate to the field of multimedia technology, and in particular to an audio processing method and electronic device.
  • a description audio is added to the audio data of the DTV service.
  • descriptive audio refers to audio data that analyzes the content of the current video screen.
  • Some embodiments of the present application provide an audio processing method and electronic equipment to meet the needs of ordinary people and blind people watching movies and television programs at the same time.
  • an audio processing method including:
  • the first decoding result corresponding to the main audio and the second decoding result corresponding to the description audio are obtained.
  • the main audio is the audio data of the audio and video file itself
  • the description audio is Audio data describing the content of the video screen in the audio and video file;
  • the first decoding result and the audio mixing result of the first decoding result and the second decoding result are respectively output through different audio output devices.
  • some embodiments of the present application provide an audio processing method, including:
  • Audio and video files where the audio and video files include main channel audio and description audio;
  • control the earphone When there is earphone access, control the earphone to output the main channel audio and the description audio, and control the speaker to output the main channel audio;
  • an electronic device including:
  • Speaker a processor connected to the speaker, the processor is used for:
  • the audio and video file includes main audio and description audio
  • the description audio is audio data describing the content of the video screen in the audio and video file
  • the main audio is only Audio data other than the description audio in the audio and video file
  • control the earphone When there is earphone access, control the earphone to output the main channel audio and the description audio, and control the speaker to output the main channel audio;
  • some embodiments of the present application provide an electronic device, including: a speaker; and a processor connected to the speaker, and the processor is configured to:
  • Audio and video files where the audio and video files include main audio and descriptive audio, the main audio is the audio data of the audio and video file itself, and the descriptive audio is a comparison of the video in the audio and video file Audio data describing the content of the screen;
  • the output of the earphone is controlled according to the main channel audio and the description audio, and the output of the speaker is controlled according to the main channel audio.
  • some embodiments of the present application provide an electronic device, including:
  • the decoder is used to obtain the first decoding result corresponding to the main audio and the second decoding result corresponding to the description audio according to the audio data in the audio and video file, the main audio being the audio data of the audio and video file itself,
  • the description audio is audio data that describes the content of the video screen in the audio and video file;
  • the audio mixer is configured to output the first decoding result and the mixing result of the first decoding result and the second decoding result through different audio output devices.
  • some embodiments of the present application provide a computer-readable storage medium, the computer-readable storage medium includes a computer program, and when the computer program is executed by a processor, the method according to any one of the first aspects is implemented .
  • the first decoding result corresponding to the main audio and the second decoding result corresponding to the description audio are obtained according to the audio data in the audio and video files, and the first decoding result is obtained by
  • the result, as well as the mixing result of the first decoding result and the second decoding result are respectively output through different audio output devices, so that in the scene where ordinary people and blind people watch audio and video at the same time, ordinary people and blind people can choose Different audio output devices perform audio listening to ensure that the two do not affect each other, which meets the needs of ordinary people and blind people to watch at the same time.
  • Figure 1 is a schematic diagram of application scenarios to which some embodiments of this application are adapted;
  • FIG. 2 is a schematic flowchart of an audio processing method provided by an embodiment of this application.
  • FIG. 3 is a schematic flowchart of an audio processing method provided by another embodiment of this application.
  • FIG. 4A is a schematic diagram of a corresponding scenario in some embodiments of this application.
  • 4B is a schematic diagram of a corresponding audio data processing process in some embodiments of this application.
  • FIG. 5A is a schematic diagram of a corresponding scene in some embodiments of this application.
  • 5B is a first schematic diagram of a corresponding audio data processing process in some embodiments of this application.
  • 5C is a second schematic diagram of the audio data processing process in some embodiments of the application.
  • FIG. 6A is a schematic diagram of corresponding scenarios in some embodiments of this application.
  • 6B is a schematic diagram of the corresponding audio data processing process in some embodiments of this application.
  • FIG. 7A is a schematic diagram of corresponding scenarios in some embodiments of this application.
  • FIG. 7B is a schematic diagram of a corresponding audio data processing process in some embodiments of this application.
  • FIG. 8 is a first structural diagram of an electronic device provided by some embodiments of the application.
  • FIG. 9 is a second structural diagram of an electronic device provided by some embodiments of the application.
  • FIG. 1 is a schematic diagram of application scenarios to which some embodiments of this application are adapted.
  • the electronic device 200 may be any electronic device with a multimedia playback function, including but not limited to: a TV, a mobile phone, a platform computer, a desktop computer, etc.
  • FIG. 1 illustrates a case where the electronic device is a television.
  • the user can use the remote control 100 to select a video program, and the remote control 100 can also use the remote control 100 to control the playback state of the video program, for example, start playback, pause playback, adjust volume, switch viewing modes, and so on.
  • the film and television programs include video data and audio data. Therefore, for ordinary users, they can listen to the audio of the program through their ears and watch the video of the program through their eyes. However, for blind people, they can only listen to the audio of the program and cannot watch the video screen, which reduces the perception experience of the film and television program.
  • the audio data of the film and television program itself is called main channel audio
  • the audio data used to explain the content of the current video screen is called descriptive audio. That is to say, in the DTV service, the audio data of the film and television program includes two parts, namely the main channel audio and the description audio.
  • the user watching the movie and TV program in FIG. 1 is an ordinary person
  • the user can set the watching mode of the movie and TV program to the ordinary person watching mode through the remote control.
  • the electronic device only outputs the main audio of the film and television program. That is, only the main channel audio is decoded and output.
  • the user watching the film and television program in FIG. 1 is a blind person
  • the user can set the viewing mode of the film and television program to the blind person viewing mode through the remote control.
  • the electronic device simultaneously outputs the main audio and description audio of the film and television program.
  • the blind can also perceive the video screen of the film and television program by listening to the description audio.
  • Some embodiments of the present application provide an audio processing method, which can meet the needs of ordinary people and blind people to watch TV at the same time.
  • FIG. 2 is a schematic flowchart of an audio processing method provided by an embodiment of this application.
  • the method in some embodiments may be executed by the electronic device in FIG. 1.
  • the electronic device may be any electronic device with multimedia playback function, including but not limited to: TV, mobile phone, platform computer, desktop computer, etc.
  • the method of some embodiments includes:
  • S201 Obtain a first decoding result corresponding to the main audio and a second decoding result corresponding to the description audio according to the audio data in the audio and video file, where the main audio is the audio data of the audio and video file itself, and the description Audio is audio data that describes the content of the video screen in the audio and video file.
  • the main channel audio includes audio data that exists during the audio and video recording process, and describes the audio data added to help the blind person understand the content of the video screen during the post-production process of the audio.
  • the description audio is removed from the data of the audio and video file, and the remaining audio data is the main channel audio.
  • S202 Output the first decoding result and the audio mixing result of the first decoding result and the second decoding result through different audio output devices.
  • audio and video files refer to multimedia files that include audio data and video data. Audio and video files can be played through electronic devices for users to watch. Exemplarily, the audio and video file may be a short audio and video film taken by the user, or a film and television program. For ease of description, in the subsequent description of some embodiments of the present application, a video program is taken as an example for description.
  • electronic devices that support multimedia playback usually include the following devices: controller, demodulator, decoder, and mixer.
  • controller is connected to the demodulator, decoder, and mixer separately for demodulation.
  • Control the processing of the decoder, decoder and mixer.
  • the following describes the data processing process when the electronic device plays a video program in conjunction with the above-mentioned devices.
  • the controller determines the movie and TV program to be played according to the playing instruction input by the user through the remote control.
  • Obtain the code stream of the video program from the server and the code stream is obtained by encoding the original data of the video program according to the protocol type.
  • the code stream of the film and television program obtained by the electronic device from the server is the code stream encoded by the DVT protocol.
  • the electronic device cannot directly play the stream data, but needs to perform a series of processing on the stream data.
  • the stream data of the film and television program is obtained by modulating video data and audio data using a preset modulation method. Since electronic devices process and play video data and audio data in different ways, it is necessary to demodulate the code stream data through a demodulator to obtain video data and audio data.
  • the controller controls the decoder to decode the audio data to obtain audio suitable for playback.
  • the controller controls the mixer to perform mixing processing on the decoded audio, and then output it through the audio output device.
  • the audio mixing process may be a synthesis process on the decoded audio, or a sound effect process on the decoded audio.
  • the decoder obtains the first decoding result corresponding to the main channel audio and the second decoding result corresponding to the description audio according to the audio data corresponding to the movie and TV program.
  • the main channel audio is the audio data of the movie and TV program itself
  • the description audio is the audio data describing the content of the video screen of the movie and TV program.
  • the decoder in some embodiments may be specifically a digital signal processing (DSP) chip responsible for audio decoding work.
  • DSP digital signal processing
  • the decoder can be used to decode audio from multiple input sources, such as DTV input audio, USB input audio, HDMI input audio, etc.
  • the decoder in some embodiments needs to support simultaneous decoding of more than two audio channels.
  • the audio data of film and television programs includes main channel audio and description audio.
  • the decoder decodes the main channel audio and the description audio respectively to obtain the first decoding result corresponding to the main channel audio and the second decoding result corresponding to the description audio.
  • the audio mixer outputs the first decoding result and the mixing result of the first decoding result and the second decoding result through different audio output devices.
  • the main channel audio is output through the audio output device 1, and the mixing result of the main channel audio and the description audio is output through the audio output device 2. Therefore, ordinary people can listen through the audio output device 1, and blind people can listen through the audio output device 2. It can be seen that even if the ordinary person and the blind person watch movies and television programs at the same time, they will not affect each other.
  • the audio output device includes: an audio output device built into the electronic device, and/or an external audio output device connected to the electronic device.
  • the audio output device built into the electronic device may be, for example, a speaker
  • the external audio output device connected to the electronic device may be, for example, a headset.
  • the earphone may be a wired earphone or a wireless earphone.
  • the headset in some embodiments may be a Bluetooth headset.
  • the mixer may output the main channel audio through the speakers, and output the result of mixing the main channel audio and the description audio through the headphones.
  • ordinary people listen to audio data through speakers, and blind people listen to audio data through headphones, and the two do not affect each other.
  • the audio mixer may also output the main channel audio through headphones, and output the result of mixing the main channel audio and the description audio through a speaker.
  • ordinary people listen to audio data through headphones, and blind people listen to audio data through speakers, and the two do not affect each other.
  • what the blind person hears through earphones or speakers is the result of mixing the main audio and the description audio. Since the description audio is the audio data for explaining the video screen of the film and television program, that is, the description audio and the video screen are synchronized. At the same time, the main channel audio and video images in the film and television program are also synchronized. Therefore, in some embodiments, the main channel audio and the description audio are mixed and output to ensure the synchronization of the two channels of audio heard by the blind.
  • the electronic device can also receive the audio output strategy set by the user, that is, the user can manually set according to actual needs in which audio output device to output the main audio, and which audio output device to output the mixed audio. Sound result.
  • the first decoding result corresponding to the main channel audio and the second decoding result corresponding to the description audio are obtained according to the audio data corresponding to the film and television program.
  • the mixing results of the first decoding result and the second decoding result are output through different audio output devices, so that in the scene where ordinary people and blind people watch movies and television programs at the same time, ordinary people and blind people can choose different audio output devices to perform Audio listening ensures that the two do not affect each other, meeting the needs of ordinary people and blind people to watch at the same time.
  • the terminal responds that there is no earphone access, and outputs the main channel audio and the description audio through the speaker. At this time, it may be that only blind people are watching the video, and there is no need to process the data separately.
  • FIG. 3 is a schematic flowchart of an audio processing method provided by another embodiment of this application. Some embodiments are refined on the basis of the above embodiments.
  • the audio output device in some embodiments is described by taking speakers and headphones as examples.
  • the method of some embodiments includes:
  • the demodulator obtains a code stream corresponding to the audio and video file, and demodulates the code stream to obtain video data and audio data in the audio and video file.
  • the decoder performs protocol analysis on the audio data in the audio and video file, and determines whether the audio data includes descriptive audio. If yes, execute S303, if not, execute S304.
  • the decoder decodes the audio data in the audio and video file to obtain the first decoding result corresponding to the main audio and the second decoding result corresponding to the description audio, and stores the first decoding result in In the first buffer, the second decoding result is stored in the second buffer.
  • the decoder decodes the audio data in the audio and video files to obtain the first decoding result corresponding to the main audio, and stores the first decoding result in the first buffer.
  • S305 The controller judges whether the electronic device is connected with a headset, if yes, execute S306, if not, execute S307.
  • the controller can detect whether the electronic device is connected to the earphone according to the connection mode and the connection protocol between the electronic device and the earphone.
  • the controller may detect whether there is a headset device among the Bluetooth devices that establish a pairing relationship with the electronic device.
  • the audio mixer obtains the first decoding result from the first buffer, and sends the first decoding result to the speaker. If the second buffer is not empty, then Acquire the second decoding result in a second buffer, mix the first decoding result and the second decoding result, and send the obtained mixing result to the headset.
  • the audio mixer obtains the first decoding result from the first buffer, and if the second buffer is not empty, obtains the second decoding result from the second buffer, and Mixing the first decoding result and the second decoding result, and sending the obtained mixing result to a speaker.
  • the main audio is output through the speaker
  • the mixing result of the main audio and the description audio is output through the earphone to meet the viewing needs of ordinary people and blind people.
  • the electronic device is not connected to the earphone, the result of mixing the main audio and the description audio is output through the speaker, so that the blind person can meet the viewing needs of the blind person as much as possible without wearing the earphone.
  • audio may not be described for a certain period of time. At this time, the buffer result in the second buffer is empty, and the audio mixer directly outputs the first decoding result.
  • the controller before S305, it may further include: the controller obtains the current viewing mode of the film and television program, and when it is determined that the current viewing mode is a mode that supports blind viewing, execute S305 to S307 .
  • the viewing modes of film and television programs include: a mode that supports viewing by the blind and a mode that does not support viewing by the blind.
  • the user can make settings through the remote control.
  • the electronic device When the user selects a mode that supports blind viewing, the electronic device will consider the blind viewing needs when performing audio mixing processing. Exemplarily, after the main channel audio and the description audio are mixed, the result of the mixing is output through headphones.
  • the electronic device When the user selects a mode that does not support blind viewing, the electronic device does not need to consider the blind viewing requirements when performing audio mixing processing, and directly outputs the main channel audio through speakers and headphones.
  • the electronic equipment in the following scenarios is an example of a TV that supports DTV services.
  • FIG. 4A is a schematic diagram of corresponding scenes in some embodiments of this application
  • FIG. 4B is a schematic diagram of corresponding audio data processing procedures in some embodiments of this application.
  • Figure 4A in this scene, only ordinary people watch the movie and TV program. In this scenario, the user sets the viewing mode of the film and television program to a mode that does not support blind viewing.
  • the DTV code stream of the film and television program is demodulated by the demodulator to obtain audio data and video data.
  • the audio data is sent to the decoder for decoding.
  • the decoder decodes the audio data to obtain the decoding result corresponding to the main audio and the decoding result corresponding to the descriptive audio.
  • the decoder decodes the audio data to obtain the corresponding decoding result of the main audio.
  • the mixer regardless of whether the audio data includes descriptive audio or not, the mixer only outputs the decoding result corresponding to the main audio to the speaker. If the electronic device is also connected to a headset, the decoding result corresponding to the main audio will be output to the headset at the same time.
  • FIG. 5A is a schematic diagram of corresponding scenes in some embodiments of this application
  • FIG. 5B is a schematic diagram 1 of corresponding audio data processing in some embodiments of this application
  • FIG. 5C is a schematic diagram 2 of corresponding audio data processing in some embodiments of this application.
  • Fig. 5A in this scenario, only blind people are watching movies and TV programs, and the user is not wearing headphones. In this scenario, the user sets the viewing mode of the film and television program to a mode that supports viewing by the blind.
  • the DTV code stream of the film and television program is demodulated by the demodulator to obtain audio data and video data.
  • the audio data is sent to the decoder for decoding.
  • FIG. 5A is a schematic diagram of corresponding scenes in some embodiments of this application
  • FIG. 5B is a schematic diagram 1 of corresponding audio data processing in some embodiments of this application
  • FIG. 5C is a schematic diagram 2 of corresponding audio data processing in some embodiments of this application.
  • Fig. 5A in
  • the decoder decodes the audio data to obtain the decoding result corresponding to the main audio.
  • the mixer outputs the decoding result corresponding to the main audio to the speaker.
  • the decoder decodes the audio data to obtain a decoding result corresponding to the main audio and a decoding result corresponding to the descriptive audio.
  • the mixer mixes the decoding result corresponding to the main audio and the decoding result corresponding to the description audio, and then outputs the mixed result to the speaker.
  • FIG. 6A is a schematic diagram of corresponding scenes in some embodiments of this application
  • FIG. 6B is a schematic diagram of corresponding audio data processing procedures in some embodiments of this application.
  • the DTV code stream of the video program is demodulated by the demodulator to obtain audio data and video data.
  • the audio data is sent to the decoder for decoding.
  • the decoder decodes the audio data to obtain the decoding result corresponding to the main audio and the decoding result corresponding to the description audio.
  • the mixer outputs the decoding result corresponding to the main audio to the speaker, mixes the decoding result corresponding to the main audio and the decoding result corresponding to the description audio, and outputs the mixed result to the earphone.
  • FIG. 7A is a schematic diagram of corresponding scenes in some embodiments of this application
  • FIG. 7B is a schematic diagram of corresponding audio data processing procedures in some embodiments of this application.
  • the DTV code stream of the video program is demodulated by the demodulator to obtain audio data and video data.
  • the audio data is sent to the decoder for decoding.
  • the decoder decodes the audio data to obtain the decoding result corresponding to the main audio and the decoding result corresponding to the description audio.
  • the mixer outputs the decoding result corresponding to the main channel audio to the earphone, mixes the decoding result corresponding to the main channel audio and the decoding result corresponding to the description audio, and outputs the mixed result to the speaker.
  • the audio processing methods of some embodiments can meet the viewing needs of ordinary and blind people in different application scenarios. Especially in a scene where the blind and ordinary people watch at the same time, it can be achieved that the two do not interfere with each other, which improves the viewing experience of various users.
  • the present application also provides an audio processing method including: receiving audio and video files, wherein the audio and video files include main audio and description audio, and the main audio is the audio and video file itself
  • the description audio is audio data describing the content of the video screen in the audio and video file; when there is earphone access, the earphone is controlled to output the main channel audio and the description audio, Control the speaker to output the main channel audio; when there is no earphone access, control the speaker to output the main channel audio and the description audio.
  • FIG. 8 is a first structural diagram of an electronic device provided by some embodiments of the application.
  • the electronic device of some embodiments may be any electronic device that supports multimedia playback.
  • the electronic device is a television.
  • the electronic device 800 of some embodiments includes: a decoder 801 and a mixer 802.
  • the decoder 801 is configured to obtain a first decoding result corresponding to the main audio and a second decoding result corresponding to the description audio according to the audio data in the audio and video file, the main audio being the audio and video file itself Audio data, where the description audio is audio data describing the content of the video screen in the audio and video file;
  • the audio mixer 802 is configured to output the first decoding result and the mixing result of the first decoding result and the second decoding result through different audio output devices.
  • the electronic devices of some embodiments can be used to implement the technical solutions of the method embodiment shown in FIG. 2, and the implementation principles and technical effects are similar, and will not be repeated here.
  • FIG. 9 is a second structural diagram of an electronic device provided by some embodiments of this application. As shown in FIG. 9, the electronic device 800 of some embodiments may further include a controller 803 and a demodulator 801.
  • the audio output device includes a speaker and an earphone, and the controller 803 is used to determine whether the electronic device is connected to the earphone;
  • the audio mixer 802 is specifically configured to output the audio mixing result of the first decoding result and the second decoding result through the earphone if the earphone is connected to the electronic device, and output the The first decoding result is output through the speaker.
  • the decoder 801 is specifically configured to decode audio data in an audio and video file to obtain a first decoding result corresponding to the main audio and a second decoding result corresponding to the description audio. And store the first decoding result in a first buffer, and store the second decoding result in a second buffer.
  • the audio mixer 802 is specifically used for:
  • the second buffer is not empty, the second decoding result is obtained from the second buffer, and the first decoding result and the second decoding result are mixed to obtain The result of the mixing is sent to the headset.
  • the decoder 801 is configured to perform protocol analysis on audio data in an audio and video file, and determine that the audio data includes the main audio and the description audio.
  • the controller 803 is further configured to obtain the current viewing mode of the audio and video file, and determine that the current viewing mode is a mode that supports viewing by the blind.
  • the electronic device further includes a demodulator for:
  • the code stream is demodulated to obtain video data and audio data in the audio and video file.
  • the electronic equipment of some embodiments can be used to implement the technical solutions of any of the foregoing method embodiments, and its implementation principles and technical effects are similar, and will not be repeated here.
  • some embodiments of the present application also provide an electronic device, an electronic device, a speaker; a processor connected to the speaker, and the processor is configured to receive audio and video files, wherein the audio The video file includes main channel audio and description audio, the main channel audio is audio data of the audio and video file itself, and the description audio is audio data describing the content of the video screen in the audio and video file; When there is earphone access, control the earphone to output the main channel audio and the description audio, and control the speaker to output the main channel audio; when there is no earphone access, control the speaker to output the main channel audio And the description audio.
  • controlling the earphone to output the main channel audio and the description audio, and controlling the speaker to output the main channel audio includes: according to the audio and video files To obtain the first decoding result corresponding to the main audio data and the second decoding result corresponding to the description audio; when a headset is connected, pass the mixing result of the first decoding result and the second decoding result The earphone outputs; the first decoding result is output through the speaker.
  • the controlling the speaker to output the main channel audio and the description audio when there is no earphone access includes: obtaining the first channel audio corresponding to the main channel audio according to the audio data in the audio and video file A decoding result and a second decoding result corresponding to the description audio; when no earphone is connected, the audio mixing result of the first decoding result and the second decoding result is output through the speaker.
  • the processor includes a controller, a decoder, and a mixer.
  • the processor controls the headset to output the main audio and the description audio, and controls the The speaker outputs the main channel audio; when there is no earphone access, controlling the speaker to output the main channel audio and the description audio includes: the decoder obtains the main channel audio according to the audio data in the audio and video files The corresponding first decoding result and the second decoding result corresponding to the description audio; when a headset is connected, the controller controls the mixer to mix the first decoding result with the second decoding result The result is output through the headphones, and the first decoding result is output through the speaker; when no headphones are connected, the controller controls the mixer to combine the first decoding result with the first decoding result. The audio mixing result of the second decoding result is output through the speaker.
  • the electronic device further includes a first buffer and a second buffer, the first buffer is used to buffer the first decoding result, and the second buffer is used to buffer the second decoding result. result.
  • the present application also provides an electronic device, including a speaker; a processor connected to the speaker, the processor is configured to: receive audio and video files, where the audio and video files include main audio and description Audio, the main channel audio is the audio data of the audio and video file itself, and the description audio is the audio data describing the content of the video screen in the audio and video file; The main channel audio and the description audio control the output of the earphones, and the speaker output is controlled according to the main channel audio; when there is no earphone access, the main channel audio and the description audio control the Speaker output.
  • the audio and video files include main audio and description Audio
  • the main channel audio is the audio data of the audio and video file itself
  • the description audio is the audio data describing the content of the video screen in the audio and video file
  • the main channel audio and the description audio control the output of the earphones, and the speaker output is controlled according to the main channel audio; when there is no earphone access, the main channel audio and the description audio control the Speaker output.
  • An embodiment of the present application further provides an electronic device including: a processor and a memory; wherein the memory is used to store a computer program; the processor is used to execute the computer program stored in the memory to implement the audio processing method in the above embodiment .
  • an electronic device including: a processor and a memory; wherein the memory is used to store a computer program; the processor is used to execute the computer program stored in the memory to implement the audio processing method in the above embodiment .
  • the memory may be independent or integrated with the processor.
  • the electronic device may further include: a bus for connecting the memory and the processor.
  • the electronic equipment provided in some embodiments can be used to implement the technical solutions in any of the foregoing method embodiments, and its implementation principles and technical effects are similar, and some embodiments will not be repeated here.
  • Some embodiments of the present application further provide a computer-readable storage medium, the computer-readable storage medium includes a computer program, and the computer program is used to implement the technical solution in any of the above method embodiments.
  • Some embodiments of the present application further provide a chip, including: a memory, a processor, and a computer program.
  • the computer program is stored in the memory, and the processor runs the computer program to execute any of the above method embodiments.
  • the disclosed device and method may be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the modules is only a logical function division, and there may be other divisions in actual implementation, for example, multiple modules can be combined or integrated. To another system, or some features can be ignored, or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or modules, and may be in electrical, mechanical or other forms.
  • modules described as separate components may or may not be physically separated, and the components displayed as modules may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the modules may be selected according to actual needs to achieve the objectives of some embodiments.
  • the functional modules in the various embodiments of the present application may be integrated into one processing unit, or each module may exist alone physically, or two or more modules may be integrated into one unit.
  • the units formed by the above-mentioned modules can be realized in the form of hardware, or in the form of hardware plus software functional units.
  • the above-mentioned integrated modules implemented in the form of software function modules may be stored in a computer readable storage medium.
  • the above-mentioned software function module is stored in a storage medium and includes several instructions to make a computer device (which can be a personal computer, a server, or a network device, etc.) or a processor (English: processor) to execute the various embodiments of the present application Part of the method.
  • processor may be a central processing unit (English: Central Processing Unit, abbreviated: CPU), or other general-purpose processors, digital signal processors (English: Digital Signal Processor, abbreviated: DSP), and application-specific integrated circuits (English: Application Specific Integrated Circuit, referred to as ASIC) etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the steps of the method disclosed in combination with the application can be directly embodied as executed by a hardware processor, or executed by a combination of hardware and software modules in the processor.
  • the memory may include a high-speed RAM memory, and may also include a non-volatile storage NVM, such as at least one disk storage, and may also be a U disk, a mobile hard disk, a read-only memory, a magnetic disk, or an optical disk.
  • NVM non-volatile storage
  • the bus can be an Industry Standard Architecture (ISA) bus, Peripheral Component (PCI) bus, or Extended Industry Standard Architecture (EISA) bus, etc.
  • ISA Industry Standard Architecture
  • PCI Peripheral Component
  • EISA Extended Industry Standard Architecture
  • the bus can be divided into address bus, data bus, control bus, etc.
  • the buses in the drawings of this application are not limited to only one bus or one type of bus.
  • the above-mentioned storage medium can be realized by any type of volatile or non-volatile storage device or their combination, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable Except for programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disks or optical disks.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read-only memory
  • EPROM erasable except for programmable read only memory
  • PROM programmable read only memory
  • ROM read only memory
  • magnetic memory flash memory
  • flash memory magnetic disks or optical disks.
  • optical disks any available medium that can be accessed by a general-purpose or special-purpose computer.
  • An exemplary storage medium is coupled to the processor, so that the processor can read information from the storage medium and can write information to the storage medium.
  • the storage medium may also be an integral part of the processor.
  • the processor and the storage medium may be located in Application Specific Integrated Circuits (ASIC for short).
  • ASIC Application Specific Integrated Circuits
  • the processor and the storage medium may also exist as discrete components in the electronic device or the main control device.
  • a person of ordinary skill in the art can understand that all or part of the steps in the foregoing method embodiments can be implemented by a program instructing relevant hardware.
  • the aforementioned program can be stored in a computer readable storage medium.
  • the steps including the foregoing method embodiments are executed; and the foregoing storage medium includes: ROM, RAM, magnetic disk, or optical disk and other media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Educational Administration (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

本申请一些实施例提供的音频处理方法及电子设备,根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,通过将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出,使得在普通人和盲人同时观看音视频的场景中,普通人和盲人可以选择不同的音频输出设备进行音频收听,保证二者不互相影响,满足了普通人和盲人同时观看的需求。

Description

音频处理方法及电子设备
本申请要求于2019年05月23日提交中国专利局、申请号为201910433524.2、申请名称为″音频数据的处理方法及电子设备″的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请一些实施例涉及多媒体技术领域,尤其涉及一种音频处理方法及电子设备。
背景技术
随着多媒体技术的发展,人们越来越喜欢通过影视节目获取资讯。对于普通人员,可以通过耳朵收听节目的音频,通过眼睛观看节目的视频画面。但是,对于盲人而言,却只能收听节目的音频,无法观看到视频画面,使得对影视节目的感知体验降低。
现有的数字电视(Digital television,DTV)业务,为了提升盲人对影视节目的感知体验,在DTV业务的音频数据中添加一路描述音频。其中,描述音频是指对当前视频画面的内容进行解析的音频数据。当盲人″观看″电视时,在收听影视节目的主路音频的同时,还可以通过收听描述音频来感知影视节目的视频画面。
但是,现有的DTV业务,无法满足普通人员和盲人同时观看电视的需求。
发明内容
本申请一些实施例提供一种音频处理方法及电子设备,用以满足普通人和盲人同时观看影视节目的需求。
第一方面,本申请一些实施例提供一种音频处理方法,包括:
根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据, 所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
第二方面,本申请一些实施例提供一种音频处理方法,包括:
接收音视频文件,其中,所述音视频文件包括主路音频和描述音频;
在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;
在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
第三方面,本申请一些实施例提供一种电子设备,包括:
扬声器;和扬声器连接的处理器,所述处理器用于:
接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据,所述主路音频是只音视频文件中除所述描述音频外的音频数据;
在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;
在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
第四方面,本申请一些实施例提供一种电子设备,包括:扬声器;和扬声器连接的处理器,所述处理器用于:
接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
在有耳机接入时,根据所述主路音频和所述描述音频控制所述耳机的输出,根据所述主路音频控制所述扬声器的输出。
第五方面,本申请一些实施例提供一种电子设备,包括:
解码器,用于根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
混音器,用于将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
第六方面,本申请一些实施例提供一种计算机可读存储介质,所述计算机可读存储介质包括计算机程序,所述计算机程序被处理器执行时实现如第一方面任一项所述的方法。
本申请一些实施例提供的音频处理方法及电子设备,根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,通过将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出,使得在普通人和盲人同时观看音视频的场景中,普通人和盲人可以选择不同的音频输出设备进行音频收听,保证二者不互相影响,满足了普通人和盲人同时观看的需求。
附图说明
为了更清楚地说明本申请一些实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本申请一些实施例适应的应用场景示意图;
图2为本申请一实施例提供的音频处理方法的流程示意图;
图3为本申请另一实施例提供的音频处理方法的流程示意图;
图4A为本申请一些实施例中对应的场景示意图;
图4B为本申请一些实施例中对应的音频数据处理过程示意图;
图5A为本申请一些实施例中对应的场景示意图;
图5B为本申请一些实施例中对应的音频数据处理过程示意图一;
图5C为本申请一些实施例中对应的音频数据处理过程示意图二;
图6A为本申请一些实施例中对应的场景示意图;
图6B为本申请一些实施例中对应的音频数据处理过程示意图;
图7A为本申请一些实施例中对应的场景示意图;
图7B为本申请一些实施例中对应的音频数据处理过程示意图;
图8为本申请一些实施例提供的电子设备的结构示意图一;
图9为本申请一些实施例提供的电子设备的结构示意图二。
具体实施方式
下面将结合本申请一些实施例中的附图,对本申请一些实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
本申请的说明书和权利要求书及上述附图中的术语″第一″、″第二″、″第三″″第四″等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本申请的实施例例如能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语″包括″和″具有″以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。
图1为本申请一些实施例适应的应用场景示意图。如图1所示,该应用场景中,至少一个用户通过电子设备200观看影视节目。其中,电子设备200可以是具有多媒体播放功能的任意电子设备,包括但不限于:电视机、手机、平台电脑、台式电脑等。示例性的,图1中示例了电子设备为电视机的情况。用户在通过电子设备观看影视节目时,可以使用遥控器100选择影视节目,还可以使用遥控器100对影视节目的播放状态进行控制,例如:开始播放、暂停播放、调节音量、切换观看模式等。
可以理解的,为了丰富用户对影视节目的感知体验,影视节目通过包括视频数据和音频数据。因此,对于普通用户来说,可以通过耳朵收听节目的音频,通过眼睛观看节目的视频画面。但是,对于盲人而言,却只能收听节目的音频,无法观看到视频画面,使得对影视节目的感知体验降低。
现有的数字电视(Digital television,DTV)业务,为了提升盲人对影视节目的感知体验,在DTV业务的音频数据中添加一路特殊音频,该特殊音频用于对当前视频画面的内容进行解说。为了描述方便,本申请一些实施例中将影视节目本身的音频数据称为主路音频,将用于对当前视频画面的内容进行解说的音频数据称为描述音频。也就是说,DTV业务中,影视节目的音频数据包括两部分,分别为主路音频和描述音频。
一种可能的场景中,若图1中观看影视节目的用户为普通人,用户可以通过遥控器将影视节目的观看模式设置为普通人观看模式。该模式下,电子设备只输出影视节目的主路音频。即只针对主路音频进行解码和输出。
另一种可能的场景中,若图1中观看影视节目的用户为盲人,用户可以通过遥控器将影视节目的观看模式设置为盲人观看模式。该模式下,电子设备同时输出影视节目的主路音频和描述音频。也就是说,盲人在收听影视节目的主路音频的同时,还可以通过收听描述音频来感知影视节目的视频画面。
然而,图1所示的应用场景中,还可能存在另外一种场景,例如:观看影视节目的用户既有普通人,也有盲人。现有的DTV业务,无法满足普通人员和盲人同时观看电视的需求。
本申请一些实施例提供一种音频处理方法,能够满足普通人和盲人同时观看电视的需求。
下面以具体地实施例对本申请的技术方案进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例不再赘述。
图2为本申请一实施例提供的音频处理方法的流程示意图,一些实施例的方法可以由图1中的电子设备执行。该电子设备可以是具有多媒体播放功能的任意电子设备,包括但不限于:电视机、手机、平台电脑、台式电脑等。
如图2所示,一些实施例的方法,包括:
S201:根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据。
在一些实施例中,主路音频包括音视频录制过程中即存在的音频数据,描述音频时后期制作过程中为了帮助盲人了解视频画面的内容而增加的音频数据。
在一些实施例中,音视频文件的数据除去描述音频,剩下的音频数据就是主路音频。
S202:将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
首先,本申请一些实施例中,音视频文件是指包括音频数据和视频数据的 多媒体文件。音视频文件可以通过电子设备进行播放以供用户观看。示例性的,音视频文件可以是用户自拍摄的音视频短片,还可以是影视节目。为了描述方便,本申请一些实施例的后续描述中,以影视节目为例进行说明。
通常,支持多媒体播放的电子设备通常包括如下器件:控制器、解调器、解码器和混音器,其中,控制器与解调器、解码器、混音器分别连接,用于对解调器、解码器和混音器的处理过程进行控制。下面结合上述各器件描述电子设备对影视节目进行播放时的数据处理过程。
电子设备对影视节目进行播放时,控制器根据用户通过遥控器输入的播放指令,确定待播放影视节目。从服务器获取该影视节目的码流,该码流是根据协议类型对影视节目的原始数据进行编码后得到的。以DTV业务为例,电子设备从服务器获取到的影视节目的码流是采用DVT协议编码后的码流。电子设备不能直接对该码流数据进行播放,而是需要对该码流数据进行一系列的处理。
在一些实施例中,影视节目的码流数据是对视频数据和音频数据采用预设的调制方式进行调制得到的。由于电子设备对视频数据和音频数据的处理和播放方式不同,因此,需要首先通过解调器对该码流数据进行解调,得到视频数据和音频数据。针对解调得到的音频数据,控制器控制解码器对该音频数据进行解码,得到适合播放的音频。在一些实施例中,控制器控制混音器对解码得到的音频进行混音处理,然后通过音频输出设备进行输出。其中,混音处理可以是对解码后的音频进行合成处理,还可以是对解码后的音频进行音效处理。
一些实施例S201中,解码器根据影视节目对应的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果。其中,所述主路音频是所述影视节目本身的音频数据,所述描述音频是对所述影视节目的视频画面的内容进行描述的音频数据。
可以理解的,一些实施例中的解码器可以具体是负责音频解码工作的数字信号处理(Digital Singnal Processo,DSP)芯片。解码器可以用于对多种输入源的音频进行解码,如:DTV输入音频、USB输入音频、HDMI输入音频等。
一些实施例中的解码器需要支持同时进行两路以上音频的解码。以DTV输入音频为例,影视节目的音频数据中包括主路音频和描述音频。解码器获取音频数据后,对主路音频和描述音频分别进行解码,得到主路音频对应的第一解码结果和描述音频对应的第二解码结果。
一些实施例S202中,混音器将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
可以理解的,普通人观看影视节目时,只需要收听主路音频,而盲人观看影视节目时,需要同时收听主路音频和描述音频。一些实施例中,为了兼顾普通人和盲人的观看需求,将主路音频通过音频输出设备1进行输出,将主路音频和描述音频的混音结果通过音频输出设备2进行输出。从而,普通人可以通过音频输出设备1进行收听,盲人可以通过音频输出设备2进行收听,可见,即使普通人和盲人同时观看影视节目,也不会相互影响。
其中,音频输出设备包括:电子设备内置的音频输出设备,和/或,与电子设备连接的外置音频输出设备。电子设备内置的音频输出设备例如可以是扬声器,与电子设备连接的外置音频输出设备例如可以是耳机。其中,耳机可以为有线耳机或者无线耳机。在一些实施例中,一些实施例中的耳机可以为蓝牙耳机。
一些实施例中,音频输出设备可以有多种。以扬声器和耳机为例,一种可能的实施方式中,混音器可以将主路音频通过扬声器进行输出,将主路音频与描述音频的混音结果通过耳机进行输出。该场景下,普通人通过扬声器收听音频数据,盲人通过耳机收听音频数据,两者互不影响。
另一种可能的实施方式中,混音器还可以将主路音频通过耳机进行输出,将主路音频与描述音频的混音结果通过扬声器进行输出。该场景下,普通人通过耳机收听音频数据,盲人通过扬声器收听音频数据,两者互不影响。
在一些实施例中,一些实施例中,盲人通过耳机或者扬声器收听到的是主路音频与描述音频混音后得到的混音结果。由于描述音频是对影视节目的视频画面进行解说的音频数据,即描述音频与视频画面具有同步性。同时,影视节目中的主路音频与视频画面也具有同步性,因此,一些实施例中将主路音频与描述音频进行混音后输出,保证了盲人收听到的两路音频的同步性。
可以理解的,一些实施例中,对于具体采用哪种音频输出设备进行输出并不作具体限定,只要主路音频、以及主路音频与描述音频的混音结果通过不同的音频输出设备进行输出即可。一种可能的实施方式中,电子设备还可以接收用户设置的音频输出策略,即,用户可以根据实际需要手动设置在哪路音频输出设备中输出主路音频,在哪路音频输出设备中输出混音结果。
一些实施例提供的音频处理方法,根据影视节目对应的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,通过将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出,使得在普通人和盲人同时观看影视节目的场景中,普通人和盲人可以选择不同的音频输出设备进行音频收听,保证二者不互相影响,满足了普通人和盲人同时观看的需求。
在一些实施例中,如果音频数据中包含主路音频和描述音频,终端响应于没有耳机接入,则将主路音频和描述音频通过扬声器输出。此时,可以是仅盲人在观看视频,无需将数据进行分别处理。
图3为本申请另一实施例提供的音频处理方法的流程示意图。一些实施例在上述实施例的基础上进行细化。一些实施例中的音频输出设备以扬声器和耳机为例进行描述。
如图3所示,一些实施例的方法,包括:
S301:解调器获取音视频文件对应的码流,对所述码流进行解调,得到音视频文件中的视频数据和音频数据。
S302:解码器对音视频文件中的音频数据进行协议解析,判断所述音频数据中是否包括描述音频。若是,则执行S303,若否,则执行S304。
可以理解的,当根据协议解析结果,确定音频数据中不包括描述音频时,则无需对描述音频进行解码,以提高解码器的处理效率。
S303:解码器对音视频文件中的音频数据进行解码,得到所述主路音频对应的第一解码结果,以及所述描述音频对应的第二解码结果,并将所述第一解码结果存储至第一缓存器中,将所述第二解码结果存储至第二缓存器中。
S304:解码器对音视频文件中的音频数据进行解码,得到所述主路音频对应的第一解码结果,并将所述第一解码结果存储至第一缓存器中。
S305:控制器判断电子设备是否连接有耳机,若是,则执行S306,若否,则执行S307。
在一些实施例中,控制器可以根据电子设备与耳机之间的连接方式以及连接协议,检测电子设备是否连接有耳机。示例性的,以蓝牙耳机为例,控制器可以检测与电子设备建立配对关系的蓝牙设备中是否存在耳机设备。
S306:混音器从所述第一缓存器中获取所述第一解码结果,并将所述第一 解码结果发送给所述扬声器,若所述第二缓存器不为空,则从所述第二缓存器中获取所述第二解码结果,并对所述第一解码结果和所述第二解码结果进行混音,将得到的混音结果发送给所述耳机。
S307:混音器从所述第一缓存器中获取所述第一解码结果,若所述第二缓存器不为空,则从所述第二缓存器中获取所述第二解码结果,并对所述第一解码结果和所述第二解码结果进行混音,将得到的混音结果发送给扬声器。
一些实施例中,在电子设备连接有耳机时,则将主路音频通过扬声器输出,将主路音频和描述音频的混音结果通过耳机输出,以同时满足普通人和盲人的观看需求。在电子设备没有连接耳机时,将主路音频和描述音频的混音结果通过扬声器输出,以使盲人在不佩戴耳机的情况下,尽可能满足盲人的观看需求。
在一些实施例中,某一段时间可能没有描述音频,此时第二缓存器中的缓存结果为空,则混音器直接将第一解码结果进行输出。
在一些实施例中,一些实施例中,在S305之前,还可以包括:控制器获取所述影视节目的当前观看模式,在确定所述当前观看模式为支持盲人观看的模式时,执行S305至S307。
其中,影视节目的观看模式包括:支持盲人观看的模式和不支持盲人观看的模式。用户可以通过遥控器进行设置。当用户选择支持盲人观看的模式时,则电子设备在进行混音处理时,会考虑盲人观看需求。示例性的,将主路音频和描述音频进行混音处理后,将混音结果通过耳机进行输出。当用户选择不支持盲人观看的模式时,则电子设备在进行混音处理时,无须考虑盲人观看需求,直接将主路音频通过扬声器和耳机进行输出即可。
下面结合几个具体的场景分别进行描述,下述各场景中的电子设备是以支持DTV业务的电视机为例。
图4A为本申请一些实施例中对应的场景示意图,图4B为本申请一些实施例中对应的音频数据处理过程示意图。如图4A所示,该场景中,只有普通人观看影视节目。该场景下,用户将影视节目的观看模式设置为不支持盲人观看的模式。如图4B所示,影视节目的DTV码流经过解调器解调后,得到音频数据和视频数据。音频数据被送入解码器进行解码。当音频数据中包括描述音频时,解码器对音频数据进行解码,得到主路音频对应的解码结果和描述音频对应的解码结果。当音频数据中不包括描述音频时,解码器对音频数据进行解码,得 到主路音频对应的解码结果。该场景下,无论音频数据中是否包括描述音频,混音器都只将主路音频对应的解码结果输出给扬声器。如果电子设备还连接有耳机,将同时将主路音频对应的解码结果输出给耳机。
图5A为本申请一些实施例中对应的场景示意图,图5B为本申请一些实施例中对应的音频数据处理过程示意图一,图5C为本申请一些实施例中对应的音频数据处理过程示意图二。如图5A所示,该场景中,只有盲人观看影视节目,且用户未佩戴耳机。该场景下,用户将影视节目的观看模式设置为支持盲人观看的模式。如图5B和图5C所示,影视节目DTV码流经过解调器解调后,得到音频数据和视频数据。音频数据被送入解码器进行解码。如图5B所示,当音频数据中不包括描述音频时,解码器对音频数据进行解码,得到主路音频对应的解码结果。混音器将主路音频对应的解码结果输出给扬声器。如图5C所示,当音频数据中包括描述音频时,解码器对音频数据进行解码,得到主路音频对应的解码结果和描述音频对应的解码结果。混音器对主路音频对应的解码结果和描述音频对应的解码结果进行混音后,将混音结果输出给扬声器。
图6A为本申请一些实施例中对应的场景示意图,图6B为本申请一些实施例中对应的音频数据处理过程示意图。如图6A所示,该场景中,普通人和盲人观看影视节目,且盲人佩戴耳机。该场景下,用户将影视节目的观看模式设置为支持盲人观看的模式。如图6B所示,影视节目DTV码流经过解调器解调后,得到音频数据和视频数据。音频数据被送入解码器进行解码。解码器对音频数据进行解码,得到主路音频对应的解码结果和描述音频对应的解码结果。混音器将主路音频对应的解码结果输出给扬声器,并对主路音频对应的解码结果和描述音频对应的解码结果进行混音,将混音结果输出给耳机。
图7A为本申请一些实施例中对应的场景示意图,图7B为本申请一些实施例中对应的音频数据处理过程示意图。如图7A所示,该场景中,普通人和盲人观看影视节目,且普通人佩戴耳机。该场景下,用户将影视节目的观看模式设置为支持盲人观看的模式。如图7B所示,影视节目DTV码流经过解调器解调后,得到音频数据和视频数据。音频数据被送入解码器进行解码。解码器对音频数据进行解码,得到主路音频对应的解码结果和描述音频对应的解码结果。混音器将主路音频对应的解码结果输出给耳机,并对主路音频对应的解码结果和描述音频对应的解码结果进行混音,将混音结果输出给扬声器。
由此可见,一些实施例的音频处理方法,在不同的应用场景中均能满足普通和盲人的观看需求。尤其在盲人和普通人同时观看的场景中,能够实现二者互不干扰,提高了各种用户的观看体验。
在一些实施例中,本申请还提供了一种音频处理方法包括:接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
图8为本申请一些实施例提供的电子设备的结构示意图一。一些实施例的电子设备可以是支持多媒体播放的任一电子设备。在一些实施例中,该电子设备为电视机。
如图8所示,一些实施例的电子设备800,包括:解码器801和混音器802。
其中,解码器801,用于根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
混音器802,用于将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
一些实施例的电子设备,可用于执行上述如图2所示方法实施例的技术方案,其实现原理和技术效果类似,此处不再赘述。
图9为本申请一些实施例提供的电子设备的结构示意图二,如图9所示,一些实施例的电子设备800,还可以包括:控制器803和解调器801。
在一些实施例中,所述音频输出设备包括扬声器和耳机,所述控制器803,用于判断电子设备是否连接有所述耳机;
所述混音器802,具体用于若所述电子设备连接有所述耳机,则将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出。
在一些实施例中,所述解码器801具体用于对音视频文件中的音频数据进行解码,得到所述主路音频对应的第一解码结果,以及所述描述音频对应的第 二解码结果,并将所述第一解码结果存储至第一缓存器中,将所述第二解码结果存储至第二缓存器中。
在一些实施例中,所述混音器802具体用于:
从所述第一缓存器中获取所述第一解码结果,并将所述第一解码结果发送给所述扬声器;
若所述第二缓存器不为空,则从所述第二缓存器中获取所述第二解码结果,并对所述第一解码结果和所述第二解码结果进行混音,将得到的混音结果发送给所述耳机。
在一些实施例中,所述解码器801,用于对音视频文件中的音频数据进行协议解析,确定所述音频数据包括所述主路音频和所述描述音频。
在一些实施例中,所述控制器803还用于获取所述音视频文件的当前观看模式,确定所述当前观看模式为支持盲人观看的模式。
在一些实施例中,所述电子设备还包括:解调器,用于:
获取音视频文件的码流;
对所述码流进行解调,得到音视频文件中的视频数据和音频数据。
一些实施例的电子设备,可用于执行上述任一方法实施例的技术方案,其实现原理和技术效果类似,此处不再赘述。
在一些实施例中,本申请一些实施例中还提供了一种电子设备,一种电子设备,扬声器;和扬声器连接的处理器,所述处理器用于:接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
在一些实施例中,所述在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频包括:根据所述音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;在有耳机接入时,将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出;将所述第一解码结果通过所述扬声器进行输出。
在一些实施例中,所述在无耳机接入时,控制所述扬声器输出所述主路音 频和所述描述音频包括:根据所述音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;在无耳机接入时,将所述第一解码结果与所述第二解码结果的混音结果通过所述扬声器进行输出。
在一些实施例中,所述处理器包括控制器,解码器,混音器,所述处理器在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频包括:所述解码器根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;在有耳机接入时,所述控制器控制所述混音器将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出;在无耳机接入时,所述控制器控制所述混音器将所述第一解码结果与所述第二解码结果的混音结果通过所述扬声器进行输出。
在一些实施例中,所述电子设备还包括第一缓存器和第二缓存器,所述第一缓存器用于缓存所述第一解码结果,所述第二缓存器用户缓存所述第二解码结果。
在一些实施例中,本申请还提供了一种电子设备,包括扬声器;和扬声器连接的处理器,所述处理器用于:接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;在有耳机接入时,根据所述主路音频和所述描述音频控制所述耳机的输出,根据所述主路音频控制所述扬声器的输出;在无耳机接入时,根据所述主路音频和所述描述音频控制所述扬声器的输出。
本申请一实施例还提供一种电子设备包括:处理器以及存储器;其中,存储器,用于存储计算机程序;处理器,用于执行存储器存储的计算机程序,以实现上述实施例中的音频处理方法。具体可以参见前述方法实施例中的相关描述。
在一些实施例中,存储器既可以是独立的,也可以跟处理器集成在一起。
当所述存储器是独立于处理器之外的器件时,所述电子设备还可以包括:总线,用于连接所述存储器和处理器。
一些实施例提供的电子设备,可用于执行上述任一方法实施例中的技术方 案,其实现原理和技术效果类似,一些实施例此处不再赘述。
本申请一些实施例还提供一种计算机可读存储介质,所述计算机可读存储介质包括计算机程序,所述计算机程序用于实现如上任一方法实施例中的技术方案。
本申请一些实施例还提供一种芯片,包括:存储器、处理器以及计算机程序,所述计算机程序存储在所述存储器中,所述处理器运行所述计算机程序执行上述任一方法实施例中的技术方案。
在本申请所提供的几个实施例中,应该理解到,所揭露的设备和方法,可以通过其它的方式实现。例如,以上所描述的设备实施例仅仅是示意性的,例如,所述模块的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个模块可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或模块的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的模块可以是或者也可以不是物理上分开的,作为模块显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现一些实施例方案的目的。
另外,在本申请各个实施例中的各功能模块可以集成在一个处理单元中,也可以是各个模块单独物理存在,也可以两个或两个以上模块集成在一个单元中。上述模块成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能单元的形式实现。
上述以软件功能模块的形式实现的集成的模块,可以存储在一个计算机可读取存储介质中。上述软件功能模块存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器(英文:processor)执行本申请各个实施例所述方法的部分步骤。
应理解,上述处理器可以是中央处理单元(英文:Central Processing Unit,简称:CPU),还可以是其他通用处理器、数字信号处理器(英文:Digital Signal Processor,简称:DSP)、专用集成电路(英文:Application Specific Integrated Circuit,简称:ASIC)等。通用处理器可以是微处理器或者该处理器也可以是 任何常规的处理器等。结合申请所公开的方法的步骤可以直接体现为硬件处理器执行完成,或者用处理器中的硬件及软件模块组合执行完成。
存储器可能包含高速RAM存储器,也可能还包括非易失性存储NVM,例如至少一个磁盘存储器,还可以为U盘、移动硬盘、只读存储器、磁盘或光盘等。
总线可以是工业标准体系结构(Industry Standard Architecture,ISA)总线、外部设备互连(Peripheral Component,PCI)总线或扩展工业标准体系结构(Extended Industry Standard Architecture,EISA)总线等。总线可以分为地址总线、数据总线、控制总线等。为便于表示,本申请附图中的总线并不限定仅有一根总线或一种类型的总线。
上述存储介质可以是由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。存储介质可以是通用或专用计算机能够存取的任何可用介质。
一种示例性的存储介质耦合至处理器,从而使处理器能够从该存储介质读取信息,且可向该存储介质写入信息。当然,存储介质也可以是处理器的组成部分。处理器和存储介质可以位于专用集成电路(Application Specific Integrated Circuits,简称:ASIC)中。当然,处理器和存储介质也可以作为分立组件存在于电子设备或主控设备中。
本领域普通技术人员可以理解:实现上述各方法实施例的全部或部分步骤可以通过程序指令相关的硬件来完成。前述的程序可以存储于一计算机可读取存储介质中。该程序在执行时,执行包括上述各方法实施例的步骤;而前述的存储介质包括:ROM、RAM、磁碟或者光盘等各种可以存储程序代码的介质。
最后应说明的是:以上各实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述各实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分或者全部技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的范围。

Claims (17)

  1. 一种电子设备,其特征在于,包括:
    扬声器;
    和扬声器连接的处理器,
    所述处理器用于:
    接收音视频文件,其中,所述音视频文件包括主路音频和描述音频;
    在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;
    在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
  2. 根据权利要求1所述的电子设备,其特征在于,
    所述在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频包括:
    根据所述音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;
    在有耳机接入时,将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出;将所述第一解码结果通过所述扬声器进行输出。
  3. 根据权利要求1所述的电子设备,其特征在于,
    所述在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频包括:
    根据所述音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;
    在无耳机接入时,将所述第一解码结果与所述第二解码结果的混音结果通过所述扬声器进行输出。
  4. 根据权利要求1所述的电子设备,其特征在于,
    所述处理器包括控制器,解码器,混音器,所述处理器在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频包括:
    所述解码器根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果;
    在有耳机接入时,所述控制器控制所述混音器将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出;
    在无耳机接入时,所述控制器控制所述混音器将所述第一解码结果与所述第二解码结果的混音结果通过所述扬声器进行输出。
  5. 根据权利要求4所述的电子设备,其特征在于,所述电子设备还包括第一缓存器和第二缓存器,所述第一缓存器用于缓存所述第一解码结果,所述第二缓存器用户缓存所述第二解码结果。
  6. 一种电子设备,其特征在于,包括:
    扬声器;
    和扬声器连接的处理器,
    所述处理器用于:
    接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据,所述主路音频是只音视频文件中除所述描述音频外的音频数据;
    在有耳机接入时,根据所述主路音频和所述描述音频控制所述耳机的输出,根据所述主路音频控制所述扬声器的输出;
    在无耳机接入时,根据所述主路音频和所述描述音频控制所述扬声器的输出。
  7. 一种电子设备,其特征在于,包括:
    解码器,用于根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
    混音器,用于将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
  8. 根据权利要求7所述的电子设备,其特征在于,所述音频输出设备包括扬声器和耳机,所述电子设备还包括:
    控制器,用于判断电子设备是否连接有所述耳机;
    所述混音器,具体用于若所述电子设备连接有所述耳机,则将所述第一解 码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出。
  9. 一种音频处理方法,其特征在于,包括:
    接收音视频文件,其中,所述音视频文件包括主路音频和描述音频,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
    在有耳机接入时,控制所述耳机输出所述主路音频和所述描述音频,控制所述扬声器输出所述主路音频;
    在无耳机接入时,控制所述扬声器输出所述主路音频和所述描述音频。
  10. 一种音频处理方法,其特征在于,包括:
    根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,所述主路音频是所述音视频文件本身的音频数据,所述描述音频是对所述音视频文件中的视频画面的内容进行描述的音频数据;
    将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出。
  11. 根据权利要求10所述的方法,其特征在于,所述音频输出设备包括扬声器和耳机,所述将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出,包括:
    判断电子设备是否连接有所述耳机,若所述电子设备连接有所述耳机,则将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出。
  12. 根据权利要求11所述的方法,其特征在于,所述根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果,包括:
    对音视频文件中的音频数据进行解码,得到所述主路音频对应的第一解码结果,以及所述描述音频对应的第二解码结果,并将所述第一解码结果存储至第一缓存器中,将所述第二解码结果存储至第二缓存器中。
  13. 根据权利要求12所述的方法,其特征在于,所述将所述第一解码结果与所述第二解码结果的混音结果通过所述耳机进行输出,将所述第一解码结果通过所述扬声器进行输出,包括:
    从所述第一缓存器中获取所述第一解码结果,并将所述第一解码结果发送给所述扬声器;
    若所述第二缓存器不为空,则从所述第二缓存器中获取所述第二解码结果,并对所述第一解码结果和所述第二解码结果进行混音,将得到的混音结果发送给所述耳机。
  14. 根据权利要求12所述的方法,其特征在于,所述对音视频文件中的音频数据进行解码之前,还包括:
    对音视频文件中的音频数据进行协议解析,确定所述音频数据包括所述主路音频和所述描述音频。
  15. 根据权利要求10至14任一项所述的方法,其特征在于,所述将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出之前,还包括:
    获取所述音视频文件的当前观看模式,确定所述当前观看模式为支持盲人观看的模式。
  16. 根据权利要求10至14任一项所述的方法,其特征在于,所述根据音视频文件中的音频数据,获取主路音频对应的第一解码结果和描述音频对应的第二解码结果之前,还包括:
    获取音视频文件的码流;
    对所述码流进行解调,得到所述音视频文件中的视频数据和音频数据。
  17. 根据权利要求11所述的方法,其特征在于,所述将所述第一解码结果、以及所述第一解码结果与所述第二解码结果的混音结果分别通过不同的音频输出设备进行输出,还包括:
    若所述电子设备未连接有所述耳机,则将所述第一解码结果与所述第二解码结果的混音结果通过所述扬声器进行输出。
PCT/CN2020/083613 2019-05-23 2020-04-08 音频处理方法及电子设备 WO2020233263A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/347,372 US11956497B2 (en) 2019-05-23 2021-06-14 Audio processing method and electronic device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910433524.2A CN110139151A (zh) 2019-05-23 2019-05-23 音频数据的处理方法及电子设备
CN201910433524.2 2019-05-23

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/347,372 Continuation US11956497B2 (en) 2019-05-23 2021-06-14 Audio processing method and electronic device

Publications (1)

Publication Number Publication Date
WO2020233263A1 true WO2020233263A1 (zh) 2020-11-26

Family

ID=67572680

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/083613 WO2020233263A1 (zh) 2019-05-23 2020-04-08 音频处理方法及电子设备

Country Status (3)

Country Link
US (1) US11956497B2 (zh)
CN (1) CN110139151A (zh)
WO (1) WO2020233263A1 (zh)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110139151A (zh) * 2019-05-23 2019-08-16 青岛海信传媒网络技术有限公司 音频数据的处理方法及电子设备
US11669295B2 (en) * 2020-06-18 2023-06-06 Sony Group Corporation Multiple output control based on user input
CN113783645A (zh) * 2021-09-30 2021-12-10 上海电影技术厂有限公司 一种服务于视障人士的无障碍放映设备
FR3131058B1 (fr) * 2021-12-21 2024-08-09 Sagemcom Broadband Sas Boitier décodeur pour la restitution d’une piste audio additionnelle.

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2144438A1 (en) * 2008-07-07 2010-01-13 Sony United Kingdom Limited Television device with user interface
CN101877806A (zh) * 2009-04-28 2010-11-03 St微电子(格勒诺布尔2)有限公司 控制音频内容通过两个扬声器的广播的方法和设备
CN201663649U (zh) * 2010-04-27 2010-12-01 胡志超 一种具有盲人功能的电视接收机
CN103179451A (zh) * 2013-03-19 2013-06-26 深圳市九洲电器有限公司 基于dvb标准的双音频混合输出方法、装置及机顶盒
CN105245938A (zh) * 2015-09-30 2016-01-13 努比亚技术有限公司 播放多媒体文件的装置和方法
CN110139151A (zh) * 2019-05-23 2019-08-16 青岛海信传媒网络技术有限公司 音频数据的处理方法及电子设备

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101458951A (zh) * 2008-12-30 2009-06-17 胡礼斌 具有多功能的影音节目信号处理系统
KR102531886B1 (ko) * 2016-08-17 2023-05-16 삼성전자주식회사 전자장치 및 그 제어방법
CN109275027A (zh) * 2018-09-26 2019-01-25 Tcl海外电子(惠州)有限公司 视频的语音输出方法、电子播放设备及存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2144438A1 (en) * 2008-07-07 2010-01-13 Sony United Kingdom Limited Television device with user interface
CN101877806A (zh) * 2009-04-28 2010-11-03 St微电子(格勒诺布尔2)有限公司 控制音频内容通过两个扬声器的广播的方法和设备
CN201663649U (zh) * 2010-04-27 2010-12-01 胡志超 一种具有盲人功能的电视接收机
CN103179451A (zh) * 2013-03-19 2013-06-26 深圳市九洲电器有限公司 基于dvb标准的双音频混合输出方法、装置及机顶盒
CN105245938A (zh) * 2015-09-30 2016-01-13 努比亚技术有限公司 播放多媒体文件的装置和方法
CN110139151A (zh) * 2019-05-23 2019-08-16 青岛海信传媒网络技术有限公司 音频数据的处理方法及电子设备

Also Published As

Publication number Publication date
US11956497B2 (en) 2024-04-09
CN110139151A (zh) 2019-08-16
US20210306699A1 (en) 2021-09-30

Similar Documents

Publication Publication Date Title
WO2020233263A1 (zh) 音频处理方法及电子设备
EP3562163B1 (en) Audio-video synthesis method and system
CN109445740B (zh) 音频播放方法、装置、电子设备及存储介质
US7894854B2 (en) Image/audio playback device of mobile communication terminal
US9189985B2 (en) Mobile information terminal
US9055265B2 (en) Accessibility improvement for hearing impaired
WO2015139547A1 (zh) 一种设置应用场景的方法、设备和系统
KR20180069576A (ko) 영상제공장치, 그의 제어 방법 및 영상제공시스템
US20120317594A1 (en) Method and system for providing an improved audio experience for viewers of video
CN113050916A (zh) 音频播放方法、装置及存储介质
US20200194016A1 (en) Audio playing method and apparatus, and terminal
CN108650592B (zh) 一种实现颈带式环绕立体声的方法及立体声控制系统
CN115767158A (zh) 同步播放方法、终端设备及存储介质
CN102833520A (zh) 一种视频会议信号处理的方法、视频会议服务器及系统
TW201015997A (en) System for listening multiple sound tracks
EP4456568A1 (en) Audio processing method and terminals
JP5581437B1 (ja) 映像提供システムおよびプロラム
KR20140012463A (ko) 방송 수신기 및 오디오 신호 처리 방법
TWI631853B (zh) 影音控制裝置及其方法
US11533523B2 (en) Interconnected system for high-quality wireless transmission of audio and video between electronic consumer devices
JP2024100180A (ja) 再生制御装置、受信装置、およびプログラム
EP3785443A1 (en) System and method for rendering stitched video media stream on a display device
KR100620682B1 (ko) 이동통신 단말기의 음향 재생 장치 및 방법
JP2010041220A (ja) データ処理装置、データ処理方法、及びプログラム
CN115348466A (zh) 播放节目的方法、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20809075

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20809075

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 28/03/2022)

122 Ep: pct application non-entry in european phase

Ref document number: 20809075

Country of ref document: EP

Kind code of ref document: A1