WO2020207132A1 - 应用程序的音频数据的输出方法及显示设备 - Google Patents

应用程序的音频数据的输出方法及显示设备 Download PDF

Info

Publication number
WO2020207132A1
WO2020207132A1 PCT/CN2020/076488 CN2020076488W WO2020207132A1 WO 2020207132 A1 WO2020207132 A1 WO 2020207132A1 CN 2020076488 W CN2020076488 W CN 2020076488W WO 2020207132 A1 WO2020207132 A1 WO 2020207132A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
audio data
display device
application
input instruction
Prior art date
Application number
PCT/CN2020/076488
Other languages
English (en)
French (fr)
Inventor
孙永瑞
Original Assignee
海信视像科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 海信视像科技股份有限公司 filed Critical 海信视像科技股份有限公司
Priority to US16/847,100 priority Critical patent/US11288033B2/en
Publication of WO2020207132A1 publication Critical patent/WO2020207132A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Definitions

  • the present disclosure relates to the field of audio technology, and in particular, to a method for outputting audio data of an application program and a display device.
  • display devices can not only provide users with various traditional live program content received and sent through broadcast signals, but also provide users with various content provided by content service providers such as video on demand, online games, and video chats.
  • content service providers such as video on demand, online games, and video chats.
  • the display device can be controlled to perform the functions associated with the above-mentioned various contents based on the user's operation of the physical hard keys or virtual keys on the control device such as remote control, mobile terminal, etc., or through The own microphone or the voice input by the user received by the microphone on the control device is controlled to perform the above-mentioned various content-related functions.
  • the embodiments of the present disclosure provide a method for outputting audio data of an application program and a display device, so as to realize that the audio data of only one application program is played at the same time in the display device, so as to improve the user's listening experience.
  • embodiments of the present disclosure provide a method for outputting audio data of an application in a display device.
  • the method includes: receiving the first audio data for playing the second audio data of the second application during the process of the display device playing the first audio data of one or more first application programs through the audio output interface of the display device.
  • An input instruction in response to the first input instruction, the output volume of the first audio data is controlled to be muted, and the output volume of the second audio data is controlled to remain; through the audio output interface, after the output is muted
  • the method further includes: after receiving the first input instruction, determining whether the second application program needs to acquire the audio focus of the operating system in the display device; Second, the application needs to obtain the audio focus and modify the audio stream type corresponding to the second audio data.
  • the determining whether the second application program needs to acquire the audio focus of the operating system in the display device includes: determining the package name of the second application program; determining the pre-stored need to acquire the Whether the package name of the second application exists in the whitelist of the audio focus application; if the package name of the second application exists in the whitelist, it is determined that the second application needs to obtain the Audio focus.
  • the second application includes a voice assistant application.
  • the method further includes: after receiving the first input instruction, determining whether there is a third application program whose audio stream type corresponding to the third audio data is modified;
  • the application program retains the volume of the audio stream type corresponding to the third audio data of the third application program, and mutes the volume of the audio stream type corresponding to the audio data of other applications except the third application program.
  • the third application is the second application.
  • the method further includes: receiving a second input instruction for stopping playing the second audio data; in response to the second input instruction, controlling the output volume of the first audio data, and Mute is restored to the volume before receiving the first input instruction for playing the second audio data; and the first audio data after the output volume is restored through the audio output interface is output.
  • embodiments of the present disclosure provide a method for outputting audio data of an application in a display device.
  • the method includes: receiving the first audio data for playing the second audio data of the second application during the process of the display device playing the first audio data of one or more first application programs through the audio output interface of the display device.
  • An input instruction in response to the first input instruction, the second audio data is output through the audio output interface, and the first audio data is not output.
  • the method further includes: after receiving the first input instruction, determining whether the second application program needs to acquire the audio focus of the operating system in the display device; in response to the second The application needs to obtain the audio focus, and mark the audio focus of the operating system as being occupied by the second application.
  • the method further includes: after receiving the first input instruction, determining whether the audio focus of the operating system currently in the display device is occupied by other applications; in response to the audio focus being occupied If the other application program is occupied, the second audio data is not output to the audio output interface; in response to the audio focus is not occupied by the other application program, the second audio data is output to the audio output interface.
  • the method further includes: receiving a second input instruction for stopping playing the second audio data; and in response to the second input instruction, controlling the output of the second audio data to resume through the audio output interface.
  • the first audio data is a second input instruction for stopping playing the second audio data.
  • inventions of the present disclosure provide a display device.
  • the display device includes a user interface, an audio output interface, a memory, and a controller.
  • the memory in communication with the user interface and the audio output interface, is configured to store a computer program, instructions associated with the user interface, and audio data associated with the audio output interface.
  • the controller in communication with the user interface, the audio output interface, and the memory, is configured to run the computer program and/or the instruction to make the display device: the display device passes through the display device
  • the audio output interface receives a first input instruction for playing second audio data of a second application during the process of playing the first audio data of one or more first application programs; in response to the first input instruction , Controlling the output volume of the first audio data to be muted, and controlling the output volume of the second audio data to remain; through the audio output interface, after outputting the muted first audio data and retaining the output volume The superimposed data of the second audio data.
  • the controller is further configured to run the computer program and/or the instruction to cause the display device to: after receiving the first input instruction, determine whether the second application program The audio focus of the operating system in the display device needs to be acquired; in response to the second application program needing to acquire the audio focus, the audio stream type corresponding to the second audio data is modified.
  • the controller is further configured to run the computer program and/or the instruction to make the display device: determine the package name of the second application; determine the pre-stored need to obtain the audio Whether the package name of the second application exists in the whitelist of the focused application; if the package name of the second application exists in the whitelist, it is determined that the second application needs to obtain the audio focus.
  • the second application includes a voice assistant application.
  • the controller is further configured to run the computer program and/or the instruction to make the display device: after receiving the first input instruction, determine whether there is a third audio data corresponding The third application whose audio stream type is modified; in response to the existence of the third application, the volume of the audio stream type corresponding to the third audio data of the third application is retained, and the third application is muted.
  • the third application is the second application.
  • the controller is further configured to run the computer program and/or the instruction to cause the display device to: receive a second input instruction for stopping playing the second audio data; in response to The second input instruction controls the output volume of the first audio data, and restores the volume from mute to the volume before receiving the first input instruction for playing the second audio data; the output is restored through the audio output interface The first audio data after the volume is output.
  • an embodiment of the present disclosure provides a display device.
  • the display device includes a user interface, an audio output interface, a memory, and a controller.
  • the memory in communication with the user interface and the audio output interface, is configured to store a computer program, instructions associated with the user interface, and audio data associated with the audio output interface.
  • the controller in communication with the user interface, the audio output interface, and the memory, is configured to run the computer program and/or the instruction to make the display device: the display device passes through the display device
  • the audio output interface receives a first input instruction for playing second audio data of a second application during the process of playing the first audio data of one or more first application programs; in response to the first input instruction , Outputting the second audio data through the audio output interface, but not outputting the first audio data.
  • the controller is further configured to run the computer program and/or the instruction to cause the display device to: after receiving the first input instruction, determine whether the second application program requires Acquire the audio focus of the operating system in the display device; and in response to the second application needing to acquire the audio focus, mark the audio focus of the operating system as being occupied by the second application.
  • the controller is further configured to run the computer program and/or the instruction so that the display device: after receiving the first input instruction, determine the current operation in the display device Whether the audio focus of the system is occupied by other applications; in response to the audio focus being occupied by the other application, the second audio data is not output to the audio output interface; in response to the audio focus not being occupied by the audio output interface Occupied by other application programs, output the second audio data to the audio output interface.
  • the controller is further configured to run the computer program and/or the instruction to cause the display device to: receive a second input instruction for stopping playing the second audio data; in response to The second input instruction controls to resume outputting the first audio data through the audio output interface.
  • embodiments of the present disclosure provide a non-volatile computer-readable storage medium having a computer program stored on the storage medium.
  • the computer program is executed by a processor of a display device, the following Or the method described in the second aspect.
  • FIG. 1A exemplarily shows a schematic diagram of an operation scene between a display device 200 and a control device 100;
  • FIG. 1B exemplarily shows a configuration block diagram of the control device 100 in FIG. 1A;
  • FIG. 1C exemplarily shows a configuration block diagram of the display device 200 in FIG. 1A;
  • FIG. 1D exemplarily shows a block diagram of the architecture configuration of the operating system in the memory of the display device 200
  • Figure 2 shows a schematic diagram of the audio data output of each application in the display device
  • Figure 3A shows a flow chart of a first method for controlling audio data output of an application in a display device
  • FIG. 3B shows a flowchart of a second method for controlling the output of audio data of an application program in a display device
  • 4A shows a flow chart of a third method for controlling the output of audio data of an application in a display device
  • FIG. 4B shows a flowchart of a fourth method for controlling the output of audio data of an application in a display device.
  • module used in the present disclosure refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic or a combination of hardware or/and software code that can perform functions related to the element.
  • gesture refers to a user's behavior that is used to express a predetermined idea, action, purpose, and/or result through an action such as a change of hand shape or hand movement.
  • FIG. 1A exemplarily shows a schematic diagram of an operation scene between the display device 200 and the control device 100.
  • the control device 100 and the display device 200 can communicate in a wired or wireless manner.
  • the control device 100 is configured to control the display device 200.
  • the control device 100 can receive an operation instruction input by the user, and convert the operation instruction into an instruction that the display device 200 can recognize and respond to, and play an intermediary role in realizing information interaction between the user and the display device 200.
  • the user sends an instruction to the display device 200 to control channel addition and subtraction by operating the channel increase or channel reduction key on the control device 100, and the display device 200 receives and responds to the instruction to perform the channel addition and subtraction operation .
  • the control device 100 may be a remote controller 100A.
  • the remote controller 100A can implement communication methods such as infrared protocol communication, Bluetooth protocol communication, or other short-distance communication.
  • the remote controller 100A controls the display device 200 in a wireless or other wired manner.
  • the user can control the display device 200 by inputting user instructions through physical or virtual buttons on the remote control 100A, a voice interface, a control panel interface, etc.
  • the user can use physical or virtual keys such as volume plus and minus keys, channel control keys, up/down/left/right movement keys, voice input keys, menu keys, and power on/off keys on the remote control 100A to input
  • the corresponding control instruction is used to control the display device 200.
  • the control device 100 may also be a smart device, such as a mobile terminal 100B, a tablet computer, a computer, a notebook computer, etc.
  • the user can control the display device 200 through an application program running on a smart device.
  • the user can configure the application.
  • the configured application program can be projected onto the screen associated with the smart device, so that the user can control the display device 200 through an intuitive user interface (UI).
  • UI intuitive user interface
  • the mobile terminal 100B and the display device 200 install the same and/or related software applications, so that the two can communicate through a network communication protocol, thereby achieving one-to-one control operation and data communication.
  • a control instruction protocol may be established between the mobile terminal 100B and the display device 200.
  • the user realizes corresponding functions by operating various physical or virtual keys on the user interface of the mobile terminal 100B.
  • the user operates a physical button arranged on the remote control 100A to realize the function corresponding to the physical button; or, the audio and video content displayed on the mobile terminal 100B can be transmitted to the display device 200, thereby realizing the synchronous display function.
  • the display device 200 may provide a broadcast receiving function, a computer support function and other network TV functions.
  • the display device 200 may be implemented as digital TV, network TV, Internet Protocol TV (IPTV), and so on.
  • IPTV Internet Protocol TV
  • the display device 200 may be a liquid crystal display, an organic light emitting display, or a projection device.
  • the specific display device type, size, resolution, etc. are not limited in the embodiment of the present application.
  • the display device 200 and the server 300 perform data communication through multiple communication methods.
  • the display device 200 is communicatively connected through a local area network (LAN), a wireless local area network (WLAN), and other networks.
  • LAN local area network
  • WLAN wireless local area network
  • the server 300 provides various contents and interactions to the display device 200.
  • the display device 200 can send and receive information.
  • the display device 200 receives electronic program guide (Electronic Program Guide, EPG) data, software program updates, etc., or accesses a remotely stored digital media library.
  • EPG Electronic Program Guide
  • the server 300 can be one group or multiple groups, and can be one type or multiple types of servers.
  • the server 300 provides network service content such as video-on-demand and advertising services.
  • FIG. 1B exemplarily shows a configuration block diagram of the control device 100.
  • the control device 100 includes a controller 110, a memory 120, a communicator 130, a user input interface 140, an output interface 150, and a power supply 160.
  • the controller 110 includes a random access memory (Random Access Memory, RAM) 111, a read-only memory (Read-Only Memory, ROM) 112, a processor 113, a communication interface (not shown), and a communication bus (not shown).
  • RAM Random Access Memory
  • ROM Read-Only Memory
  • the controller 110 is used to control the operation and operation of the control device 100, as well as the communication cooperation between internal components, and external and internal data processing.
  • the controller 110 may generate a control corresponding to the detected interaction. Signal and send the control signal to the display device 200.
  • the memory 120 is used to store various operating programs, data, and applications for driving and controlling the control device 100 under the control of the controller 110.
  • the memory 120 can store various control signal instructions input by the user.
  • the communicator 130 realizes communication of control signals and data signals with the display device 200 under the control of the controller 110.
  • control device 100 sends a touch signal, a button signal, and other control signals to the display device 200 through the communicator 130.
  • the control device 100 receives the signal sent by the display device 200 through the communicator 130.
  • the communicator 130 includes an infrared signal interface 131 and a radio frequency signal interface 132.
  • the communicator 130 when the communicator 130 includes an infrared signal interface, the communicator 130 needs to convert a user input instruction into an infrared control signal according to an infrared control protocol, and then send it to the display device 200 via an infrared sending circuit.
  • the communicator 130 when the communicator 130 includes a radio frequency signal interface, the communicator 130 needs to convert user input instructions into digital signals, then modulate them according to the radio frequency control signal modulation protocol, and send them to the display device 200 via the radio frequency transmitting terminal.
  • the user input interface 140 includes at least one of a microphone 141, a touch panel 142, a sensor 143, a button 144, etc., so that the user can input a user instruction for controlling the display device 200 to the control device through voice, touch, gesture, pressing, etc. 100.
  • the output interface 150 outputs a user instruction received by the user input interface 140 to the display device 200, or outputs an image or voice signal received by the display device 200.
  • the output interface 150 includes an LED interface 151, a vibration interface 152 that generates vibration, a sound output interface 153 that outputs sound, and a display 154 that outputs images, and the like.
  • the remote controller 100A can receive output signals such as audio, video, or data from the output interface 150, and display the output signals as images on the display 154, output as audio on the sound output interface 153, or output on the vibration interface 152 In the form of vibration.
  • output signals such as audio, video, or data from the output interface 150
  • display the output signals as images on the display 154, output as audio on the sound output interface 153, or output on the vibration interface 152 In the form of vibration.
  • the power supply 160 is used to provide operating power support for the components in the control device 100 under the control of the controller 110.
  • the power supply 160 is configured as a battery and related control circuits.
  • FIG. 1C exemplarily shows a block diagram of the hardware configuration of the display device 200.
  • the display device 200 may include a tuner and demodulator 210, a communicator 220, a detector 230, an external device interface 240, a controller 250, a memory 260, a user interface 265, a video processor 270, a display 275, Audio processor 280, audio input interface 285, and power supply 290.
  • the tuner and demodulator 210 which receives broadcast television signals through wired or wireless means, can perform modulation and demodulation processing such as amplification, mixing and resonance, and is used to demodulate the television selected by the user from multiple wireless or cable broadcast television signals Audio and video signals carried in the frequency of the channel, and additional information (such as EPG data).
  • modulation and demodulation processing such as amplification, mixing and resonance
  • the tuner and demodulator 210 can be selected by the user and controlled by the controller 250 to respond to the frequency of the television channel selected by the user and the television signal carried by the frequency.
  • the tuner and demodulator 210 can receive signals in many ways according to different broadcasting formats of TV signals, such as terrestrial broadcasting, cable broadcasting, satellite broadcasting or Internet broadcasting; and according to different modulation types, it can be digital modulation or analog Modulation method; and according to different types of received TV signals, analog signals and digital signals can be demodulated.
  • different broadcasting formats of TV signals such as terrestrial broadcasting, cable broadcasting, satellite broadcasting or Internet broadcasting
  • modulation types it can be digital modulation or analog Modulation method
  • received TV signals, analog signals and digital signals can be demodulated.
  • the tuner demodulator 210 may also be in an external device, such as an external set-top box.
  • the set-top box outputs a television signal after modulation and demodulation, and inputs it to the display device 200 through the external device interface 240.
  • the communicator 220 is a component used to communicate with external devices, such as external servers, according to various communication protocols.
  • the display device 200 may transmit content data to an external device connected via the communicator 220, or browse and download content data from an external device connected via the communicator 220.
  • the communicator 220 includes a network communication protocol module such as a WIFI module 221, a Bluetooth communication protocol module 222, and a wired Ethernet communication protocol module 223 or a near field communication protocol module, so that the communicator 220 can receive the control device 100 according to the control of the controller 250 Control signals, and implement the control signals as WIFI signals, Bluetooth signals, radio frequency signals, etc.
  • a network communication protocol module such as a WIFI module 221, a Bluetooth communication protocol module 222, and a wired Ethernet communication protocol module 223 or a near field communication protocol module
  • the detector 230 is a component of the display device 200 for collecting signals from the external environment or interacting with the outside.
  • the detector 230 may include a sound collector 231, such as a microphone.
  • the detector 230 is used to receive the voice signal corresponding to the user's voice, such as the voice signal corresponding to the control instruction of the user to control the display device 200; or, the detector 230 can collect the environmental sound used to identify the type of environmental scene, so that the display device 200 can automatically Adapt to environmental noise.
  • the detector 230 includes an image collector 232, such as a camera, a camera, and the like.
  • the detector 230 is used to collect external environment scenes to adaptively change the display parameters of the display device 200; and to collect attributes of the user or perform gesture interaction with the user to realize the function of interaction between the display device and the user.
  • the detector 230 includes a light receiver for collecting the intensity of ambient light, so as to adapt to changes in display parameters of the display device 200 and so on.
  • the detector 230 includes a temperature sensor. For example, by sensing the ambient temperature, the display device 200 can adaptively adjust the display color temperature of the image.
  • the display device 200 in a high temperature environment, can adjust the display image color temperature to be colder; in a low temperature environment, the display device 200 can adjust the display image color temperature to be warmer.
  • the external device interface 240 is a component that provides the controller 210 to control data transmission between the display device 200 and external devices.
  • the external device interface 240 can be connected to external devices such as set-top boxes, game devices, notebook computers, etc. in a wired/wireless manner, and can receive external devices such as video signals (such as moving images), audio signals (such as music), and additional information (such as EPG) and other data.
  • external devices such as set-top boxes, game devices, notebook computers, etc. in a wired/wireless manner, and can receive external devices such as video signals (such as moving images), audio signals (such as music), and additional information (such as EPG) and other data.
  • the external device interface 240 includes: High Definition Multimedia Interface (HDMI) terminal 241, Composite Video Broadcast Signal (CVBS) terminal 242, analog or digital component terminal 243, Universal Serial Bus ( Any one or more of a USB terminal 244, a component terminal (not shown), a red-green-blue (RGB) terminal (not shown), and the like.
  • HDMI High Definition Multimedia Interface
  • CVBS Composite Video Broadcast Signal
  • USB Universal Serial Bus
  • the controller 250 controls the work of the display device 200 and responds to user operations by running various software control programs (such as an operating system and various application programs) stored on the memory 260.
  • various software control programs such as an operating system and various application programs
  • the controller 250 includes a RAM 251, a ROM 252, a graphics processor 253, a central processing unit (CPU) processor 254, a communication interface 255, and a communication bus 256.
  • a RAM 251 a ROM 252
  • a graphics processor 253 a graphics processor 253
  • a central processing unit (CPU) processor 254 a communication interface 255, and a communication bus 256.
  • CPU central processing unit
  • the RAM 251, the ROM 252, the graphics processor 253, and the CPU processor 254 communication interface 255 are connected through a communication bus 256.
  • ROM252 used to store various system startup instructions. For example, when the power-on signal is received, the power of the display device 200 starts to start, and the CPU processor 254 runs the system startup instruction in the ROM 252, and copies the operating system stored in the memory 260 to the RAM 251 to start the startup operating system. After the operating system is started up, the CPU processor 254 copies various application programs in the memory 260 to the RAM 251, and then starts to run and start various application programs.
  • the graphics processor 253 is used to generate various graphics objects, such as icons, operating menus, and user input instructions to display graphics.
  • the graphics processor 253 may include an arithmetic unit, which is used to perform operations by receiving various interactive instructions input by the user, and then display various objects according to the display attributes; and includes a renderer, which is used to generate various objects obtained based on the arithmetic unit, and perform operations.
  • the rendered result is displayed on the display 275.
  • the CPU processor 254 is configured to execute operating system and application program instructions stored in the memory 260. And according to the received user input instructions, to execute various applications, data and content processing, so as to finally display and play various audio and video content.
  • the CPU processor 254 may include multiple processors.
  • the multiple processors may include a main processor and multiple or one sub-processor.
  • the main processor is configured to perform some initialization operations of the display device 200 in the display device preloading mode, and/or, to display screen operations in the normal mode. Multiple or one sub-processor, used to perform an operation in the standby mode of the display device.
  • the communication interface 255 may include the first interface to the nth interface. These interfaces may be network interfaces connected to external devices via a network.
  • the controller 250 may control the overall operation of the display device 200. For example, in response to receiving a user input command for selecting a Graphical User Interface (GUI) object displayed on the display 275, the controller 250 may perform an operation related to the object selected by the user input command.
  • GUI Graphical User Interface
  • the object can be any one of the selectable objects, such as a hyperlink or an icon.
  • the operation related to the selected object for example, the operation of displaying the page, document, image, etc. connected to the hyperlink, or the operation of executing the program corresponding to the object.
  • the user input command for selecting the GUI object may be an input command or a voice command corresponding to the voice spoken by the user through various input devices connected to the display device 200.
  • the input device may be a mouse, a keyboard, a touch pad, etc.
  • the memory 260 is used to store various types of data, software programs or application programs for driving and controlling the operation of the display device 200.
  • the memory 260 may include volatile and/or nonvolatile readable memory.
  • the term “memory” includes the memory 260, the RAM 251 and ROM 252 of the controller 250, or the memory card in the display device 200.
  • the memory 260 is specifically used to store operating programs that drive the controller 250 in the display device 200; to store various application programs built in the display device 200 and downloaded from external devices by the user; and to store configuration provided by the display 275 Data such as various GUIs, various objects related to the GUI, and visual effect images of the selector used to select GUI objects.
  • the memory 260 is specifically used to store the drivers and related data of the tuner and demodulator 210, the communicator 220, the detector 230, the external device interface 240, the video processor 270, the display 275, the audio processor 280, etc. , External data (such as audio and video data) received from the external device interface or user data (such as button information, voice information, touch information, etc.) received from the user interface.
  • External data such as audio and video data
  • user data such as button information, voice information, touch information, etc.
  • the memory 260 specifically stores software and/or programs used to represent an operating system (OS). These software and/or programs may include, for example: kernel, middleware, application programming interface (API), and/or application.
  • OS operating system
  • these software and/or programs may include, for example: kernel, middleware, application programming interface (API), and/or application.
  • the kernel can control or manage system resources, and functions implemented by other programs (such as the middleware, API, or application program); at the same time, the kernel can provide interfaces to allow middleware, API, or application program access control To control or manage system resources.
  • FIG. 1D exemplarily shows a block diagram of the architecture configuration of the operating system in the memory of the display device 200.
  • the operating system architecture from top to bottom is the application layer, the framework layer and the kernel layer.
  • Application layer system built-in applications and non-system-level applications belong to the application layer. Responsible for direct interaction with users.
  • the application layer may include multiple applications, such as live TV applications, video-on-demand applications, media center applications, screenshot applications, and so on.
  • Live TV applications can provide live TV through different sources.
  • a live TV application may use input from cable TV, wireless broadcasting, satellite services, or other types of live TV services to provide TV signals.
  • the live TV application can display the video of the live TV signal on the display device 200.
  • Video-on-demand applications can provide videos from different storage sources. Unlike live TV applications, VOD provides video display from certain storage sources. For example, the video on demand can come from the server side of cloud storage, and from the local hard disk storage that contains the stored video programs.
  • Media center applications can provide various multimedia content playback applications.
  • the media center may be different from live TV or video on demand, and users can access various images or audio and video stored in the memory through the media center application.
  • the screenshot application can take a screenshot of the current display on the monitor, and mark the identifiers (such as people, channel logos, buildings, etc.) contained in the screenshot image, such as the identification frame, name, etc., to provide users with a display Display function of various identifiers contained in the screen.
  • the currently displayed screen can be at least one of text, image, and video.
  • the framework layer is responsible for providing the API required by the application layer.
  • live TV applications, video-on-demand applications, and media center applications can call decoders to decode audio and video through the interface provided by the framework layer.
  • the screenshot application can call the captured screenshot image of the current display screen through the interface provided by the framework layer.
  • the kernel layer provides core system services, such as file management, memory management, process management, network management, system security authority management and other services.
  • the kernel layer may be implemented as a kernel based on various operating systems, for example, a kernel based on the Android operating system.
  • the kernel also provides communication between system software and hardware, and device driver services for various hardware, such as: providing display drivers for displays, camera drivers for cameras, button drivers for remote controls, and WIFI modules WiFi driver, audio driver for audio output interface, power management driver for power management (PM) module, etc.
  • the user interface 265 receives various user interactions. Specifically, it is used to send the user's input signal to the controller 250, or to transmit the output signal from the controller 250 to the user.
  • the remote control 100A may send input signals such as a power switch signal, a channel selection signal, and a volume adjustment signal input by the user to the user interface 265, and then the user interface 265 transfers to the controller 250; or the remote control 100A may Receive output signals such as audio, video, or data output from the user interface 265 through the controller 250 and display the received output signal or output the received output signal as audio or vibration.
  • the user can input user commands on the GUI displayed on the display 275, and the user interface 265 receives the user input commands through the GUI.
  • the user interface 265 may receive a user input command for controlling the position of the selector in the GUI to select different objects or items.
  • the user may input a user command by inputting a specific sound or gesture, and the user interface 265 recognizes the sound or gesture through the sensor to receive the user input command.
  • the video processor 270 is used to receive external video signals, and perform video data processing such as decompression, decoding, scaling, noise reduction, frame rate conversion, resolution conversion, and image synthesis according to the standard codec protocol of the input signal.
  • the video signal displayed or played directly on the display 275.
  • the video processor 270 includes a demultiplexing module, a video decoding module, an image synthesis module, a frame rate conversion module, a display formatting module, and the like.
  • the demultiplexing module is used to demultiplex the input audio and video data stream, such as input MPEG-2 stream (based on the compression standard of digital storage media moving images and voice), then the demultiplexing module will demultiplex it Multiplexed into video signals and audio signals, etc.
  • input audio and video data stream such as input MPEG-2 stream (based on the compression standard of digital storage media moving images and voice)
  • the video decoding module is used to process the demultiplexed video signal, including decoding and scaling.
  • An image synthesis module such as an image synthesizer, is used to superimpose and mix the GUI signal generated by the graphics generator with the zoomed video image according to user input or itself to generate an image signal for display.
  • the frame rate conversion module is used to convert the frame rate of the input video, such as converting the frame rate of the input 60Hz video to a frame rate of 120Hz or 240Hz, and the usual format is implemented in a frame-insertion mode.
  • the display formatting module is used to change the signal output by the frame rate conversion module into a signal conforming to the display format such as a display, for example, format the signal output by the frame rate conversion module to output RGB data signals.
  • the display 275 is used to receive the image signal input from the video processor 270 to display video content, images, and a menu control interface.
  • the displayed video content can be from the video content in the broadcast signal received by the tuner and demodulator 210, or from the video content input by the communicator 220 or the external device interface 240.
  • the display 275 simultaneously displays a user manipulation interface UI generated in the display device 200 and used for controlling the display device 200.
  • the display 275 may include a display screen component for presenting a picture and a driving component for driving image display.
  • the display 275 may also include a projection device and a projection screen.
  • the audio processor 280 is used to receive external audio signals, and perform decompression and decoding according to the standard codec protocol of the input signal, as well as audio data processing such as noise reduction, digital-to-analog conversion, and amplification processing.
  • the audio signal to be played is used to receive external audio signals, and perform decompression and decoding according to the standard codec protocol of the input signal, as well as audio data processing such as noise reduction, digital-to-analog conversion, and amplification processing.
  • the audio processor 280 may support various audio formats. Such as MPEG-2, MPEG-4, Advanced Audio Coding (AAC), High Efficiency AAC (HE-AAC) and other formats.
  • AAC Advanced Audio Coding
  • HE-AAC High Efficiency AAC
  • the audio output interface 285 is used to receive the audio signal output by the audio processor 280 under the control of the controller 250.
  • the audio output interface 285 may include a speaker 286 or output to an external audio output terminal 287 of a generator of an external device, such as a headset Output terminal.
  • the video processor 270 may include one or more chips.
  • the audio processor 280 may also include one or more chips.
  • the video processor 270 and the audio processor 280 may be separate chips, or may be integrated with the controller 250 in one or more chips.
  • the power supply 290 is used to provide power supply support for the display device 200 with power input from an external power supply under the control of the controller 250.
  • the power supply 290 may be a built-in power supply circuit installed inside the display device 200, or may be a power supply installed outside the display device 200. It should be noted that, on the basis of the above-mentioned Figures 1A-1D, the display device can receive the voice input by the user through its own sound collector 231 (which can be implemented as a microphone) or the microphone 141 on the control device, and be controlled to perform its provision. The functions required for various content.
  • a voice assistant application can be set in the display device to provide voice interaction between the display device and the user, so that the user can control the display device to perform various functions through voice.
  • the voice button on the control device when the user presses the voice button on the control device, the voice of "inquire about the weather forecast of a certain place" is issued.
  • the microphone on the control device receives and recognizes the voice and transmits it to the display device.
  • the display device can respond to the voice command to start Voice assistant application program, and query the weather forecast information of a certain place from the server through the voice assistant application program, and then display the queried weather forecast information on the display and broadcast the queried weather forecast information on the audio output interface.
  • the user makes the voice "I want to watch a movie of a certain star", and after the microphone on the display device receives and recognizes the voice, it can respond to the voice command to start the voice assistant application and send it to the server through the voice assistant application. Query the movie information played by a star, and then display the queried movie information on the display, and broadcast the queried movie information on the audio output interface.
  • the audio output interface usually simultaneously outputs the audio and video content provided by the video-on-demand application.
  • the sound of audio and video content and the sound generated by the interaction between the user and the display device through the voice assistant application In this way, in fact, the user pays more attention to the voice interaction with the display device at this time instead of watching the audio and video content.
  • the voice assistant application It is difficult for the user to distinguish the sound output by the voice assistant application, which does not meet the user's intention of the user currently performing the operation.
  • the microphone not only collects the user's input voice, but also the sound of audio and video content, it affects the accuracy and efficiency of voice recognition to a certain extent.
  • this embodiment provides a method for controlling the output of audio data of an application in a display device, aiming to control the audio output interface in the above example to only play the audio data of the voice assistant application, but not the audio data of the video-on-demand application , So as to meet the user's operating intentions and improve the user's listening experience.
  • the operating system in the memory in FIG. 1D is an Android operating system as an example to describe the output control method of the audio data of each application program in the display device.
  • Figure 2 shows a schematic diagram of the audio data output of each application in the display device.
  • An application program which refers to an application program that needs to output audio data to an audio output interface (such as a speaker).
  • Audiotrack The playback path refers to the path through which the application applies to the display device to transmit audio data.
  • each application program needs to apply for its own corresponding channel to transmit audio data.
  • AudioMixer Audio mixing unit, used to superimpose the audio data transmitted by each application through the corresponding AudioTrack.
  • StreamType Stream type, used to indicate the audio stream type corresponding to the audio data transmitted in AudioTrack. For example, music type (Stream_music), notification type (Stream_notification).
  • Volume Volume, specifically the volume of the audio stream type; here, you can call the setStreamVolume() method to directly set the volume of the audio stream type.
  • Db decibel (decibel), a unit of measurement used to measure audio intensity.
  • App1 to AppN that need to play audio data first apply to the operating system of the display device to transmit AudioTrack1 to AudioTrackN of their respective audio data. Afterwards, the respective audio data is transmitted to AudioMixer for mixing through the applied AudioTrack; during the mixing process, AudioMixer superimposes the volume (Volume) of the audio stream type (StreamType) according to the decibel value (Db) of the audio data transmitted by each AudioTrack Processing (the superposition processing can be briefly described as the Db*Volume algorithm). Finally, AudioMixer outputs the mixed audio data from an audio output interface (such as a speaker) through an audio driver.
  • an audio output interface such as a speaker
  • 3A-3B show a flowchart of a method for controlling the output of audio data of an application in a display device.
  • the method includes the following steps:
  • Step S41 In the process of playing App1 and App2, a play instruction of App3 is received.
  • Step S42 App3 applies to the operating system for AudioTrack that transmits its audio data.
  • Step S43 Determine whether App3 needs to acquire the audio focus; if so, execute step S44; otherwise, end the process.
  • Step S44 Modify StreamType3 of the audio data transmitted in AudioTrack3 created by the operating system for App3.
  • the framework layer obtains the process number of the AudioTrack process currently applying to the operating system, and then queries the package name of App3 corresponding to the process number.
  • the display device can pre-store the application that needs to obtain the audio focus into the white list, for example, set the package name of the voice assistant application in the white list; and the applications in the white list can be dynamically adjusted.
  • the framework layer compares the package name of the application pre-stored in the whitelist with the package name of App3 corresponding to the AudioTrack application process to determine whether App3 needs to obtain audio focus.
  • the package name of the voice assistant application pre-stored in the whitelist. Then, when the framework layer determines that the package name of App3 corresponding to the current application AudioTrack process is a voice assistant application, it determines that App3 needs to obtain audio focus, and then modifies it to StreamType3 of the audio data transmitted in AudioTrack3 created by App3, such as from Stream_music to Stream_special .
  • the StreamType of the audio data transmitted in the AudioTrack created by the operating system for App1 to App3 may all be of the same type, for example, all Stream_music.
  • the StreamType3 of the audio data transmitted in the AudioTrack3 created for App3 is modified from Stream_music to Stream_special, which facilitates the subsequent identification of the AudioTrack whose StreamType has been modified during audio mixing in AudioMixer, and corresponding processing.
  • the framework layer determines that the package name of App3 corresponding to the currently applied AudioTrack process is not a voice assistant application, it determines that App3 does not need to obtain audio focus, and it does not need to perform the step of modifying the StreamType3 of the audio data transmitted in AudioTrack3 created by App3.
  • Step S45 Determine whether there is an AudioTrack3 whose StreamType3 created for App3 is modified in each AudioTrack participating in the mixing; if so, perform step S46 and then perform step S47; otherwise, directly perform step S47.
  • Step S46 Mute the volume of StreamType in other AudioTracks except AudioTrack3 whose StreamType3 is modified.
  • Step S47 superimpose the audio data in each AudioTrack participating in the audio mixing in AudioMixer.
  • Step S48 output the mixed audio data to the audio output interface.
  • the framework layer determines that the StreamType3 of the audio data is modified to the AudioTrack3 created by App3, and when the AudioTrack3 is active, the audio data in AudioTrack1
  • the volume (Volume) of the StreamType2 of the audio data in the StreamType1 and AudioTrack2 is muted (that is, the setStreamVolume() method is called to set the volume of StreamType1 and StreamType2 to 0), and the volume of the StreamType3 of the audio data in AudioTrack3 is retained.
  • AudioTrack1 ⁇ AudioTrack3 participating in the mixing are superimposed according to the Db*Volume algorithm, a mixed data (including the audio data of App1 ⁇ AudioTrack2 after muting and the audio data of App3 without muting) is obtained, thus The user can only hear the audio data of App3 from the audio output interface, which conforms to the user's current operation intention and improves the user's listening experience.
  • the framework layer judges that there is no audio data StreamType3 is modified to be the AudioTrack3 created by App3, it directly superimposes the audio data in AudioTrack1 ⁇ AudioTrack3 participating in the mixing according to the Db*Volume algorithm to obtain the audio of multiple applications Data, so that the user hears audio data of multiple applications from the audio output interface.
  • the framework layer judges whether the AudioTrack is in the active state, specifically refers to the framework layer judging whether the App transmits audio data to the AudioTrack it has applied for. If the App transmits audio data to the AudioTrack it has applied for, the AudioTrack is in the active state; otherwise, the AudioTrack is in the inactive state.
  • step S45 the framework layer determines that AudioTrack3 whose StreamType3 created by App3 is modified is switched to inactive or destroyed (for example, after the display device receives an instruction to stop playing App3 and executes the step of exiting App3), then the framework layer Unmute the volume of StreamType in other AudioTrack in step S46, that is, restore the volume of StreamType in other AudioTrack to the level before mute (that is, call setStreamVolume() method to restore the volume of StreamType1 and StreamType2), and will participate in mixing. After the audio data in each AudioTrack of the audio is superimposed, it is output to the audio output interface.
  • each application in the application layer normally transmits its own audio data to the AudioTrack that it applies for.
  • the framework layer is in AudioMixer, and the output volume of the audio data of the target application that needs to obtain the audio focus is retained by controlling the audio data.
  • the audio data of each application is mixed and output to the audio output interface, so that the user can only listen to the audio output interface to retain the output volume
  • the audio data of the target application to meet the user's current operating intentions.
  • 4A and 4B show a flowchart of an output control method of audio data of another application in a display device. With reference to Figure 4A and Figure 4B, the method includes the following steps.
  • Step S51 In the process of playing App1 to App2, a play instruction of App3 is received.
  • Step S52 App3 applies to the operating system for AudioTrack that transmits its audio data.
  • Step S53 Determine whether App3 needs to acquire the audio focus; if so, execute step S54; otherwise, end the process.
  • Step S54 Modify the flag focus flag for indicating that the audio focus is currently acquired to the process number of AudioTrack3 created by App3.
  • the framework layer obtains the process number of the AudioTrack process currently applying to the operating system, and then queries the package name of App3 corresponding to the process number.
  • the display device can pre-store the application that needs to obtain the audio focus into the white list, for example, set the package name of the voice assistant application in the white list; and the applications in the white list can be dynamically adjusted.
  • the framework layer compares the package name of the application pre-stored in the whitelist with the package name of App3 corresponding to the AudioTrack application process to determine whether App3 needs to obtain audio focus.
  • the package name of the voice assistant application pre-stored in the whitelist. Then, when the framework layer determines that the package name of App3 corresponding to the current application AudioTrack process is a voice assistant application, it determines that App3 needs to obtain audio focus, and then modifies the focus flag that indicates the current audio focus of the operating system to be created by App3
  • the process number of AudioTrack3, that is, the focus flag of the audio focus of the current operating system is activated by the AudioTrack3 requested by App3 to indicate that App3 is currently occupying the audio focus of the operating system.
  • the framework layer judges that the package name of App3 corresponding to the currently applied AudioTrack process is not a voice assistant application, and determines that App3 does not need to obtain audio focus, it may not perform the step of modifying the focus flag indicating the current audio focus acquisition.
  • Step S55 Determine whether the flag focus flag used to indicate the current acquisition of the audio focus is activated by another AudioTrack; if so, perform step S56; otherwise, perform step S57.
  • Step S56 Transmit empty audio data to AudioMixer.
  • Step S57 Transmit the real audio data to AudioMixer.
  • Step S58 superimpose the audio data in each AudioTrack participating in the audio mixing in AudioMixer.
  • Step S59 output the mixed audio data to the audio output interface.
  • the framework layer determines whether the focus flag indicating the current acquisition of the audio focus of the operating system is activated by other AudioTracks, thereby determining whether to transmit real audio data to AudioMixer.
  • App3 such as a voice assistant application
  • the operating system creates AudioTrack3 for it, and changes the focus flag indicating the current audio focus of the operating system to the process number of AudioTrack3 to mark App3 Occupy the audio focus of the operating system.
  • AudioTrack1 applied by App1 transmits its audio data to AudioMixer
  • the framework layer judges that the flag focus used to indicate the current audio focus of the operating system is activated by AudioTrack3, indicating that App1 currently loses the audio focus of the operating system, and then transmits it to AudioMixer For empty audio data, you can write 0x0 data to AudioMixer.
  • AudioTrack2 applied by App2 transmits its audio data to AudioMixer
  • the framework layer judges that the flag focus used to indicate that the audio focus of the operating system is currently acquired is activated by AudioTrack3, indicating that App2 currently loses the audio focus of the operating system, and then sends it to AudioMixer To transmit empty audio data, you can write 0x0 data to AudioMixer.
  • the framework layer determines that the focus flag indicating the current audio focus of the operating system is occupied by AudioTrack3, that is, it is not activated by other AudioTracks, indicating that App3 currently occupies the operating system’s
  • the real audio data sent by App3 to AudioTrack3 is transmitted to AudioMixer.
  • AudioTrack1 AudioTrack1 ⁇ AudioTrack3 participating in the mixing
  • AudioTrack1 applied by App1 and AudioTrack2 applied by App2 both transmit empty audio data to AudioMixer
  • only AudioTrack3 applied by App3 will be real
  • the audio data is transmitted to AudioMixer, so only the audio data of App3 is obtained after mixing, so that the user can only hear the audio data of App3 from the audio output interface, which conforms to the current user's operation intention and improves the user's listening experience.
  • step S53 when the framework layer determines that App3 no longer uses the audio focus (for example, after the display device receives an instruction to stop playing App 3 and executes the step of exiting App3), it will use the flag focus to indicate that the audio focus is currently acquired.
  • the flag is deactivated, that is, the flag focus flag used to indicate the current audio focus is restored to before App3 uses the audio focus, and the audio data in each AudioTrack participating in the mixing is superimposed, and then output to the audio output interface.
  • each application in the application layer normally transmits its audio data to the AudioTrack that it applies for.
  • AudioTrack transmits audio data to AudioMixer by controlling the target application that needs to obtain audio focus.
  • the audio data of the program and the audio data of other applications are not transmitted to AudioMixer through control, and then the audio data of the target application is mixed and output to the audio output interface, so that only what the user listens to from the audio output interface is reserved.
  • the display device only outputs the audio data of the target application by controlling the audio output interface, and does not output the audio data of other applications, so that the user hears only the audio data of the target application through the audio output interface, which improves The user’s listening experience.
  • the embodiments of the present disclosure also provide a computer-readable non-volatile storage medium.
  • the storage medium stores a computer program executable by a computer. When the program is in the When running on a computer, any one of the methods in the foregoing embodiments can be realized when the computer is executed.
  • the aforementioned computer-readable non-volatile storage medium may be any available medium or data storage device that can be accessed by the processor in the server, including but not limited to magnetic storage such as floppy disks, hard disks, magnetic tapes, magneto-optical disks (MO), etc., optical Memory such as compact disc (CD), digital versatile disc (DVD), Blu-ray disc (BD), high-definition versatile disc (HVD), etc., as well as semiconductor memory such as ROM, erasable programmable read-only memory (EPROM), charged erasable Programmable read-only memory (EEPROM), non-volatile memory (NAND FLASH), solid state drive (SSD), etc.
  • magnetic storage such as floppy disks, hard disks, magnetic tapes, magneto-optical disks (MO), etc.
  • optical Memory such as compact disc (CD), digital versatile disc (DVD), Blu-ray disc (BD), high-definition versatile disc (HVD), etc.
  • semiconductor memory such as ROM, erasable
  • the embodiments of the present application can be provided as methods, systems, or computer program products. Therefore, the present application may adopt the form of a complete hardware embodiment, a complete application embodiment, or an embodiment combining applications and hardware. Moreover, this application may adopt the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes.
  • a computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions can also be stored in a computer-readable memory that can guide a computer or other programmable data processing equipment to work in a specific manner, so that the instructions stored in the computer-readable memory produce an article of manufacture including the instruction device.
  • the device implements the functions specified in one process or multiple processes in the flowchart and/or one block or multiple blocks in the block diagram.
  • These computer program instructions can also be loaded on a computer or other programmable data processing equipment, so that a series of operation steps are executed on the computer or other programmable equipment to produce computer-implemented processing, so as to execute on the computer or other programmable equipment.
  • the instructions provide steps for implementing functions specified in a flow or multiple flows in the flowchart and/or a block or multiple blocks in the block diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

本公开实施例公开了一种应用程序的音频数据的输出方法及显示设备。该方法包括:在显示设备通过其音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;响应于所述第一输入指令,控制所述第一音频数据的输出音量被静音,并控制所述第二音频数据的输出音量保留;通过所述音频输出接口,输出被静音后的所述第一音频数据与保留输出音量后的所述第二音频数据的叠加数据。

Description

应用程序的音频数据的输出方法及显示设备
相关申请的交叉引用
本申请要求在2019年04月09日提交中国专利局、申请号为201910281617.8、申请名称为“应用程序的音频数据的输出方法及显示设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本公开涉及音频技术领域,尤其涉及一种应用程序的音频数据的输出方法及显示设备。
背景技术
为了满足用户个性化需求,显示设备不仅可以为用户提供通过广播信号接收发送的各种传统直播节目内容,而且可以为用户提供由内容服务商提供的诸如视频点播、网络游戏、视频聊天等各种应用和服务内容。
在提供上述各种内容的同时,显示设备可以基于用户对诸如遥控器、移动终端等控制装置上的物理硬键或虚拟键的操作而被控制执行与上述各种内容关联的功能,也可以通过自身的麦克风或控制装置上的麦克风接收的用户输入的语音而被控制执行上述各种内容关联的功能。
发明内容
本公开实施例提供一种应用程序的音频数据的输出方法及显示设备,用以实现显示设备中同一时刻仅播放一个应用程序的音频数据,以提高用户听感体验。
第一方面,本公开实施例提供了一种显示设备中应用程序的音频数据的输出方法。
该方法包括:在显示设备通过所述显示设备的音频输出接口,播放一个 或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;响应于所述第一输入指令,控制所述第一音频数据的输出音量被静音,并控制所述第二音频数据的输出音量保留;通过所述音频输出接口,输出被静音后的所述第一音频数据与保留输出音量后的所述第二音频数据的叠加数据。
在一些实施例中,所述方法还包括:在接收到所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;响应于所述第二应用程序需要获取所述音频焦点,修改所述第二音频数据对应的音频流类型。
在一些实施例中,所述确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点,包括:确定所述第二应用程序的包名;判断预存的需要获取所述音频焦点的应用程序的白名单中,是否存在所述第二应用程序的包名;若所述白名单中存在所述第二应用程序的包名,确定所述第二应用程序需要获取所述音频焦点。
在一些实施例中,所述第二应用程序包括语音助手应用程序。
在一些实施例中,所述方法还包括:在接收到所述第一输入指令之后,确定是否存在第三音频数据对应的音频流类型被修改的第三应用程序;响应于存在所述第三应用程序,保留所述第三应用程序的所述第三音频数据对应的音频流类型的音量,静音除所述第三应用程序之外的其它应用程序的音频数据对应的音频流类型的音量。
在一些实施例中,所述第三应用程序为所述第二应用程序。
在一些实施例中,所述方法还包括:接收用于停止播放所述第二音频数据的第二输入指令;响应于所述第二输入指令,控制所述第一音频数据的输出音量,由静音恢复至接收用于播放所述第二音频数据的第一输入指令之前的音量;通过所述音频输出接口,输出恢复输出音量后的所述第一音频数据。
第二方面,本公开实施例提供了一种显示设备中应用程序的音频数据的输出方法。
该方法包括:在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;响应于所述第一输入指令,通过所述音频输出接口,输出所述第二音频数据,不输出所述第一音频数据。
在一些实施例中,所述方法还包括:在接收所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;响应于所述第二应用程序需要获取所述音频焦点,将所述操作系统的音频焦点标记为被所述第二应用程序占用。
在一些实施例中,所述方法还包括:在接收所述第一输入指令之后,确定当前所述显示设备中的操作系统的音频焦点是否被其他应用程序占用;响应于所述音频焦点被所述其他应用程序占用,不向所述音频输出接口输出所述第二音频数据;响应于所述音频焦点没有被所述其他应用程序占用,向所述音频输出接口输出所述第二音频数据。
在一些实施例中,所述方法还包括:接收用于停止播放所述第二音频数据的第二输入指令;响应于所述第二输入指令,通过所述音频输出接口,控制恢复输出所述第一音频数据。
第三方面,本公开实施例提供一种显示设备。该显示设备包括用户接口、音频输出接口、存储器,以及控制器。
所述存储器,与所述用户接口和所述音频输出接口通信,被配置为存储计算机程序、所述用户接口关联的指令,以及所述音频输出接口关联的音频数据。
所述控制器,与所述用户接口、所述音频输出接口以及所述存储器通信,被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;响应于所述第一输入指令,控制所述第一音频数据的输出音量被静音,并控制所述第二音频数据的输出音量保留;通过所述音频输出接口,输出被 静音后的所述第一音频数据与保留输出音量后的所述第二音频数据的叠加数据。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在接收到所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;响应于所述第二应用程序需要获取所述音频焦点,修改所述第二音频数据对应的音频流类型。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:确定所述第二应用程序的包名;判断预存的需要获取所述音频焦点的应用程序的白名单中,是否存在所述第二应用程序的包名;若所述白名单中存在所述第二应用程序的包名,确定所述第二应用程序需要获取所述音频焦点。
在一些实施例中,所述第二应用程序包括语音助手应用程序。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在接收到所述第一输入指令之后,确定是否存在第三音频数据对应的音频流类型被修改的第三应用程序;响应于存在所述第三应用程序,保留所述第三应用程序的所述第三音频数据对应的音频流类型的音量,静音除所述第三应用程序之外的其它应用程序的音频数据对应的音频流类型的音量。
在一些实施例中,所述第三应用程序为所述第二应用程序。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:接收用于停止播放所述第二音频数据的第二输入指令;响应于所述第二输入指令,控制所述第一音频数据的输出音量,由静音恢复至接收用于播放所述第二音频数据的第一输入指令之前的音量;通过所述音频输出接口,输出恢复输出音量后的所述第一音频数据。
第四方面,本公开实施例提供了一种显示设备。该显示设备包括用户接口、音频输出接口、存储器,以及控制器。
所述存储器,与所述用户接口和所述音频输出接口通信,被配置为存储有计算机程序、所述用户接口关联的指令,以及所述音频输出接口关联的音频数据。
所述控制器,与所述用户接口、所述音频输出接口以及所述存储器通信,被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;响应于所述第一输入指令,通过所述音频输出接口,输出所述第二音频数据,不输出所述第一音频数据。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在接收所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;响应于所述第二应用程序需要获取所述音频焦点,将所述操作系统的音频焦点标记为被所述第二应用程序占用。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:在接收所述第一输入指令之后,确定当前所述显示设备中的操作系统的音频焦点是否被其他应用程序占用;响应于所述音频焦点被所述其他应用程序占用,不向所述音频输出接口输出所述第二音频数据;响应于所述音频焦点没有被所述其他应用程序占用,向所述音频输出接口输出所述第二音频数据。
在一些实施例中,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:接收用于停止播放所述第二音频数据的第二输入指令;响应于所述第二输入指令,通过所述音频输出接口,控制恢复输出所述第一音频数据。
第五方面,本公开实施例提供了一种非易失性计算机可读的存储介质,所述存储介质上存储有计算机程序,所述计算机程序被显示设备的处理器执行时实现如第一方面或第二方面所述的方法。
附图说明
为了更清楚地说明本公开实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简要介绍,显而易见地,下面描述中的附图仅仅是本公开的一些实施例,对于本领域的普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1A示例性示出了显示设备200与控制装置100之间操作场景的示意图;
图1B示例性示出了图1A中控制装置100的配置框图;
图1C示例性示出了图1A中显示设备200的配置框图;
图1D示例性示出了显示设备200存储器中操作系统的架构配置框图;
图2示出了显示设备中各应用程序的音频数据输出的架构示意图;
图3A示出了第一种显示设备中应用程序的音频数据的输出控制方法流程图;
图3B示出了第二种显示设备中应用程序的音频数据的输出控制方法的流程图;
图4A示出了第三种显示设备中应用程序的音频数据的输出控制方法的流程图;
图4B示出了第四种显示设备中应用程序的音频数据的输出控制方法的流程图。
具体实施方式
为了使本公开的目的、技术方案和优点更加清楚,下面将结合附图对本公开作进一步地详细描述。显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。
基于本公开中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其它实施例,都属于本公开保护的范围。此外,虽然本公开中公开内容按照示范性一个或几个实例来介绍,但应理解,可以就这些公 开内容的各个方面也可以单独构成一个完整技术方案。
应当理解,本公开中使用的术语“包括”和“具有”以及他们的任何变形,意图在于覆盖但不排他的包含,例如,包含了一系列组件的产品或设备不必限于清楚地列出的那些组件,而是可包括没有清楚地列出的或对于这些产品或设备固有的其它组件。
本公开中使用的术语“模块”,是指任何已知或后来开发的硬件、软件、固件、人工智能、模糊逻辑或硬件或/和软件代码的组合,能够执行与该元件相关的功能。
本公开中使用的术语“手势”,是指用户通过一种手型的变化或手部运动等动作,用于表达预定想法、动作、目的和/或结果的用户行为。
图1A示例性示出了显示设备200与控制装置100之间操作场景的示意图。如图1A所示,控制装置100和显示设备200之间可以有线或无线方式进行通信。
控制装置100被配置为控制显示设备200。控制装置100可接收用户输入的操作指令,并将该操作指令转换为显示设备200可识别和响应的指令,起着实现用户与显示设备200之间信息交互的中介作用。
在一些实施例中,用户通过操作控制装置100上的频道增加或频道减少键,来向显示设备200发出控制频道加减的指令,而显示设备200接收并响应该指令,执行频道加减的操作。
控制装置100可以是遥控器100A。遥控器100A可以实现红外协议通信,蓝牙协议通信,或其他短距离通信等通信方式。遥控器100A通过无线或其他有线方式来控制显示设备200。用户可以通过遥控器100A上的物理或者虚拟按键、语音接口、控制面板接口等,输入用户指令,来控制显示设备200。
一些实施例中,用户可以通过遥控器100A上的音量加减键、频道控制键、上/下/左/右的移动按键、语音输入按键、菜单键、开关机按键等物理或虚拟按键,输入相应的控制指令,来控制显示设备200。
控制装置100也可以是智能设备,如移动终端100B、平板电脑、计算机、 笔记本电脑等。例如,用户可以通过智能设备上运行的应用程序,来控制显示设备200。用户可以对该应用程序进行配置。配置后的应用程序可以被投射到与智能设备相关联的屏幕上,使得用户可以通过直观的用户界面(UI),控制显示设备200。
在一些实施例中,移动终端100B与显示设备200安装相同和/或相关联的软件应用,使得两者可以通过网络通信协议实现通信,进而达到一对一控制操作和数据通信的目的。
在一些实施例中,移动终端100B与显示设备200之间可以建立控制指令协议。用户通过操作移动终端100B的用户界面的各种物理或虚拟按键,来实现相应功能。例如,用户操作遥控器100A上布置的物理按键,实现与该物理按键对应的的功能;或者,移动终端100B上显示的音视频内容能够传输到显示设备200中,从而实现同步显示功能。
显示设备200可提供广播接收功能,计算机支持功能等网络电视功能。显示设备200可以实施为数字电视、网络电视、互联网协议电视(IPTV)等。
显示设备200可以是液晶显示器、有机发光显示器、投影设备。具体显示设备类型、尺寸大小和分辨率等,本申请实施例中不作限定。
显示设备200与服务器300通过多种通信方式进行数据通信。在一些实施例中,显示设备200通过局域网(LAN)、无线局域网(WLAN)和其他网络进行通信连接。
服务器300向显示设备200提供各种内容和互动。
在一些实施例中,显示设备200可以发送和接收信息。
例如,显示设备200接收电子节目指南(Electronic Program Guide,EPG)数据、软件程序更新等,或访问远程储存的数字媒体库。
服务器300可以一组,也可以多组,可以一类或多类服务器。通过服务器300提供视频点播和广告服务等网络服务内容。
图1B中示例性示出了控制装置100的配置框图。如图1B所示,控制装置100包括控制器110、存储器120、通信器130、用户输入接口140、输出 接口150、供电电源160。
控制器110包括随机存取存储器(Random Access Memory,RAM)111、只读存储器(Read-Only Memory,ROM)112、处理器113、通信接口(未示出)以及通信总线(未示出)。
控制器110用于控制控制装置100的运行和操作,以及内部各部件之间的通信协作、外部和内部的数据处理。
在一些实施例中,当检测到用户按压在遥控器100A上布置的按键的交互,或触摸在遥控器100A上布置的触摸面板的交互时,控制器110可产生与检测到的交互相应的控制信号,并将该控制信号发送到显示设备200。
存储器120,用于在控制器110的控制下,存储驱动和控制控制装置100的各种运行程序、数据和应用。
存储器120,可以存储用户输入的各类控制信号指令。
通信器130在控制器110的控制下,实现与显示设备200之间控制信号和数据信号的通信。
在一些实施例中,控制装置100通过通信器130向显示设备200发送触摸信号,按钮信号等控制信号。控制装置100通过通信器130接收由显示设备200发出的信号。
通信器130包括红外信号接口131和射频信号接口132。
在一些实施例中,当通信器130包括红外信号接口时,通信器130需要将用户输入指令,按照红外控制协议转化为红外控制信号,然后经红外发送电路发送至显示设备200。
在一些实施例中,当通信器130包括射频信号接口时,通信器130需将用户输入指令转化为数字信号,然后按照射频控制信号调制协议进行调制后,由射频发送端子发送至显示设备200。
用户输入接口140包括麦克风141、触摸板142、传感器143、按键144等中至少一者,从而用户可以通过语音、触摸、手势、按压等操作,将控制显示设备200的用户指令,输入到控制装置100。
输出接口150,通过将用户输入接口140接收的用户指令输出至显示设备200,或者,输出由显示设备200接收的图像或语音信号。
输出接口150包括LED接口151、产生振动的振动接口152、输出声音的声音输出接口153,和输出图像的显示器154等。
例如,遥控器100A可从输出接口150接收音频、视频或数据等输出信号,并且将该输出信号在显示器154上显示为图像形式,在声音输出接口153输出为音频形式,或在振动接口152输出为振动形式。
供电电源160,用于在控制器110的控制下为控制装置100内的各元件提供运行电力支持。供电电源160被配置为电池及相关控制电路。
图1C中示例性示出了显示设备200的硬件配置框图。如图1C所示,显示设备200中可以包括调谐解调器210、通信器220、检测器230、外部装置接口240、控制器250、存储器260、用户接口265、视频处理器270、显示器275、音频处理器280、音频输入接口285、供电电源290。
调谐解调器210,通过有线或无线方式接收广播电视信号,可以进行放大、混频和谐振等调制解调处理,用于从多个无线或有线广播电视信号中解调出用户所选择的电视频道的频率中所携带的音视频信号,以及附加信息(例如EPG数据)。
调谐解调器210,可根据用户选择,以及由控制器250控制,响应用户选择的电视频道的频率以及该频率所携带的电视信号。
调谐解调器210,根据电视信号的广播制式不同,可以接收信号的途径有很多种,诸如:地面广播、有线广播、卫星广播或互联网广播等;以及根据调制类型不同,可以数字调制方式或模拟调制方式;以及根据接收电视信号的种类不同,可以解调模拟信号和数字信号。
在其他一些示例性实施例中,调谐解调器210也可在外部设备中,如外部机顶盒等。这样,机顶盒通过调制解调后输出电视信号,经过外部装置接口240输入至显示设备200中。
通信器220,是用于根据各种通信协议与外部设备,如外部服务器等,进 行通信的组件。
例如显示设备200可将内容数据发送至经由通信器220连接的外部设备,或者,从经由通信器220连接的外部设备浏览和下载内容数据。
通信器220包括WIFI模块221、蓝牙通信协议模块222、有线以太网通信协议模块223等网络通信协议模块或近场通信协议模块,从而通信器220可根据控制器250的控制,接收控制装置100的控制信号,并将控制信号实现为WIFI信号、蓝牙信号、射频信号等。
检测器230,是显示设备200用于采集外部环境或与外部交互的信号的组件。检测器230可以包括声音采集器231,如麦克风。检测器230用于接收用户声音对应的语音信号,如用户控制显示设备200的控制指令对应的语音信号;或者,检测器230可以采集用于识别环境场景类型的环境声音,实现显示设备200可以自适应环境噪声。
在一些实施例中,检测器230,包括图像采集器232,如相机、摄像头等。检测器230用于采集外部环境场景,以自适应变化显示设备200的显示参数;以及用于采集用户的属性,或与用户进行手势交互,以实现显示设备与用户之间互动的功能。
在一些实施例中,检测器230,包括光接收器,用于采集环境光线强度,以自适应显示设备200的显示参数变化等。
在一些实施例中,检测器230,包括温度传感器,如通过感测环境温度,显示设备200可自适应调整图像的显示色温。
在一些实施例中,在温度偏高的环境中,可调整显示设备200显示图像色温偏冷色调;在温度偏低的环境中,可以调整显示设备200显示图像色温偏暖色调。
外部装置接口240,是提供控制器210,控制显示设备200与外部设备间数据传输的组件。
外部装置接口240可按照有线/无线方式,与诸如机顶盒、游戏装置、笔记本电脑等外部设备连接,可接收外部设备的诸如视频信号(例如运动图像)、 音频信号(例如音乐)、附加信息(例如EPG)等数据。
其中,外部装置接口240包括:高清多媒体接口(High Definition Multimedia Interface,HDMI)端子241、复合视频消隐同步(Composite Video Broadcast Signal,CVBS)端子242、模拟或数字分量端子243、通用串行总线(USB)端子244、组件(Component)端子(未示出)、红绿蓝(RGB)端子(未示出)等中的任一个或多个。
控制器250,通过运行存储在存储器260上的各种软件控制程序(如操作系统和各种应用程序),来控制显示设备200的工作和响应用户的操作。
如图1C所示,控制器250包括RAM251、ROM252、图形处理器253、中央处理器(Central Processing Unit,CPU)处理器254、通信接口255、以及通信总线256。
其中,RAM251、ROM252以及图形处理器253、CPU处理器254通信接口255通过通信总线256相连接。
ROM252,用于存储各种系统启动指令。如在接收到开机信号时,显示设备200的电源开始启动,CPU处理器254运行ROM252中的系统启动指令,将存储在存储器260的操作系统拷贝至RAM251中,以开始运行启动操作系统。当操作系统启动完成后,CPU处理器254再将存储器260中各种应用程序拷贝至RAM251中,然后,开始运行启动各种应用程序。
图形处理器253,用于产生各种图形对象,如图标、操作菜单、以及用户输入指令显示图形等。图形处理器253可以包括运算器,用于通过接收用户输入各种交互指令进行运算,进而根据显示属性显示各种对象;以及包括渲染器,用于产生基于运算器得到的各种对象,将进行渲染的结果显示在显示器275上。
CPU处理器254,用于执行存储在存储器260中的操作系统和应用程序指令。以及根据接收的用户输入指令,来执行各种应用程序、数据和内容的处理,以便最终显示和播放各种音视频内容。
在一些示例性实施例中,CPU处理器254,可以包括多个处理器。多个 处理器可包括一个主处理器以及多个或一个子处理器。主处理器,用于在显示设备预加载模式中执行显示设备200的一些初始化操作,和/或,在正常模式下显示画面的操作。多个或一个子处理器,用于执行在显示设备待机模式等状态下的一种操作。
通信接口255,可包括第一接口到第n接口。这些接口可以是经由网络被连接到外部设备的网络接口。
控制器250可以控制显示设备200的整体操作。例如:响应于接收到用于选择在显示器275上显示的图形用户界面(Graphical User Interface,GUI)对象的用户输入命令,控制器250便可以执行与由用户输入命令选择的对象有关的操作。
其中,该对象可以是可选对象中的任何一个,例如超链接或图标。该与所选择的对象有关的操作,例如显示连接到超链接页面、文档、图像等操作,或者执行与对象相对应的程序的操作。该用于选择GUI对象的用户输入命令,可以是通过连接到显示设备200的各种输入装置,输入命令或者与由用户说出语音相对应的语音命令。例如,输入装置可以是鼠标、键盘、触摸板等。
存储器260,用于存储驱动和控制显示设备200运行的各种类型的数据、软件程序或应用程序。存储器260可以包括易失性和/或非易失性可读存储器。而术语“存储器”包括存储器260、控制器250的RAM251和ROM252、或显示设备200中的存储卡。
在一些实施例中,存储器260具体用于存储驱动显示设备200中控制器250的运行程序;存储显示设备200内置的和用户从外部设备下载的各种应用程序;存储用于配置由显示器275提供的各种GUI、与GUI相关的各种对象及用于选择GUI对象的选择器的视觉效果图像等数据。
在一些实施例中,存储器260具体用于存储调谐解调器210、通信器220、检测器230、外部装置接口240、视频处理器270、显示器275、音频处理器280等的驱动程序和相关数据,从外部装置接口接收的外部数据(例如音视频数据)或用户接口接收的用户数据(例如按键信息、语音信息、触摸信息等)。
在一些实施例中,存储器260具体存储用于表示操作系统(OS)的软件和/或程序,这些软件和/或程序可包括,例如:内核、中间件、应用编程接口(API)和/或应用程序。示例性的,内核可控制或管理系统资源,以及其它程序所实施的功能(如所述中间件、API或应用程序);同时,内核可以提供接口,以允许中间件、API或应用程序访问控制器,以实现控制或管理系统资源。
图1D中示例性示出了显示设备200存储器中操作系统的架构配置框图。该操作系统架构从上到下依次是应用层、框架层和内核层。
应用层,系统内置的应用程序以及非系统级的应用程序都是属于应用层。负责与用户进行直接交互。应用层可包括多个应用程序,如直播电视应用程序、视频点播应用程序、媒体中心应用程序、截图应用程序等。
直播电视应用程序,可以通过不同的信号源提供直播电视。例如,直播电视应用程可以使用来自有线电视、无线广播、卫星服务或其他类型的直播电视服务的输入提供电视信号。以及,直播电视应用程序可在显示设备200上显示直播电视信号的视频。
视频点播应用程序,可以提供来自不同存储源的视频。不同于直播电视应用程序,视频点播提供来自某些存储源的视频显示。例如,视频点播可以来自云存储的服务器端、来自包含已存视频节目的本地硬盘储存器。
媒体中心应用程序,可以提供各种多媒体内容播放的应用程序。例如,媒体中心,可以为不同于直播电视或视频点播,用户可通过媒体中心应用程序访问存储器内存储的各种图像或音视频。
截图应用程序,可以对显示器上当前显示画面进行截图,并对画面截图图像中包含的识别物(如人物、频道台标、建筑等)进行诸如标识识别框、名称等的标注,以为用户提供显示画面中包含的各种识别物信息的展示功能。当前显示画面,可以是文字、图像、视频中的至少一个。
框架层,负责提供应用层所需要的API。例如,直播电视应用程序、视频点播应用程序、媒体中心应用程序可通过框架层提供的接口调用解码器进行音视频解码。又如,截图应用程序可通过框架层提供的接口调用已截取的当 前显示画面的截图图像。
内核层,提供核心系统服务,例如:文件管理、内存管理、进程管理、网络管理、系统安全权限管理等服务。内核层可以被实现为基于各种操作系统的内核,例如,基于安卓操作系统的内核。
内核也同时提供系统软件和硬件之间的通信,为各种硬件提供设备驱动服务,例如:为显示器提供显示驱动程序、为摄像头提供摄像头驱动程序、为遥控器提供按键驱动程序、为WIFI模块提供WiFi驱动程序、为音频输出接口提供音频驱动程序、为电源管理(PM)模块提供电源管理驱动等。
用户接口265,接收各种用户交互。具体的,用于将用户的输入信号发送给控制器250,或者,将从控制器250的输出信号传送给用户。示例性的,遥控器100A可将用户输入的诸如电源开关信号、频道选择信号、音量调节信号等输入信号发送至用户接口265,再由用户接口265转送至控制器250;或者,遥控器100A可接收经控制器250处理从用户接口265输出的音频、视频或数据等输出信号,并且显示接收的输出信号或将接收的输出信号输出为音频或振动形式。
在一些实施例中,用户可在显示器275上显示的GUI输入用户命令,则用户接口265通过GUI接收用户输入命令。确切的说,用户接口265可接收用于控制选择器在GUI中的位置以选择不同的对象或项目的用户输入命令。
或者,用户可通过输入特定的声音或手势进行输入用户命令,则用户接口265通过传感器识别出声音或手势,来接收用户输入命令。视频处理器270,用于接收外部的视频信号,根据输入信号的标准编解码协议,进行解压缩、解码、缩放、降噪、帧率转换、分辨率转换、图像合成等视频数据处理,可得到直接在显示器275上显示或播放的视频信号。
示例的,视频处理器270,包括解复用模块、视频解码模块、图像合成模块、帧率转换模块、显示格式化模块等。
其中,解复用模块,用于对输入音视频数据流进行解复用处理,如输入MPEG-2流(基于数字存储媒体运动图像和语音的压缩标准),则解复用模块 将其进行解复用成视频信号和音频信号等。
视频解码模块,用于对解复用后的视频信号进行处理,包括解码和缩放处理等。
图像合成模块,如图像合成器,其用于将图形生成器根据用户输入或自身生成的GUI信号,与缩放处理后视频图像进行叠加混合处理,以生成可供显示的图像信号。
帧率转换模块,用于对输入视频的帧率进行转换,如将输入的60Hz视频的帧率转换为120Hz或240Hz的帧率,通常的格式采用如插帧方式实现。
显示格式化模块,用于将帧率转换模块输出的信号,改变为符合诸如显示器显示格式的信号,如将帧率转换模块输出的信号进行格式转换以输出RGB数据信号。
显示器275,用于接收源自视频处理器270输入的图像信号,进行显示视频内容、图像以及菜单操控界面。显示视频内容,可以来自调谐解调器210接收的广播信号中的视频内容,也可以来自通信器220或外部装置接口240输入的视频内容。显示器275,同时显示显示设备200中产生且用于控制显示设备200的用户操控界面UI。
以及,显示器275可以包括用于呈现画面的显示屏组件以及驱动图像显示的驱动组件。或者,倘若显示器275为一种投影显示器,还可以包括一种投影装置和投影屏幕。
音频处理器280,用于接收外部的音频信号,根据输入信号的标准编解码协议,进行解压缩和解码,以及降噪、数模转换、和放大处理等音频数据处理,得到可以在扬声器286中播放的音频信号。
示例性的,音频处理器280可以支持各种音频格式。例如MPEG-2、MPEG-4、高级音频编码(AAC)、高效AAC(HE-AAC)等格式。
音频输出接口285,用于在控制器250的控制下接收音频处理器280输出的音频信号,音频输出接口285可包括扬声器286,或输出至外接设备的发生装置的外接音响输出端子287,如耳机输出端子。
在一些实施例中,视频处理器270可以包括一个或多个芯片组成。音频处理器280,也可以包括一个或多个芯片组成。
以及,在其他一些示例性实施例中,视频处理器270和音频处理器280,可以为单独的芯片,也可以与控制器250一起集成在一个或多个芯片中。
供电电源290,用于在控制器250的控制下,将外部电源输入的电力为显示设备200提供电源供电支持。供电电源290可以是安装在显示设备200内部的内置电源电路,也可以是安装在显示设备200外部的电源。需要说明的是,在上述图1A-1D的基础上,显示设备可以通过自身的声音采集器231(可实现为麦克风)或控制装置上的麦克风141接收用户输入的语音,而被控制执行其提供的各种内容所需的功能。
其中,显示设备内可设置语音助手应用程序,用于提供显示设备和用户之间的语音交互,从而实现用户通过语音控制显示设备执行各种功能。
例如,用户通过按压控制装置上的语音键的同时发出“查询某地天气预报”的语音,控制装置上的麦克风接收并识别该语音后传送至显示设备,显示设备可以响应于该语音指令,启动语音助手应用程序,并通过语音助手应用程序向服务器查询某地天气预报信息,进而在显示器上展示查询的天气预报信息,以及在音频输出接口播报查询的天气预报信息。
又如,用户发出“我要看某明星的电影”的语音,显示设备上的麦克风接收并识别该语音后,可以响应于该语音指令,启动语音助手应用程序,并通过语音助手应用程序向服务器查询某明星饰演的电影信息,进而在显示器上展示查询的电影信息,以及在音频输出接口播报查询的电影信息。
然而,例如,在显示设备正在播放一视频点播应用程序提供的音视频内容的过程中,用户需要对显示设备执行上述语音控制功能,此时,音频输出接口通常同时输出通过视频点播应用程序提供的音视频内容的声音、以及用户与显示设备之间通过语音助手应用程序进行交互而产生的声音,这样,实际上用户此时更关注与显示设备的语音交互,而非观看音视频内容,从而使得用户不易区分出语音助手应用程序输出的声音,不符合用户当前执行操作 的用户意图。同时,由于麦克风不仅仅采集到用户输入语音,也可以采集到音视频内容的声音,一定程度上影响语音识别的准确率和效率。
因此,本实施例提供一种显示设备中应用程序的音频数据的输出控制方法,旨在上述示例中控制音频输出接口仅播放语音助手应用程序的音频数据,而不播放视频点播应用程序的音频数据,从而符合用户操作意图,提高用户听感体验。
本实施例中以图1D中存储器中的操作系统是安卓操作系统为例,对显示设备中各应用程序的音频数据的输出控制方法进行描述。
图2示出了显示设备中各应用程序的音频数据输出的架构示意图。
下面首先对本实施中涉及的安卓操作系统中的相关术语进行解释。
App:应用程序,指需要将音频数据输出到音频输出接口(如扬声器)的应用程序。
Audiotrack:回放通路,指应用程序向显示设备申请的传输音频数据的通路。这里,每个应用程序需要申请自身对应的传输音频数据的通路。
AudioMixer:混音单元,用于将各应用程序通过相应的AudioTrack传输的音频数据进行叠加处理。
StreamType:流类型,用于表示AudioTrack中传输的音频数据对应的音频流类型。例如音乐类型(Stream_music)、通知类型(Stream_notification)。
Volume:音量,具体为音频流类型的音量;这里,可以调用setStreamVolume()方法来直接设置音频流类型的音量大小。
Db:分贝(decibel),用于度量音频强度的计量单位。
还需说明的是,安卓操作系统中,在显示设备播放一App的音频数据时,也即该App使用诸如AudioMixer、音频输出接口等音频资源时,需要先向显示设备的操作系统申请音频焦点(AudioFocus),在该App获得音频焦点之后才可以使用音频资源。
如图2所示,需要播放音频数据的App1~AppN,首先向显示设备的操作系统申请传输各自音频数据的AudioTrack1~AudioTrackN。之后通过申请的 AudioTrack将各自音频数据传输至AudioMixer进行混音;在混音过程中,AudioMixer依据各AudioTrack传输的音频数据的分贝值(Db)与音频流类型(StreamType)的音量(Volume)进行叠加处理(该叠加处理可以简述为Db*Volume算法)。最后,AudioMixer将混音后的音频数据通过音频驱动程序从音频输出接口(如扬声器)输出。
图3A-3B示出了显示设备中应用程序的音频数据的输出控制方法流程图。
结合图3A-3B来说,该方法包括以下步骤:
步骤S41:在播放App1和App2的过程中,接收到App3的播放指令。
步骤S42:App3向操作系统申请传输其音频数据的AudioTrack。
步骤S43:判断App3是否需要获取音频焦点;若是,则执行步骤S44;否则,结束流程。
步骤S44:修改操作系统为App3创建的AudioTrack3中传输的音频数据的StreamType3。
具体的,App3向操作系统申请AudioTrack时,框架层获取当前向操作系统申请AudioTrack进程的进程号,进而查询到该进程号对应的App3的包名。显示设备内可以预存需要获取音频焦点的应用程序至白名单中,例如将语音助手应用程序的包名设置在白名单中;且该白名单中的应用程序可以动态调整。这样,框架层通过将白名单中预存的应用程序的包名与申请AudioTrack进程对应的App3的包名进行比较,从而判断App3是否需要获取音频焦点。
例如,白名单中预存语音助手应用程序的包名。那么,框架层判断当前申请AudioTrack进程对应的App3的包名是语音助手应用程序时,确定App3需要获取音频焦点,进而修改为App3创建的AudioTrack3中传输的音频数据的StreamType3,如从Stream_music修改为Stream_special。
在一些实施例中,操作系统为App1到App3创建的AudioTrack中传输的音频数据的StreamType可以均为相同类型,如均为Stream_music。这样,将为App3创建的AudioTrack3中传输的音频数据的StreamType3由Stream_music 修改为Stream_special,可以便于后续在AudioMixer中混音时对该StreamType被修改的AudioTrack进行识别,从而作出相应处理。
框架层判断当前申请AudioTrack进程对应的App3的包名不是语音助手应用程序时,确定App3不需要获取音频焦点,则可不执行修改为App3创建的AudioTrack3中传输的音频数据的StreamType3的步骤。
步骤S45:判断参与混音的各AudioTrack中是否存在为App3创建的StreamType3被修改的AudioTrack3;若是,则执行步骤S46后再执行步骤S47;否则,直接执行步骤S47。
步骤S46:将除StreamType3被修改的AudioTrack3之外的其它AudioTrack中StreamType的音量静音。
步骤S47:将参与混音的各AudioTrack中的音频数据在AudioMixer中进行叠加。
步骤S48:将混音后的音频数据输出至音频输出接口。
具体的,在AudioMixer中进行混音时,参与混音的AudioTrack1~AudioTrack3中,框架层判断存在音频数据的StreamType3被修改的为App3创建的AudioTrack3、同时该AudioTrack3处于激活状态时,将AudioTrack1中音频数据的StreamType1和AudioTrack2中音频数据的StreamType2的音量(Volume)进行静音(即分别调用setStreamVolume()方法将StreamType1和StreamType2的音量设置为0),将AudioTrack3中音频数据的StreamType3的音量进行保留,这样在将参与混音的AudioTrack1~AudioTrack3中的音频数据按照Db*Volume算法进行叠加处理时,得出一混音数据(包括静音后的App1~AudioTrack2的音频数据和未作静音的App3的音频数据),从而用户从音频输出接口仅能听到App3的音频数据,符合用户当前的操作意图,提高了用户的听感体验。
框架层判断不存在音频数据的StreamType3被修改的为App3创建的AudioTrack3时,直接将参与混音的AudioTrack1~AudioTrack3中的音频数据按照Db*Volume算法进行叠加处理,这样得出多个应用程序的音频数据,从 而用户从音频输出接口听到多个应用程序的音频数据。
这里,框架层判断AudioTrack是否处于激活状态,具体指框架层判断App是否向其申请的AudioTrack传输音频数据。若App向其申请的AudioTrack传输音频数据,则该AudioTrack处于激活状态;否则,该AudioTrack处于非激活状态。
此外,在步骤S45中,框架层判断为App3创建的StreamType3被修改的AudioTrack3切换为非激活状态或者已销毁(如显示设备接收到停止播放App3的指令时执行退出App3的步骤后),则框架层将步骤S46中其它AudioTrack中StreamType的音量进行解静音,也即将其它AudioTrack中StreamType的音量恢复为静音之前的级别(即分别调用setStreamVolume()方法将StreamType1和StreamType2的音量恢复之前),并且将参与混音的各AudioTrack中的音频数据进行叠加后,再输出至音频输出接口。
如上面实施例所述,应用层中各应用程序正常向其申请的AudioTrack中传输各自的音频数据,框架层在AudioMixer中,通过控制需要获取音频焦点的目标应用程序的音频数据的输出音量被保留、以及通过控制其他应用程序的音频数据的输出音量被静音,之后将各应用程序的音频数据进行混音处理后输出至音频输出接口,这样用户从音频输出接口收听到的仅有保留输出音量的目标应用程序的音频数据,以符合用户当前的操作意图。图4A和图4B示出了显示设备中另一应用程序的音频数据的输出控制方法的流程图。结合图4A和图4B来说,该方法包括以下步骤。
步骤S51:在播放App1~App2的过程中,接收到App3的播放指令。
步骤S52:App3向操作系统申请传输其音频数据的AudioTrack。
步骤S53:判断App3是否需要获取音频焦点;若是,则执行步骤S54;否则,结束流程。
步骤S54:将用于指示当前获取音频焦点的标志focus flag修改成为App3创建的AudioTrack3的进程号。
具体的,App3向操作系统申请AudioTrack时,框架层获取当前向操作系 统申请AudioTrack进程的进程号,进而查询到该进程号对应的App3的包名。显示设备内可以预存需要获取音频焦点的应用程序至白名单中,例如将语音助手应用程序的包名设置在白名单中;且该白名单中的应用程序可以动态调整。这样,框架层通过将白名单中预存的应用程序的包名与申请AudioTrack进程对应的App3的包名进行比较,从而判断App3是否需要获取音频焦点。
例如,白名单中预存语音助手应用程序的包名。那么,框架层判断当前申请AudioTrack进程对应的App3的包名是语音助手应用程序时,确定App3需要获取音频焦点,进而将用于指示当前获取操作系统的音频焦点的标志focus flag修改成为App3创建的AudioTrack3的进程号,也即当前操作系统的音频焦点的focus flag被App3申请的AudioTrack3所激活,以表明当前是App3占用操作系统的音频焦点。
框架层判断当前申请AudioTrack进程对应的App3的包名不是语音助手应用程序时,确定App3不需要获取音频焦点,则可不执行修改用于指示当前获取音频焦点的标志focus flag的步骤。
步骤S55:判断用于指示当前获取音频焦点的标志focus flag是否被其他AudioTrack激活;若是,则执行步骤S56;否则,执行步骤S57。
步骤S56:向AudioMixer中传输空的音频数据。
步骤S57:向AudioMixer中传输真实的音频数据。
步骤S58:将参与混音的各AudioTrack中的音频数据在AudioMixer中进行叠加。
步骤S59:将混音后的音频数据输出至音频输出接口。
具体的,各AudioTrack在向AudioMixer传输音频数据时,框架层判断用于指示当前获取操作系统的音频焦点的标志focus flag是否被其他AudioTrack激活,从而判断是否向AudioMixer中传输真实的音频数据。
例如,App3(如语音助手应用程序)向操作系统申请AudioTrack后,操作系统为其创建AudioTrack3,并将用于指示当前获取操作系统的音频焦点的标志focus flag修改成AudioTrack3的进程号,以标记App3占用操作系统的 音频焦点。
那么,App1申请的AudioTrack1向AudioMixer传输其音频数据时,框架层判断用于指示当前获取操作系统的音频焦点的标志focus flag被AudioTrack3激活,表明App1当前失去操作系统的音频焦点,则向AudioMixer中传输空的音频数据,具体可向AudioMixer中写入0x0数据。
同理,App2申请的AudioTrack2向AudioMixer传输其音频数据时,框架层判断用于指示当前获取操作系统的音频焦点的标志focus flag被AudioTrack3激活,表明App2当前失去操作系统的音频焦点,则向AudioMixer中传输空的音频数据,具体可向AudioMixer中写入0x0数据。
然而,App3申请的AudioTrack3向AudioMixer传输其音频数据时,框架层判断用于指示当前获取操作系统的音频焦点的标志focus flag被AudioTrack3占用,也即未被其他AudioTrack激活,表明App3当前占用操作系统的音频焦点,则向AudioMixer中传输由App3送至AudioTrack3的真实的音频数据。
这样,在将参与混音的AudioTrack1~AudioTrack3中的音频数据在AudioMixer中进行叠加时,由于App1申请的AudioTrack1和App2申请的AudioTrack2均将空的音频数据传输至AudioMixer,仅App3申请的AudioTrack3将真实的音频数据传输至AudioMixer,所以,混音后仅得出App3的音频数据,从而用户从音频输出接口仅能听到App3的音频数据,符合当前用户的操作意图,提高了用户的听感体验。
此外,在步骤S53中,框架层判断App3不再使用音频焦点(如显示设备接收到停止播放App 3的指令时执行退出App3的步骤后)时,则将用于指示当前获取音频焦点的标志focus flag去激活,也即将用于指示当前获取音频焦点的标志focus flag恢复为App3使用音频焦点之前,并且将参与混音的各AudioTrack中的音频数据进行叠加后,再输出至音频输出接口。
如上面实施例所述,应用层中各应用程序正常向其申请的AudioTrack中传输各自的音频数据,框架层中AudioTrack向AudioMixer传输音频数据过程 中,通过控制向AudioMixer传输需要获取音频焦点的目标应用程序的音频数据、以及通过控制向AudioMixer不传输其他应用程序的音频数据,之后将目标应用程序的音频数据进行混音处理后输出至音频输出接口,这样用户从音频输出接口收听到的仅有保留的真实的目标应用程序的音频数据,以符合用户当前的操作意图。
在上述实施例中,显示设备通过控制音频输出接口仅输出目标应用程序的音频数据,不输出其他应用程序的音频数据,从而用户通过音频输出接口听到的仅有目标应用程序的音频数据,提高用户的听感体验。
在上述各实施例的基础上,本公开实施例还提供了一种计算机可读的非易失性存储介质,所述存储介质内存储有可由计算机执行的计算机程序,当所述程序在所述计算机上运行时,使得所述计算机执行时实现上述实施例中的任一一种方法。
上述计算机可读非易失性存储介质可以是服务器中的处理器能够存取的任何可用介质或数据存储设备,包括但不限于磁性存储器如软盘、硬盘、磁带、磁光盘(MO)等、光学存储器如光盘(CD)、数字通用光盘(DVD)、蓝光光盘(BD)、高清通用光盘(HVD)等、以及半导体存储器如ROM、可擦除可编程只读存储器(EPROM)、带电可擦除可编程只读存储器(EEPROM)、非易失性存储器(NAND FLASH)、固态硬盘(SSD)等。
对于系统/装置实施例而言,由于其基本相似于方法实施例,所以描述的比较简单,相关之处参见方法实施例的部分说明即可。
需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者一个操作与另一个实体或者另一个操作区分开来,而不一定要求或者暗示这些实体或者操作之间存在任何这种实际的关系或者顺序。
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全应用实施例、或结合应用和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘 存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。
本公开是参照根据本公开实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
尽管已描述了本公开的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本公开范围的所有变更和修改。
显然,本领域的技术人员可以对本公开进行各种改动和变型而不脱离本公开的精神和范围。这样,倘若本公开的这些修改和变型属于本公开权利要求及其等同技术的范围之内,则本公开也意图包括这些改动和变型在内。

Claims (23)

  1. 一种显示设备中应用程序的音频数据的输出方法,其特征在于,所述方法包括:
    在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;
    响应于所述第一输入指令,控制所述第一音频数据的输出音量被静音,并控制所述第二音频数据的输出音量保留;
    通过所述音频输出接口,输出被静音后的所述第一音频数据与保留输出音量后的所述第二音频数据的叠加数据。
  2. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    在接收到所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;
    响应于所述第二应用程序需要获取所述音频焦点,修改所述第二音频数据对应的音频流类型。
  3. 根据权利要求2所述的方法,其特征在于,所述确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点,包括:
    确定所述第二应用程序的包名;
    判断预存的需要获取所述音频焦点的应用程序的白名单中,是否存在所述第二应用程序的包名;
    若所述白名单中存在所述第二应用程序的包名,确定所述第二应用程序需要获取所述音频焦点。
  4. 根据权利要求3所述的方法,其特征在于,所述第二应用程序包括语音助手应用程序。
  5. 根据权利要求3所述的方法,其特征在于,所述方法还包括:
    在接收到所述第一输入指令之后,确定是否存在第三音频数据对应的音频流 类型被修改的第三应用程序;
    响应于存在所述第三应用程序,保留所述第三应用程序的所述第三音频数据对应的音频流类型的音量,静音除所述第三应用程序之外的其它应用程序的音频数据对应的音频流类型的音量。
  6. 根据权利要求5所述的方法,其特征在于,所述第三应用程序为所述第二应用程序。
  7. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    接收用于停止播放所述第二音频数据的第二输入指令;
    响应于所述第二输入指令,控制所述第一音频数据的输出音量,由静音恢复至接收用于播放所述第二音频数据的第一输入指令之前的音量;
    通过所述音频输出接口,输出恢复输出音量后的所述第一音频数据。
  8. 一种显示设备中应用程序的音频数据的输出方法,其特征在于,所述方法包括:
    在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;
    响应于所述第一输入指令,通过所述音频输出接口,输出所述第二音频数据,不输出所述第一音频数据。
  9. 根据权利要求8所述的方法,其特征在于,所述方法还包括:
    在接收所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;
    响应于所述第二应用程序需要获取所述音频焦点,将所述操作系统的音频焦点标记为被所述第二应用程序占用。
  10. 根据权利要求8所述的方法,其特征在于,所述方法还包括:
    在接收所述第一输入指令之后,确定当前所述显示设备中的操作系统的音频焦点是否被其他应用程序占用;
    响应于所述音频焦点被所述其他应用程序占用,不向所述音频输出接口 输出所述第二音频数据;
    响应于所述音频焦点没有被所述其他应用程序占用,向所述音频输出接口输出所述第二音频数据。
  11. 根据权利要求8所述的方法,其特征在于,所述方法还包括:
    接收用于停止播放所述第二音频数据的第二输入指令;
    响应于所述第二输入指令,通过所述音频输出接口,控制恢复输出所述第一音频数据。
  12. 一种显示设备,其特征在于,包括用户接口、音频输出接口、存储器,以及控制器,其中,
    所述存储器,与所述用户接口和所述音频输出接口通信,被配置为存储计算机程序、所述用户接口关联的指令,以及所述音频输出接口关联的音频数据;
    所述控制器,与所述用户接口、所述音频输出接口以及所述存储器通信,被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;
    响应于所述第一输入指令,控制所述第一音频数据的输出音量被静音,并控制所述第二音频数据的输出音量保留;
    通过所述音频输出接口,输出被静音后的所述第一音频数据与保留输出音量后的所述第二音频数据的叠加数据。
  13. 根据权利要求12所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在接收到所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;
    响应于所述第二应用程序需要获取所述音频焦点,修改所述第二音频数据对应的音频流类型。
  14. 根据权利要求13所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    确定所述第二应用程序的包名;
    判断预存的需要获取所述音频焦点的应用程序的白名单中,是否存在所述第二应用程序的包名;
    若所述白名单中存在所述第二应用程序的包名,确定所述第二应用程序需要获取所述音频焦点。
  15. 根据权利要求14所述的显示设备,其特征在于,所述第二应用程序包括语音助手应用程序。
  16. 根据权利要求14所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在接收到所述第一输入指令之后,确定是否存在第三音频数据对应的音频流类型被修改的第三应用程序;
    响应于存在所述第三应用程序,保留所述第三应用程序的所述第三音频数据对应的音频流类型的音量,静音除所述第三应用程序之外的其它应用程序的音频数据对应的音频流类型的音量。
  17. 根据权利要求16所述的显示设备,其特征在于,所述第三应用程序为所述第二应用程序。
  18. 根据权利要求12所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    接收用于停止播放所述第二音频数据的第二输入指令;
    响应于所述第二输入指令,控制所述第一音频数据的输出音量,由静音恢复至接收用于播放所述第二音频数据的第一输入指令之前的音量;
    通过所述音频输出接口,输出恢复输出音量后的所述第一音频数据。
  19. 一种显示设备,其特征在于,包括用户接口、音频输出接口、存储器,以及控制器,其中,
    所述存储器,与所述用户接口和所述音频输出接口通信,被配置为存储 有计算机程序、所述用户接口关联的指令,以及所述音频输出接口关联的音频数据;
    所述控制器,与所述用户接口、所述音频输出接口以及所述存储器通信,被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在显示设备通过所述显示设备的音频输出接口,播放一个或多个第一应用程序的第一音频数据的过程中,接收用于播放第二应用程序的第二音频数据的第一输入指令;
    响应于所述第一输入指令,通过所述音频输出接口,输出所述第二音频数据,不输出所述第一音频数据。
  20. 根据权利要求19所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在接收所述第一输入指令之后,确定所述第二应用程序是否需要获取所述显示设备中的操作系统的音频焦点;
    响应于所述第二应用程序需要获取所述音频焦点,将所述操作系统的音频焦点标记为被所述第二应用程序占用。
  21. 根据权利要求19所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    在接收所述第一输入指令之后,确定当前所述显示设备中的操作系统的音频焦点是否被其他应用程序占用;
    响应于所述音频焦点被所述其他应用程序占用,不向所述音频输出接口输出所述第二音频数据;
    响应于所述音频焦点没有被所述其他应用程序占用,向所述音频输出接口输出所述第二音频数据。
  22. 根据权利要求19所述的显示设备,其特征在于,所述控制器还被配置为运行所述计算机程序和/或所述指令使得所述显示设备:
    接收用于停止播放所述第二音频数据的第二输入指令;
    响应于所述第二输入指令,通过所述音频输出接口,控制恢复输出所述 第一音频数据。
  23. 一种非易失性计算机可读的存储介质,其特征在于,所述存储介质上存储有计算机程序,所述计算机程序被显示设备的处理器执行时实现如权利要求1-11中任一项所述的方法。
PCT/CN2020/076488 2019-04-09 2020-02-24 应用程序的音频数据的输出方法及显示设备 WO2020207132A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/847,100 US11288033B2 (en) 2019-04-09 2020-04-13 Method for outputting audio data of applications and display device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910281617.8A CN110032357B (zh) 2019-04-09 2019-04-09 应用程序的音频数据的输出方法及显示设备
CN201910281617.8 2019-04-09

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/847,100 Continuation US11288033B2 (en) 2019-04-09 2020-04-13 Method for outputting audio data of applications and display device

Publications (1)

Publication Number Publication Date
WO2020207132A1 true WO2020207132A1 (zh) 2020-10-15

Family

ID=67237792

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/076488 WO2020207132A1 (zh) 2019-04-09 2020-02-24 应用程序的音频数据的输出方法及显示设备

Country Status (2)

Country Link
CN (1) CN110032357B (zh)
WO (1) WO2020207132A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112911381A (zh) * 2021-01-18 2021-06-04 海信视像科技股份有限公司 一种显示设备、模式调整方法、设备及介质
CN113423006A (zh) * 2021-05-31 2021-09-21 惠州华阳通用电子有限公司 一种基于主副声道的多音频流混音播放方法及系统
CN113794928A (zh) * 2021-09-14 2021-12-14 青岛海信传媒网络技术有限公司 一种音频播放方法及显示设备

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11288033B2 (en) 2019-04-09 2022-03-29 Hisense Visual Technology Co., Ltd. Method for outputting audio data of applications and display device
CN110032357B (zh) * 2019-04-09 2022-05-10 海信视像科技股份有限公司 应用程序的音频数据的输出方法及显示设备
CN111343498B (zh) * 2020-03-10 2021-09-10 海信视像科技股份有限公司 一种静音控制方法、装置及智能电视
CN111601158B (zh) * 2020-05-14 2021-11-02 青岛海信传媒网络技术有限公司 一种流媒体管道切音轨的优化方法及显示设备
CN112086091A (zh) * 2020-09-18 2020-12-15 南京孝德智能科技有限公司 一种智能化养老服务系统及方法
CN112346699A (zh) * 2020-10-29 2021-02-09 深圳Tcl新技术有限公司 音频应用的音量控制方法、装置、终端及可读存储介质
CN113225495B (zh) * 2021-04-23 2022-09-02 深圳创维-Rgb电子有限公司 消除视频播放杂音的方法以及装置
CN114302248B (zh) * 2021-04-30 2024-04-12 海信视像科技股份有限公司 一种显示设备及多窗口语音播报方法
CN113518258B (zh) * 2021-05-14 2023-06-30 北京天籁传音数字技术有限公司 一种低延迟全场景音频实现方法、装置和电子设备
CN113076082B (zh) * 2021-06-03 2021-09-03 杭州星犀科技有限公司 音频源处理的方法、系统、电子装置和存储介质
CN114416011B8 (zh) * 2021-11-12 2024-04-05 青岛海信移动通信技术有限公司 一种终端、音频控制方法及存储介质
CN117201665B (zh) * 2023-08-26 2024-03-29 哈尔滨师范大学 一种数据处理方法、装置及系统

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160342384A1 (en) * 2012-09-28 2016-11-24 Pantech Co., Ltd. Mobile terminal and method for controlling sound output
CN106648534A (zh) * 2016-12-26 2017-05-10 三星电子(中国)研发中心 一种对互斥的音频实现同时播放的方法
CN107145328A (zh) * 2017-04-10 2017-09-08 北京猎户星空科技有限公司 一种音频切换方法、装置及智能设备
CN110032357A (zh) * 2019-04-09 2019-07-19 青岛海信电器股份有限公司 应用程序的音频数据的输出方法及显示设备

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9342516B2 (en) * 2011-05-18 2016-05-17 Microsoft Technology Licensing, Llc Media presentation playback annotation
CN106598539B (zh) * 2016-12-15 2020-02-14 广州酷狗计算机科技有限公司 一种应用程序内音频的处理方法和装置
CN107040660A (zh) * 2017-04-18 2017-08-11 广东欧珀移动通信有限公司 移动终端播放音量的控制方法、装置及移动终端
CN107832034A (zh) * 2017-10-26 2018-03-23 深圳市沃特沃德股份有限公司 音频播放控制方法和装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160342384A1 (en) * 2012-09-28 2016-11-24 Pantech Co., Ltd. Mobile terminal and method for controlling sound output
CN106648534A (zh) * 2016-12-26 2017-05-10 三星电子(中国)研发中心 一种对互斥的音频实现同时播放的方法
CN107145328A (zh) * 2017-04-10 2017-09-08 北京猎户星空科技有限公司 一种音频切换方法、装置及智能设备
CN110032357A (zh) * 2019-04-09 2019-07-19 青岛海信电器股份有限公司 应用程序的音频数据的输出方法及显示设备

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112911381A (zh) * 2021-01-18 2021-06-04 海信视像科技股份有限公司 一种显示设备、模式调整方法、设备及介质
CN113423006A (zh) * 2021-05-31 2021-09-21 惠州华阳通用电子有限公司 一种基于主副声道的多音频流混音播放方法及系统
CN113794928A (zh) * 2021-09-14 2021-12-14 青岛海信传媒网络技术有限公司 一种音频播放方法及显示设备
CN113794928B (zh) * 2021-09-14 2023-07-25 Vidaa(荷兰)国际控股有限公司 一种音频播放方法及显示设备

Also Published As

Publication number Publication date
CN110032357B (zh) 2022-05-10
CN110032357A (zh) 2019-07-19

Similar Documents

Publication Publication Date Title
WO2020207132A1 (zh) 应用程序的音频数据的输出方法及显示设备
WO2021109487A1 (zh) 音频数据流输出的控制方法及显示设备
CN111277884B (zh) 视频播放方法及设备
US11288033B2 (en) Method for outputting audio data of applications and display device
WO2021031629A1 (zh) 显示设备和控制装置按键复用方法
WO2021164177A1 (zh) 播放媒体资源的方法、显示设备及移动终端
US11425466B2 (en) Data transmission method and device
CN112188279A (zh) 一种频道切换方法和显示设备
CN112214189A (zh) 图像显示方法及显示设备
CN113938724A (zh) 显示设备及录屏分享方法
CN112135180A (zh) 一种内容显示方法和显示设备
WO2021169168A1 (zh) 一种视频文件预览方法及显示设备
CN111277891B (zh) 一种节目录制提示方法及显示设备
CN112203154A (zh) 一种显示设备
CN111954059A (zh) 屏保的展示方法及显示设备
CN112073787B (zh) 显示设备及首页显示方法
CN110602540B (zh) 显示设备音量控制方法及显示设备
WO2020248699A1 (zh) 一种声音处理法及显示设备
CN111954043A (zh) 一种信息栏显示方法及显示设备
CN112118476B (zh) 一种快速显示节目预约图标的方法及显示设备
CN112040285B (zh) 界面显示方法及显示设备
CN111885415B (zh) 一种音频数据快速输出方法及显示设备
WO2021159658A1 (zh) 一种界面显示方法及显示设备
WO2021253592A1 (zh) 一种启动体育模式的方法及显示设备
WO2020147507A1 (zh) 显示设备和显示方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20788483

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20788483

Country of ref document: EP

Kind code of ref document: A1