WO2022237464A1 - 音频合成方法、装置、设备、介质及程序产品 - Google Patents

音频合成方法、装置、设备、介质及程序产品 Download PDF

Info

Publication number
WO2022237464A1
WO2022237464A1 PCT/CN2022/087491 CN2022087491W WO2022237464A1 WO 2022237464 A1 WO2022237464 A1 WO 2022237464A1 CN 2022087491 W CN2022087491 W CN 2022087491W WO 2022237464 A1 WO2022237464 A1 WO 2022237464A1
Authority
WO
WIPO (PCT)
Prior art keywords
control
audio
setting
user
live broadcast
Prior art date
Application number
PCT/CN2022/087491
Other languages
English (en)
French (fr)
Inventor
陈映宜
Original Assignee
北京字节跳动网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字节跳动网络技术有限公司 filed Critical 北京字节跳动网络技术有限公司
Publication of WO2022237464A1 publication Critical patent/WO2022237464A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Definitions

  • Embodiments of the present disclosure relate to the technical field of radio and television and Internet live broadcasting, and in particular to an audio synthesis method, device, equipment, medium and program product.
  • Live broadcast technology enables people to break through the limitations of space and obtain various information on the live broadcast site in real time.
  • webcasting has also become a platform for people to show themselves, and has gained great social popularity.
  • Mobile live broadcasting will lower the technical threshold of live broadcasting, allowing ordinary people without professional skills to simply broadcast live broadcasting.
  • a large number of peripheral auxiliary products have been derived around mobile live broadcast devices such as mobile phones.
  • a high-quality mobile live broadcast requires hardware equipment including: desktop computers, sound cards, several intermediate converters/connectors, mobile phones, earphones, microphones, etc. Real-time addition of a live broadcast effect.
  • the existing technology has the technical problem of only relying on complex professional equipment and a team of professional technicians to add synthetic audio when facing high-quality live broadcast requirements. This undoubtedly raised the technical threshold for live broadcasting and increased the cost of live broadcasting for anchors.
  • Embodiments of the present disclosure provide an audio synthesis method, device, device, medium, and program product to solve the problem that the existing technology can only rely on complex professional equipment and a team of professional technicians when facing high-quality live broadcast requirements.
  • Technical issues with synthesizing audio are described below.
  • an embodiment of the present disclosure provides an audio synthesis method, including:
  • the live broadcast device In the live broadcast device installed with the live broadcast application, synthesize the audio to be synthesized and the initial audio, and use the synthesized target audio as the input audio of the live broadcast application, wherein the audio to be synthesized is the audio indicated by the audio synthesis operation, and the initial audio
  • the live broadcast device receives the sound signal sent by the anchor and/or in the live broadcast environment through the audio collection device.
  • an audio synthesis device including:
  • the display module is used to display the composite control on the display screen in a first form when the preset display condition is met, and the visual impact of the first form on the content displayed on the display screen meets the preset visual requirements;
  • the receiving module is used to receive the user's audio synthesis operation on the synthesis control
  • the processing module is used to synthesize the audio to be synthesized with the initial audio in the live broadcast device installed with the live broadcast application, and use the synthesized target audio as the input audio of the live broadcast application, wherein the audio to be synthesized is indicated by the audio synthesis operation
  • the audio, the initial audio is the sound signal from the host received by the live broadcast device through the audio collection device, and/or, the sound signal in the live broadcast environment.
  • an electronic device including:
  • processors at least one processor and memory
  • the memory stores a computer program
  • the at least one processor executes the computer program stored in the memory, so that the at least one processor executes the audio synthesis method described in the above first aspect and various possible designs of the first aspect.
  • an embodiment of the present disclosure provides a live streaming all-in-one machine, including any possible electronic device provided in the third aspect.
  • the embodiments of the present disclosure provide a computer-readable storage medium, where a computer program is stored in the computer-readable storage medium, and when the processor executes the computer program, the above first aspect and each of the first aspect are realized. possible design of the described audio synthesis method.
  • an embodiment of the present disclosure provides a computer program product, including a computer program.
  • the computer program is executed by a processor, the audio synthesis method described in the above first aspect and various possible designs of the first aspect is implemented.
  • an embodiment of the present disclosure provides a computer program.
  • the computer program is executed by a processor, the audio synthesis method described in the above-mentioned first aspect and various possible designs of the first aspect is implemented.
  • Embodiments of the present disclosure provide an audio synthesis method, device, equipment, medium, and program product.
  • the synthesis control is displayed on the display screen in a first form, and then the synthesis control is received from the user.
  • the audio synthesis operation of the control finally, in the live broadcast device installed with the live broadcast application, synthesize the audio to be synthesized with the initial audio, and use the synthesized target audio as the input audio of the live broadcast application.
  • the embodiments of the present disclosure solve the technical problem in the prior art that only complex professional equipment and a team of professional technicians can be used to add synthesized audio in the face of high-quality live broadcast requirements in the prior art through the above method. It achieves the technical effect that the anchor can conveniently and quickly add synthetic audio to the live audio stream on the live broadcast terminal, reduces the cost and technical threshold of live broadcast, and improves the user experience.
  • FIG. 1 is a schematic structural diagram of a live broadcast equipment kit used by an existing indoor anchor provided by an embodiment of the present disclosure
  • FIG. 2 is a first schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure
  • FIG. 3 is a second schematic flow diagram of an audio synthesis method provided by an embodiment of the present disclosure.
  • 4a-4c are schematic diagrams of displaying composite controls on the screen and user interaction operations provided by the embodiments of the present disclosure
  • FIG. 5 is a third schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure.
  • 6a-6b are schematic diagrams of application scenarios for setting pop-up windows through composite controls provided by embodiments of the present disclosure.
  • FIG. 7 is a fourth schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of a sound setting interface provided by an embodiment of the present disclosure.
  • FIG. 9 is a schematic diagram of a scene of an external shortcut key keyboard connected to a live broadcast integrated device provided by an embodiment of the present disclosure.
  • 10a-10c are schematic diagrams of displaying the playback progress of a synthesis control provided by an embodiment of the present disclosure when the audio to be synthesized is played;
  • FIG. 11 is a structural block diagram of an audio synthesis device provided by an embodiment of the present disclosure.
  • Fig. 12 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • live broadcast technology solves an efficient means for people to receive live broadcast information in different geographical spaces.
  • the live webcast that has emerged in recent years has added a lot of fun to people's entertainment life and has been sought after by people.
  • a large number of anchors have joined the ranks of the webcast.
  • Live broadcasting through mobile devices is favored by many anchors for its simplicity.
  • the anchor needs to perform audio synthesis when performing live broadcast, such as adding audio effects such as "entry sound”, "warming laughter” and “applause” to the live broadcast in application.
  • FIG. 1 is a schematic structural diagram of a live broadcast equipment kit used by an existing indoor anchor provided by an embodiment of the present disclosure.
  • the live broadcast equipment kit includes: a mobile terminal 101 , an intermediate adapter 102 , an intermediate converter 103 , a sound card 104 , and a desktop computer 105 .
  • the live application is installed in the mobile terminal 101 .
  • the intermediate adapter 102 is used to solve the problem that the charging interface of the mobile terminal 101 cannot be charged because the charging interface is occupied by the data line when the mobile terminal 101 is live broadcasting for a long time.
  • the intermediate converter 103 is used to transfer the special effect data in the desktop computer 105 to the mobile terminal 101, and after connecting the earphones, it can provide the customer with an ear return function, and it can also solve the problem that the sound card 104 can only output a single channel and cannot receive live broadcast in reverse A problem with the audio data output by the app.
  • the sound card 104 is connected with the microphone to convert the audio analog signal into a digital signal, and input it into the desktop computer 105 for various sound effect processing.
  • the method shown in Figure 1 not only increases the cost of live broadcasting, but also requires the configuration of multiple auxiliary equipment, and the vast majority of anchors do not have professional technology and cannot configure this set of equipment by themselves, so a dedicated team of technical personnel is also required To build, operate and maintain this set of professional live broadcast equipment, these have raised the live broadcast threshold that has been lowered through mobile terminals such as mobile phones.
  • the inventive concept of the present application aims at:
  • the anchor can quickly synthesize the required audio into the live audio stream through the screen or buttons during the live broadcast.
  • the live broadcast does not require multiple devices, only a small device can be used to complete the audio synthesis, and there is no need for a team of professional and technical personnel to participate in operation and maintenance.
  • the anchor can add synthesized audio in real time during the live broadcast.
  • FIG. 2 is a first schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure.
  • the method of this embodiment is applied in a live broadcast integrated terminal, and the audio synthesis method includes:
  • the preset display conditions include: at least one of: detection of an instruction to turn on the camera device, detection of a live broadcast terminal moving in a preset movement mode, receipt of a preset voice instruction, and receipt of a preset button switch turn-on instruction .
  • the synthetic control includes a floating window control, and displays the synthetic control on the display screen in a first form, including:
  • the visual impact of the first form on the content displayed on the display screen meets the preset visual requirements, including: the occlusion rate of the audio synthesis control on the live interface is less than or equal to the preset occlusion threshold, and/or, The comprehensive transparency of the audio synthesis control is greater than or equal to the preset transparency threshold;
  • the integrated transparency is determined according to the transparency of each display element in the audio synthesis control by using a preset model. Specifically, the transparency of each display element on a certain control is firstly multiplied by its corresponding preset weight value, and then each product is added to obtain a comprehensive weight value.
  • the transparency of its background color and the transparency of the text on the button control are multiplied by the corresponding weight value, and then the sum of the two is the comprehensive transparency of the button control.
  • the display screen includes: a touch screen and a non-touch screen.
  • the integrated live broadcast terminal includes at least one display screen.
  • the all-in-one live broadcast terminal can be connected to multiple display screens through an extension interface, and the extension interface includes a wireless interface and a wired interface.
  • S202 Receive a user's audio synthesis operation on the synthesis control.
  • the audio synthesis operation on the synthesis control includes: click, double click, shake, slide according to the preset path, etc.
  • the display screen is a touch screen
  • Preset operating objects include: any part of the user's body, touch objects (such as stylus pens), heat-sensitive objects (such as heating rods), electromagnetic induction objects (electromagnetic induction pens), RFID (Radio Frequency Identification, radio frequency identification) tags keys etc.
  • the sensor installed on the live broadcast integrated terminal will confirm the receipt of the user's audio synthesis operation when it senses that the heat source is close to the preset sensing area.
  • the sensor installed on the live broadcast integrated terminal will confirm the receipt of the user's audio synthesis operation when it senses that a magnetic field is close to the preset sensing area.
  • the card reader or card reading circuit installed on the live broadcast integrated terminal will recognize the audio information to be synthesized contained in the RFID tag, and the audio synthesis operation will be received. .
  • S203 In the live broadcast device installed with the live broadcast application, synthesize the audio to be synthesized with the initial audio, and use the synthesized target audio as the input audio of the live broadcast application.
  • the audio to be synthesized is the audio indicated by the audio synthesis operation
  • the initial audio is the sound signal received by the live broadcast device from the host through the audio collection device, and/or in the live broadcast environment.
  • the live broadcast application codes the target audio, it transmits it to the live broadcast platform server through the network, and then the live broadcast platform service sends it to each audience.
  • the audio to be synthesized includes: laughter, exit/entry sounds, bird calls (such as crow calls), applause, slapping, gunshots, and the like.
  • live broadcast equipment installed with the live broadcast application includes: live camera equipment (professional camera for news media live broadcast), personal computer (Personal Computer, referred to as PC), mobile terminal (mobile phone, tablet computer), webcast Integrated equipment, live broadcast vehicles (such as live broadcast vehicles, live broadcast shopping carts), etc.
  • live camera equipment professional camera for news media live broadcast
  • PC Personal Computer
  • mobile terminal mobile phone, tablet computer
  • webcast Integrated equipment live broadcast vehicles (such as live broadcast vehicles, live broadcast shopping carts), etc.
  • the audio synthesis method provided by the embodiment of the present disclosure, the method firstly displays the synthesis control on the display screen in the first form when the preset display condition is met, then receives the user's audio synthesis operation on the synthesis control, and finally installs the live application
  • the audio to be synthesized is synthesized with the initial audio
  • the synthesized target audio is used as the input audio of the live broadcast application.
  • FIG. 3 is a second schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure.
  • the method of this embodiment is applied in a live broadcast integrated terminal, and the audio synthesis method includes:
  • the visual impact of the first form on the content displayed on the display screen meets preset visual requirements.
  • the live broadcast application starts the camera on the live broadcast integrated terminal.
  • the all-in-one live broadcast terminal detects that the camera is turned on, it displays a composite control on the edge of the display screen.
  • the synthetic control includes a floating window control, and the synthetic control is displayed on the display screen in a first form, including:
  • FIGS. 4a-4c are schematic diagrams of displaying composite controls on a screen and user interaction operations provided by an embodiment of the present disclosure.
  • the composite control is a floating window 41.
  • the floating window 41 is set to have a background color transparency of more than 50%, so that the floating window 41 is opposite to the live broadcast. The visual impact of other content on the interface is reduced to meet the preset requirements.
  • the synthesis control includes multiple sub-controls. As shown in FIG. 4 b , the synthesis control includes: a pop-up setting sub-control 411 , multiple audio sub-controls 412 corresponding to the audio to be synthesized, and a collapse control 413 . After detecting that the camera is turned on, all sub-controls are displayed directly.
  • the synthesis control includes a plurality of sub-controls, but the first form does not display all the sub-controls, as shown in Figure 4c, the first form only displays part of the audio sub-controls corresponding to the audio to be synthesized 412 , setting the sub-control 411 and the expansion control 414 in the pop-up window.
  • the state setting operation includes: an operation of changing the display form of the synthetic control and an operation of changing the position of the synthetic control.
  • Operations that change the display form of composite controls including: expand/collapse operations, operations that change the order or position of sub-controls, compression, enlargement, etc.
  • the operation of changing the position of the composite control includes: dragging the composite control to another position of the display screen.
  • the composite control always needs to be displayed on the side.
  • the user drags the composite control and releases it, calculate the distance from the composite control to the four edges of the display, and snap the composite control to the minimum distance. displayed near the edges.
  • S303 Switch the form and/or display position of the composite control according to the state setting operation.
  • S304 Detect in real time that the user approaches or touches the synthesized control on the touch screen by using a preset operator.
  • the display screen includes at least one touch screen
  • the preset operating objects include: any part of the user's body, a touch object (such as a stylus), a heat sensitive object (such as a heating rod), an electromagnetic induction object ( electromagnetic induction pen), RFID tag key, etc.
  • the synthesis control includes multiple sub-controls corresponding to the audio to be synthesized, as shown in Figure 4b, and this step specifically includes:
  • a composite operation instruction corresponding to the sub-control is determined.
  • S305 In the live broadcast device installed with the live broadcast application, synthesize the audio to be synthesized with the initial audio, and use the synthesized target audio as the input audio of the live broadcast application.
  • the audio synthesis method provided by the embodiment of the present disclosure, the method firstly displays the synthesis control on the display screen in the first form when the preset display condition is met, then receives the user's audio synthesis operation on the synthesis control, and finally installs the live application
  • the audio to be synthesized is synthesized with the initial audio
  • the synthesized target audio is used as the input audio of the live broadcast application.
  • the audio synthesis method provided by the present disclosure can also use the synthesis control to set various audio control parameters of the live broadcast integrated device in the form of a pop-up window during live broadcast,
  • the following is illustrated with specific embodiments.
  • FIG. 5 is a third schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure. The method of this embodiment is applied in a live broadcast integrated terminal, and the audio synthesis method includes:
  • the pop-up window setting sub-control 411 shown in FIG. 4b or FIG. 4c after the user clicks on this control, the integrated live broadcast device obtains the control instruction of the pop-up window setting operation.
  • S502 In response to the pop-up window setting operation, superimpose and display the setting window on the display screen.
  • the setting window includes a sound setting tab
  • the display area corresponding to the sound setting tab includes: an ear return switch control, a dodge switch control, a speaker selection control, a microphone selection control, and a microphone volume setting control At least one control among controls such as a sound effect selection control and a sound effect volume control.
  • the setting window includes a shortcut key setting tab
  • the shortcut key setting tab includes: the selected shortcut key array and the shortcut key array to be selected, then this step also includes:
  • FIGS. 6a-6b are schematic diagrams of application scenarios for setting pop-up windows through synthetic controls provided by embodiments of the present disclosure.
  • An implementation of the sound setting tab is shown in Figure 6a.
  • a setting window pops up at the bottom of the display screen.
  • Turn on or off the return effect or dodge effect by using the switch of the return or dodge control.
  • Select Speaker or Microphone by tapping the corresponding control.
  • the selection control can also be selected by ticking or by pulling down the selection control.
  • the volume bar can also be set in forms such as an open arc, a closed ring, and a knob.
  • FIG. 6b An implementation of the sound effect shortcut key tab 422 is shown in Figure 6b.
  • the shortcut keys corresponding to the audio to be synthesized that have been selected and displayed are displayed in the box.
  • up to 8 shortcut keys are allowed to be added , click the "trash can" button on the right side of the selected shortcut key, that is, the delete button, to set the shortcut key not to be displayed.
  • the audio to be synthesized option can also be downloaded from the cloud server or Internet server through the network interface, as shown in Figure 6b, the audio to be synthesized corresponding to "whistle” and "scream" is not pre-stored in the integrated live broadcast terminal. Then just click the download button on the right side of the control to download from the server to the local storage.
  • the audio shortcut key tab 422 is also updated synchronously. After the user sees the new option, You can click the download button to download. Of course, the user can also set automatic download or batch download to download all the updated audio to be synthesized to the local storage at one time.
  • the setting of the sound effect shortcut key can also be selected by using a drop-down box, or other selection methods, which can be selected by those skilled in the art according to the actual situation, which is not limited in this application.
  • the audio synthesis method acquires the user’s pop-up window setting operation on the synthesis control, and then superimposes and displays the settings on the display screen in response to the pop-up window setting operation Window, and then respond to the user’s operation instructions for each control, and set various audio control parameters correspondingly, which solves the technical problem that the anchor cannot easily and quickly change various sound-related setting parameters during live broadcast, and requires the assistance of professional and technical personnel. It makes it easy and convenient to set the shortcut keys for sound settings and synthesis controls, without changing the direction of sight, affecting the live broadcast effect, and without pausing the live broadcast to set, which improves the user's experience of using the live broadcast integrated terminal.
  • the audio synthesis method provided by the present disclosure can also set the display status of the synthesis control and various sound control parameters through a special sound setting interface on the live broadcast integrated terminal before the live broadcast, as follows Specific steps are described and introduced with specific embodiments.
  • FIG. 7 is a fourth schematic flowchart of an audio synthesis method provided by an embodiment of the present disclosure. The method of this embodiment is applied in a live broadcast integrated terminal, and the audio synthesis method includes:
  • the all-in-one live broadcast terminal detects that the user clicks the sound setting button, or opens the sound setting interface through a preset setting command at the beginning of the live broadcast application.
  • S702 Display a sound setting interface on the display screen in response to the user's sound setting operation.
  • the sound setting interface includes: ear return switch control, dodge switch control, speaker selection control, microphone selection control, microphone volume setting control, sound effect selection control, sound effect volume control, synthesis control (such as floating window) display At least one control among controls such as the switch control, the selected shortcut key array, and the to-be-selected shortcut key array.
  • S703 Correspondingly setting each audio control parameter according to the user's setting instruction for each parameter control on the sound setting interface.
  • FIG. 8 for a possible implementation manner.
  • FIG. 8 is a schematic diagram of a sound setting interface provided by an embodiment of the present disclosure.
  • the user selects the control through the speaker, and the user can choose to use the "wired headset” or the speaker that comes with the live broadcast integrated device, that is, the "local speaker”; through the microphone selection control, the user can Select “wired headset”, “local microphone”, “canon microphone”; through the ear return switch control, the user can turn on or off the ear return effect by toggling the switch control to the right/left, and on the ear return switch control, It also includes a prompt text, such as "The speaker of this machine does not support the ear return function", reminding the user that the ear return function and the speaker selection are directly related to the setting, so as to avoid improper selection of user parameters and affect the use effect; similarly, the dodge switch control and The floating window switch control also has similar operations, see Figure 8 for details.
  • Sound effect beautification effects include: original sound, KTV, magnetic, old records, ethereal, theater and other sound effects. Each sound effect corresponds to multiple sound effect setting parameters. By selecting a specific sound effect, you can quickly set each parameter, reducing the professional threshold for sound effect setting , so that the anchor who does not have professional skills can also perform simple and convenient sound effect settings.
  • the bottom of Figure 8 is the shortcut key selection array of the floating window, that is, the synthesis control.
  • the user only needs to select each button control, and then set and display the corresponding shortcut key of the audio to be synthesized, so that the user can pass the shortcut key in the live broadcast interface after the live broadcast starts. key to perform audio synthesis directly.
  • the audio synthesis method provided by the embodiments of the present disclosure is based on the embodiments shown in FIG. 2 , FIG. 3 and FIG. 5 , by obtaining the user’s voice setting operation before the camera is turned on or the live broadcast is started, and then responding to the user’s voice Setting operation, the sound setting interface is displayed on the display screen, and then according to the user's setting instructions for each parameter control on the sound setting interface, each audio control parameter is set correspondingly, which solves the problem that the anchor cannot easily and quickly change various sound-related setting parameters , technical issues that require the assistance of professional and technical personnel.
  • the setting of shortcut keys for sound setting and synthesis control is simple and convenient, and the user's experience of using the live broadcast integrated terminal is improved.
  • the all-in-one live broadcast machine that is, the all-in-one live broadcast device, is connected to at least one shortcut key keyboard through an external interface
  • the corresponding audio to be synthesized is determined according to the operation instruction.
  • the external interfaces include: wireless interfaces and wired interfaces.
  • FIG. 9 is a schematic diagram of a scene of an external shortcut keyboard connected to an integrated live broadcast device provided by an embodiment of the present disclosure.
  • the all-in-one live broadcast machine 91 is connected to the shortcut key keyboard 92 through a wired interface or a wireless interface, and multiple sub-controls in the synthesis control 911 used to implement the audio synthesis method in various embodiments of the present disclosure are aligned with the keys on the keyboard.
  • the corresponding relationship between each key on the shortcut key keyboard and each sub-control on the display screen can be set arbitrarily.
  • Shortcut keyboards include: mechanical button keyboards, touch screens, etc. Clicking the shortcut key by the user is equivalent to clicking the corresponding sub-control on the display screen, and at the same time, the sub-control switches to the selected state.
  • the sub-control can synchronously display the playback progress of the audio to be synthesized in the form of a bar-shaped progress bar or a fan-shaped progress graph.
  • FIG. 10a-10c are schematic diagrams of a composition control displaying a playback progress when the audio to be synthesized is played according to an embodiment of the present disclosure.
  • a playback progress bar for playing the audio corresponding to "Laughter 1" is displayed on the sub-control.
  • the background color or background pattern of the sub-control displays the playback progress bar of the audio corresponding to "Laughter 1" in the form of a rotating gradient.
  • the background color or background pattern of the sub-control changes from left to right, from right to left, from top to bottom, from bottom to top, from bottom left to
  • the playing progress bar of the audio corresponding to "Laughter 1" is displayed in any one of the forms of gradient from upper right, from lower right to upper left, etc.
  • the anchor can know the playback progress of the audio to be synthesized. Further, you can click the sub-control again to end the playback, or click other sub-controls to mix and play multiple audios to be synthesized, which increases the operational flexibility of live audio synthesis and improves the live broadcast. Real-time live audio effects to improve user experience.
  • FIG. 11 is a structural block diagram of an audio synthesis apparatus 1100 provided in an embodiment of the present disclosure. For ease of description, only the parts related to the embodiments of the present disclosure are shown. Referring to Figure 11, the device includes:
  • the display module 1101 is configured to display the synthesized control on the display screen in a first form when the preset display condition is met, and the visual impact of the first form on the content displayed on the display screen meets the preset visual requirements;
  • a receiving module 1102 configured to receive the user's audio synthesis operation on the synthesis control
  • the processing module 1103 is configured to synthesize the audio to be synthesized with the initial audio in the live broadcast device installed with the live broadcast application, and use the synthesized target audio as the input audio of the live broadcast application, wherein the audio to be synthesized is For the audio indicated by the audio synthesis operation, the initial audio is the sound signal received by the live broadcast device from the host through the audio collection device, and/or in the live broadcast environment.
  • the display screen includes a touch screen
  • the receiving module 1102 is configured to detect in real time that the user approaches or touches the touch screen on the touch screen with a preset operator.
  • Composite controls are configured to detect in real time that the user approaches or touches the touch screen on the touch screen with a preset operator.
  • the synthesis control includes a sub-control corresponding to the audio to be synthesized, and the receiving module 1102 is specifically configured to:
  • a composite operation instruction corresponding to the sub-control is determined.
  • the preset display conditions include: detection of an instruction to turn on the camera device, detection of a live broadcast terminal moving in a preset motion mode, receipt of a preset voice command, and receipt of a preset button switch being turned on at least one of the instructions.
  • the composite control includes a floating window control
  • the display module 1101 is configured to display the floating window control on the edge of the display screen according to preset transparency, preset size and preset shape.
  • the receiving module 1102 is further configured to obtain the user's state setting operation on the synthetic control
  • the display module 1101 is further configured to switch the form and/or display position of the composite control according to the state setting operation.
  • the display module 1101 is further configured to superimpose and display a setting window on the display screen in response to the user's pop-up window setting operation on the synthetic control;
  • the processing module 1103 is further configured to correspondingly set each audio control parameter according to the user's setting instruction for each parameter control on the setting window.
  • the setting window includes a sound setting tab
  • the sound setting tab includes: an ear return switch control, and/or a dodge switch control
  • the processing module 1103 is specifically used for:
  • the setting window includes a sound setting tab
  • the sound setting tab includes: a speaker selection control
  • the processing module 1103 is specifically used for:
  • the setting window includes a sound setting tab
  • the sound setting tab includes: a microphone selection control
  • the processing module 1103 is specifically used for:
  • the setting window includes a sound setting tab
  • the sound setting tab includes: a microphone-connected volume setting control
  • the processing module 1103 is specifically used for:
  • the setting window includes a sound setting tab
  • the sound setting tab includes: a sound effect selection control, and/or, a sound effect volume control
  • the processing module 1103 is specifically used for:
  • the setting window includes a shortcut key setting tab
  • the shortcut key setting tab includes: an array of shortcut keys selected and an array of shortcut keys to be selected
  • the processing module 1103 is specifically used to :
  • the display module 1101 before displaying the synthetic control in the first form on the display screen when the preset display condition is met, is further configured to respond to the user's voice to set Operation, the sound setting interface is displayed on the display;
  • the processing module 1103 is further configured to correspondingly set each audio control parameter according to the user's setting instruction for each parameter control on the sound setting interface.
  • the sound setting interface includes: an ear return switch control, and/or a dodge switch control
  • the processing module 1103 is further configured to:
  • the sound setting interface includes: a speaker selection control, and the processing module 1103 is further configured to:
  • the sound setting interface includes: a microphone selection control, and the processing module 1103 is further configured to:
  • the sound setting interface includes: a microphone-connected volume setting control, and the processing module 1103 is further configured to:
  • the sound setting interface includes: a sound effect selection control, and/or, a sound effect volume control
  • the processing module 1103 is further configured to:
  • the sound setting interface includes: an array of shortcut keys selected and an array of shortcut keys to be selected, and the processing module 1103 is further configured to:
  • the sound setting interface includes: a synthetic control display switch, and the processing module 1103 is further configured to:
  • the audio synthesis device 1100 is connected to at least one shortcut keyboard through an external interface, and the receiving the user's audio synthesis operation on the synthesis control includes;
  • the receiving module 1102 is further configured to receive an operation instruction sent by the shortcut keyboard, the operation instruction is determined by the shortcut keyboard in response to the user's operation;
  • the processing module 1103 is further configured to determine the corresponding audio to be synthesized according to the operation instruction.
  • the device 1100 provided in this embodiment can be used to execute the steps related to the user end in the technical solutions of the above method embodiments, and its implementation principle and technical effect are similar, so this embodiment will not repeat them here.
  • the electronic device 1200 may be a terminal device or a server.
  • the terminal equipment may include but not limited to mobile phones, notebook computers, digital broadcast receivers, personal digital assistants (Personal Digital Assistant, PDA for short), tablet computers (Portable Android Device, PAD for short), portable multimedia players (Portable Media Player, referred to as PMP), mobile terminals such as vehicle-mounted terminals (such as vehicle-mounted navigation terminals), and fixed terminals such as digital TVs, desktop computers, etc.
  • PDA Personal Digital Assistant
  • PMP portable multimedia players
  • mobile terminals such as vehicle-mounted terminals (such as vehicle-mounted navigation terminals)
  • fixed terminals such as digital TVs, desktop computers, etc.
  • the electronic device shown in FIG. 12 is only an example, and should not limit the functions and application scope of the embodiments of the present disclosure.
  • an electronic device 1200 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) 1201, which may be stored in a read-only memory (Read Only Memory, referred to as ROM) 1202 or from a storage device. 1208 loads the programs in the random access memory (Random Access Memory, RAM for short) 1203 to execute various appropriate actions and processes. In the RAM 1203, various programs and data necessary for the operation of the electronic device 1200 are also stored.
  • the processing device 1201, ROM 1202, and RAM 1203 are connected to each other through a bus 1204.
  • An input/output (I/O) interface 1205 is also connected to the bus 1204 .
  • the following devices can be connected to the I/O interface 1205: input devices 12012 including, for example, touch screens, touch pads, keyboards, mice, cameras, microphones, accelerometers, gyroscopes, etc.; including, for example, liquid crystal displays (Liquid Crystal Display, referred to as LCD ), a speaker, a vibrator, etc.; a storage device 1208 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 1209.
  • the communication means 1209 may allow the electronic device 1200 to perform wireless or wired communication with other devices to exchange data. While FIG. 12 shows electronic device 1200 having various means, it is to be understood that implementing or having all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program codes for executing the methods shown in the flowcharts.
  • the computer program may be downloaded and installed from a network via communication means 1209, or from storage means 1208, or from ROM 1202.
  • the processing device 1201 the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof.
  • Computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programming read-only memory (Erasable Programmable Read-Only Memory, referred to as EPROM or flash memory), optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, referred to as CD-ROM), optical storage device, magnetic storage device, or the above any suitable combination.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave carrying computer-readable program code therein. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code contained on the computer readable medium can be transmitted by any appropriate medium, including but not limited to: electric wire, optical cable, radio frequency (Radio Frequency, RF for short), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device is made to execute the methods shown in the above-mentioned embodiments.
  • Computer program code for carrying out the operations of the present disclosure can be written in one or more programming languages, or combinations thereof, including object-oriented programming languages—such as Java, Smalltalk, C++, and conventional Procedural Programming Language - such as "C" or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or it can be connected to an external A computer (connected via the Internet, eg, using an Internet service provider).
  • LAN Local Area Network
  • WAN Wide Area Network
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented by software or by hardware. Wherein, the name of the unit does not constitute a limitation of the unit itself under certain circumstances, for example, the first obtaining unit may also be described as "a unit for obtaining at least two Internet Protocol addresses".
  • exemplary types of hardware logic components include: Field Programmable Gate Array (Field Programmable Gate Array, FPGA for short), Application Specific Integrated Circuit (ASIC for short), application specific standard product ( Application Specific Standard Product (ASSP for short), System on a Chip (SOC for short), Complex Programmable Logic Device (CPLD for short), etc.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • Embodiments of the present disclosure also provide a computer program product, including a computer program, wherein the computer program implements the methods in the foregoing embodiments when executed by a processor.
  • An embodiment of the present disclosure also provides a live broadcast integrated machine or a live broadcast integrated device, including the electronic device corresponding to FIG. 12 .
  • the control circuit of the live broadcast all-in-one machine includes: a main control module and an audio processing module, wherein the live application is installed on the main control module, and the audio processing module is used to synthesize the audio to be synthesized into the original audio, so as to Form the target audio.
  • the audio to be synthesized can be stored in the audio processing module, also can be stored in the main control module, and can also be stored in the memory of the integrated live broadcast machine.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • General Engineering & Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本公开实施例提供的音频合成方法、装置、设备、介质及程序产品,在该方法中,首先在满足预设显示条件时,以第一形态在显示屏上显示合成控件,然后接收用户对合成控件的音频合成操作,最后在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。本公开实施例通过上述方法,解决了现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。达到了主播在直播终端上即可便捷快速地在直播音频流中添加合成音频的技术效果,降低了直播的成本和技术门槛,提高了用户的使用体验感。

Description

音频合成方法、装置、设备、介质及程序产品
相关申请交叉引用
本申请要求于2021年5月13日提交中国专利局、申请号为202110523959.3、发明名称为“音频合成方法、装置、设备、介质及程序产品”的中国专利申请的优先权,其全部内容通过引用并入本文。
技术领域
本公开实施例涉及广播电视及互联网直播技术领域,尤其涉及一种音频合成方法、装置、设备、介质及程序产品。
背景技术
随着社会信息化的不断深入发展,人们对于信息获取的及时性需求也不断提高。直播技术能够使得人们突破空间的限制,实时获取到直播现场的各类信息。而近年来兴起的网络直播,也成为了人们展示自我的一个平台,获得了极大的社会热度。手机直播更是将降低了直播的技术门槛,让不具备专业技术的普通人也能简单地进行直播。然而,随着观众对于直播内容及质量的要求不断提高,围绕着手机这类移动直播设备,衍生出了大量的周边辅助产品。
目前,一个高质量的手机直播,所需要的硬件设备包括:台式电脑、声卡、若干中间转换器/衔接器、手机、耳机、麦克风等,除了主播外,还需要配备专门的技术人员来负责各种直播效果的实时添加。
即现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。这无疑又提高了直播的技术门槛,增加了主播们的直播成本。
发明内容
本公开实施例提供一种音频合成方法、装置、设备、介质及程序产品,以解决现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。
第一方面,本公开实施例提供一种音频合成方法,包括:
在满足预设显示条件时,以第一形态在显示屏上显示合成控件,第一形态对显示屏上显示内容的视觉影响满足预设视觉要求;
接收用户对合成控件的音频合成操作;
在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频,其中,待合成音频为音频合成操作所指示的音频,初始音频为直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
第二方面,本公开实施例提供一种音频合成装置,包括:
显示模块,用于在满足预设显示条件时,以第一形态在显示屏上显示合成控件,第一形态对显示屏上显示内容的视觉影响满足预设视觉要求;
接收模块,用于接收用户对合成控件的音频合成操作;
处理模块,用于在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频,其中,待合成音频为音频合成操作所指示的音频,初始音频为直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
第三方面,本公开实施例提供一种电子设备,包括:
至少一个处理器和存储器;
所述存储器存储计算机程序;
所述至少一个处理器执行所述存储器存储的计算机程序,使得所述至少一个处理器执行如上第一方面以及第一方面各种可能的设计所述的音频合成方法。
第四方面,本公开实施例提供一种直播一体机,包括第三方面所提供的任意一种可能的电子设备。
第五方面,本公开实施例提供一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机程序,当处理器执行所述计算机程序时,实现如上第一方面以及第一方面各种可能的设计所述的音频合成方法。
第六方面,本公开实施例提供一种计算机程序产品,包括计算机程序,该计算机程序被处理器执行时,实现如上第一方面以及第一方面各种可能的设计所述的音频合成方法。
第七方面,本公开实施例提供一种计算机程序,所述计算机程序被处理器执行时,实现如上述第一方面以及第一方面各种可能的设计所述的音频合成方法。
本公开实施例提供了音频合成方法、装置、设备、介质及程序产品,在该方法中,首先在满足预设显示条件时,以第一形态在显示屏上显示合成控件,然后接收用户对合成控件的音频合成操作,最后在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。本公开实施例通过上述方法,解决了现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。达到了主播在直播终端上即可便捷快速地在直播音频流中添加合成音频的技术效果,降低了直播的成本和技术门槛,提高了用户的使用体验感。
附图说明
为了更清楚地说明本公开实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本公开实施例提供的现有室内主播使用的直播设备套件结构示意图;
图2为本公开实施例提供的音频合成方法的流程示意图一;
图3为本公开实施例提供的音频合成方法的流程示意图二;
图4a-4c为本公开实施例提供的在屏幕上显示合成控件及用户交互操作示意图;
图5为本公开实施例提供的音频合成方法的流程示意图三;
图6a-6b为本公开实施例提供的通过合成控件进行弹窗设置的应用场景示意图;
图7为本公开实施例提供的音频合成方法的流程示意图四;
图8为本公开实施例提供的一种声音设置界面示意图;
图9为本公开实施例提供的一种直播一体化设备外接快捷键键盘的场景示意图;
图10a-10c为本公开实施例提供的一种合成控件在待合成音频播放时显示播放进度的示意图;
图11为本公开实施例提供的音频合成装置的结构框图;
图12为本公开实施例提供的电子设备的结构示意图。
具体实施方式
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本公开一部分实施例,而不是全部的实施例。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,包括但不限于对多个实施例的组合,都属于本公开保护的范围。
本公开的说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”、“第四”等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本公开的实施例例如能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。
随着移动互联网技术的不断发展,人们的日常生活已经高度信息化,而直播技术解决了人们在不同地理空间上异地接收直播地信息的一种高效手段。近年来兴起的网络直播,更是给人们的娱乐生活增加了很多乐趣,收到了人们的追捧。大量主播纷纷加入了网络直播的行列。而通过移动设备来进行直播,以其简便性受到了很多主播的青睐。但是,随着观众对于直播质量要求的不断提高,主播在进行直播的时候,需要进行音频合成,如添加“入场音”、“暖场笑声”、“鼓掌声”等等音频效果到直播应用中。
面对这种需求,对于大多数主播来说,只能够通过购置一套专业的直播设备,包括:台式电脑、专业声卡、若干个中间转换器/衔接器等专业直播辅助设备来实现直播中的音频合成需求。
图1为本公开实施例提供的现有室内主播使用的直播设备套件结构示意图。如图1所示,直播设备套件包括:移动终端101、中间衔接器102、中间转换器103、声卡104、台式电脑105。
其中,移动终端101中安装有直播应用。
中间衔接器102用于解决移动终端101在长时间直播时,充电接口被数据线占用而无法充电的问题。
中间转换器103用于将台式电脑105中的特效数据转接到移动终端101中,并且连接耳机后可以给客户提供耳返功能,还能够解决声卡104只能单路输出而不能反向接收直播应用输出的音频数据的问题。
声卡104与麦克风连接,实现音频模拟信号转换为数字信号,输入台式电脑105中进行各种音效处理。
但是,图1所示的方式不但造成直播成本的提高,需要配置多个辅助设备,而绝大多数主播不具备专业技术,无法自己去配置这一套设备,因此还需要配备专门的技术人员团队来构建、运营和维护这套专业直播设备,这些都使得本已通过手机等移动终端降低了的直播门槛,又被提高了起来。
并且,由于连接了多个设备,使得整套直播设备的稳定性容易受到各种因素的影响,任何一个设备的接插件插口出现问题,或者线路由于某种原因断路,都会使得系统无法工作。
基于上述技术问题,本申请的发明构思旨在:
通过一个直播终端直接让主播在直播的时候通过屏幕或者按键快速合成所需要的音频到直播的音频流中。使得直播无需多个设备,仅用一个小型设备就能完成音频合成,也无需专业技术人员团队参与运维,主播一人即可在直播时实时添加合成音频。
下面结合附图对本申请提供的音频合成方法进行详细介绍。
参考图2,图2为本公开实施例提供的音频合成方法的流程示意图一。本实施例的方法应用在直播一体式终端中,该音频合成方法包括:
S201:在满足预设显示条件时,以第一形态在显示屏上显示合成控件。
在本步骤中,预设显示条件包括:检测到摄像设备开启指令、检测到直播终端按预设运动方式进行运动、接收到预设语音指令和接收到预设按钮开关开启指令中的至少一种。
在一种可能的设计中,合成控件包括悬浮窗控件,以第一形态在显示屏上显示合成控件,包括:
根据预设透明度、预设大小以及预设形状在显示屏的边缘显示悬浮窗控件。
在另一种可能的设计中,第一形态对显示屏上显示内容的视觉影响满足预设视觉要求,包括:音频合成控件对直播界面的遮挡率小于或等于预设遮挡阈值,和/或,音频合成控件的综合透明度大于或等于预设透明阈值;
其中,综合透明度是利用预设模型,根据音频合成控件中各个显示元素的透明度来确定的。具体的,某个控件上各个显示元素的透明度先乘以其对应的预设权重值,然后将各个乘积相加即为综合权重值。
例如:一个按钮控件,其背景色的透明度与按钮控件上的文本的透明度分别乘以对应的权重值,再求两者的和,即为按钮控件的综合透明度。
在一种可能的设计中,所述显示屏包括:触控屏和非触控屏。
在本实施例中,直播一体式终端上包括至少一个显示屏。在一种可能的设计中,直播一体式终端可以通过扩展接口连接多个显示屏,扩展接口包括无线接口和有线接口。
S202:接收用户对合成控件的音频合成操作。
在本步骤中,对合成控件的音频合成操作包括:单击、双击、摇一摇、按预设路径滑动等
在本实施例中,当显示屏为触控屏时,具体的,实时检测所述用户利用预设操作物靠近 或触碰到所述触控屏上的所述合成控件。
预设操作物包括:用户的身体任意部位,触控物(如触控笔)、热感应物(如发热棒)、电磁感应物(电磁感应笔),RFID(Radio Frequency Identification,射频识别)标签钥匙等。
触控物直接触控合成控件。
热感应物只要靠近触控屏所对应的预设感应区域,则安装在直播一体式终端上的传感器,感应到热源靠近预设感应区域时,确认接收到用户的音频合成操作。
电磁感应物只要靠近触控屏所对应的预设感应区域,则安装在直播一体式终端上的传感器,感应到有磁场靠近预设感应区域时,确认接收到用户的音频合成操作。
RFID标签钥匙只要靠近触控屏所对应的预设感应区域,则安装在直播一体式终端上的读卡器或读卡电路,识别RFID标签中包含的待合成音频信息,即接收到了音频合成操作。
S203:在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。
在本步骤中,待合成音频为音频合成操作所指示的音频,初始音频为直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
直播应用将目标音频编码后,通过网络传输给直播平台服务器,再由直播平台服务发送给各个观众。
需要说明的是,待合成音频包括:笑声、出场/入场声、鸟叫声(如乌鸦叫声)、掌声、打耳光声、枪声等等。
需要说明的是,在安装了直播应用的直播设备包括:直播摄像设备(用于新闻媒体直播的专业摄像机)、个人电脑(Personal Computer,简称PC)、移动终端(手机、平板电脑)、网络直播一体化设备、直播载具(如直播车辆、直播购物车)等。
本公开实施例提供的音频合成方法,该方法首先在满足预设显示条件时,以第一形态在显示屏上显示合成控件,然后接收用户对合成控件的音频合成操作,最后在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。本公开实施例通过,解决了现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。达到了主播在直播终端上即可便捷快速地在直播音频流中添加合成音频的技术效果,降低了直播的成本和技术门槛,提高了用户的使用体验感。
参考图3,图3为本公开实施例提供的音频合成方法的流程示意图二。本实施例的方法应用在直播一体式终端中,该音频合成方法包括:
S301:检测到摄像设备开启指令时,以第一形态在显示屏上显示合成控件。
在本步骤中,第一形态对所述显示屏上显示内容的视觉影响满足预设视觉要求。
在本实施例中,用户在启动直播一体式终端上的直播应用后,点击开始直播,则直播应用开启直播一体式终端上的摄像头。而当直播一体式终端检测到摄像头被打开后,则在显示屏的边缘显示合成控件。
在一种可能的设计中,合成控件包括悬浮窗控件,则以第一形态在显示屏上显示合成控件,包括:
根据预设透明度、预设大小以及预设形状在显示屏的边缘显示悬浮窗控件。
图4a-4c为本公开实施例提供的在屏幕上显示合成控件及用户交互操作示意图。如图4a 所示,合成控件为悬浮窗41,在检测到摄像头打开后,以一个小悬浮窗的形态显示合成控件,并且悬浮窗41设置为底色透明度50%以上,使得悬浮窗41对直播界面上其它内容的视觉影响降低到满足预设要求。
在一种可能的设计中,合成控件包括多个子控件,如图4b所示,合成控件包括:弹窗设置子控件411、多个与待合成音频对应的音频子控件412、收起控件413。在检测到摄像头打开后,直接显示所有的子控件。
在另一种可能的设计中,合成控件包括多个子控件并,但是第一形态并不展示所有的子控件,如图4c所示,第一形态只显示部分与待合成音频对应的音频子控件412,弹窗设置子控件411以及展开控件414。
S302:获取用户对合成控件的状态设置操作。
在本步骤中,状态设置操作包括:改变合成控件显示形态的操作以及改变合成控件所在位置的操作。
改变合成控件显示形态的操作,包括:展开/收起操作、调换子控件顺序或位置的操作、压缩、放大等等。
改变合成控件所在位置的操作,包括:将合成控件拖拽到显示屏的其它位置。
需要说明的是,在一种可能的设计中,合成控件始终需要靠边显示,当用户拖拽合成控件松开后,计算合成控件到显示屏四个边缘的距离,将合成控件吸附到最小距离的边缘附近显示。
S303:根据状态设置操作切换合成控件的形态,和/或,显示位置。
例如,如图4a-4c所示,用户在图4c中点击了展开控件414后,合成控件从图4c的形态转换成图4b的形态,用户再点击收起控件413后,合成控件从图4b的形态转换成图4a的形态。
S304:实时检测用户利用预设操作物靠近或触碰到触控屏上的合成控件。
在本实施例中,显示屏包括至少一个触控屏,预设操作物包括:用户的身体任意部位,触控物(如触控笔)、热感应物(如发热棒)、电磁感应物(电磁感应笔),RFID标签钥匙等。
在一种可能的设计中,合成控件包括多个与待合成音频对应的子控件,如图4b所示,则本步骤具体包括:
若检测到所述预设操作物靠近或触碰到所述子控件,则将所述子控件切换为被选中形态;
确定与所述子控件对应的合成操作指令。
S305:在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。
本步骤的具体原理与名词解释介绍参考S203,在此不再赘述。
本公开实施例提供的音频合成方法,该方法首先在满足预设显示条件时,以第一形态在显示屏上显示合成控件,然后接收用户对合成控件的音频合成操作,最后在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为直播应用的输入音频。本公开实施例通过,解决了现有技术在面对高质量的直播需求时,存在只能够依赖复杂的专业设备和专业技术人员团队来添加合成音频的技术问题。达到了主播在直播终端上即可便捷快速地在直播音频流中添加合成音频的技术效果,降低了直播的成本和技术门 槛,提高了用户的使用体验感。
参考图5,在上述两个实施例的基础上,本公开提供的音频合成方法还能够利用合成控件在直播时,通过弹窗设置的形式对直播一体式设备的各项音频控制参数进行设置,下面以具体实施例进行举例说明。
图5为本公开实施例提供的音频合成方法的流程示意图三。本实施例的方法应用在直播一体式终端中,该音频合成方法包括:
S501:获取用户对合成控件的弹窗设置操作。
在本实施例中,如图4b或图4c中所示的弹窗设置子控件411,用户在点击了该控件后,直播一体式设备获取到弹窗设置操作的控制指令。
S502:响应于弹窗设置操作,在显示屏上叠加显示设置窗口。
在本实施例中,设置窗口包括声音设置选项卡,该声音设置选项卡所对应的显示区域内,包括:耳返开关控件、闪避开关控件、扬声器选择控件、麦克风选择控件、连麦音量设置控件、音效选择控件、音效音量控件等控件中的至少一个控件。
S503:响应于用户对各个控件的操作指令,对应设置各项音频控制参数。
在本实施例中,具体的:
S5031、耳返开关控件的开关指令,设置耳返效果开关状态所对应的控制参数,并对应切换耳返开关控件的开关显示状态。
和/或,
S5032、响应于用户对闪避开关控件的开关指令,设置闪避效果开关状态所对应的控制参数,并对应切换闪避开关控件的开关显示状态。
和/或,
S5033、响应于用户对扬声器选择控件的选择指令,设置直播设备中各个扬声器的工作状态,并对应切换扬声器选择控件的扬声器选中状态。
和/或,
S5034、响应于用户对麦克风选择控件的选择指令,设置直播设备中各个麦克风的工作状态,并对应切换麦克风选择控件的麦克风选中状态。
和/或,
S5035、响应于用户对连麦音量设置控件的第一音量设置操作,设置连麦时对方音量的大小控制参数,并对应改变连麦音量设置控件的显示状态;
和/或,
S5036、响应于用户对音效选择控件的选择指令,设置目标音频合成时的音效控制参数,并对应切换音效选择控件的音效选中状态。
和/或,
S5037、响应于用户对音效音量控件的第二音量设置操作,设置音效对应的音量大小控制参数,并对应改变音效音量控件的显示状态。
在一种可能的设计中,设置窗口包括快捷键设置选项卡,该快捷键设置选项卡包括:已选择快捷键阵列以及待选快捷键阵列,则本步骤还包括:
S5038、响应于用户对已选择快捷键阵列的删除指令,更新已选择快捷键阵列以及待选快捷键阵列的展示状态。
和/或,
S5039、响应于用户对待选快捷键阵列的选择指令,更新已选择快捷键阵列以及待选快捷键阵列的展示状态。
需要说明的是,S5031-S5039并没有先后顺序的要求,用户可以同时设置几个参数。
图6a-6b为本公开实施例提供的通过合成控件进行弹窗设置的应用场景示意图。声音设置选项卡的一种实现方式如图6a所示,用户点击了弹窗设置子控件411后,在显示屏下方弹出设置窗口,在设置窗口的声音设置选项卡421上,用户通过拨动耳返或闪避控件的开关来设置开启或关闭耳返效果或闪避效果。通过点击对应的控件来选择扬声器或麦克风。通过拨动音量条来设置各类音量大小。通过点击对应的音效控件来切换音效模式。
需要说明的是,选择控件除了图6a的每种选项单独一个按钮控件外,还可以是通过勾选,或者是通过下拉选择控件来选择。音量条也可以设置为开口弧线、封闭圆环、旋钮等形式。
音效快捷键选项卡422的一种实现方式如图6b所示,方框中显示的是已经被选中显示的待合成音频所对应的快捷键,在本实施例中,最多允许添加8个快捷键,点击已被选中的快捷键右侧的“垃圾箱”状按钮即删除按钮,即可设置不显示此快捷键。
下方以阵列形式排布着多个待合成音频选项,点击右侧的“+”号图标,即可将此待合成音频添加到上方的处于显示状态的快捷键方框中,并且,选中后,切换该选择控件的背景色,并且将“+”改为“√”表明,此选项已被选择。
此外,待合成音频选项还可以通过网络接口从云端服务器或互联网服务器中下载,如图6b中“口哨”,“尖叫”对应的待合成音频,在直播一体式终端中没有预先存储有的,则只需点击控件右侧的下载按钮,即可从服务器下载到本地存储中。
值得说明的是,在一种可能的设计中,只要云端服务器或互联网服务器中更新了待合成音频选项列表,则在音效快捷键选项卡422中也进行同步更新,用户看到有新选项后,可以点击下载按钮下载。当然,用户也可以设置自动下载,或批量下载,一次性将所有更新的待合成音频都下载到本地存储中。
需要说明的是,音效快捷键的设置也可以采用下拉框选择,或者是其它的选择方式,本领域技术人员可以根据实际情况进行选用,本申请不做限定。
本公开实施例提供的音频合成方法,在图2和图3所示实施例的基础上,获取用户对合成控件的弹窗设置操作,然后响应于弹窗设置操作,在显示屏上叠加显示设置窗口,再响应于用户对各个控件的操作指令,对应设置各项音频控制参数,解决了主播在直播时无法简便快捷地更改各项声音相关的设置参数,需要专业技术人员辅助的技术问题。使得声音设置和合成控件的快捷键设置简单方便,无需改变视线方向,影响直播效果,也无需暂停直播来设置,提高了用户对于直播一体式终端的使用体验感。
在上述三个实施例的基础上,本公开提供的音频合成方法还能够在直播前,在直播一体式终端上,通过专门的声音设置界面对合成控件的显示状态以及各项声音控制参数,下面以具体实施例来对具体步骤进行说明介绍。
图7为本公开实施例提供的音频合成方法的流程示意图四。本实施例的方法应用在直播一体式终端中,该音频合成方法包括:
S701:在没有开启摄像头或者直播开启之前,获取用户的声音设置操作。
在本步骤中,直播一体式终端检测到用户点击声音设置按钮,或者是通过预设的直播应用开始时的设置指令,进行了声音设置界面的开启操作。
S702:响应于用户的声音设置操作,在显示屏上显示声音设置界面。
在本实施例中,声音设置界面包括:耳返开关控件、闪避开关控件、扬声器选择控件、麦克风选择控件、连麦音量设置控件、音效选择控件、音效音量控件、合成控件(如悬浮窗)显示开关控件、已选择快捷键阵列以及待选快捷键阵列等控件中的至少一个控件。
S703:根据用户对声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数。
在本实施例中,具体的:
S7031、耳返开关控件的开关指令,设置耳返效果开关状态所对应的控制参数,并对应切换耳返开关控件的开关显示状态。
和/或,
S7032、响应于用户对闪避开关控件的开关指令,设置闪避效果开关状态所对应的控制参数,并对应切换闪避开关控件的开关显示状态。
和/或,
S7033、响应于用户对扬声器选择控件的选择指令,设置直播设备中各个扬声器的工作状态,并对应切换扬声器选择控件的扬声器选中状态。
和/或,
S7034、响应于用户对麦克风选择控件的选择指令,设置直播设备中各个麦克风的工作状态,并对应切换麦克风选择控件的麦克风选中状态。
和/或,
S7035、响应于用户对连麦音量设置控件的第一音量设置操作,设置连麦时对方音量的大小控制参数,并对应改变连麦音量设置控件的显示状态;
和/或,
S7036、响应于用户对音效选择控件的选择指令,设置目标音频合成时的音效控制参数,并对应切换音效选择控件的音效选中状态。
和/或,
S7037、响应于用户对音效音量控件的第二音量设置操作,设置音效对应的音量大小控制参数,并对应改变音效音量控件的显示状态。
和/或,
S7038、响应于用户对已选择快捷键阵列的删除指令,更新已选择快捷键阵列以及待选快捷键阵列的展示状态。
和/或,
S7039、响应于用户对待选快捷键阵列的选择指令,更新已选择快捷键阵列以及待选快捷键阵列的展示状态。
和/或,
S70310、响应于用户对合成控件(如悬浮窗)显示开关的开关指令,对应设置合成控件是否在显示屏上显示。
需要说明的是,S7031-S70310并没有先后顺序的要求,用户可以同时设置几个参数。
具体的,一种可能的实现方式可以参考图8。
图8为本公开实施例提供的一种声音设置界面示意图。如图8所示,用户在声音设置界面上,通过扬声器选择控件,用户可以选择用“有线耳机”或者是直播一体设备自带的扬声器,即“本机扬声器”;通过麦克风选择控件,用户可以选择“有线耳麦”、“本机麦克风”、“卡农麦克风”;通过耳返开关控件,用户通过向右/向左拨动开关控件,开启或关闭耳返效果,并且在耳返开关控件上,还包括提示文本,如“本机扬声器不支持耳返功能”,提示用户耳返功能与扬声器选择直接是具备关联设置的,以避免用户参数选择不当,影响使用效果;同理,闪避开关控件和悬浮窗开关控件也具备类似的操作,具体内容参见图8。
图8中的美声选项即为音效美化效果的选择控件,通过单击右侧展开图标“>”,即可展开音效列表,供用户选择对应的音效美化效果,在本公开的各个实施例中,音效美化效果包括:原声、KTV、磁性、老唱片、空灵、剧场等音效特性,每种音效对应着多个音效设置参数,通过选择特定音效,可以快速设置各个参数,降低了音效设置的专业门槛,使得不具备专业技术的主播也可以进行简单方便的音效设置。
图8的最下方为悬浮窗即合成控件的快捷键选择阵列,用户只需要选中各个按钮控件,则设置显示对应的待合成音频的快捷键,便于用户在直播开始后,在直播界面中通过快捷键直接进行音频合成。
本公开实施例提供的音频合成方法,在图2、图3和图5所示实施例的基础上,通过在没有开启摄像头或者直播开启之前,获取用户的声音设置操作,然后响应于用户的声音设置操作,在显示屏上显示声音设置界面,再根据用户对声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,解决了主播无法简便快捷地更改各项声音相关的设置参数,需要专业技术人员辅助的技术问题。使得声音设置和合成控件的快捷键设置简单方便,提高了用户对于直播一体式终端的使用体验感。
在一种可能的设计中,在本公开的各个实施例的基础上,直播一体机即直播一体化设备通过外接接口连接至少一个快捷键键盘,步骤S202:接收用户对合成控件的音频合成操作,包括;
接收快捷键键盘发送的操作指令,操作指令是快捷键键盘响应于用户的操作而确定的;
根据操作指令确定对应的待合成音频。
值得说明的是,外接接口包括:无线接口以及有线接口。
图9为本公开实施例提供的一种直播一体化设备外接快捷键键盘的场景示意图。如图9所示,直播一体机91通过有线接口或无线接口连接快捷键键盘92,用于实现本公开各个实施例的音频合成方法的合成控件911中的多个子控件与键盘上的按键呈一一对应关系。
在一种可能的设计中,快捷键键盘上各个按键与显示屏上各个子控件的对应关系是可以任意设置的。快捷键键盘包括:机械按钮式键盘、触控屏等。用户点击快捷键相当于在显示屏上点击了对应的子控件,同时,子控件切换显示为被选中的状态。
还需要说明的是,子控件在被点击选中后,可以用条形进度条或扇形进度图的形式,同步显示待合成音频的播放进度。
图10a-10c为本公开实施例提供的一种合成控件在待合成音频播放时显示播放进度的示意图。如图10a所示,在用户点击了“笑声1”子控件后,在子控件上显示播放“笑声1”所对应的音频的播放进度条。
如图10b所示,在用户点击了“笑声1”子控件后,子控件的背景颜色或背景图案以旋转渐变的形式来显示“笑声1”所对应的音频的播放进度条。
如图10c所示,在用户点击了“笑声1”子控件后,子控件的背景颜色或背景图案以从左至右、从右至左、从上至下、从下至上、从左下至右上、从右下至左上等方式中的任意一种渐变的形式来显示“笑声1”所对应的音频的播放进度条。
这样主播就能够知道待合成音频的播放进度,进一步的,可以通过再次点击子控件结束播放,或者是点击其它子控件混合播放多个待合成音频,增加了直播音频合成的操作灵活性,提升直播时的现场音频效果,提高用户体验。
对应于上文实施例的音频合成方法,图11为本公开实施例提供的音频合成装置1100的结构框图。为了便于说明,仅示出了与本公开实施例相关的部分。参照图11,装置包括:
显示模块1101,用于在满足预设显示条件时,以第一形态在显示屏上显示合成控件,所述第一形态对所述显示屏上显示内容的视觉影响满足预设视觉要求;
接收模块1102,用于接收用户对所述合成控件的音频合成操作;
处理模块1103,用于在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为所述直播应用的输入音频,其中,所述待合成音频为所述音频合成操作所指示的音频,所述初始音频为所述直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
在本公开的一个实施例中,所述显示屏包括触控屏,所述接收模块1102,用于实时检测所述用户利用预设操作物靠近或触碰到所述触控屏上的所述合成控件。
在本公开的一个实施例中,所述合成控件包括与所述待合成音频对应的子控件,所述接收模块1102,具体用于:
若检测到所述预设操作物靠近或触碰到所述子控件,则将所述子控件切换为被选中形态;
确定与所述子控件对应的合成操作指令。
在本公开的一个实施例中,所述预设显示条件包括:检测到摄像设备开启指令、检测到直播终端按预设运动方式进行运动、接收到预设语音指令和接收到预设按钮开关开启指令中的至少一种。
在本公开的一个实施例中,所述合成控件包括悬浮窗控件,所述显示模块1101,用于根据预设透明度、预设大小以及预设形状在显示屏的边缘显示所述悬浮窗控件。
在本公开的一个实施例中,所述接收模块1102,还用于获取所述用户对所述合成控件的状态设置操作;
所述显示模块1101,还用于根据所述状态设置操作切换所述合成控件的形态,和/或,显示位置。
在本公开的一个实施例中,所述显示模块1101,还用于响应于所述用户对所述合成控件的弹窗设置操作,在显示屏上叠加显示设置窗口;
所述处理模块1103,还用于根据所述用户对所述设置窗口上各个参数控件的设置指令,对应设置各项音频控制参数。
在本公开的一个实施例中,所述设置窗口包括声音设置选项卡,所述声音设置选项卡包括:耳返开关控件,和/或,闪避开关控件,所述处理模块1103,具体用于:
响应于所述用户对所述耳返开关控件的开关指令,设置耳返效果开关状态所对应的控制 参数,并对应切换所述耳返开关控件的开关显示状态;
和/或,
响应于所述用户对所述闪避开关控件的开关指令,设置闪避效果开关状态所对应的控制参数,并对应切换所述闪避开关控件的开关显示状态。
在本公开的一个实施例中,所述设置窗口包括声音设置选项卡,所述声音设置选项卡包括:扬声器选择控件,所述处理模块1103,具体用于:
响应于所述用户对所述扬声器选择控件的选择指令,设置所述直播设备中各个扬声器的工作状态,并对应切换所述扬声器选择控件的扬声器选中状态。
在本公开的一个实施例中,所述设置窗口包括声音设置选项卡,所述声音设置选项卡包括:麦克风选择控件,所述处理模块1103,具体用于:
响应于所述用户对所述麦克风选择控件的选择指令,设置所述直播设备中各个麦克风的工作状态,并对应切换所述麦克风选择控件的麦克风选中状态。
在本公开的一个实施例中,所述设置窗口包括声音设置选项卡,所述声音设置选项卡包括:连麦音量设置控件,所述处理模块1103,具体用于:
响应于所述用户对所述连麦音量设置控件的第一音量设置操作,设置连麦时对方音量的大小控制参数,并对应改变所述连麦音量设置控件的显示状态。
在本公开的一个实施例中,所述设置窗口包括声音设置选项卡,所述声音设置选项卡包括:音效选择控件,和/或,音效音量控件,所述处理模块1103,具体用于:
响应于所述用户对所述音效选择控件的选择指令,设置所述目标音频合成时的音效控制参数,并对应切换所述音效选择控件的音效选中状态;
和/或,
响应于所述用户对所述音效音量控件的第二音量设置操作,设置音效对应的音量大小控制参数,并对应改变所述音效音量控件的显示状态。
在本公开的一个实施例中,所述设置窗口包括快捷键设置选项卡,所述快捷键设置选项卡包括:已选择快捷键阵列以及待选快捷键阵列,所述处理模块1103,具体用于:
响应于所述用户对所述已选择快捷键阵列的删除指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态;
和/或,
响应于所述用户对所述待选快捷键阵列的选择指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态。
在本公开的一个实施例中,在所述在满足预设显示条件时,以第一形态在显示屏上显示合成控件之前,所述显示模块1101,还用于响应于所述用户的声音设置操作,在显示屏上显示声音设置界面;
所述处理模块1103,还用于根据所述用户对所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数。
在本公开的一个实施例中,所述声音设置界面包括:耳返开关控件,和/或,闪避开关控件,所述处理模块1103,还用于:
响应于所述用户对所述耳返开关控件的开关指令,设置耳返效果开关状态所对应的控制参数,并对应切换所述耳返开关控件的开关显示状态;
和/或,
响应于所述用户对所述闪避开关控件的开关指令,设置闪避效果开关状态所对应的控制参数,并对应切换所述闪避开关控件的开关显示状态。
在本公开的一个实施例中,所述声音设置界面包括:扬声器选择控件,所述处理模块1103,还用于:
响应于所述用户对所述扬声器选择控件的选择指令,设置所述直播设备中各个扬声器的工作状态,并对应切换所述扬声器选择控件的扬声器选中状态。
在本公开的一个实施例中,所述声音设置界面包括:麦克风选择控件,所述处理模块1103,还用于:
响应于所述用户对所述麦克风选择控件的选择指令,设置所述直播设备中各个麦克风的工作状态,并对应切换所述麦克风选择控件的麦克风选中状态。
在本公开的一个实施例中,所述声音设置界面包括:连麦音量设置控件,所述处理模块1103,还用于:
响应于所述用户对所述连麦音量设置控件的第一音量设置操作,设置连麦时对方音量的大小控制参数,并对应改变所述连麦音量设置控件的显示状态。
在本公开的一个实施例中,所述声音设置界面包括:音效选择控件,和/或,音效音量控件,所述处理模块1103,还用于:
响应于所述用户对所述音效选择控件的选择指令,设置所述目标音频合成时的音效控制参数,并对应切换所述音效选择控件的音效选中状态;
和/或,
响应于所述用户对所述音效音量控件的第二音量设置操作,设置音效对应的音量大小控制参数,并对应改变所述音效音量控件的显示状态。
在本公开的一个实施例中,所述声音设置界面包括:已选择快捷键阵列以及待选快捷键阵列,所述处理模块1103,还用于:
响应于所述用户对所述已选择快捷键阵列的删除指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态;
和/或,
响应于所述用户对所述待选快捷键阵列的选择指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态。
在本公开的一个实施例中,所述声音设置界面包括:合成控件显示开关,所述处理模块1103,还用于:
响应于所述用户对所述合成控件显示开关的开关指令,对应设置所述合成控件是否在显示屏上显示。
在本公开的一个实施例中,所述音频合成装置1100通过外接接口连接至少一个快捷键键盘,所述接收用户对所述合成控件的音频合成操作,包括;
所述接收模块1102,还用于接收所述快捷键键盘发送的操作指令,所述操作指令是所述快捷键键盘响应于所述用户的操作而确定的;
所述处理模块1103,还用于根据所述操作指令确定对应的所述待合成音频。
本实施例提供的装置1100,可用于执行上述方法实施例的技术方案中用户端相关的步 骤,其实现原理和技术效果类似,本实施例此处不再赘述。
参考图12,其示出了适于用来实现本公开实施例的电子设备的结构示意图,该电子设备1200可以为终端设备或服务器。其中,终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,简称PDA)、平板电脑(Portable Android Device,简称PAD)、便携式多媒体播放器(Portable Media Player,简称PMP)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图12示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图12所示,电子设备1200可以包括处理装置(例如中央处理器、图形处理器等)1201,其可以根据存储在只读存储器(Read Only Memory,简称ROM)1202中的程序或者从存储装置1208加载到随机访问存储器(Random Access Memory,简称RAM)1203中的程序而执行各种适当的动作和处理。在RAM 1203中,还存储有电子设备1200操作所需的各种程序和数据。处理装置1201、ROM 1202以及RAM 1203通过总线1204彼此相连。输入/输出(I/O)接口1205也连接至总线1204。
通常,以下装置可以连接至I/O接口1205:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置12012;包括例如液晶显示器(Liquid Crystal Display,简称LCD)、扬声器、振动器等的输出装置1207;包括例如磁带、硬盘等的存储装置1208;以及通信装置1209。通信装置1209可以允许电子设备1200与其他设备进行无线或有线通信以交换数据。虽然图12示出了具有各种装置的电子设备1200,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置1209从网络上被下载和安装,或者从存储装置1208被安装,或者从ROM 1202被安装。在该计算机程序被处理装置1201执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Erasable Programmable Read-Only Memory,简称EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(Compact Disc Read-Only Memory,简称CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用 于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequency,简称RF)等等,或者上述的任意合适的组合。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备执行上述实施例所示的方法。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(Local Area Network,简称LAN)或广域网(Wide Area Network,简称WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(Field Programmable Gate Array,简称FPGA)、专用集成电路(Application Specific Intergrated Circuit,简称ASIC)、专用标准产品(Application Specific Standard Product,简称ASSP)、片上系统(System on a Chip,简称SOC)、复杂可编程逻辑设备(Complex Programmable Logic Device,简称CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器 (EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
本公开实施例中还提供了一种计算机程序产品,包括计算机程序,其特征在于,该计算机程序被处理器执行时实现上述各实施例中的方法。
本公开实施例中还提供了一种直播一体机或直播一体化设备,包括图12所对应的电子设备。还需要说明的是,该直播一体机的控制电路包括:主控模块和音频处理模块,其中,主控模块上安装了直播应用,音频处理模块用于将待合成音频合成到初始音频中,以形成目标音频。待合成音频可以存储在音频处理模块中,也可以存储在主控模块中,还可以存储在直播一体机的存储器中。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (21)

  1. 一种音频合成方法,其特征在于,包括:
    在满足预设显示条件时,以第一形态在显示屏上显示合成控件,所述第一形态对所述显示屏上显示内容的视觉影响满足预设视觉要求;
    接收用户对所述合成控件的音频合成操作;
    在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为所述直播应用的输入音频,其中,所述待合成音频为所述音频合成操作所指示的音频,所述初始音频为所述直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
  2. 根据权利要求1所述的音频合成方法,其特征在于,所述显示屏包括触控屏,所述接收用户对所述合成控件的音频合成操作,包括:
    实时检测所述用户利用预设操作物靠近或触碰到所述触控屏上的所述合成控件。
  3. 根据权利要求2所述的音频合成方法,其特征在于,所述合成控件包括与所述待合成音频对应的子控件,所述实时检测所述用户利用预设操作物靠近或触碰到所述触控屏上的所述合成控件,包括:
    若检测到所述预设操作物靠近或触碰到所述子控件,则将所述子控件切换为被选中形态;
    确定与所述子控件对应的合成操作指令。
  4. 根据权利要求1至3中任意一项所述的音频合成方法,其特征在于,所述预设显示条件包括:检测到摄像设备开启指令、检测到直播终端按预设运动方式进行运动、接收到预设语音指令和接收到预设按钮开关开启指令中的至少一种。
  5. 根据权利要求1至4中任意一项所述的音频合成方法,其特征在于,所述合成控件包括悬浮窗控件,所述以第一形态在显示屏上显示合成控件,包括:
    根据预设透明度、预设大小以及预设形状在显示屏的边缘显示所述悬浮窗控件。
  6. 根据权利要求1至5中任意一项所述的音频合成方法,其特征在于,还包括:
    获取所述用户对所述合成控件的状态设置操作;
    根据所述状态设置操作切换所述合成控件的形态,和/或,显示位置。
  7. 根据权利要求1至6中任意一项所述的音频合成方法,其特征在于,还包括:
    响应于所述用户在直播时对所述合成控件的弹窗设置操作,在显示屏上叠加显示设置窗口;
    和/或,响应于所述用户在直播前触发的声音设置操作,在显示屏上显示声音设置界面;
    根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数。
  8. 根据权利要求7所述的音频合成方法,其特征在于,所述设置窗口包括声音设置选项卡,所述声音设置选项卡,和/或,所述声音设置界面中包括:耳返开关控件,和/或,闪避开关控件,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述耳返开关控件的开关指令,设置耳返效果开关状态所对应的控制参数,并对应切换所述耳返开关控件的开关显示状态;
    和/或,
    响应于所述用户对所述闪避开关控件的开关指令,设置闪避效果开关状态所对应的控制参数,并对应切换所述闪避开关控件的开关显示状态。
  9. 根据权利要求7或8所述的音频合成方法,其特征在于,所述设置窗口包括声音设置选项卡,所述声音设置选项卡,和/或,所述声音设置界面中包括:扬声器选择控件,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述扬声器选择控件的选择指令,设置所述直播设备中各个扬声器的工作状态,并对应切换所述扬声器选择控件的扬声器选中状态。
  10. 根据权利要求7至9中任意一项所述的音频合成方法,其特征在于,所述设置窗口包括声音设置选项卡,所述声音设置选项卡,和/或,所述声音设置界面中包括:麦克风选择控件,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述麦克风选择控件的选择指令,设置所述直播设备中各个麦克风的工作状态,并对应切换所述麦克风选择控件的麦克风选中状态。
  11. 根据权利要求7至10中任意一项所述的音频合成方法,其特征在于,所述设置窗口包括声音设置选项卡,所述声音设置选项卡,和/或,所述声音设置界面中包括:连麦音量设置控件,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述连麦音量设置控件的第一音量设置操作,设置连麦时对方音量的大小控制参数,并对应改变所述连麦音量设置控件的显示状态。
  12. 根据权利要求7至11中任意一项所述的音频合成方法,其特征在于,所述设置窗口包括声音设置选项卡,所述声音设置选项卡,和/或,所述声音设置界面中包括:音效选择控件,和/或,音效音量控件,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述音效选择控件的选择指令,设置所述目标音频合成时的音效控制参数,并对应切换所述音效选择控件的音效选中状态;
    和/或,
    响应于所述用户对所述音效音量控件的第二音量设置操作,设置音效对应的音量大小控制参数,并对应改变所述音效音量控件的显示状态。
  13. 根据权利要求7至12中任意一项所述的音频合成方法,其特征在于,所述设置窗口包括快捷键设置选项卡,所述快捷键设置选项卡,和/或,所述声音设置界面中包括:已选择快捷键阵列以及待选快捷键阵列,所述根据所述用户对所述设置窗口或所述声音设置界面上各个参数控件的设置指令,对应设置各项音频控制参数,包括:
    响应于所述用户对所述已选择快捷键阵列的删除指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态;
    和/或,
    响应于所述用户对所述待选快捷键阵列的选择指令,更新所述已选择快捷键阵列以及所述待选快捷键阵列的展示状态。
  14. 根据权利要求1至13中任意一项所述的音频合成方法,其特征在于,所述方法应用 于直播一体机,所述直播一体机包括:主控模块以及音频处理模块,所述直播应用安装于所述主控模块中,所述音频处理模块用于合成所述目标音频。
  15. 根据权利要求14所述的音频合成方法,其特征在于,所述直播一体机通过外接接口连接至少一个快捷键键盘,所述接收用户对所述合成控件的音频合成操作,包括;
    接收所述快捷键键盘发送的操作指令,所述操作指令是所述快捷键键盘响应于所述用户的操作而确定的;
    根据所述操作指令确定对应的所述待合成音频。
  16. 一种音频合成装置,其特征在于,包括:
    显示模块,用于在满足预设显示条件时,以第一形态在显示屏上显示合成控件,所述第一形态对所述显示屏上显示内容的视觉影响满足预设视觉要求;
    接收模块,用于接收用户对所述合成控件的音频合成操作;
    处理模块,用于在安装了直播应用的直播设备中,将待合成音频与初始音频进行合成,并将合成后的目标音频作为所述直播应用的输入音频,其中,所述待合成音频为所述音频合成操作所指示的音频,所述初始音频为所述直播设备通过音频采集设备接收到主播所发出的,和/或,直播环境中的声音信号。
  17. 一种电子设备,其特征在于,包括:
    至少一个处理器和存储器;
    所述存储器用于存储计算机程序;
    所述至少一个处理器执行所述存储器存储的所述计算机程序,使得所述至少一个处理器执行如权利要求1至15中任意一项所述的音频合成方法。
  18. 一种直播一体机,其特征在于,包括:权利要求17所述的电子设备。
  19. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机程序,当处理器执行所述计算机程序时,实现如权利要求1至15中任意一项所述的音频合成方法。
  20. 一种计算机程序产品,包括计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1至15中任意一项所述的音频合成方法。
  21. 一种计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1至15中任意一项所述的音频合成方法。
PCT/CN2022/087491 2021-05-13 2022-04-18 音频合成方法、装置、设备、介质及程序产品 WO2022237464A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110523959.3 2021-05-13
CN202110523959.3A CN113112986A (zh) 2021-05-13 2021-05-13 音频合成方法、装置、设备、介质及程序产品

Publications (1)

Publication Number Publication Date
WO2022237464A1 true WO2022237464A1 (zh) 2022-11-17

Family

ID=76722536

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/087491 WO2022237464A1 (zh) 2021-05-13 2022-04-18 音频合成方法、装置、设备、介质及程序产品

Country Status (2)

Country Link
CN (1) CN113112986A (zh)
WO (1) WO2022237464A1 (zh)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113157244A (zh) * 2021-05-13 2021-07-23 北京字节跳动网络技术有限公司 音效调节方法、装置、设备、介质及程序产品
CN113112986A (zh) * 2021-05-13 2021-07-13 北京字节跳动网络技术有限公司 音频合成方法、装置、设备、介质及程序产品
CN113542792B (zh) * 2021-07-14 2023-04-07 北京字节跳动网络技术有限公司 音频合流方法、音频上传方法、设备及程序产品
CN114095480B (zh) * 2022-01-24 2022-04-15 北京麦颂文化传播有限公司 Ktv直播连麦方法、装置和系统
CN115086729B (zh) * 2022-06-10 2024-04-26 北京字跳网络技术有限公司 一种连麦展示方法、装置、电子设备、计算机可读介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111309214A (zh) * 2020-03-17 2020-06-19 网易(杭州)网络有限公司 一种视频界面的设置方法及装置、电子设备、存储介质
CN111510738A (zh) * 2020-04-26 2020-08-07 北京字节跳动网络技术有限公司 一种直播中音频的传输方法及装置
CN111880694A (zh) * 2020-07-22 2020-11-03 维沃移动通信有限公司 显示方法、装置、设备及存储介质
CN112347273A (zh) * 2020-11-05 2021-02-09 北京字节跳动网络技术有限公司 音频播放方法、装置、电子设备及存储介质
CN112423009A (zh) * 2020-11-09 2021-02-26 珠海格力电器股份有限公司 一种控制直播音频的方法及设备
CN113112986A (zh) * 2021-05-13 2021-07-13 北京字节跳动网络技术有限公司 音频合成方法、装置、设备、介质及程序产品

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150078584A1 (en) * 2013-09-16 2015-03-19 Nancy Diane Moon Live Sound Mixer User Interface
CN105786439B (zh) * 2014-12-26 2019-05-07 腾讯科技(深圳)有限公司 一种终端中声音的控制方法和装置
CN105872253B (zh) * 2016-05-31 2020-07-07 腾讯科技(深圳)有限公司 一种直播声音处理方法及移动终端
CN207010753U (zh) * 2017-07-28 2018-02-13 徐文波 一种具有硬件音效处理功能的智能移动终端
CN109257493B (zh) * 2018-09-15 2021-04-16 深圳市创成微电子有限公司 一种基于app的音频处理装置控制系统
CN109814798B (zh) * 2019-01-17 2021-03-12 Oppo广东移动通信有限公司 耳返功能控制方法、装置以及移动终端
CN111208969A (zh) * 2020-01-03 2020-05-29 海信视像科技股份有限公司 一种声音输出设备的选择控制方法及显示设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111309214A (zh) * 2020-03-17 2020-06-19 网易(杭州)网络有限公司 一种视频界面的设置方法及装置、电子设备、存储介质
CN111510738A (zh) * 2020-04-26 2020-08-07 北京字节跳动网络技术有限公司 一种直播中音频的传输方法及装置
CN111880694A (zh) * 2020-07-22 2020-11-03 维沃移动通信有限公司 显示方法、装置、设备及存储介质
CN112347273A (zh) * 2020-11-05 2021-02-09 北京字节跳动网络技术有限公司 音频播放方法、装置、电子设备及存储介质
CN112423009A (zh) * 2020-11-09 2021-02-26 珠海格力电器股份有限公司 一种控制直播音频的方法及设备
CN113112986A (zh) * 2021-05-13 2021-07-13 北京字节跳动网络技术有限公司 音频合成方法、装置、设备、介质及程序产品

Also Published As

Publication number Publication date
CN113112986A (zh) 2021-07-13

Similar Documents

Publication Publication Date Title
WO2022237464A1 (zh) 音频合成方法、装置、设备、介质及程序产品
CN109076125B (zh) 一种显示方法及设备
WO2021052255A1 (zh) 热点推荐弹窗的控制方法、装置、介质及电子设备
WO2019105438A1 (zh) 视频特效添加方法、装置及智能移动终端
US20170185373A1 (en) User terminal device, and mode conversion method and sound system for controlling volume of speaker thereof
WO2016124095A1 (zh) 生成视频的方法、装置及终端
CN107003818A (zh) 在设备之间共享屏幕的方法及使用该方法的设备
JP4908360B2 (ja) 携帯情報端末連係システムおよび連係処理プログラム並びに連係処理装置
KR20150007897A (ko) 외부 기기에 디스플레이된 정보와 관련된 기능을 실행하는 휴대용 기기, 휴대용 기기에서 외부 기기에 디스플레이된 정보와 관련된 기능을 실행하는 방법 및 컴퓨터 판독 가능한 기록 매체
CN111770366A (zh) 一种消息补发方法、服务器及显示设备
JP2021505092A (ja) オーディオデータを再生する方法及び装置
KR20150079371A (ko) 단말기에서 근전도 검사 장치로 데이터를 전송하기 위한 장치, 시스템 및 방법
EP4124052A1 (en) Video production method and apparatus, and device and storage medium
KR20140070218A (ko) 종료 효과를 표시하는 모바일 장치 및 그 제어방법
CN103677377A (zh) 实现终端展示的系统、装置
WO2022042001A1 (zh) 信息显示方法、装置、设备及存储介质
WO2023202359A1 (zh) 直播内容显示方法、装置、设备、可读存储介质及产品
CN103778549A (zh) 移动应用推广系统及方法
WO2023185647A1 (zh) 媒体内容的显示方法、装置、设备、存储介质和程序产品
US11886484B2 (en) Music playing method and apparatus based on user interaction, and device and storage medium
CN105721904B (zh) 显示装置和控制显示装置的内容输出的方法
CN112437334A (zh) 一种显示设备
KR20150025302A (ko) 외부 단말과의 연동을 실행하는 방법, 장치 및 기록매체
CN112506859A (zh) 一种维护硬盘数据的方法及显示设备
CN110022445B (zh) 一种内容输出方法及终端设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22806434

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE