WO2022242351A1 - 一种多媒体处理方法、装置、设备及介质 - Google Patents

一种多媒体处理方法、装置、设备及介质 Download PDF

Info

Publication number
WO2022242351A1
WO2022242351A1 PCT/CN2022/085468 CN2022085468W WO2022242351A1 WO 2022242351 A1 WO2022242351 A1 WO 2022242351A1 CN 2022085468 W CN2022085468 W CN 2022085468W WO 2022242351 A1 WO2022242351 A1 WO 2022242351A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
multimedia interface
multimedia
interface
target audio
Prior art date
Application number
PCT/CN2022/085468
Other languages
English (en)
French (fr)
Inventor
陈可蓉
周伊诺
龚彪
杨晶生
赵田
刘敬晖
吕大千
杨耀
成涛
潘灶烽
史田辉
唐荣意
贡国栋
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Priority to US18/262,301 priority Critical patent/US20240121479A1/en
Publication of WO2022242351A1 publication Critical patent/WO2022242351A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • H04N21/4415Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8166Monomedia components thereof involving executable data, e.g. software
    • H04N21/8173End-user applications, e.g. Web browser, game

Definitions

  • the present disclosure relates to the field of multimedia technology, and in particular to a multimedia processing method, device, equipment and medium.
  • the multimedia file of the information recording can be played back for further review.
  • the way of playing back multimedia files is relatively fixed, and the way is single, and the flexibility is not high.
  • the present disclosure provides a multimedia processing method, device, device and medium.
  • An embodiment of the present disclosure provides a multimedia processing method, the method comprising:
  • the first multimedia interface includes first content
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio
  • An embodiment of the present disclosure also provides a multimedia processing device, the device comprising:
  • the first interface module is configured to display a first multimedia interface, and the first multimedia interface includes first content
  • a request module configured to receive an interface switching request from a user in the first multimedia interface
  • the second interface module is used to switch the currently displayed first multimedia interface to a second multimedia interface, and display second content in the second multimedia interface;
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio
  • An embodiment of the present disclosure also provides an electronic device, which includes: a processor; a memory for storing instructions executable by the processor; and the processor, for reading the instruction from the memory.
  • the instructions can be executed, and the instructions are executed to implement the multimedia processing method provided by the embodiments of the present disclosure.
  • the embodiment of the present disclosure also provides a computer-readable storage medium, the storage medium stores a computer program, and the computer program is used to execute the multimedia processing method provided by the embodiment of the present disclosure.
  • the embodiment of the present disclosure also provides a computer program, including instructions, and when the instruction is executed by a processor, the processor executes the multimedia processing method provided by the embodiment of the present disclosure.
  • An embodiment of the present disclosure also provides a computer program product, including instructions, which, when executed by a processor, cause the processor to execute the multimedia processing method provided by the embodiments of the present disclosure.
  • FIG. 1 is a schematic flowchart of a multimedia processing method provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic flowchart of another multimedia processing method provided by an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of a multimedia interface provided by an embodiment of the present disclosure.
  • FIG. 4 is a schematic diagram of another multimedia interface provided by an embodiment of the present disclosure.
  • Fig. 5 is a schematic diagram of a floating window assembly provided by an embodiment of the present disclosure.
  • FIG. 6 is a schematic structural diagram of a multimedia processing device provided by an embodiment of the present disclosure.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the term “comprise” and its variations are open-ended, ie “including but not limited to”.
  • the term “based on” is “based at least in part on”.
  • the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one further embodiment”; the term “some embodiments” means “at least some embodiments.” Relevant definitions of other terms will be given in the description below.
  • the technical solutions provided by the embodiments of the present disclosure have the following advantages: the multimedia processing solutions provided by the embodiments of the present disclosure display the first multimedia interface, and the first multimedia interface includes the first content; Interface switching request in the first multimedia interface; switch the currently displayed first multimedia interface to the second multimedia interface, and display the second content in the second multimedia interface; wherein, the first content It includes second content and other content associated with the second content, and the second content includes target audio and target subtitles corresponding to the target audio.
  • FIG. 1 is a schematic flowchart of a multimedia processing method provided by an embodiment of the present disclosure.
  • the method can be executed by a multimedia processing device, where the device can be implemented by software and/or hardware, and generally can be integrated into an electronic device. As shown in Figure 1, the method includes:
  • Step 101 Display a first multimedia interface, where the first multimedia interface includes first content.
  • the multimedia interface refers to an interface for displaying various types of multimedia information, and the multimedia information may include audio, video, text, etc., without limitation.
  • the first multimedia interface refers to one of the multimedia interfaces, and the first content refers to the content displayed on the first multimedia interface, which may include a variety of multimedia information.
  • the first content may be Including recording and drawing audio and/or video, corresponding subtitle content, meeting minutes and other content related to the meeting.
  • the client may obtain the first content according to the user's request, and display the first multimedia interface and display the first content in the first multimedia interface. Since the first content can include multiple types of information, different display areas can be set for each type of information in the first multimedia interface for display, for example, audio and video areas and subtitle areas can be set in the first multimedia interface And the minutes display area and other areas are used to display audio, video, subtitle content and minutes respectively.
  • Step 102 Receive an interface switching request from the user in the first multimedia interface.
  • the interface switching request refers to a request for switching between different interfaces.
  • the interface switching button may be a virtual button preset in the first multimedia interface, and the specific position and expression form are not limited.
  • Step 103 switch the currently displayed first multimedia interface to a second multimedia interface, and display the second content in the second multimedia interface.
  • the second multimedia interface is a multimedia interface different from the display content of the above-mentioned first multimedia interface.
  • the first content displayed in the first multimedia interface may include the second content and other content associated with the second content, that is, the second content may be a part of the first content in the first multimedia interface.
  • the second content can include target audio and target subtitles corresponding to the target audio.
  • the target audio can be any audio data that records information.
  • the target audio can be the audio data during the recording meeting.
  • ASR Automatic Speech Recognition
  • the currently displayed first multimedia interface can be closed and the second multimedia interface can be opened, and the second content can be displayed in the second multimedia interface to realize interface switching. switch. Since only audio and subtitles are included in the second multimedia interface, it is helpful for the user to focus on the multimedia content in complex scenes.
  • the floating window component of the second multimedia interface may be displayed for the user to quickly switch back to the second multimedia interface.
  • the target audio can continue to be played.
  • the display of the second multimedia interface may be returned.
  • the multimedia processing solution provided by the embodiment of the present disclosure displays the first multimedia interface, and the first multimedia interface includes the first content; receives the user's interface switching request in the first multimedia interface; and displays the currently displayed first multimedia interface.
  • a multimedia interface is switched to a second multimedia interface, and the second content is displayed in the second multimedia interface; wherein, the first content includes the second content and other content associated with the second content, and the second content Including the target audio and the target subtitle corresponding to the target audio.
  • the multimedia processing method may further include: receiving a playback trigger operation to the target audio; playing the target audio, and during the playback of the target audio, based on the timestamp of the subtitle sentence included in the target subtitle, matching with the target audio The subtitle sentences corresponding to the audio playback progress are highlighted.
  • the play trigger operation refers to a trigger operation for playing multimedia, and there may be various specific forms of the play trigger operation, which are not limited.
  • Target subtitles belong to structured text, including three layers of paragraphs, sentences and words.
  • a subtitle sentence is a sentence in the target subtitle, and a subtitle sentence can include at least one word or word. Since the target subtitle is obtained by performing speech recognition on the target audio, each subtitle sentence has a corresponding speech sentence, and each speech sentence corresponds to a time stamp in the target audio.
  • Target subtitles can be obtained after the target audio is recognized for speech, and the corresponding speech sentences in the target audio can be determined for each subtitle sentence in the target subtitles.
  • each speech sentence corresponds to a playback time of the target audio, it can be based on The corresponding relationship between the speech sentence and the playback time of the target audio determines the time stamp of each subtitle sentence in the target subtitle.
  • the specific manner of highlighting is not limited in this embodiment of the disclosure. For example, highlighting may be performed in one or more ways of highlighting, bolding, increasing display font size, changing display font, and adding underline.
  • the target audio after receiving the user's play trigger operation on the target audio, the target audio can be played, and in the process of playing the target audio, according to the time stamps of each subtitle sentence included in the target subtitle, the subtitles corresponding to the playback progress can be sequentially
  • the subtitle sentences are displayed prominently, that is, as the target audio is played, the subtitle sentences in the target subtitle are sequentially displayed prominently as the playback progresses.
  • the corresponding subtitle sentences can also be associated and highlighted, which can realize the associated interaction between multimedia and subtitles, so that users can better understand multimedia content and improve user experience.
  • the multimedia processing method may further include: in response to the end of playing the target audio, acquiring the next audio associated with the target audio, and switching to play the next audio.
  • the end of playing the target audio can be determined based on the user's operation, or the end of playing the target audio can also be determined based on its playing progress reaching the moment when the playing is completed.
  • the next audio refers to a pre-set audio associated with the attribute information of the target audio
  • the attribute information is not limited, for example, the attribute information may be time, user or other key information.
  • the next audio may be the audio of the next conference adjacent to the ending time of the conference.
  • the next audio associated with the target audio may be determined, the next audio is acquired, and the next audio is played.
  • the next audio may also be the next audio in the playlist determined based on the attribute values of one or more attribute information. For example, if the attribute information includes meeting date, then the user can determine a playlist based on the meeting date.
  • the next audio is continued to be played.
  • the advantage of this setting is that the next audio can be seamlessly played after the current audio is played, so that the user can learn more about relevant content, avoid the abrupt feeling caused by the stop, and improve the multimedia information playback experience.
  • the multimedia processing method may further include: determining non-mute segments in the target audio; and playing the target audio, including: playing only the non-mute segments when playing the target audio.
  • the multimedia processing method may also include: determining a mute segment and a non-mute segment in the target audio; playing the target audio, including: playing the mute segment at a first playback speed, and playing the non-mute segment at a second playback speed, Wherein, the first playback speed is greater than the second playback speed.
  • the silent segment refers to an audio segment whose volume is zero in the target audio
  • the non-muted segment refers to an audio segment whose volume is not zero in the target audio.
  • the mute segment and the non-mute segment of the target audio may also be determined through volume recognition, and the mute segment is played at the first playback speed, and the non-mute segment is played at the second playback speed.
  • the first playback speed and the second playback speed can be determined according to actual conditions, as long as the first playback speed is greater than the second playback speed, for example, the first playback speed can be set to twice the second playback speed.
  • the multimedia processing method may further include: receiving a user's interaction trigger operation on the second multimedia interface; and determining interaction content based on the interaction trigger operation.
  • determining the interaction content based on the interaction trigger operation includes: displaying the interaction component on the second multimedia interface in response to the interaction trigger operation; acquiring the interaction content based on the interaction component, and displaying the interaction content on the second multimedia interface ;
  • the interactive components include emoticon components and/or comment components, and the interactive content includes interactive emoticons and/or comments.
  • the multimedia processing method may further include: displaying interactive content on the first multimedia interface.
  • the interactive input trigger operation refers to the trigger operation that the user wants to perform interactive input on the current multimedia content.
  • the interactive input trigger operation may be included in the playback time axis of the target audio on the second multimedia interface or
  • the interaction button may be a button preset on the second multimedia interface, and the specific position and style of the button are not limited.
  • Interactive components refer to functional components used for operations such as inputting, editing and publishing interactive content.
  • the interactive component may include emoticon parts and/or comment parts.
  • the emoticon parts are functional parts for emoticon input and may include a set number of emoticons. The set number can be set according to the actual situation, for example, the set number can be 5 .
  • Emoticons can include likes, hearts, and various emotional expressions, etc., and the details are not limited.
  • the interactive component can be displayed to the user, and the interactive component is displayed with emoticon components and/or comment components, and the interaction components selected by the user in the emoticon components can be obtained.
  • Emoticons can also acquire the comments input by the user in the comment component, and display the interactive emoticons and/or comments on the second multimedia interface, and the specific display position is not limited.
  • the interactive expressions and/or comments can also be displayed on the first multimedia interface, and the specific display position is not limited.
  • the multimedia processing method may further include: determining the interaction time point corresponding to the interactive input trigger operation; on the playing time axis of the target audio on the second multimedia interface and/or the first multimedia interface,
  • the interactive prompt logo is displayed at the position of the time point.
  • the interaction time point refers to the corresponding time point in the target audio when the user performs an interactive input trigger operation.
  • the interactive prompt logo refers to the prompt logo used to remind the user of the existence of interactive content.
  • the interactive prompt logo corresponding to different interactive content can be different.
  • the interactive prompt logo corresponding to the emoticon can be the expression itself, and the interactive prompt logo corresponding to the comment can be set. Dialog ID.
  • the real-time moment of the interactive input trigger operation can be determined, and the playback time point of the target audio at the real-time moment can be determined as the interaction time point, and then can be displayed on the second multimedia interface and/or the first
  • an interactive prompting sign corresponding to the interactive content is displayed to remind the user that there is interactive content here.
  • the corresponding interactive prompt signs can be superimposed and displayed.
  • the prompt logo of the interactive content can be displayed on the playback time axis of the two multimedia interfaces, so that the content displayed on the two multimedia interfaces is synchronized, prompting other users that there is interactive content here, so that
  • the user-to-interaction is not limited to itself, and the interaction methods are more diversified, which further improves the user's interactive experience.
  • the multimedia processing method may further include: receiving a modification operation on the target subtitle displayed on the first multimedia interface; and synchronously modifying the target subtitle displayed on the second multimedia interface.
  • the modified target subtitles can be displayed in the first multimedia interface; 2.
  • the target subtitle displayed in the multimedia interface is modified synchronously, and the modified target subtitle is displayed.
  • both the first multimedia interface and the second multimedia interface are interfaces of the first application program
  • the multimedia processing method may further include: receiving a program switching request; switching the first application program to the background to run and Start the second application program, display the display interface of the second application program; display the floating window component of the second multimedia interface in the display interface of the second application program.
  • the second application program may be any program different from the first application program.
  • the first application program and the second application program are two different functional modules of the same application program, or two different application programs.
  • the floating window component can be an entry component for quickly returning to the second multimedia interface in the first application program, that is, the first application program can be quickly switched from running in the background to running in the foreground through the floating window component.
  • the specific expression form of the floating window component is not limited, for example, the floating window component may be a circular or square display window.
  • the program switching request can be received, and then the first application program can be switched to run in the background and the second application program can be started to display the display interface of the second application program, and the display interface of the second application program can be In addition to displaying relevant content of the second application program, a floating window component of the second multimedia interface may also be displayed.
  • the floating window component can be suspended on the top layer of the display interface of the second application program, so that the user can trigger the floating window component when operating the current display interface.
  • the specific position of the floating window component in the display interface of the second application program can be set according to actual conditions, for example, the floating window component can be displayed at any position where the current display content is blocked by cloth.
  • the floating window component includes a cover picture and/or playing information of the target audio.
  • the playback information includes playback progress; the cover image is displayed in association with the playback progress.
  • the playback progress is displayed around the cover image.
  • the cover picture is determined based on the first content.
  • the floating window component may include information related to the target audio, for example, the floating window component may include a cover image and/or playing information, and the playing information may include playing progress, playing time, and the like.
  • the cover picture may be determined according to the first content included in the first multimedia interface. For example, when the first content includes a video corresponding to the target audio, an image may be captured from the video as the cover picture.
  • the cover picture can also be displayed in association with the playback progress. When there are multiple cover pictures in specific display, the cover picture can be changed with the change of the play progress, that is, with the change of the play progress, the display corresponding to the current play progress will be displayed in real time. cover image. In addition, the playback progress can also be displayed around the cover image, just for example.
  • the user can also understand the playback status of the audio when operating other applications, further improving the audio playback effect.
  • the multimedia processing method may further include: receiving a trigger operation on the floating window component, switching the first application program from running in the background to running in the foreground, and returning to displaying the second multimedia interface.
  • the current second application program can be switched to run in the background, and the first application program can be run from the background to the foreground to display the second multimedia interface.
  • the multimedia processing method may further include: continuing to play the target audio based on the floating window component. If the target audio is being played, after receiving the program switching request and displaying the floating window component in the display interface of the second application program, the target audio can be played continuously based on the floating window component, so as to realize the seamless playback of the audio.
  • the floating window component of the multimedia interface can also be displayed, through which the floating window component can quickly return to the multimedia interface, and continue to play audio data and display the playback status. Improve the efficiency of returning to the multimedia interface, better meet the needs of users, and improve the user experience.
  • the multimedia processing method may further include: switching the currently displayed third multimedia interface to the second multimedia interface in response to the user's interface switching request in the currently displayed third multimedia interface ;
  • the third multimedia interface includes third content
  • the third content includes attribute information of the second content.
  • the attribute information of the second content includes at least one of title information, time information and source information of the second content.
  • the third multimedia interface is a multimedia interface different from the display content of the above-mentioned first multimedia interface and the second multimedia interface.
  • the third content included in the third multimedia interface has an association relationship with the second content, and may include attribute information of the second content, and the attribute information of the second content may be determined according to actual conditions, for example, may include title information of the second content , time information, source information and other information.
  • the third multimedia interface may be an interface including an information list, and the information list includes attribute information of multiple audios, one of which is the attribute information of the target audio.
  • the terminal displays the third multimedia interface. After receiving the user's interface switching request in the third multimedia interface, it can close the currently displayed third multimedia interface and open the second multimedia interface.
  • the second content is displayed on the second multimedia interface to realize switching of the interface.
  • both can be switched to a multimedia interface including only audio and subtitles, which further improves the switching flexibility of multimedia interfaces in different modes, and improves the user interface. switching efficiency.
  • FIG. 2 is a schematic flowchart of another multimedia processing method provided by an embodiment of the present disclosure. This embodiment further optimizes the foregoing multimedia processing method on the basis of the foregoing embodiments. As shown in Figure 2, the method includes:
  • Step 201 displaying a first multimedia interface or a third multimedia interface.
  • the first multimedia interface includes the first content
  • the third multimedia interface includes the third content
  • Step 202 Receive an interface switching request from the user on the first multimedia interface or the third multimedia interface.
  • Step 203 Switch the currently displayed first multimedia interface or the third multimedia interface to the second multimedia interface, and display the second content on the second multimedia interface.
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio.
  • the third content includes attribute information of the second content, and the attribute information of the second content includes at least one of title information, time information, and source information of the second content.
  • FIG. 3 is a schematic diagram of a multimedia interface provided by an embodiment of the present disclosure.
  • a schematic diagram of a second multimedia interface is shown, and an audio and The corresponding subtitles, the timeline of the audio and the positions of multiple control buttons acting on the audio are just examples in Figure 3, which also shows the audio cover picture and the name "Team Review Meeting”.
  • step 203 After step 203, step 204-step 206, step 207-step 211 and/or step 212-step 216 may be executed, and the specific order of execution is not limited, and FIG. 2 is only an example.
  • Step 204 receiving a trigger operation for playing the target audio.
  • Step 205 Play the target audio, and during the playback of the target audio, highlight and display the subtitle sentences corresponding to the playback progress of the target audio based on the timestamps of the subtitle sentences included in the target subtitles.
  • the multimedia processing method further includes: determining a non-silent segment in the target audio; and playing the target audio, including: playing only the non-silent segment when playing the target audio.
  • the multimedia processing method also includes: determining the mute segment and the non-mute segment in the target audio; playing the target audio, including: playing the mute segment at a first playback speed, and playing the non-mute segment at a second playback speed, wherein, the first A playback speed is greater than a second playback speed.
  • Step 206 in response to the end of playing the target audio, acquire the next audio associated with the target audio, and switch to play the next audio.
  • Step 207 receiving an interaction trigger operation of the user on the second multimedia interface.
  • Step 208 in response to the interaction trigger operation, display the interactive component on the second multimedia interface, and acquire the interactive content based on the interactive component.
  • the interactive components include emoticon components and/or comment components
  • the interactive content includes interactive emoticons and/or comments.
  • FIG. 4 is a schematic diagram of another multimedia interface provided by an embodiment of the present disclosure.
  • FIG. 4 another schematic diagram of a second multimedia interface is shown in the figure.
  • FIG. 4 Part of the content and the location of the buttons are different.
  • the underlined subtitle sentence in the target subtitle in FIG. 4 represents the subtitle sentence corresponding to the playback progress of the target audio.
  • the target video is played, other subtitle sentences will also be highlighted by underlining.
  • the figure shows an example of the emoticon part 11 in the interactive component.
  • the default interactive emoticon can be sent and displayed in the second multimedia interface, such as "like" in the middle of the interface. .
  • an emoticon panel may also be displayed, and the emoticon panel may include multiple emoticons for the user to choose (not shown in the figure).
  • an exit button is shown below the second multimedia interface in FIG. 4 , and when the user triggers the exit button, the user can exit from the second multimedia interface to the first multimedia interface.
  • the second multimedia interface shown in FIG. 3 and FIG. 4 are examples and should not be used as limitations.
  • Step 209 display the interactive content on the second multimedia interface and/or the first multimedia interface.
  • Step 210 Determine the interaction time point corresponding to the interaction input trigger operation.
  • Step 211 on the playing time axis of the target audio on the second multimedia interface and/or the first multimedia interface, an interaction prompt logo is displayed at the position of the interaction time point.
  • Step 212 receiving a program switching request.
  • Step 213 switch the first application program to run in the background and start the second application program, and display the display interface of the second application program.
  • both the first multimedia interface and the second multimedia interface are interfaces of the first application program.
  • Step 214 display the floating window component of the second multimedia interface in the display interface of the second application program.
  • the floating window component includes a cover picture and/or playing information of the target audio.
  • the playback information includes playback progress; the cover image is displayed in association with the playback progress.
  • the playback progress is displayed around the cover image.
  • the cover picture is determined based on the first content.
  • FIG. 5 is a schematic diagram of a floating window component provided by an embodiment of the present disclosure.
  • the floating window component 12 under other application programs is shown in the figure, and the other application programs are the second application program
  • the floating window component 12 may be set in an area close to the boundary of the display interface of other application programs.
  • the floating window component 12 also shows the cover picture and the playing progress of the second multimedia interface.
  • the black filled area on the edge of the floating window component in the figure represents the playing progress, and the playing progress in the figure is close to two-thirds.
  • the floating window assembly shown in FIG. 5 is only an example, and floating window assemblies of other shapes or styles are also applicable.
  • step 214 step 215 and/or step 216 may be performed.
  • Step 215 Receive a trigger operation on the floating window component, switch the first application program from running in the background to running in the foreground, and return to display the second multimedia interface.
  • Step 216 continue to play the target audio based on the floating window component.
  • step 210 if the target audio is being played normally, the target audio can be played continuously based on the floating window component.
  • the multimedia processing method may further include: receiving a modification operation on the target subtitle displayed on the first multimedia interface; and synchronously modifying the target subtitle displayed on the second multimedia interface.
  • the multimedia processing solution provided by the embodiment of the present disclosure displays the first multimedia interface, and the first multimedia interface includes the first content; receives the user's interface switching request in the first multimedia interface; and displays the currently displayed first multimedia interface.
  • a multimedia interface is switched to a second multimedia interface, and the second content is displayed in the second multimedia interface; wherein, the first content includes the second content and other content associated with the second content, and the second content Including the target audio and the target subtitle corresponding to the target audio.
  • FIG. 6 is a schematic structural diagram of a multimedia processing device provided by an embodiment of the present disclosure.
  • the device may be implemented by software and/or hardware, and may generally be integrated into an electronic device. As shown in Figure 6, the device includes:
  • the first interface module 301 is configured to display a first multimedia interface, where the first multimedia interface includes first content;
  • a request module 302 configured to receive an interface switching request from a user in the first multimedia interface
  • the second interface module 303 is configured to switch the currently displayed first multimedia interface to a second multimedia interface, and display second content in the second multimedia interface;
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio
  • the device also includes a playback module for:
  • the playing module is specifically used for:
  • the device also includes a first audio recognition module, configured to:
  • the playback module is specifically used for:
  • the device also includes a second audio recognition module, configured to:
  • the playback module is specifically used for:
  • the device also includes an interaction module, configured to:
  • the interaction module is used for:
  • the interactive components include emoticon components and/or comment components
  • the interactive content includes interactive emoticons and/or comments.
  • the interaction module is used for:
  • the interaction module is used for:
  • an interaction prompt logo is displayed at the position of the interaction time point.
  • the device also includes a modification module for:
  • both the first multimedia interface and the second multimedia interface are interfaces of a first application program
  • the device further includes a floating window module, configured to:
  • the floating window component of the second multimedia interface is displayed in the display interface of the second application program.
  • the floating window component includes a cover picture and/or playing information of the target audio.
  • the playing information includes playing progress; the cover image is displayed in association with the playing progress.
  • cover pictures there are multiple cover pictures, and the cover pictures change as the playing progress changes.
  • the playback progress is displayed around the cover image.
  • the cover picture is determined based on the first content.
  • the device also includes a return module for:
  • the floating window module is also used for:
  • the device also includes a third interface module, configured to:
  • the third multimedia interface includes third content, and the third content includes attribute information of the second content.
  • the attribute information of the second content includes at least one of title information, time information and source information of the second content.
  • the multimedia processing device provided by the embodiment of the present disclosure can execute the multimedia processing method provided by any embodiment of the present disclosure, and has corresponding functional modules and beneficial effects for executing the method.
  • An embodiment of the present disclosure further provides a computer program product, including a computer program/instruction, and when the computer program/instruction is executed by a processor, the multimedia processing method provided in any embodiment of the present disclosure is implemented.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure. Referring to FIG. 7 in detail below, it shows a schematic structural diagram of an electronic device 400 suitable for implementing an embodiment of the present disclosure.
  • the electronic device 400 in the embodiment of the present disclosure may include, but is not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (Personal Digital Assistants), PADs (Tablet Computers), PMPs (Portable Multimedia Players), vehicle-mounted terminals ( Mobile terminals such as car navigation terminals) and stationary terminals such as digital TVs, desktop computers and the like.
  • the electronic device shown in FIG. 7 is only an example, and should not limit the functions and application scope of the embodiments of the present disclosure.
  • an electronic device 400 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) 401, which may be randomly accessed according to a program stored in a read-only memory (ROM) 402 or loaded from a storage device 408.
  • ROM read-only memory
  • RAM random access memory
  • various appropriate actions and processes are executed by programs in the memory (RAM) 403 .
  • RAM random access memory
  • various programs and data necessary for the operation of the electronic device 400 are also stored.
  • the processing device 401, the ROM 402, and the RAM 403 are connected to each other through a bus 404.
  • An input/output (I/O) interface 405 is also connected to bus 404 .
  • the following devices can be connected to the I/O interface 405: input devices 406 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speaker, vibration an output device 407 such as a computer; a storage device 408 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 409.
  • the communication means 409 may allow the electronic device 400 to perform wireless or wired communication with other devices to exchange data. While FIG. 7 shows electronic device 400 having various means, it should be understood that implementing or having all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product, which includes a computer program carried on a non-transitory computer readable medium, where the computer program includes program code for executing the method shown in the flowchart.
  • the computer program may be downloaded and installed from a network via communication means 409, or from storage means 408, or from ROM 402.
  • the processing device 401 When the computer program is executed by the processing device 401, the above functions defined in the multimedia processing method of the embodiment of the present disclosure are executed.
  • the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples of computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave carrying computer-readable program code therein. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • Program code embodied on a computer readable medium may be transmitted by any appropriate medium, including but not limited to wires, optical cables, RF (radio frequency), etc., or any suitable combination of the above.
  • the client and the server can communicate using any currently known or future network protocols such as HTTP (HyperText Transfer Protocol, Hypertext Transfer Protocol), and can communicate with digital data in any form or medium
  • HTTP HyperText Transfer Protocol
  • the communication eg, communication network
  • Examples of communication networks include local area networks (“LANs”), wide area networks (“WANs”), internetworks (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed network of.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: displays a first multimedia interface, and the first multimedia interface includes The first content; receiving the user's interface switching request in the first multimedia interface; switching the currently displayed first multimedia interface to a second multimedia interface, and displaying the first multimedia interface on the second multimedia interface
  • the second content is displayed in ; wherein, the first content includes the second content and other content associated with the second content, and the second content includes target audio and target subtitles corresponding to the target audio.
  • Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, or combinations thereof, including but not limited to object-oriented programming languages—such as Java, Smalltalk, C++, and Includes conventional procedural programming languages - such as the "C" language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer can be connected to the user computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (such as through an Internet service provider). Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider such as AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented by software or by hardware. Wherein, the name of a unit does not constitute a limitation of the unit itself under certain circumstances.
  • FPGAs Field Programmable Gate Arrays
  • ASICs Application Specific Integrated Circuits
  • ASSPs Application Specific Standard Products
  • SOCs System on Chips
  • CPLD Complex Programmable Logical device
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • the present disclosure provides a multimedia processing method, including:
  • the first multimedia interface includes first content
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio
  • the method further includes:
  • the method further includes:
  • the method further includes:
  • Said playing said target audio includes:
  • the method further includes: determining a mute segment and a non-mute segment in the target audio;
  • Said playing said target audio includes:
  • the method further includes:
  • the determining the interaction content based on the interaction triggering operation includes:
  • the interactive components include emoticon components and/or comment components
  • the interactive content includes interactive emoticons and/or comments.
  • the method further includes:
  • the method further includes:
  • an interaction prompt logo is displayed at the position of the interaction time point.
  • the method further includes:
  • both the first multimedia interface and the second multimedia interface are interfaces of the first application program, and the method also includes include:
  • the floating window component of the second multimedia interface is displayed in the display interface of the second application program.
  • the floating window component includes a cover picture and/or playing information of the target audio.
  • the playing information includes playing progress; the cover image is displayed in association with the playing progress.
  • the multimedia processing method provided in the present disclosure there are multiple cover pictures, and the cover pictures change as the playing progress changes.
  • the playback progress is displayed around the cover image.
  • the cover picture is determined based on the first content.
  • the method further includes:
  • the method further includes:
  • the method further includes:
  • the third multimedia interface includes third content, and the third content includes attribute information of the second content.
  • the attribute information of the second content includes at least one of title information, time information and source information of the second content.
  • the present disclosure provides a multimedia processing device, including:
  • the first interface module is configured to display a first multimedia interface, and the first multimedia interface includes first content
  • a request module configured to receive an interface switching request from a user in the first multimedia interface
  • the second interface module is configured to switch the currently displayed first multimedia interface to a second multimedia interface, and display second content in the second multimedia interface;
  • the first content includes the second content and other content associated with the second content
  • the second content includes target audio and target subtitles corresponding to the target audio
  • the device further includes a playback module, configured to:
  • the playing module is specifically configured to:
  • the device further includes a first audio recognition module, configured to:
  • the playback module is specifically used for:
  • the device also includes a second audio recognition module, configured to:
  • the playback module is specifically used for:
  • the device further includes an interaction module, configured to:
  • the interaction module is used for:
  • the interactive components include emoticon components and/or comment components
  • the interactive content includes interactive emoticons and/or comments.
  • the interaction module is used for:
  • the interaction module is used for:
  • an interaction prompt logo is displayed at the position of the interaction time point.
  • the device further includes a modification module, configured to:
  • both the first multimedia interface and the second multimedia interface are interfaces of a first application program, and the device further Includes floating window module for:
  • the floating window component of the second multimedia interface is displayed in the display interface of the second application program.
  • the floating window component includes a cover picture and/or playing information of the target audio.
  • the playing information includes playing progress
  • the cover picture is displayed in association with the playback progress.
  • the multimedia processing device provided in the present disclosure there are multiple cover pictures, and the cover pictures change as the playing progress changes.
  • the playback progress is displayed around the cover image.
  • the cover picture is determined based on the first content.
  • the device further includes a return module, configured to:
  • the floating window module is further configured to:
  • the device further includes a third interface module, configured to:
  • the third multimedia interface includes third content, and the third content includes attribute information of the second content.
  • the attribute information of the second content includes at least one of title information, time information and source information of the second content.
  • the present disclosure provides an electronic device, including:
  • the processor is configured to read the executable instructions from the memory, and execute the instructions to implement any one of the multimedia processing methods provided in the present disclosure.
  • the present disclosure provides a computer-readable storage medium, the storage medium stores a computer program, and the computer program is used to execute any one of the multimedia programs provided by the present disclosure. Approach.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本公开实施例涉及一种多媒体处理方法、装置、设备及介质,其中该方法包括:展示第一多媒体界面,第一多媒体界面中包括第一内容;接收用户在第一多媒体界面中的界面切换请求;将当前展示的第一多媒体界面切换为第二多媒体界面,并在第二多媒体界面中展示第二内容;其中,第一内容包括第二内容以及与第二内容关联的其他内容,第二内容包括目标音频以及目标音频对应的目标字幕。采用上述技术方案,可以实现包括两种不同内容的界面的切换,其中一个界面可以仅包括音频和字幕,有助于用户在复杂场景下专注于多媒体内容,并且提升了多媒体内容播放的灵活性,可以满足多种场景的需求,进而提升了用户的体验效果。

Description

一种多媒体处理方法、装置、设备及介质
本申请要求于2021年5月19日递交、申请号为202110547916.9、名称为“一种多媒体处理方法、装置、设备及介质”的中国专利申请的优先权,其全部内容通过引用并入本文。
技术领域
本公开涉及多媒体技术领域,尤其涉及一种多媒体处理方法、装置、设备及介质。
背景技术
随着智能设备和多媒体技术的不断发展,通过智能设备进行信息记录越来越多地被应用在日常生活和办公生活中。
在一些相关产品中,可以对信息记录的多媒体文件进行回放以供再次回顾。目前回放多媒体文件的方式较为固定,并且方式单一,灵活性不高。
发明内容
为了解决上述技术问题或者至少部分地解决上述技术问题,本公开提供了一种多媒体处理方法、装置、设备及介质。
本公开实施例提供了一种多媒体处理方法,所述方法包括:
展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
接收用户在第一多媒体界面中的界面切换请求;
将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
本公开实施例还提供了一种多媒体处理装置,所述装置包括:
第一界面模块,用于展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
请求模块,用于接收用户在第一多媒体界面中的界面切换请求;
第二界面模块,用于将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在 所述第二多媒体界面中展示第二内容;
其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
本公开实施例还提供了一种电子设备,所述电子设备包括:处理器;用于存储所述处理器可执行指令的存储器;所述处理器,用于从所述存储器中读取所述可执行指令,并执行所述指令以实现如本公开实施例提供的多媒体处理方法。
本公开实施例还提供了一种计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序用于执行如本公开实施例提供的多媒体处理方法。
本公开实施例还提供了一种计算机程序,包括指令,所述指令当由处理器执行时使所述处理器执行如本公开实施例提供的多媒体处理方法。
本公开实施例还提供了一种计算机程序产品,包括指令,所述指令当由处理器执行时使所述处理器执行如本公开实施例提供的多媒体处理方法。
附图说明
结合附图并参考以下具体实施方式,本公开各实施例的上述和其他特征、优点及方面将变得更加明显。贯穿附图中,相同或相似的附图标记表示相同或相似的元素。应当理解附图是示意性的,原件和元素不一定按照比例绘制。
图1为本公开实施例提供的一种多媒体处理方法的流程示意图;
图2为本公开实施例提供的另一种多媒体处理方法的流程示意图;
图3为本公开实施例提供的一种多媒体界面的示意图;
图4为本公开实施例提供的另一种多媒体界面的示意图;
图5为本公开实施例提供的一种浮窗组件的示意图;
图6为本公开实施例提供的一种多媒体处理装置的结构示意图;
图7为本公开实施例提供的一种电子设备的结构示意图。
具体实施方式
下面将参照附图更详细地描述本公开的实施例。虽然附图中显示了本公开的某些实施例,然而应当理解的是,本公开可以通过各种形式来实现,而且不应该被解释为限于这里阐述的实施例,相反提供这些实施例是为了更加透彻和完整地理解本公开。应当理解的是,本公开的附图及实施例仅用于示例性作用,并非用于限制本公开的保护范围。
应当理解,本公开的方法实施方式中记载的各个步骤可以按照不同的顺序执行,和/或并行执行。此外,方法实施方式可以包括附加的步骤和/或省略执行示出的步骤。本公开的范围在此方面不受限制。
本文使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”;术语“另一实施例”表示“至少一个另外的实施例”;术语“一些实施例”表示“至少一些实施例”。其他术语的相关定义将在下文描述中给出。
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对不同的装置、模块或单元进行区分,并非用于限定这些装置、模块或单元所执行的功能的顺序或者相互依存关系。
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的,而并不是用于对这些消息或信息的范围进行限制。
本公开实施例提供的技术方案与相关技术相比具有如下优点:本公开实施例提供的多媒体处理方案,展示第一多媒体界面,第一多媒体界面中包括第一内容;接收用户在第一多媒体界面中的界面切换请求;将当前展示的第一多媒体界面切换为第二多媒体界面,并在第二多媒体界面中展示第二内容;其中,第一内容包括第二内容以及与第二内容关联的其他内容,第二内容包括目标音频以及目标音频对应的目标字幕。采用上述技术方案,可以实现包括两种不同内容的界面的切换,其中一个界面可以仅包括音频和字幕,有助于用户在复杂场景下专注于多媒体内容,并且提升了多媒体内容播放的灵活性,可以满足多种场景的需求,进而提升了用户的体验效果。
图1为本公开实施例提供的一种多媒体处理方法的流程示意图,该方法可以由多媒体处理装置执行,其中该装置可以采用软件和/或硬件实现,一般可集成在电子设备中。如图1所示,该方法包括:
步骤101、展示第一多媒体界面,第一多媒体界面中包括第一内容。
其中,多媒体界面是指用于对多种类型的多媒体信息进行展示的界面,多媒体信息可以包括音频、视频和文本等,具体不限。第一多媒体界面是指多媒体界面中的一个,第一内容是指在第一多媒体界面中展示的内容,可以包括多种多媒体信息,例如在会议录制的 场景下,第一内容可以包括录制绘制的音频和/或视频、对应的字幕内容以及会议纪要等与会议相关的内容。
本公开实施例中,客户端根据用户的请求,可以获取第一内容,并展示第一多媒体界面以及在第一多媒体界面中展示第一内容。由于第一内容中可以包括多种类型的信息,第一多媒体界面中可以针对各类型的信息设置不同的展示区域进行展示,例如第一多媒体界面中可以设置音视频区域、字幕区域以及纪要展示区域等区域,分别用于展示音频、视频、字幕内容以及纪要等。
步骤102、接收用户在第一多媒体界面中的界面切换请求。
其中,界面切换请求是指用于进行不同界面之间切换的请求。
本公开实施例中,展示第一多媒体界面之后,可以检测用户在第一媒体界面上的触发操作,当检测到用户对界面切换按键的触发操作之后,可以确定接收到界面切换请求。其中,界面切换按键可以为预先设置在第一多媒体界面中的虚拟按键,具体位置和表现形式不限。
步骤103、将当前展示的第一多媒体界面切换为第二多媒体界面,并在第二多媒体界面中展示第二内容。
其中,第二多媒体界面是与上述第一多媒体界面展示内容不同的一个多媒体界面。上述第一多媒体界面中展示的第一内容可以包括第二内容以及与第二内容关联的其他内容,也即第二内容可以为上述第一多媒体界面的第一内容的一部分。第二内容可以包括目标音频以及目标音频对应的目标字幕,目标音频可以为任意一个记录信息的音频数据,例如目标音频可以为记录会议过程中的音频数据,目标字幕是指对目标音频采用语音识别(Automatic Speech Recognition,ASR)技术识别和处理之后得到的文本内容,本公开实施例中对具体的语音识别技术不作限定,例如可以采用随机模型法或人工神经网络的方法等。
本公开实施例中,接收到界面切换请求之后,可以将当前展示的第一多媒体界面关闭并打开第二多媒体界面,在第二多媒体界面中展示第二内容,实现界面的切换。由于第二多媒体界面中仅包括音频和字幕,有助于用户在复杂场景下专注于多媒体内容。
可以理解的是,在第二多媒体界面中展示第二内容之后,还可以基于用户对第二多媒体界面中退出按键的触发操作,返回第一多媒体界面中,实现两种不同模式的多媒体界面 的灵活切换,用户可以根据实际需求进行切换。在返回展示第一多媒体界面后,可以展示第二多媒体界面的浮窗组件,以供用户快速切换回第二多媒体界面。在第二多媒体界面的浮窗组件中,可以继续播放目标音频。在用户对第二多媒体界面的浮窗组件进行触发操作后,可以返回展示第二多媒体界面。
本公开实施例提供的多媒体处理方案,展示第一多媒体界面,第一多媒体界面中包括第一内容;接收用户在第一多媒体界面中的界面切换请求;将当前展示的第一多媒体界面切换为第二多媒体界面,并在第二多媒体界面中展示第二内容;其中,第一内容包括第二内容以及与第二内容关联的其他内容,第二内容包括目标音频以及目标音频对应的目标字幕。采用上述技术方案,可以实现包括两种不同内容的界面的切换,其中一个界面可以仅包括音频和字幕,有助于用户在复杂场景下专注于多媒体内容。此外,以多种形式(例如,在第一多媒体界面播放和在第二多媒体界面播放)来播放多媒体内容,提升了多媒体内容播放的灵活性,可以满足多种场景的需求,进而提升了用户的体验效果。
在一些实施例中,多媒体处理方法还可以包括:接收对目标音频的播放触发操作;播放目标音频,并在目标音频的播放过程中,基于目标字幕中包括的字幕语句的时间戳,对与目标音频的播放进度对应的字幕语句进行突出展示。
其中,播放触发操作是指用于播放多媒体的触发操作,播放触发操作的具体形式可以为多种,具体不限。目标字幕属于结构化文本,包括段、句和词三层结构,字幕语句是目标字幕中的句,一个字幕语句可以包括至少一个字或词。由于目标字幕是通过对目标音频进行语音识别得到的,每个字幕语句均具有对应的语音语句,每个语音语句对应于目标音频中的一个时间戳。在对目标音频进行语音识别之后可以得到目标字幕,对目标字幕中的每个字幕语句可以确定其在目标音频中对应的语音语句,由于每个语音语句对应于目标音频的一个播放时间,可以根据语音语句和目标音频的播放时间之间的对应关系,确定目标字幕中每个字幕语句的时间戳。突出展示的具体方式本公开实施例不作限定,例如突出展示可以为采用高亮、加粗、增大显示字号、更改显示字体和添加下划线等方式中的一种或多种进行展示。
具体的,当接收到用户对目标音频的播放触发操作之后,可以播放目标音频,并在播放目标音频的过程中,根据目标字幕中包括的各字幕语句的时间戳,依次对与播放进度对应的字幕语句进行突出展示,也即随着目标音频的播放,目标字幕中的字幕语句随着播放 的进行,依次进行突出展示。
上述方案中,当音频播放的过程中对应的字幕语句也可以关联突出展示,可以实现多媒体和字幕之间的关联互动,以使用户更好地了解多媒体内容,提高了用户的体验效果。
在一些实施例中,多媒体处理方法还可以包括:响应于目标音频播放结束,获取与目标音频关联的下一音频,切换播放下一音频。在这里,目标音频的播放结束可以基于用户的操作确定,或者,目标音频的播放结束也可以基于其播放进度达到播放完成时刻而确定。
其中,下一音频是指预先设置的与目标音频的属性信息关联的一个音频,属性信息不限,例如属性信息可以为时间、用户或其他关键信息等。例如当目标音频为一个会议的记录音频,则下一音频可以为与该会议的结束时刻相邻的后一个会议音频。具体的,在目标视频的播放结束之后,可以确定与目标音频关联的下一音频,获取该下一音频并播放该下一音频。或者,下一音频还可以是基于一项或多项属性信息的属性值确定出的播放列表中的下一个音频。例如,若属性信息包括会议日期,那么,用户可以基于会议日期来确定一播放列表。从而,在该播放列表中的目标音频播放结束之后,继续播放下一音频。这样设置的好处在于,当前音频播放结束之后可以无缝播放下一个音频,使用户了解更多地相关内容,并且避免停止造成的突兀感,提升多媒体信息回放体验。
在一些实施例中,多媒体处理方法还可以包括:确定目标音频中的非静音片段;播放目标音频,包括:在播放目标音频时仅播放非静音片段。在一些实施例中,多媒体处理方法还可以包括:确定目标音频中的静音片段和非静音片段;播放目标音频,包括:以第一播放速度播放静音片段,以第二播放速度播放非静音片段,其中,第一播放速度大于第二播放速度。
其中,静音片段是指目标音频中音量为零的音频片段,非静音片段是指目标音频中音量不为零的音频片段。具体的,通过对目标音频的音量识别,可以确定其中的非静音片段,并在播放目标音频时仅播放非静音片段。可选的,还可以通过音量识别确定目标音频的静音片段和非静音片段,并采用第一播放速度播放静音片段,采用第二播放速度播放非静音片段。第一播放速度和第二播放速度可以根据实际情况确定,只要满足第一播放速度大于第二播放速度即可,例如第一播放速度可以设置为第二播放速度的两倍。
上述方案中,播放音频时可以跳过静音片段,只播放重点内容,;也可以通过两种不同的速度播放静音片段和非静音片段,上述两种方式均可以提升用户了解音频内容的速 度,提升音频播放的灵活性。
在一些实施例中,多媒体处理方法还可以包括:接收用户在第二多媒体界面上的交互触发操作;基于交互触发操作,确定交互内容。可选的,基于交互触发操作,确定交互内容,包括:响应于交互触发操作,在第二多媒体界面展示交互组件;基于交互组件获取交互内容,并在第二多媒体界面展示交互内容;其中,交互组件包括表情部件和/或评论部件,交互内容包括交互表情和/或评论。可选的,多媒体处理方法还可以包括:在第一多媒体界面上展示交互内容。
其中,交互输入触发操作是指用户想要对当前的多媒体内容进行交互输入的触发操作,本公开实施例中交互输入触发操作可以包括在第二多媒体界面上目标音频的播放时间轴上或在交互按钮上的触发操作,交互按钮可以为预先设置在第二多媒体界面上的按钮,按钮的具体位置和样式不作限定。交互组件是指用于进行交互内容输入、编辑和发布等操作的功能组件。交互组件可以包括表情部件和/或评论部件,表情部件为用于进行表情输入的功能部件,可以包括设定数量的表情,设定数量可以根据实际情况进行设定,例如设定数量可以为5。表情可以包括点赞、爱心和多种情绪表情等,具体不限。
具体的,当接收到用户在第二多媒体界面上的交互触发操作之后,可以展示交互组件给用户,交互组件中展示有表情部件和/或评论部件,可以获取表情部件中用户选择的交互表情,也可以获取评论部件中用户输入的评论,将交互表情和/或评论展示在第二多媒体界面上,具体展示位置不限。可选的,还可以将交互表情和/或评论展示在第一多媒体界面上,具体展示位置不限。
上述方案中,在多媒体界面上展示多媒体内容的同时,还支持用户的交互,并可以在第一多媒体界面和/或第二多媒体界面中展示交互内容,提升了用户的参与体验效果。
在一些实施例中,多媒体处理方法还可以包括:确定交互输入触发操作对应的交互时间点;在第二多媒体界面和/或第一多媒体界面上目标音频的播放时间轴上,交互时间点所在位置展示交互提示标识。其中,交互时间点是指用户进行交互输入触发操作时在目标音频中对应的时间点。交互提示标识是指用于提醒用户存在交互内容的提示标识,不同交互内容对应的交互提示标识可以不同,例如表情对应的交互提示标识可以为表情本身,评论对应的交互提示标识可以为设定的对话框标识。
接收到用户的交互输入触发操作之后,可以确定交互输入触发操作的实时时刻,将目 标音频在实时时刻的播放时间点确定为交互时间点,之后可以在第二多媒体界面和/或第一多媒体界面上目标音频的播放时间轴上展示交互内容对应的交互提示标识,以提示用户此处具有交互内容。当播放时间轴上一个时间点包括多个交互内容时,可以将对应的交互提示标识叠加展示。
上述方案中,在用户输入交互内容之后,可以在两个多媒体界面的播放时间轴上均展示交互内容的提示标识,使得两个多媒体界面展示的内容同步,提示其他用户此处具有交互内容,使得用户对交互不仅局限于自身,交互方式更加多样化,进一步提高了用户的交互体验。
在一些实施例中,多媒体处理方法还可以包括:接收对第一多媒体界面中展示的目标字幕的修改操作;对第二多媒体界面中展示的目标字幕进行同步修改。当用户在第一多媒体界面中对展示的目标字幕中的字、词或句中的至少一个进行修改之后,可以在第一多媒体界面中展示修改后的目标字幕;并且同时对第二多媒体界面中展示的目标字幕进行同步修改,展示修改后的目标字幕。上述方案中,针对两个多媒体界面中展示的内容,当一个多媒体界面中的内容修改之后,另一个多媒体界面的内容也可以同步修改,避免用户在不同多媒体界面查看相同内容时存在错误。
在一些实施例中,第一多媒体界面和第二多媒体界面均为第一应用程序的界面,多媒体处理方法还可以包括:接收程序切换请求;将第一应用程序切换至后台运行并启动第二应用程序,展示第二应用程序的展示界面;在第二应用程序的展示界面中展示第二多媒体界面的浮窗组件。
其中,第二应用程序可以为任意一个与第一应用程序不同的程序。例如,第一应用程序和第二应用程序分别为同一应用程序的两个不同功能模块,或两个不同的应用程序。浮窗组件可以为用于快速返回第一应用程序中第二多媒体界面的入口组件,也即通过浮窗组件可以将第一应用程序快速从后台运行切换至前台运行。并且浮窗组件的具体表现形式不限,例如浮窗组件可以为圆形或方形的展示小窗。
具体的,基于用户的触发操作可以接收程序切换请求,之后可以将第一应用程序切换到后台运行并启动第二应用程序,展示第二应用程序的展示界面,并且第二应用程序的展示界面可以除了展示第二应用程序的相关内容还可以展示第二多媒体界面的浮窗组件。该浮窗组件可以悬浮于第二应用程序的展示界面的最上层,以供用户在对当前显示界面操作 时可以对该浮窗组件进行触发操作。浮窗组件在第二应用程序的展示界面中的具体位置可以根据实际情况设置,例如浮窗组件可以展示在任何布遮挡当前显示内容的位置。
可选的,浮窗组件包括目标音频的封面图片和/或播放信息。可选的,播放信息包括播放进度;封面图片和播放进度关联展示。可选的,封面图片的数量为多个,封面图片随着播放进度的变化而变化。可选的,播放进度环绕封面图片显示。可选的,封面图片基于第一内容确定。
浮窗组件中可以包括目标音频的相关信息,例如浮窗组件可以包括封面图片和/或播放信息,播放信息可以包括播放进度、播放时间点等。封面图片可以根据第一多媒体界面中包括的第一内容确定,例如当第一内容中包括目标音频对应的视频时,可以通过在视频中截取图像作为封面图片。封面图片还可以与播放进度关联展示,具体展示时当封面图片为多个时,可以将封面图片随着播放进度的变换而变换,也即随着播放进度的变换,实时展示当前播放进度对应的封面图片。此外,播放进度还可以环绕封面图片进行展示,仅为示例。
上述方案中,通过在浮窗组件中展示目标音频的图片或播放信息等相关信息,可以使用户在对其他应用程序操作时还可以了解音频的播放情况,进一步提升了音频播放效果。
可选的,多媒体处理方法还可以包括:接收对浮窗组件的触发操作,将第一应用程序从后台运行切换至前台运行,并返回展示第二多媒体界面。当接收到用户对浮窗组件的点击操作之后,可以将当前的第二应用程序切换到后台运行,并将第一应用程序从后台运行且至前台运行,展示第二多媒体界面。
可选的,在接收程序切换请求之前若目标音频正在播放过程中,多媒体处理方法还可以包括:基于浮窗组件继续播放目标音频。如果目标音频正在播放过程中,接收到程序切换请求并在第二应用程序的展示界面中展示浮窗组件之后,可以基于浮窗组件继续播放目标音频,以实现音频的无缝播放。
上述方案中,在展示多媒体界面的基础上,当切换到其他应用程序之后还可以展示多媒体界面的浮窗组件,通过该浮窗组件可以快速返回多媒体界面,并且继续播放音频数据和展示播放情况,提升了返回多媒体界面的效率,更能满足用户的需求,提升了用户的体验效果。
在一些实施例中,多媒体处理方法还可以包括:响应于用户在当前展示的第三多媒体 界面中的界面切换请求,将当前展示的第三多媒体界面切换为第二多媒体界面;其中,第三多媒体界面中包括第三内容,第三内容包括第二内容的属性信息。可选的,第二内容的属性信息包括第二内容的标题信息、时间信息和来源信息中的至少一个。
其中,第三多媒体界面是与上述第一多媒体界面、第二多媒体界面展示内容不同的一个多媒体界面。第三多媒体界面中包括的第三内容与第二内容具有关联关系,可以包括第二内容的属性信息,第二内容的属性信息可以根据实际情况确定,例如可以包括第二内容的标题信息、时间信息和来源信息等信息中的至少一个。示例性的,第三多媒体界面可以为包括一个信息列表的界面,信息列表中包括多个音频的属性信息,其中一个为目标音频的属性信息。终端展示第三多媒体界面,当接收到用户在第三多媒体界面中的界面切换请求之后,可以将当前展示的第三多媒体界面关闭并打开第二多媒体界面,在第二多媒体界面中展示第二内容,实现界面的切换。
上述方案中,可以通过两种展示内容不同的多媒体界面的切换操作,均可以切换至仅包括音频和字幕的多媒体界面,进一步提升了不同模式的多媒体界面的切换灵活性,并且提升了用户的界面切换效率。
图2为本公开实施例提供的另一种多媒体处理方法的流程示意图,本实施例在上述实施例的基础上,进一步优化了上述多媒体处理方法。如图2所示,该方法包括:
步骤201、展示第一多媒体界面或第三多媒体界面。
其中,第一多媒体界面中包括第一内容,第三多媒体界面中包括第三内容。
步骤202、接收用户在第一多媒体界面或第三多媒体界面中的界面切换请求。
步骤203、将当前展示的第一多媒体界面或第三多媒体界面切换为第二多媒体界面,并在第二多媒体界面中展示第二内容。
其中,第一内容包括第二内容以及与第二内容关联的其他内容,第二内容包括目标音频以及目标音频对应的目标字幕。第三内容包括第二内容的属性信息,第二内容的属性信息包括第二内容的标题信息、时间信息和来源信息中的至少一个。
示例性的,图3为本公开实施例提供的一种多媒体界面的示意图,如图3所示,展示了第二多媒体界面的一个示意图,第二多媒体界面中展示了一个音频和对应的字幕,音频的时间轴和作用于音频的多个控制按键的位置在图3中仅为示例,图中还展示了音频的封 面图片和名称“团队回顾会议”。
步骤203之后,可以执行步骤204-步骤206、步骤207-步骤211和/或步骤212-步骤216,具体执行顺序不限,图2中仅为示例。
步骤204、接收对目标音频的播放触发操作。
步骤205、播放目标音频,并在目标音频的播放过程中,基于目标字幕中包括的字幕语句的时间戳,对与目标音频的播放进度对应的字幕语句进行突出展示。
可选的,多媒体处理方法还包括:确定目标音频中的非静音片段;播放目标音频,包括:在播放目标音频时仅播放非静音片段。
可选的,多媒体处理方法还包括:确定目标音频中的静音片段和非静音片段;播放目标音频,包括:以第一播放速度播放静音片段,以第二播放速度播放非静音片段,其中,第一播放速度大于第二播放速度。
步骤206、响应于目标音频播放结束,获取与目标音频关联的下一音频,切换播放下一音频。
步骤207、接收用户在第二多媒体界面上的交互触发操作。
步骤208、响应于交互触发操作,在第二多媒体界面展示交互组件,基于交互组件获取交互内容。
其中,交互组件包括表情部件和/或评论部件,交互内容包括交互表情和/或评论。
示例性的,图4为本公开实施例提供的另一种多媒体界面的示意图,如图4所示,图中展示了第二多媒体界面的另一个示意图,与图3相比,图4中部分内容和按键的位置不同。图4中目标字幕中具有下划线的字幕语句表征目标音频的播放进度对应的字幕语句,随着目标视频的播放,其他字幕语句也会通过下划线的方式进行突出展示。并且图中示例性的展示了交互组件中的表情部件11,用户点击该表情部件11时,可以发送默认的交互表情并在第二多媒体界面中进行展示,如界面中间的“点赞”。可选的,用户点击该表情部件11时,还可以展示表情面板,表情面板中可以包括多个表情以供用户选择(图中未展示)。并且,图4中第二多媒体界面的下方展示了退出按键,当用户触发该退出按键时,可以从第二多媒体界面退出至第一多媒体界面。图3和图4中展示的第二多媒体界面均为示例,不应作为限定。
步骤209、在第二多媒体界面和/或第一多媒体界面上展示交互内容。
步骤210、确定交互输入触发操作对应的交互时间点。
步骤211、在第二多媒体界面和/或第一多媒体界面上目标音频的播放时间轴上,交互时间点所在位置展示交互提示标识。
步骤212、接收程序切换请求。
步骤213、将第一应用程序切换至后台运行并启动第二应用程序,展示第二应用程序的展示界面。
其中,第一多媒体界面和第二多媒体界面均为第一应用程序的界面。
步骤214、在第二应用程序的展示界面中展示第二多媒体界面的浮窗组件。
可选的,浮窗组件中包括目标音频的封面图片和/或播放信息。可选的,播放信息包括播放进度;封面图片和播放进度关联展示。可选的,封面图片的数量为多个,封面图片随着播放进度的变化而变化。可选的,播放进度环绕封面图片显示。可选的,封面图片基于第一内容确定。
示例性的,图5为本公开实施例提供的一种浮窗组件的示意图,如图5所示,图中展示了其他应用程序下的浮窗组件12,其他应用程序即为第二应用程序,浮窗组件12可以设置在其他应用程序的展示界面的靠近边界的区域。并且浮窗组件12中还展示了第二多媒体界面的封面图片以及播放进度,图中浮窗组件边缘的黑色填充区域表征了播放进度,图中播放进度为接近三分之二。图5展示的浮窗组件仅为示例,其他形状或样式的浮窗组件也可适用。
步骤214之后,可以执行步骤215和/或步骤216。
步骤215、接收对浮窗组件的触发操作,将第一应用程序从后台运行切换至前台运行,并返回展示第二多媒体界面。
步骤216、基于浮窗组件继续播放目标音频。
在执行步骤210之前,若目标音频正则播放过程中,则可以基于浮窗组件继续播放目标音频。
在一些实施例中,多媒体处理方法还可以包括:接收对第一多媒体界面中展示的目标字幕的修改操作;对第二多媒体界面中展示的目标字幕进行同步修改。
本公开实施例提供的多媒体处理方案,展示第一多媒体界面,第一多媒体界面中包括第一内容;接收用户在第一多媒体界面中的界面切换请求;将当前展示的第一多媒体界面 切换为第二多媒体界面,并在第二多媒体界面中展示第二内容;其中,第一内容包括第二内容以及与第二内容关联的其他内容,第二内容包括目标音频以及目标音频对应的目标字幕。采用上述技术方案,可以实现包括两种不同内容的界面的切换,其中一个界面可以仅包括音频和字幕,有助于用户在复杂场景下专注于多媒体内容,并且提升了多媒体内容播放的灵活性,可以满足多种场景的需求,进而提升了用户的体验效果。
图6为本公开实施例提供的一种多媒体处理装置的结构示意图,该装置可由软件和/或硬件实现,一般可集成在电子设备中。如图6所示,该装置包括:
第一界面模块301,用于展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
请求模块302,用于接收用户在第一多媒体界面中的界面切换请求;
第二界面模块303,用于将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
可选的,所述装置还包括播放模块,用于:
接收对所述目标音频的播放触发操作;
播放所述目标音频,并在所述目标音频的播放过程中,基于所述目标字幕中包括的字幕语句的时间戳,对与所述目标音频的播放进度对应的字幕语句进行突出展示。
可选的,所述播放模块具体用于:
响应于所述目标音频播放结束,获取与所述目标音频关联的下一音频,切换播放所述下一音频。
可选的,所述装置还包括第一音频识别模块,用于:
确定所述目标音频中的非静音片段;
所述播放模块具体用于:
在播放所述目标音频时仅播放所述非静音片段。
可选的,所述装置还包括第二音频识别模块,用于:
确定所述目标音频中的静音片段和非静音片段;
所述播放模块具体用于:
以第一播放速度播放所述静音片段,以第二播放速度播放所述非静音片段,其中,所述第一播放速度大于所述第二播放速度。
可选的,所述装置还包括交互模块,用于:
接收用户在所述第二多媒体界面上的交互触发操作;
基于所述交互触发操作,确定交互内容。
可选的,所述交互模块用于:
响应于所述交互触发操作,在所述第二多媒体界面展示交互组件;
基于所述交互组件获取交互内容,并在所述第二多媒体界面展示所述交互内容;
其中,所述交互组件包括表情部件和/或评论部件,所述交互内容包括交互表情和/或评论。
可选的,所述交互模块用于:
在所述第一多媒体界面上展示所述交互内容。
可选的,所述交互模块用于:
确定所述交互输入触发操作对应的交互时间点;
在所述第二多媒体界面和/或所述第一多媒体界面上所述目标音频的播放时间轴上,所述交互时间点所在位置展示交互提示标识。
可选的,所述装置还包括修改模块,用于:
接收对所述第一多媒体界面中展示的所述目标字幕的修改操作;
对所述第二多媒体界面中展示的所述目标字幕进行同步修改。
可选的,所述第一多媒体界面和所述第二多媒体界面均为第一应用程序的界面,所述装置还包括浮窗模块,用于:
接收程序切换请求;
将所述第一应用程序切换至后台运行并启动第二应用程序,展示所述第二应用程序的展示界面;
在所述第二应用程序的展示界面中展示所述第二多媒体界面的浮窗组件。
可选的,所述浮窗组件包括所述目标音频的封面图片和/或播放信息。
可选的,所述播放信息包括播放进度;所述封面图片和所述播放进度关联展示。
可选的,所述封面图片的数量为多个,所述封面图片随着播放进度的变化而变化。
可选的,所述播放进度环绕所述封面图片显示。
可选的,所述封面图片基于所述第一内容确定。
可选的,所述装置还包括返回模块,用于:
接收对所述浮窗组件的触发操作,将所述第一应用程序从后台运行切换至前台运行,并返回展示所述第二多媒体界面。
可选的,在所述接收程序切换请求之前若所述目标音频正在播放过程中,所述浮窗模块还用于:
基于所述浮窗组件继续播放所述目标音频。
可选的,所述装置还包括第三界面模块,用于:
响应于用户在当前展示的第三多媒体界面中的界面切换请求,将当前展示的所述第三多媒体界面切换为所述第二多媒体界面;
其中,所述第三多媒体界面中包括第三内容,所述第三内容包括所述第二内容的属性信息。
可选的,所述第二内容的属性信息包括所述第二内容的标题信息、时间信息和来源信息中的至少一个。
本公开实施例所提供的多媒体处理装置可执行本公开任意实施例所提供的多媒体处理方法,具备执行方法相应的功能模块和有益效果。
本公开实施例还提供了一种计算机程序产品,包括计算机程序/指令,该计算机程序/指令被处理器执行时实现本公开任意实施例所提供的多媒体处理方法。
图7为本公开实施例提供的一种电子设备的结构示意图。下面具体参考图7,其示出了适于用来实现本公开实施例中的电子设备400的结构示意图。本公开实施例中的电子设备400可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图7示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图7所示,电子设备400可以包括处理装置(例如中央处理器、图形处理器等)401,其可以根据存储在只读存储器(ROM)402中的程序或者从存储装置408加载到随机访问存 储器(RAM)403中的程序而执行各种适当的动作和处理。在RAM 403中,还存储有电子设备400操作所需的各种程序和数据。处理装置401、ROM 402以及RAM 403通过总线404彼此相连。输入/输出(I/O)接口405也连接至总线404。
通常,以下装置可以连接至I/O接口405:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置406;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置407;包括例如磁带、硬盘等的存储装置408;以及通信装置409。通信装置409可以允许电子设备400与其他设备进行无线或有线通信以交换数据。虽然图7示出了具有各种装置的电子设备400,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在非暂态计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置409从网络上被下载和安装,或者从存储装置408被安装,或者从ROM 402被安装。在该计算机程序被处理装置401执行时,执行本公开实施例的多媒体处理方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读 介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。
在一些实施方式中,客户端、服务器可以利用诸如HTTP(HyperText Transfer Protocol,超文本传输协议)之类的任何当前已知或未来研发的网络协议进行通信,并且可以与任意形式或介质的数字数据通信(例如,通信网络)互连。通信网络的示例包括局域网(“LAN”),广域网(“WAN”),网际网(例如,互联网)以及端对端网络(例如,ad hoc端对端网络),以及任何当前已知或未来研发的网络。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:展示第一多媒体界面,所述第一多媒体界面中包括第一内容;接收用户在第一多媒体界面中的界面切换请求;将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括但不限于面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实 际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、片上系统(SOC)、复杂可编程逻辑设备(CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
根据本公开的一个或多个实施例,本公开提供了一种多媒体处理方法,包括:
展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
接收用户在第一多媒体界面中的界面切换请求;
将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
接收对所述目标音频的播放触发操作;
播放所述目标音频,并在所述目标音频的播放过程中,基于所述目标字幕中包括的字幕语句的时间戳,对与所述目标音频的播放进度对应的字幕语句进行突出展示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
响应于所述目标音频播放结束,获取与所述目标音频关联的下一音频,切换播放所述下一音频。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
确定所述目标音频中的非静音片段;
所述播放所述目标音频,包括:
在播放所述目标音频时仅播放所述非静音片段。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:确定所述目标音频中的静音片段和非静音片段;
所述播放所述目标音频,包括:
以第一播放速度播放所述静音片段,以第二播放速度播放所述非静音片段,其中,所述第一播放速度大于所述第二播放速度。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
接收用户在所述第二多媒体界面上的交互触发操作;
基于所述交互触发操作,确定交互内容。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述基于所述交互触发操作,确定交互内容,包括:
响应于所述交互触发操作,在所述第二多媒体界面展示交互组件;
基于所述交互组件获取交互内容,并在所述第二多媒体界面展示所述交互内容;
其中,所述交互组件包括表情部件和/或评论部件,所述交互内容包括交互表情和/或评论。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
在所述第一多媒体界面上展示所述交互内容。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
确定所述交互输入触发操作对应的交互时间点;
在所述第二多媒体界面和/或所述第一多媒体界面上所述目标音频的播放时间轴上, 所述交互时间点所在位置展示交互提示标识。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
接收对所述第一多媒体界面中展示的所述目标字幕的修改操作;
对所述第二多媒体界面中展示的所述目标字幕进行同步修改。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述第一多媒体界面和所述第二多媒体界面均为第一应用程序的界面,所述方法还包括:
接收程序切换请求;
将所述第一应用程序切换至后台运行并启动第二应用程序,展示所述第二应用程序的展示界面;
在所述第二应用程序的展示界面中展示所述第二多媒体界面的浮窗组件。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述浮窗组件包括所述目标音频的封面图片和/或播放信息。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述播放信息包括播放进度;所述封面图片和所述播放进度关联展示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述封面图片的数量为多个,所述封面图片随着播放进度的变化而变化。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述播放进度环绕所述封面图片显示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述封面图片基于所述第一内容确定。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
接收对所述浮窗组件的触发操作,将所述第一应用程序从后台运行切换至前台运行,并返回展示所述第二多媒体界面。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,在所述接收程序切换请求之前若所述目标音频正在播放过程中,所述方法还包括:
基于所述浮窗组件继续播放所述目标音频。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述方法还包括:
响应于用户在当前展示的第三多媒体界面中的界面切换请求,将当前展示的所述第三 多媒体界面切换为所述第二多媒体界面;
其中,所述第三多媒体界面中包括第三内容,所述第三内容包括所述第二内容的属性信息。
根据本公开的一个或多个实施例,本公开提供的多媒体处理方法中,所述第二内容的属性信息包括所述第二内容的标题信息、时间信息和来源信息中的至少一个。
根据本公开的一个或多个实施例,本公开提供了一种多媒体处理装置,包括:
第一界面模块,用于展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
请求模块,用于接收用户在第一多媒体界面中的界面切换请求;
第二界面模块,用于将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括播放模块,用于:
接收对所述目标音频的播放触发操作;
播放所述目标音频,并在所述目标音频的播放过程中,基于所述目标字幕中包括的字幕语句的时间戳,对与所述目标音频的播放进度对应的字幕语句进行突出展示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述播放模块具体用于:
响应于所述目标音频播放结束,获取与所述目标音频关联的下一音频,切换播放所述下一音频。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括第一音频识别模块,用于:
确定所述目标音频中的非静音片段;
所述播放模块具体用于:
在播放所述目标音频时仅播放所述非静音片段。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括 第二音频识别模块,用于:
确定所述目标音频中的静音片段和非静音片段;
所述播放模块具体用于:
以第一播放速度播放所述静音片段,以第二播放速度播放所述非静音片段,其中,所述第一播放速度大于所述第二播放速度。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括交互模块,用于:
接收用户在所述第二多媒体界面上的交互触发操作;
基于所述交互触发操作,确定交互内容。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述交互模块用于:
响应于所述交互触发操作,在所述第二多媒体界面展示交互组件;
基于所述交互组件获取交互内容,并在所述第二多媒体界面展示所述交互内容;
其中,所述交互组件包括表情部件和/或评论部件,所述交互内容包括交互表情和/或评论。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述交互模块用于:
在所述第一多媒体界面上展示所述交互内容。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述交互模块用于:
确定所述交互输入触发操作对应的交互时间点;
在所述第二多媒体界面和/或所述第一多媒体界面上所述目标音频的播放时间轴上,所述交互时间点所在位置展示交互提示标识。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括修改模块,用于:
接收对所述第一多媒体界面中展示的所述目标字幕的修改操作;
对所述第二多媒体界面中展示的所述目标字幕进行同步修改。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述第一多媒体 界面和所述第二多媒体界面均为第一应用程序的界面,所述装置还包括浮窗模块,用于:
接收程序切换请求;
将所述第一应用程序切换至后台运行并启动第二应用程序,展示所述第二应用程序的展示界面;
在所述第二应用程序的展示界面中展示所述第二多媒体界面的浮窗组件。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述浮窗组件包括所述目标音频的封面图片和/或播放信息。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述播放信息包括播放进度;
所述封面图片和所述播放进度关联展示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述封面图片的数量为多个,所述封面图片随着播放进度的变化而变化。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述播放进度环绕所述封面图片显示。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述封面图片基于所述第一内容确定。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括返回模块,用于:
接收对所述浮窗组件的触发操作,将所述第一应用程序从后台运行切换至前台运行,并返回展示所述第二多媒体界面。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,在所述接收程序切换请求之前若所述目标音频正在播放过程中,所述浮窗模块还用于:
基于所述浮窗组件继续播放所述目标音频。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述装置还包括第三界面模块,用于:
响应于用户在当前展示的第三多媒体界面中的界面切换请求,将当前展示的所述第三多媒体界面切换为所述第二多媒体界面;
其中,所述第三多媒体界面中包括第三内容,所述第三内容包括所述第二内容的属性 信息。
根据本公开的一个或多个实施例,本公开提供的多媒体处理装置中,所述第二内容的属性信息包括所述第二内容的标题信息、时间信息和来源信息中的至少一个。
根据本公开的一个或多个实施例,本公开提供了一种电子设备,包括:
处理器;
用于存储所述处理器可执行指令的存储器;
所述处理器,用于从所述存储器中读取所述可执行指令,并执行所述指令以实现如本公开提供的任一所述的多媒体处理方法。
根据本公开的一个或多个实施例,本公开提供了一种计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序用于执行如本公开提供的任一所述的多媒体处理方法。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (26)

  1. 一种多媒体处理方法,包括:
    展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
    接收用户在第一多媒体界面中的界面切换请求;
    将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
    其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
  2. 根据权利要求1所述的方法,还包括:
    接收对所述目标音频的播放触发操作;
    播放所述目标音频,并在所述目标音频的播放过程中,基于所述目标字幕中包括的字幕语句的时间戳,对与所述目标音频的播放进度对应的字幕语句进行突出展示。
  3. 根据权利要求2所述的方法,还包括:
    响应于所述目标音频播放结束,获取与所述目标音频关联的下一音频,切换播放所述下一音频。
  4. 根据权利要求2所述的方法,还包括:
    确定所述目标音频中的非静音片段;
    所述播放所述目标音频,包括:
    在播放所述目标音频时仅播放所述非静音片段。
  5. 根据权利要求2所述的方法,还包括:确定所述目标音频中的静音片段和非静音片段;
    所述播放所述目标音频,包括:
    以第一播放速度播放所述静音片段,以第二播放速度播放所述非静音片段,其中,所述第一播放速度大于所述第二播放速度。
  6. 根据权利要求1所述的方法,还包括:
    接收用户在所述第二多媒体界面上的交互触发操作;
    基于所述交互触发操作,确定交互内容。
  7. 根据权利要求6所述的方法,其中,所述基于所述交互触发操作,确定交互内容,包括:
    响应于所述交互触发操作,在所述第二多媒体界面展示交互组件;
    基于所述交互组件获取交互内容,并在所述第二多媒体界面展示所述交互内容;
    其中,所述交互组件包括表情部件和/或评论部件,所述交互内容包括交互表情和/或评论。
  8. 根据权利要求6所述的方法,还包括:
    在所述第一多媒体界面上展示所述交互内容。
  9. 根据权利要求6所述的方法,还包括:
    确定所述交互输入触发操作对应的交互时间点;
    在所述第二多媒体界面和/或所述第一多媒体界面上所述目标音频的播放时间轴上,所述交互时间点所在位置展示交互提示标识。
  10. 根据权利要求1所述的方法,还包括:
    接收对所述第一多媒体界面中展示的所述目标字幕的修改操作;
    对所述第二多媒体界面中展示的所述目标字幕进行同步修改。
  11. 根据权利要求1所述的方法,其中,所述第一多媒体界面和所述第二多媒体界面均为第一应用程序的界面,所述方法还包括:
    接收程序切换请求;
    将所述第一应用程序切换至后台运行并启动第二应用程序,展示所述第二应用程序的展示界面;
    在所述第二应用程序的展示界面中展示所述第二多媒体界面的浮窗组件。
  12. 根据权利要求1所述的方法,还包括:
    接收用户在第二多媒体界面中的返回操作;
    返回展示所述第一多媒体界面并在所述第一多媒体界面中展示所述第二多媒体界面的浮窗组件。
  13. 根据权利要求11或12所述的方法,其中,所述浮窗组件包括所述目标音频的封面图片和/或播放信息。
  14. 根据权利要求13所述的方法,其中,所述播放信息包括播放进度;所述封面图片和所述播放进度关联展示。
  15. 根据权利要求14所述的方法,其中,所述封面图片的数量为多个,所述封面图片随着播放进度的变化而变化。
  16. 根据权利要求14所述的方法,其中,所述播放进度环绕所述封面图片显示。
  17. 根据权利要求13所述的方法,其中,所述封面图片基于所述第一内容确定。
  18. 根据权利要求11或12所述的方法,还包括:
    接收对所述浮窗组件的触发操作,返回展示所述第二多媒体界面。
  19. 根据权利要求11或12所述的方法,还包括:
    基于所述浮窗组件继续播放所述目标音频。
  20. 根据权利要求1所述的方法,还包括:
    响应于用户在当前展示的第三多媒体界面中的界面切换请求,将当前展示的所述第三多媒体界面切换为所述第二多媒体界面;
    其中,所述第三多媒体界面中包括第三内容,所述第三内容包括所述第二内容的属性信息。
  21. 根据权利要求20所述的方法,其中,所述第二内容的属性信息包括所述第二内容的标题信息、时间信息和来源信息中的至少一个。
  22. 一种多媒体处理装置,包括:
    第一界面模块,用于展示第一多媒体界面,所述第一多媒体界面中包括第一内容;
    请求模块,用于接收用户在第一多媒体界面中的界面切换请求;
    第二界面模块,用于将当前展示的所述第一多媒体界面切换为第二多媒体界面,并在所述第二多媒体界面中展示第二内容;
    其中,所述第一内容包括所述第二内容以及与所述第二内容关联的其他内容,所述第二内容包括目标音频以及所述目标音频对应的目标字幕。
  23. 一种电子设备,包括:
    处理器;
    用于存储所述处理器可执行指令的存储器;
    所述处理器,用于从所述存储器中读取所述可执行指令,并执行所述指令以实现上述 权利要求1-21中任一所述的多媒体处理方法。
  24. 一种计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序用于执行上述权利要求1-21中任一所述的多媒体处理方法。
  25. 一种计算机程序,包括:
    指令,所述指令当由处理器执行时使所述处理器执行根据权利要求1-21中任一项所述的多媒体处理方法。
  26. 一种计算机程序产品,包括指令,所述指令当由处理器执行时使所述处理器执行根据权利要求1-21中任一项所述的多媒体处理方法。
PCT/CN2022/085468 2021-05-19 2022-04-07 一种多媒体处理方法、装置、设备及介质 WO2022242351A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/262,301 US20240121479A1 (en) 2021-05-19 2022-04-07 Multimedia processing method, apparatus, device, and medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110547916.9 2021-05-19
CN202110547916.9A CN113259740A (zh) 2021-05-19 2021-05-19 一种多媒体处理方法、装置、设备及介质

Publications (1)

Publication Number Publication Date
WO2022242351A1 true WO2022242351A1 (zh) 2022-11-24

Family

ID=77183345

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/085468 WO2022242351A1 (zh) 2021-05-19 2022-04-07 一种多媒体处理方法、装置、设备及介质

Country Status (3)

Country Link
US (1) US20240121479A1 (zh)
CN (1) CN113259740A (zh)
WO (1) WO2022242351A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024169865A1 (zh) * 2023-02-14 2024-08-22 北京字跳网络技术有限公司 媒体内容显示方法、装置、设备、可读存储介质及产品

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113259740A (zh) * 2021-05-19 2021-08-13 北京字跳网络技术有限公司 一种多媒体处理方法、装置、设备及介质
CN113721810A (zh) * 2021-09-08 2021-11-30 北京字跳网络技术有限公司 显示方法、装置、设备及存储介质
CN113849258B (zh) * 2021-10-13 2024-07-02 北京字跳网络技术有限公司 内容展示方法、装置、设备及存储介质
CN113891168B (zh) * 2021-10-19 2023-12-19 北京有竹居网络技术有限公司 字幕处理方法、装置、电子设备和存储介质
CN113885830B (zh) * 2021-10-25 2024-07-02 北京字跳网络技术有限公司 一种音效展示方法及终端设备
CN114327178B (zh) * 2021-11-18 2023-05-02 荣耀终端有限公司 界面显示方法、装置及存储介质
CN114979798B (zh) * 2022-04-21 2024-03-22 维沃移动通信有限公司 播放速度控制方法和电子设备
CN115047999B (zh) * 2022-07-27 2024-07-02 北京字跳网络技术有限公司 界面切换方法、装置、电子设备、存储介质及程序产品
CN115484490B (zh) * 2022-09-15 2024-06-11 北京字跳网络技术有限公司 一种视频处理方法、装置、设备及存储介质

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108337543A (zh) * 2017-12-27 2018-07-27 努比亚技术有限公司 一种视频播放方法、终端及计算机可读存储介质
CN306040285S (zh) * 2020-02-17 2020-09-08
CN112380365A (zh) * 2020-11-18 2021-02-19 北京字跳网络技术有限公司 一种多媒体的字幕交互方法、装置、设备及介质
CN112397104A (zh) * 2020-11-26 2021-02-23 北京字节跳动网络技术有限公司 音频与文本的同步方法、装置、可读介质和电子设备
CN112416963A (zh) * 2020-11-11 2021-02-26 北京字跳网络技术有限公司 搜索内容的匹配方法、装置、电子设备及存储介质
US20210074298A1 (en) * 2019-09-11 2021-03-11 Soundhound, Inc. Video conference captioning
CN113259740A (zh) * 2021-05-19 2021-08-13 北京字跳网络技术有限公司 一种多媒体处理方法、装置、设备及介质

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999017235A1 (en) * 1997-10-01 1999-04-08 At & T Corp. Method and apparatus for storing and retrieving labeled interval data for multimedia recordings

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108337543A (zh) * 2017-12-27 2018-07-27 努比亚技术有限公司 一种视频播放方法、终端及计算机可读存储介质
US20210074298A1 (en) * 2019-09-11 2021-03-11 Soundhound, Inc. Video conference captioning
CN306040285S (zh) * 2020-02-17 2020-09-08
CN112416963A (zh) * 2020-11-11 2021-02-26 北京字跳网络技术有限公司 搜索内容的匹配方法、装置、电子设备及存储介质
CN112380365A (zh) * 2020-11-18 2021-02-19 北京字跳网络技术有限公司 一种多媒体的字幕交互方法、装置、设备及介质
CN112397104A (zh) * 2020-11-26 2021-02-23 北京字节跳动网络技术有限公司 音频与文本的同步方法、装置、可读介质和电子设备
CN113259740A (zh) * 2021-05-19 2021-08-13 北京字跳网络技术有限公司 一种多媒体处理方法、装置、设备及介质

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024169865A1 (zh) * 2023-02-14 2024-08-22 北京字跳网络技术有限公司 媒体内容显示方法、装置、设备、可读存储介质及产品

Also Published As

Publication number Publication date
US20240121479A1 (en) 2024-04-11
CN113259740A (zh) 2021-08-13

Similar Documents

Publication Publication Date Title
WO2022242351A1 (zh) 一种多媒体处理方法、装置、设备及介质
CN111970577B (zh) 字幕编辑方法、装置和电子设备
WO2022068533A1 (zh) 互动信息处理方法、装置、设备及介质
WO2022105710A1 (zh) 一种会议纪要的交互方法、装置、设备及介质
WO2023066297A1 (zh) 消息处理方法、装置、设备及存储介质
CN113365134B (zh) 音频分享方法、装置、设备及介质
WO2022105760A1 (zh) 一种多媒体浏览方法、装置、设备及介质
WO2023036277A1 (zh) 内容显示方法、装置、设备及介质
EP4124052B1 (en) Video production method and apparatus, and device and storage medium
WO2022105709A1 (zh) 多媒体的交互方法、信息交互方法、装置、设备及介质
US20220406311A1 (en) Audio information processing method, apparatus, electronic device and storage medium
CN113778419B (zh) 多媒体数据的生成方法、装置、可读介质及电子设备
WO2023035835A1 (zh) 一种信息展示方法、装置、设备及介质
CN112380365A (zh) 一种多媒体的字幕交互方法、装置、设备及介质
CN112165652A (zh) 视频处理方法、装置、设备及计算机可读存储介质
CN113253885A (zh) 目标内容的显示方法、装置、设备、可读存储介质及产品
CN110650376B (zh) 视频播放时的转场动画的实现方法、装置、移动终端、存储介质
WO2024140239A1 (zh) 页面显示方法、装置、设备、计算机可读存储介质及产品
CN112380362B (zh) 基于用户交互的音乐播放方法、装置、设备及存储介质
WO2023088484A1 (zh) 用于多媒体资源剪辑场景的方法、装置、设备及存储介质
WO2022257777A1 (zh) 多媒体处理方法、装置、设备及介质
CN113132789B (zh) 一种多媒体的交互方法、装置、设备及介质
JP7572108B2 (ja) 議事録のインタラクション方法、装置、機器及び媒体
WO2024165040A1 (zh) 信息显示方法、装置、设备及存储介质
EP4383698A1 (en) Multimedia data processing method, apparatus, device and medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22803676

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 18262301

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22803676

Country of ref document: EP

Kind code of ref document: A1