CN110166801B - Media file processing method and device and storage medium - Google Patents

Media file processing method and device and storage medium Download PDF

Info

Publication number
CN110166801B
CN110166801B CN201910340609.6A CN201910340609A CN110166801B CN 110166801 B CN110166801 B CN 110166801B CN 201910340609 A CN201910340609 A CN 201910340609A CN 110166801 B CN110166801 B CN 110166801B
Authority
CN
China
Prior art keywords
data
media file
media
extension
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910340609.6A
Other languages
Chinese (zh)
Other versions
CN110166801A (en
Inventor
高小猛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN201910340609.6A priority Critical patent/CN110166801B/en
Publication of CN110166801A publication Critical patent/CN110166801A/en
Application granted granted Critical
Publication of CN110166801B publication Critical patent/CN110166801B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/251Learning process for intelligent management, e.g. learning user preferences for recommending movies
    • H04N21/252Processing of multiple end-users' preferences to derive collaborative data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25891Management of end-user data being end-user preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47202End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages

Abstract

The embodiment of the invention discloses a method, a device and a storage medium for processing a media file, wherein the method comprises the following steps: receiving a playing instruction of a media file; acquiring media data of the media file based on the playing instruction; acquiring extended data for extending the presentation form of the media file and configuration data corresponding to the extended data and indicating to realize target extension; and displaying the media data and the extension data in a mode of fusing the display form of the extension data and the display form of the media data based on the configuration data. Therefore, the display forms of the media data are richer and more diversified, and the user experience is improved.

Description

Media file processing method and device and storage medium
Technical Field
The present invention relates to data processing technologies, and in particular, to a method and an apparatus for processing a media file, and a storage medium.
Background
With the development of multimedia technology, during the playing process of online multimedia, only single presentation of the information of the online multimedia is performed, and the presentation form is single, so that the diversified requirements of users on the information presentation cannot be met.
Disclosure of Invention
The embodiment of the invention provides a media file processing method, a media file processing device and a storage medium, which can expand the presentation form in the playing process of a media file.
The technical scheme of the embodiment of the invention is realized as follows:
the embodiment of the invention provides a media file processing method, which comprises the following steps:
receiving a playing instruction of a media file;
acquiring media data of the media file based on the playing instruction;
acquiring extended data for extending the presentation form of the media file and configuration data corresponding to the extended data and indicating to realize target extension;
and displaying the media data and the extension data in a mode of fusing the display form of the extension data and the display form of the media data based on the configuration data.
An embodiment of the present invention further provides a device for processing a media file, where the device includes:
the receiving unit is used for receiving a playing instruction of the media file;
the acquisition unit is used for acquiring the media data of the media file based on the playing instruction;
the media file expansion system is used for acquiring expansion data used for expanding the presentation form of the media file and configuration data corresponding to the expansion data and indicating the realization of target expansion;
and the playing unit is used for displaying the media data and the extended data in a mode of fusing the display form of the extended data and the display form of the media data based on the configuration data.
In the foregoing solution, the obtaining unit is further configured to obtain at least one of audio data, video data, and picture data for prompting the switching of the key information of the media file.
In the above solution, the obtaining unit is further configured to determine a file identifier of the media file;
and sending an extended data request carrying the file identifier;
and receiving the returned extension data corresponding to the file identifier.
In the above scheme, the obtaining unit is further configured to obtain a corresponding user identifier in response to that the current user login state is logged in;
and determining corresponding social network information based on the user identification;
and obtaining the extension data associated with the media file based on the social network information.
In the above scheme, the obtaining unit is further configured to obtain a user identifier of a registered user, where the registered user corresponds to the playing client of the media file;
and determining extension data associated with the registered user based on the user identifier of the registered user;
and screening the determined extension data to obtain the extension data related to the media file.
In the above scheme, the obtaining unit is further configured to receive a form extension instruction corresponding to the media file in a playing process of the media file;
and acquiring extension data for extending the presentation form of the media file in response to the form extension instruction.
In the above scheme, the obtaining unit is further configured to obtain historical behavior data of the target user;
and determining, based on the historical behavior data, a user representation indicative of an interest classification of the target user;
and obtaining the extension data corresponding to the user representation.
In the above scheme, the playing unit is further configured to fuse the extended data and the media data based on a target extension manner indicated by the configuration data to obtain fused media data;
and displaying the obtained fusion media data through a playing window.
In the above scheme, the playing unit is further configured to display the media data through a first playing window;
and displaying the extended data through a second playing window based on the configuration data, so that the display form of the extended data is fused with the display form of the media data.
An embodiment of the present invention further provides a device for processing a media file, including:
a memory configured to store a processing program for a media file;
and the processor is configured to run the program, wherein the program executes the processing method of the media file provided by the embodiment of the invention when running.
The embodiment of the invention also provides a storage medium which stores an executable program, and when the executable program is executed by a processor, the processing method of the media file provided by the embodiment of the invention is realized.
The application of the embodiment of the invention has the following beneficial effects:
the media data and the extended data are displayed in a mode of fusing the display form of the extended data and the display form of the media data, the playing form of the media file is extended, the display form of the media data is richer and more diverse, and the user experience is improved.
Drawings
FIG. 1 is a block diagram of a media file processing system according to an embodiment of the present invention;
fig. 2 is a schematic hardware structure diagram of a media file processing apparatus according to an embodiment of the present invention;
FIG. 3 is a flowchart illustrating a method for processing a media file according to an embodiment of the present invention;
FIG. 4 is a flowchart illustrating a method for processing a media file according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of an interface for expanding a user trigger format according to an embodiment of the present invention;
FIG. 6 is a flowchart illustrating a method for processing a media file according to an embodiment of the present invention;
fig. 7 is a schematic diagram illustrating that a terminal and a server cooperatively implement media file processing according to an embodiment of the present invention;
FIG. 8 is a schematic diagram of sound effect enhancement for an on-line movie according to an embodiment of the present invention;
fig. 9 is a schematic diagram of a software structure of a media file processing apparatus according to an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail below with reference to the drawings and examples. It should be understood that the examples provided herein are merely illustrative of the present invention and are not intended to limit the present invention. In addition, the following embodiments are provided as partial embodiments for implementing the present invention, not all embodiments for implementing the present invention, and the technical solutions described in the embodiments of the present invention may be implemented in any combination without conflict.
It should be noted that, in the embodiments of the present invention, the terms "comprises", "comprising" or any other variation thereof are intended to cover a non-exclusive inclusion, so that a method or apparatus including a series of elements includes not only the explicitly recited elements but also other elements not explicitly listed or inherent to the method or apparatus. Without further limitation, an element defined by the phrase "comprising a … …" does not exclude the presence of other related elements in a method or apparatus that comprises the element (e.g., a step in a method or a unit in an apparatus, such as a unit that may be part of a circuit, part of a processor, part of a program or software, etc.).
For example, the media file processing method provided by the embodiment of the present invention includes a series of steps, but the media file processing method provided by the embodiment of the present invention is not limited to the described steps, and similarly, the media file processing apparatus provided by the embodiment of the present invention includes a series of units, but the apparatus provided by the embodiment of the present invention is not limited to include the explicitly described units, and may also include units that are required to obtain related information or perform processing based on the information.
In the description that follows, references to the terms "first", "second", and the like, are intended only to distinguish between similar objects and not to indicate a particular ordering for the objects, it being understood that "first", "second", and the like may be interchanged under certain circumstances or sequences of events to enable embodiments of the invention described herein to be practiced in other than the order illustrated or described herein.
Before further detailed description of the present invention, terms and expressions referred to in the embodiments of the present invention are described, and the terms and expressions referred to in the embodiments of the present invention are applicable to the following explanations.
1) The user representation refers to a virtual representation of a real user, is a target user model established on a series of attribute data, and refers to a hierarchical interest model of a corresponding user abstracted according to historical behavior data of the user, and is used for indicating interest classification of the user.
2) Media files, media available in the internet in various forms (e.g., video, audio, teletext, etc.) such as video files, audio files, articles including teletext forms, etc. presented in a client.
3) In response to the condition or state on which the performed operation depends, one or more of the performed operations may be in real-time or may have a set delay when the dependent condition or state is satisfied; there is no restriction on the order of execution of the operations performed unless otherwise specified.
Fig. 1 is an alternative architecture diagram of a media file processing system according to an embodiment of the present invention, and referring to fig. 1, to support an exemplary application, terminals (including a terminal 400-1 and a terminal 400-2) are connected to a server 200 through a network 300, where the network 300 may be a wide area network or a local area network, or a combination of the two, and data transmission is implemented using a wireless link.
A terminal (terminal 400-1 and/or terminal 400-2) for receiving a play instruction of the media file, and requesting media data of the media file from the server 200 based on the play instruction;
and, for requesting extension data for extending the presentation form of the media file and configuration data indicating implementation of the target extension corresponding to the extension data from the server 200;
a server 200 for returning media data, extension data and configuration data based on a request of a terminal;
in practical applications, the server 200 may be a single server configured to support various services, or may be a server cluster.
The terminal (terminal 400-1 and/or terminal 400-2) is further configured to present the media data and the extension data in a manner of fusing a presentation form of the extension data and a presentation form of the media data based on the configuration data.
In practical applications, the terminal may be various types of user terminals such as a smart phone, a tablet computer, a notebook computer, and the like, and may also be a wearable computing device, a Personal Digital Assistant (PDA), a desktop computer, a cellular phone, a media player, a navigation device, a game console, a television, or a combination of any two or more of these data processing devices or other data processing devices.
In some embodiments, a media playing client (e.g., a video playing client, an audio playing client) is disposed on the terminal, and when a user triggers playing of a media file based on the media playing client, expansion of a media file presentation form is automatically implemented.
When receiving a play instruction of a media file triggered by a user, the media play client requests the server 200 for media data of the media file, expansion data for expanding a presentation form of the media file, and configuration data indicating that target expansion is realized corresponding to the expansion data, and presents the media data and the expansion data in a manner of fusing the presentation form of the expansion data and the presentation form of the media data based on the configuration data.
In some embodiments, during the playing process of the media file, when the user triggers the format extension instruction of the media file based on the media playing client, the media file presentation format extension is realized.
When a user triggers a playing instruction of a media file based on a media playing client, the media playing client requests media data of the media file from the server 200 to play the media file, in the playing process of the media file, the media playing client receives a form expansion instruction of the media file triggered by the user, requests expansion data of a display form used for expanding the media file from the server 200, and configuration data corresponding to the expansion data and indicating realization of target expansion, and displays the media data and the expansion data in a mode of fusing the display form of the expansion data and the display form of the media data based on the configuration data.
Next, a media file processing apparatus according to an embodiment of the present invention will be described. The media file processing device provided by the embodiment of the present invention may be implemented as hardware or a combination of hardware and software, and various exemplary implementations of the device provided by the embodiment of the present invention are described below.
In the following, a hardware structure of a media file processing apparatus according to an embodiment of the present invention is described in detail, fig. 2 is a schematic diagram of a hardware structure of a media file processing apparatus according to an embodiment of the present invention, and in an actual implementation, the media file processing apparatus may be disposed in a terminal, for example, implemented by the terminal 400-1 in fig. 1, and the media file processing apparatus may also be disposed in a server and a terminal, for example, implemented by the terminal 400-1 and the server 200 in fig. 1 in a cooperation manner, it can be understood that fig. 2 only shows an exemplary structure of the media file processing apparatus, and not a whole structure, and a part of the structure or a whole structure shown in fig. 2 may be implemented as needed.
Referring to fig. 2, an apparatus for processing a media file according to an embodiment of the present invention includes: at least one processor 201, memory 202, user interface 203, and at least one network interface 204. The various components in the processing device 20 of the media file are coupled together by a bus system 205. It will be appreciated that the bus system 205 is used to enable communications among the components. The bus system 205 includes a power bus, a control bus, and a status signal bus in addition to a data bus. For clarity of illustration, however, the various buses are labeled as bus system 205 in fig. 2.
The user interface 203 may include, among other things, a display, a keyboard, a mouse, a trackball, a click wheel, a key, a button, a touch pad, or a touch screen.
It will be appreciated that the memory 202 can be either volatile memory or nonvolatile memory, and can include both volatile and nonvolatile memory.
The memory 202 in embodiments of the present invention is used to store various types of data to support the operation of the processing device 20 for media files. Examples of such data include: any executable instructions for operating on the processing device 20 of the media file, such as executable instructions, may be included in the executable instructions, and the program implementing the processing method of the media file of the embodiment of the present invention may be included in the executable instructions.
The method for processing the media file disclosed by the embodiment of the invention can be applied to the processor 201, or can be implemented by the processor 201. The processor 201 may be an integrated circuit chip having signal processing capabilities. In implementation, the steps of the media file processing method may be implemented by integrated logic circuits of hardware or instructions in the form of software in the processor 201. The Processor 201 may be a general purpose Processor, a Digital Signal Processor (DSP), or other programmable logic device, discrete gate or transistor logic device, discrete hardware components, or the like. The processor 201 may implement or perform the methods, steps, and logic blocks disclosed in embodiments of the present invention. A general purpose processor may be a microprocessor or any conventional processor or the like. The steps of the method disclosed by the embodiment of the invention can be directly implemented by a hardware decoding processor, or can be implemented by combining hardware and software modules in the decoding processor. The software module may be located in a storage medium located in the memory 202, and the processor 201 reads the information in the memory 202, and performs the steps of the media file processing method provided by the embodiment of the present invention in combination with hardware thereof.
Next, a method for processing a media file according to an embodiment of the present invention will be described. Fig. 3 is a flowchart illustrating a method for processing a media file according to an embodiment of the present invention, in some embodiments, the method for processing the media file may be implemented by a terminal, for example, by the terminal 400-1 in fig. 1, where the terminal 400-1 is provided with a media playing client, and with reference to fig. 1 and fig. 3, the method for processing the media file according to the embodiment of the present invention includes:
step 301: the media playing client receives a playing instruction of the media file.
In practical applications, the media playing client is used for playing media files (such as video files and audio files), and specifically may be a video playing client, an audio playing client, and the like. The user triggers a play instruction (e.g., a click operation) indicating the media file to be played based on the media playing client, so as to play the media file.
Step 302: and acquiring the media data of the media file based on the playing instruction.
In some embodiments, the packaging format of the media file is a streaming media format, that is, the media file can be decoded and played without complete downloading and additional transcoding, such as an FLV (Flash Video) file; in other embodiments, the encapsulated format of the Media file is a non-streaming format, that is, the Media file needs to be completely downloaded before being decoded and played, such as a window Media Video (WMV, windows Media Video) file, an MKV file format (MKV, MKV file format) file, and the like.
For different media file types, the media data may be at least one of video data and audio data, and in practical applications, the media file further includes metadata, i.e. data describing the media data, and the metadata carries media information ensuring that the media data is decoded correctly.
Step 303: the method comprises the steps of obtaining expansion data used for expanding the presentation form of the media file and configuration data corresponding to the expansion data and indicating that target expansion is achieved.
In some embodiments, the expansion of the media file presentation form may be to highlight key content of the media file, such as highlighting (bracketing) key episodes of the media file, such as highlighting puzzles (laughter points) in the video, bracketing emotions (puncta) in the video, and so forth; in other embodiments, the media file presentation format may be expanded to prompt switching of key information in the media file, such as prompting a scene change of a video, prompting a user to rest, simulating a performance, and the like.
The extended data will be explained. In some embodiments, the extension data may be at least one of audio data, video data, and picture data.
Illustratively, the extension data is audio data including at least one of: real voice data, synthetic voice data and non-voice data; the real voice data is collected/recorded real voice data, for example, for a certain movie, the sound of audiences in the movie playing process is collected, and the sound of the audiences is used as extension data; the synthesized voice data is synthesized virtual voice data simulating the voice of the user, such as synthesized voice data simulating a public character (such as a movie star); the non-human voice data includes collected animal sounds (such as cat and dog) and virtual sounds virtually realized through software, such as virtual applause, virtual laughter and the like.
In practical applications, the video data may be animation data highlighting key content of the video or prompting switching of key information of the video, such as simulating a funny plot in a video (media file) by using an animated character image to highlight a funny point of the video, or the video data may be still lifelike video data taking a real life as an authoring material and taking a real person as an expression object, such as advocating an environmentally-friendly plot in the video (media file) by using real environmental pollution video data.
The acquisition timing of the extension data will be explained. In some embodiments, after the user triggers the playing of the media file, the acquisition of the extension data is automatically realized, that is, the media playing client requests the server to acquire the extension data and the configuration data corresponding to the extension data while acquiring the media data; in other embodiments, the obtaining of the extension data is performed after being triggered by the user, for example, in the process of playing the media file, the user triggers the form extension instruction by clicking operation to instruct the media playing client to request the server to obtain the extension data and the configuration data corresponding to the extension data.
The manner of acquiring the extension data will be described. In some embodiments, the media playing client may obtain the extension data based on the file identification of the media file by: the media playing client determines the file identification of the media file, sends an extended data request carrying the file identification to the server, and receives the extended data corresponding to the file identification returned by the server. Taking a media file as an example of a video file, a media playing client determines that a video Identifier (ID) sends an extended data request carrying the video ID to a server, and receives extended data corresponding to the video ID and configuration data corresponding to the extended data returned by the server. Therefore, the extended data corresponds to a specific file identifier (such as a video ID), the relevance between the extended data and the currently played media file is ensured, the acceptance of the user on the extended data in the playing process of the media file is improved, and the click rate of the media file is improved.
In other embodiments, the media playing client may further obtain the extension data based on the user identifier by: and responding to the fact that the current user login state is logged in, the media playing client side obtains a corresponding user identification, determines corresponding social network information based on the user identification, and obtains the extended data of the associated media file based on the social network information. Taking a media file as a video file as an example, a media playing client acquires a user account of a current login user, determines a friend account associated with the user account based on the user account, namely determines a friend of the login user, and acquires extension data corresponding to the video file associated with the friend account based on the determined friend account; still taking a media file as an example of a video file, the media playing client acquires a user account of a current login user, determines a user account concerned by the user account based on the user account, that is, determines a user concerned by the login user, and acquires extension data corresponding to the video file associated with the concerned user account based on the concerned user account. Therefore, the extension data come from the extension data of friends of the current login user or the extension data of the concerned user, so that when the extension data is used for extending the presentation form of the media file, the watching/listening experience of the user is improved, and the viscosity of the user is enhanced.
In other embodiments, the media playing client may further obtain the extension data based on the user identifier by: the media playing client acquires the user identification of the registered user, the registered user corresponds to the media playing client, the extension data associated with the registered user is determined based on the user identification of the registered user, and the extension data associated with the media file is obtained by screening from the extension data associated with the registered user. Therefore, the extension data come from the extension data provided by the registered user in the current media playing client, the interaction enthusiasm of the user based on the media playing client is improved, and the user experience is good.
In other embodiments, the media playing client may further obtain the extension data based on the user portrait by: the media playing client acquires historical behavior data of a target user, determines a user portrait indicating interest classification of the target user based on the historical behavior data, and acquires extension data corresponding to the user portrait. Here, the target user may be a user currently logged in by the media playing client, and obtain, through the user representation, extension data provided by a user (such as a star) interested by the user, so as to enhance the interest level of the user in the media file presentation and improve the playing rate of the media file.
Step 304: and displaying the media data and the extended data in a mode of fusing the display form of the extended data and the display form of the media data based on the configuration data.
In some embodiments, the media playing client may implement fusion of the presentation form of the extension data and the presentation form of the media data, and presentation of the media data and the extension data by: and the media playing client fuses the expanded data and the media data based on the target expansion mode indicated by the configuration data to obtain fused media data, and displays the obtained fused media data through the playing window.
Here, the expansion method will be explained. In some embodiments, the expansion mode may include at least one of: audio fusion, small window fusion and barrage fusion; the audio fusion corresponding expansion data comprises audio data, and the audio data in the media file are subjected to audio fusion; the small window fusion corresponding expansion data comprises video data, and the video data is synchronously played with the media file in a small window form; the corresponding extended data of the bullet screen fusion comprises character data, and the character data is displayed through the bullet screen in the process of displaying the media data.
In other embodiments, the media playing client may further implement fusion of the presentation form of the extension data and the presentation form of the media data, and presentation of the media data and the extension data by: the media playing client displays the media data through the first playing window, and displays the extended data through the second playing window based on the configuration data, so that the display form of the extended data is fused with the display form of the media data. Taking the example that the extension data is audio data and the media data includes audio data, the audio data in the extension data and the audio data in the media data are played through two independent playing windows (two audio tracks), so that the fusion of the audio data is realized.
By applying the embodiment of the invention, the media data and the extended data are displayed in a mode of fusing the display form of the extended data and the display form of the media data, the playing form of the media file is extended, the highlighting of key contents in the media file or the prompt of key information switching in the media file is realized, meanwhile, due to the diversity of the types of the extended data, the display form of the media data is richer and more diverse, and the watching and/or listening experience of a user is improved.
Next, taking a media file as a video file (such as a series), for example, a processing method of the media file according to an embodiment of the present invention is described, and fig. 4 is a flowchart illustrating the processing method of the media file according to the embodiment of the present invention, in some embodiments, the processing method of the media file may be implemented by a terminal and a server in cooperation, for example, by the terminal 400-1 and the server 200 in fig. 1, the terminal 400-1 is provided with a media playing client, and with reference to fig. 1 and fig. 4, the processing method of the media file according to the embodiment of the present invention includes:
step 401: and the media playing client receives a playing instruction of the video file.
Step 402: and sending a video data acquisition request to the server based on the playing instruction.
In actual implementation, the media playing client sends a video data acquisition request carrying a video ID to the server based on a playing instruction triggered by a user, so as to acquire video data of the video file.
Step 403: and receiving the video data of the video file returned by the server.
Step 404: in the playing process of the video file, a form expansion instruction is received.
Fig. 5 is an interface schematic diagram of user-triggered form expansion provided in an embodiment of the present invention, and referring to fig. 5, when a user clicks a form expansion key presented in a play interface, a form expansion instruction corresponding to a currently played video file is triggered to instruct form expansion of the currently played video file, and a presentation form of expansion data is fused on the basis of presentation of the current video file.
Step 405: and sending a form expansion request to the server based on the form expansion instruction.
Here, in actual implementation, the media playing client sends a format extension request carrying a video ID to the server based on a format extension instruction triggered by a user, so as to obtain extension data for extending a presentation format of the video file and configuration data indicating that target extension is implemented.
Step 406: and receiving the extended data returned by the server and the configuration data corresponding to the extended data.
In some embodiments, the extension data is audio data corresponding to the video ID, and in practical applications, the audio data may be pre-collected/recorded voice of a specific person, such as pre-collected voice (e.g., laughing, crying, comment, etc.) fed back by a specific user while watching the video, pre-recorded voice of an animal, etc.; in practical applications, the audio data may also be synthesized to simulate the voice of a particular person, such as a pre-synthesized voice simulating a public character (star).
Correspondingly, the configuration data corresponding to the audio data is data indicating that audio fusion is implemented, and in practical applications, the configuration data may include: audio fusion mode (i.e., data encoding mode for obtaining fusion data), fusion time (i.e., time for performing audio fusion), fusion volume (volume corresponding to extension data and volume corresponding to video data in fusion data obtained by audio fusion, or volume ratio corresponding to extension data and volume corresponding to video data in fusion data obtained by audio fusion), and the like.
Step 407: and fusing the extended data and the video data based on the configuration data to obtain fused data.
In some embodiments, before the extension data is fused with the video data, the media playing client needs to cut the extension data so that the playing start point of the extension data is aligned with the playing start point of the video data, that is, the playing start point of the extension data is consistent with the playing start point of the video data.
In practical implementation, the media playing client performs audio fusion, such as mixing, on the extension data and the audio data in the video data based on the configuration data, so that the sound corresponding to the extension data exists as a background sound of the sound in the video.
Step 408: and displaying the fusion data through a playing window.
In actual implementation, the media playing client displays the video through the video window, and simultaneously plays the fusion data obtained after audio fusion through the audio player, so that the presentation of the extension data is fused with the presentation of the audio data in the video file.
Taking a media file as a video file (such as a movie) as an example, the method for processing the media file provided by the embodiment of the present invention is continuously described, fig. 6 is a schematic flow diagram of the method for processing the media file provided by the embodiment of the present invention, in some embodiments, the method for processing the media file may be implemented cooperatively by a terminal and a server, for example, implemented by the terminal 400-1 and the server 200 in fig. 1, the terminal 400-1 is provided with a media playing client, the media playing client includes an obtaining unit and a playing unit, the server includes a service module and a storage module, fig. 7 is a schematic diagram of cooperatively implementing media file processing by the terminal (front end) and the server (background) provided by the embodiment of the present invention, and with reference to fig. 1, fig. 6, and fig. 7, the method for processing the media file provided by the embodiment of the present invention includes:
step 501: and the media playing client receives a form expansion instruction triggered by a user through the acquisition unit in the process of playing the movie.
In practical application, the form expansion instruction is used for realizing sound effect enhancement of a played movie, and a user triggers and controls the form expansion instruction for realizing the sound effect enhancement of the movie by clicking a sound effect enhancement button displayed on a current movie playing interface.
Step 502: in response to the format extension instruction, the acquisition unit sends a format extension request carrying the movie ID to the server.
Step 503: the server analyzes the form extension instruction through the service module to obtain the movie ID.
Step 504: the service module acquires the audience sound file corresponding to the movie ID stored in the storage module based on the movie ID.
Here, in practical applications, the storage module stores in advance the audience sound file corresponding to the movie, for example, the collected sounds (such as laughter, crying, etc.) of the audience watching the movie during playing the movie in a certain cinema, and the audience sound file corresponds to the movie ID, and the movie ID is used as the retrieval index, which is convenient for the server to retrieve. The audience sound file comprises audience sound data and configuration data indicating the realization of audio fusion.
In practical applications, the service module may further obtain a subtitle file corresponding to the movie ID based on the movie ID.
Step 505: and the service module sends the acquired audience sound file to a playing unit of the media playing client.
Step 506: the playing unit performs sound mixing processing on the audience sound and the movie sound based on the audience sound file and plays the sound.
Fig. 8 is a schematic diagram illustrating sound effect enhancement of an online movie according to an embodiment of the present invention, and referring to fig. 8, a form extension function is turned on by a form extension instruction triggered by a user to enhance the sound effect of the played movie, so that a movie watching user can hear cinema sound accompanied by surround sound of audience sound, cinema atmosphere when the user watches the movie online is improved, and user experience is improved.
The software implementation of the media file processing apparatus provided in the embodiment of the present invention is explained. Fig. 9 is a schematic diagram of a software structure of a media file processing apparatus according to an embodiment of the present invention, and referring to fig. 9, a media file processing apparatus 90 according to an embodiment of the present invention includes:
a receiving unit 91 for receiving a play instruction of a media file;
an obtaining unit 92, configured to obtain media data of the media file based on the playing instruction;
the media file expansion method comprises the steps of obtaining expansion data used for expanding the presentation form of the media file and configuration data corresponding to the expansion data and indicating to achieve target expansion;
a playing unit 93, configured to display the media data and the extended data in a manner of fusing a display form of the extended data and a display form of the media data based on the configuration data.
In some embodiments, the obtaining unit is further configured to obtain at least one of audio data, video data, and picture data for highlighting key content of the media file;
wherein the audio data comprises at least one of: real voice data, synthetic voice data, non-voice data.
In some embodiments, the obtaining unit is further configured to obtain at least one of audio data, video data, and picture data for prompting key information switching of the media file.
In some embodiments, the obtaining unit is further configured to determine a file identifier of the media file;
and sending an extended data request carrying the file identifier;
and receiving the returned extension data corresponding to the file identifier.
In some embodiments, the obtaining unit is further configured to obtain a corresponding user identifier in response to that the current user login status is logged in;
and determining corresponding social network information based on the user identification;
and acquiring the extended data associated with the media file based on the social network information.
In some embodiments, the obtaining unit is further configured to obtain a user identifier of a registered user, where the registered user corresponds to a playing client of the media file;
and determining extension data associated with the registered user based on the user identifier of the registered user;
and screening the determined expansion data to obtain the expansion data related to the media file.
In some embodiments, the obtaining unit is further configured to receive a form extension instruction corresponding to the media file during the playing of the media file;
and acquiring extension data for extending the presentation form of the media file in response to the form extension instruction.
In some embodiments, the obtaining unit is further configured to obtain historical behavior data of the target user;
and determining, based on the historical behavior data, a user representation indicative of an interest classification of the target user;
and obtaining the extension data corresponding to the user representation.
In some embodiments, the playing unit is further configured to fuse the extension data and the media data based on a target extension manner indicated by the configuration data to obtain fused media data;
and displaying the obtained fusion media data through a playing window.
In some embodiments, the playing unit is further configured to present the media data through a first playing window;
and displaying the extended data through a second playing window based on the configuration data, so that the display form of the extended data is fused with the display form of the media data.
The embodiment of the invention has the following beneficial technical effects:
1. the media data and the extended data are displayed in a mode of fusing the display form of the extended data and the display form of the media data, so that the expansion of the playing form of the media file is realized, the display form of the media data is richer and more diverse, and the watching and/or listening experience of a user is improved;
2. the acquired extension data has the function of highlighting the key content of the media file or prompting the switching of the key information, so that the key content of the media file is highlighted or the switching of the key information of the media file is prompted in the process of showing the media data and the extension data, and a user can accurately master the key information of the media file;
3. the media playing client acquires the extended data of the associated media file based on the social network information, so that the extended data comes from the social network and the user viscosity is enhanced;
4. and the media playing client filters the extended data associated with the media file from the extended data associated with the registered user. Therefore, the interactive enthusiasm of the user based on the media playing client is improved;
5. the media playing client acquires the extended data based on the user portrait, so that the interestingness of the user in the media file display is enhanced, and the playing rate of the media file is improved.
An embodiment of the present invention further provides a device for processing a media file, where the device includes:
a memory for storing an executable program;
the processor is configured to implement the method for processing the media file according to the embodiment of the present invention when executing the executable program stored in the memory.
The embodiment of the invention also provides a storage medium which stores an executable program, and the executable program is executed by a processor to realize the media file processing method provided by the embodiment of the invention.
Here, it should be noted that: similar to the above description of the method, the beneficial effects of the method are described, and no further description is given, for the technical details not disclosed in the media file processing apparatus according to the embodiment of the present invention, please refer to the description of the method embodiment of the present invention.
All or part of the steps of the embodiments may be implemented by hardware associated with program instructions, and the program may be stored in a computer-readable storage medium, and when executed, performs the steps including the method embodiments; and the aforementioned storage medium includes: various media that can store program codes, such as a removable Memory device, a Random Access Memory (RAM), a Read-Only Memory (ROM), a magnetic disk, and an optical disk.
Alternatively, the integrated unit of the present invention may be stored in a computer-readable storage medium if it is implemented in the form of a software functional module and sold or used as a separate product. Based on such understanding, the technical solutions of the embodiments of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the methods described in the embodiments of the present invention. And the aforementioned storage medium includes: a removable storage device, a RAM, a ROM, a magnetic or optical disk, or various other media that can store program code.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the appended claims.

Claims (12)

1. A method for processing a media file, the method comprising:
receiving a playing instruction of a media file;
acquiring media data of the media file based on the playing instruction;
receiving a form expansion instruction corresponding to the media file in the playing process of the media file;
responding to the form extension instruction, acquiring extension data and configuration data corresponding to the extension data and indicating to achieve target extension, wherein the extension data is used for highlighting key contents of the media file or prompting switching of key information in the media file, and the extension data comprises at least one of audio data, video data and picture data;
fusing the extended data and the media data based on a target extension mode indicated by the configuration data to obtain fused media data, wherein the target extension mode comprises at least one of the following modes: audio fusion, small window fusion and barrage fusion, wherein the audio fusion is to perform audio fusion on audio data in the extended data and audio data in the media file, the small window fusion is to play video data in the extended data synchronously with the media file in a small window mode, and the barrage fusion is to display text data in the extended data through a barrage in the process of displaying the media data;
and displaying the obtained fusion media data through a playing window.
2. The method of claim 1, wherein the obtaining the extension data comprises:
acquiring at least one of audio data, video data and picture data for highlighting key content of the media file;
wherein the audio data comprises at least one of: real voice data, synthetic voice data and non-voice data.
3. The method of claim 1, wherein the obtaining the extension data comprises:
and acquiring at least one of audio data, video data and picture data for prompting the key information switching of the media file.
4. The method of claim 1, wherein the obtaining the extension data comprises:
determining a file identifier of the media file;
sending an extended data request carrying the file identifier;
and receiving the returned extension data corresponding to the file identification.
5. The method of claim 1, wherein the obtaining the extension data comprises:
responding to the fact that the current user login state is logged in, and acquiring a corresponding user identifier;
determining corresponding social network information based on the user identification;
and acquiring the extended data associated with the media file based on the social network information.
6. The method of claim 1, wherein the obtaining the extension data comprises:
acquiring a user identifier of a registered user, wherein the registered user corresponds to a playing client of the media file;
determining extension data associated with the registered user based on the user identification of the registered user;
and screening the determined expansion data to obtain the expansion data related to the media file.
7. The method of claim 1, wherein the obtaining the extension data comprises:
acquiring historical behavior data of a target user;
determining, based on the historical behavior data, a user representation indicative of an interest classification of the target user;
obtaining the extension data corresponding to the user representation.
8. The method of claim 1, wherein the method further comprises:
displaying the media data through a first playing window;
and displaying the extended data through a second playing window based on the configuration data, so that the display form of the extended data is fused with the display form of the media data.
9. An apparatus for processing a media file, the apparatus comprising:
the receiving unit is used for receiving a playing instruction of the media file;
an obtaining unit, configured to obtain media data of the media file based on the play instruction;
the acquiring unit is further configured to receive a form extension instruction corresponding to the media file in the playing process of the media file; responding to the form extension instruction, acquiring extension data and configuration data corresponding to the extension data and indicating to achieve target extension, wherein the extension data is used for highlighting key contents of the media file or prompting switching of key information in the media file, and the extension data comprises at least one of audio data, video data and picture data;
a playing unit, configured to fuse the extended data and the media data based on a target extension manner indicated by the configuration data to obtain fused media data, and display the obtained fused media data through a playing window, where the target extension manner includes at least one of: audio fusion, small window fusion and barrage fusion, wherein the audio fusion is to perform audio fusion on audio data in the extended data and audio data in the media file, the small window fusion is to play video data in the extended data synchronously with the media file in a small window mode, and the barrage fusion is to display text data in the extended data through a barrage in the process of displaying the media data.
10. The apparatus of claim 9,
the acquiring unit is further configured to acquire at least one of audio data, video data, and picture data for highlighting key content of the media file;
wherein the audio data comprises at least one of: real voice data, synthetic voice data, non-voice data.
11. An apparatus for processing a media file, the apparatus comprising:
a memory configured to store a processing program for a media file;
a processor configured to run the program, wherein the program when executed performs the method of processing a media file of any of claims 1 to 8.
12. A computer-readable storage medium comprising a stored program, wherein the program when executed performs the method of processing a media file of any of claims 1 to 8.
CN201910340609.6A 2019-04-25 2019-04-25 Media file processing method and device and storage medium Active CN110166801B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910340609.6A CN110166801B (en) 2019-04-25 2019-04-25 Media file processing method and device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910340609.6A CN110166801B (en) 2019-04-25 2019-04-25 Media file processing method and device and storage medium

Publications (2)

Publication Number Publication Date
CN110166801A CN110166801A (en) 2019-08-23
CN110166801B true CN110166801B (en) 2022-12-20

Family

ID=67640223

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910340609.6A Active CN110166801B (en) 2019-04-25 2019-04-25 Media file processing method and device and storage medium

Country Status (1)

Country Link
CN (1) CN110166801B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114816608A (en) * 2021-01-29 2022-07-29 腾讯科技(深圳)有限公司 Media file playing method and device, electronic equipment and storage medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103517158B (en) * 2012-06-25 2017-02-22 华为技术有限公司 Method, device and system for generating videos capable of showing video notations
CN104683703B (en) * 2013-11-27 2018-12-21 腾讯科技(北京)有限公司 A kind of video broadcasting method and device
CN105828100A (en) * 2016-03-21 2016-08-03 乐视网信息技术(北京)股份有限公司 Audio and video files simultaneous playing method, device and system
CN107682713B (en) * 2017-04-11 2020-11-03 腾讯科技(北京)有限公司 Media file playing method and device
CN107690084B (en) * 2017-04-11 2019-11-08 腾讯科技(北京)有限公司 The playing method and device of media file
CN109379636B (en) * 2018-09-20 2022-06-17 京东方科技集团股份有限公司 Bullet screen processing method, device and system

Also Published As

Publication number Publication date
CN110166801A (en) 2019-08-23

Similar Documents

Publication Publication Date Title
CN108066986B (en) Streaming media determining method and device and storage medium
CN108184144B (en) Live broadcast method and device, storage medium and electronic equipment
US20190373322A1 (en) Interactive Video Content Delivery
WO2017121362A1 (en) Information processing method, and server and computer storage medium
WO2018028533A1 (en) Media information publishing method, terminal, server, system and storage medium
CN105635764B (en) Method and device for playing push information in live video
CN108900854B (en) Live broadcast microphone room switching method, storage medium, equipment and system
US20070250775A1 (en) Methods, systems, and computer program products for providing hyperlinked video
WO2016004240A1 (en) Interactive distributed multimedia system
TW200921539A (en) Embedded video player advertisement display
US20130198321A1 (en) Content associated with primary content
CN111436008A (en) Method, device, equipment and storage medium for displaying information on video
CN111444415B (en) Barrage processing method, server, client, electronic equipment and storage medium
JP2019091014A (en) Method and apparatus for reproducing multimedia
CN103686454A (en) Information acquisition method and apparatus
CN111131848A (en) Video live broadcast data processing method, client and server
CN111512635A (en) Method and system for selectively skipping media content
CN111881395A (en) Page presenting method, device, equipment and computer readable storage medium
CN111800668A (en) Bullet screen processing method, device, equipment and storage medium
JP2017538328A (en) Promotion information processing method, apparatus, device, and computer storage medium
CN112801684A (en) Advertisement playing method and device
CN115190366A (en) Information display method and device, electronic equipment and computer readable medium
US11838576B2 (en) Video distribution system, method, computing device and user equipment
CN110166801B (en) Media file processing method and device and storage medium
CN114025188A (en) Live broadcast advertisement display method, system, device, terminal and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant