WO2022228557A1 - 剪辑模板搜索方法及装置 - Google Patents
剪辑模板搜索方法及装置 Download PDFInfo
- Publication number
- WO2022228557A1 WO2022228557A1 PCT/CN2022/090348 CN2022090348W WO2022228557A1 WO 2022228557 A1 WO2022228557 A1 WO 2022228557A1 CN 2022090348 W CN2022090348 W CN 2022090348W WO 2022228557 A1 WO2022228557 A1 WO 2022228557A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dimension
- template
- target
- multimedia resource
- editing
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 85
- 239000000463 material Substances 0.000 claims abstract description 14
- 238000012216 screening Methods 0.000 claims description 51
- 230000000007 visual effect Effects 0.000 claims description 23
- 238000004590 computer program Methods 0.000 claims description 18
- 238000004364 calculation method Methods 0.000 claims description 17
- 230000002708 enhancing effect Effects 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 8
- 238000012545 processing Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 230000007704 transition Effects 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 4
- 238000004891 communication Methods 0.000 description 4
- 238000001914 filtration Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 239000002131 composite material Substances 0.000 description 1
- 235000009508 confectionery Nutrition 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/435—Filtering based on additional data, e.g. user or group profiles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/438—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/483—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
Definitions
- the present disclosure relates to the field of Internet technologies, and in particular, to a method and device for searching for clip templates.
- APP application
- the application usually provides a wealth of clip templates, and users can use the clip templates and select their favorite photos or videos to obtain a composite video.
- how to quickly and accurately find the clip template desired by the user is an urgent problem to be solved at present.
- the present disclosure provides a clip template search method and apparatus.
- an embodiment of the present disclosure provides a method for searching for a clip template, including:
- the search result is used to indicate whether a target editing template matching the first multimedia resource is found, and the target editing template is used to indicate that the multimedia material to be edited is edited into a second multimedia material according to the target editing method.
- the target editing method is the editing method adopted by the first multimedia resource.
- the searching according to the first multimedia resource to obtain search results includes:
- the identification result of the candidate clip template in the target dimension is obtained according to the first multimedia resource and the feature of the candidate clip template in the target dimension.
- the method before acquiring the search result according to the identification result of the candidate clip template in the target dimension, the method further includes:
- the at least one editing template includes the candidate editing template.
- the target dimension includes one or more of: a music style dimension, an audio fingerprint dimension, a video size dimension, a video segment feature dimension, and a visual effect dimension.
- the target dimension includes a plurality of the music genre dimension, the audio fingerprint dimension, the video size dimension, the video segment feature dimension, and the visual effects dimension;
- the obtaining the search result according to the identification result of the candidate clip template in the target dimension includes:
- the search result is acquired according to the weighted calculation result corresponding to the candidate clip template.
- the determining of the candidate editing template according to the characteristics of the first multimedia resource and the at least one editing template on the target dimension respectively includes:
- the characteristics of the first multimedia resource and the first screening result respectively in the current dimension obtain the first recognition result of each editing template in the first screening result in the current dimension respectively;
- the second screening result includes: one or more editing templates, the first screening result
- the initial state includes: the at least one clip template
- the obtaining the first multimedia resource specified by the user includes: obtaining a target link input by the user, and analyzing the target link to obtain the first multimedia resource.
- the method further includes sending the search results to the user.
- an embodiment of the present disclosure provides a clip template search device, including:
- an acquisition module for acquiring the first multimedia resource specified by the user
- a search module configured to perform a search according to the first multimedia resource, and obtain a search result; wherein, the search result is used to indicate whether a target clip template matching the first multimedia resource is found, and the The target editing template is used to indicate that the multimedia material to be edited is edited into a second multimedia resource according to a target editing method, and the target editing method is the editing method adopted by the first multimedia resource.
- embodiments of the present disclosure provide an electronic device, including: a memory, a processor, and a computer program;
- the memory is configured to store the computer program
- the processor is configured to execute the computer program to implement the method of any one of the first aspects.
- an embodiment of the present disclosure provides a readable storage medium, including: a program
- an embodiment of the present disclosure further provides a program product, including: a computer program, where the computer program is stored in a readable storage medium, and at least one processor of an electronic device can read from the readable storage medium The computer program, executed by the at least one processor, causes the electronic device to implement the method according to any one of the first aspects.
- Embodiments of the present disclosure provide a method and apparatus for searching for a clip template, wherein the method includes: in a video editing scenario, the server device obtains a first multimedia resource specified by a user, and searches according to the first multimedia resource specified by the user.
- the target editing template is used to instruct the multimedia material to be edited to be edited into the second multimedia resource according to the target editing method.
- searching according to the first multimedia resource improves the accuracy of search results, can better meet the needs of users for video creation, and can also improve the utilization rate of the target editing template.
- FIG. 1 is a schematic diagram of an application scenario of a method for searching for a clip template provided by an embodiment of the present disclosure
- FIG. 2 is a flowchart of a clip template search method provided by an embodiment of the present disclosure
- FIG. 3 is a schematic structural diagram of a search model provided by an embodiment of the present disclosure.
- FIG. 4 is a flowchart of a method for searching for a clip template provided by another embodiment of the present disclosure
- FIG. 5 is a schematic structural diagram of a search model provided by another embodiment of the present disclosure.
- FIG. 6 is a flowchart of a method for searching for a clip template provided by another embodiment of the present disclosure.
- FIG. 7 is a flowchart of a clip template search method provided by another embodiment of the present disclosure.
- 8A to 8K are schematic diagrams of human-computer interaction interfaces provided by the present disclosure.
- FIG. 9 is a schematic structural diagram of a clip template search apparatus provided by an embodiment of the present disclosure.
- FIG. 10 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure.
- the existing APP usually supports searching for editing templates by keywords.
- the user may not be able to obtain the correct keyword, and the clip template found by the keyword input by the user is not the clip template that the user wants, so that the user may not want to continue video creation.
- the present disclosure provides a clip template search method, the core idea of which is: by acquiring the first multimedia resource specified by the user, and analyzing the first multimedia resource and the candidate clip template, so as to quickly
- the target editing template that the user wants is located from the candidate editing templates the accuracy of the search results is improved, the user's needs for video creation can be better met, and the utilization rate of the target editing template can also be improved.
- FIG. 1 is a schematic diagram of an application scenario of a method for searching for a clip template provided by an embodiment of the present disclosure.
- the clip template search method provided in this embodiment can be applied to the scene shown in FIG. 1 .
- the scenario includes: a server device 101 and a terminal device 102.
- the server device 101 and the terminal device 102 may be connected through a wired or wireless network.
- the server device 101 may be implemented in any software and/or hardware manner.
- the server device 101 may be a server, and the server may be an independent server, a server cluster composed of multiple independent servers, or a cloud server.
- the server device 101 may also be a software program integrated in the electronic device.
- the software program is executed by at least one processor of the electronic device, the technical solution executed by the server device in the clip template search method provided by the embodiment of the present disclosure can be executed.
- the server device can interact with one or more terminal devices at the same time, and send the same or different data to the terminal devices.
- the terminal device 102 may be implemented in any software and/or hardware manner.
- the terminal device 102 may be, but not limited to, a notebook computer, a desktop computer, a smart phone, a portable terminal device, a wearable device, a personal digital assistant (personal digital assistant, PDA) and other devices. No restrictions apply.
- the terminal device 102 may also be a software program integrated in the electronic device. When the software program is executed by the processor of the electronic device, the technical solution executed by the terminal device in the clip template search method provided by the embodiment of the present disclosure may be executed.
- FIG. 1 exemplarily shows a scenario in which one server device interacts with one terminal device.
- the server device can interact with more terminal devices in parallel.
- FIG. 2 is a flowchart of a method for searching a clip template according to an embodiment of the present disclosure.
- the execution body of this embodiment may be a server device. Referring to Figure 2, this embodiment includes:
- the server device acquires the first multimedia resource designated by the user through the terminal device.
- the first multimedia resource may be an audio file or a video file.
- the first multimedia resource may also be a multimedia resource (eg, a short video) obtained by performing video creation according to the target editing template.
- a multimedia resource eg, a short video
- the first multimedia resource may be actively reported by the user through the terminal device.
- it may also be specified by the user in other ways, for example, the user reports the target link through the terminal device, and the server device obtains the target link by parsing.
- multiple application programs are installed on the terminal device, wherein the multiple application programs include: a video editing application program and other application programs.
- the target link may be copied by the user from any one of the above other applications and pasted into the video editing application.
- the above other applications can be, but are not limited to, music applications, video applications, social applications, and so on.
- the target link may also be manually input by the user to the video editing application installed on the terminal device.
- the present disclosure does not limit the source of the target link and the manner in which the terminal device obtains the target link.
- the video editing application After the video editing application obtains the target link, it parses the target link, obtains the webpage corresponding to the target link, and obtains the uniform resource locator (URL) of the pointed first multimedia resource from the webpage. ;
- the terminal device accesses and downloads the first multimedia resource according to the URL of the first multimedia resource; Afterwards, the terminal device through the established communication link with the server device, the first multimedia resource obtained The resource is uploaded to the server device.
- URL uniform resource locator
- the video editing application program when it obtains the target link, it can send the target link to the server device through the established communication link between the terminal device and the server device; after the server device receives the target link, The target link is parsed, the webpage corresponding to the target link is obtained, and the URL of the pointed first multimedia resource is obtained from the webpage; the terminal device accesses and downloads the first multimedia resource according to the URL of the first multimedia resource.
- the purpose of this step is to search for a target editing template that is the same as the editing method used by the first multimedia resource.
- the search result is used to indicate whether a target clip template matching the first multimedia resource is searched.
- a template library deployed on the server device is searched to obtain a search result.
- the template library deployed on the server device includes: at least one clip template.
- Each editing template is used to provide a preset editing mode, and the multimedia material to be edited selected or imported by the user can be edited into a new multimedia resource according to the preset editing mode provided by the editing template.
- the server device may obtain the search result by performing an analysis on the target dimension on the first multimedia resource and each clip template in the template library.
- the target dimension may include one or more dimensions of music style dimension, audio fingerprint dimension, video size dimension, video segment feature dimension, and visual effect dimension.
- a trained search model may be pre-deployed on the server device, and after the server device acquires the first multimedia resource specified by the user, the search model may be used to perform the above search.
- the search model may be used to perform the above search.
- other methods or algorithms can also be used to perform the above search.
- the server device obtains the first multimedia resource specified by the user, and searches the template library according to the first multimedia resource specified by the user, and obtains the same value as the first multimedia resource.
- Matching target clip template In this solution, searching through the first multimedia resource improves the accuracy of search results, can better meet the needs of users for video creation, and can also improve the utilization rate of the target editing template.
- the search model searches according to the first multimedia resource and each clip template included in the template library will be described in detail respectively in two cases that the first multimedia resource is an audio file and a video file.
- server device performing the search and the server device storing the template library may be the same device or different devices, which are not limited in this embodiment of the present disclosure.
- Case 1 The first multimedia resource is an audio file
- FIG. 3 is a schematic structural diagram of a search model provided by an embodiment of the present disclosure
- FIG. 4 is a flowchart of a method for searching a clip template provided by another embodiment of the present disclosure.
- the search model 300 includes: a music style identification sub-model 301 and an audio fingerprint identification sub-model 302 .
- the music style recognition sub-model 301 is used to output the first recognition result corresponding to each clip template in the music style dimension according to the first multimedia resource and the audio files contained in each template video, and the first recognition result is used to indicate the first recognition result. Similarity in musical style between a multimedia resource and an audio file contained in a template video.
- the music styles can be pre-divided into multiple styles, for example, the music styles include: sad, quiet, popular, cheerful, relaxed, sweet, happy and so on.
- the specific classification of music styles is not limited in the embodiments of the present disclosure.
- the music style recognition sub-model 301 obtains the first feature vector corresponding to the first multimedia resource and the second feature vector of the audio file included in the clip template through a specific algorithm. Then, calculate the distance between the first feature vector and the second feature vector (for example: Euclidean distance, of course, it can also be the distance obtained by other algorithms), and obtain the first recognition corresponding to the clip template according to the distance obtained by calculation result.
- Euclidean distance for example: Euclidean distance, of course, it can also be the distance obtained by other algorithms
- the audio fingerprint identification sub-model 302 is used to output the second identification result corresponding to each clip template in the audio fingerprint dimension according to the first multimedia resource and the audio files contained in each template video, and the second identification result is used to indicate the multimedia The similarity of the audio fingerprint between the file and the audio file contained in the template video.
- the audio fingerprinting sub-model 302 specifically adopts an audio fingerprinting technology (audio fingerprinting technology) to analyze the multimedia files and the audio files contained in each clip template.
- audio fingerprinting technology refers to using a specific algorithm to extract the data features of the audio files that need to be identified, such as sound spectrum features, spectral features, etc. Compared.
- the clip template search method provided by this embodiment includes:
- the server device obtains the target audio file.
- the target audio file in this embodiment is the first multimedia resource.
- the specific implementation manner of acquiring the target audio file by the server device may refer to the description in the embodiment shown in FIG. 2 , which will not be repeated here.
- the target dimensions in this step exemplarily include: a music style dimension and an audio fingerprint dimension.
- the duration of the first multimedia resource (in this embodiment, the first multimedia resource is the target audio file) is too long, and the server device may
- the audio files of the clip templates contained in the resource and template library are respectively sliced, and the first audio sub-file corresponding to the first multimedia resource and the first audio sub-file corresponding to the audio file contained in each of the clip templates are obtained. Two audio subfiles.
- the server device may perform slicing processing at fixed intervals, so that the audio durations of each first audio sub-file and each second audio sub-file are consistent, which is convenient for the music style identification sub-model and the audio fingerprint identification sub-model segment by segment analysis.
- the durations of the audio files included in each editing template may not be exactly the same, the duration of the audio files included in some editing templates is longer, and the duration of the audio files included in some editing templates is shorter. If the audio file included in the clip template has a long duration, slice processing can be performed at the above-mentioned fixed interval. If the duration of the audio files included in the clip template satisfies the above-mentioned fixed interval, the slicing process may not be performed. Whether to perform slicing processing can be flexibly set according to actual needs.
- the first audio sub-file corresponding to the first multimedia resource and the second audio sub-file corresponding to the audio files contained in each of the clip templates can be input into the music style recognition sub-model and audio fingerprint respectively.
- the sub-model is identified, so as to obtain the first identification result and the second identification result corresponding to each clip template output by the music style identification sub-model.
- the search result is acquired according to the first recognition result and the second recognition result corresponding to each candidate clip template respectively.
- the music style recognition sub-model and the audio fingerprint recognition sub-model may perform recognition tasks in parallel, or may perform recognition tasks in sequence.
- the candidate clipping template is a part of the at least one clipping template
- the method further includes: S402', determining the candidate editing template according to the target audio file and the characteristics of at least one editing template in the target dimension respectively.
- the number of candidate clip templates may be one or multiple.
- the “at least one editing template” in this step may include part of the editing templates in the template library, and may also include all the editing templates in the template library.
- the at least one editing template may be determined according to factors such as the release time, usage, and collection of the editing template; alternatively, it may be determined randomly; or determined by any other means.
- a second filtering result corresponding to the musical style dimension is obtained.
- the second filtering result corresponding to the music style dimension includes one or more editing templates.
- the first audio sub-file corresponding to the first multimedia resource and the second audio sub-file corresponding to each clip template included in the second screening result corresponding to the music style dimension are input into the audio fingerprint identification sub-model to obtain the music style.
- the second identification result corresponding to each clip template included in the second screening result corresponding to the dimension is input into the audio fingerprint identification sub-model to obtain the music style.
- the first filter result can be obtained according to the filter condition and the second identification result corresponding to each clip template included in the second filter result corresponding to the music style dimension.
- the editing template included in the first screening result is the aforementioned candidate editing template.
- the second screening result corresponding to the audio fingerprint dimension is obtained.
- the second screening result corresponding to the audio fingerprint dimension includes one or more editing templates.
- the first audio sub-file corresponding to the first multimedia resource and the second audio sub-file corresponding to each clip template included in the second screening result corresponding to the audio fingerprint dimension are input into the music style recognition sub-model to obtain the audio fingerprint.
- the first identification result corresponding to each clip template included in the second screening result corresponding to the dimension is input into the music style recognition sub-model to obtain the audio fingerprint.
- the first filter result can be obtained according to the filter condition and the second recognition result corresponding to each clip template included in the second filter result corresponding to the audio fingerprint dimension.
- the clip template included in the screening result is the aforementioned candidate clip template.
- the filter condition corresponding to the music style dimension or the audio fingerprint dimension may be empty.
- Obtaining the search result according to the first recognition result and the second recognition result corresponding to each candidate editing template can be achieved in the following manner:
- the weighting calculation result corresponding to each candidate editing template is obtained according to the first identification result and the second identification result corresponding to each candidate editing template, the respective weight coefficients corresponding to the music style dimension and the audio fingerprint dimension.
- the respective weight coefficients corresponding to the music style dimension and the audio fingerprint dimension can be flexibly configured according to requirements.
- the pre-trained search model is first used to analyze the audio file specified by the user (ie, the first multimedia resource) and the audio files contained in each clip template in the template library in the dimension of music style and audio fingerprint. , and comprehensively evaluates whether the clip template is the target clip template the user wants to find according to the analysis results in the two dimensions, which ensures the accuracy of the search results.
- the above-mentioned search task is performed through the search model, which can improve the search efficiency.
- the first multimedia resource is a video file
- FIG. 5 is a schematic structural diagram of a search model provided by an embodiment of the present disclosure
- FIG. 6 is a flowchart of a method for searching a clip template provided by another embodiment of the present disclosure.
- the search model 500 includes five sub-models, namely: a music style recognition sub-model 501, a video size recognition sub-model 502, a video segment feature recognition sub-model 503, a visual effect recognition sub-model 504, and an audio fingerprint recognition. Submodel 505.
- the music style identification sub-model 501 included in the search model 500 provided by the embodiment shown in FIG. 5 is similar to the music style sub-model 301 included in the search model 300 provided by the embodiment shown in FIG. 3;
- the audio fingerprint identification sub-model 505 included in the search model 500 is similar to the audio identification sub-model 302 included in the search model 300 provided by the embodiment shown in FIG. 3; for details, please refer to the detailed description in the embodiment shown in FIG. 3, which is not repeated here. Repeat.
- the video size recognition sub-model 502 is used to output the third recognition result corresponding to each editing template in the video size dimension according to the size feature of the first multimedia resource and the size feature of the editing template. It is used to indicate the similarity of the video size between the first multimedia resource and the clip template.
- the above-mentioned size features include: duration and/or aspect ratio of video frames.
- the size feature of the first multimedia resource includes: the duration of the multimedia file and/or the aspect ratio of the video frames included in the multimedia file.
- the size characteristics of the clip template include: the duration of the clip template and/or the aspect ratio of the video frame of the clip template.
- the aspect ratio of the video frame may be obtained by dividing the length of the video frame by the width of the video frame, or may be obtained by dividing the width of the video frame by the length of the video frame.
- the video clip feature identification sub-model 503 is used to output the fourth recognition result corresponding to the clip template on the feature dimension of the video clip according to the video clip feature of the first multimedia resource and the video clip feature of the clip template, and the fourth recognition result is used for Indicates the similarity between the video segment of the first multimedia resource and the video segment of the clip template.
- the video segment feature identification sub-model 503 segments the first multimedia resource according to the transition moment of the first multimedia resource, and obtains a plurality of first video sub-segments corresponding to the first multimedia resource; According to the transition moment of each editing template, the editing template is segmented to obtain a plurality of second video sub-segments corresponding to each editing template.
- the video segment feature identification sub-model 503 obtains the first video sub-segment corresponding to each video template file according to the sequence of the first video sub-segment, the sequence of the second video sub-segment, and the duration, transition mode and other characteristics of the video sub-segment. 4. Identification results.
- the visual effect recognition sub-model 504 is used for outputting the fifth recognition result corresponding to the editing template in the visual effect dimension according to the visual effect of the first multimedia resource and the visual effect of the editing template, and the fifth recognition result is used to indicate the first multiple recognition result. How similar the visuals of the media asset are to the visuals of the clip template.
- the visual effect recognition sub-model 504 specifically identifies one or more of the sticker material style, sticker material size, text material style, text material size, filter effect, etc. used by the first multimedia resource and the clip template respectively. Multiple items are obtained to obtain the fifth recognition result in the dimension of visual effects.
- the clip template search method provided by this embodiment includes:
- the server device obtains the target video file.
- the above-mentioned target video file is the first multimedia resource.
- the specific implementation manner of acquiring the target video file by the server device may refer to the description in the embodiment shown in FIG. 2 , which will not be repeated here.
- the target dimensions include: a music style dimension, an audio fingerprint dimension, a video size dimension, a video clip feature dimension, and a visual effect dimension.
- the server device may perform slice processing on the audio files of the clip templates contained in the first multimedia resource and the template library, respectively, to obtain the first multimedia resource.
- the first multimedia resource and the at least one editing template can be input into the five sub-models included in the search model 500, respectively, and the first recognition results corresponding to the editing templates output by the five sub-models can be obtained to The fifth identification result.
- the search result is acquired according to the first identification result to the fifth identification result corresponding to each candidate editing template respectively.
- the candidate clipping template is a part of the at least one clipping template
- the method further includes: S602', determining the candidate editing template according to the target video file and the characteristics of at least one editing template in the target dimension respectively.
- the number of candidate clip templates may be one or multiple.
- the “at least one editing template” in this step may include part of the editing templates in the template library, and may also include all the editing templates in the template library.
- the at least one editing template may be determined according to factors such as the release time, usage, and collection of the editing template; alternatively, it may be determined randomly; or determined by any other means.
- S602' may include the following steps:
- Step a Determine the current dimension according to the priority order of the above five sub-models.
- the priority order of the five sub-models included in the search model 500 is the priority order of each dimension corresponding to the target dimension, and the current dimension is determined according to the order of priorities of the five sub-models included in the search model 500 from high to bottom.
- Step b Acquire a first identification result of each clip template in the first screening result in the current dimension according to the first multimedia resource and the characteristics of the first screening result in the current dimension respectively.
- the first screening result is the screening result corresponding to the previous dimension.
- Step c obtaining a second screening result corresponding to the current dimension according to the first identification result and the screening condition corresponding to the current dimension;
- the second screening result includes: one or more editing templates,
- the An initial state of a screening result includes: the at least one clip template.
- Step d determining that the second screening result is the first screening result.
- step a determines that the editing template included in the second screening result corresponding to the last current dimension is the candidate editing template.
- the filter conditions corresponding to some dimensions can be configured to be empty.
- the priority order of the five sub-models included in the search model is: music style recognition sub-model > video size recognition sub-model > video segment feature recognition sub-model > audio fingerprint recognition sub-model > visual effect recognition sub-model.
- the filter condition corresponding to the audio fingerprint dimension is empty, then directly input the audio file contained in the first multimedia resource and the audio file of the clip template contained in the second filtering result corresponding to the dimension of the video clip into the audio fingerprint identifier. in the model. That is to say, the clip template included in the second screening result corresponding to the feature dimension of the video clip is the same as the clip template included in the second screening result corresponding to the audio fingerprint dimension.
- the priority order of each sub-model included in the search model can be flexibly configured, and is not limited to the above example.
- the music style identification sub-model and the audio fingerprint identification sub-model included in the search model analyze the audio files included in the multimedia files and the audio files included in the clip templates, and the audio files included in the multimedia files and the clip templates include audio files. If there is a duration process, the method in the embodiment shown in FIG. 4 can be used to perform slice processing on the audio file with an excessively long duration. For details, refer to the description in the embodiment shown in FIG. 4 , which will not be repeated here.
- Obtaining the search result according to the first identification result to the fifth identification result corresponding to each of the editing templates can be achieved in the following manner:
- the music style dimension, the video size dimension, the video segment feature dimension, the audio fingerprint dimension and the visual effect dimension corresponding to each editing template respectively obtain the corresponding corresponding to each editing template.
- the search result according to the weighted calculation result corresponding to the candidate editing template and the second preset threshold; wherein, if the highest score in the weighted calculation result corresponding to the candidate editing template is greater than the second preset threshold, it is determined that the template is in the template.
- the target editing template matching the target multimedia is searched in the library, and the target editing template is the candidate editing template corresponding to the highest score in the weighted calculation result; if the highest score in the corresponding weighted calculation result of the candidate editing template is less than or equal to the If the preset threshold is two, it is determined that no target clip template matching the first multimedia resource is found in the template library.
- the values of the first preset threshold and the second preset threshold may be the same or different.
- the pre-trained search model is first used to perform the music style dimension, audio fingerprint dimension, video size dimension, and video segment feature dimension on the target video file (ie, the first multimedia resource) and the candidate clip template specified by the user. And the analysis in the dimension of visual effects, and comprehensively evaluate whether the candidate editing template is the target editing template that the user wants to find according to the analysis results in the above five dimensions, so as to ensure the accuracy of the search results.
- the search model may be adjusted based on different scenarios and purposes.
- the search model may only include a sub-model for music style identification.
- the embodiments shown in FIG. 3 and FIG. 5 do not limit the specific implementation of the search model.
- FIG. 7 is a flowchart of a method for searching for a clip template provided by another embodiment of the present disclosure. Referring to FIG. 7 , the method of this embodiment includes:
- a server device acquires a first multimedia resource.
- the server device performs a search according to the first multimedia resource, and obtains a search result.
- the server device returns the search result to the terminal device.
- the terminal device receives the search result returned by the server device.
- the server device acquires the target clip template from the template library according to the search result.
- the server device sends the target editing template to the terminal device.
- the server device can extract all data corresponding to the target clip template from the template library according to the identifier of the target clip template, and package it, and then sent to the end device.
- the terminal device displays the target editing template.
- the terminal device After receiving the data of the target clip template, the terminal device performs decoding and display, so that the user holding the terminal device can view the detailed information of the target clip template.
- the target editing template is more in line with the user's requirements for video creation, the utilization rate of the target editing template can be improved, and at the same time, the target editing template can be improved. Improve the enthusiasm of users for video creation.
- the terminal device may display prompt information that the target clip template is not found to the user according to the search result.
- the clip template search method provided by the embodiments of the present disclosure is introduced in detail with reference to the accompanying drawings and application scenarios.
- the terminal device as a mobile phone
- a video editing APP referred to as Application 1
- a video APP referred to as Application 2
- mobile phones and other terminal devices provide the function of intelligently reading the clipboard, which can automatically fill in the copied content automatically.
- the method for searching for clip templates provided by the present disclosure will be respectively introduced in two scenarios of enabling the smart clipboard reading function and turning off the smart clipboard reading function.
- the user can copy the target link through other video applications or music applications, and when the smart clipboard reading function of the mobile phone is enabled, when the user opens application 1, application 1 can obtain the target link from the clipboard.
- application 2 can display the user interface 11 exemplarily shown in FIG. 8A on the mobile phone, the user interface 11 is used to display the video playing page of application 2, and application 2 can perform some function sets in the video playing page, such as playing Multimedia files (such as short videos), share multimedia files.
- the user interface 11 includes: a control 1101, where the control 1101 is used to copy the link of the multimedia file currently being played.
- the application 1 After the user exits the application 2 and opens the application 1 within a preset time period, the application 1 sends the target link to the server through the background.
- the preset duration is, for example, 5 seconds, 10 seconds, and so on.
- the application 1 may display the user interface 12 exemplarily shown in FIG. 8B on the mobile phone, wherein the user interface 12 is used to display the waiting page of the application 1 .
- the current search progress information may be displayed in the user interface 12 , for example, “75% in recognition” is exemplarily displayed in the waiting page displayed by the user interface 12 .
- the user interface 12 also includes a control 1201 for abandoning the current search task.
- application 1 After application 1 receives that the user performs an operation such as clicking the control 1201 in the user interface 12 shown in FIG. 8B , application 1 can display the default home page of application 1 on the mobile phone.
- the application 1 When the application 1 receives the search result returned by the server and the data of the matched target clip template, the application 1 can display the user interface 13 exemplarily shown in FIG. 8C on the mobile phone, wherein the user interface 13 includes: a display window 1301.
- the display window is used to display the cover of the target editing template, and the cover may be any video frame in the video included in the target editing template, or a specific video frame.
- the user interface 13 further includes: a control 1302, wherein the control 1302 is used to enter the details page of the target clip template.
- the application 1 receives that the user performs an operation such as clicking on the control 1302 in the user interface 13 shown in FIG. 8C
- the application 1 displays the user interface 14 shown in FIG. 8D on the mobile phone, wherein the user interface 14 is used to display the
- the video playing page application 1 can execute some function sets in the video playing page, such as playing multimedia files, providing a visual entry to the video authoring page, and so on.
- the user interface 13 further includes a control 1303 , wherein the control 1303 is used to close the presentation window 1301 .
- the application 1 After the application 1 receives the user's click operation on the control 1302 in the user interface 13 shown in FIG. 8C , the application 1 can send a clip template acquisition request to the server through the mobile phone, and the clip template acquisition request is used to request to acquire all the target clip templates. data.
- the user interface 14 includes: a control 1401, the control 1401 is used to enter a video creation page using the target clip template as a creation template.
- the application 1 can send a candidate clip template acquisition request to the server through the mobile phone, and the candidate clip template acquisition request is used to request the weighted score only times All data of candidate clip templates in the target clip template.
- the server may send all the data of the candidate clip template whose weighted score ranks second to the target clip template to the mobile phone according to the search result.
- the search result returned by the server indicates that the template library does not contain a target clip template that matches the target link
- application 1 can display the template shown in FIG. 8E on the mobile phone
- the user interface 15 shown in the figure wherein the user interface 15 includes a window 1501 , wherein the window 1501 includes a text information display area 1502 and a control 1503 .
- the text information display area 1502 can display the relevant content of the search result. For example, the text information display area 1502 displays the text "The same template has not been found, try another link ⁇ ".
- Control 1503 is used to close window 1501.
- application 1 may display the default home page of application 1 on the mobile phone.
- the application 1 can obtain the target link through manual input by the user.
- the application 1 displays the user interface 16 exemplarily shown in FIG. 8F on the mobile phone, wherein the user interface 16 includes: an input window 1601, wherein the input window 1601 includes a control 1602, and the control 1602 is used to enter the link search page.
- the application 1 After application 1 receives that the user performs an operation such as clicking on the control 1602 in the user interface 16 shown in FIG. 8F , the application 1 displays the user interface 17 exemplarily shown in FIG. 8G on the mobile phone.
- the link search page displayed on the user interface 17 includes an input window 1701, and prompt information can be displayed in the input window 1701 to remind the user to paste video links or music links from other applications to find the same template.
- application 1 can display the user interface 18 shown in FIG. 8H on the mobile phone after receiving the user to perform an operation such as long-pressing the input window 1701 in the user interface 17 shown in FIG. 8G , and the user interface 18 includes controls. 1801 , the control 1801 is used to paste the content in the clipboard into the input window 1701 .
- the application 1 After the application 1 receives the user's operation such as clicking the control 1801 in the user interface 18 shown in FIG. 8H, the target link is displayed in the input window 1701, and the application 1 correspondingly displays the user interface 19 shown in FIG. 8I on the mobile phone.
- the user interface 17 may further include: an input method software disk area 1702 , and the user can manually input a target link into the input window 1701 by operating the input method software keyboard area 1702 .
- the user interface 17 may also include a control 1703 that allows the user to close the link search page.
- User interface 17 also includes controls 1704 for generating search tasks based on target links.
- the mobile phone sends the target link to the server according to the search task.
- the user interface 19 may further include: a control 1705 , where the control 1705 is used to delete all content in the input window 1701 .
- the control 1705 is used to delete all content in the input window 1701 .
- the application 1 displays the user interface 17 shown in FIG. 8G on the mobile phone.
- the user manually operates the input method soft keyboard area 1702 included in the user interface 17 to input part or all of the content of the target link into the input window 1701, he can also operate the control 1705 to delete all the content in the input window 1701 .
- control 1704 When application 1 detects that the correct target link has not been entered in the input window 1701 shown in the user interface 17 and user interface 18, the control 1704 is in the first state; when application 1 detects that the correct target link has been entered in the input window 1701 , the control 1704 is in the second state.
- the first state is an inactive state
- the second state is an active state.
- the operation control 1704 In an inactive state, the operation control 1704 cannot generate a search task; in an activated state, the operation control 1704 can generate a search task according to the target link in the input window 1701 .
- user interface 18 and user interface 19 also include controls 1704 in which user interface 18 is in a first state and user interface 19 in which control 1704 is in a second state.
- the application 1 After the application 1 receives that the user performs an operation such as clicking the control 1704 in the user interface 19 as shown in FIG. 8I, the application 1 generates a search task, and sends the target link to the server through the mobile phone, so that the server searches according to the target link. And after the application 1 receives the user's operation such as clicking the control 1704 in the user interface 19, the application 1 can display the user interface 12 shown in FIG. 8B on the mobile phone.
- the application 1 can display the user interface 13 to the user interface 15 shown in FIG. 8C to FIG. 8E on the mobile phone according to the search result.
- a user interface 20 as shown in Figure 8J is illustratively displayed.
- the user interface 20 includes a window 2001, wherein the window 2001 is used to display guiding information, for example, the guiding information is "support for finding templates through links".
- the communication quality between the mobile phone and the server is poor.
- the server can match the target clip template according to the target link, the mobile phone cannot obtain the target clip from the server due to the poor communication quality between the mobile phone and the server.
- template data The application 1 can display the user interface 21 shown in FIG. 8K on the mobile phone, wherein the user interface 21 includes a window 2101, and the window 2101 is used to display a loading failure page.
- the loading failure page may include an area 2102, a window 2103, and a control 2104, where the area 2102 is used to display the prompt information of the loading failure; the window 2103 contains a control 2105, and the control 2105 is used to generate a new data loading task; the control 2104 is used to Cancel the data load task.
- FIGS. 8A to 8K From the schematic diagrams of the human-computer interaction interface shown in the above-mentioned FIGS. 8A to 8K, combined with the actual application scenarios, it can be seen that the editing template search method provided by the embodiment of the present disclosure is convenient for the user to operate the controls on the user interface displayed by the terminal device. Quickly obtain the desired target editing template, which can better meet the needs of users for video creation.
- FIG. 9 is a schematic structural diagram of a clip template search apparatus according to an embodiment of the present disclosure.
- the clip template search apparatus 900 provided in this embodiment includes: an acquisition module 901 and a search module 902 .
- the obtaining module 901 is configured to obtain the first multimedia resource designated by the user.
- a search module 902 configured to perform a search according to the first multimedia resource, and obtain a search result
- the search result is used to indicate whether a target editing template matching the first multimedia resource is found, and the target editing template is used to indicate that the multimedia material to be edited is edited into a second multimedia material according to the target editing method.
- the target editing method is the editing method adopted by the first multimedia resource.
- the search module 902 is specifically configured to obtain the search result according to the identification result of the candidate clip template in the target dimension; wherein, the identification result of the candidate clip template in the target dimension is based on the first The characteristics of the multimedia resource and the candidate clip template in the target dimension are obtained.
- the search module 902 is further configured to determine the candidate editing template according to the characteristics of the first multimedia resource and the at least one editing template respectively in the target dimension; wherein the at least one editing template is Templates include the candidate clip templates.
- the target dimension includes one or more of: a music style dimension, an audio fingerprint dimension, a video size dimension, a video segment feature dimension, and a visual effect dimension.
- the target dimension includes a plurality of the music genre dimension, the audio fingerprint dimension, the video size dimension, the video segment feature dimension, and the visual effects dimension;
- the search module 902 is specifically configured to obtain, according to the identification results of each dimension of the candidate editing template in the target dimension and the corresponding weight coefficients of each dimension in the target dimension, the corresponding data of the candidate editing template. Weighted calculation result; obtain the search result according to the weighted calculation result corresponding to the candidate clip template.
- the search module 902 is specifically configured to determine the current dimension according to the priority order of each dimension in the target dimension; obtain the current dimension according to the characteristics of the first multimedia resource and the first screening result respectively in the current dimension.
- the first screening result the first recognition result of each editing template in the current dimension is obtained; according to the first recognition result and the screening condition corresponding to the current dimension, obtain the second screening result corresponding to the current dimension ;
- the second screening result includes: one or more editing templates, and the initial state of the first screening result includes: the at least one editing template; determine that the second screening result is the first screening result;
- the current dimension is determined until the second screening result corresponding to the last described current dimension is obtained, and the editing template included in the second screening result corresponding to the last described current dimension is determined to be: the candidate clip template.
- the obtaining module 901 is specifically configured to obtain a target link input by a user, and parse the target link to obtain the first multimedia resource.
- the clip template searching apparatus 900 further includes: a sending module 903 .
- the sending module 903 is configured to send the search result to the user.
- the sending module 903 is further configured to send the target clip template to the user according to the search result.
- the clip template search apparatus provided in this embodiment can be used to execute the technical solution executed by the server device in any of the foregoing embodiments, and its implementation principle and technical effect are similar, and reference may be made to the foregoing detailed description, which will not be repeated here.
- FIG. 10 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure.
- the electronic device 1000 provided in this embodiment includes: a memory 1001 and a processor 1002 .
- the memory 1001 may be an independent physical unit, and may be connected to the processor 1002 through a bus 1003 .
- the memory 1001 and the processor 1002 can also be integrated together, implemented by hardware, and the like.
- the memory 1001 is used to store program instructions, and the processor 1002 invokes the program instructions to execute the operations performed by the server device or the terminal device in any of the above method embodiments.
- the foregoing electronic device 1000 may also include only the processor 1002 .
- the memory 1001 for storing programs is located outside the electronic device 1000, and the processor 1002 is connected to the memory through circuits/wires for reading and executing the programs stored in the memory.
- the processor 1002 may be a central processing unit (CPU), a network processor (NP), or a combination of CPU and NP.
- CPU central processing unit
- NP network processor
- the processor 1002 may further include hardware chips.
- the above-mentioned hardware chip may be an application-specific integrated circuit (ASIC), a programmable logic device (PLD) or a combination thereof.
- ASIC application-specific integrated circuit
- PLD programmable logic device
- the above-mentioned PLD can be a complex programmable logic device (CPLD), a field-programmable gate array (FPGA), a general array logic (generic array logic, GAL) or any combination thereof.
- CPLD complex programmable logic device
- FPGA field-programmable gate array
- GAL general array logic
- the memory 1001 may include volatile memory (volatile memory), such as random-access memory (RAM); the memory may also include non-volatile memory (non-volatile memory), such as flash memory (flash memory) ), a hard disk drive (HDD) or a solid-state drive (SSD); the memory may also include a combination of the above-mentioned types of memory.
- volatile memory such as random-access memory (RAM)
- non-volatile memory such as flash memory (flash memory)
- HDD hard disk drive
- SSD solid-state drive
- the memory may also include a combination of the above-mentioned types of memory.
- the present disclosure further provides a computer-readable storage medium, where the computer-readable storage medium includes computer program instructions, when executed by at least one processor of an electronic device, the computer program instructions are used to perform the service in any of the above method embodiments.
- the present disclosure also provides a program product comprising a computer program, the computer program being stored in a readable storage medium, from which at least one processor of the electronic device can read the data
- the computer program is executed by the at least one processor, so that the electronic device executes the technical solution executed by the server device or the terminal device in any of the above method embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Library & Information Science (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
Claims (11)
- 一种剪辑模板搜索方法,其特征在于,包括:获取用户指定的第一多媒体资源;根据所述第一多媒体资源进行搜索,获取搜索结果;其中,所述搜索结果用于指示是否搜索到与所述第一多媒体资源匹配的目标剪辑模板,所述目标剪辑模板用于指示待剪辑多媒体素材按照目标剪辑方式被剪辑成第二多媒体资源,所述目标剪辑方式为所述第一多媒体资源所采用的剪辑方式。
- 根据权利要求1所述的方法,其特征在于,所述根据所述第一多媒体资源进行搜索,获取搜索结果,包括:根据候选剪辑模板在目标维度的识别结果,获取所述搜索结果;其中,所述候选剪辑模板在目标维度的识别结果是根据所述第一多媒体资源和所述候选剪辑模板在所述目标维度的特征获得的。
- 根据权利要求2所述的方法,其特征在于,所述根据所述候选剪辑模板在目标维度的识别结果,获取所述搜索结果之前,还包括:根据所述第一多媒体资源以及至少一个剪辑模板分别在目标维度上的特征,确定所述候选剪辑模板;其中,所述至少一个剪辑模板包括所述候选剪辑模板。
- 根据权利要求2或3所述的方法,其特征在于,所述目标维度包括:音乐风格维度、音频指纹维度、视频尺寸维度、视频片段特征维度以及视觉效果维度中的一个或多个。
- 根据权利要求4所述的方法,其特征在于,所述目标维度包括所述音乐风格维度、所述音频指纹维度、所述视频尺寸维度、所述视频片段特征维度以及所述视觉效果维度中的多个维度;所述根据所述候选剪辑模板在所述目标维度的识别结果,获取所述搜索结果,包括:根据所述候选剪辑模板分别在所述目标维度中各维度的识别结果、以及所述目标维度中各维度对应的权重系数,获取所述候选剪辑模板对应的加权计算结果;根据所述候选剪辑模板对应的加权计算结果,获取所述搜索结果。
- 根据权利要求4所述的方法,其特征在于,所述根据所述第一多媒体资源以及至少一个剪辑模板分别在目标维度上的特征,确定所述候选剪辑模板,包括:按照目标维度中的各维度的优先级顺序,确定当前维度;根据所述第一多媒体资源以及第一筛选结果分别在所述当前维度的特征,获取所述第一筛选结果中每个剪辑模板分别在所述当前维度的第一识别结果;根据所述第一识别结果以及所述当前维度对应的筛选条件,获取所述当前维度对应的第二筛选结果;所述第二筛选结果包括:一个或者多个剪辑模板,所述第一筛选结果的初始状态包括:所述至少一个剪辑模板;确定所述第二筛选结果为第一筛选结果;返回执行所述按照目标维度中的各维度的优先级顺序,确定当前维度,直到获取最后一个所述当前维度对应的第二筛选结果,确定最后一个所述当前维度对应的第二筛选结果 包括的剪辑模板为所述候选剪辑模板。
- 根据权利要求1至3任一项所述的方法,其特征在于,所述获取用户指定的第一多媒体资源,包括:获取用户输入的目标链接,并对所述目标链接进行解析,获取所述第一多媒体资源。
- 根据权利要求1至3任一项所述的方法,其特征在于,所述方法还包括:向所述用户发送所述搜索结果。
- 根据权利要求8所述的方法,其特征在于,所述方法还包括:根据所述搜索结果,向所述用户发送所述目标剪辑模板。
- 一种电子设备,其特征在于,包括:存储器、处理器以及计算机程序;所述存储器被配置为存储所述计算机程序;所述处理器被配置为执行所述计算机程序,以实现如权利要求1至9任一项所述的方法。
- 一种程序产品,其特征在于,包括:计算机程序,所述计算机程序存储在可读存储介质中,电子设备的至少一个处理器可以从所述可读存储介质中读取所述计算机程序,所述至少一个处理器执行所述计算机程序使得所述电子设备实现如权利要求1至9任一项所述的方法。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2023566920A JP2024516836A (ja) | 2021-04-30 | 2022-04-29 | クリップテンプレート検索方法及び装置 |
EP22795026.8A EP4322025A4 (en) | 2021-04-30 | 2022-04-29 | METHOD AND APPARATUS FOR SEARCHING CUTTING TEMPLATE |
US18/484,933 US20240037134A1 (en) | 2021-04-30 | 2023-10-11 | Method and apparatus for searching for clipping template |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110485269.3A CN115269889B (zh) | 2021-04-30 | 2021-04-30 | 剪辑模板搜索方法及装置 |
CN202110485269.3 | 2021-04-30 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/484,933 Continuation US20240037134A1 (en) | 2021-04-30 | 2023-10-11 | Method and apparatus for searching for clipping template |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022228557A1 true WO2022228557A1 (zh) | 2022-11-03 |
Family
ID=83744887
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/090348 WO2022228557A1 (zh) | 2021-04-30 | 2022-04-29 | 剪辑模板搜索方法及装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20240037134A1 (zh) |
EP (1) | EP4322025A4 (zh) |
JP (1) | JP2024516836A (zh) |
CN (1) | CN115269889B (zh) |
WO (1) | WO2022228557A1 (zh) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118175348A (zh) * | 2022-12-09 | 2024-06-11 | 北京字跳网络技术有限公司 | 视频模板的推送方法、装置、介质及设备 |
CN116506694B (zh) * | 2023-06-26 | 2023-10-27 | 北京达佳互联信息技术有限公司 | 视频剪辑方法、装置、电子设备及存储介质 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090254565A1 (en) * | 2008-04-03 | 2009-10-08 | Thumbplay, Inc. | Subsequent tailoring of a sign-up page based on a search engine query |
CN110177219A (zh) * | 2019-07-01 | 2019-08-27 | 百度在线网络技术(北京)有限公司 | 视频的模板推荐方法和装置 |
CN111541936A (zh) * | 2020-04-02 | 2020-08-14 | 腾讯科技(深圳)有限公司 | 视频及图像处理方法、装置、电子设备、存储介质 |
CN111914523A (zh) * | 2020-08-19 | 2020-11-10 | 腾讯科技(深圳)有限公司 | 基于人工智能的多媒体处理方法、装置及电子设备 |
CN112203140A (zh) * | 2020-09-10 | 2021-01-08 | 北京达佳互联信息技术有限公司 | 一种视频剪辑方法、装置、电子设备及存储介质 |
CN112449231A (zh) * | 2019-08-30 | 2021-03-05 | 腾讯科技(深圳)有限公司 | 多媒体文件素材的处理方法、装置、电子设备及存储介质 |
CN113840099A (zh) * | 2020-06-23 | 2021-12-24 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、设备及计算机可读存储介质 |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AUPP624698A0 (en) * | 1998-09-29 | 1998-10-22 | Canon Kabushiki Kaisha | Method and apparatus for multimedia editing |
US7921156B1 (en) * | 2010-08-05 | 2011-04-05 | Solariat, Inc. | Methods and apparatus for inserting content into conversations in on-line and digital environments |
US10691885B2 (en) * | 2016-03-30 | 2020-06-23 | Evernote Corporation | Extracting structured data from handwritten and audio notes |
US10444946B2 (en) * | 2016-12-13 | 2019-10-15 | Evernote Corporation | Shared user driven clipping of multiple web pages |
US20190042574A1 (en) * | 2017-08-01 | 2019-02-07 | Samsung Electronics Co., Ltd. | Electronic device and method for controlling the electronic device |
CN109299445A (zh) * | 2018-08-01 | 2019-02-01 | 政采云有限公司 | 获取文件模板的方法、装置、计算设备及存储介质 |
CN109255053B (zh) * | 2018-09-14 | 2021-08-20 | 北京奇艺世纪科技有限公司 | 资源搜索方法、装置、终端、服务器、计算机可读存储介质 |
CN110139159B (zh) * | 2019-06-21 | 2021-04-06 | 上海摩象网络科技有限公司 | 视频素材的处理方法、装置及存储介质 |
CN110536177B (zh) * | 2019-09-23 | 2020-10-09 | 北京达佳互联信息技术有限公司 | 视频生成方法、装置、电子设备及存储介质 |
CN111105819B (zh) * | 2019-12-13 | 2021-08-13 | 北京达佳互联信息技术有限公司 | 剪辑模板的推荐方法、装置、电子设备及存储介质 |
CN111246300B (zh) * | 2020-01-02 | 2022-04-22 | 北京达佳互联信息技术有限公司 | 剪辑模板的生成方法、装置、设备及存储介质 |
CN111243632B (zh) * | 2020-01-02 | 2022-06-24 | 北京达佳互联信息技术有限公司 | 多媒体资源的生成方法、装置、设备及存储介质 |
CN111460183B (zh) * | 2020-03-30 | 2024-02-13 | 北京金堤科技有限公司 | 多媒体文件生成方法和装置、存储介质、电子设备 |
CN111522863B (zh) * | 2020-04-15 | 2023-07-25 | 北京百度网讯科技有限公司 | 一种主题概念挖掘方法、装置、设备以及存储介质 |
CN111835986B (zh) * | 2020-07-09 | 2021-08-24 | 腾讯科技(深圳)有限公司 | 视频编辑处理方法、装置及电子设备 |
CN111930994A (zh) * | 2020-07-14 | 2020-11-13 | 腾讯科技(深圳)有限公司 | 视频编辑的处理方法、装置、电子设备及存储介质 |
CN111741331B (zh) * | 2020-08-07 | 2020-12-22 | 北京美摄网络科技有限公司 | 一种视频片段处理方法、装置、存储介质及设备 |
CN112015926B (zh) * | 2020-08-27 | 2022-03-04 | 北京字节跳动网络技术有限公司 | 搜索结果的展示方法、装置、可读介质和电子设备 |
CN112711937B (zh) * | 2021-01-18 | 2022-06-24 | 腾讯科技(深圳)有限公司 | 一种模板推荐方法、装置、设备及存储介质 |
US12099544B2 (en) * | 2022-07-21 | 2024-09-24 | Google Llc | Systems and methods for generating stories for live events using a scalable pipeline |
-
2021
- 2021-04-30 CN CN202110485269.3A patent/CN115269889B/zh active Active
-
2022
- 2022-04-29 WO PCT/CN2022/090348 patent/WO2022228557A1/zh active Application Filing
- 2022-04-29 JP JP2023566920A patent/JP2024516836A/ja active Pending
- 2022-04-29 EP EP22795026.8A patent/EP4322025A4/en active Pending
-
2023
- 2023-10-11 US US18/484,933 patent/US20240037134A1/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090254565A1 (en) * | 2008-04-03 | 2009-10-08 | Thumbplay, Inc. | Subsequent tailoring of a sign-up page based on a search engine query |
CN110177219A (zh) * | 2019-07-01 | 2019-08-27 | 百度在线网络技术(北京)有限公司 | 视频的模板推荐方法和装置 |
CN112449231A (zh) * | 2019-08-30 | 2021-03-05 | 腾讯科技(深圳)有限公司 | 多媒体文件素材的处理方法、装置、电子设备及存储介质 |
CN111541936A (zh) * | 2020-04-02 | 2020-08-14 | 腾讯科技(深圳)有限公司 | 视频及图像处理方法、装置、电子设备、存储介质 |
CN113840099A (zh) * | 2020-06-23 | 2021-12-24 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、设备及计算机可读存储介质 |
CN111914523A (zh) * | 2020-08-19 | 2020-11-10 | 腾讯科技(深圳)有限公司 | 基于人工智能的多媒体处理方法、装置及电子设备 |
CN112203140A (zh) * | 2020-09-10 | 2021-01-08 | 北京达佳互联信息技术有限公司 | 一种视频剪辑方法、装置、电子设备及存储介质 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4322025A4 |
Also Published As
Publication number | Publication date |
---|---|
JP2024516836A (ja) | 2024-04-17 |
CN115269889A (zh) | 2022-11-01 |
US20240037134A1 (en) | 2024-02-01 |
EP4322025A1 (en) | 2024-02-14 |
CN115269889B (zh) | 2024-07-02 |
EP4322025A4 (en) | 2024-08-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107256267B (zh) | 查询方法和装置 | |
US10860811B2 (en) | Method and device for generating review article of hot news, and terminal device | |
US10275499B2 (en) | Tag selection, clustering, and recommendation for content hosting services | |
WO2022228557A1 (zh) | 剪辑模板搜索方法及装置 | |
US9292519B2 (en) | Signature-based system and method for generation of personalized multimedia channels | |
CN105009118B (zh) | 定制的内容消费界面 | |
US8666749B1 (en) | System and method for audio snippet generation from a subset of music tracks | |
US10402407B2 (en) | Contextual smart tags for content retrieval | |
TW201214173A (en) | Methods and apparatus for displaying content | |
US20140164371A1 (en) | Extraction of media portions in association with correlated input | |
US9449027B2 (en) | Apparatus and method for representing and manipulating metadata | |
EP2210196A2 (en) | Generating metadata for association with a collection of content items | |
CN104281656B (zh) | 在应用程序中加入标签信息的方法和装置 | |
US11048736B2 (en) | Filtering search results using smart tags | |
JP2017535860A (ja) | マルチメディア内容の提供方法および装置 | |
CN111723289B (zh) | 信息推荐方法及装置 | |
CN111680254A (zh) | 一种内容推荐方法及装置 | |
US20140161423A1 (en) | Message composition of media portions in association with image content | |
US20170357712A1 (en) | Method and system for searching and identifying content items in response to a search query using a matched keyword whitelist | |
WO2023016349A1 (zh) | 一种文本输入方法、装置、电子设备和存储介质 | |
US20130346385A1 (en) | System and method for a purposeful sharing environment | |
CN110290199A (zh) | 内容推送方法、装置及设备 | |
WO2023128877A2 (zh) | 视频生成方法、装置、电子设备及可读存储介质 | |
US20140163956A1 (en) | Message composition of media portions in association with correlated text | |
US10241988B2 (en) | Prioritizing smart tag creation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22795026 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023566920 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2022795026 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2022795026 Country of ref document: EP Effective date: 20231106 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |