WO2022087826A1 - Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible - Google Patents

Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible Download PDF

Info

Publication number
WO2022087826A1
WO2022087826A1 PCT/CN2020/123998 CN2020123998W WO2022087826A1 WO 2022087826 A1 WO2022087826 A1 WO 2022087826A1 CN 2020123998 W CN2020123998 W CN 2020123998W WO 2022087826 A1 WO2022087826 A1 WO 2022087826A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
target
picture
preset
processed
Prior art date
Application number
PCT/CN2020/123998
Other languages
English (en)
Chinese (zh)
Inventor
周娴
李鑫超
刘志鹏
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to CN202080044431.1A priority Critical patent/CN114026874A/zh
Priority to PCT/CN2020/123998 priority patent/WO2022087826A1/fr
Publication of WO2022087826A1 publication Critical patent/WO2022087826A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234381Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the temporal resolution, e.g. decreasing the frame rate by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440281Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the temporal resolution, e.g. by frame skipping

Definitions

  • the present invention belongs to the field of network technologies, and in particular, relates to a video processing method, device, removable device and readable storage medium.
  • the present invention provides a video processing method, device, removable device and readable storage medium, so as to solve the problems of high screening cost and low screening efficiency.
  • an embodiment of the present invention provides a video processing method, which includes:
  • the target information For the target picture in the video to be processed, determine the target information corresponding to the picture parameter of the target picture; the target information is used to represent whether the picture parameter of the target picture meets the preset requirements;
  • an embodiment of the present invention provides a video processing apparatus, and the apparatus includes a memory and a processor;
  • the memory for storing program codes
  • the processor calls the program code, and when the program code is executed, is configured to perform the following operations:
  • the target information For the target picture in the video to be processed, determine the target information corresponding to the picture parameter of the target picture; the target information is used to represent whether the picture parameter of the target picture meets the preset requirements;
  • an embodiment of the present invention provides a movable device, where the movable device is configured to execute the steps in the video processing method described in the first aspect.
  • an embodiment of the present invention provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the following operations are implemented:
  • the target information For the target picture in the video to be processed, determine the target information corresponding to the picture parameter of the target picture; the target information is used to represent whether the picture parameter of the target picture meets the preset requirements;
  • target information corresponding to the picture parameters of the target picture is determined, wherein the target information is used to indicate whether the picture parameters of the target picture meet the preset requirements.
  • the target information of the included target picture is screened for the video to be processed, wherein the screening process includes deleting the target picture that does not meet the preset requirements or one of the video clips to which the target picture belongs, and finally, according to the screened and processed video to be processed , generate the target video.
  • FIG. 1 is a flowchart of steps of a video processing method provided by an embodiment of the present invention.
  • FIG. 2 is a schematic diagram of a clipping process provided by an embodiment of the present invention.
  • FIG. 3 is a block diagram of a video processing apparatus provided by an embodiment of the present invention.
  • FIG. 4 is a block diagram of a computing processing device according to an embodiment of the present invention.
  • FIG. 5 is a block diagram of a portable or fixed storage unit according to an embodiment of the present invention.
  • FIG. 1 is a flowchart of steps of a video processing method provided by an embodiment of the present invention. As shown in FIG. 1 , the method may include:
  • Step 101 For a target picture in the video to be processed, determine target information corresponding to a picture parameter of the target picture; the target information is used to represent whether the picture parameter of the target picture meets a preset requirement.
  • the video to be processed may be a video that needs to be screened for pictures with poor quality.
  • the video to be processed may be a video shot by the user, or a video downloaded from a network, which is not limited in this embodiment of the present invention.
  • a video can be essentially understood as a picture sequence composed of multiple frames of images, and the target picture in the video to be processed may be all pictures in the picture sequence, or a part of the pictures in the picture sequence.
  • the picture parameter of the target picture may be a parameter that can characterize the picture quality, and the specific number and type of the picture parameter may be set according to actual requirements.
  • the picture parameters may be set as the clarity of the picture, the exposure level of the picture, and the like.
  • the preset requirement may be a quality requirement that needs to be satisfied by the picture parameters of the picture. If the picture parameters of the picture meet the preset requirements, it means that the picture quality is high and can meet the quality requirements. On the contrary, if the picture parameters of the picture do not meet the preset requirements, it means that the quality of the picture is poor and cannot meet the quality requirements.
  • the specific content of the preset requirements can be set according to actual needs. For example, the preset requirement may be that the degree of sharpness is greater than the preset sharpness threshold, or the degree of exposure is within the range of the preset exposure degree, and so on.
  • the target information may be a tag that can characterize whether the picture parameters of the target picture meet the preset requirements.
  • the specific content of the label can be set according to actual needs. For example, numbers, letters, special symbols, etc. can be used as labels.
  • the specific content of the label when the picture parameters meet the preset requirements is different from the specific content of the label when the picture parameters do not meet the preset requirements. For example, "0" may be used as a label indicating that the picture parameters of the target picture do not meet the preset requirements, and "1" may be used as a label indicating that the picture parameters of the target picture meet the preset requirements.
  • Step 102 Perform screening processing on the video to be processed according to the target information of the target picture contained in the video to be processed; wherein, the screening processing includes deleting target pictures or the target pictures that do not meet the preset requirements One of the video clips the image belongs to.
  • whether the picture parameters of the target picture satisfy a preset condition may be first determined according to the target information of the target picture. For example, it may be determined that the picture parameter does not meet the preset requirement when the target information is 0, and it can be determined that the picture parameter meets the preset requirement when the target information is 1. Further, if the picture parameters of the target picture do not meet the preset requirements, the target picture may be determined as the target picture that does not meet the preset requirements. Correspondingly, the target picture may be deleted or one of the video clips to which the target picture that does not meet the preset requirements belongs may be directly deleted. One of the video clips to which the target picture belongs may be a video clip that includes the target picture in the video to be processed, and the specific length of the video clip can be set according to actual requirements, which is not limited in this embodiment of the present invention.
  • Step 103 generating a target video according to the to-be-processed video after screening and processing.
  • the video to be processed is screened, which can reduce the pictures of poor quality included in the video to be processed, and further improve the image of the video to be processed after the screening process. quality.
  • the target video is generated according to the video to be processed after screening, which can ensure that the final generated target video has higher image quality.
  • the screened video to be processed may be directly used as the target video.
  • the video processing method can determine the target information corresponding to the picture parameter of the target picture for the target picture in the video to be processed, wherein the target information is used to indicate whether the picture parameter of the target picture satisfies the According to the preset requirements, the to-be-processed video is screened according to the target information of the target picture contained in the to-be-processed video, wherein the screening process includes deleting the target picture that does not meet the preset requirements or one of the video clips to which the target picture belongs, and finally , and generate the target video according to the to-be-processed video after screening.
  • the screening cost can be reduced to a certain extent and the screening efficiency can be improved.
  • the above picture parameters may include one or more of a blur degree, a shake degree, an exposure degree, and a color change degree of the picture.
  • the picture parameters may also include other types of parameters, which are not limited in this embodiment of the present invention.
  • the degree of blurring of the picture it can be determined whether the overall picture is clear, based on the degree of shaking of the picture, it can be determined whether the picture of the picture has a shaking problem, based on the degree of exposure of the picture, it can be determined whether the brightness of the picture is appropriate, and the degree of color change of the picture can be determined. Is the color difference of the picture appropriate? However, in practical application scenarios, when the picture is blurred and unclear, the picture is shaken, the brightness is too high or too low, and the color difference is too high or too low, it often means that the quality of the picture is poor.
  • the pictures in the video to be repaired can be screened from these parameters in the future, and then certain To a certain extent, it can ensure that pictures with poor quality can be deleted more accurately.
  • the calculation amount that the hardware device can bear may be determined first, and the number of selected picture parameters is determined according to the amount of calculation that the hardware device can bear, so as to ensure that the hardware device can run normally. Wherein, the number of selected picture parameters is positively related to the loadable calculation amount.
  • the fuzzy label information corresponding to the degree of blur the information of the shaking label corresponding to the degree of shaking, the information of the exposure label corresponding to the degree of exposure, and the corresponding degree of color change can be obtained.
  • color difference label information the fuzzy label information, the shake label information, the exposure label information and the color difference label information are the target information.
  • the fuzzy label information can be used to characterize whether the blur degree of the target image meets the preset requirement corresponding to the blur degree
  • the jitter label information can be used to characterize whether the jitter degree of the target image meets the preset requirement corresponding to the jitter degree
  • the exposure label information is used to characterize the target Whether the exposure degree of the picture meets the preset requirements corresponding to the exposure degree
  • the color difference label information is used to indicate whether the color change degree of the target image meets the preset requirements corresponding to the color change degree.
  • the above-mentioned operation of screening the video to be processed according to the target information of the target picture contained in the video to be processed may be implemented by the following sub-steps:
  • Sub-step (1) segment the to-be-processed video to obtain multiple video segments.
  • the video to be processed may be divided into video segments containing the same number of frames by using a preset segmentation algorithm in a manner of dividing at equal intervals.
  • a preset segmentation algorithm may be used to randomly divide the video to be processed into video segments containing different numbers of frames in a manner of unequal interval division, which is not limited in this embodiment of the present invention.
  • other division manners may also be used to implement division, which is not limited in this embodiment of the present invention.
  • Sub-step (2) For each of the video clips, perform screening processing on the video clips according to the target information of the target picture contained in the video clips.
  • a video clip may be used as a processing unit, and screening processing may be performed for each video clip.
  • the screening process for the video clip may include deleting the target picture or the video clip that does not meet the preset requirements in the video clip.
  • the video to be processed is first divided to obtain a plurality of video segments.
  • the video clip is used as a processing unit to filter the video clip.
  • the video clips after the screening process can be merged to obtain the target video. Since the screening process deletes pictures with poor quality in each video clip, the target video is obtained by merging the video clips after the screening process, which can ensure the quality of the target video to a certain extent. Specifically, when merging, the filtered video segments may be merged according to the sequence in the video to be processed, so as to ensure that the final target video can be played normally.
  • the video to be processed may be independently selected by the user. Specifically, before the target information corresponding to the picture parameter of the target picture is determined, a user's selection operation on an optional video may be received; the optional video selected by the selection operation may be determined as the to-be-processed video; the optional video For videos stored in electronic devices. Correspondingly, after the target video is generated, the target video can be displayed to the user, so as to ensure that the user can obtain the processing result in time and improve the interaction effect.
  • each video stored in the electronic device can be displayed in the client, and the user can click the optional video to be processed to realize the input selection operation.
  • the electronic device may receive the selection operation, and determine the selected optional video as the video to be processed.
  • target information may be added to each target picture and displayed to the user.
  • the user can select the target image to be deleted according to the requirements and in combination with the target information of the target image. Accordingly, the electronic device can delete the target picture selected by the user. In this way, by displaying the target picture with the target information added to the user, the user can independently decide the specific target picture to be deleted, which can improve the flexibility of the operation.
  • the following steps may also be performed:
  • Step A Determine the target picture according to the pictures contained in the video to be processed; extract the target picture from the video to be processed, and add timestamp information to each of the target pictures; the timestamp information Used to characterize the order of the target picture in the target picture.
  • the appearance time point of each target picture in the video to be processed may be used as timestamp information.
  • the target picture includes picture a, picture b, and picture c.
  • the appearance time points of the picture a, the picture b, and the picture c are respectively: the 10th second, the 15th second, and the 21st second.
  • the 10th second, the 15th second, and the 21st second can be used as the timestamp information including the picture a, the picture b, and the picture c, respectively.
  • an association relationship between each target picture and the corresponding time stamp information may be established, or, the corresponding time stamp information may also be written into the target picture. .
  • Step B According to the timestamp information of the target picture, the The target picture is merged into the to-be-processed video.
  • the location of the target picture may be determined according to the timestamp information first, and then the target picture is inserted into the corresponding location. For example, picture a may be inserted into the video to be processed at the 10th second, picture b may be inserted into the video to be processed at the 15th second, and picture c may be inserted into the video to be processed at the 21st second.
  • the preset requirement may include one or more of normal blur degree, normal jitter degree, normal exposure degree and normal color change degree.
  • the target information corresponding to the picture parameter is a label used to indicate that the picture parameter does not meet the preset requirements, and if so, it can be determined that none of the picture parameters meet the preset requirements. .
  • the third target picture can be directly deleted before being merged. In this way, the burden of subsequent screening processing can be reduced, and unnecessary target pictures can be avoided from performing merging operations, thereby ensuring processing efficiency.
  • the target picture before the operation of determining the target information corresponding to the picture parameters of the target picture, the target picture is determined and the target picture is extracted. In this way, by extracting the target picture separately, it is possible to avoid interference from other pictures in the video to be processed when the target information corresponding to the picture parameters of the target picture is subsequently determined, thereby ensuring the processing effect to a certain extent.
  • the target picture is combined into the video to be processed, so as to ensure that the subsequent steps can be performed normally. And by adding time stamp information to the target picture, the target picture can be conveniently combined into the video to be processed according to the time stamp information, thereby ensuring the processing effect.
  • the operation of determining the target picture can be implemented by the following sub-steps:
  • Sub-step (3) when the total number of pictures contained in the video to be processed is greater than the first preset number threshold, select m frames of pictures from the video to be processed according to a preset frame rate as the target picture; the m is not greater than the first preset number threshold.
  • the first preset number threshold and the preset frame rate may be set according to actual requirements.
  • the first preset number threshold may be the number of pictures required to implement screening.
  • the first preset number threshold may be 50, 60, and so on.
  • the preset frame rate can be 4 frames per second (fps).
  • the total number of pictures included in the video to be processed may be determined first, for example, the configuration information of the video to be processed may be read to obtain the total number.
  • selection can be made according to the preset frame rate, and accordingly, these target pictures can be correspondingly extracted subsequently. That is, frame extraction can be performed according to a preset frame rate to obtain m target pictures, and the m target pictures form a picture sequence.
  • Sub-step (4) under the condition that the total number is not greater than the first preset number threshold, use all pictures included in the video to be processed as the target picture.
  • the total number is not greater than the first preset number threshold, it means that if all the pictures are directly used as the target pictures, too much calculation amount will not be introduced. Therefore, all pictures can be directly used as target pictures without other processing. In this way, it is possible to avoid performing unnecessary other processing while realizing the determination of the target picture.
  • all pictures included in the video to be processed are used as target pictures only when the total number is not greater than the first preset number threshold, and when the total number is greater than the first preset number threshold , select only part of the image as the target image. In this way, it can be ensured that there are enough target pictures to a certain extent while avoiding too many target pictures, so as to ensure that more information can be provided for subsequent screening, and the screening effect can be ensured.
  • the preset requirements in this embodiment of the present invention may include preset requirements corresponding to the degree of blur: normal degree of blur, preset requirements corresponding to the degree of jitter: normal degree of jitter and preset requirements corresponding to the degree of exposure: normal degree of exposure,
  • the preset requirements corresponding to the degree of color change the degree of color change is normal.
  • the target information corresponding to the blur degree of the target picture can be determined through the following sub-steps:
  • Sub-step (5) determine the first ratio of the first number of pixels with the fuzzy confidence greater than the preset reliability threshold in the target picture to the total number of pixels of the target picture; according to the first ratio and the first ratio; A magnitude relationship between preset ratio thresholds determines blur label information; the blur label information is used to represent whether the blur degree of the target image is normal.
  • the fuzzy confidence level can be used to represent the probability that the pixel point is a fuzzy pixel point.
  • the target image may be input into a preset blur detection module, and the blur detection module may be implemented based on a semantic segmentation network composed of convolutional neural networks (Convolutional Neural Networks, CNN).
  • CNN convolutional Neural Networks
  • sample images with different degrees of blur can be obtained, and the neural network can be trained by using these sample images to generate a blur detection module.
  • the blur detection module can extract the picture features of the input target picture, and then determine the blur confidence of each pixel in the target picture according to the extracted picture features.
  • the preset reliability threshold and the first preset ratio threshold may be set according to actual requirements.
  • the preset reliability threshold may be 80%.
  • the first preset ratio threshold may be 0.2.
  • the fuzzy confidence of the pixel point is greater than the preset confidence threshold, it can be considered that the pixel point has a fuzzy problem, and the pixel point is a fuzzy pixel point.
  • the fuzzy confidence of each pixel can be compared with a preset confidence threshold to determine the first number. Then, the first number is divided by the total number of pixels of the target image to obtain a first ratio.
  • the blurring label information of the target image may be set as the first blurring label.
  • the first fuzzy label indicates that the degree of blurring of the target image is abnormal.
  • the blur tag information of the target image can be set as the second blur tag.
  • the second fuzzy label indicates that the degree of blurring of the target image is normal.
  • the first fuzzy label may be "unclear"
  • the second fuzzy label may be "clear".
  • the background area in the target image can also be determined, and then the coincidence ratio of the blurred area and the background area can be detected; wherein, the blurred area is an area composed of pixels whose blur confidence is greater than a preset confidence threshold.
  • the first ratio is greater than the first preset ratio threshold and the overlap ratio is not greater than the preset overlap ratio threshold
  • the blur tag information of the target image is set as the first blur tag.
  • the preset coincidence ratio threshold can be set according to actual needs. For example, the preset coincidence ratio threshold can be 90%.
  • the coincidence ratio is greater than the preset coincidence ratio threshold, it can be considered that the blurred area in the target image is a normal phenomenon. If the coincidence ratio is not greater than the preset coincidence ratio threshold, it can be considered that the blurred area in the target image is caused by abnormal factors.
  • the target image sets a first blur label representing an abnormal blur degree. In this way, in the case of blurred background caused by focusing, it can be avoided that the target picture is misjudged as a picture with an abnormal degree of blurring, and then inappropriate fuzzy label information is set for the target picture.
  • the target information corresponding to the degree of shaking of the target picture can be determined by the following sub-steps:
  • Sub-step (6) take the target picture as the input of the first preset classification model, and determine the jitter label information according to the output category of the first preset classification model; the first preset classification model is used for Whether the degree of jitter is normal for image classification.
  • the shaking label information is used to represent whether the shaking degree of the target image is normal.
  • the first preset classification model may be based on a CNN neural network, and the first preset classification model may be obtained by training sample pictures with different degrees of jitter (including normal and abnormal degrees of jitter), and the first preset classification model may be. During the training process, through deep learning to learn the ability to distinguish whether the jitter of the picture is normal or not. Specifically, after the target picture is input into the first preset classification model, the first preset classification model can extract the picture features of the input target picture, and then according to the extracted picture features, determine whether the degree of shaking of the target picture is normal, and output category.
  • the shaking label information of the target image may be set as the first fuzzy label.
  • the output category of the first preset classification model is a category representing a normal degree of shaking
  • the shaking label information of the target image may be set as the second fuzzy label.
  • the first jitter tag may be "tremble”
  • the second jitter tag may be "untremble”.
  • the jitter label information is determined by the output category of the first preset classification model. In this way, it is only necessary to input the target image into the first preset classification model to conveniently determine whether the target image jitters, which in turn can facilitate Set the jitter label information later to improve the setting efficiency.
  • the target information corresponding to the exposure degree of the target picture can be determined by the following sub-steps:
  • Sub-step (7) take the target picture as the input of the second preset classification model, and determine the exposure label information according to the output category of the second preset classification model; the second preset classification model is used for Whether the exposure level is normal for image classification.
  • the exposure label information is used to represent whether the exposure degree of the target image is normal.
  • the second preset classification model may be based on a CNN neural network, and the second preset classification model may be obtained by training sample pictures with different exposure levels (including normal exposure and abnormal exposure).
  • the second preset classification model The ability to distinguish whether the exposure level of a picture is normal can be learned during the training process. Specifically, after the target picture is input into the second preset classification model, the second preset classification model can extract the picture features of the input target picture, and then judge whether the exposure degree of the target picture is normal according to the extracted picture features, and output category.
  • the exposure label information of the target image may be set as the first exposure label.
  • the output category of the second preset classification model is a category representing a normal exposure degree
  • the exposure label information of the target image may be set as the second exposure label.
  • the first exposure tag may be "expose F”
  • the second exposure tag may be "expose R”.
  • the exposure label information is determined by the output category of the second preset classification model. In this way, it is only necessary to input the target image into the second preset classification model to conveniently determine whether the target image has abnormal exposure ( Overexposure and overdarkness), which can facilitate subsequent setting of exposure label information and improve setting efficiency.
  • abnormal exposure Overexposure and overdarkness
  • the target information corresponding to the color change degree of the target picture can be determined by the following sub-steps:
  • Sub-step (8) determine the second ratio between the second number of pixels whose color value exceeds the preset color value range in the target picture and the total number of pixels; according to the second ratio and the second preset ratio The size relationship between the thresholds determines the color difference label information; the color difference label information is used to represent whether the color change degree of the target picture is normal.
  • the color value of the pixel point may be the color channel value of the pixel point, for example, the red, green and blue (RGB) color channel value.
  • the preset color value range and the second preset ratio threshold can be set according to actual needs. For example, the preset color value range may be determined according to the lowest color value and the highest color value in multiple pictures with normal color difference. The second preset ratio threshold may be based on the lowest ratio threshold that will affect the user's viewing experience. If the color value falls within the preset color value range, it can be considered that the color of the pixel point is normal, and if the color value does not fall within the preset color value range, it can be considered that the color of the pixel point is abnormal.
  • a color value detection algorithm may be used to determine the color value of each pixel in the target image. The color value is then compared to a preset range of color values to determine the second quantity. The second ratio is then divided by the total number of pixels in the target image.
  • the color difference label information of the target picture may be set as the first color difference label.
  • the first color difference label indicates that the color change degree of the target picture is abnormal.
  • the second ratio is not greater than the second preset ratio threshold, it can be considered that the color change degree of the target picture is normal, and accordingly, the color difference label information of the target picture can be set as the second color difference label.
  • the second color difference label indicates that the color change degree of the target picture is normal.
  • the first color difference label may be "Color F"
  • the second color difference label may be "Color R".
  • the present invention by determining the proportion of pixels with abnormal colors in the target picture, determining whether the color change degree of the target picture is normal according to the proportion of the proportion, and setting the corresponding color difference label information, to a certain extent, it is possible to ensure that the picture quality The accuracy of the color difference judgment result, and the reliability of the set color difference label information.
  • the above-mentioned step of screening the video clips according to the target information of the target pictures included in the video clips may include:
  • the target pictures that appear continuously refer to the pictures that are adjacent to the target picture in the forward direction and/or the backward direction in the video segment are also target pictures.
  • the target pictures whose picture parameters do not meet the preset requirements may be determined according to the target information of each target picture in the video clip.
  • the pictures of the target pictures whose picture parameters do not meet the preset requirements that appear continuously are used as the first target pictures.
  • the picture parameters do not meet the preset requirements may be that all the picture parameters do not meet the preset requirements, or some picture parameters do not meet the preset requirements.
  • the target picture that all picture parameters do not meet the preset requirements and appears continuously can be used as the first target picture
  • the target information can be: the first fuzzy label "unclear”, the first jitter label "tremble", The consecutively appearing target pictures of the first exposure label "expose F” and the first color difference label "Color F" are used as the first target picture.
  • a target picture in which some of the picture parameters do not meet the preset requirements and which appear continuously may also be used as the first target picture.
  • the video clip includes picture 1 to picture 20.
  • the target pictures whose picture parameters do not meet the preset requirements include: picture 1, picture 10, picture 11, and picture 13. Because the picture 1 to the picture 10 are the target pictures whose parameters that appear continuously do not meet the preset requirements. Therefore, picture 1 to picture 10 can be determined as the first target picture.
  • the second preset number threshold may be set according to actual needs. If the third number is greater than the second preset number threshold, it can be considered that if all the first target pictures are directly deleted, there is a high probability that the video will appear in the video later. The front and back of the screen are not harmoniously connected, and the video playback is not smooth. Therefore, only part of the first target picture can be culled. Specifically, n first target pictures may be randomly selected for deletion.
  • the present invention by determining the first target pictures that appear continuously and the picture parameters do not meet the preset requirements, and in the case of a large number of first target pictures, n partial first target pictures are eliminated, and the continuous low-quality pictures can be In the case of occurrence, avoid the problem that the subsequent video playback is not smooth due to culling too many consecutive pictures.
  • a picture transition effect may be added to the remaining first target pictures, wherein the picture transition effect may be used to control the first target picture
  • the display effect when it appears can include one or more of right out, left in, spin out and spin in, and fade in and fade out.
  • the picture transition effect may also include other types of effects, which are not limited in this embodiment of the present invention.
  • an operation of adding a picture transition effect may be performed when an adding instruction sent by a user is received. In this way, unnecessary adding operations can be avoided, resulting in that the final generated target video cannot satisfy the user’s needs. question of needs.
  • a transition video frame may be generated according to the picture content of the remaining first target pictures; wherein, the picture parameters of the transition video frame meet preset requirements and The picture similarity is greater than the preset similarity threshold, and the picture similarity is the similarity between the picture content of the transition video frame and the picture content of the remaining first target picture.
  • the preset similarity threshold may be set according to actual requirements. For example, the preset similarity threshold may be 99%.
  • transition video frames are added to the remaining first target pictures.
  • the image quality of the video can be ensured to a certain extent, and at the same time, the culling of the first target picture can be avoided to a greater extent.
  • the front and back of the video are not harmoniously connected, and the video playback is not smooth.
  • a target picture whose picture parameters do not meet the preset requirements can be determined, and then a second target picture is obtained.
  • the number of the determined second target pictures may be counted to obtain a fourth number.
  • a ratio of the fourth number to the total number of pictures in the video clip may be calculated to obtain a third ratio.
  • the third preset ratio threshold and the third preset quantity threshold may be set according to actual requirements. If the third ratio is greater than the third preset ratio threshold, it may be considered that the low-quality pictures in the video clip occupy the proportion of If the fourth number is greater than the third preset number threshold, it can be considered that the number of low-quality pictures in the video clip is large, and it can be determined that the overall quality of the video clip is poor. Therefore, the video clip can be directly discarded. video clips.
  • the overall quality of the video clip is measured in terms of the proportion of low-quality second target pictures and the specific quantity, which can avoid the existence of multiple second targets in the video clip due to the large number of pictures in the video clip.
  • the video clips whose third ratio is greater than the third preset ratio threshold, and/or the fourth number is greater than the third preset number threshold may also be displayed to the user, and the user receives the video clips for these video clips. select operation, and then delete the video clip selected by the selection operation to ensure the flexibility of user operation.
  • the proportion or specific number of the second target pictures is relatively high, that is, the When the overall quality of the video clip is poor, the video clip is directly discarded, thereby reducing the computation amount of subsequent related operations to a certain extent and saving processing resources.
  • the above-mentioned sub-steps may be executed before sub-step (2a), or may be executed after sub-step (2a).
  • it can be performed before sub-step (2a), so that unnecessary culling operations can be avoided, and processing resources can be saved to a greater extent.
  • the above-mentioned step of determining the third preset ratio threshold and the third preset quantity threshold may include: sub-step (2d1): determining, according to the video clip template corresponding to the video to be processed, the third preset ratio threshold and the third preset quantity threshold corresponding to the video clip templates; wherein, the third preset ratio thresholds corresponding to different video clip templates are different, and different video clip templates correspond to The third preset number threshold is different, and the third preset ratio threshold and the third preset number threshold are related to the video content type corresponding to the video clip template.
  • the video processing method in the embodiment of the present invention can be applied to an automatic video editing scene.
  • waste films pictures with lower quality
  • the automatic rejection method provided by the embodiments of the present invention can be performed online in real time, thereby reducing the time-consuming of rejection to a certain extent and ensuring the timeliness of processing.
  • FIG. 2 is a schematic diagram of an editing process provided by an embodiment of the present invention.
  • a user may first select a material (video to be processed), and then perform frame extraction according to a preset frame rate to obtain a picture sequence composed of target pictures. Then, through the CNN network, the fuzzy label information, jitter label information and exposure label information of the target image are determined. Select the clip template, select the clip suitable for clipping.
  • the clips suitable for editing may be the remaining video clips after being discarded through the above sub-step (2e).
  • the video clip template corresponding to the video to be processed may be selected from optional clip templates.
  • the optional editing template may be preset in the electronic device and may be used for video editing.
  • the optional editing template may include a sports template, a gourmet template, a dynamic template, and the like.
  • the editing template commonly used by the user may be determined according to the user's historical usage record. Then, the commonly used editing template is determined as the video editing template corresponding to the video to be processed.
  • the optional editing template may be directly displayed to the user, and the optional editing template selected by the user is determined as the video editing template corresponding to the video to be processed.
  • the third preset ratio threshold and the third preset quantity threshold corresponding to the video clip template corresponding to the video to be processed may be searched in the preset correspondence between the video clip template and the threshold.
  • the video is often edited according to the video editing template.
  • the editing methods of the video editing templates are often different.
  • the corresponding video content type is dynamic content, such as a video editing template that highlights the motion process of the moving subject in the video, such as motion template and dynamic template.
  • the focus is often on the coherence of the video screen. Since the content in such video images is often in motion, it is difficult for users to perceive low-quality images.
  • a video editing template that highlights the appearance of a static subject in a video, such as a gourmet template
  • the focus is often on the appearance details of the subject in the video screen.
  • the motion range of the content in the video picture is often small, so it is easier for the user to perceive that there are more low-quality pictures.
  • different thresholds may be set for different video editing templates according to the picture characteristics of each video editing template.
  • a higher third preset ratio threshold and a third preset number threshold are set, and in the case where the video content type corresponding to the video clip template is dynamic content, A lower third preset ratio threshold and third preset quantity threshold are set.
  • different third preset ratio thresholds and third preset quantity thresholds are correspondingly set for different video editing templates according to the video content types corresponding to the video editing templates, and when editing, according to the currently used Video clip template, select the corresponding threshold to filter video clips, and then to a certain extent, the screening operation can be more adapted to the current clipping needs, thereby improving the effect of clip screening.
  • FIG. 3 is a block diagram of a video processing apparatus provided by an embodiment of the present invention.
  • the apparatus may include: a memory 301 and a processor 302 .
  • the memory 301 is used to store program codes.
  • the processor 302 calls the program code, and when the program code is executed, is configured to perform the following operations: for the target picture in the video to be processed, determine the target information corresponding to the picture parameter of the target picture;
  • the target information is used to represent whether the picture parameters of the target picture meet the preset requirements; according to the target information of the target picture contained in the video to be processed, the video to be processed is screened; wherein, the screening process It includes deleting a target picture that does not meet the preset requirements or one of the video clips to which the target picture belongs; and generating a target video according to the video to be processed after screening.
  • the processor 302 includes deleting a target picture that does not meet the preset requirements or one of the video clips to which the target picture belongs; and generating a target video according to the
  • the picture parameters include one or more of a blur degree, a shake degree, an exposure degree, and a color change degree of the picture.
  • the processor 302 is further configured to: segment the video to be processed to obtain multiple video segments; for each of the video segments, according to the target information of the target picture included in the video segment, Screening processing is performed on the video clips.
  • the processor 302 is further configured to: for the target picture in the video to be processed, before determining the target information corresponding to the picture parameter of the target picture, determine the target picture according to the pictures included in the video to be processed.
  • the target picture is extracted; the target picture is extracted from the video to be processed, and timestamp information is added to each of the target pictures; the timestamp information is used to characterize the order of the target picture in the target picture; For the target picture in the video to be processed, after determining the target information corresponding to the picture parameter of the target picture, the target picture is combined into the video to be processed according to the timestamp information of the target picture.
  • the processor 302 is further configured to: in the case that the total number of pictures included in the to-be-processed video is greater than a first preset number threshold, extract data from the to-be-processed video according to a preset frame rate.
  • the processor 302 is further configured to: determine the first ratio of the first number of pixels with a fuzzy confidence level greater than a preset confidence threshold in the target picture to the total number of pixels in the target picture; According to the magnitude relationship between the first ratio and the first preset ratio threshold, the fuzzy label information is determined; the fuzzy label information is used to indicate whether the fuzzy degree of the target picture is normal; and/or, the target image is The picture is used as the input of the first preset classification model, and according to the output category of the first preset classification model, the shaking label information is determined; the first preset classification model is used for classifying pictures according to whether the shaking degree is normal; and/ Or, taking the target picture as the input of the second preset classification model, and determining the exposure label information according to the output category of the second preset classification model;
  • the processor 302 is further configured to: determine a third number of first target pictures in the video clip; the first target picture is target information indicating that the picture parameters do not meet the preset requirements and are continuous The target pictures that appear; if the third number is greater than the second preset number threshold, remove n of the first target pictures; the n is less than the third number.
  • the processor 302 is further configured to: add a picture transition effect to the remaining first target picture; wherein, the picture transition effect includes right-out, left-in, spin-out, spin-in, and fade-in and fade-out one or more of.
  • the processor 302 is further configured to: determine a third ratio between the fourth number of second target pictures included in the video clip and the total number of pictures in the video clip; the second target The picture is a target picture whose target information indicates that the picture parameters do not meet the preset requirements; determine a third preset ratio threshold and a third preset number threshold; if the third ratio is greater than the third preset ratio threshold, and /or, if the fourth quantity is greater than the third preset quantity threshold, the video clip is discarded.
  • the processor 302 is further configured to: determine, according to the video clip template corresponding to the video to be processed, the third preset ratio threshold and the third preset number corresponding to the video clip template threshold; wherein the third preset ratio threshold corresponding to different video clip templates is different, the third preset number threshold corresponding to different video clip templates is different, the third preset ratio threshold and the third preset ratio threshold
  • the three preset quantity thresholds are related to the video content type corresponding to the video clip template.
  • the processor 302 is further configured to: determine a third target picture whose all picture parameters do not meet the preset requirements; delete the third target picture; wherein the preset requirements include the One or more of normal blur, normal jitter, normal exposure, and normal color variation.
  • the processor 302 is further configured to: combine the screened and processed video clips to obtain the target video.
  • the processor 302 is further configured to: for the target picture in the video to be processed, before determining the target information corresponding to the picture parameter of the target picture, receive the user's selection operation on the optional video; The optional video selected by the selection operation is determined to be the video to be processed; the optional video is the video stored in the electronic device; after the target video is generated according to the screened video to be processed, the target video is displayed to the user. video.
  • the video processing apparatus determines, for the target picture in the video to be processed, target information corresponding to the picture parameters of the target picture, wherein the target information is used to indicate whether the picture parameters of the target picture meet the predetermined requirements.
  • Setting requirements according to the target information of the target picture contained in the video to be processed, the video to be processed is screened, wherein the screening process includes deleting the target picture that does not meet the preset requirements or one of the video clips to which the target picture belongs, and finally, Generate a target video according to the to-be-processed video after screening.
  • an embodiment of the present invention further provides a movable device, where the movable device includes a video capture device, and the movable device is configured to capture a video to be processed through the video capture device.
  • the method processes the video to be processed.
  • the movable device is a drone and/or an unmanned vehicle.
  • an embodiment of the present invention also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, each step in the above video processing method is implemented, and can To achieve the same technical effect, in order to avoid repetition, details are not repeated here.
  • the device embodiments described above are only illustrative, wherein the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in One place, or it can be distributed over multiple network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution in this embodiment. Those of ordinary skill in the art can understand and implement it without creative effort.
  • Various component embodiments of the present invention may be implemented in hardware, or in software modules running on one or more processors, or in a combination thereof.
  • a microprocessor or a digital signal processor may be used in practice to implement some or all of the functions of some or all of the components in the computing processing device according to the embodiments of the present invention.
  • the present invention can also be implemented as apparatus or apparatus programs (eg, computer programs and computer program products) for performing part or all of the methods described herein.
  • Such a program implementing the present invention may be stored on a computer-readable medium, or may be in the form of one or more signals. Such signals may be downloaded from Internet sites, or provided on carrier signals, or in any other form.
  • FIG. 4 is a block diagram of a computing processing device provided by an embodiment of the present invention. As shown in FIG. 4 , FIG. 4 shows a computing processing device that can implement the method according to the present invention.
  • the computing processing device traditionally includes a processor 710 and a computer program product or computer readable medium in the form of a memory 720 .
  • the memory 720 may be electronic memory such as flash memory, EEPROM (electrically erasable programmable read only memory), EPROM, hard disk, or ROM.
  • the memory 720 has storage space 730 for program code for performing any of the method steps in the above-described methods.
  • the storage space 730 for program codes may include various program codes for implementing various steps in the above methods, respectively.
  • These program codes can be read from or written to one or more computer program products.
  • These computer program products include program code carriers such as hard disks, compact disks (CDs), memory cards or floppy disks.
  • Such computer program products are typically portable or fixed storage units as described with reference to FIG. 5 .
  • the storage unit may have storage segments, storage spaces, etc. arranged similarly to the memory 720 in the computing processing device of FIG. 4 .
  • the program code may, for example, be compressed in a suitable form.
  • the storage unit includes computer readable code, ie code readable by a processor such as 710 for example, which when executed by a computing processing device, causes the computing processing device to perform each of the methods described above. step.
  • any reference signs placed between parentheses shall not be construed as limiting the claim.
  • the word “comprising” does not exclude the presence of elements or steps not listed in a claim.
  • the word “a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
  • the invention can be implemented by means of hardware comprising several different elements and by means of a suitably programmed computer. In a unit claim enumerating several means, several of these means can be embodied by one and the same item of hardware.
  • the use of the words first, second, and third, etc. do not denote any order. These words can be interpreted as names.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

La présente invention concerne un procédé et un appareil de traitement vidéo, un dispositif mobile et un support de stockage lisible. Le procédé comprend : pour une image cible dans une vidéo à traiter, la détermination d'informations cibles correspondant à un paramètre d'image de l'image cible, les informations cibles étant utilisées pour représenter si le paramètre d'image de l'image cible satisfait une exigence prédéfinie (101) ; selon les informations cibles de l'image cible comprise dans la vidéo à traiter, la réalisation d'un traitement de filtrage sur la vidéo à traiter, le traitement de filtrage comprenant la suppression de l'image cible, qui ne satisfait pas l'exigence prédéfinie, ou d'un des clips vidéo auxquels l'image cible appartient (102) ; et finalement, selon la vidéo à traiter soumise au traitement de filtrage, la génération d'une vidéo cible (103). Au moyen de la présente invention, durant un traitement vidéo, des images dans une vidéo sont automatiquement filtrées selon des paramètres d'image des images de sorte que les coûts de filtrage peuvent être réduits dans une certaine mesure, et l'efficacité du filtrage est améliorée.
PCT/CN2020/123998 2020-10-27 2020-10-27 Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible WO2022087826A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202080044431.1A CN114026874A (zh) 2020-10-27 2020-10-27 视频处理方法、装置、可移动设备及可读存储介质
PCT/CN2020/123998 WO2022087826A1 (fr) 2020-10-27 2020-10-27 Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/123998 WO2022087826A1 (fr) 2020-10-27 2020-10-27 Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible

Publications (1)

Publication Number Publication Date
WO2022087826A1 true WO2022087826A1 (fr) 2022-05-05

Family

ID=80053986

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/123998 WO2022087826A1 (fr) 2020-10-27 2020-10-27 Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible

Country Status (2)

Country Link
CN (1) CN114026874A (fr)
WO (1) WO2022087826A1 (fr)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114598925B (zh) * 2022-03-18 2023-10-20 脸萌有限公司 一种视频编辑方法、装置、设备及存储介质
CN114446331B (zh) * 2022-04-07 2022-06-24 深圳爱卓软科技有限公司 一种快速剪裁视频的视频编辑软件系统
CN114979705A (zh) * 2022-04-12 2022-08-30 杭州电子科技大学 一种基于深度学习、自注意力机制与符号推理的自动剪辑方法
CN116433538A (zh) * 2023-06-15 2023-07-14 加之创(厦门)科技有限公司 一种用于视频图像健康监测的图像处理方法、介质及设备

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105704398A (zh) * 2016-03-11 2016-06-22 咸阳师范学院 一种视频处理方法
CN107977463A (zh) * 2017-12-21 2018-05-01 广东欧珀移动通信有限公司 图片处理方法、装置、存储介质及终端
CN110650367A (zh) * 2019-08-30 2020-01-03 维沃移动通信有限公司 视频处理方法、电子设备及介质

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100132007A1 (en) * 2008-11-25 2010-05-27 Cisco Technology, Inc. Accelerating channel change time with external picture property markings
JP5332773B2 (ja) * 2009-03-18 2013-11-06 ソニー株式会社 画像処理装置および方法
CN105681825B (zh) * 2015-12-31 2018-12-07 深圳市瑞彩电子技术有限公司 一种视频码流输出控制装置和方法
EP3364342A1 (fr) * 2017-02-17 2018-08-22 Cogisen SRL Procédé de traitement d'image et de compression vidéo
CN109862394A (zh) * 2019-03-27 2019-06-07 北京周同科技有限公司 视频内容的审核方法、装置、设备和存储介质
CN110290320B (zh) * 2019-06-27 2021-01-22 Oppo广东移动通信有限公司 视频预览图生成方法和装置、电子设备、计算机可读存储介质
CN110929070A (zh) * 2019-12-09 2020-03-27 北京字节跳动网络技术有限公司 图像处理方法、装置、电子设备及存储介质
CN111090778B (zh) * 2019-12-26 2023-06-27 北京百度网讯科技有限公司 一种图片生成方法、装置、设备及存储介质
CN111143613B (zh) * 2019-12-30 2024-02-06 携程计算机技术(上海)有限公司 选择视频封面的方法、系统、电子设备与存储介质
CN111356016B (zh) * 2020-03-11 2022-04-22 北京小米松果电子有限公司 视频处理方法、视频处理装置及存储介质
CN111416950B (zh) * 2020-03-26 2023-11-28 腾讯科技(深圳)有限公司 视频处理方法、装置、存储介质及电子设备

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105704398A (zh) * 2016-03-11 2016-06-22 咸阳师范学院 一种视频处理方法
CN107977463A (zh) * 2017-12-21 2018-05-01 广东欧珀移动通信有限公司 图片处理方法、装置、存储介质及终端
CN110650367A (zh) * 2019-08-30 2020-01-03 维沃移动通信有限公司 视频处理方法、电子设备及介质

Also Published As

Publication number Publication date
CN114026874A (zh) 2022-02-08

Similar Documents

Publication Publication Date Title
WO2022087826A1 (fr) Procédé et appareil de traitement vidéo, dispositif mobile et support de stockage lisible
US11093754B2 (en) Method, system and apparatus for selecting frames of a video sequence
CN111866585B (zh) 一种视频处理方法及装置
CN111327945B (zh) 用于分割视频的方法和装置
US8379154B2 (en) Key-frame extraction from video
US10170157B2 (en) Method and apparatus for finding and using video portions that are relevant to adjacent still images
US9576202B1 (en) Systems and methods for identifying a scene-change/non-scene-change transition between frames
US8195038B2 (en) Brief and high-interest video summary generation
US8818037B2 (en) Video scene detection
CN107430780B (zh) 用于基于视频内容特性的输出创建的方法
CA3039239C (fr) Conformite du contenu media a la source de camera originale au moyen de reconnaissance optique de caracteres
WO2021017406A1 (fr) Procédé et appareil d'extraction de clip vidéo, dispositif et support de stockage
US20110255844A1 (en) System and method for parsing a video sequence
KR101709085B1 (ko) 컨볼루션 신경망을 이용한 샷 경계 검출 방법 및 장치
CN110139169B (zh) 视频流的质量评估方法及其装置、视频拍摄系统
CN110996183B (zh) 视频摘要的生成方法、装置、终端及存储介质
CN113542909A (zh) 视频处理方法、装置、电子设备及计算机存储介质
Husa et al. Automatic thumbnail selection for soccer videos using machine learning
CN114302226B (zh) 一种视频画幅智能裁剪方法
Husa et al. HOST-ATS: automatic thumbnail selection with dashboard-controlled ML pipeline and dynamic user survey
CN108985244B (zh) 一种电视节目类型识别方法及装置
US10923154B2 (en) Systems and methods for determining highlight segment sets
CN103517150B (zh) 蓝光播放器用以表示网络视频正在加载的方法及系统
Tsao et al. Thumbnail image selection for VOD services
CN113762016A (zh) 关键帧选取方法和装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20958991

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20958991

Country of ref document: EP

Kind code of ref document: A1