CN108235116B - Feature propagation method and apparatus, electronic device, and medium - Google Patents

Feature propagation method and apparatus, electronic device, and medium Download PDF

Info

Publication number
CN108235116B
CN108235116B CN201711455916.6A CN201711455916A CN108235116B CN 108235116 B CN108235116 B CN 108235116B CN 201711455916 A CN201711455916 A CN 201711455916A CN 108235116 B CN108235116 B CN 108235116B
Authority
CN
China
Prior art keywords
current frame
frame
low
feature
features
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711455916.6A
Other languages
Chinese (zh)
Other versions
CN108235116A (en
Inventor
石建萍
李玉乐
林达华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sensetime Technology Development Co Ltd
Original Assignee
Beijing Sensetime Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sensetime Technology Development Co Ltd filed Critical Beijing Sensetime Technology Development Co Ltd
Priority to CN201711455916.6A priority Critical patent/CN108235116B/en
Publication of CN108235116A publication Critical patent/CN108235116A/en
Application granted granted Critical
Publication of CN108235116B publication Critical patent/CN108235116B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the invention discloses a feature propagation method and device, electronic equipment and a medium, wherein the method comprises the following steps: judging whether the current frame is a key frame; responding to the fact that the current frame is a non-key frame in a video, and acquiring high-layer characteristics of the current frame according to the low-layer characteristics of an adjacent previous key frame and the low-layer characteristics of the current frame; in the neural network, the network depth of the first network layer corresponding to the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame. The embodiment of the invention utilizes the consistency information between the video frames and the characteristic that the semantic labels between adjacent frames are similar to spread the semantic features of the video from the adjacent previous key frame to the current frame, thereby reducing the repeated calculation time and improving the accuracy of semantic segmentation.

Description

Feature propagation method and apparatus, electronic device, and medium
Technical Field
The present invention relates to computer vision technology, and in particular, to a feature propagation method and apparatus, an electronic device, a program, and a medium.
Background
Video semantic segmentation is an important issue in computer vision and video semantic understanding tasks. The video semantic segmentation model has important application in many fields, such as automatic driving, video monitoring, video target analysis and the like.
Currently, although semantic segmentation techniques for images are studied more, semantic segmentation techniques for videos are studied less. The video semantic segmentation requires high real-time performance, and meanwhile, enough precision can be guaranteed.
Disclosure of Invention
The embodiment of the invention provides a technical scheme for characteristic propagation in a video.
According to an aspect of the embodiments of the present invention, there is provided a feature propagation method, including:
judging whether the current frame is a key frame;
responding to the fact that the current frame is a non-key frame in a video, and acquiring high-level features of the current frame according to low-level features of a previous key frame adjacent to the current frame and the low-level features of the current frame; in the neural network, the network depth of the first network layer corresponding to the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
Optionally, in any one of the above method embodiments of the present invention, the obtaining, according to the low-layer feature of the previous key frame adjacent to the current frame and the low-layer feature of the current frame, the high-layer feature of the current frame from the high-layer feature of the previous key frame includes:
acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the adjacent previous key frame and the low-layer feature of the current frame;
and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight.
Optionally, in any of the above method embodiments of the present invention, in response to that the current frame is a non-key frame in a video, the method further includes:
and performing semantic segmentation on the current frame at least based on the high-level features of the current frame to obtain a semantic label of the current frame.
Optionally, in any one of the above method embodiments of the present invention, the performing semantic segmentation on the current frame based on at least a high-layer feature of the current frame includes:
and performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame to obtain a semantic label of the current frame.
Optionally, in any one of the above method embodiments of the present invention, performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame includes:
converting the low-layer features of the current frame to obtain features consistent with the number of channels of the high-layer features of the current frame;
splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame;
and performing semantic segmentation on the current frame based on the current frame characteristics.
Optionally, in any one of the method embodiments of the present invention, the determining whether the current frame is a key frame includes:
and judging whether the current frame is a key frame or not by using a key frame scheduling strategy.
Optionally, in any one of the method embodiments of the present invention, the determining whether the current frame is a key frame by using a key frame scheduling policy includes: judging whether the current frame is a key frame by using a fixed-length scheduling method;
in response to the current frame being a non-key frame in a video, the method further comprises: and extracting the features of the current frame to obtain the low-level features of the current frame.
Optionally, in any of the method embodiments of the present invention, determining whether the current frame is a key frame by using a key frame scheduling policy includes:
extracting the features of the current frame to obtain the low-level features of the current frame;
acquiring a scheduling probability value of the current frame scheduled as a key frame according to the low-layer characteristics of the previous key frame and the low-layer characteristics of the current frame;
and determining whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
Optionally, in any of the above method embodiments of the present invention, acquiring a scheduling probability value that the current frame is scheduled as a key frame according to the low-layer feature of the previous key frame and the low-layer feature of the current frame includes:
splicing the low-level features of the previous key frame and the low-level features of the current frame to obtain splicing features;
and acquiring a scheduling probability value of whether the current frame is scheduled as a key frame or not based on the splicing characteristics through a key frame scheduling network.
Optionally, in any of the above method embodiments of the present invention, the method further includes:
responding to the fact that the current frame is a key frame in a video, extracting features of the current frame, obtaining low-layer features of the current frame and caching the low-layer features;
and extracting the features of the low-level features of the current frame to obtain the high-level features of the current frame and caching the high-level features.
Optionally, in any of the above method embodiments of the present invention, the method further includes:
and responding to the fact that the current frame is a key frame in the video, performing semantic segmentation on the current frame based on the high-level features of the current frame, and obtaining a semantic label of the current frame.
According to another aspect of the embodiments of the present invention, there is provided a feature propagation device including:
the judging module is used for judging whether the current frame is a key frame;
the feature propagation module is used for responding to the fact that the current frame is a non-key frame in the video according to the judgment result of the judgment module, and acquiring the high-level feature of the current frame according to the low-level feature of a previous key frame adjacent to the current frame and the low-level feature of the current frame and the high-level feature of the previous key frame; in the neural network, the network depth of the first network layer of the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
Optionally, in any one of the apparatus embodiments of the present invention, the feature propagation module is specifically configured to:
acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the previous key frame and the low-layer feature of the current frame; and
and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight.
Optionally, in any one of the apparatus embodiments of the present invention, the apparatus further includes:
and the semantic segmentation module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, performing semantic segmentation on the current frame at least based on the high-level features of the current frame, and obtaining the semantic label of the current frame.
Optionally, in any one of the apparatus embodiments of the present invention, when the semantic segmentation module performs semantic segmentation on the current frame based on at least a high-level feature of the current frame, the semantic segmentation module is specifically configured to: and performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame.
Optionally, in any one of the apparatus embodiments of the present invention, when performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame, the semantic segmentation module is specifically configured to:
converting the low-layer features of the current frame to obtain features consistent with the number of channels of the high-layer features of the current frame;
splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame; and
and performing semantic segmentation on the current frame based on the current frame characteristics.
Optionally, in any one of the apparatus embodiments of the present invention, the determining module is specifically configured to determine whether the current frame is a key frame by using a key frame scheduling policy.
Optionally, in any one of the apparatus embodiments of the present invention, the determining module is specifically configured to determine whether the current frame is a key frame by using a fixed-length scheduling method;
the device further comprises:
and the first feature extraction module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, and performing feature extraction on the current frame to obtain the low-level features of the current frame.
Optionally, in any one of the apparatus embodiments of the present invention, the apparatus further includes:
the first feature extraction module is used for extracting features of the current frame to obtain low-level features of the current frame;
the acquisition module is used for acquiring the scheduling probability value of the current frame scheduled as the key frame according to the low-layer characteristics of the adjacent previous key frame and the low-layer characteristics of the current frame;
the judging module is specifically configured to determine whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
Optionally, in any one of the apparatus embodiments of the present invention, the obtaining module includes:
the splicing unit is used for splicing the low-layer feature of the previous key frame and the low-layer feature of the current frame to obtain a splicing feature;
and the key frame scheduling network is used for acquiring a scheduling probability value of whether the current frame is scheduled as a key frame or not based on the splicing characteristics.
Optionally, in any one of the apparatus embodiments of the present invention, the first feature extraction module is further configured to, according to a determination result of the determination module, respond that a current frame is a key frame in a video, perform feature extraction on the current frame, obtain a low-level feature of the current frame, and cache the low-level feature;
the device further comprises:
and the second feature extraction module is used for extracting features of the low-level features of the key frames, obtaining the high-level features of the key frames and caching the high-level features.
Optionally, in any one of the apparatus embodiments of the present invention, the semantic segmentation module is further configured to, according to a determination result of the determination module, respond that a current frame is a key frame in a video, and perform semantic segmentation on the current frame based on a high-level feature of the current frame to obtain a semantic tag of the current frame.
According to still another aspect of the embodiments of the present invention, there is provided an electronic apparatus including: the feature propagation device according to any of the above embodiments of the present invention.
According to another aspect of the embodiments of the present invention, there is provided another electronic device, including:
a processor and a feature propagation apparatus according to any of the above embodiments of the invention;
when the processor runs the feature propagation device, the units in the feature propagation device according to any of the above embodiments of the present invention are run.
According to still another aspect of the embodiments of the present invention, there is provided still another electronic device including: a processor and a memory;
the memory is used for storing at least one executable instruction, and the executable instruction enables the processor to execute the operation of each step in the feature propagation method according to any one of the above embodiments of the invention.
According to yet another aspect of the embodiments of the present invention, there is provided a computer program, including computer readable code, characterized in that when the computer readable code runs on a device, a processor in the device executes instructions for implementing the steps in the feature propagation method according to any one of the above-mentioned embodiments of the present invention.
According to another aspect of the embodiments of the present invention, there is provided a computer-readable medium for storing computer-readable instructions, which when executed, implement the operations of the steps in the feature propagation method according to any one of the above-mentioned embodiments of the present invention.
Based on the feature propagation method and apparatus, the electronic device, the program, and the medium provided in the above embodiments of the present invention, when the current frame is a non-key frame in a video, according to a low-level feature of a previous key frame adjacent to the current frame and a low-level feature of the current frame, a high-level feature of the current frame is obtained from a high-level feature of the previous key frame, so that semantic segmentation is performed on the non-key frame based on the high-level feature. The embodiment of the invention utilizes the consistency information between video frames and the characteristic that semantic labels between adjacent frames are similar, and spreads the high-level features for video semantic segmentation from the adjacent previous key frame to the current frame so as to carry out semantic segmentation on the current frame based on the high-level features of the current frame, and the high-level features for semantic segmentation are not required to be extracted frame by frame from the continuous frames of the video, so that the repeated calculation time is reduced compared with the mode of extracting the high-level features for semantic segmentation frame by frame; in addition, the embodiment of the invention transmits the high-level characteristics of the previous key frame to the current frame for semantic segmentation instead of directly transmitting the semantic labels, and improves the accuracy of semantic segmentation compared with the method of transmitting the semantic labels of the key frame by optical flow.
The technical solution of the present invention is further described in detail by the accompanying drawings and embodiments.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description, serve to explain the principles of the invention.
The invention will be more clearly understood from the following detailed description, taken with reference to the accompanying drawings, in which:
FIG. 1 is a flow chart of one embodiment of a feature propagation method of the present invention.
FIG. 2 is a flow chart of another embodiment of a feature propagation method of the present invention.
Fig. 3 is a flow chart of another embodiment of the feature propagation method of the present invention.
Fig. 4 is a schematic structural diagram of an embodiment of the feature dissemination device of the present invention.
Fig. 5 is a schematic structural view of another embodiment of the feature dissemination device of the present invention.
Fig. 6 is a schematic structural view of another embodiment of a feature dissemination device of the present invention.
Fig. 7 is a schematic structural diagram of an embodiment of an electronic device according to the present invention.
Detailed Description
Various exemplary embodiments of the present invention will now be described in detail with reference to the accompanying drawings. It should be noted that: the relative arrangement of the components and steps, the numerical expressions and numerical values set forth in these embodiments do not limit the scope of the present invention unless specifically stated otherwise.
Meanwhile, it should be understood that the sizes of the respective portions shown in the drawings are not drawn in an actual proportional relationship for the convenience of description.
The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the invention, its application, or uses.
Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail but are intended to be part of the specification where appropriate.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, further discussion thereof is not required in subsequent figures.
Embodiments of the invention are operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the computer system/server include, but are not limited to: personal computer systems, server computer systems, thin clients, thick clients, hand-held or laptop devices, microprocessor-based systems, set-top boxes, programmable consumer electronics, networked personal computers, minicomputer systems, mainframe computer systems, distributed cloud computing environments that include any of the above, and the like.
The computer system/server may be described in the general context of computer system-executable instructions, such as program modules, being executed by a computer system. Generally, program modules may include routines, programs, objects, components, logic, data structures, etc. that perform particular tasks or implement particular abstract data types. The computer system/server may be practiced in distributed cloud computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed cloud computing environment, program modules may be located in both local and remote computer system storage media including memory storage devices.
In the process of implementing the invention, the inventor finds out through research that in the existing video semantic segmentation method, a model for image semantic segmentation is directly applied to a video, and because continuous frames of the video have much redundant information, the information is not utilized in frame-by-frame processing, so that the calculation complexity is high; in another video semantic segmentation method, optical flow is used to propagate features from key frames to non-key frames, which uses a deep neural network to calculate semantic tags of key frames, then uses a small network to calculate optical flows of key frames and current frames, i.e. pixel-by-pixel displacement vectors in key frames and current frames, and then propagates semantic tags from key frames to current frames through optical flow, i.e.: the key frame propagation semantic tags are processed based on the pixel-by-pixel displacement vectors to obtain semantic tags of a current frame, and due to the fact that the images of the images in the video shake and blur caused by target motion and the like in the video, the obtained light stream is not accurate, and therefore semantic segmentation precision is reduced.
FIG. 1 is a flow chart of one embodiment of a feature propagation method of the present invention. As shown in fig. 1, the feature propagation method of this embodiment includes:
and 102, judging whether the current frame is a key frame.
For example, a key frame scheduling policy may be utilized to determine whether the current frame is a key frame.
And 104, responding to that the current frame is a non-key frame in the video, and acquiring the high-layer feature of the current frame according to the low-layer feature of the previous key frame adjacent to the current frame and the low-layer feature of the current frame.
In the neural network, the network depth of a first network layer for extracting the low-level features of the previous key frame and the current frame and the low-level features of the current frame is shallower than the network depth of a second network layer for extracting the low-level features in the neural network to obtain the high-level features.
In each embodiment of the invention, the neural network comprises more than two network layers with different network depths, wherein the network layer for extracting features in the network layers included in the neural network can be called as a feature layer, after the neural network receives a frame, the first feature layer is used for extracting the features of the input frame and inputting the frame into the second feature layer, each feature layer sequentially extracts the features of the input frame from the second feature layer, and the extracted features are input into the next network layer for feature extraction until the features for performing semantic segmentation are obtained. The network depth of each feature layer in the neural network is from shallow to deep according to the sequence of feature extraction, and the feature layer for feature extraction in the neural network can be divided into a low-layer feature layer and a high-layer feature layer according to the network depth, namely the first network layer and the second network layer. The features which are sequentially subjected to feature extraction by each feature layer in the low-layer feature layers and are finally output are called low-layer features, and the features which are sequentially subjected to feature extraction by each feature layer in the high-layer feature layers and are finally output are called high-layer features. Compared with a feature layer with shallow network depth in the same neural network, the feature layer with the deep network depth has a larger visual field and focuses on more spatial structure information, and when the extracted features are used for semantic segmentation, the semantic segmentation is more accurate, however, the deeper the network depth is, the higher the calculation difficulty and complexity are. In practical applications, the feature layers in the neural network may be divided into a lower-layer feature layer and a higher-layer feature layer according to a preset standard, such as a calculated amount, where the preset standard may be adjusted according to actual requirements. For example, for a neural network including 101 feature layers connected in sequence, the first 30 (or other numbers) feature layers from 1 st to 30 th among the 100 feature layers may be set as the lower feature layer, and the last 70 feature layers from 31 st to 100 th may be set as the upper feature layer according to the preset setting. For example, for a Pyramid Scene Parsing Network (PSPN), the neural Network may include four partial convolutional networks (conv1 to conv4) and a classification layer, each partial convolutional Network further includes a plurality of convolutional layers, and according to the size of the calculated amount, a convolutional layer from conv1 to conv4_3 in the PSPN may be used as a lower layer feature layer, which occupies the calculated amount of the PSPN about 1/8, and each convolutional layer from conv4_4 to the last classification layer in the PSPN may be used as an upper layer feature layer, which occupies the calculated amount of the PSPN about 7/8; and the classification layer is used for carrying out semantic segmentation on the high-level features output by the high-level feature layer.
Based on the feature propagation and display method provided by the above embodiment of the present invention, when the current frame is a non-key frame in the video, according to the low-level feature of the previous key frame adjacent to the current frame and the low-level feature of the current frame, the high-level feature of the current frame is obtained from the high-level feature of the previous key frame, so as to perform semantic segmentation on the non-key frame based on the high-level feature. The embodiment of the invention utilizes the consistency information between video frames and the characteristic that semantic labels between adjacent frames are similar, and spreads the high-level features for video semantic segmentation from the adjacent previous key frame to the current frame so as to carry out semantic segmentation on the current frame based on the high-level features of the current frame, and the high-level features for semantic segmentation are not required to be extracted frame by frame from the continuous frames of the video, so that the repeated calculation time is reduced compared with the mode of extracting the high-level features for semantic segmentation frame by frame; in addition, the embodiment of the invention transmits the high-level characteristics of the previous key frame to the current frame for semantic segmentation instead of directly transmitting the semantic labels, and improves the accuracy of semantic segmentation compared with the method of transmitting the semantic labels of the key frame by optical flow.
In one implementation manner of the embodiments of the present invention, in the operation 102, acquiring a high-level feature of the current frame from a high-level feature of a previous key frame according to a low-level feature of a previous key frame adjacent to the current frame and a low-level feature of the current frame may include:
acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the adjacent previous key frame and the low-layer feature of the current frame;
and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features and the conversion weight values of the previous key frame, wherein the features are propagated from the previous key frame and are also called propagation features.
In one alternative example, the transform weights for transforming from the low-level features of the previous key frame to the low-level features of the current frame may be obtained through a plurality of convolutional layers.
In another embodiment of the feature propagation method of the present invention, the method may further include: and responding to the fact that the current frame is a non-key frame in the video, and performing semantic segmentation on the current frame at least based on the high-level features of the current frame to obtain a semantic label of the current frame.
In one embodiment, performing semantic segmentation on the current frame based on at least the high-level features of the current frame may include: and performing semantic segmentation on the current frame based on the low-layer features and the high-layer features of the current frame to obtain a semantic label of the current frame.
In practical applications, the number of channels of the second network layer from which the high-level features are extracted is usually greater than the number of channels of the first network layer from which the low-level features are extracted, and in order to fuse the low-level features and the high-level features of the current frame, in one example, semantic segmentation is performed on the current frame based on the low-level features and the high-level features of the current frame, which may include:
converting the low-layer characteristics of the current frame to obtain characteristics consistent with the number of channels of the high-layer characteristics of the current frame;
splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame;
and performing semantic segmentation on the current frame based on the current frame characteristics.
In the embodiment of the invention, the high-level features of the current frame and the features of the current frame, which are acquired by the high-level features of the previous key frame, are fused for semantic segmentation, and the features of the non-key frame are acquired without using a single-frame model with high calculation cost, so that the calculation amount is reduced, and the accuracy of the semantic segmentation is ensured.
In addition, in another embodiment of the feature propagation method of the present invention, high-level features of non-key frames after a previous key frame may be cached, when a current frame is a non-key frame, features obtained by conversion of the current frame are spliced or fused with the high-level features of the current frame, the high-level features of the previous key frame, and the high-level features of the previous key frame and the non-key frames before the current frame, so as to obtain the features of the current frame, and based on the features of the current frame, the current frame is semantically segmented.
Based on the embodiment, all cached high-level features between the previous key frame and the current frame can be transmitted to the current frame, and the concatenation or fusion is carried out to carry out semantic segmentation, so that a more robust semantic segmentation effect can be obtained at a very low fusion cost.
In an implementation manner of each embodiment of the present invention, the key frame scheduling policy may be a fixed length scheduling method, for example, determining that a key frame is determined every l to 5 frames, that is: whether the current frame is a key frame can be judged by using a fixed-length scheduling method.
FIG. 2 is a flow chart of another embodiment of a feature propagation method of the present invention. As shown in fig. 2, the feature propagation method of this embodiment includes:
202, using a fixed length scheduling method to determine whether the current frame is a key frame.
If the current frame is a key frame, operation 212 is performed. Otherwise, if the current frame is a non-key frame in the video, operation 204 is performed.
And 204, extracting the features of the current frame (also called as the current non-key frame) to obtain the low-level features of the current frame.
In one example of the embodiments of the present invention, feature extraction may be performed on a current frame through a lower layer feature layer (i.e., a first network layer) of a neural network, so as to obtain a lower layer feature of the current frame.
And 206, acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the previous key frame adjacent to the current frame and the low-layer feature of the current frame.
The conversion weight may be a conversion matrix between two features, namely, a low-level feature of a previous key frame and a low-level feature of a current frame, and includes conversion elements between pixel-by-pixel features in the low-level feature of the previous key frame and the low-level feature of the current frame.
208, according to the high-level feature of the previous key frame and the conversion weight, converting the high-level feature of the previous key frame into the high-level feature of the current frame.
And 210, performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame to obtain a semantic label of the current frame.
The flow to semantic segmentation in step 210 is finished, and then the subsequent flow of this embodiment is not executed.
And 212, extracting features of the current frame (also called as the current key frame), obtaining low-level features of the current frame and caching the low-level features.
In one example, the current frame may be feature extracted by a lower feature layer (i.e., the first network layer) of the neural network.
214, extracting the features of the low layer of the current frame, obtaining the features of the high layer of the current frame and caching the features.
In one example, the feature extraction of the low-level features of the current frame may be performed by a high-level feature layer (i.e., a second network layer) of the neural network.
And 216, performing semantic segmentation on the current frame based on the high-level features of the current frame to obtain a semantic label of the current frame.
In the embodiments of the present invention, the key frames and the non-key frames may share a low-level feature layer of a neural network for low-level feature extraction, where the neural network may adopt PSPN, and the neural network may include four parts of convolutional networks (conv1 to conv4) and a classification layer, each part of convolutional network is further divided into a plurality of convolutional layers, where the low-level feature layer of the neural network may include convolutional layers from conv1 to conv4_3 in PSPN, and occupies a computation amount of PSPN about 7/8; the high-level feature layer of the neural network can comprise convolution layers from conv4_4 to the layer before the last classification layer, occupies the calculation amount of about 1/8 of the PSPN, and is used for extracting the high-level features of the key frames; the classification layer is used for correspondingly identifying the category of at least one pixel in the key frame or the non-key frame based on the high-level features of the key frame or the non-key frame, so that the semantic segmentation of the key frame or the non-key frame is realized.
In each embodiment of the invention, for the key frame, a single frame model with high calculation cost can be called, for example, PSPN is used for semantic segmentation, so that a high-precision semantic segmentation result is obtained. For non-key frames, the high-level features of the key frames can be adaptively transmitted to the current frame to obtain the high-level features of the current frame, consistency information between continuous frames of the video is fully utilized, repeated calculation time is avoided, and the current frame is subjected to semantic segmentation based on the low-level features and the high-level features of the current frame to obtain the semantic label of the current frame. The embodiment ensures the semantic segmentation precision of the key frame, does not need to adopt a single-frame model with high calculation cost to perform semantic segmentation on non-key frames frame by frame, reduces the calculation complexity and the calculation time, and saves the calculation resources.
Fig. 3 is a flow chart of another embodiment of the feature propagation method of the present invention. As shown in fig. 3, the feature propagation method of this embodiment includes:
and 302, extracting the features of the current frame to obtain the low-level features of the current frame.
In one example of the embodiments of the present invention, feature extraction may be performed on a current frame through a lower layer feature layer (i.e., a first network layer) of a neural network, so as to obtain a lower layer feature of the current frame.
And 304, acquiring a scheduling probability value of the current frame scheduled as the key frame according to the low-layer characteristics of the previous key frame adjacent to the current frame and the low-layer characteristics of the current frame.
In one example, the low-level features of the previous key frame and the low-level features of the current frame may be spliced, and the obtained splicing features are input into a key frame scheduling network, and the key frame scheduling network obtains a scheduling probability value of whether the current frame should be scheduled as a key frame based on the splicing features.
And 306, determining whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
If the current frame is a key frame, operation 314 is performed. Otherwise, if the current frame is a non-key frame in the video, operation 308 is performed.
308, according to the low-level features of the previous key frame adjacent to the current frame (also called current non-key frame) and the low-level features of the current frame, obtaining the conversion weight from the low-level features of the previous key frame to the low-level features of the current frame.
And 310, converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight values.
312, based on the low-level features and the high-level features of the current frame, performing semantic segmentation on the current frame to obtain a semantic tag of the current frame.
Thereafter, the subsequent flow of the present embodiment is not executed.
And 314, extracting features of the current frame (also called as the current key frame), obtaining low-layer features of the current frame and caching the low-layer features.
In one example, the current frame may be feature extracted by a lower feature layer (i.e., the first network layer) of the neural network.
And 316, extracting the features of the low layer of the current frame, obtaining the high layer of the current frame and caching.
In one example, the feature extraction of the low-level features of the current frame may be performed by a high-level feature layer (i.e., a second network layer) of the neural network.
And 318, performing semantic segmentation on the current frame based on the high-level features of the current frame to obtain a semantic label of the current frame.
The embodiment of the invention can be used for internet entertainment products such as automatic driving scenes, video monitoring scenes, portrait segmentation and the like, for example:
1, under the scene of automatic driving, the embodiment of the invention can be utilized to rapidly segment the targets in the video, such as people and vehicles;
2, in a video monitoring scene, people can be quickly segmented;
3, in the internet entertainment products such as portrait segmentation and the like, people can be rapidly segmented from video frames.
Those of ordinary skill in the art will understand that: all or part of the steps for implementing the method embodiments may be implemented by hardware related to program instructions, and the program may be stored in a computer readable storage medium, and when executed, the program performs the steps including the method embodiments; and the aforementioned storage medium includes: various media that can store program codes, such as ROM, RAM, magnetic or optical disks.
Fig. 4 is a schematic structural diagram of an embodiment of the feature dissemination device of the present invention. The feature propagation device of each embodiment of the present invention can be used to implement the feature propagation method of each embodiment described above. As shown in fig. 4, the feature propagation device of one embodiment includes: a judging module and a characteristic transmission module. Wherein:
and the judging module is used for judging whether the current frame is a key frame.
And the feature propagation module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, and acquiring the high-level feature of the current frame from the high-level feature of the previous key frame according to the low-level feature of the previous key frame adjacent to the current frame and the low-level feature of the current frame.
In the neural network, the network depth of the first network layer of the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
Based on the feature transmission playing device provided in the above embodiment of the present invention, when the current frame is a non-key frame in the video, according to the low-level feature of the previous key frame adjacent to the current frame and the low-level feature of the current frame, the high-level feature of the previous key frame is used to obtain the high-level feature of the current frame, so as to perform semantic segmentation on the non-key frame based on the high-level feature. The embodiment of the invention utilizes the consistency information between video frames and the characteristic that semantic labels between adjacent frames are similar, and spreads the high-level features for video semantic segmentation from the adjacent previous key frame to the current frame so as to carry out semantic segmentation on the current frame based on the high-level features of the current frame, and the high-level features for semantic segmentation are not required to be extracted frame by frame from the continuous frames of the video, so that the repeated calculation time is reduced compared with the mode of extracting the high-level features for semantic segmentation frame by frame; in addition, the embodiment of the invention transmits the high-level characteristics of the previous key frame to the current frame for semantic segmentation instead of directly transmitting the semantic labels, and improves the accuracy of semantic segmentation compared with the method of transmitting the semantic labels of the key frame by optical flow.
In one embodiment, the feature propagation module is specifically configured to: acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the previous key frame and the low-layer feature of the current frame; and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight.
Fig. 5 is a schematic structural view of another embodiment of the feature dissemination device of the present invention. As shown in fig. 5, compared with the embodiment shown in fig. 4, the feature propagation device of this embodiment further includes: and the semantic segmentation module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, performing semantic segmentation on the current frame at least based on the high-level features of the current frame, and obtaining a semantic label of the current frame.
In one embodiment, the semantic segmentation module is specifically configured to perform semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame when performing semantic segmentation on the current frame based on at least the high-layer feature of the current frame.
In one optional example, when performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame, the semantic segmentation module is specifically configured to: converting the low-layer characteristics of the current frame to obtain characteristics consistent with the number of channels of the high-layer characteristics of the current frame; splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame; and performing semantic segmentation on the current frame based on the current frame characteristics.
In an implementation manner of each of the above feature propagation device embodiments of the present invention, the determining module is specifically configured to determine whether the current frame is a key frame by using a key frame scheduling policy.
In one optional example, the determining module is specifically configured to determine whether the current frame is a key frame by using a fixed-length scheduling method. Accordingly, referring again to fig. 5, the feature propagation device of the further embodiment may further include: and the first feature extraction module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, and extracting features of the current frame to obtain low-level features of the current frame.
Alternatively, referring to fig. 6, in the feature propagation apparatus according to still another embodiment, a first feature extraction module and an acquisition module may be further included. Wherein: and the first feature extraction module is used for extracting the features of the current frame to obtain the low-level features of the current frame. And the obtaining module is used for obtaining the scheduling probability value of the current frame scheduled as the key frame according to the low-layer characteristics of the adjacent previous key frame and the low-layer characteristics of the current frame. Accordingly, in this embodiment, the determining module is specifically configured to determine whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
In one embodiment, the obtaining module may include: the splicing unit is used for splicing the low-layer characteristics of the previous key frame and the low-layer characteristics of the current frame to obtain splicing characteristics; and the key frame scheduling network is used for acquiring a scheduling probability value of whether the current frame is scheduled as the key frame or not based on the splicing characteristics.
For example, in the feature propagation device in each of the above embodiments, the first feature extraction module may be further configured to, according to the determination result of the determination module, perform feature extraction on the current frame in response to that the current frame is a key frame in the video, obtain a low-level feature of the current frame, and buffer the low-level feature. Referring again to fig. 5 or 6, in a further embodiment of the feature propagation device, it may further include: and the second feature extraction module is used for extracting features of the low-level features of the key frames according to the judgment result of the judgment module, obtaining the high-level features of the key frames and caching the high-level features.
Optionally, in the feature propagation device in each of the embodiments, the semantic segmentation module may be further configured to, according to a determination result of the determination module, respond that the current frame is a key frame in the video, perform semantic segmentation on the current frame based on a high-level feature of the current frame, and obtain a semantic tag of the current frame.
In addition, an embodiment of the present invention further provides an electronic device, including the feature propagation device according to any of the above embodiments of the present invention.
In addition, another electronic device is provided in an embodiment of the present invention, including:
a memory for storing executable instructions; and
one or more processors in communication with the memory for executing the executable instructions to perform the operations of the feature propagation method of any of the above embodiments of the present invention.
In addition, an embodiment of the present invention further provides another electronic device, including:
a processor and feature propagation apparatus of any of the above embodiments of the invention;
the elements of the feature propagation apparatus of any of the above embodiments of the invention are operated when the processor operates the feature propagation apparatus.
Fig. 7 is a schematic structural diagram of an embodiment of an electronic device according to the present invention. Referring now to fig. 7, shown is a schematic diagram of an electronic device suitable for use in implementing a terminal device or server of an embodiment of the present application. As shown in fig. 7, the electronic device includes one or more processors, a communication section, and the like, for example: one or more Central Processing Units (CPUs), and/or one or more image processors (GPUs), etc., which may perform various appropriate actions and processes according to executable instructions stored in a Read Only Memory (ROM) or loaded from a storage section into a Random Access Memory (RAM). The communication part may include, but is not limited to, a network card, which may include, but is not limited to, an ib (infiniband) network card, and the processor may communicate with the read-only memory and/or the random access memory to execute the executable instructions, connect with the communication part through the bus, and communicate with other target devices through the communication part, so as to complete operations corresponding to any method provided by the embodiments of the present application, for example, determine whether the current frame is a key frame; responding to the fact that the current frame is a non-key frame in a video, and acquiring high-level features of the current frame according to low-level features of a previous key frame adjacent to the current frame and the low-level features of the current frame; in the neural network, the network depth of the first network layer corresponding to the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
In addition, in the RAM, various programs and data necessary for the operation of the apparatus can also be stored. The CPU, ROM, and RAM are connected to each other via a bus. In the case of RAM, ROM is an optional module. The RAM stores executable instructions or writes executable instructions into the ROM during operation, and the executable instructions cause the processor to execute operations corresponding to any one of the methods of the invention. An input/output (I/O) interface is also connected to the bus. The communication unit may be integrated, or may be provided with a plurality of sub-modules (e.g., a plurality of IB network cards) and connected to the bus link.
The following components are connected to the I/O interface: an input section including a keyboard, a mouse, and the like; an output section including a display such as a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and the like, and a speaker; a storage section including a hard disk and the like; and a communication section including a network interface card such as a LAN card, a modem, or the like. The communication section performs communication processing via a network such as the internet. The drive is also connected to the I/O interface as needed. A removable medium such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive as necessary, so that a computer program read out therefrom is mounted into the storage section as necessary.
It should be noted that the architecture shown in fig. 7 is only an optional implementation manner, and in a specific practical process, the number and types of the components in fig. 7 may be selected, deleted, added or replaced according to actual needs; in different functional component settings, separate settings or integrated settings may also be used, for example, the GPU and the CPU may be separately set or the GPU may be integrated on the CPU, the communication part may be separately set or integrated on the CPU or the GPU, and so on. These alternative embodiments are all within the scope of the present disclosure.
In addition, the embodiment of the present invention further provides a computer storage medium for storing computer readable instructions, which when executed, implement the operations of the feature propagation method according to any one of the above embodiments of the present invention.
In addition, the embodiment of the present invention further provides a computer program, which includes computer readable instructions, and when the computer readable instructions are run in a device, a processor in the device executes executable instructions for implementing the steps in the feature propagation method according to any one of the above embodiments of the present invention.
In an alternative embodiment, the computer program is embodied in a Software product, such as a Software Development Kit (SDK), or the like.
In one or more alternative embodiments, the embodiment of the present invention further provides a computer program product for storing computer readable instructions, which when executed, make a computer execute the feature propagation method described in any one of the above possible implementation manners.
The computer program product may be embodied in hardware, software or a combination thereof. In an alternative example, the computer program product is embodied as a computer storage medium, and in another alternative example, the computer program product is embodied as a software product, such as an SDK or the like.
In the present specification, the embodiments are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same or similar parts in the embodiments are referred to each other. For the device embodiment, since it basically corresponds to the method embodiment, the description is relatively simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
The method and apparatus of the present invention may be implemented in a number of ways. For example, the methods and apparatus of the present invention may be implemented in software, hardware, firmware, or any combination of software, hardware, and firmware. The above-described order for the steps of the method is for illustrative purposes only, and the steps of the method of the present invention are not limited to the order specifically described above unless specifically indicated otherwise. Furthermore, in some embodiments, the present invention may also be embodied as a program recorded in a recording medium, the program including machine-readable instructions for implementing a method according to the present invention. Thus, the present invention also covers a recording medium storing a program for executing the method according to the present invention.
The description of the present invention has been presented for purposes of illustration and description, and is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to practitioners skilled in this art. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.

Claims (26)

1. A feature propagation method, comprising:
judging whether the current frame is a key frame;
responding to the fact that the current frame is a non-key frame in a video, and acquiring high-level features of the current frame according to low-level features of a previous key frame adjacent to the current frame and the low-level features of the current frame; in the neural network, the network depth of the first network layer corresponding to the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
2. The method according to claim 1, wherein said obtaining the high-level features of the current frame from the high-level features of the previous key frame according to the low-level features of the previous key frame adjacent to the current frame and the low-level features of the current frame comprises:
acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the adjacent previous key frame and the low-layer feature of the current frame;
and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight.
3. The method of claim 2, wherein in response to the current frame being a non-key frame in a video, further comprising:
and performing semantic segmentation on the current frame at least based on the high-level features of the current frame to obtain a semantic label of the current frame.
4. The method of claim 3, wherein the semantically segmenting the current frame based on at least the high-level features of the current frame comprises:
and performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame to obtain a semantic label of the current frame.
5. The method of claim 4, wherein performing semantic segmentation on the current frame based on the low-layer features and the high-layer features of the current frame comprises:
converting the low-layer features of the current frame to obtain features consistent with the number of channels of the high-layer features of the current frame;
splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame;
and performing semantic segmentation on the current frame based on the current frame characteristics.
6. The method according to any one of claims 1-5, wherein said determining whether the current frame is a key frame comprises:
and judging whether the current frame is a key frame or not by using a key frame scheduling strategy.
7. The method of claim 6, wherein the determining whether the current frame is a key frame using a key frame scheduling policy comprises: judging whether the current frame is a key frame by using a fixed-length scheduling method;
in response to the current frame being a non-key frame in a video, the method further comprises: and extracting the features of the current frame to obtain the low-level features of the current frame.
8. The method of claim 6, wherein determining whether the current frame is a key frame using a key frame scheduling policy comprises:
extracting the features of the current frame to obtain the low-level features of the current frame;
acquiring a scheduling probability value of the current frame scheduled as a key frame according to the low-layer characteristics of the previous key frame and the low-layer characteristics of the current frame;
and determining whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
9. The method of claim 8, wherein obtaining a scheduling probability value that the current frame is scheduled as a key frame according to the low-level features of the previous key frame and the low-level features of the current frame comprises:
splicing the low-level features of the previous key frame and the low-level features of the current frame to obtain splicing features;
and acquiring a scheduling probability value of whether the current frame is scheduled as a key frame or not based on the splicing characteristics through a key frame scheduling network.
10. The method of claim 6, further comprising:
responding to the fact that the current frame is a key frame in a video, extracting features of the current frame, obtaining low-layer features of the current frame and caching the low-layer features;
and extracting the features of the low-level features of the current frame to obtain the high-level features of the current frame and caching the high-level features.
11. The method of any of claims 1-5, further comprising:
and responding to the fact that the current frame is a key frame in the video, performing semantic segmentation on the current frame based on the high-level features of the current frame, and obtaining a semantic label of the current frame.
12. A feature propagation device, comprising:
the judging module is used for judging whether the current frame is a key frame;
the feature propagation module is used for responding to the fact that the current frame is a non-key frame in the video according to the judgment result of the judgment module, and acquiring the high-level feature of the current frame according to the low-level feature of a previous key frame adjacent to the current frame and the low-level feature of the current frame and the high-level feature of the previous key frame; in the neural network, the network depth of the first network layer of the low-level features of the previous key frame is extracted and obtained, and is shallower than the network depth of the second network layer corresponding to the high-level features of the previous key frame.
13. The apparatus of claim 12, wherein the feature propagation module is specifically configured to:
acquiring a conversion weight value converted from the low-layer feature of the previous key frame to the low-layer feature of the current frame according to the low-layer feature of the previous key frame and the low-layer feature of the current frame; and
and converting the high-level features of the previous key frame into the high-level features of the current frame according to the high-level features of the previous key frame and the conversion weight.
14. The apparatus of claim 13, further comprising:
and the semantic segmentation module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, performing semantic segmentation on the current frame at least based on the high-level features of the current frame, and obtaining the semantic label of the current frame.
15. The apparatus according to claim 14, wherein the semantic segmentation module, when performing semantic segmentation on the current frame based on at least the high-layer features of the current frame, is specifically configured to: and performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame.
16. The apparatus of claim 15, wherein the semantic segmentation module, when performing semantic segmentation on the current frame based on the low-layer feature and the high-layer feature of the current frame, is specifically configured to:
converting the low-layer features of the current frame to obtain features consistent with the number of channels of the high-layer features of the current frame;
splicing or fusing the features obtained by converting the current frame with the high-level features of the current frame to obtain the features of the current frame; and
and performing semantic segmentation on the current frame based on the current frame characteristics.
17. The apparatus according to any of claims 12-16, wherein the determining module is specifically configured to determine whether the current frame is a key frame by using a key frame scheduling policy.
18. The apparatus according to claim 17, wherein the determining module is specifically configured to determine whether the current frame is a key frame by using a fixed length scheduling method;
the device further comprises:
and the first feature extraction module is used for responding to the current frame as a non-key frame in the video according to the judgment result of the judgment module, and performing feature extraction on the current frame to obtain the low-level features of the current frame.
19. The apparatus of claim 17, further comprising:
the first feature extraction module is used for extracting features of the current frame to obtain low-level features of the current frame;
the acquisition module is used for acquiring the scheduling probability value of the current frame scheduled as the key frame according to the low-layer characteristics of the adjacent previous key frame and the low-layer characteristics of the current frame;
the judging module is specifically configured to determine whether the current frame is scheduled as a key frame according to the scheduling probability value of the current frame.
20. The apparatus of claim 19, wherein the obtaining module comprises:
the splicing unit is used for splicing the low-layer feature of the previous key frame and the low-layer feature of the current frame to obtain a splicing feature;
and the key frame scheduling network is used for acquiring a scheduling probability value of whether the current frame is scheduled as a key frame or not based on the splicing characteristics.
21. The apparatus according to claim 18, wherein the first feature extraction module is further configured to, according to the determination result of the determination module, perform feature extraction on a current frame in response to that the current frame is a key frame in a video, obtain a low-level feature of the current frame, and buffer the low-level feature;
the device further comprises:
and the second feature extraction module is used for extracting features of the low-level features of the key frames, obtaining the high-level features of the key frames and caching the high-level features.
22. The apparatus according to any one of claims 14 to 16, wherein the semantic segmentation module is further configured to, according to the determination result of the determination module, in response to that a current frame is a key frame in a video, perform semantic segmentation on the current frame based on a high-level feature of the current frame, and obtain a semantic tag of the current frame.
23. An electronic device, comprising: the feature propagation device of any one of claims 12 to 22.
24. An electronic device, comprising:
a processor and the feature propagation apparatus of any one of claims 12 to 22;
the elements of the signature propagation device as claimed in any one of claims 12-22 are operated when the processor operates the signature propagation device.
25. An electronic device, comprising: a processor and a memory;
the memory is configured to store at least one executable instruction that causes the processor to perform the operations of the steps of the feature propagation method of any of claims 1-11.
26. A computer-readable medium storing computer-readable instructions that, when executed, perform the operations of the steps of the feature propagation method of any one of claims 1-11.
CN201711455916.6A 2017-12-27 2017-12-27 Feature propagation method and apparatus, electronic device, and medium Active CN108235116B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711455916.6A CN108235116B (en) 2017-12-27 2017-12-27 Feature propagation method and apparatus, electronic device, and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711455916.6A CN108235116B (en) 2017-12-27 2017-12-27 Feature propagation method and apparatus, electronic device, and medium

Publications (2)

Publication Number Publication Date
CN108235116A CN108235116A (en) 2018-06-29
CN108235116B true CN108235116B (en) 2020-06-16

Family

ID=62649228

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711455916.6A Active CN108235116B (en) 2017-12-27 2017-12-27 Feature propagation method and apparatus, electronic device, and medium

Country Status (1)

Country Link
CN (1) CN108235116B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109151615B (en) * 2018-11-02 2022-01-25 湖南双菱电子科技有限公司 Video processing method, computer device, and computer storage medium
CN111383245B (en) * 2018-12-29 2023-09-22 北京地平线机器人技术研发有限公司 Video detection method, video detection device and electronic equipment
CN109919044A (en) * 2019-02-18 2019-06-21 清华大学 The video semanteme dividing method and device of feature propagation are carried out based on prediction
CN110060264B (en) * 2019-04-30 2021-03-23 北京市商汤科技开发有限公司 Neural network training method, video frame processing method, device and system
CN112465826B (en) * 2019-09-06 2023-05-16 上海高德威智能交通系统有限公司 Video semantic segmentation method and device
CN110738108A (en) * 2019-09-09 2020-01-31 北京地平线信息技术有限公司 Target object detection method, target object detection device, storage medium and electronic equipment
CN110929605A (en) * 2019-11-11 2020-03-27 中国建设银行股份有限公司 Video key frame storage method, device, equipment and storage medium
CN111062395B (en) * 2019-11-27 2020-12-18 北京理工大学 Real-time video semantic segmentation method
CN111654724B (en) * 2020-06-08 2021-04-06 上海纽菲斯信息科技有限公司 Low-bit-rate coding transmission method of video conference system
CN112016513B (en) * 2020-09-08 2024-01-30 北京达佳互联信息技术有限公司 Video semantic segmentation method, model training method, related device and electronic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101650728A (en) * 2009-08-26 2010-02-17 北京邮电大学 Video high-level characteristic retrieval system and realization thereof
CN103065300A (en) * 2012-12-24 2013-04-24 安科智慧城市技术(中国)有限公司 Method for video labeling and device for video labeling
CN106156747A (en) * 2016-07-21 2016-11-23 四川师范大学 The method of the monitor video extracting semantic objects of Behavior-based control feature
CN106934352A (en) * 2017-02-28 2017-07-07 华南理工大学 A kind of video presentation method based on two-way fractal net work and LSTM

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10062412B2 (en) * 2015-06-05 2018-08-28 Apple Inc. Hierarchical segmentation and quality measurement for video editing
US9818032B2 (en) * 2015-10-28 2017-11-14 Intel Corporation Automatic video summarization
CN105677735B (en) * 2015-12-30 2020-04-21 腾讯科技(深圳)有限公司 Video searching method and device
US10303984B2 (en) * 2016-05-17 2019-05-28 Intel Corporation Visual search and retrieval using semantic information

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101650728A (en) * 2009-08-26 2010-02-17 北京邮电大学 Video high-level characteristic retrieval system and realization thereof
CN103065300A (en) * 2012-12-24 2013-04-24 安科智慧城市技术(中国)有限公司 Method for video labeling and device for video labeling
CN106156747A (en) * 2016-07-21 2016-11-23 四川师范大学 The method of the monitor video extracting semantic objects of Behavior-based control feature
CN106934352A (en) * 2017-02-28 2017-07-07 华南理工大学 A kind of video presentation method based on two-way fractal net work and LSTM

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Deep Feature Flow for Video Recognition;Xizhou Zhu; Yuwen Xiong; Jifeng Dai; Lu Yuan; Yichen Wei;《IEEE Conference on Computer Vision and Pattern Recognition》;20171109;第3页第1列第7段-第6页第1列第6段,以及图1、2 *

Also Published As

Publication number Publication date
CN108235116A (en) 2018-06-29

Similar Documents

Publication Publication Date Title
CN108235116B (en) Feature propagation method and apparatus, electronic device, and medium
JP6837158B2 (en) Video identification and training methods, equipment, electronic devices and media
CN110458918B (en) Method and device for outputting information
CN108229341B (en) Classification method and device, electronic equipment and computer storage medium
CN108830235B (en) Method and apparatus for generating information
WO2018166438A1 (en) Image processing method and device and electronic device
WO2020000879A1 (en) Image recognition method and apparatus
CN111476871B (en) Method and device for generating video
US11164004B2 (en) Keyframe scheduling method and apparatus, electronic device, program and medium
US11811853B2 (en) Systems and methods for content delivery acceleration of virtual reality and augmented reality web pages
CN112862877B (en) Method and apparatus for training an image processing network and image processing
CN108337505B (en) Information acquisition method and device
CN112215171B (en) Target detection method, device, equipment and computer readable storage medium
CN113159091B (en) Data processing method, device, electronic equipment and storage medium
CN111681177B (en) Video processing method and device, computer readable storage medium and electronic equipment
CN109377508B (en) Image processing method and device
CN112784765B (en) Method, apparatus, device and storage medium for recognizing motion
US20210064919A1 (en) Method and apparatus for processing image
WO2020093724A1 (en) Method and device for generating information
CN114511041B (en) Model training method, image processing method, device, equipment and storage medium
WO2020205003A1 (en) Techniques to capture and edit dynamic depth images
CN115205150A (en) Image deblurring method, device, equipment, medium and computer program product
CN113033677A (en) Video classification method and device, electronic equipment and storage medium
CN109816023B (en) Method and device for generating picture label model
CN110852250B (en) Vehicle weight removing method and device based on maximum area method and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant