CN115866295A - Video key frame secondary extraction method and system for terminal row of convertor station - Google Patents

Video key frame secondary extraction method and system for terminal row of convertor station Download PDF

Info

Publication number
CN115866295A
CN115866295A CN202211474126.3A CN202211474126A CN115866295A CN 115866295 A CN115866295 A CN 115866295A CN 202211474126 A CN202211474126 A CN 202211474126A CN 115866295 A CN115866295 A CN 115866295A
Authority
CN
China
Prior art keywords
video
frame
value
difference
definition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211474126.3A
Other languages
Chinese (zh)
Inventor
谭林林
王嘉琦
程鑫
陈中
曹卫国
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202211474126.3A priority Critical patent/CN115866295A/en
Publication of CN115866295A publication Critical patent/CN115866295A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention provides a secondary extraction method and a secondary extraction system for a video key frame of a terminal row of a convertor station, and relates to the field of video processing. The secondary extraction method of the video key frame facing to the terminal strip of the convertor station comprises the following steps: carrying out video acquisition on a real object of a terminal row of the converter station; carrying out feature extraction and graying processing on each frame of picture in the video; subtracting the gray values of the corresponding pixel points of the two adjacent frames of images, and taking the absolute value of the gray values to obtain a difference value; comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to become a key frame of the video, and establishing a primary selection picture set; the problems that for a terminal row video shot manually in a current convertor station, the number of each frame of picture is large, the similarity of each frame of picture is high, and the definition of the picture is uneven are solved.

Description

Video key frame secondary extraction method and system for terminal row of convertor station
Technical Field
The invention relates to the technical field of video processing, in particular to a secondary extraction method and a secondary extraction system of a video key frame facing a terminal row of a convertor station.
Background
For video, the video is composed of many still pictures, which are called frames. Due to the large number of video frames and the high similarity between adjacent frames, the computer needs to spend much time on processing the video. In order to effectively reduce the operation time of a computer, video frames need to be screened and extracted, and after video key frames which can contain effective information are extracted, the key frames are processed, so that the operation time can be greatly reduced.
In an application scene of the converter station, in order to improve the workload of data acquisition of the terminal block on site, the data is acquired in a video mode. Due to the uncertainty of the definition of manual shooting and the similarity of shot objects, the conventional terminal row-oriented video key frame extraction technology has great limitations, and has the problems of large number of pictures, high similarity of each frame of picture and uneven definition of the pictures.
Disclosure of Invention
Technical problem to be solved
Aiming at the defects of the prior art, the invention provides a method and a system for extracting a video key frame twice for a terminal strip of a converter station, which solve the problems of huge number of pictures per frame, high similarity of the pictures per frame and uneven definition of the pictures for a terminal strip video shot manually in the current converter station.
(II) technical scheme
In order to realize the purpose, the invention is realized by the following technical scheme:
on the one hand, the method for extracting the video key frame secondarily facing to the convertor station terminal strip is provided, and comprises the following steps:
carrying out video acquisition on a real object of a terminal row of the converter station;
performing feature extraction and graying processing on each frame of picture in the video;
subtracting the gray values of the corresponding pixel points of the two adjacent frames of images, and taking the absolute value of the gray values to obtain a differential value;
comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to become a key frame of the video, and establishing a primary selection picture set;
performing convolution operation on the gray value of the pixel point corresponding to each image in the primary selection image set through a Laplace mask to obtain Conv;
calculating a standard deviation between Conv and a gray value of a corresponding pixel point of each image, wherein the standard deviation represents the definition of each image;
and setting a definition threshold, defining the pictures with the definition lower than the definition threshold as fuzzy, and defining the pictures with the definition higher than the definition threshold as clear.
Preferably, the video acquisition of the converter station terminal strip real object is realized through a high-definition camera, and the high-definition camera is installed in a place where the converter station indoor picture of the terminal strip real object can be shot clearly.
Preferably, the performing of feature extraction and graying processing on each frame of picture in the video specifically includes:
let the n frame and the n-1 frame in the video sequence be f n And f n-1 Gray levels of corresponding pixel points of two frames of images after feature extraction and gray level processingThe values are respectively denoted as f n (x, y) and f n-1 (x,y)。
Preferably, the subtraction is performed on the gray values of the corresponding pixels of the two adjacent frames of images, and the absolute value of the gray values is taken to obtain the difference value, wherein the specific implementation formula is as follows:
D n (x,y)=|f n (x,y)-f n-1 (x,y)|
wherein D is n Representing the differential value.
Preferably, the comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to be a key frame of the video, and establishing a primary selection picture set, specifically including:
comparing the difference value with a set difference threshold value, setting a difference value threshold value zeta between two adjacent frames when the difference value between two adjacent frames is larger and the difference between two adjacent frames is larger, and meeting the requirement that D is met if the difference value between one frame and the adjacent previous frame is higher than the set threshold value zeta n >ζ, the frame is extracted to become a key frame of the video, the first extraction of the key frame of the video is realized, and an initially selected picture set omega is established 1
Preferably, the performing convolution operation on the gray value of the pixel point corresponding to each image in the initially selected image set through a laplacian mask to obtain Conv specifically includes:
in the process of performing convolution operation on the gray value of the pixel point corresponding to each image in the primarily selected image set through a Laplace mask, the set omega is subjected to convolution operation 1 Setting the gray value of the corresponding pixel point of each image as f (x, y), and performing convolution operation through a Laplace mask to obtain Conv, wherein the Conv is represented by the following formula:
Figure BDA0003957688820000031
wherein
Figure BDA0003957688820000032
K is selected so that each array element is an integer and the sum of all array elements is zero.
In another aspect, a video key frame secondary extraction system facing a converter station terminal row is provided, and the system comprises a multi-lens primary video key frame extraction unit considering frame difference correlation and a multi-lens secondary key frame extraction unit considering definition complementarity;
the multi-shot primary video key frame extraction unit considering frame difference correlation comprises: performing feature extraction and graying processing on each frame of picture in the video, then performing difference on the gray values of two adjacent frames, comparing the difference value with a set difference threshold value, realizing the extraction of the key frame of the video for the first time, and establishing a primary selected picture set omega 1;
the multi-shot secondary key frame extraction unit considering the definition complementarity comprises a set omega 1 Performing convolution operation on the gray value of the pixel point corresponding to each frame of image through a Laplace mask to obtain Conv, and then calculating the standard deviation of the Conv and the gray value of the pixel point corresponding to each frame of image to obtain a value representing the definition of each frame of image; setting a definition threshold from the set omega 1 And screening out clear pictures to realize secondary extraction of the video key frames facing the terminal row of the convertor station.
In yet another aspect, an apparatus is provided, comprising:
one or more processors;
a memory for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to perform the method for video keyframe secondary extraction for a converter-oriented sub-row.
In yet another aspect, a computer readable storage medium is provided, which stores a computer program, which when executed by a processor implements a method for secondary extraction of video keyframes for a converter-oriented sub-strip as described.
(III) advantageous effects
The invention discloses a method and a system for extracting a video key frame twice for a terminal strip of a converter station, which solve the problems of huge number of pictures per frame, high similarity of the pictures per frame and uneven picture definition of a terminal strip video shot manually in the current converter station.
Drawings
Fig. 1 is a schematic flow chart of a secondary extraction method of video key frames for a terminal row of a converter station according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings of the present invention, and it is to be understood that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Examples
As shown in fig. 1, in an aspect, an embodiment of the present invention provides a method for extracting a video key frame twice for a terminal block of a converter station, including:
carrying out video acquisition on a real object of a terminal row of the converter station;
carrying out feature extraction and graying processing on each frame of picture in the video;
subtracting the gray values of the corresponding pixel points of the two adjacent frames of images, and taking the absolute value of the gray values to obtain a difference value;
comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to become a key frame of the video, and establishing a primary selection picture set;
performing convolution operation on the gray value of the pixel point corresponding to each image in the primary selection image set through a Laplace mask to obtain Conv;
calculating a standard deviation between Conv and a gray value of a corresponding pixel point of each image, wherein the standard deviation represents the definition of each image;
and setting a definition threshold, defining the pictures with the definition lower than the definition threshold as fuzzy, and defining the pictures with the definition higher than the definition threshold as clear.
Preferably, the video acquisition of the converter station terminal strip real object is realized through a high-definition camera, and the high-definition camera is installed in a place where the converter station indoor picture of the terminal strip real object can be shot clearly.
Preferably, the performing the feature extraction and the graying processing on each frame of picture in the video specifically includes:
let the n frame and the n-1 frame image in the video sequence be f respectively n And f n-1 The gray values of the corresponding pixel points of the two frames of images after the feature extraction and the graying processing are respectively recorded as f n (x, y) and f n-1 (x,y)。
Preferably, the subtraction is performed on the gray values of the corresponding pixels of the two adjacent frames of images, and the absolute value of the gray values is taken to obtain the difference value, wherein the specific implementation formula is as follows:
D n (x,y)=|F n (x,y)-F n-1 (x,y)|
wherein D is n Representing the differential value.
Preferably, the comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to become a key frame of the video, and establishing an initial selection picture set, specifically including:
comparing the difference value with a set difference threshold value, setting a difference value threshold value zeta between two adjacent frames when the difference value between two adjacent frames is larger and the difference between two adjacent frames is larger, and meeting the requirement that D is met if the difference value between one frame and the adjacent previous frame is higher than the set threshold value zeta n >Zeta, the frame is extracted to become a key frame of the video, the first extraction of the key frame of the video is realized, and a primary selection picture set omega is established 1
Preferably, the performing convolution operation on the gray value of the pixel point corresponding to each image in the initially selected image set through a laplacian mask to obtain Conv specifically includes:
in the process of performing convolution operation on the gray value of the pixel point corresponding to each image in the primarily selected image set through a Laplace mask, the set omega is subjected to convolution operation 1 The gray value of the corresponding pixel point of each image is set as f (x, y), generalConvolution operation is performed through the laplace mask to obtain Conv, which is shown as the following formula:
Figure BDA0003957688820000061
wherein
Figure BDA0003957688820000062
K is selected so that each array element is an integer and the sum of all array elements is zero.
As still another embodiment of the present invention, there is provided a video key frame secondary extraction system oriented to a substation terminal block, the system including a multi-shot primary video key frame extraction unit considering frame difference correlation and a multi-shot secondary key frame extraction unit considering sharpness complementarity;
the multi-shot primary video key frame extraction unit considering frame difference correlation comprises: performing feature extraction and graying processing on each frame of picture in the video, then performing difference on the gray values of two adjacent frames, comparing the difference value with a set difference threshold value, realizing the extraction of the key frame of the video for the first time, and establishing a primary selected picture set omega 1;
the multi-shot secondary key frame extraction unit considering sharpness complementarity comprises a set omega 1 Performing convolution operation on the gray value of the pixel point corresponding to each frame of image through a Laplace mask to obtain Conv, and then calculating the standard deviation of the Conv and the gray value of the pixel point corresponding to each frame of image to obtain a value representing the definition of each frame of image; setting a definition threshold from the set omega 1 And screening out clear pictures to realize secondary extraction of the video key frames facing the terminal row of the converter station.
As still another embodiment of the present invention, there is provided an apparatus including:
one or more processors;
a memory for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to perform a method of secondary video key frame extraction for a converter station sub-row as in the above embodiments.
As a further embodiment of the present invention, a computer readable storage medium storing a computer program is provided, which when executed by a processor implements a video key frame secondary extraction method for a terminal row of a converter station in the above embodiments.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrases "comprising a," "8230," "8230," or "comprising" does not exclude the presence of additional like elements in a process, method, article, or apparatus that comprises the element.

Claims (9)

1. A secondary extraction method of video key frames facing a convertor station terminal row is characterized by comprising the following steps:
carrying out video acquisition on a real object of a terminal row of the converter station;
carrying out feature extraction and graying processing on each frame of picture in the video;
subtracting the gray values of the corresponding pixel points of the two adjacent frames of images, and taking the absolute value of the gray values to obtain a differential value;
comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to become a key frame of the video, and establishing a primary selection picture set;
performing convolution operation on the gray value of the pixel point corresponding to each image in the primary selection image set through a Laplace mask to obtain Conv;
calculating a standard deviation between Conv and a gray value of a corresponding pixel point of each image, wherein the standard deviation represents the definition of each image;
and setting a definition threshold, defining the pictures with the definition lower than the definition threshold as fuzzy, and defining the pictures with the definition higher than the definition threshold as clear.
2. The method for secondary extraction of video keyframes from a terminal row of converter stations as claimed in claim 1, wherein: the video acquisition of the converter station terminal strip real object is realized through a high-definition camera, and the high-definition camera is arranged in a place which can clearly shoot the terminal strip real object picture in the converter station chamber.
3. The method for secondary extraction of video keyframes from a terminal row of converter stations as claimed in claim 1, wherein: the specific steps of carrying out feature extraction and graying processing on each frame of picture in the video comprise:
let the n frame and the n-1 frame in the video sequence be f n And f n-1 The gray values of the corresponding pixel points of the two frames of images after the feature extraction and the graying processing are respectively recorded as f n (x, y) and f n-1 (x,y)。
4. The method for secondary extraction of video keyframes from a terminal row of converter stations as claimed in claim 3, wherein: the gray values of the corresponding pixel points of the two adjacent frames of images are subtracted, and the absolute value of the gray values is taken to obtain a differential value, wherein the specific implementation formula is as follows:
D n (x,y)=|f n (x,y)-f n-1 (x,y)|
wherein D is n Representing the differential value.
5. The method for extracting video key frames secondarily from convertor station terminal blocks as claimed in claim 4, wherein the method comprises the following steps: comparing the difference value with a set difference threshold, if the difference value is higher than the set difference threshold, extracting the frame to be a key frame of the video, and establishing a primary selection picture set, which specifically comprises:
comparing the difference value with a set difference threshold value, setting a difference value threshold value zeta between two adjacent frames when the difference value between two adjacent frames is larger and the difference between two adjacent frames is larger, and meeting the requirement that D is met if the difference value between one frame and the adjacent previous frame is higher than the set threshold value zeta n >ζ, the frame is extracted to become a key frame of the video, the first extraction of the key frame of the video is realized, and an initially selected picture set omega is established 1
6. The method for secondary extraction of video keyframes from a terminal row of converter stations as claimed in claim 5, wherein: performing convolution operation on the gray value of the pixel point corresponding to each image in the initially selected image set through a laplacian mask to obtain Conv, specifically comprising:
in the process of performing convolution operation on the gray value of the pixel point corresponding to each image in the primarily selected image set through a Laplace mask, the set omega is subjected to convolution operation 1 Setting the gray value of the corresponding pixel point of each image as f (x, y), and performing convolution operation through a Laplace mask to obtain Conv, wherein the Conv is represented by the following formula:
Figure FDA0003957688810000021
/>
wherein
Figure FDA0003957688810000022
K is selected so that each array element is an integer and the sum of all array elements is zero.
7. A video key frame secondary extraction system facing a convertor station terminal row is characterized by comprising a multi-lens primary video key frame extraction unit considering frame difference correlation and a multi-lens secondary key frame extraction unit considering definition complementarity;
the multi-shot primary video key frame extraction unit considering frame difference correlation comprises: performing feature extraction and graying processing on each frame of picture in the video, then performing difference on the gray values of two adjacent frames, comparing the difference value with a set difference threshold value, realizing the extraction of the key frame of the video for the first time, and establishing a primary selected picture set omega 1;
the multi-shot secondary key frame extraction unit considering sharpness complementarity comprises a set omega 1 Performing convolution operation on the gray values of the corresponding pixel points of each frame of image through a Laplacian mask to obtain Conv, then calculating the standard deviation between the Conv and the gray values of the corresponding pixel points of each frame of image, and obtaining a value representing the definition of each frame of image; setting a definition threshold from the set omega 1 And screening out clear pictures to realize secondary extraction of the video key frames facing the terminal row of the converter station.
8. An apparatus, comprising:
one or more processors;
a memory for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to perform a method of video keyframe secondary extraction as recited in any one of claims 1-6 directed to a substation sub-row.
9. A computer-readable storage medium storing a computer program, wherein the program, when executed by a processor, implements a method for secondary extraction of video keyframes oriented to a converter station sub-row as claimed in any one of claims 1 to 6.
CN202211474126.3A 2022-11-22 2022-11-22 Video key frame secondary extraction method and system for terminal row of convertor station Pending CN115866295A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211474126.3A CN115866295A (en) 2022-11-22 2022-11-22 Video key frame secondary extraction method and system for terminal row of convertor station

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211474126.3A CN115866295A (en) 2022-11-22 2022-11-22 Video key frame secondary extraction method and system for terminal row of convertor station

Publications (1)

Publication Number Publication Date
CN115866295A true CN115866295A (en) 2023-03-28

Family

ID=85665335

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211474126.3A Pending CN115866295A (en) 2022-11-22 2022-11-22 Video key frame secondary extraction method and system for terminal row of convertor station

Country Status (1)

Country Link
CN (1) CN115866295A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116939197A (en) * 2023-09-15 2023-10-24 海看网络科技(山东)股份有限公司 Live program head broadcasting and replay content consistency monitoring method based on audio and video

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111275626A (en) * 2018-12-05 2020-06-12 深圳市炜博科技有限公司 Video deblurring method, device and equipment based on ambiguity
CN112149495A (en) * 2020-08-07 2020-12-29 中国矿业大学(北京) Video key frame extraction method based on parallax tracking
US20210142069A1 (en) * 2018-05-18 2021-05-13 Cambricon Technologies Corporation Limited Video retrieval method, and method and apparatus for generating video retrieval mapping relationship
CN112990191A (en) * 2021-01-06 2021-06-18 中国电子科技集团公司信息科学研究院 Shot boundary detection and key frame extraction method based on subtitle video

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210142069A1 (en) * 2018-05-18 2021-05-13 Cambricon Technologies Corporation Limited Video retrieval method, and method and apparatus for generating video retrieval mapping relationship
CN111275626A (en) * 2018-12-05 2020-06-12 深圳市炜博科技有限公司 Video deblurring method, device and equipment based on ambiguity
CN112149495A (en) * 2020-08-07 2020-12-29 中国矿业大学(北京) Video key frame extraction method based on parallax tracking
CN112990191A (en) * 2021-01-06 2021-06-18 中国电子科技集团公司信息科学研究院 Shot boundary detection and key frame extraction method based on subtitle video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
蓝章礼;帅丹;李益才;: "基于相关系数的道路监控视频关键帧提取算法", 重庆交通大学学报(自然科学版), no. 01, 15 February 2016 (2016-02-15) *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116939197A (en) * 2023-09-15 2023-10-24 海看网络科技(山东)股份有限公司 Live program head broadcasting and replay content consistency monitoring method based on audio and video

Similar Documents

Publication Publication Date Title
US6912313B2 (en) Image background replacement method
US8582915B2 (en) Image enhancement for challenging lighting conditions
US9912839B2 (en) Method for conversion of a saturated image into a non-saturated image
CN105469375B (en) Method and device for processing high dynamic range panorama
US20080031339A1 (en) Image matching device and method for motion pictures
WO2003036557A1 (en) Method and apparatus for background segmentation based on motion localization
CN107135401B (en) Key frame selection method and system
CN108335272B (en) Method and device for shooting picture
EP1542152A1 (en) Object detection
CN110276769B (en) Live broadcast content positioning method in video picture-in-picture architecture
EP1665806A1 (en) Motion vector field re-timing
CN113242428B (en) Post-processing acceleration method based on ROI (region of interest) in video conference scene
CN115866295A (en) Video key frame secondary extraction method and system for terminal row of convertor station
US11373279B2 (en) Image processing method and device
CN117333398A (en) Multi-scale image denoising method and device based on self-supervision
CN111460964A (en) Moving target detection method under low-illumination condition of radio and television transmission machine room
CN114419102A (en) Multi-target tracking detection method based on frame difference time sequence motion information
RU2669470C1 (en) Device for removing logos and subtitles from video sequences
CN111654747B (en) Bullet screen display method and device
CN113965814A (en) Multi-meeting-place key frame extraction method and system based on video meeting scene
CN108898566B (en) Low-illumination color video enhancement method using space-time illumination map
US8582882B2 (en) Unit for and method of segmentation using average homogeneity
CN107292803B (en) Reversible robust digital image watermarking method according with human eye observation rule
WO2016113407A1 (en) Methods and apparatus for groupwise contrast enhancement
CN112465853B (en) Background conversion method and device for video picture, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination