CN105741322B - A kind of field of view dividing method based on the fusion of video features layer - Google Patents

A kind of field of view dividing method based on the fusion of video features layer Download PDF

Info

Publication number
CN105741322B
CN105741322B CN201610072608.4A CN201610072608A CN105741322B CN 105741322 B CN105741322 B CN 105741322B CN 201610072608 A CN201610072608 A CN 201610072608A CN 105741322 B CN105741322 B CN 105741322B
Authority
CN
China
Prior art keywords
pixel
video
color
feature
coordinate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201610072608.4A
Other languages
Chinese (zh)
Other versions
CN105741322A (en
Inventor
张睿
童玉娟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Quzhou University
Original Assignee
Quzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Quzhou University filed Critical Quzhou University
Priority to CN201610072608.4A priority Critical patent/CN105741322B/en
Publication of CN105741322A publication Critical patent/CN105741322A/en
Application granted granted Critical
Publication of CN105741322B publication Critical patent/CN105741322B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of field of view dividing methods based on the fusion of video features layer.It includes the following steps:Calculate the color characteristic of each pixel in video;Calculate the dynamic feature of each pixel in video;Calculate the textural characteristics of each pixel in video;The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, region segmentation is carried out to the visual field in video according to fusion gained feature.The invention comprehensively utilizes the dynamic feature on video image vegetarian refreshments time dimension and the color characteristics and textural characteristics on Spatial Dimension, improve the validity and correctness of field of view segmentation.

Description

A kind of field of view dividing method based on the fusion of video features layer
Technical field
The present invention relates to video analysis processing technology field more particularly to a kind of visual field areas based on the fusion of video features floor Domain splitting method.
Background technology
With video technique it is continuous it is ripe decline with cost, video analysis and treatment technology have been widely used for scientific research, The numerous areas of production and social life.Carrying out region segmentation to the visual field presented in video helps to extract in video Valuable information, be a kind of important video analysis and treatment technology.
Currently, what is mainly used for reference for the region segmentation method of visual field in video is image region segmentation technology.Commonly Image region segmentation technology has the method based on color characteristic, the method based on textural characteristics, and the side based on shape feature Method etc..Obviously, image region segmentation technology is grafted directly in the video object, has ignored the abundant dynamic for including in video The time variation of feature and video content necessarily leads to the deficiency of field of view segmentation validity and correctness.
Invention content
The abundant dynamic for including in video the purpose of the present invention is overcoming existing field of view dividing method to have ignored is special The time variation of sign and video content leads to the technical problem of field of view segmentation validity and correctness deficiency, provides one The field of view dividing method that kind is merged based on video features layer, fully utilizes the dynamic on video image vegetarian refreshments time dimension Property the feature and color characteristic on Spatial Dimension and textural characteristics, improve the validity and correctness of field of view segmentation.
To solve the above-mentioned problems, the present invention is achieved by the following scheme:
A kind of field of view dividing method based on the fusion of video features layer of the present invention, includes the following steps:
S1:Calculate the color characteristic of each pixel in video;
S2:Calculate the dynamic feature of each pixel in video;
S3:Calculate the textural characteristics of each pixel in video;
S4:The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, root Region segmentation is carried out to the visual field in video according to fusion gained feature.
In the technical scheme, it using the field of view dividing method merged based on video features layer, fully utilizes and regards Dynamic feature on frequency pixel time dimension and the color characteristic on Spatial Dimension and textural characteristics, are utilized in video The Space Time Joint Distribution feature of visual information can not utilize vision to believe when overcoming to video capture image region segmentation method The deficiency of the temporal dynamic property feature of breath, the color video that this method is suitable for various resolution ratio fixed to visual field carry out visual field Region segmentation.
Preferably, the step S1 includes the following steps:
S11:One color feature vector based on RGB color, the RGB face are generated to each pixel of video Color characteristic vector is as follows:
f1(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t)
Wherein, R (i, j) |tPixel of the video in t frames at coordinate (i, j) is represented on red color channel Pixel value, G (i, j) |tRepresent pixel pixel on green color channel of the video in t frames at coordinate (i, j) Value, B (i, j) |tRepresent pixel value of pixel of the video in t frames at coordinate (i, j) on blue color channels;
S12:The RGB color of video is converted into hsv color space;
S13:One color feature vector based on hsv color space, the HSV face are generated to each pixel of video Color characteristic vector is as follows:
f2(i, j) |t=(H (i, j) |t, S (i, j) |t, V (i, j) |t)
Wherein, H (i, j) |tRepresent pixel pixel on tone channel of the video in t frames at coordinate (i, j) Value, S (i, j) |tRepresent pixel pixel value on saturation degree channel of the video in t frames at coordinate (i, j), V (i, j)|tRepresent pixel pixel value on luminance channel of the video in t frames at coordinate (i, j);
S14:By color feature vector and the color feature vector based on hsv color space based on RGB color into Row series connection, generates a color feature vector based on double color spaces, which indicates as follows:
f3(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t, H (i, j) |t, S (i, j) |t, V (i, j) |t)。
Preferably, the step S2 includes the following steps:
S21:Convert video to gray processing video;
S22:Background model is built to each pixel in gray processing video;
S23:The number of the conspicuousness gray-value variation occurred on each pixel in statistics gray processing video, conspicuousness Gray-value variation is defined as:Gray-value variation amplitude on one pixel is beyond the ash set by background model on the pixel Angle value normal variation range;
S24:The dynamic of each pixel in gray processing video is calculated, the calculation formula of the dynamic of pixel is:
Wherein, Ψ (i, j) |tRepresent pixel of the gray processing video from start frame at t frame time internal coordinates (i, j) The number of the conspicuousness gray-value variation of upper generation;D (i, j) |tRepresent pixel at coordinate (i, j) gray processing video from The frequency of conspicuousness gray-value variation, i.e., the dynamic of pixel at coordinate (i, j) occur in start frame to t frame times.
Preferably, the step S3 includes the following steps:
S31:Convert video to gray processing video, calculate gray processing video using original LBP operators is sitting in t frames The LBP texture values of the pixel at (i, j) are marked, and as the 1st texture eigenvalue W of the pixel1(i, j) |t
S32:The LBP of pixel of the gray processing video in t frames at coordinate (i, j) is calculated using round LBP operators Texture value, and as the 2nd texture eigenvalue W of the pixel2(i, j) |t
S33:By the 1st texture eigenvalue and the 2nd line of pixel of the gray processing video in t frames at coordinate (i, j) Reason eigenvalue cluster is combined into the texture feature vector of the pixel, i.e.,:f4(i, j) |t=(W1(i, j) |t, W2(i, j) |t)。
Preferably, the step S4 includes the following steps:
S41:The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, Obtain fusion feature vector;
S42:Automatically gathered using the fusion feature vector on all pixels point in video when clustering method pair t frames Alanysis;
S43:All pixels being classified as corresponding to the fusion feature vector of one kind by clustering are divided into together The region segmentation to visual field in video is completed in one region.
Substantial effect of the invention is:A variety of visual signatures of the video on time and Spatial Dimension are fully utilized, It is quiet due to increasing including the pixel dynamic feature on time dimension and the color on Spatial Dimension and textural characteristics Pixel dynamic feature this key message not having in state image uses image region segmentation method to overcome The problem of validity and correctness deficiency caused by region segmentation is carried out to video.
Description of the drawings
Fig. 1 is the work flow diagram of the present invention;
Fig. 2 is the schematic diagram of original LBP operators in step S31;
Fig. 3 is the schematic diagram of circle LBP operators in step S32.
Specific implementation mode
Below with reference to the embodiments and with reference to the accompanying drawing the technical solutions of the present invention will be further described.
Embodiment:A kind of field of view dividing method based on the fusion of video features layer of the present embodiment, as shown in Figure 1, Include the following steps:
S1:Calculate the color characteristic of each pixel in video;
S2:Calculate the dynamic feature of each pixel in video;
S3:Calculate the textural characteristics of each pixel in video;
S4:The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, root Region segmentation is carried out to the visual field in video according to fusion gained feature.
Step S1 includes the following steps:
S11:One color feature vector based on RGB color, the RGB face are generated to each pixel of video Color characteristic vector is as follows:
f1(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t)
Wherein, R (i, j) |tPixel of the video in t frames at coordinate (i, j) is represented on red color channel Pixel value, G (i, j) |tRepresent pixel pixel on green color channel of the video in t frames at coordinate (i, j) Value, B (i, j) |tRepresent pixel value of pixel of the video in t frames at coordinate (i, j) on blue color channels;
S12:The RGB color of video is converted into hsv color space;
S13:One color feature vector based on hsv color space, the HSV face are generated to each pixel of video Color characteristic vector is as follows:
f2(i, j) |t=(H (i, j) |t, S (i, j) |t, V (i, j) |t)
Wherein, H (i, j) |tRepresent pixel pixel on tone channel of the video in t frames at coordinate (i, j) Value, S (i, j) |tRepresent pixel pixel value on saturation degree channel of the video in t frames at coordinate (i, j), V (i, j)|tRepresent pixel pixel value on luminance channel of the video in t frames at coordinate (i, j);
S14:By color feature vector and the color feature vector based on hsv color space based on RGB color into Row series connection, generates a color feature vector based on double color spaces, which indicates as follows:
f3(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t, H (i, j) |t, S (i, j) |t, V (i, j) |t)。
Step S2 includes the following steps:
S21:Gray processing processing is carried out to video, converts video to gray processing video;
S22:Background model is built to each pixel in gray processing video;
S23:The number of the conspicuousness gray-value variation occurred on each pixel in statistics gray processing video, conspicuousness Gray-value variation is defined as:Gray-value variation amplitude on one pixel is beyond the ash set by background model on the pixel Angle value normal variation range, i.e. gray-value variation amplitude on a pixel is beyond set by background model on the pixel Gray value normal variation range is primary, and the conspicuousness gray-value variation number of the pixel adds 1;
S24:The dynamic of each pixel in gray processing video is calculated, the calculation formula of the dynamic of pixel is:
Wherein, Ψ (i, j) |tRepresent pixel of the gray processing video from start frame at t frame time internal coordinates (i, j) The number of the conspicuousness gray-value variation of upper generation;D (i, j) |tRepresent pixel at coordinate (i, j) gray processing video from The frequency of generation conspicuousness gray-value variation in start frame to t frame times, i.e., the dynamic of pixel at coordinate (i, j), as The dynamic of vegetarian refreshments refers to that the frequency of conspicuousness gray-value variation occurs on pixel, the pixel in the low expression video of dynamic The scene changes at place are small, and dynamic height indicates that the scene changes in video at the pixel are big.
Step S3 includes the following steps:
S31:Convert video to gray processing video, calculate gray processing video using original LBP operators is sitting in t frames The LBP texture values of the pixel at (i, j) are marked, and as the 1st texture eigenvalue W of the pixel1(i, j) |t, original LBP operators are as shown in Figure 2;
S32:The LBP of pixel of the gray processing video in t frames at coordinate (i, j) is calculated using round LBP operators Texture value, and as the 2nd texture eigenvalue W of the pixel2(i, j) |t, round LBP operators are as shown in Figure 3;
S33:By the 1st texture eigenvalue and the 2nd line of pixel of the gray processing video in t frames at coordinate (i, j) Reason eigenvalue cluster is combined into the texture feature vector of the pixel, i.e.,:f4(i, j) |t=(W1(i, j) |t, W2(i, j) |t)。
Step S4 includes the following steps:
S41:The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, Obtain fusion feature vector:
F (i, j) |t=(D (i, j) |t, R (i, j) |t, G (i, j) |t, B (i, j) |t, H (i, j) |t, S (i, j) |t, V (i, j) |t, W1(i, j) |t, W2(i, j) |t);
S42:Using the fusion feature vector f (i, j) on all pixels point in video when clustering method pair t frames |tIt carries out Automatic clustering;
S43:All pixels being classified as corresponding to the fusion feature vector of one kind by clustering are divided into together The region segmentation to visual field in video is completed in one region.
The dynamic of pixel refers to that the frequency of conspicuousness gray-value variation, the low expression video of dynamic occur on pixel In scene changes at the pixel it is small, dynamic height indicates that the scene changes in video at the pixel are big, using based on regarding The field of view dividing method of frequency Feature-level fusion, fully utilize dynamic feature on video image vegetarian refreshments time dimension and Color characteristic on Spatial Dimension and textural characteristics are utilized the Space Time Joint Distribution feature of visual information in video, overcome The deficiency of the temporal dynamic property feature of visual information, this method can not be utilized to be applicable in when to video capture image region segmentation method Field of view segmentation is carried out in the color video to the fixed various resolution ratio of visual field.

Claims (4)

1. a kind of field of view dividing method based on the fusion of video features layer, which is characterized in that include the following steps:
S1:Calculate the color characteristic of each pixel in video;
S2:Calculate the dynamic feature of each pixel in video;
S3:Calculate the textural characteristics of each pixel in video;
S4:The dynamic feature of each pixel in video, color characteristic and textural characteristics are subjected to Feature-level fusion, according to melting It closes gained feature and region segmentation is carried out to the visual field in video;
The step S2 includes the following steps:
S21:Convert video to gray processing video;
S22:Background model is built to each pixel in gray processing video;
S23:The number of the conspicuousness gray-value variation occurred on each pixel in statistics gray processing video, conspicuousness gray scale Value variation is defined as:Gray-value variation amplitude on one pixel is beyond the gray value set by background model on the pixel Normal variation range;
S24:The dynamic of each pixel in gray processing video is calculated, the calculation formula of the dynamic of pixel is:
Wherein, ψ (i, j) |tGray processing video is represented from start frame to occur on pixel at t frame time internal coordinates (i, j) Conspicuousness gray-value variation number;D (i, j) |tPixel at coordinate (i, j) is represented in gray processing video from start frame The frequency of conspicuousness gray-value variation, i.e., the dynamic of pixel at coordinate (i, j) occur in t frame times.
2. a kind of field of view dividing method based on the fusion of video features layer according to claim 1, which is characterized in that The step S1 includes the following steps:
S11:One color feature vector based on RGB color is generated to each pixel of video, the RGB color is special Sign vector is as follows:
f1(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t)
Wherein, R (i, j) |tRepresent pixel pixel on red color channel of the video in t frames at coordinate (i, j) Value, G (i, j) |tRepresent pixel pixel value on green color channel of the video in t frames at coordinate (i, j), B (i, j) |tRepresent pixel value of pixel of the video in t frames at coordinate (i, j) on blue color channels;
S12:The RGB color of video is converted into hsv color space;
S13:One color feature vector based on hsv color space is generated to each pixel of video, the hsv color is special Sign vector is as follows:
f2(i, j) | t=(H (i, j) |t, S (i, j) |t, V (i, j) |t)
Wherein, H (i, j) |tRepresent pixel pixel value on tone channel of the video in t frames at coordinate (i, j), S (i, j) |tPixel pixel value on saturation degree channel of the video in t frames at coordinate (i, j) is represented, V (i, j) |tGeneration Pixel pixel value on luminance channel of the table video in t frames at coordinate (i, j);
S14:Color feature vector based on RGB color is gone here and there with the color feature vector based on hsv color space Connection, generates a color feature vector based on double color spaces, which indicates as follows:
f3(i, j) |t=(R (i, j) |t, G (i, j) |t, B (i, j) |t, H (i, j) |t, S (i, j) |t, V (i, j) |t)。
3. a kind of field of view dividing method based on the fusion of video features layer according to claim 1 or 2, feature exist In the step S3 includes the following steps:
S31:It converts video to gray processing video, gray processing video is calculated in t frames in coordinate using original LBP operators The LBP texture values of pixel at (i, j), and as the 1st texture eigenvalue W of the pixel1(i, j) |t
S32:The LBP textures of pixel of the gray processing video in t frames at coordinate (i, j) are calculated using round LBP operators Value, and as the 2nd texture eigenvalue W of the pixel2(i, j) |t
S33:1st texture eigenvalue of pixel of the gray processing video in t frames at coordinate (i, j) and the 2nd texture is special Value indicative is combined as the texture feature vector of the pixel, i.e.,:f4(i, j) |t=(W1(i, j) |t, W2(i, j) |t)。
4. a kind of field of view dividing method based on the fusion of video features layer according to claim 1 or 2, feature exist In the step S4 includes the following steps:
S41:The dynamic feature of each pixel, color characteristic and textural characteristics in video are subjected to Feature-level fusion, are obtained Fusion feature vector;
S42:Automatically cluster minute is carried out using the fusion feature vector on all pixels point in video when clustering method pair t frames Analysis;
S43:All pixels being classified as by clustering corresponding to a kind of fusion feature vector are divided into same The region segmentation to visual field in video is completed in region.
CN201610072608.4A 2016-02-01 2016-02-01 A kind of field of view dividing method based on the fusion of video features layer Expired - Fee Related CN105741322B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610072608.4A CN105741322B (en) 2016-02-01 2016-02-01 A kind of field of view dividing method based on the fusion of video features layer

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610072608.4A CN105741322B (en) 2016-02-01 2016-02-01 A kind of field of view dividing method based on the fusion of video features layer

Publications (2)

Publication Number Publication Date
CN105741322A CN105741322A (en) 2016-07-06
CN105741322B true CN105741322B (en) 2018-08-03

Family

ID=56242193

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610072608.4A Expired - Fee Related CN105741322B (en) 2016-02-01 2016-02-01 A kind of field of view dividing method based on the fusion of video features layer

Country Status (1)

Country Link
CN (1) CN105741322B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108509830B (en) * 2017-02-28 2020-12-01 华为技术有限公司 Video data processing method and device
CN110807398A (en) * 2019-10-28 2020-02-18 衢州学院 Method and device for dividing field area
CN110910398B (en) * 2019-10-28 2021-07-20 衢州学院 Video complex scene region segmentation method and device based on decision layer fusion
CN110826445B (en) * 2019-10-28 2021-04-23 衢州学院 Method and device for detecting specific target area in colorless scene video
CN110866460B (en) * 2019-10-28 2020-11-27 衢州学院 Method and device for detecting specific target area in complex scene video
CN110827293B (en) * 2019-10-28 2020-09-08 衢州学院 Method and device for segmenting achromatic scene area based on decision-making layer fusion
CN110826446B (en) * 2019-10-28 2020-08-21 衢州学院 Method and device for segmenting field of view region of texture-free scene video
CN110910399B (en) * 2019-10-28 2020-09-15 衢州学院 Non-texture scene region segmentation method and device based on decision layer fusion
CN110796073B (en) * 2019-10-28 2021-05-25 衢州学院 Method and device for detecting specific target area in non-texture scene video
CN110807783B (en) * 2019-10-28 2023-07-18 衢州学院 Efficient visual field region segmentation method and device for achromatic long video
CN111028262A (en) * 2019-12-06 2020-04-17 衢州学院 Multi-channel composite high-definition high-speed video background modeling method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101316328A (en) * 2007-05-29 2008-12-03 中国科学院计算技术研究所 News anchor lens detection method based on space-time strip pattern analysis
CN102426583A (en) * 2011-10-10 2012-04-25 北京工业大学 Chinese medicine tongue manifestation retrieval method based on image content analysis
CN102915544A (en) * 2012-09-20 2013-02-06 武汉大学 Video image motion target extracting method based on pattern detection and color segmentation
CN105118049A (en) * 2015-07-22 2015-12-02 东南大学 Image segmentation method based on super pixel clustering

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101316328A (en) * 2007-05-29 2008-12-03 中国科学院计算技术研究所 News anchor lens detection method based on space-time strip pattern analysis
CN102426583A (en) * 2011-10-10 2012-04-25 北京工业大学 Chinese medicine tongue manifestation retrieval method based on image content analysis
CN102915544A (en) * 2012-09-20 2013-02-06 武汉大学 Video image motion target extracting method based on pattern detection and color segmentation
CN105118049A (en) * 2015-07-22 2015-12-02 东南大学 Image segmentation method based on super pixel clustering

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
基于SOFM的视频对象分割算法的研究;康达辉;《中国优秀硕士学位论文全文数据库 信息科技辑》;20061015(第2006年第10期);第22-30页 *
基于颜色特征和纹理特征的磨粒彩色图像分割;郭恒光 等;《润滑与密封》;20130630;第38卷(第6期);第94-97页 *
多目标矿业复杂图像特征提取与分类;王兰莎;《中国优秀硕士学位论文全文数据库 信息科技辑》;20120515(第2012年第05期);第24-25页 *

Also Published As

Publication number Publication date
CN105741322A (en) 2016-07-06

Similar Documents

Publication Publication Date Title
CN105741322B (en) A kind of field of view dividing method based on the fusion of video features layer
CN102063706B (en) Rapid defogging method
CN106897972A (en) A kind of self-adapting histogram underwater picture Enhancement Method of white balance and dark primary
CN110599415A (en) Image contrast enhancement implementation method based on local adaptive gamma correction
CN103248906B (en) Method and system for acquiring depth map of binocular stereo video sequence
CN106897981A (en) A kind of enhancement method of low-illumination image based on guiding filtering
EP3249618B1 (en) Banknote classification and identification method and device based on lab color space
CN105825503A (en) Visual-saliency-based image quality evaluation method
CN108830800A (en) The luminance raising Enhancement Method of image under a kind of half-light scene
CN107610093B (en) Full-reference image quality evaluation method based on similarity feature fusion
CN106875358A (en) Image enchancing method and image intensifier device based on Bayer format
CN110349097B (en) Color enhancement method for image significance and image processing device
US20170193644A1 (en) Background removal
CN105184757A (en) Food image color enhancement method based on color space characteristics
CN103164847A (en) Method for eliminating shadow of moving target in video image
CN108550124B (en) Illumination compensation and image enhancement method based on bionic spiral
CN103106671A (en) Method for detecting interested region of image based on visual attention mechanism
CN102768758B (en) Improved color image unsharp masking (USM) method
CN102542526A (en) Image decolorizing method
CN101873506B (en) Image processing method for providing depth information and image processing system thereof
US8351724B2 (en) Blue sky color detection technique
TWI707307B (en) Image adjusting method
CN110399886B (en) Screen image JND model construction method
CN110415816B (en) Skin disease clinical image multi-classification method based on transfer learning
CN111105383A (en) Image fusion color enhancement method for three-color vision

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20180803

Termination date: 20210201