CN102521805B - Video word processing method based on interframe information - Google Patents

Video word processing method based on interframe information Download PDF

Info

Publication number
CN102521805B
CN102521805B CN 201110391472 CN201110391472A CN102521805B CN 102521805 B CN102521805 B CN 102521805B CN 201110391472 CN201110391472 CN 201110391472 CN 201110391472 A CN201110391472 A CN 201110391472A CN 102521805 B CN102521805 B CN 102521805B
Authority
CN
China
Prior art keywords
frame
new
present image
text filed
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN 201110391472
Other languages
Chinese (zh)
Other versions
CN102521805A (en
Inventor
田岩
许毅平
文灏
陈柱
孙福生
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN 201110391472 priority Critical patent/CN102521805B/en
Publication of CN102521805A publication Critical patent/CN102521805A/en
Application granted granted Critical
Publication of CN102521805B publication Critical patent/CN102521805B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a video word processing method based on interframe information. The method comprises the following steps of: detecting a text area R(x, y, t) of a current image; checking the text area R(x, y, t) of a current image f(x, y, t); performing background restoration on the text area R(x, y, t) from the first frame to the last frame; and performing the background restoration from the last frame to the first frame. The method has the advantages of ensuring the accuracy of detection positioning, ensuring the correctness of a restoration result, and ensuring the time continuity of a video result.

Description

Video text disposal route based on inter-frame information
Technical field
The invention belongs to video image and handle application, be specifically related to a kind of video text disposal route based on inter-frame information.
Background technology
During video image was handled, literal extracted and the image information reparation all has very application prospects, therefore, also more and more is subject to people's attention in recent years.
Yet the method in positioning image Chinese version zone also is difficult to accomplish complete detection and extraction for the literal in some complex background automatically at present.Present image repair algorithm mainly can be divided into two big classes based on the restorative procedure of PDE with based on the restorative procedure of texture.The main two kinds of different thinkings of supposition and optimizing that embodied respectively.These methods have been ignored frame of video correlativity in time often based on Flame Image Process, thereby it is inaccurate to cause repairing the result, and repair the result occur flicker, phenomenons such as saltus step easily when playing.
Summary of the invention
The object of the present invention is to provide a kind of video text disposal route based on inter-frame information, it can solve and repair inaccurate and reparation result easy flicker, the problems such as saltus step of occurring when playing of result in the existing method.
The present invention is achieved by the following technical solutions:
A kind of video text disposal route based on inter-frame information comprises the steps:
(1) (t), establish present image is f (x to the text filed R of detection present image for x, y, y, t), the N width of cloth image of its front is followed successively by f (x, y, t-1) ..., f (x, y, t-N), this N width of cloth image correspondence text filed be respectively R (x, y, t-1),, R (x, y, t-N), wherein x represents the horizontal ordinate of present image, and y represents the ordinate of present image, and t represents the frame number of present image;
(2) to present image f (x, y, text filed R t) (x, y t) carry out verification, specifically comprise following substep:
(21) judgement is text filed
Figure BDA0000114642060000021
Whether the subregion among the t is text filed
Figure BDA0000114642060000022
In occurred;
(22) if do not occur, can determine that then this subregion is the false-alarm zone, and this subregion is got rid of, and (x, y t) are updated to R with text filed R New(x, y, t), R wherein New(x, y t) are area to be repaired in the present image;
(23) if occurred, then continue to text filed R (x, y, t) in other subregions carry out verification;
(3) from front to back to text filed R (x, y t) carry out the background reparation, specifically comprise following substep:
(31) (x, y t) and the motion excursion of preceding N frame, if motion is bigger, directly handle next frame to analyze present image f; Less if move, then carry out step (32);
(32) with the text filed R of present image New(x, y, t) and the text filed R of preceding N frame (x, y, t-1) ..., (x, y t-N) compare R, and the zone that obtains utilizing inter-frame information to repair in the present image is R New(x, y, t)-R (x, y, t-1) ..., R New(x, y, t)-R (x, y, t-N);
(33) (x, y is in t) to present image f The zone, the search best matching blocks is finished reparation in preceding N frame, and upgrades R New(x, y, t);
(34) repeating step (3) is handled back one frame, until processes complete section video.(4) begin to carry out the background reparation forward from last frame, specifically comprise following substep:
(41) (x, y t) with the motion excursion of back N frame, if motion is bigger, directly handle front one frame, if it is less to move, then carry out step (42) to analyze present image f;
(42) with the text filed R of present image New(x, y, t) with the text filed R of back N frame (x, y, t+1) ..., (x, y t+N) compare R, and the zone that obtains utilizing inter-frame information to repair in the present image is R New(x, y, t)-R (x, y, t+1) ..., R New(x, y, t)-R (x, y, t+N);
(43) (x, y is in t) to present image f
Figure BDA0000114642060000031
Figure BDA0000114642060000032
The zone, the search best matching blocks is finished reparation in the N frame of back, and upgrades R New(x, y, t);
(44) if this moment To remaining text filed, the match block that search is best in this frame is repaired so; Otherwise repeating step (4) is handled front one frame, up to processes complete section video.
Method of the present invention has the following advantages: (1) based on the continuity of video text captions, the inventive method proposes to utilize inter-frame information verification surveyed area, improves the accuracy of zone location; (2) based on the continuity of video content, the inventive method utilizes the interframe available information to finish reparation, can find optimum matching information fast like this, has guaranteed to repair result's correctness again; (3) the inventive method result that will at every turn repair is as the available information of repairing next time, constantly accumulates and continue the continuity of repairing the result, and the interframe that can reduce image processing method to a great extent and produced is glimmered and saltus step.Generally speaking, the present invention can guarantee the accuracy of detection and location, guarantees to repair result's correctness, guarantees the time continuity of results for video.
Description of drawings
Fig. 1 (a) is an original image.
Fig. 1 (b) illustrates through the image after the art methods processing.
Fig. 1 (c) illustrates through the image after the inventive method processing.
Fig. 2 is the process flow diagram that the present invention is based on the video text disposal route of inter-frame information.
Fig. 3 is the refinement process flow diagram of step in the inventive method (2).
Fig. 4 is the refinement process flow diagram of step in the inventive method (3).
Fig. 5 is the refinement process flow diagram of step in the inventive method (4).
Embodiment
Below in conjunction with accompanying drawing and concrete case study on implementation the present invention is further described.
Below at first terms more of the present invention are used to make an explanation:
Image text filed: image Chinese version region.
False-alarm zone: be detected as text filed non-text filed in the image.
Background is repaired: removes text filedly, repair also and reduce by text filed background of blocking.
Motion excursion: the variable quantity of position in the frame of front and back, a certain zone.
Best matching blocks: in the hunting zone with the most close zone on certain tolerance mode, area to be repaired.
Shown in Fig. 2,3,4,5, the video text disposal route that the present invention is based on inter-frame information may further comprise the steps:
(1) (t), establish present image is f (x to the text filed R of detection present image for x, y, y, t), the N width of cloth image of its front is followed successively by f (x, y, t-1) ..., f (x, y, t-N), this N width of cloth image correspondence text filed be respectively R (x, y, t-1),, R (x, y, t-N), wherein x represents the horizontal ordinate of present image, and y represents the ordinate of present image, and t represents the frame number of present image;
(2) to present image f (x, y, text filed R t) (x, y t) carry out verification, specifically comprise following substep:
(21) judgement is text filed
Figure BDA0000114642060000041
In subregion whether text filed
Figure BDA0000114642060000042
In occurred;
(22) if do not occur, can determine that then this subregion is the false-alarm zone, and this subregion is got rid of, and (x, y t) are updated to R with text filed R New(x, y, t), R wherein New(x, y t) are area to be repaired in the present image;
(23) if occurred, then continue to text filed R (x, y, t) in other subregions carry out verification;
(3) from front to back to text filed R (x, y t) carry out the background reparation, specifically comprise following substep:
(31) analyze present image f (x, y, t) and the motion excursion of preceding N frame; If it is bigger to move, directly next frame is handled; Less if move, then carry out step (32);
(32) with the text filed R of present image New(x, y, t) and the text filed R of preceding N frame (x, y, t-1) ..., (x, y t-N) compare R, and the zone that obtains utilizing inter-frame information to repair in the present image is R New(x, y, t)-R (x, y, t-1) ..., R New(x, y, t)-R (x, y, t-N);
(33) (x, y is in t) to present image f
Figure BDA0000114642060000051
Figure BDA0000114642060000052
The zone, the search best matching blocks is finished reparation in preceding N frame, and upgrades R New(x, y, t);
(34) repeating step (3) is handled back one frame, until processes complete section video.(4) begin to carry out the background reparation forward from last frame, specifically comprise following substep:
(41) analyze present image f (x, y is t) with the motion excursion of back N frame; If it is bigger to move, directly front one frame is handled; Less if move, then carry out step (42);
(42) with the text filed R of present image New(x, y, t) with the text filed R of back N frame (x, y, t+1) ..., (x, y t+N) compare R, and the zone that obtains utilizing inter-frame information to repair in the present image is R New(x, y, t)-R (x, y, t+1) ..., R New(x, y, t)-R (x, y, t+N);
(43) (x, y is in t) to present image f The zone, the search best matching blocks is finished reparation in the N frame of back, and upgrades R New(x, y, t);
(44) if this moment
Figure BDA0000114642060000055
To remaining text filed, the match block that search is best in this frame is repaired so; Otherwise repeating step (4) is handled front one frame, up to processes complete section video.
For the validity based on the video text disposal route of inter-frame information is analyzed, the inventor is respectively from literal type, the background complexity, and whether background motion speed, caption character length occurs changes these 4 aspects making video test sequence of starting with; Wherein the difference of literal type is represented the character features difference, mainly influences the universality of algorithm in the test section; Background complexity difference mainly influences algorithm in the accuracy of test section and the correctness of eliminating false-alarm regions mechanism; The background motion speed, can the reparation part that have influence on algorithm use inter-frame information; Whether caption character length occurs changes, and inter-frame information that the reparation part of algorithm can utilize what have influence on.
According to the independent component analysis as can be known, these 4 factors all are independently of one another in video; They have reflected that respectively algorithm detects and repair the performance of each link, and these 4 features of video can directly visually distinguish, therefore for selecting and to make video test sequence all very useful.
7 video test sequence have been made according to the various combination of 4 factors in the experiment, according to statistical testing of business cycles to multitude of video, independent factor in these 7 videos meets the combinational logic of ordinary video, and array mode contained most of kind of common video, and its array mode is as shown in table 1.Each cycle tests time span is 10 minutes, and frame per second was for 24 frame/seconds.
The feature of table 17 video test sequence
Figure BDA0000114642060000061
Fig. 1 (a) is an original image, and Fig. 1 (b) is the result that the Criminisi method obtains, and Fig. 1 (c) is the result that this disposal route obtains; The texture information complexity of the literal part of blocking as can be seen among Fig. 1 (a), details is abundant.Fig. 1 (b) can't recover the detailed information partly that is blocked fully owing to adopted this frame restorative procedure of Criminisi, causes the texture information disappearance, so the result has significantly blur effect.Fig. 1 (c) is than Fig. 1 (b), and details is resumed fully, and does not almost have repairing mark.
For the advantage of this disposal route better is described, further carry out quantitative judge below by the objective evaluation parameter; Experimental result is shown in table 2 and 3, and experimental result shows that method of the present invention can guarantee the accuracy of detection and location, guarantees to repair result's correctness, guarantees the time continuity of results for video.
Table 2 text detection result's quality evaluation
Figure BDA0000114642060000071
Table 3 background is repaired result's quality evaluation
The evaluation index explanation:
For the effect to text detection is carried out quantitative test, the present invention considers to adopt following two kinds of quality objective evaluation indexs to carry out quantitative test;
(1) false alarm rate
Refer to the non-text filed text filed ratio that totally detects that accounts among the image detection result:
Figure BDA0000114642060000073
The false alarm rate of video is that the false alarm rate to frame of video is averaged and obtains;
(2) loss
Refer to the text filed totally text filed ratio that accounts for that is not detected in the image:
The loss of video is that the loss to frame of video is averaged and obtains.
For the effect to the video reparation is carried out quantitative test, the present invention considers to adopt following four kinds of quality objective evaluation indexs to carry out quantitative test:
(1) signal to noise ratio (S/N ratio) (Snr)
The quality that can reflect the image repair effect.Generally calculate with following formula:
Snr ( f O , f ) = Σ i , j ( f O ( i , j ) - μ ( f O ) ) 2 / Σ i , j ( f O ( i , j ) - f ( i , j ) ) 2
Wherein f is an original image, f OBe the image after repairing, the average of μ representative image.
(2) square error (Rmse)
Being original image and the quadratic sum of repairing the difference of image corresponding point, is the judgement to an object fidelity of image repair performance;
Rmse ( f 0 , f ) = Σ i , j ( f 0 ( i , j ) - f ( i , j ) ) 2
(3) the universal qualities factor (UQI)
By original image and the comparison of repairing image, the quality of reflection image repair quality.
UQI ( f 0 , f ) = 4 σ f 0 f μ ( f 0 ) μ ( f ) ( σ f 0 2 + σ f 2 ) [ μ ( f 0 ) 2 + μ ( f ) 2 ]
Wherein f is an original image, f OBe the image after repairing, the average of μ representative image, the variance of σ representative image, Presentation video f OCovariance with f.
(4) temporal correlation (Corr)
Correlativity is to weigh the standard of the degree of correlation of the gray scale of repairing image front and back frame correspondence position on time, and the temporal correlation of the big more explanation restored video of correlativity is good more;
Corr ( f t , f t + 1 ) = Σ i , j ( f t ( i , j ) - μ ( f t ) ) ( f t + 1 ( i , j ) - μ ( f t + 1 ) ) Σ i , j ( f t ( i , j ) - μ ( f t ) ) 2 Σ i , j ( f t + 1 ( i , j ) - μ ( f t + 1 ) ) 2
F wherein tAnd f T+1Be respectively and repair back t and t+1 image constantly, the average of μ representative image.

Claims (2)

1. video text disposal route based on inter-frame information may further comprise the steps:
(1) (t), establish described present image is f (x to the text filed R of detection present image for x, y, y, t), the N width of cloth image of its front is followed successively by f (x, y, t-1) ..., f (x, y, t-N), described N width of cloth image correspondence text filed be respectively R (x, y, t-1),, R (x, y, t-N), wherein x represents the horizontal ordinate of described present image, and y represents the ordinate of described present image, and t represents the frame number of described present image;
(2) to described present image f (x, y, described text filed R t) (x, y t) carry out verification, specifically comprise following substep:
(21) judge described text filed R (x, y, t) subregion in whether text filed R (x, y, t-1) ..., (x, y occurred in t-N) R;
(22) if do not occur, can determine that then described subregion is the false-alarm zone, described subregion is got rid of, and (x, y t) are updated to R with described text filed R New(x, y, t), R wherein New(x, y t) are area to be repaired in the described present image;
(23) if occurred, then continue to described text filed R (x, y, t) in other subregions carry out verification;
(3) from front to back to described text filed R (x, y t) carry out the background reparation, specifically comprise following substep:
(31) (x, y t) and the motion excursion of preceding N frame, if motion is bigger, then directly handle next frame, if it is less to move, then carry out step (32) to analyze described present image f;
(32) with the text filed R of described present image New(x, y, t) and the text filed R of preceding N frame (x, y, t-1) ..., (x, y t-N) compare R, and the zone that obtains utilizing inter-frame information to repair in the described present image is R New(x, y, t)-R (x, y, t-1) ..., R New(x, y, t)-R (x, y, t-N);
(33) to described present image f (x, y, t) R in New(x, y, t)-R (x, y, t-1) ..., R New(x, y, t)-R (x, y, t-N) zone, the search best matching blocks is finished reparation in the N frame before described, and upgrades R New(x, y, t);
(34) repeating step (3) is handled back one frame, until processes complete section video;
(4) begin to carry out the background reparation forward from last frame, specifically comprise following substep:
(41) analyze described present image f (x, y is t) with the motion excursion of back N frame; If it is bigger to move, directly front one frame is handled; Less if move, then carry out step (42);
(42) with the text filed R of described present image New(x, y, t) with the text filed R of back N frame (x, y, t+1) ..., (x, y t+N) compare R, and the zone that obtains utilizing inter-frame information to repair in the present image is R New(x, y, t)-R (x, y, t+1) ..., R New(x, y, t)-R (x, y, t+N);
(43) to described present image f (x, y, t) R in New(x, y, t)-R (x, y, t+1) ..., R New(x, y, t)-R (x, y, t+N) zone, the search best matching blocks is finished reparation in the N frame of back, and upgrades R New(x, y, t);
(44) if this moment To remaining text filed, the match block that search is best in this frame is repaired so; Otherwise repeating step (4) is handled front one frame, up to processes complete section video.
2. video text disposal route according to claim 1 is characterized in that: utilize inter-frame information to finish range check and background reparation in step (2)~(4).
CN 201110391472 2011-11-30 2011-11-30 Video word processing method based on interframe information Expired - Fee Related CN102521805B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 201110391472 CN102521805B (en) 2011-11-30 2011-11-30 Video word processing method based on interframe information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 201110391472 CN102521805B (en) 2011-11-30 2011-11-30 Video word processing method based on interframe information

Publications (2)

Publication Number Publication Date
CN102521805A CN102521805A (en) 2012-06-27
CN102521805B true CN102521805B (en) 2013-07-24

Family

ID=46292710

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 201110391472 Expired - Fee Related CN102521805B (en) 2011-11-30 2011-11-30 Video word processing method based on interframe information

Country Status (1)

Country Link
CN (1) CN102521805B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069766B (en) * 2015-07-24 2017-12-08 北京航空航天大学 A kind of an inscription on a tablet restorative procedure based on the description of Chinese character image contour feature
CN109872277A (en) * 2017-12-04 2019-06-11 北京京东尚科信息技术有限公司 Information processing method and device

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101572778B (en) * 2008-04-28 2011-11-30 北大方正集团有限公司 Method and system for broadcasting programs
CN101360193A (en) * 2008-09-04 2009-02-04 北京中星微电子有限公司 Video subtitle processing apparatus and method
MX2011003076A (en) * 2009-06-17 2011-04-19 Panasonic Corp Information recording medium for reproducing 3d video, and reproduction device.

Also Published As

Publication number Publication date
CN102521805A (en) 2012-06-27

Similar Documents

Publication Publication Date Title
KR101734829B1 (en) Voice data recognition method, device and server for distinguishing regional accent
CN103297851A (en) Method and device for quickly counting and automatically examining and verifying target contents in long video
CN103970733B (en) A kind of Chinese new word identification method based on graph structure
CN112884747A (en) Automatic bridge crack detection system integrating cyclic residual convolution and context extractor network
CN108304328B (en) Text description generation method, system and device for crowdsourcing test report
CA2519908A1 (en) Target detection improvements using temporal integrations and spatial fusion
CN101324927A (en) Method and apparatus for detecting shadows
CN104951842A (en) Novel method for predicting oil field output
CN103985106B (en) Apparatus and method for carrying out multiframe fusion to very noisy image
US20110081049A1 (en) Detector and method for identifying a road lane boundary
CN102521805B (en) Video word processing method based on interframe information
CN104639714A (en) Test method of mobile phone response time
US20100141774A1 (en) Apparatus for Determining Information in Order to Temporally Align Two Information Signals
CN107180293B (en) Exploration target-oriented geological evaluation level determination method
CN102685547B (en) Low-bit-rate video quality detection method based on blocking effects and noises
CN103903258B (en) Method for detecting change of remote sensing image based on order statistic spectral clustering
CN110728582A (en) Information processing method, device, storage medium and processor
CN116255896A (en) Landslide early warning method and device based on satellite and electronic equipment
CN110020686A (en) A kind of road surface method for detecting abnormality based on intelligent perception sensing data
CN111813996B (en) Video searching method based on sampling parallelism of single frame and continuous multi-frame
CN104200815A (en) Audio noise real-time detection method based on correlation analysis
CN105989095B (en) Take the correlation rule significance test method and device of data uncertainty into account
CN113239727A (en) Person detection and identification method
CN114140511A (en) Method and device for estimating dimension specification of object, electronic device and storage medium
CN116434197A (en) Abnormal action detection closed loop iterative optimization method, system, medium and equipment

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20130724

Termination date: 20141130

EXPY Termination of patent right or utility model