CN106407969A - Robust complex background video text positioning and extracting method - Google Patents

Robust complex background video text positioning and extracting method Download PDF

Info

Publication number
CN106407969A
CN106407969A CN201610778073.2A CN201610778073A CN106407969A CN 106407969 A CN106407969 A CN 106407969A CN 201610778073 A CN201610778073 A CN 201610778073A CN 106407969 A CN106407969 A CN 106407969A
Authority
CN
China
Prior art keywords
text
angle point
area
positioning
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610778073.2A
Other languages
Chinese (zh)
Inventor
汤景凡
王兴起
姜明
李志涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Hangzhou Electronic Science and Technology University
Original Assignee
Hangzhou Electronic Science and Technology University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Electronic Science and Technology University filed Critical Hangzhou Electronic Science and Technology University
Priority to CN201610778073.2A priority Critical patent/CN106407969A/en
Publication of CN106407969A publication Critical patent/CN106407969A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • G06V10/225Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition based on a marking or identifier characterising the area
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/752Contour matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Character Input (AREA)

Abstract

The invention discloses a robust complex background video text positioning and extracting method, and mainly solves the problems that the present video text positioning and extracting method is not sufficiently robust under the complex background. In the text positioning phase, angular points are selected to act as the basic characteristics of characters, and then positioning of text rows is realized by the method of combination of coarse positioning and fine positioning. Candidate text areas are obtained in coarse positioning by utilizing the four characteristics of Area, Saturation, Ratio and Position of a binary text area distribution graph. Finally, the candidate text areas are segmented into the text rows by using the horizontal projection of the angular points and the method of angular point density fusion so that accurate positioning of the text rows can be realized and pseudo text rows can be eliminated. The text extracting phase is completed through polarity judgment based on binary images and an improved local OTSU method so that the problems of optimal threshold selection and stroke detail loss under the complex background can be effectively solved. The robust complex background video text positioning and extracting method has high recall rate in video text positioning and extracting of various programs.

Description

A kind of healthy and strong complex background Video Text Location and abstracting method
Technical field
The invention belongs to field of video image processing is and in particular to a kind of healthy and strong complex background Video Text Location and taking out Take method.
Background technology
With the fast development of modern science and technology, a lot of information in life are all to be transmitted by multimedia form.Wherein, Word in video is one of most useful information type, and these texts provide much valuable information, and such as program is situated between Continue, scene location, especially bulletin, the title of speaker, contest scores, date and time, Real Estate Trend, media event and regard Frequency content etc..Text identification has had a lot of real world applications, such as visual classification, document analysis, the video inspection based on video content Rope, help blind person, automatic marking, Car license recognition etc..So extracting to the text message of video, to the deep layer understanding video Semantic information is significant.
There has been now a lot of video text positioning and the algorithm extracting both at home and abroad, can be largely classified into based on connected domain , based on texture, based on edge and based on study method.Method locating speed ratio wherein based on connected domain is very fast, But easily by disturbing that picture contrast changes;Method positioning based on texture and edge is more stable, but has the time complicated Spend high shortcoming;And the method positioning quality based on study depends entirely on the training of sample.
Content of the invention
The present invention is directed in prior art the unhealthy and strong problem of Video Text Location under complex background it is proposed that a kind of video The positioning of middle complex background text and extraction robust method.
The purpose of the present invention is achieved through the following technical solutions:A kind of healthy and strong complex background Video Text Location And abstracting method, comprise the following steps:
Pretreatment:Input video frame, frame of video is converted into gray level image, carries out Corner Detection to frame of video, obtains angle Point two-value scattergram;
Text coarse positioning:First with angle point two-value scattergram, region merging technique is carried out by sliding window, form two-value literary composition One's respective area scattergram, then filters out the angle point in non-textual region using four attributes of two-value text regional distribution chart, real Existing text filed coarse positioning;
Text fine positioning:By the method for angle point floor projection and angle point density fusion, realize line of text fine positioning;
Text extracts:The polarity that line of text image after positioning is carried out based on bianry image judges, then using improvement Local OTSU algorithm calculate threshold value in each piece of region, and optimal threshold is adjusted to according to polarity judged result, finally complete Become the binaryzation of line of text.
Further, through region, fusion obtains described two-value text regional distribution chart, specifically selects a n*n's Rectangle frame, 5≤n≤15, centered on angle point, angle point color is set to the color in rectangle frame region, has traveled through all angle points, Finally give two-value text regional distribution chart.
Further, described two-value text regional distribution chart has multiple connected regions.
Further, the attribute of described two-value text regional distribution chart includes:Area, Saturation, Ratio and Position.
Further, the described angle point filtering out in non-textual region refers to by four attribute character Area, Saturation, Ratio and Position are progressively filtered, specifically:
First, Area attribute filter refer to each of current two-value text regional distribution chart connected region Area by It is ranked up according to size, filter out relatively small region Area;Then, Saturation filters and refers to calculate each even The ratio of the area A (Rect) of area A (Area) and boundary rectangle frame Rect of logical region Area, Saturation=A (Area)/A (Rect), (Saturation ∈ (0,1)), filter out the less connected region of Saturation;Then, Ratio Filter and refer to calculate the depth-width ratio of each connected region boundary rectangle frame, filter out Ratio and be more than 1:2.5 connected region;? Afterwards, Position filter method specifically calculates the positional information of connected region, filters out the connected region of part in frame of video 2/3 Domain.
Further, the method for described angle point floor projection and angle point density fusion refers to:
First, the angle point rectangular histogram often gone by statistics, is divided into literary composition using histogrammic Wave crest and wave trough by text filed One's own profession, its trough basis for estimation is the angle point number of continuous q row to be less than 1/4 or the 1/3 of angle point meansigma methodss number be considered as trough, 3≤q ≤6.Then the method utilizing angle point density fusion removes the background area of line of text or pseudo- line of text.
Further, the method for described angle point density fusion refers to filter out the background of line of text remaining or removes pseudo-text One's own profession, its filtering rule utilizes H*1/2H (H is the height of line of text) horizontal sliding window mouth to carry out level slip, removes angle point close Less than the region of threshold value C, C is angle point number to degree, and the rectangle frame finally again rectangle frame being smaller than H is fused into new text OK.
Further, the described polarity based on bianry image judges it is the polarity carrying out based on local OTSU bianry image Judge, first, carry out the binaryzation of line of text using local OTSU method, then four borders after above-mentioned bianry image are made For sub-pixel point, carry out four connected region seed fill algorithm filling, Filling power is p, 0<p<255, finally calculate black and white two-value Shared ratio, heavy color is then the polarity of word.
Further, described optimal threshold is adjusted to according to polarity judged result refers to calculate using local OTSU algorithm Go out threshold value T in each block, but do not carry out binary conversion treatment.Then utilize text polarity judged result, then change present threshold value For optimal threshold, amended optimal threshold is finally utilized to execute binarization operation.
Further, described modification present threshold value specifically refers to for optimal threshold, if polarity judges text color being Black, then new threshold value T1=T-T*0.1, conversely, during white text, new threshold value T1=T+T*0.1.
The invention has the beneficial effects as follows:In the String localization stage, it is basic as word that the present invention chooses healthy and strong angle point Feature, then completes the text filed positioning of candidate by coarse positioning, and reservation as much as possible is text filed;Recycle fine positioning Realize segmentation and the verification of text filed line of text.Text extraction stage, is judged and office by the polarity based on bianry image The method that portion OTSU combines completes optimal threshold under complex background and chooses difficult problem.Through the many experiments to the present invention Test, the String localization of the present invention and extraction algorithm have good vigorousness to complex background video.
Brief description
Fig. 1 is the inventive method flow chart.
Specific embodiment
Below in conjunction with accompanying drawing, the invention will be further described.
As shown in figure 1, a kind of healthy and strong complex background Video Text Location that provides of the present invention and abstracting method, including with Lower step:
Pretreatment:Input video frame, frame of video is converted into gray level image, carries out Corner Detection to frame of video, obtains angle Point two-value scattergram;
Text coarse positioning:First with angle point two-value scattergram, region merging technique is carried out by sliding window, form two-value literary composition One's respective area scattergram, then filters out the angle point in non-textual region using four attributes of two-value text regional distribution chart, real Existing text filed coarse positioning;
Text fine positioning:By the method for angle point floor projection and angle point density fusion, realize line of text fine positioning;
Text extracts:The polarity that line of text image after positioning is carried out based on bianry image judges, then using improvement Local OTSU algorithm calculate threshold value in each piece of region, and optimal threshold is adjusted to according to polarity judged result, finally complete Become the binaryzation of line of text.
Further, through region, fusion obtains described two-value text regional distribution chart, specifically selects a n*n's Rectangle frame, 5≤n≤15, centered on angle point, angle point color is set to the color in rectangle frame region, has traveled through all angle points, Finally give two-value text regional distribution chart.
Further, described two-value text regional distribution chart has multiple connected regions.
Further, the attribute of described two-value text regional distribution chart includes:Area, Saturation, Ratio and Position.
Further, the described angle point filtering out in non-textual region refers to by four attribute character Area, Saturation, Ratio and Position are progressively filtered, specifically:
First, Area attribute filter refer to each of current two-value text regional distribution chart connected region Area by It is ranked up according to size, filter out relatively small region Area;Then, Saturation filters and refers to calculate each even The ratio of the area A (Rect) of area A (Area) and boundary rectangle frame Rect of logical region Area, Saturation=A (Area)/A (Rect), (Saturation ∈ (0,1)), filter out the less connected region of Saturation;Then, Ratio Filter and refer to calculate the depth-width ratio of each connected region boundary rectangle frame, filter out Ratio and be more than 1:2.5 connected region;? Afterwards, Position filter method specifically calculates the positional information of connected region, filters out the connected region of part in frame of video 2/3 Domain.
Further, the method for described angle point floor projection and angle point density fusion refers to:
First, the angle point rectangular histogram often gone by statistics, is divided into literary composition using histogrammic Wave crest and wave trough by text filed One's own profession, its trough basis for estimation is the angle point number of continuous q row to be less than 1/4 or the 1/3 of angle point meansigma methodss number be considered as trough, 3≤q ≤6.Then the method utilizing angle point density fusion removes the background area of line of text or pseudo- line of text.
Further, the method for described angle point density fusion refers to filter out the background of line of text remaining or removes pseudo-text One's own profession, its filtering rule utilizes H*1/2H (H is the height of line of text) horizontal sliding window mouth to carry out level slip, removes angle point close Less than the region of threshold value C, C is angle point number to degree, and the rectangle frame finally again rectangle frame being smaller than H is fused into new text OK.
Further, the described polarity based on bianry image judges it is the polarity carrying out based on local OTSU bianry image Judge, first, carry out the binaryzation of line of text using local OTSU method, then four borders after above-mentioned bianry image are made For sub-pixel point, carry out four connected region seed fill algorithm filling, Filling power is p, 0<p<255, finally calculate black and white two-value Shared ratio, heavy color is then the polarity of word.
Further, described optimal threshold is adjusted to according to polarity judged result refers to calculate using local OTSU algorithm Go out threshold value T in each block, but do not carry out binary conversion treatment.Then utilize text polarity judged result, then change present threshold value For optimal threshold, amended optimal threshold is finally utilized to execute binarization operation.
Further, described modification present threshold value specifically refers to for optimal threshold, if polarity judges text color being Black, then new threshold value T1=T-T*0.1, conversely, during white text, new threshold value T1=T+T*0.1.
Embodiment
The realization of the present embodiment, comprises the following steps:
1st, input video frame, carries out pretreatment to frame of video, for example, be converted into gray level image;Using Harris algorithm to regarding Frequency frame carries out Corner Detection, obtains the angle point two-value scattergram that background is that black, angle point are white;
2nd, the rectangle frame of 6*6, centered on angle point, the face consistent by being set to angle point in the rectangle frame of each angle point are utilized Color, obtains two-value text regional distribution chart;
3rd, carry out String localization using by the thick text positioning method to essence;
Coarse positioning:Four attributes removals choosing two-value text regional distribution chart are pseudo- text filed, obtain candidate's text area Domain.It is area, saturation, depth-width ratio and position respectively, be designated as Area, Saturation, Ratio, Position.
Area:Find each connected region Area, then according to the size of connected region is ranked up, filter out Relatively small region.Because the less region of area is not necessarily the main contents of frame of video reflection, and small area region It is easy to be filtered.
Saturation:The i.e. saturation feature of angle point, counts the boundary rectangle frame Rect of each connected region.Filter out The less connected region of Saturation.
Saturation=A (Area)/A (Rect), (Saturation ∈ (0,1));
Wherein A (Area) is the area of connected domain, and A (Rect) is the area of the boundary rectangle frame of connected domain.Due to video In overlapping text be all level, the therefore value of Saturation is close to 1, and the Saturation of pseudo- text connected region It is worth close to 0.
Ratio:Connected region boundary rectangle frame depth-width ratio, according to the own characteristic of Chinese character, it is big that the present invention filters out Ratio In 1:2.5 connected region.
Position:The positional information of connected region, filters out in frame of video 2/3 connected region.Because in frame of video Overlapping text information be typically all in the lower section of frame of video.
Fine positioning:Due to multiline text or pseudo- text may be comprised in the connected region of each candidate, therefore, it can Carry out floor projection using angle point, the connected region of each candidate is divided into line of text and realizes being accurately positioned;Finally utilize Angle point density fusion method filters further to the background of line of text or pseudo- line of text, and filter window size is that (H is H*1/2H The height of line of text), the rectangle frame again rectangle frame being smaller than H afterwards is fused into new line of text, and line of text positioning completes;
4th, text extracts.The present invention judges mutually to tie with improved local OTSU binaryzation using the polarity based on bianry image The method closed completes the extraction of line of text.
Local OTSU binaryzation:Line of text image division such as is at the region of size, the size of each zonule is H*H (height of line of text).Each region carries out local OTSU binaryzation;
Polarity based on bianry image judges:Using four borders after above-mentioned bianry image as sub-pixel point, carry out Four connected region seed fill algorithm is filled, and the value of filling is 128, finally calculates the ratio shared by black and white two-value, heavy face Color is then the polarity of word;
Improved local OTSU binaryzation:First to line of text image procossing as the OTSU algorithm of local, by line of text It is divided into the region that multiple sizes are H*H, then execution OTSU algorithm calculates threshold value T in each block, but does not now carry out two Value is processed.Because T now is not the optimal threshold of text segmentation.It is thus desirable to according to text polarity judged result, then repair Change this threshold value (if polarity judges text color for black, new threshold value T1=T-T*0.1, conversely, during white text, new threshold Value T1=T+T*0.1), finally utilize amended optimal threshold to execute binarization operation.

Claims (10)

1. a kind of healthy and strong complex background Video Text Location and abstracting method are it is characterised in that comprise the following steps:
Pretreatment:Input video frame, frame of video is converted into gray level image, carries out Corner Detection to frame of video, obtains angle point two Distribution value figure;
Text coarse positioning:First with angle point two-value scattergram, region merging technique is carried out by sliding window, form two-value text area Domain scattergram, then filters out the angle point in non-textual region using four attributes of two-value text regional distribution chart, realizes literary composition One's respective area coarse positioning;
Text fine positioning:By the method for angle point floor projection and angle point density fusion, realize line of text fine positioning;
Text extracts:The polarity that line of text image after positioning is carried out based on bianry image judges, then utilizes improved office Portion's OTSU algorithm calculates the threshold value in each piece of region, and is adjusted to optimal threshold according to polarity judged result, finally completes literary composition The binaryzation of one's own profession.
2. method according to claim 1 is it is characterised in that described two-value text regional distribution chart merges through region Obtain, specifically select the rectangle frame of a n*n, 5≤n≤15, centered on angle point, the color in rectangle frame region is set to Angle point color, has traveled through all angle points, finally gives two-value text regional distribution chart.
3. method according to claim 2 is it is characterised in that described two-value text regional distribution chart has multiple connections Region.
4. method according to claim 1 is it is characterised in that the attribute of described two-value text regional distribution chart includes: Area, Saturation, Ratio and Position.
5. method according to claim 1 is it is characterised in that the described angle point filtering out in non-textual region refers to lead to Cross four attribute character Area, Saturation, Ratio and Position are progressively filtered, specifically:
First, Area attribute filters and refers to each of current two-value text regional distribution chart connected region Area according to face Long-pending size is ranked up, and filters out relatively small region Area;Then, Saturation filters and refers to calculate each connected region The ratio of the area A (Rect) of area A (Area) and boundary rectangle frame Rect of domain Area, Saturation=A (Area)/A (Rect), (Saturation ∈ (0,1)), filters out the less connected region of Saturation;Then, Ratio filters and refers to Calculate the depth-width ratio of each connected region boundary rectangle frame, filter out Ratio and be more than 1:2.5 connected region;Finally, Position filter method specifically calculates the positional information of connected region, filters out the connected region of part in frame of video 2/3.
6. method according to claim 1 is it is characterised in that the side of described angle point floor projection and angle point density fusion Method refers to:
First, the angle point rectangular histogram often gone by statistics, is divided into line of text using histogrammic Wave crest and wave trough by text filed, Its trough basis for estimation is the angle point number of continuous q row to be less than 1/4 or the 1/3 of angle point meansigma methodss number be considered as trough, 3≤q≤6. Then the method utilizing angle point density fusion removes the background area of line of text or pseudo- line of text.
7. method according to claim 6 is it is characterised in that the method for described angle point density fusion refers to filter out literary composition The background of one's own profession remaining or the pseudo- line of text of removal, its filtering rule utilizes H*1/2H (H is the height of line of text) horizontal sliding window Mouth carries out level slip, removes the region that angle point density is less than threshold value C, and C is angle point number, is finally smaller than rectangle frame again The rectangle frame of H is fused into new line of text.
8. method according to claim 1 is it is characterised in that the described polarity based on bianry image judges it is based on office The polarity that portion's OTSU bianry image is carried out judges, first, carries out the binaryzation of line of text using local OTSU method, then upper State four borders after bianry image as sub-pixel point, carry out four connected region seed fill algorithm filling, Filling power is p, 0 <p<255, finally calculate the ratio shared by black and white two-value, heavy color is then the polarity of word.
9. method according to claim 1 is it is characterised in that described is adjusted to optimal threshold according to polarity judged result Refer to calculate threshold value T in each block using local OTSU algorithm, but do not carry out binary conversion treatment.Then utilize text polarity Judged result, then change present threshold value for optimal threshold, finally utilize amended optimal threshold to execute binarization operation.
10. method according to claim 9 it is characterised in that described modification present threshold value be optimal threshold specifically Refer to, if polarity judges text color for black, new threshold value T1=T-T*0.1, conversely, during white text, new threshold value T1=T +T*0.1.
CN201610778073.2A 2016-08-30 2016-08-30 Robust complex background video text positioning and extracting method Pending CN106407969A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610778073.2A CN106407969A (en) 2016-08-30 2016-08-30 Robust complex background video text positioning and extracting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610778073.2A CN106407969A (en) 2016-08-30 2016-08-30 Robust complex background video text positioning and extracting method

Publications (1)

Publication Number Publication Date
CN106407969A true CN106407969A (en) 2017-02-15

Family

ID=58003924

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610778073.2A Pending CN106407969A (en) 2016-08-30 2016-08-30 Robust complex background video text positioning and extracting method

Country Status (1)

Country Link
CN (1) CN106407969A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107302718A (en) * 2017-08-17 2017-10-27 河南科技大学 A kind of video caption area positioning method based on Corner Detection
CN107688788A (en) * 2017-08-31 2018-02-13 平安科技(深圳)有限公司 Document charts abstracting method, electronic equipment and computer-readable recording medium
CN109993165A (en) * 2019-03-28 2019-07-09 永康市几米电子科技有限公司 The identification of tablet plate medicine name and tablet plate information acquisition method, device and system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101593277A (en) * 2008-05-30 2009-12-02 电子科技大学 A kind of complicated color image Chinese version zone automatic positioning method and device
CN103268481A (en) * 2013-05-29 2013-08-28 焦点科技股份有限公司 Method for extracting text in complex background image
CN104182750A (en) * 2014-07-14 2014-12-03 上海交通大学 Extremum connected domain based Chinese character detection method in natural scene image

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101593277A (en) * 2008-05-30 2009-12-02 电子科技大学 A kind of complicated color image Chinese version zone automatic positioning method and device
CN103268481A (en) * 2013-05-29 2013-08-28 焦点科技股份有限公司 Method for extracting text in complex background image
CN104182750A (en) * 2014-07-14 2014-12-03 上海交通大学 Extremum connected domain based Chinese character detection method in natural scene image

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
张洋: "电视视频字幕文字的提取方法研究", 《中国优秀硕士学位论文全文数据库_信息科技辑》 *
王刚: "新闻视频字幕的自动提取和识别", 《中国优秀硕士学位论文全文数据库_信息科技辑》 *
胡倩 等: "自然场景下的文本定位", 《电脑知识与技术》 *
黄晓冬: "基于特征融合的视频文本获取研究", 《中国博士学位论文全文数据库_信息科技辑》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107302718A (en) * 2017-08-17 2017-10-27 河南科技大学 A kind of video caption area positioning method based on Corner Detection
CN107302718B (en) * 2017-08-17 2019-12-10 河南科技大学 video subtitle area positioning method based on angular point detection
CN107688788A (en) * 2017-08-31 2018-02-13 平安科技(深圳)有限公司 Document charts abstracting method, electronic equipment and computer-readable recording medium
CN107688788B (en) * 2017-08-31 2021-01-08 平安科技(深圳)有限公司 Document chart extraction method, electronic device and computer readable storage medium
CN109993165A (en) * 2019-03-28 2019-07-09 永康市几米电子科技有限公司 The identification of tablet plate medicine name and tablet plate information acquisition method, device and system

Similar Documents

Publication Publication Date Title
CN104751142B (en) A kind of natural scene Method for text detection based on stroke feature
CN107679502A (en) A kind of Population size estimation method based on the segmentation of deep learning image, semantic
CN101183425B (en) Guangdong and Hong Kong license plate locating method
CN105205488B (en) Word area detection method based on Harris angle points and stroke width
CN109255350B (en) New energy license plate detection method based on video monitoring
CN102542268B (en) Method for detecting and positioning text area in video
Yang et al. Lecture video indexing and analysis using video ocr technology
CN104244073B (en) Automatic detecting and recognizing method of scroll captions in videos
CN1312625C (en) Character extracting method from complecate background color image based on run-length adjacent map
CN104077577A (en) Trademark detection method based on convolutional neural network
CN101515325A (en) Character extracting method in digital video based on character segmentation and color cluster
CN101593277A (en) A kind of complicated color image Chinese version zone automatic positioning method and device
CN104598907B (en) Lteral data extracting method in a kind of image based on stroke width figure
CN102208023A (en) Method for recognizing and designing video captions based on edge information and distribution entropy
CN104050684B (en) A kind of video frequency motion target sorting technique based on on-line training and system
CN103336961A (en) Interactive natural scene text detection method
CN107977645B (en) Method and device for generating video news poster graph
CN104766076A (en) Detection method and device for video images and texts
CN104299009A (en) Plate number character recognition method based on multi-feature fusion
CN101106716A (en) A shed image division processing method
KR20120019425A (en) Image processing device, method, and program
CN106407969A (en) Robust complex background video text positioning and extracting method
CN113191358B (en) Metal part surface text detection method and system
CN102073872A (en) Image-based method for identifying shape of parasite egg
CN106295627A (en) For identifying the method and device of word psoriasis picture

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170215

RJ01 Rejection of invention patent application after publication