TW319856B - Graph-text segmentation method of geographical image data - Google Patents
Graph-text segmentation method of geographical image data Download PDFInfo
- Publication number
- TW319856B TW319856B TW84107775A TW84107775A TW319856B TW 319856 B TW319856 B TW 319856B TW 84107775 A TW84107775 A TW 84107775A TW 84107775 A TW84107775 A TW 84107775A TW 319856 B TW319856 B TW 319856B
- Authority
- TW
- Taiwan
- Prior art keywords
- text
- cluster
- image
- feature point
- clustering
- Prior art date
Links
Landscapes
- Character Input (AREA)
Abstract
Description
S856 A7 B7 經濟部中央標準局員工消費合作社印製 五、發明説明(1 ) 本發明係關於一種圖文分離之方法,特別是關於—種可 將經電腦化以地圖影像資料處理後之地圖影像資料内的圖 文分離出來’此方法可有效地解析影像資料,尤其能解析 圖文黏接或圖文重疊的部分,使解析後的影像之失誤率最 小 ΰ 發明之背景説明: 在資訊未發達的年代,地理資訊管理技術(Geographic Information Management Technology)爲幫助人們 解決大量資訊爆炸的問題,提供一套有效管理大量地理資 料的方法,然而地理資訊系統(GIS)的使用者,需投入大 .量的成本’從原始地圖獲取重要的地圖資訊,因爲傳統的 人工數化方法是不準確且昂貴的。 從1 9 8 0年代早期起,人們以掃瞄器提供一種經濟的方式 ,將已存在的原始地圖掃瞄成網路地圖,並將之轉換成向 量資料’然而,並沒有一套有效且可靠的方法,能提供高 度準確性與自動化的轉換,特別是大部分的產品,並無地 ,圖文字辨識的能力。圖形向量化的相關論文在最近幾年相 繼被提出,但是由於各個國家在地圖上所表現的特徵,使 得現存地圖自動化處理系統没有辨法完善地處理中文地團 ’因此發展一套中文地圖的自動化處理系統是有其必要性 的。然而在發展此一系統時,我們發現到圖文分離是地圖 自動化的一大瓶頸’唯有將圖文予以分離後,再分別作文 字辨識與圖形解譯的工作,才能建立起整個自動化的系統。 _40083.DQC/SLH - 4 - 本紙張尺度適用中國國家標率(CNS ) Λ4規格(210X297公釐) (請先聞讀背面之注意事項再填寫本頁)S856 A7 B7 Printed by the Employee Consumer Cooperative of the Central Bureau of Standards of the Ministry of Economy V. Description of the invention (1) The present invention relates to a method for separating images and texts, in particular to a map image that can be computerized and processed with map image data Separation of graphics and text in the data 'This method can effectively analyze image data, especially the part where graphics and text are glued or overlapped, so that the error rate of the analyzed image is minimized. Background of the invention: Information is not developed In the era of Geographic Information Management Technology (Geographic Information Management Technology) to help people solve the problem of a large number of information explosions, it provides a set of methods for effectively managing large amounts of geographic data, but users of geographic information systems (GIS) need to invest a large amount. The cost of obtaining important map information from the original map, because the traditional manual digitization method is inaccurate and expensive. Since the early 1980s, people have used scanners to provide an economical way to scan existing original maps into network maps and convert them into vector data. However, there is no set of effective and reliable The method can provide a high degree of accuracy and automatic conversion, especially for most products, without the ability to recognize land and pictures. Related papers on graphic vectorization have been successively proposed in recent years, but due to the characteristics of various countries on the map, the existing map automation processing system does not have a discriminatory method to deal with Chinese plots perfectly. Therefore, a set of Chinese map automation has been developed The processing system is necessary. However, in the development of this system, we found that the separation of graphics and text is a major bottleneck in map automation. Only after separating graphics and text, and then separately performing text recognition and graphics interpretation, can the entire automated system be established . _40083.DQC / SLH-4-This paper scale is applicable to China National Standard Rate (CNS) Λ4 specification (210X297mm) (please read the precautions on the back before filling this page)
3.δδ56 Α7 Β7_ 五、發明説明(2 ) (請先閱讀背面之注意事項再填寫本頁) 過去有多篇論文相繼討論圖文分離這個專題,譬如 Fletcher使用霍氏轉換(Hough Transform)配合字元區 塊法(Character Blocking),將一電子文件圖上的文字 與線條分離,其方法大致如下: ⑴先計算出連結區塊(Connected Component),將圖上 相連的區塊均找出來。 (2)求出每一連結區)鬼中的點密度,依據密度大小,簡單的 區分出文字和線條圖形。 ⑶對於文字區塊,則依中心座標,應用霍氏轉換原理,將 位於同一直線上的字母找出.來,再依字母間距來找出字 (Word)。 此方法的好處在於能夠清楚分辨字串、字與字元,文字最 大與最小的比値可以差到五倍,文字的轉向亦不受限制, 但是,在地圖上並非所有的文字均呈直線排列,而且地圖 上圖文黏接,圖文重疊的部分常常被誤判爲一區塊,因此 失眞率很高,因而並不能有效地應用在地圖的圖文分離上 。(見於L. A. Fletcher and R. Kasturi,”A Robust Algorithm 經濟部中央標準局負工消費合作社印製 for Text String Separation from Mixed Text/Graphics3.δδ56 Α7 Β7_ V. Description of invention (2) (please read the notes on the back before filling out this page) There have been many papers in the past that discussed the topic of graphic separation, such as Fletcher using Hough Transform (Hough Transform) with the word Character Blocking (Character Blocking), to separate the text and lines on an electronic document map, the method is roughly as follows: (1) First calculate the connected block (Connected Component), and find out all the connected blocks on the map. (2) Find the density of dots in ghosts in each connected area, and simply distinguish text and line graphics according to the density. (3) For the text block, according to the central coordinates, applying the principle of Huo's transformation, find the letters on the same straight line, and then find the word (Word) according to the letter spacing. The advantage of this method is that it can clearly distinguish between strings, characters and characters. The maximum and minimum ratio of the text can be up to five times, and the direction of the text is not limited. However, not all text on the map is arranged in a straight line In addition, the graphics and text on the map are glued together, and the overlapping part of the graphics and text is often mistaken as a block, so the rate of miss is very high, so it cannot be effectively applied to the separation of graphics and text on the map. (See L. A. Fletcher and R. Kasturi, "A Robust Algorithm Printed by the Consumer Labor Cooperative of the Central Standards Bureau of the Ministry of Economy for Text String Separation from Mixed Text / Graphics
Images”,IEEE Trans. Pattern Anal. Machine Intell, pp 910-918,Vol. 10,1988) 爲解決圖文黏接的問題,Kasturi在1990年提出之方 法亦針對圖文黏接的問題,採用連結區塊方式,其方法 如下: ⑴先完成Fletcher所提方法,做好初步囷文分離。 40083.D0C/SLH · 5 ~ 本紙張尺度適用中國國家標準(CNS ) A4規格(210X297公釐) --- 經濟部中央標準局員工消費合作社印製 A7 五、發明説明(3 ⑵在連結區塊中找到疑有圖文黏接㈣方,由此點恢其八 鄰點向周邊找尋可疑字元,直到所有點都被找到爲止。 ⑶檢查此可疑字元是否能夠插入某個字串中。 然而此方法雖可以稣決簡單的圖文黏接問題,卻仍不能解 決圖又重疊的問題。(見於R. Kasturi and S.T. B〇w,,,A System for Interpretation of Line Drawings11 JEEE Trans, Pattern Anal.,Vol'· 12, No. 10, pp 978·991,199〇) , 另外,Wahl亦針對文件分析提出了分離文字、圖形, 以及影像的方法,該方法之步驟計有: ⑴依據跑長碼(Run-length code)原理,將每一列的黑 點編碼爲1,白點爲〇。 ⑵务某歹]中的一個黑點之間的白點個數小於某一臨界値, 則將〇改爲1。 ⑶同理,再對每一行(γ _方向)重覆(丨)與(2 )的步驟。 ⑷將兩個結果做且(A N D )運算。 ⑸將(4 )所得的區塊,計算其點密度。 ⑹依點密度的大小區分文字、圖形與影像。 、這個方法雖然能夠有效地區分文字、圖形、影像,但是 應用到地圖上,仍然無法解決圖文黏接與圖文重疊的問 題。(見於 F. M. Wahl,Μ. K. Wang,and R, G. Casey, MBlock Segmentation and Text Extraction in MixedImages ”, IEEE Trans. Pattern Anal. Machine Intell, pp 910-918, Vol. 10, 1988) In order to solve the problem of image and text bonding, the method proposed by Kasturi in 1990 also aimed at the problem of image and text bonding. The block method is as follows: ⑴ First complete the method proposed by Fletcher, and do a preliminary separation of text. 40083.D0C / SLH · 5 ~ This paper scale is applicable to the Chinese National Standard (CNS) A4 specification (210X297 mm)- -A7 printed by the Staff Consumer Cooperative of the Central Bureau of Standards of the Ministry of Economic Affairs 5. Description of the invention (3 ⑵ Find the suspected image in the link block, and then restore its eight neighbors to find suspicious characters in the surrounding area until All points have been found. (3) Check whether the suspicious character can be inserted into a string. However, although this method can solve the simple problem of image and text bonding, it still cannot solve the problem of overlapping images. (See R . Kasturi and ST B〇w ,,, A System for Interpretation of Line Drawings11 JEEE Trans, Pattern Anal., Vol '· 12, No. 10, pp 978 · 991, 199〇), in addition, Wahl also proposed for document analysis Separate text The method of graphics and images, the steps of this method include: (1) According to the principle of run-length code, the black point of each column is coded as 1 and the white point is 0. (2) One of services] If the number of white dots between black dots is less than a certain critical value, change 〇 to 1. ⑶Similarly, repeat the steps (丨) and (2) for each row (γ _direction). ⑷ AND results. ⑸ Calculate the point density of the block obtained in (4). ⑹ Distinguish between text, graphics and images according to the size of the point density. Although this method can effectively distinguish text, graphics and images , But applied to the map, still can not solve the problem of image and text adhesion and image overlap. (See FM Wahl, Μ.K. Wang, and R, G. Casey, MBlock Segmentation and Text Extraction in Mixed
Text/Image Documents",Comput. Vision,Graphics,and Image Processing,V〇l. 20, pp. 375-390,1982)Text / Image Documents ", Comput. Vision, Graphics, and Image Processing, V〇l. 20, pp. 375-390, 1982)
Boatto et al.亦提出一個偵側與分離「圖文黏接」和厂 圖文重#」的方法’不過應用到中文地圖上,對於複雜 40083.D0C/SLH _ β _ Μ氏張尺度適射酬家標準(CNS ) Α4規格(21Qx 297公楚 (請先閱讀背面之注意事項再填寫本f ) 訂 經濟部中央標隼局負工消費合作社印製 A7 B7 五、發明説明(4 ) 的中文字與圖形重疊的問題,仍然無法有效的處理。(見 L. Boatto et al. ''Detection and Separation of SymbolsBoatto et al. Also proposed a method to detect the side and separate the "picture and text bonding" and the factory picture and text # "but it is applied to the Chinese map, for the complex 40083.D0C / SLH _ β _ Μ's Zhang scale fit Compensation Standard (CNS) Α4 specifications (21Qx 297 public (please read the precautions on the back before filling in this f)) Printed A7 B7 printed by the Central Standard Falcon Bureau of the Ministry of Economic Affairs Consumer Cooperative V. Invention description (4) in Chinese The problem of overlapping characters and graphics is still not effectively dealt with (see L. Boatto et al. '' Detection and Separation of Symbols
Connected to Graphics in Line Drawings丨',InternationalConnected to Graphics in Line Drawings 丨 ', International
Conference on Pattern Recognition, Vol. 2, pp. 545-548, 1992) 至於其它相關理論和技術上,已有些專利如中華民國 專利3 0 9 8 7號,” Η卩刷體框字處理法”揭示印刷字體的圖 文處理方法,但是該方法只限於框出印刷字體,對於地 圖上的手寫多變化性的字體,甚至與圖形相連,就束手 無策了 3 有鑑於目前的圖文處理方法仍不能有效解決圖文黏接 以及圖文重疊的問題,本發明乃提供一種新的圖文分離 方法’該方法以抽取特徵點的方式,然後根據特徵點的 密度來判斷文字、線條,此法能夠分離不同大小的文字 或文字有旋轉的情況,JL能夠有效地處理「圖文黏接」 和「圖文重疊」等問題。 根據本發明,在圖文分離前,須先做一些前置處理, 如二値化、細線化及特徵點擷取等,二値化的目的在於 除去不必要的雜訊,並增進地圖品質,細線化則在降低 ,影像的储存空間,所有圖形都變成一個圖素(1_pixel的 寬度),因此任何線條的交點都可確定是一個圖素,這樣 一來,在地圖上文字符號的特徵,都可明顯地凸顯出來 ,因爲文字符號的特徵點群聚性較圖形高,利用此原理 即可做圖文分離’不論文字的大小、文字不同的轉向, 或文^座落的位置皆可處理,對,,圖文黏接”與”圖文重 疊J等文件分析最常見的問題,都可以輕易地被解決。Conference on Pattern Recognition, Vol. 2, pp. 545-548, 1992) As for other related theories and technologies, some patents have been published, such as the Republic of China Patent No. 3 0 9 8 7, "H-Brush Frame Word Processing" has been revealed Graphic processing method of printed fonts, but this method is limited to framed printed fonts. For handwritten fonts on the map, even connected with graphics, it is helpless. 3 Given the current graphic processing methods still cannot be effectively solved The problem of image and text sticking and image and text overlapping, the present invention provides a new method of image and text separation. This method extracts feature points, and then judges text and lines according to the density of feature points. This method can separate different sizes. The text or the text is rotated, JL can effectively deal with the problems of "graphic and text bonding" and "graphic and text overlapping". According to the present invention, before the image and text are separated, some pre-processing must be done, such as binarization, thinning, and feature point extraction. The purpose of binarization is to remove unnecessary noise and improve the quality of the map. Thinning is decreasing, the storage space of the image, all graphics become a pixel (1_pixel width), so the intersection of any line can be determined to be a pixel, so that the characteristics of the text symbols on the map are all It can be clearly highlighted, because the feature point clustering of text symbols is higher than that of graphics, and this principle can be used to separate graphics and text. Regardless of the size of the text, the different direction of the text, or the location of the text ^, it can be processed. Yes, the most common problems in the analysis of documents such as “text” and “text” overlap J can be easily solved.
^0085.D0C/SLH^ 0085.D0C / SLH
本纸張尺度適用中國國家樣隼(CNS (請先閱讀背面之a意事項再填寫本頁)This paper scale is applicable to the Chinese National Falcon (CNS (please read the notes on the back before filling this page)
3 i&S56 A7 〜._____B7____ 友、發明説明(5 ) 一*~~ 如説明本發明之原理及其優點,現以一較佳實施例配 合下列圖式説明於後,其中: _ 一係本發明流程方塊圖 圖一係圖一中前置處理之細部流程圖 圖三A係圖二中二値化處理之實例原圖. 圖三B係圖三A之二値化之結果 圖四係圖一中細線,化結果所用的鏈結碼表示法 圖五係圖一中圖文分離之細部流程圖 圖六係定義圖五中特徵點種類 圖七係用以説明文字是一個一個的特徵點群聚 圖八係說明圖五之群聚分類會因歸類錯誤導致線段ab消失 圖九係説明圖五之線段修補是採用分枝追縱之法 圖十係説明圖五之線段修補之對象是採夾角最小者 (d 1 >d2>d3) 圖十一 A係本發明實施例之輸入地囷影像 圖十一 B係圖--A之細線化圖像 圖十一 C係圖^--B經細線化修補之結果 圖十一D係圖十一C中所抽取之特徵點 麵濟部中夫-ιΐΐ疋消聲合作祍印製 (請先閲讀背面之‘注意事項再填寫本頁3 i & S56 A7 ~ ._____ B7____ Friends, invention description (5) 1 * ~~ If the principle and advantages of the present invention are explained, a preferred embodiment will be described in conjunction with the following drawings, among which: _ 一 系 本Invention flow block diagram Figure 1 is the detailed flow chart of the pre-processing in Figure 1 Figure 3A is the original image of the example of the two-valued processing in Figure 2 Figure 3B is the result of the two-valued conversion in Figure 3A Figure 4 is the series One thin line, the link code representation used for the conversion results. Figure 5 is the detailed flowchart of the graphic separation in Figure 1. Figure 6 is the definition of the feature point types in Figure 5. Figure 7 is used to explain that the text is a feature point group. Figure 8 shows that the clustering classification in Figure 5 will cause the line segment ab to disappear due to the classification error. Figure 9 shows that the line segment repair in Figure 5 uses the branch and chase method. Figure 10 shows that the object of line segment repair in Figure 5 is mining. The one with the smallest angle (d 1 > d2 > d3) FIG. 11A is an image of the input wall image according to an embodiment of the present invention. FIG. 11B is a thin line image of A. FIG. 11 is a C line diagram ^-B The result of thinning and repairing Figure 11D is the feature points extracted in Figure 11C. Printed by Voice Cooperative (Please read the ‘Notes on the back before filling this page
•IT 圖十一E係圖十一C經文字抽取後所剩之線條圖像 、圖十一 F係圖十一 E經線段修補之結果 發明的詳細説明: 本發明之目的係針對前述圖文分離技術之缺失,而提 供種方法,用於處理地圖時,可分離文字和圖形,針 對不同性質的部份做不同的處理,文字符號則加以辨識 ’圖形則加以向量化,使地圖輸入達成完全自動化,本 發明即提出-種全新之圖文分離技術,可解決上述所有 _^〇〇85.D〇C/SLH_ - 8 - t @ ® { CNS J A4^ (TIo'x 297^^. )------__ 經濟部中央標準局負工消費合作社印裝 A 7 B7 五、發明説明(6 ) 之困難。 本發明之圖文分離的系統架構如圖一所示,首先使用 掃瞎器(Scanner)將地圖原稿1〇1輸入個人電腦,而存 成一個二維陣列(A r r a y ),再以對上述之地圖做前置處 理l〇2(preprocessing),把圖像的品質處理得更好, 以利圖文分離的進行β此前置處理的方法可採用v i s u a i c + +語言所寫的程式處理,但不限於此。 上述前置處理102的步驟可以目前市面上可見的方法 達成根據本發明之一較佳實施例如圖二所示,包括了二 値化處理2 0 1、細線化2 〇 2、以及細線化修補2 0 3等步驟 。其中二値化2 0 1的處理原則是依據灰階與點數的統計 圖形選擇一臨界値,若灰階値大於此一臨界値,則此點 改成黑點’否則改爲白點,圖三則説明了二値化2 〇 1的 實驗結果,圖三A是原圖,圖三B則是二値化結果。 然後將二値化處理後的資料做細線化處理2 0 2,根據 此較佳實施例,細線化能採用Z h a n g & S u e η的方法, 細線化所得結果以鍵結碼(C h a i n c o d e )的方式表示, 此法乃依圖四所示的編碼方式,將所有的線不僅儲存其 起終點座標,而且儲存其鍵結碼,如此可在計算面積時• IT Figure 11E is the image of the line left after drawing 11C from the text, Figure 11F is the result of the repair of the line segment of Figure 11E. Detailed description of the invention: The purpose of the present invention is directed to the aforementioned text The lack of separation technology, and provides a method for separating text and graphics when processing maps, different treatments are performed for parts of different nature, and text symbols are recognized. Graphics are vectorized to complete the map input Automation, the present invention proposes a new type of graphic separation technology that can solve all of the above _ ^ 〇〇85.D〇C / SLH_-8-t @ ® {CNS J A4 ^ (TIo'x 297 ^^.) ------__ The Central Standards Bureau of the Ministry of Economic Affairs, the Consumer Labor Cooperative Printed A 7 B7 V. Difficulties in the description of invention (6). The system architecture of the graphic separation of the present invention is shown in FIG. 1. First, a scanner is used to input the map original 101 into a personal computer, and a two-dimensional array (Array) is stored. Map pre-processing l〇2 (preprocessing), the quality of the image is processed better, in order to separate the graphics and text to carry out β This pre-processing method can be processed by the program written in visuaic ++ language, but not Limited to this. The steps of the pre-processing 102 described above can be achieved by methods currently available on the market. A preferred embodiment according to the present invention is shown in FIG. 2, which includes a binary processing 20, a thinning 2 〇2, and a thinning repair 2 0 3 other steps. Among them, the processing principle of the binary value 201 is to select a critical value based on the statistical graphs of gray scale and number of points. If the gray scale value is greater than this critical value, the point will be changed to a black point. Otherwise, it will be changed to a white point. Three illustrate the experimental results of the two-valued 001, Figure 3A is the original image, and Figure 3B is the two-valued result. Then, the thinned data is thinned to 202. According to this preferred embodiment, the thinning can use the method of Zhang & S ue η, and the result of thinning is obtained by a bonding code (C haincode) Means that this method is based on the coding method shown in Figure 4. All lines are stored not only their starting and ending coordinates, but also their key codes, so that when calculating the area
得到較精確的値。(見於T.Y, Zhang and C.Y. Suen,,,AGet a more accurate value. (See T.Y, Zhang and C.Y. Suen ,,, A
Fast Thinning Algorithm for Thinning Digtal Patterns", Comm. ACM, Vol· 27, pp. 236-239,1984) 但上述細線化會有一些問題,如四叉點變成二個三又 點及產生冗點等,因此需再進行一細線化修補的步骤 2 0 3,該細線化修補2 0 3目的即爲使所有特徵點都可以 正確地表現出來,如此才可以用其座標來準確地計算文 40083.DOC/SLH * 9 ~ 本紙張尺度逋用中国國家標準(CNS ) A4規格(2丨OX 297公釐) (請先閱讀背面之注意事項再填寫本頁)Fast Thinning Algorithm for Thinning Digtal Patterns ", Comm. ACM, Vol. 27, pp. 236-239, 1984) However, the above thinning will have some problems, such as quadrants becoming two three-points and redundant points, etc., Therefore, a thin line repair step 2 0 3 needs to be performed. The purpose of the thin line repair 2 0 3 is to enable all feature points to be represented correctly, so that its coordinates can be used to accurately calculate the text 40083.DOC / SLH * 9 ~ The size of this paper adopts the Chinese National Standard (CNS) A4 specification (2 丨 OX 297mm) (Please read the precautions on the back before filling this page)
3iSS56 A7 〜---------—___________B7 五、發明説明。) ^ — 字的群聚性。 接下來再進仃本發明的是圖文分離步驟103,如圖五 所示,其圖文分離的步驟首先是將影像資料特徵點抽取 出來(5〇1)然後债測各特徵點的群聚性(步驟5〇2),依 …袖測的群聚性質做群聚分類(步驟5 0 3 ),由分類的群聚 性來判斷文字與圖形的部份,並將文字分離出來(步聚 504),最後再修補圖因文字抽離而造成的斷線部分,據 本發明’係先將地圖影像資料以特徵點的方式定義,而 特徵點的定義爲某一黑點,並將各特徵點的型態分成如 圖 π 所示之 degree 叫、degree = 3*degree==4 等三種 型態,當要抽取資料的特徵點時,其判斷的方式是以一 個3 X 3的遮罩(mask)判斷特徵點的型態其中degree的 數目’我們定義爲順時鐘方向丨—O變化的次數。3iSS56 A7 ~ -----------___________ B7 V. Description of the invention. ) ^ — The clustering of words. The next step to enter the invention is the graphic separation step 103. As shown in Figure 5, the graphic separation step is first to extract the feature points of the image data (5〇1) and then measure the clustering of each feature point Character (step 5〇2), according to the clustering property of the arm measurement to do the cluster classification (step 5 0 3), judge the part of the text and graphics by the clustering of the classification, and separate the text (step aggregation 504), and finally repair the broken part of the map caused by the extraction of text. According to the present invention, the map image data is first defined as a feature point, and the feature point is defined as a black dot, and each feature The types of points are divided into three types, such as degree called, degree = 3 * degree == 4, as shown in π. When the feature points of the data are to be extracted, the way to judge them is to use a 3 X 3 mask ( mask) to determine the type of feature points where the number of degrees' is defined as the number of changes in the clockwise direction -O.
至於特徵點的型態是以該特徵點的座標値來作爲計算 群聚性的依據,群聚性比較大的特徵點,我們將其歸納 於同一個群聚(c 1 u s t e r ),待所有的特徵點都完成群聚 的指派後,吾人即可簡單地利用群聚密度,來判斷某群 聚是否爲文字的群聚或是圖形的群聚Q 本發明之群聚偵測步驟5 〇 2可以利用最大一最小距離 經濟部中央標準局員工消費合作社印製 f請先聞讀背面之注意事項再填寫本頁) 群聚法則(Maximum Minimum Distance ClusteringAs for the type of the feature point, the coordinate value of the feature point is used as the basis for calculating the clustering. For the feature points with larger clustering, we will summarize them to the same cluster (c 1 uster). After all the feature points have been assigned, we can simply use the cluster density to determine whether a cluster is a text cluster or a graphic cluster. Q The cluster detection step 5 of this invention can Printed by the Employee Consumer Cooperative of the Central Standards Bureau of the Ministry of Economic Affairs. Please read the precautions on the back and then fill out this page.) Group Minimum Rule (Maximum Minimum Distance Clustering
Algorithm)來進行,正因爲文字在地囷上具備了以下之 特徵: (1) 所有的中文字在地圖上彼此有一間隔存在。 (2) 中文字經過特徵點之擷取後,發現中文字是一個"彼此 聚密結合’’的特徵點所形成之集合,如圖七所示。 因此吾人可以簡單地應用最大—最小距離群聚法則來 40083.DQC/SLH - 10 _ 本ϋ尺度4财"轉(CNS) A4規格(21QX別公楚) A7 B7 經濟部中央標準局員工消費合作社印製 五、發明説明(8 ) 計算任二特徵點之間距,允 取/ 遂冋時把特徵點予以歸類成群 聚(見於 J.T, Tou and R r r * Gonzalez, Pattern RecognitionAlgorithm), because the text has the following characteristics on the wall: (1) All Chinese characters exist at a distance from each other on the map. (2) After extracting the feature points, the Chinese characters are found to be a collection of feature points that are " combined with each other ’, as shown in Figure 7. Therefore, we can simply apply the maximum-minimum distance clustering rule to 40083. DQC / SLH-10 _ this standard 4 financial " transfer (CNS) A4 specifications (21QX Biechuchu) A7 B7 Ministry of Economic Affairs Central Standards Bureau staff consumption Printed by the cooperative. 5. Description of the invention (8) Calculate the distance between any two feature points, and classify the feature points into clusters when allowed / succeeded (see JT, Tou and R rr * Gonzalez, Pattern Recognition
Principles, Addison-Wesley Publishing Company. 1981) 〇 孩群聚法則的步驟簡單描述如下: ⑴任意地選取某-特㈣作爲群聚巾心c丨川“ Center)Z0。 ⑵在除了 ZG〇卜的所有特徵點巾選擇與^相距最遠者作 爲〇 ⑶對於所有特徵點,計算其與財㈣中心之距離。 ⑷在⑶所求得之距離中,找到一最小値MiNS。 ⑸然後在所有最小値中找到—最大値。 ⑹若此最大値MAXS>Z〇Zi/群聚半徑,即產生—新群聚 中心,並重複⑶。 ⑺把所有點歸類到其所屬之群聚。 在所有特徵點都完成歸類後,可得到數個群聚,接著做 群聚分類5 0 3,若某群聚中之特徵點個數大於一”群聚密 度臨界値(Cluster Density Threshold)",則該群聚被 歸爲文字群聚,否則爲一線條圖形之群聚,由於本發明是 以特徵點的觀點來考量文字的群聚性,因此不論文字的大 小或是旋轉、坐落的位置、圖文黏或圖文重疊等文件常見 的問題,都可輕易地被解決。Principles, Addison-Wesley Publishing Company. 1981) ○ The steps of the child grouping rule are briefly described as follows: (1) Choose a special feature arbitrarily as the grouping center (center) Z0. (2) In all but ZG〇 卜The feature point towel is the one farthest away from ^ as the ⑶. For all feature points, calculate the distance from the center of the finance. ⑷In the distance obtained by ⑶, find a minimum value MiNS. ⑸ Then among all minimum values Find—the maximum value. ⑹If the maximum value is MAXS> Z〇Zi / cluster radius, a new cluster center is generated, and repeat ⑶. ⑺Classify all points to the cluster to which they belong. At all feature points After the classification is completed, several clusters can be obtained, and then cluster classification 5 0 3, if the number of feature points in a cluster is greater than one "Cluster Density Threshold" (Cluster Density Threshold) ", then the cluster Clustering is classified as text clustering, otherwise it is a cluster of line graphics. Since the present invention considers the clustering of text from the perspective of feature points, regardless of the size or rotation of the text, the location of the placement, and the stickiness of the text Or text and text Common problems can be solved easily.
在群聚歸類的步驟中有3個參數直接影響了歸類的結果 ’第一個是選取起初群聚(Initial ClusteOZoiZ}。其 次是每一個群聚半徑的大小。最後是每一個群聚中特徵點 40083.00C/SLH 本纸張尺度適用中國國家榡準(CNS ) A4規格(2I〇X297公嫠) (請先鬩讀背面之注意事項再填寫本頁)In the clustering step, there are 3 parameters that directly affect the result of the classification. The first is to select the initial cluster (Initial ClusteOZoiZ). The second is the size of each cluster radius. The last is each cluster. Feature point 40083.00C / SLH This paper scale is applicable to China National Standard (CNS) A4 specification (2I〇X297 public daughter) (please read the precautions on the back before filling this page)
319356 A7 B7 '-------- 經濟部中央榡準局員工消費合作,社印製 五、發明説明(9 的數目(即群聚密度臨界値) 通常起初群聚中心是選取任一點作爲^,然後在其餘點 中選取與20相距最遠之點爲21,不過在實作中,這樣的 z〇,^選法並不一定能保證%2丨是最遠距,可能存在二 點X iX j又距大於z 〇 z i,使得歸類會有所差錯,即多產生 一些不必要的群聚中心,因爲群聚半徑之値會因z〇Zi之 値而變,故本發明:爲確保能產生最少的群聚數目,^及 Z 1之選取是針對所有特徵點中距離最遠的二點爲起初群 聚中心。 第二個參數是群聚半徑的選取,若太大,則許多線條的 特徵4將會被知類爲文字群聚,倘若恰好有一條線段的二 端點都被歸爲文字群聚,則該線段會被遺落,使得圖文分 離有誤,圖八説明此情況β 夕反之若群聚半徑太小,則一個中文字可能會被細分爲許 夕小群聚,而特徵點數目太少的群聚會被誤認爲線條圖形 、’ι由實作測結果,此値以最小字型的平均高度最爲合 適。 s第三個參數是群聚中的特徵點數目多少以上才能被歸納 爲文字?原則上文字群聚中特徵點個數只要大於丨即可, 反之,圖形群聚則應等於1,不過爲避免人爲的疏失,或 是細小狹長的線段所產.生的雜訊,本發明之較佳實施例係 將群聚密度臨界値設爲3,如此意謂著若某群聚中特徵點 數目大於3,則該群聚爲文字,否則爲線條圖形。 仁疋由於所選擇的群聚半徑係設爲最小字型的平均高度 _i°°e3.D0C/SLH _ 12 - 國國家榡率(CNS) A4規格(2】0><29祕嫠)----- (請先閲讀背面之:>i意事項再填寫本頁)319356 A7 B7 '-------- Employee consumption cooperation of the Central Bureau of Economics of the Ministry of Economic Affairs, printed by the company V. Description of invention (the number of 9 (that is, the critical value of cluster density) Usually, at the beginning, the cluster center is selected at any point As ^, then select the point farthest from 20 among the remaining points as 21. However, in practice, such z〇, ^ selection method does not necessarily guarantee that% 2 丨 is the farthest distance, there may be two points X iX j is greater than z 〇zi, so that the classification will be wrong, that is, some unnecessary cluster centers will be generated, because the value of the cluster radius will change due to the value of z〇Zi, so the present invention: Ensure that the minimum number of clusters can be generated. The selection of ^ and Z 1 is based on the two clusters with the farthest distance among all feature points as the initial cluster center. The second parameter is the selection of the cluster radius, if too large, many The feature 4 of the line will be known as a text cluster. If there are exactly two ends of a line segment that are classified as a text cluster, the line segment will be left behind, making the image and text separation incorrect. Figure 8 illustrates this Case β Xi Conversely, if the cluster radius is too small, a Chinese character may be subdivided into Small clusters, and cluster gatherings with too few feature points are mistaken for line graphics, and the actual measurement results, this value is the most appropriate for the average height of the smallest font. S The third parameter is in the cluster How many feature points can be summed up as text? In principle, the number of feature points in the text cluster should be greater than 丨, otherwise, the graphic cluster should be equal to 1, but to avoid artificial negligence, or small and long For the noise generated by the line segment, the preferred embodiment of the present invention sets the critical value of the cluster density to 3, which means that if the number of feature points in a cluster is greater than 3, the cluster will be text, otherwise It is a line graphic. Renzhang because the selected cluster radius is set to the average height of the smallest font _i °° e3.D0C / SLH _ 12-National National Rate (CNS) A4 specification (2) 0 > < 29 因 嫠) ----- (please read the back page first: > i matters, then fill out this page)
經濟部中夬標準局員工消費合作社印製 A 7 ______B7_ 五、發明説明(10 ) ,因此很有可能使得幾個小字被組合成一個大集合,因之 上述方法並無法保證每一個圖文分離後的文字群聚恰好代 表一個字,因此有必要進一步做文字的切割,本發明所採 用的方法是利用鄰近擴散(Neighbor-Growing)方法, 一邊找8鄰點,同時更新該文字的區塊大小,最後用一矩 形框將此文字框出。但此方法僅爲説明之例,並非用以限 制本發明。 完成文字抽離504後,所有文字群聚都會被去除,而留 下線條圖形群聚,現在將這些群聚中的特徵點連起來,圖 文分離才眞正地完成,所採用的方式是分支追蹤,以確保 各線段的存在性。本發明之分枝追蹤(Branch Tracking)是用雙向追蹤的方式,如圖九中,線段ab將 由A — B追縱一次,反向B — A再被追縱一次,這樣做的目 的是爲了確保A,B中任一點被歸爲文字時,線段AB不至 於消失。 然而,當有”圖文黏接"或”圖文重疊”情形發生時,經過 圖文分離之後勢必有線段被打斷,因此本發明尚須做線段 修補5 0 5的工作,傳統的修補線段方法是將滿足下列所有 3條件的兩線段連起來:(1 )兩線條有相同斜率;(2 )兩斷 線都有一端點degree=l ; (3)兩degree=l之端點,其距 離小於某一臨界値。但是這樣的作法,在圖文重疊發生在 狹長的線條上時,會發生錯誤,如圖十所示,當A B被一 較大字所截斷,而C D被較小的字所截斷,因此就a £而言 ,CG' FB、HD都有相同斜率,且都有一端點其 40083.DOC/SLH - 13 - 本紙张尺度適用中國國家標準(CNS ) A4说格(2丨0X297公釐)' ' (請先閲讀背面之注意事項再填寫本頁)A 7 ______B7_ printed by the employee consumer cooperative of the Central Bureau of Standards and Economics of the Ministry of Economic Affairs 5. Invention description (10), so it is possible that several small characters are combined into a large collection, so the above method cannot guarantee that each picture and text is separated. The text clusters just represent a word, so it is necessary to further cut the text. The method used in the present invention is to use the neighbor-growing method to find 8 neighbors while updating the block size of the text. Finally, use a rectangular frame to frame this text. However, this method is only an illustrative example and is not intended to limit the present invention. After completing the text extraction 504, all text clusters will be removed, leaving the line and graphic clusters. Now the feature points in these clusters are connected, and the graphic and text separation is completed successfully. The method used is branching. Trace to ensure the existence of each line segment. The branch tracking of the present invention uses a two-way tracking method. As shown in Figure 9, line segment ab will be traced once by A-B, and reverse B-A will be traced once again. The purpose of this is to ensure that When any point of A and B is classified as text, line segment AB will not disappear. However, when "picture and text stick" or "picture and text overlap" occurs, the wire segment is bound to be interrupted after the picture and text are separated, so the present invention still needs to do the work of line segment repair 5 0 5, traditional repair The line segment method is to connect two line segments that meet all the following 3 conditions: (1) the two lines have the same slope; (2) the two broken lines have an endpoint degree = l; (3) the endpoint of two degree = l, which The distance is less than a certain critical value. However, in this way, when the overlapping of pictures and text occurs on a long and narrow line, an error will occur, as shown in Figure 10, when AB is truncated by a larger character and CD is by a smaller character Truncated, so as far as a £ is concerned, CG 'FB and HD all have the same slope and have one end point. 40083.DOC / SLH-13-This paper scale is applicable to the Chinese National Standard (CNS) A4 said grid (2 丨0X297mm) '' (Please read the notes on the back before filling this page)
*1T 1 A7 B7 五、發明説明(11 ) 8 I,但因d 3最小,所以E G會被連起來,爲了避 免這種錯誤,本發明所採用的修補法則如下: (請先閱讀背面之注意事項再填寫本頁) (1) 以AE向量爲基準; (2) 求AE各與EF、M、EG之内積,並求出其爽角; (3 )夾角最+的,該向量即爲修補的對象。 經由上述特徵點抽取、群聚偵測、群聚分類以及文字抽 離和修補等步驟,厂地圖影像資料得以精確地解析出來, 由於本發明首創以「特徵點」方式定義影像資料,且依各 特徵點的群聚型態分類爲文字或是圖形,故而能有效地使 文字自地圖影像中分離出來,且圖文黏接或圖文重疊之部 分亦能有效地予以解析。本發明因此能提供一種精確有效 的圖文分離方法。 上述實施例之説明係用以解釋本發明之原理,並非用以 限制本發明之範_ ’因此’本發明尚可以其他方法予以修 正,只要其基本概念相同,皆不達背離本發明之精神。本 發明之專利範圍應如後列之申請專利範圍所限定。 經濟部中央標準局員工消費合作社印製 實驗結果 整個圖文分離的系統是在P C 3 8 6相容電腦環境下完成的 ’所有的實驗圖都是1024x768 Pixels,如此是爲了方 便利用視訊記憶體(V i d e ο M e m ο n y)作爲緩衝器 (Buffer)。本方法測試二十幾張影像。結果都相當成功, 這裏我們舉一個實例來介紹β 第一個實驗測試一張地籍圖,如圖十一(a )所示,圖十 A0083.DOC/SLH ___ U _ 本紙張尺度適用中國國家標準(CNS ) A4規格(210X297公釐) A7 A7 經濟部中央標孪局員工消費合作社印裝 五、發明説明(U ) -⑻則爲細線化的結果’圖十一⑷爲細線化的修 ’圖十一⑷爲所有特徵點賴取,圖卜⑷爲圖^ 的結果,圖十一(f)爲斷線的修補結果,其中圖文已成功 分離開,而且圖文重—叠部分的文字也分離開,線條也正: 地被修補。 本發明之優點可略述如下: (1) 計算簡單’可以很容易做成硬體,加在掃描器中,不 需增加大量額外成本。 (2) 利用文字特徵點群聚的特性,可以分離圖形與文字, 即使文字太小不同,文字有旋轉,圖文黏接或圖文重 疊,均可有效解決。 (3) 圖文重疊之文字被分離後,圖形被抽掉的部分,利用 方向性與夾角的特性,可將線條圖形補好β* 1T 1 A7 B7 Fifth, the description of the invention (11) 8 I, but because d 3 is the smallest, EG will be connected, in order to avoid this error, the repair method adopted by the present invention is as follows: (please read the notes on the back first Please fill in this page again) (1) Take the AE vector as the benchmark; (2) Find the inner product of AE and EF, M, EG, and find its cool angle; (3) The angle with the most +, the vector is the repair Object. Through the above-mentioned feature point extraction, cluster detection, cluster classification, and text extraction and repair steps, the factory map image data can be accurately parsed out. Because the invention first defines image data in a "feature point" manner, and according to each The clustering of feature points is classified as text or graphics, so it can effectively separate the text from the map image, and the part where the graphics and text are glued or overlapped can also be effectively analyzed. The present invention can thus provide an accurate and effective method for image and text separation. The descriptions of the above embodiments are used to explain the principle of the present invention, not to limit the scope of the present invention. Therefore, the present invention can be modified by other methods. As long as the basic concepts are the same, they will not depart from the spirit of the present invention. The patent scope of this invention shall be as defined in the patent application scope listed below. The experiment results printed by the Employee Consumer Cooperative of the Central Bureau of Standards of the Ministry of Economic Affairs. The entire system of image and text separation was completed in a PC 3 8 6 compatible computer environment. All experimental images are 1024x768 Pixels. This is to facilitate the use of video memory ( V ide ο M em ο ny) as a buffer (Buffer). This method tests more than 20 images. The results are quite successful, here we give an example to introduce β The first experiment tests a cadastral map, as shown in Figure 11 (a), Figure 10 A0083.DOC / SLH ___ U _ This paper scale is applicable to Chinese national standards (CNS) A4 specification (210X297mm) A7 A7 Printed by the employee consumer cooperative of the Central Standardization Bureau of the Ministry of Economic Affairs V. Invention description (U)-⑻ is the result of thinning the line "Figure 11 ⑷ is the repairing of the thin line" Eleven ⑷ is for all feature points, Figure bu ⑷ is the result of Figure ^, Figure eleven (f) is the repair result of broken line, in which the picture and text have been successfully separated, and the text is repeated-the text in the overlapping part is also Divided away, the line is: the ground is repaired. The advantages of the present invention can be briefly described as follows: (1) Simple calculation can be easily made into hardware and added to the scanner without adding a lot of extra costs. (2) Using the characteristics of the clustering of character feature points, graphics and text can be separated. Even if the text is too small and different, the text is rotated, and the text is glued or overlapped, which can be effectively solved. (3) After the text with overlapping text is separated, the part where the graphic is removed can use the characteristics of directivity and angle to make up the line graphic β
40083.0OC/SLH 本紙張尺度適用中國國家榡準(CNS)A4規格(2丨ΟΧ297公嫠) (請先閱讀背面之注意事項再填寫本頁)40083.0OC / SLH This paper scale is applicable to China National Standard (CNS) A4 (2 丨 Ο297297) (Please read the precautions on the back before filling this page)
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW84107775A TW319856B (en) | 1995-07-27 | 1995-07-27 | Graph-text segmentation method of geographical image data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW84107775A TW319856B (en) | 1995-07-27 | 1995-07-27 | Graph-text segmentation method of geographical image data |
Publications (1)
Publication Number | Publication Date |
---|---|
TW319856B true TW319856B (en) | 1997-11-11 |
Family
ID=51566983
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
TW84107775A TW319856B (en) | 1995-07-27 | 1995-07-27 | Graph-text segmentation method of geographical image data |
Country Status (1)
Country | Link |
---|---|
TW (1) | TW319856B (en) |
-
1995
- 1995-07-27 TW TW84107775A patent/TW319856B/en active
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100248917B1 (en) | Pattern recognizing apparatus and method | |
US8000529B2 (en) | System and method for creating an editable template from a document image | |
US5335290A (en) | Segmentation of text, picture and lines of a document image | |
AU2006252025B2 (en) | Recognition of parameterised shapes from document images | |
CA2598686C (en) | Intelligent importation of information from foreign application user interface using artificial intelligence | |
AU2006252019B2 (en) | Method and Apparatus for Dynamic Connector Analysis | |
WO2021143059A1 (en) | Method, apparatus, and device for determining map area, and storage medium | |
US8306325B2 (en) | Text character identification system and method thereof | |
CN109344820B (en) | Digital ammeter reading identification method based on computer vision and deep learning | |
CN112529004A (en) | Intelligent image recognition method and device, computer equipment and storage medium | |
CN112364834A (en) | Form identification restoration method based on deep learning and image processing | |
CN111626145B (en) | Simple and effective incomplete form identification and page-crossing splicing method | |
CN111753706B (en) | Complex table intersection point clustering extraction method based on image statistics | |
WO2021159802A1 (en) | Graphical captcha recognition method, apparatus, computer device, and storage medium | |
CN114694165A (en) | Intelligent PID drawing identification and redrawing method | |
US6360006B1 (en) | Color block selection | |
CN115273108A (en) | Artificial intelligence recognition automatic collection method and system | |
CN110647824A (en) | Value-added tax invoice layout extraction method based on computer vision technology | |
CN114495141A (en) | Document paragraph position extraction method, electronic equipment and storage medium | |
CN113361547A (en) | Signature identification method, device, equipment and readable storage medium | |
CN113112567A (en) | Method and device for generating editable flow chart, electronic equipment and storage medium | |
CN113177542A (en) | Method, device and equipment for identifying characters of seal and computer readable medium | |
Della Ventura et al. | Graphic symbol recognition using a signature technique | |
TW319856B (en) | Graph-text segmentation method of geographical image data | |
CN111488870A (en) | Character recognition method and character recognition device |